{ "best_metric": null, "best_model_checkpoint": null, "epoch": 3.0, "eval_steps": 500, "global_step": 2001, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0014992503748125937, "grad_norm": 0.06610345176976985, "learning_rate": 9.950248756218907e-07, "loss": 0.0231, "step": 1 }, { "epoch": 0.0029985007496251873, "grad_norm": 0.10229378949313654, "learning_rate": 1.9900497512437813e-06, "loss": 0.0302, "step": 2 }, { "epoch": 0.004497751124437781, "grad_norm": 0.09308353085324574, "learning_rate": 2.9850746268656716e-06, "loss": 0.0501, "step": 3 }, { "epoch": 0.005997001499250375, "grad_norm": 0.058628771398694095, "learning_rate": 3.980099502487563e-06, "loss": 0.0243, "step": 4 }, { "epoch": 0.0074962518740629685, "grad_norm": 0.0910650644048991, "learning_rate": 4.975124378109453e-06, "loss": 0.0383, "step": 5 }, { "epoch": 0.008995502248875561, "grad_norm": 0.12316753445017627, "learning_rate": 5.970149253731343e-06, "loss": 0.0436, "step": 6 }, { "epoch": 0.010494752623688156, "grad_norm": 0.05542672781465358, "learning_rate": 6.965174129353234e-06, "loss": 0.094, "step": 7 }, { "epoch": 0.01199400299850075, "grad_norm": 0.03735550944126593, "learning_rate": 7.960199004975125e-06, "loss": 0.0241, "step": 8 }, { "epoch": 0.013493253373313344, "grad_norm": 0.055574007146215375, "learning_rate": 8.955223880597016e-06, "loss": 0.0199, "step": 9 }, { "epoch": 0.014992503748125937, "grad_norm": 0.08334031247034454, "learning_rate": 9.950248756218906e-06, "loss": 0.0309, "step": 10 }, { "epoch": 0.01649175412293853, "grad_norm": 0.0615936672450276, "learning_rate": 1.0945273631840796e-05, "loss": 0.0135, "step": 11 }, { "epoch": 0.017991004497751123, "grad_norm": 0.04494010236736683, "learning_rate": 1.1940298507462686e-05, "loss": 0.0232, "step": 12 }, { "epoch": 0.019490254872563718, "grad_norm": 0.05796888012935532, "learning_rate": 1.2935323383084577e-05, "loss": 0.0349, "step": 13 }, { "epoch": 0.020989505247376312, "grad_norm": 0.06339185703613222, "learning_rate": 1.3930348258706468e-05, "loss": 0.0309, "step": 14 }, { "epoch": 0.022488755622188907, "grad_norm": 0.0592636851918813, "learning_rate": 1.4925373134328357e-05, "loss": 0.0256, "step": 15 }, { "epoch": 0.0239880059970015, "grad_norm": 0.060507862352465004, "learning_rate": 1.592039800995025e-05, "loss": 0.0253, "step": 16 }, { "epoch": 0.025487256371814093, "grad_norm": 0.06232454162862092, "learning_rate": 1.691542288557214e-05, "loss": 0.0341, "step": 17 }, { "epoch": 0.026986506746626688, "grad_norm": 0.08813118060803964, "learning_rate": 1.791044776119403e-05, "loss": 0.0191, "step": 18 }, { "epoch": 0.02848575712143928, "grad_norm": 0.09703603861107045, "learning_rate": 1.890547263681592e-05, "loss": 0.0334, "step": 19 }, { "epoch": 0.029985007496251874, "grad_norm": 0.12217053930995624, "learning_rate": 1.990049751243781e-05, "loss": 0.025, "step": 20 }, { "epoch": 0.031484257871064465, "grad_norm": 0.08708398557831935, "learning_rate": 2.0895522388059702e-05, "loss": 0.0385, "step": 21 }, { "epoch": 0.03298350824587706, "grad_norm": 0.05748291151872297, "learning_rate": 2.1890547263681592e-05, "loss": 0.0294, "step": 22 }, { "epoch": 0.034482758620689655, "grad_norm": 0.05267747048254378, "learning_rate": 2.2885572139303486e-05, "loss": 0.0236, "step": 23 }, { "epoch": 0.035982008995502246, "grad_norm": 0.20462262185187904, "learning_rate": 2.3880597014925373e-05, "loss": 0.0298, "step": 24 }, { "epoch": 0.037481259370314844, "grad_norm": 0.04730801516528991, "learning_rate": 2.4875621890547266e-05, "loss": 0.0231, "step": 25 }, { "epoch": 0.038980509745127435, "grad_norm": 0.07129877545543531, "learning_rate": 2.5870646766169153e-05, "loss": 0.0409, "step": 26 }, { "epoch": 0.04047976011994003, "grad_norm": 0.04787844989660117, "learning_rate": 2.6865671641791047e-05, "loss": 0.0244, "step": 27 }, { "epoch": 0.041979010494752625, "grad_norm": 0.08080289031605667, "learning_rate": 2.7860696517412937e-05, "loss": 0.0373, "step": 28 }, { "epoch": 0.043478260869565216, "grad_norm": 0.03127750279466034, "learning_rate": 2.885572139303483e-05, "loss": 0.0093, "step": 29 }, { "epoch": 0.044977511244377814, "grad_norm": 0.0672839069697597, "learning_rate": 2.9850746268656714e-05, "loss": 0.0565, "step": 30 }, { "epoch": 0.046476761619190406, "grad_norm": 0.05579809699762286, "learning_rate": 3.084577114427861e-05, "loss": 0.0262, "step": 31 }, { "epoch": 0.047976011994003, "grad_norm": 0.06764953701820348, "learning_rate": 3.18407960199005e-05, "loss": 0.0355, "step": 32 }, { "epoch": 0.049475262368815595, "grad_norm": 0.07766895860290135, "learning_rate": 3.283582089552239e-05, "loss": 0.0176, "step": 33 }, { "epoch": 0.050974512743628186, "grad_norm": 0.0877687067355835, "learning_rate": 3.383084577114428e-05, "loss": 0.0228, "step": 34 }, { "epoch": 0.05247376311844078, "grad_norm": 0.06901704426078333, "learning_rate": 3.4825870646766175e-05, "loss": 0.0301, "step": 35 }, { "epoch": 0.053973013493253376, "grad_norm": 0.050365449371918836, "learning_rate": 3.582089552238806e-05, "loss": 0.025, "step": 36 }, { "epoch": 0.05547226386806597, "grad_norm": 0.10183750100826967, "learning_rate": 3.681592039800995e-05, "loss": 0.0359, "step": 37 }, { "epoch": 0.05697151424287856, "grad_norm": 0.05576688463550469, "learning_rate": 3.781094527363184e-05, "loss": 0.0263, "step": 38 }, { "epoch": 0.05847076461769116, "grad_norm": 0.0809406763945191, "learning_rate": 3.8805970149253736e-05, "loss": 0.0446, "step": 39 }, { "epoch": 0.05997001499250375, "grad_norm": 0.03977873137524195, "learning_rate": 3.980099502487562e-05, "loss": 0.0155, "step": 40 }, { "epoch": 0.06146926536731634, "grad_norm": 0.04675680284402419, "learning_rate": 4.079601990049751e-05, "loss": 0.0213, "step": 41 }, { "epoch": 0.06296851574212893, "grad_norm": 0.13250207008437154, "learning_rate": 4.1791044776119404e-05, "loss": 0.0225, "step": 42 }, { "epoch": 0.06446776611694154, "grad_norm": 0.06061033398983015, "learning_rate": 4.27860696517413e-05, "loss": 0.0262, "step": 43 }, { "epoch": 0.06596701649175413, "grad_norm": 0.08470633670479133, "learning_rate": 4.3781094527363184e-05, "loss": 0.0403, "step": 44 }, { "epoch": 0.06746626686656672, "grad_norm": 0.06220515053666017, "learning_rate": 4.477611940298508e-05, "loss": 0.0256, "step": 45 }, { "epoch": 0.06896551724137931, "grad_norm": 0.0421105165497031, "learning_rate": 4.577114427860697e-05, "loss": 0.0233, "step": 46 }, { "epoch": 0.0704647676161919, "grad_norm": 0.056273635232726445, "learning_rate": 4.676616915422886e-05, "loss": 0.0271, "step": 47 }, { "epoch": 0.07196401799100449, "grad_norm": 0.04135843131632209, "learning_rate": 4.7761194029850745e-05, "loss": 0.0192, "step": 48 }, { "epoch": 0.0734632683658171, "grad_norm": 0.07545305238459443, "learning_rate": 4.875621890547264e-05, "loss": 0.0354, "step": 49 }, { "epoch": 0.07496251874062969, "grad_norm": 0.03485747229470634, "learning_rate": 4.975124378109453e-05, "loss": 0.0194, "step": 50 }, { "epoch": 0.07646176911544228, "grad_norm": 0.07132696830734893, "learning_rate": 5.074626865671642e-05, "loss": 0.0284, "step": 51 }, { "epoch": 0.07796101949025487, "grad_norm": 0.0476562529060188, "learning_rate": 5.1741293532338306e-05, "loss": 0.0195, "step": 52 }, { "epoch": 0.07946026986506746, "grad_norm": 0.04135610651103468, "learning_rate": 5.2736318407960206e-05, "loss": 0.0209, "step": 53 }, { "epoch": 0.08095952023988005, "grad_norm": 0.07137198475738459, "learning_rate": 5.373134328358209e-05, "loss": 0.0235, "step": 54 }, { "epoch": 0.08245877061469266, "grad_norm": 0.08351208018844225, "learning_rate": 5.472636815920398e-05, "loss": 0.0456, "step": 55 }, { "epoch": 0.08395802098950525, "grad_norm": 0.0855628538810367, "learning_rate": 5.5721393034825874e-05, "loss": 0.0383, "step": 56 }, { "epoch": 0.08545727136431784, "grad_norm": 0.07872354423296858, "learning_rate": 5.671641791044776e-05, "loss": 0.0346, "step": 57 }, { "epoch": 0.08695652173913043, "grad_norm": 0.08549249098509387, "learning_rate": 5.771144278606966e-05, "loss": 0.0436, "step": 58 }, { "epoch": 0.08845577211394302, "grad_norm": 0.08010714092690198, "learning_rate": 5.870646766169154e-05, "loss": 0.0278, "step": 59 }, { "epoch": 0.08995502248875563, "grad_norm": 0.11061139734613264, "learning_rate": 5.970149253731343e-05, "loss": 0.0303, "step": 60 }, { "epoch": 0.09145427286356822, "grad_norm": 0.08851009372866617, "learning_rate": 6.069651741293533e-05, "loss": 0.0273, "step": 61 }, { "epoch": 0.09295352323838081, "grad_norm": 0.06431070626891758, "learning_rate": 6.169154228855722e-05, "loss": 0.0476, "step": 62 }, { "epoch": 0.0944527736131934, "grad_norm": 0.04207690654899365, "learning_rate": 6.268656716417911e-05, "loss": 0.0187, "step": 63 }, { "epoch": 0.095952023988006, "grad_norm": 0.0629659489389091, "learning_rate": 6.3681592039801e-05, "loss": 0.0294, "step": 64 }, { "epoch": 0.09745127436281859, "grad_norm": 0.04739055753217489, "learning_rate": 6.46766169154229e-05, "loss": 0.0117, "step": 65 }, { "epoch": 0.09895052473763119, "grad_norm": 0.062751616457929, "learning_rate": 6.567164179104478e-05, "loss": 0.0321, "step": 66 }, { "epoch": 0.10044977511244378, "grad_norm": 0.12789865648617366, "learning_rate": 6.666666666666667e-05, "loss": 0.0285, "step": 67 }, { "epoch": 0.10194902548725637, "grad_norm": 0.09087938435304468, "learning_rate": 6.766169154228856e-05, "loss": 0.0293, "step": 68 }, { "epoch": 0.10344827586206896, "grad_norm": 0.07104633745977648, "learning_rate": 6.865671641791044e-05, "loss": 0.0248, "step": 69 }, { "epoch": 0.10494752623688156, "grad_norm": 0.04104044865025192, "learning_rate": 6.965174129353235e-05, "loss": 0.0117, "step": 70 }, { "epoch": 0.10644677661169415, "grad_norm": 0.06739350707830459, "learning_rate": 7.064676616915423e-05, "loss": 0.0259, "step": 71 }, { "epoch": 0.10794602698650675, "grad_norm": 0.12081546856137051, "learning_rate": 7.164179104477612e-05, "loss": 0.046, "step": 72 }, { "epoch": 0.10944527736131934, "grad_norm": 0.09650929672212062, "learning_rate": 7.263681592039802e-05, "loss": 0.0196, "step": 73 }, { "epoch": 0.11094452773613193, "grad_norm": 0.03209119308258964, "learning_rate": 7.36318407960199e-05, "loss": 0.0096, "step": 74 }, { "epoch": 0.11244377811094453, "grad_norm": 0.12687784195284804, "learning_rate": 7.46268656716418e-05, "loss": 0.0416, "step": 75 }, { "epoch": 0.11394302848575712, "grad_norm": 0.06383164911311603, "learning_rate": 7.562189054726369e-05, "loss": 0.0122, "step": 76 }, { "epoch": 0.11544227886056972, "grad_norm": 0.07876299012336277, "learning_rate": 7.661691542288557e-05, "loss": 0.0299, "step": 77 }, { "epoch": 0.11694152923538231, "grad_norm": 0.07838431599591475, "learning_rate": 7.761194029850747e-05, "loss": 0.0272, "step": 78 }, { "epoch": 0.1184407796101949, "grad_norm": 0.06305227657672241, "learning_rate": 7.860696517412935e-05, "loss": 0.0317, "step": 79 }, { "epoch": 0.1199400299850075, "grad_norm": 0.035442799985056346, "learning_rate": 7.960199004975125e-05, "loss": 0.015, "step": 80 }, { "epoch": 0.12143928035982009, "grad_norm": 0.05183095808072367, "learning_rate": 8.059701492537314e-05, "loss": 0.0242, "step": 81 }, { "epoch": 0.12293853073463268, "grad_norm": 0.21288021373308544, "learning_rate": 8.159203980099502e-05, "loss": 0.0657, "step": 82 }, { "epoch": 0.12443778110944528, "grad_norm": 0.14499038539812795, "learning_rate": 8.258706467661693e-05, "loss": 0.0539, "step": 83 }, { "epoch": 0.12593703148425786, "grad_norm": 0.07506130007270491, "learning_rate": 8.358208955223881e-05, "loss": 0.0357, "step": 84 }, { "epoch": 0.12743628185907047, "grad_norm": 0.06540286206867715, "learning_rate": 8.45771144278607e-05, "loss": 0.0249, "step": 85 }, { "epoch": 0.12893553223388307, "grad_norm": 0.046324410553248214, "learning_rate": 8.55721393034826e-05, "loss": 0.0211, "step": 86 }, { "epoch": 0.13043478260869565, "grad_norm": 0.04910635273005273, "learning_rate": 8.656716417910447e-05, "loss": 0.0174, "step": 87 }, { "epoch": 0.13193403298350825, "grad_norm": 0.06542227021586179, "learning_rate": 8.756218905472637e-05, "loss": 0.0306, "step": 88 }, { "epoch": 0.13343328335832083, "grad_norm": 0.04867631138137704, "learning_rate": 8.855721393034826e-05, "loss": 0.0124, "step": 89 }, { "epoch": 0.13493253373313344, "grad_norm": 0.03955791644617872, "learning_rate": 8.955223880597016e-05, "loss": 0.02, "step": 90 }, { "epoch": 0.136431784107946, "grad_norm": 0.07757595899547425, "learning_rate": 9.054726368159205e-05, "loss": 0.0349, "step": 91 }, { "epoch": 0.13793103448275862, "grad_norm": 0.05833106304874586, "learning_rate": 9.154228855721394e-05, "loss": 0.0215, "step": 92 }, { "epoch": 0.13943028485757122, "grad_norm": 0.07191436633738099, "learning_rate": 9.253731343283582e-05, "loss": 0.0351, "step": 93 }, { "epoch": 0.1409295352323838, "grad_norm": 0.074329975473276, "learning_rate": 9.353233830845772e-05, "loss": 0.0334, "step": 94 }, { "epoch": 0.1424287856071964, "grad_norm": 0.11507286561615733, "learning_rate": 9.452736318407961e-05, "loss": 0.039, "step": 95 }, { "epoch": 0.14392803598200898, "grad_norm": 0.0500122604486147, "learning_rate": 9.552238805970149e-05, "loss": 0.027, "step": 96 }, { "epoch": 0.1454272863568216, "grad_norm": 0.11474477698178565, "learning_rate": 9.65174129353234e-05, "loss": 0.0751, "step": 97 }, { "epoch": 0.1469265367316342, "grad_norm": 0.13909793248506105, "learning_rate": 9.751243781094528e-05, "loss": 0.0397, "step": 98 }, { "epoch": 0.14842578710644677, "grad_norm": 0.06883924844168375, "learning_rate": 9.850746268656717e-05, "loss": 0.0356, "step": 99 }, { "epoch": 0.14992503748125938, "grad_norm": 0.03603265842985656, "learning_rate": 9.950248756218906e-05, "loss": 0.0132, "step": 100 }, { "epoch": 0.15142428785607195, "grad_norm": 0.03810266042373329, "learning_rate": 0.00010049751243781096, "loss": 0.0142, "step": 101 }, { "epoch": 0.15292353823088456, "grad_norm": 0.053717895257444505, "learning_rate": 0.00010149253731343284, "loss": 0.0277, "step": 102 }, { "epoch": 0.15442278860569716, "grad_norm": 0.04818876674919931, "learning_rate": 0.00010248756218905473, "loss": 0.0195, "step": 103 }, { "epoch": 0.15592203898050974, "grad_norm": 0.10941924254236936, "learning_rate": 0.00010348258706467661, "loss": 0.0384, "step": 104 }, { "epoch": 0.15742128935532235, "grad_norm": 0.07867930264182815, "learning_rate": 0.0001044776119402985, "loss": 0.0162, "step": 105 }, { "epoch": 0.15892053973013492, "grad_norm": 0.08345512376505489, "learning_rate": 0.00010547263681592041, "loss": 0.0328, "step": 106 }, { "epoch": 0.16041979010494753, "grad_norm": 0.09189672805052056, "learning_rate": 0.00010646766169154229, "loss": 0.0277, "step": 107 }, { "epoch": 0.1619190404797601, "grad_norm": 0.0453433254814091, "learning_rate": 0.00010746268656716419, "loss": 0.014, "step": 108 }, { "epoch": 0.1634182908545727, "grad_norm": 0.05405715363413614, "learning_rate": 0.00010845771144278607, "loss": 0.0291, "step": 109 }, { "epoch": 0.16491754122938532, "grad_norm": 0.03927392847370858, "learning_rate": 0.00010945273631840796, "loss": 0.0156, "step": 110 }, { "epoch": 0.1664167916041979, "grad_norm": 0.08989520102953286, "learning_rate": 0.00011044776119402987, "loss": 0.0234, "step": 111 }, { "epoch": 0.1679160419790105, "grad_norm": 0.1268177748837639, "learning_rate": 0.00011144278606965175, "loss": 0.0422, "step": 112 }, { "epoch": 0.16941529235382308, "grad_norm": 0.03738615100934019, "learning_rate": 0.00011243781094527364, "loss": 0.0155, "step": 113 }, { "epoch": 0.17091454272863568, "grad_norm": 0.04597772784671858, "learning_rate": 0.00011343283582089552, "loss": 0.0228, "step": 114 }, { "epoch": 0.1724137931034483, "grad_norm": 0.059249691336683494, "learning_rate": 0.00011442786069651741, "loss": 0.0339, "step": 115 }, { "epoch": 0.17391304347826086, "grad_norm": 0.08104873791292298, "learning_rate": 0.00011542288557213932, "loss": 0.0453, "step": 116 }, { "epoch": 0.17541229385307347, "grad_norm": 0.06588671288622634, "learning_rate": 0.0001164179104477612, "loss": 0.0217, "step": 117 }, { "epoch": 0.17691154422788605, "grad_norm": 0.03522749702212235, "learning_rate": 0.00011741293532338308, "loss": 0.0179, "step": 118 }, { "epoch": 0.17841079460269865, "grad_norm": 0.0996319146350363, "learning_rate": 0.00011840796019900498, "loss": 0.0326, "step": 119 }, { "epoch": 0.17991004497751126, "grad_norm": 0.034423484759631535, "learning_rate": 0.00011940298507462686, "loss": 0.013, "step": 120 }, { "epoch": 0.18140929535232383, "grad_norm": 0.07738518975880311, "learning_rate": 0.00012039800995024876, "loss": 0.0276, "step": 121 }, { "epoch": 0.18290854572713644, "grad_norm": 0.07113916964290797, "learning_rate": 0.00012139303482587066, "loss": 0.0296, "step": 122 }, { "epoch": 0.18440779610194902, "grad_norm": 0.04801547092888486, "learning_rate": 0.00012238805970149255, "loss": 0.017, "step": 123 }, { "epoch": 0.18590704647676162, "grad_norm": 0.0708304924302058, "learning_rate": 0.00012338308457711443, "loss": 0.0224, "step": 124 }, { "epoch": 0.1874062968515742, "grad_norm": 0.08325794068925856, "learning_rate": 0.0001243781094527363, "loss": 0.0268, "step": 125 }, { "epoch": 0.1889055472263868, "grad_norm": 0.049077714829375424, "learning_rate": 0.00012537313432835822, "loss": 0.0205, "step": 126 }, { "epoch": 0.1904047976011994, "grad_norm": 0.08564942290202891, "learning_rate": 0.0001263681592039801, "loss": 0.0267, "step": 127 }, { "epoch": 0.191904047976012, "grad_norm": 0.07021233023137041, "learning_rate": 0.000127363184079602, "loss": 0.0318, "step": 128 }, { "epoch": 0.1934032983508246, "grad_norm": 0.09337476280465129, "learning_rate": 0.00012835820895522389, "loss": 0.022, "step": 129 }, { "epoch": 0.19490254872563717, "grad_norm": 0.09811732092693733, "learning_rate": 0.0001293532338308458, "loss": 0.0554, "step": 130 }, { "epoch": 0.19640179910044978, "grad_norm": 0.06961228532261528, "learning_rate": 0.00013034825870646767, "loss": 0.0356, "step": 131 }, { "epoch": 0.19790104947526238, "grad_norm": 0.08084464838848786, "learning_rate": 0.00013134328358208955, "loss": 0.0298, "step": 132 }, { "epoch": 0.19940029985007496, "grad_norm": 0.051768566679399344, "learning_rate": 0.00013233830845771146, "loss": 0.0118, "step": 133 }, { "epoch": 0.20089955022488756, "grad_norm": 0.05361967944259781, "learning_rate": 0.00013333333333333334, "loss": 0.0329, "step": 134 }, { "epoch": 0.20239880059970014, "grad_norm": 0.06705674679942888, "learning_rate": 0.00013432835820895525, "loss": 0.0237, "step": 135 }, { "epoch": 0.20389805097451275, "grad_norm": 0.1588965400271315, "learning_rate": 0.00013532338308457713, "loss": 0.0199, "step": 136 }, { "epoch": 0.20539730134932535, "grad_norm": 0.045155120039124384, "learning_rate": 0.000136318407960199, "loss": 0.0257, "step": 137 }, { "epoch": 0.20689655172413793, "grad_norm": 0.10905612081286897, "learning_rate": 0.0001373134328358209, "loss": 0.0321, "step": 138 }, { "epoch": 0.20839580209895053, "grad_norm": 0.11760170984366283, "learning_rate": 0.0001383084577114428, "loss": 0.0372, "step": 139 }, { "epoch": 0.2098950524737631, "grad_norm": 0.08582443938861739, "learning_rate": 0.0001393034825870647, "loss": 0.0426, "step": 140 }, { "epoch": 0.21139430284857572, "grad_norm": 0.07873025758417922, "learning_rate": 0.00014029850746268658, "loss": 0.025, "step": 141 }, { "epoch": 0.2128935532233883, "grad_norm": 0.10318630643872907, "learning_rate": 0.00014129353233830846, "loss": 0.0333, "step": 142 }, { "epoch": 0.2143928035982009, "grad_norm": 0.062306207851108485, "learning_rate": 0.00014228855721393034, "loss": 0.0174, "step": 143 }, { "epoch": 0.2158920539730135, "grad_norm": 0.06923837226150165, "learning_rate": 0.00014328358208955225, "loss": 0.0236, "step": 144 }, { "epoch": 0.21739130434782608, "grad_norm": 0.08009174706122868, "learning_rate": 0.00014427860696517416, "loss": 0.0377, "step": 145 }, { "epoch": 0.21889055472263869, "grad_norm": 0.07421272166337643, "learning_rate": 0.00014527363184079604, "loss": 0.0285, "step": 146 }, { "epoch": 0.22038980509745126, "grad_norm": 0.06394699190672182, "learning_rate": 0.00014626865671641792, "loss": 0.0201, "step": 147 }, { "epoch": 0.22188905547226387, "grad_norm": 0.1230962560854086, "learning_rate": 0.0001472636815920398, "loss": 0.0418, "step": 148 }, { "epoch": 0.22338830584707647, "grad_norm": 0.09792223962749103, "learning_rate": 0.00014825870646766168, "loss": 0.0428, "step": 149 }, { "epoch": 0.22488755622188905, "grad_norm": 0.08935647805738876, "learning_rate": 0.0001492537313432836, "loss": 0.0286, "step": 150 }, { "epoch": 0.22638680659670166, "grad_norm": 0.05869779630249231, "learning_rate": 0.0001502487562189055, "loss": 0.0266, "step": 151 }, { "epoch": 0.22788605697151423, "grad_norm": 0.09727307974918385, "learning_rate": 0.00015124378109452737, "loss": 0.0251, "step": 152 }, { "epoch": 0.22938530734632684, "grad_norm": 0.05489680833229503, "learning_rate": 0.00015223880597014925, "loss": 0.0256, "step": 153 }, { "epoch": 0.23088455772113944, "grad_norm": 0.08651252294071486, "learning_rate": 0.00015323383084577113, "loss": 0.0324, "step": 154 }, { "epoch": 0.23238380809595202, "grad_norm": 0.06084445807329521, "learning_rate": 0.00015422885572139304, "loss": 0.0181, "step": 155 }, { "epoch": 0.23388305847076463, "grad_norm": 0.07984776396235738, "learning_rate": 0.00015522388059701495, "loss": 0.0293, "step": 156 }, { "epoch": 0.2353823088455772, "grad_norm": 0.10663711823358574, "learning_rate": 0.00015621890547263683, "loss": 0.0224, "step": 157 }, { "epoch": 0.2368815592203898, "grad_norm": 0.09029974452506394, "learning_rate": 0.0001572139303482587, "loss": 0.0391, "step": 158 }, { "epoch": 0.2383808095952024, "grad_norm": 0.07475818991216006, "learning_rate": 0.00015820895522388059, "loss": 0.0325, "step": 159 }, { "epoch": 0.239880059970015, "grad_norm": 0.10281310121414294, "learning_rate": 0.0001592039800995025, "loss": 0.0233, "step": 160 }, { "epoch": 0.2413793103448276, "grad_norm": 0.07648721213716447, "learning_rate": 0.0001601990049751244, "loss": 0.0309, "step": 161 }, { "epoch": 0.24287856071964017, "grad_norm": 0.1353558632906228, "learning_rate": 0.00016119402985074628, "loss": 0.0412, "step": 162 }, { "epoch": 0.24437781109445278, "grad_norm": 0.11951297623629252, "learning_rate": 0.00016218905472636816, "loss": 0.0454, "step": 163 }, { "epoch": 0.24587706146926536, "grad_norm": 0.09337941334405561, "learning_rate": 0.00016318407960199004, "loss": 0.0261, "step": 164 }, { "epoch": 0.24737631184407796, "grad_norm": 0.07631986818436004, "learning_rate": 0.00016417910447761195, "loss": 0.0184, "step": 165 }, { "epoch": 0.24887556221889057, "grad_norm": 0.06382771953073095, "learning_rate": 0.00016517412935323385, "loss": 0.0229, "step": 166 }, { "epoch": 0.25037481259370314, "grad_norm": 0.0654124597455561, "learning_rate": 0.00016616915422885573, "loss": 0.0275, "step": 167 }, { "epoch": 0.2518740629685157, "grad_norm": 0.07386539118384822, "learning_rate": 0.00016716417910447761, "loss": 0.0276, "step": 168 }, { "epoch": 0.25337331334332835, "grad_norm": 0.10790605839437757, "learning_rate": 0.0001681592039800995, "loss": 0.0284, "step": 169 }, { "epoch": 0.25487256371814093, "grad_norm": 0.07520911787134094, "learning_rate": 0.0001691542288557214, "loss": 0.033, "step": 170 }, { "epoch": 0.2563718140929535, "grad_norm": 0.0437842352906429, "learning_rate": 0.00017014925373134328, "loss": 0.0164, "step": 171 }, { "epoch": 0.25787106446776614, "grad_norm": 0.1241410943091543, "learning_rate": 0.0001711442786069652, "loss": 0.0338, "step": 172 }, { "epoch": 0.2593703148425787, "grad_norm": 0.08555919176928421, "learning_rate": 0.00017213930348258707, "loss": 0.0219, "step": 173 }, { "epoch": 0.2608695652173913, "grad_norm": 0.11079300082585909, "learning_rate": 0.00017313432835820895, "loss": 0.0314, "step": 174 }, { "epoch": 0.2623688155922039, "grad_norm": 0.046045735438355655, "learning_rate": 0.00017412935323383086, "loss": 0.0111, "step": 175 }, { "epoch": 0.2638680659670165, "grad_norm": 0.10739428025622968, "learning_rate": 0.00017512437810945274, "loss": 0.0224, "step": 176 }, { "epoch": 0.2653673163418291, "grad_norm": 0.2449779327117588, "learning_rate": 0.00017611940298507464, "loss": 0.0464, "step": 177 }, { "epoch": 0.26686656671664166, "grad_norm": 0.08360762118297338, "learning_rate": 0.00017711442786069652, "loss": 0.0312, "step": 178 }, { "epoch": 0.2683658170914543, "grad_norm": 0.10593436165792847, "learning_rate": 0.00017810945273631843, "loss": 0.026, "step": 179 }, { "epoch": 0.2698650674662669, "grad_norm": 0.07842081945251449, "learning_rate": 0.0001791044776119403, "loss": 0.029, "step": 180 }, { "epoch": 0.27136431784107945, "grad_norm": 0.0919363194425753, "learning_rate": 0.0001800995024875622, "loss": 0.0246, "step": 181 }, { "epoch": 0.272863568215892, "grad_norm": 0.07122712223445765, "learning_rate": 0.0001810945273631841, "loss": 0.0175, "step": 182 }, { "epoch": 0.27436281859070466, "grad_norm": 0.12986448311978344, "learning_rate": 0.00018208955223880598, "loss": 0.0327, "step": 183 }, { "epoch": 0.27586206896551724, "grad_norm": 0.08939968166772469, "learning_rate": 0.00018308457711442789, "loss": 0.0222, "step": 184 }, { "epoch": 0.2773613193403298, "grad_norm": 0.12467738587334767, "learning_rate": 0.00018407960199004977, "loss": 0.0398, "step": 185 }, { "epoch": 0.27886056971514245, "grad_norm": 0.08880261150192895, "learning_rate": 0.00018507462686567165, "loss": 0.0276, "step": 186 }, { "epoch": 0.280359820089955, "grad_norm": 0.06274559081133477, "learning_rate": 0.00018606965174129353, "loss": 0.0162, "step": 187 }, { "epoch": 0.2818590704647676, "grad_norm": 0.06826841539115386, "learning_rate": 0.00018706467661691543, "loss": 0.0252, "step": 188 }, { "epoch": 0.28335832083958024, "grad_norm": 0.07615150472353269, "learning_rate": 0.00018805970149253734, "loss": 0.0206, "step": 189 }, { "epoch": 0.2848575712143928, "grad_norm": 0.09408473494987521, "learning_rate": 0.00018905472636815922, "loss": 0.0274, "step": 190 }, { "epoch": 0.2863568215892054, "grad_norm": 0.07069036227398094, "learning_rate": 0.0001900497512437811, "loss": 0.0196, "step": 191 }, { "epoch": 0.28785607196401797, "grad_norm": 0.09160386389460257, "learning_rate": 0.00019104477611940298, "loss": 0.0425, "step": 192 }, { "epoch": 0.2893553223388306, "grad_norm": 0.11814489122880895, "learning_rate": 0.0001920398009950249, "loss": 0.0312, "step": 193 }, { "epoch": 0.2908545727136432, "grad_norm": 0.07267374829272864, "learning_rate": 0.0001930348258706468, "loss": 0.0206, "step": 194 }, { "epoch": 0.29235382308845576, "grad_norm": 0.056069092746323, "learning_rate": 0.00019402985074626867, "loss": 0.0173, "step": 195 }, { "epoch": 0.2938530734632684, "grad_norm": 0.1072633610690865, "learning_rate": 0.00019502487562189055, "loss": 0.0356, "step": 196 }, { "epoch": 0.29535232383808097, "grad_norm": 0.10655067222684735, "learning_rate": 0.00019601990049751244, "loss": 0.0381, "step": 197 }, { "epoch": 0.29685157421289354, "grad_norm": 0.12811931955318204, "learning_rate": 0.00019701492537313434, "loss": 0.0328, "step": 198 }, { "epoch": 0.2983508245877061, "grad_norm": 0.0896744841892114, "learning_rate": 0.00019800995024875625, "loss": 0.0201, "step": 199 }, { "epoch": 0.29985007496251875, "grad_norm": 0.08188692710670159, "learning_rate": 0.00019900497512437813, "loss": 0.026, "step": 200 }, { "epoch": 0.30134932533733133, "grad_norm": 0.06429287812218996, "learning_rate": 0.0002, "loss": 0.0236, "step": 201 }, { "epoch": 0.3028485757121439, "grad_norm": 0.0889591592714688, "learning_rate": 0.00019999984769132877, "loss": 0.018, "step": 202 }, { "epoch": 0.30434782608695654, "grad_norm": 0.09099341079961838, "learning_rate": 0.00019999939076577905, "loss": 0.0341, "step": 203 }, { "epoch": 0.3058470764617691, "grad_norm": 0.1056474393511896, "learning_rate": 0.00019999862922474268, "loss": 0.0438, "step": 204 }, { "epoch": 0.3073463268365817, "grad_norm": 0.06324696346281937, "learning_rate": 0.00019999756307053948, "loss": 0.02, "step": 205 }, { "epoch": 0.30884557721139433, "grad_norm": 0.12201171857861728, "learning_rate": 0.00019999619230641713, "loss": 0.0357, "step": 206 }, { "epoch": 0.3103448275862069, "grad_norm": 0.16019354966510604, "learning_rate": 0.00019999451693655123, "loss": 0.0308, "step": 207 }, { "epoch": 0.3118440779610195, "grad_norm": 0.08064191205932211, "learning_rate": 0.00019999253696604522, "loss": 0.0231, "step": 208 }, { "epoch": 0.31334332833583206, "grad_norm": 0.1735406201730255, "learning_rate": 0.00019999025240093044, "loss": 0.0531, "step": 209 }, { "epoch": 0.3148425787106447, "grad_norm": 0.1270042673971674, "learning_rate": 0.00019998766324816607, "loss": 0.0372, "step": 210 }, { "epoch": 0.31634182908545727, "grad_norm": 0.07759625937069312, "learning_rate": 0.00019998476951563915, "loss": 0.0215, "step": 211 }, { "epoch": 0.31784107946026985, "grad_norm": 0.06823751969137144, "learning_rate": 0.00019998157121216444, "loss": 0.0188, "step": 212 }, { "epoch": 0.3193403298350825, "grad_norm": 0.08549856995516077, "learning_rate": 0.00019997806834748456, "loss": 0.0254, "step": 213 }, { "epoch": 0.32083958020989506, "grad_norm": 0.0625354342920355, "learning_rate": 0.00019997426093226986, "loss": 0.014, "step": 214 }, { "epoch": 0.32233883058470764, "grad_norm": 0.0760895159675007, "learning_rate": 0.00019997014897811833, "loss": 0.0136, "step": 215 }, { "epoch": 0.3238380809595202, "grad_norm": 0.1679998356790272, "learning_rate": 0.00019996573249755572, "loss": 0.0546, "step": 216 }, { "epoch": 0.32533733133433285, "grad_norm": 0.10916710981504504, "learning_rate": 0.00019996101150403543, "loss": 0.0275, "step": 217 }, { "epoch": 0.3268365817091454, "grad_norm": 0.10201374094804785, "learning_rate": 0.00019995598601193842, "loss": 0.0289, "step": 218 }, { "epoch": 0.328335832083958, "grad_norm": 0.08058975597656892, "learning_rate": 0.00019995065603657316, "loss": 0.0249, "step": 219 }, { "epoch": 0.32983508245877063, "grad_norm": 0.15985586700330837, "learning_rate": 0.00019994502159417573, "loss": 0.0345, "step": 220 }, { "epoch": 0.3313343328335832, "grad_norm": 0.14812712527123406, "learning_rate": 0.0001999390827019096, "loss": 0.0377, "step": 221 }, { "epoch": 0.3328335832083958, "grad_norm": 0.1178146028681012, "learning_rate": 0.00019993283937786563, "loss": 0.0304, "step": 222 }, { "epoch": 0.3343328335832084, "grad_norm": 0.10003216574781101, "learning_rate": 0.0001999262916410621, "loss": 0.0342, "step": 223 }, { "epoch": 0.335832083958021, "grad_norm": 0.10739566491501087, "learning_rate": 0.0001999194395114446, "loss": 0.0387, "step": 224 }, { "epoch": 0.3373313343328336, "grad_norm": 0.12471587478993325, "learning_rate": 0.00019991228300988585, "loss": 0.0271, "step": 225 }, { "epoch": 0.33883058470764615, "grad_norm": 0.10534548813143461, "learning_rate": 0.0001999048221581858, "loss": 0.0338, "step": 226 }, { "epoch": 0.3403298350824588, "grad_norm": 0.06743838763365886, "learning_rate": 0.00019989705697907149, "loss": 0.02, "step": 227 }, { "epoch": 0.34182908545727136, "grad_norm": 0.0757865381180341, "learning_rate": 0.00019988898749619702, "loss": 0.0215, "step": 228 }, { "epoch": 0.34332833583208394, "grad_norm": 0.14709355458702295, "learning_rate": 0.0001998806137341434, "loss": 0.0227, "step": 229 }, { "epoch": 0.3448275862068966, "grad_norm": 0.06430710408458629, "learning_rate": 0.0001998719357184186, "loss": 0.0122, "step": 230 }, { "epoch": 0.34632683658170915, "grad_norm": 0.09018751183295319, "learning_rate": 0.0001998629534754574, "loss": 0.0217, "step": 231 }, { "epoch": 0.34782608695652173, "grad_norm": 0.14650415427998667, "learning_rate": 0.0001998536670326212, "loss": 0.0493, "step": 232 }, { "epoch": 0.3493253373313343, "grad_norm": 0.06334988631041422, "learning_rate": 0.00019984407641819812, "loss": 0.0239, "step": 233 }, { "epoch": 0.35082458770614694, "grad_norm": 0.20886697619030714, "learning_rate": 0.00019983418166140285, "loss": 0.0564, "step": 234 }, { "epoch": 0.3523238380809595, "grad_norm": 0.14466113183818535, "learning_rate": 0.00019982398279237655, "loss": 0.0204, "step": 235 }, { "epoch": 0.3538230884557721, "grad_norm": 0.0982852108329072, "learning_rate": 0.0001998134798421867, "loss": 0.0273, "step": 236 }, { "epoch": 0.3553223388305847, "grad_norm": 0.11164221678350923, "learning_rate": 0.00019980267284282717, "loss": 0.0282, "step": 237 }, { "epoch": 0.3568215892053973, "grad_norm": 0.10368498660671552, "learning_rate": 0.0001997915618272179, "loss": 0.0299, "step": 238 }, { "epoch": 0.3583208395802099, "grad_norm": 0.09881979045943046, "learning_rate": 0.000199780146829205, "loss": 0.0255, "step": 239 }, { "epoch": 0.3598200899550225, "grad_norm": 0.1379397238385945, "learning_rate": 0.00019976842788356055, "loss": 0.0217, "step": 240 }, { "epoch": 0.3613193403298351, "grad_norm": 0.07438604732967513, "learning_rate": 0.00019975640502598244, "loss": 0.0243, "step": 241 }, { "epoch": 0.36281859070464767, "grad_norm": 0.07473607637141821, "learning_rate": 0.0001997440782930944, "loss": 0.0258, "step": 242 }, { "epoch": 0.36431784107946025, "grad_norm": 0.0882005313753972, "learning_rate": 0.00019973144772244582, "loss": 0.0268, "step": 243 }, { "epoch": 0.3658170914542729, "grad_norm": 0.14911031648636602, "learning_rate": 0.00019971851335251158, "loss": 0.0378, "step": 244 }, { "epoch": 0.36731634182908546, "grad_norm": 0.07841452122968684, "learning_rate": 0.00019970527522269205, "loss": 0.0174, "step": 245 }, { "epoch": 0.36881559220389803, "grad_norm": 0.09528161070909874, "learning_rate": 0.0001996917333733128, "loss": 0.0295, "step": 246 }, { "epoch": 0.37031484257871067, "grad_norm": 0.1547071536848402, "learning_rate": 0.00019967788784562473, "loss": 0.0352, "step": 247 }, { "epoch": 0.37181409295352325, "grad_norm": 0.13481821743931183, "learning_rate": 0.00019966373868180365, "loss": 0.0375, "step": 248 }, { "epoch": 0.3733133433283358, "grad_norm": 0.14639591924374834, "learning_rate": 0.00019964928592495045, "loss": 0.0408, "step": 249 }, { "epoch": 0.3748125937031484, "grad_norm": 0.22468937977099931, "learning_rate": 0.00019963452961909063, "loss": 0.0487, "step": 250 }, { "epoch": 0.37631184407796103, "grad_norm": 0.10317429082367974, "learning_rate": 0.00019961946980917456, "loss": 0.0353, "step": 251 }, { "epoch": 0.3778110944527736, "grad_norm": 0.11239999568993565, "learning_rate": 0.00019960410654107697, "loss": 0.0198, "step": 252 }, { "epoch": 0.3793103448275862, "grad_norm": 0.12839126097254944, "learning_rate": 0.00019958843986159704, "loss": 0.0271, "step": 253 }, { "epoch": 0.3808095952023988, "grad_norm": 0.16134712680852636, "learning_rate": 0.00019957246981845822, "loss": 0.0939, "step": 254 }, { "epoch": 0.3823088455772114, "grad_norm": 0.10694136315654908, "learning_rate": 0.00019955619646030802, "loss": 0.0191, "step": 255 }, { "epoch": 0.383808095952024, "grad_norm": 0.13584383708557063, "learning_rate": 0.00019953961983671788, "loss": 0.0391, "step": 256 }, { "epoch": 0.3853073463268366, "grad_norm": 0.07484156837632686, "learning_rate": 0.0001995227399981831, "loss": 0.0218, "step": 257 }, { "epoch": 0.3868065967016492, "grad_norm": 0.11474869317330276, "learning_rate": 0.00019950555699612267, "loss": 0.0185, "step": 258 }, { "epoch": 0.38830584707646176, "grad_norm": 0.08204674574283162, "learning_rate": 0.00019948807088287883, "loss": 0.0167, "step": 259 }, { "epoch": 0.38980509745127434, "grad_norm": 0.08466516343662758, "learning_rate": 0.00019947028171171742, "loss": 0.0228, "step": 260 }, { "epoch": 0.391304347826087, "grad_norm": 0.10295941699523375, "learning_rate": 0.00019945218953682734, "loss": 0.035, "step": 261 }, { "epoch": 0.39280359820089955, "grad_norm": 0.10075799964627083, "learning_rate": 0.00019943379441332047, "loss": 0.0214, "step": 262 }, { "epoch": 0.39430284857571213, "grad_norm": 0.06840271813941545, "learning_rate": 0.00019941509639723155, "loss": 0.0134, "step": 263 }, { "epoch": 0.39580209895052476, "grad_norm": 0.18490087281473808, "learning_rate": 0.000199396095545518, "loss": 0.0315, "step": 264 }, { "epoch": 0.39730134932533734, "grad_norm": 0.3398552153859189, "learning_rate": 0.00019937679191605963, "loss": 0.0447, "step": 265 }, { "epoch": 0.3988005997001499, "grad_norm": 0.11181599880625129, "learning_rate": 0.00019935718556765876, "loss": 0.0252, "step": 266 }, { "epoch": 0.4002998500749625, "grad_norm": 0.15176000894386518, "learning_rate": 0.00019933727656003963, "loss": 0.0439, "step": 267 }, { "epoch": 0.4017991004497751, "grad_norm": 0.11515411944836743, "learning_rate": 0.00019931706495384863, "loss": 0.0317, "step": 268 }, { "epoch": 0.4032983508245877, "grad_norm": 0.07747512816332972, "learning_rate": 0.0001992965508106537, "loss": 0.029, "step": 269 }, { "epoch": 0.4047976011994003, "grad_norm": 0.16798336539466618, "learning_rate": 0.00019927573419294456, "loss": 0.072, "step": 270 }, { "epoch": 0.4062968515742129, "grad_norm": 0.14842393353919936, "learning_rate": 0.00019925461516413223, "loss": 0.0437, "step": 271 }, { "epoch": 0.4077961019490255, "grad_norm": 0.1450131939805502, "learning_rate": 0.00019923319378854887, "loss": 0.0498, "step": 272 }, { "epoch": 0.40929535232383807, "grad_norm": 0.17363621713440097, "learning_rate": 0.0001992114701314478, "loss": 0.0611, "step": 273 }, { "epoch": 0.4107946026986507, "grad_norm": 0.11435400082136321, "learning_rate": 0.000199189444259003, "loss": 0.0227, "step": 274 }, { "epoch": 0.4122938530734633, "grad_norm": 0.16518050629422557, "learning_rate": 0.00019916711623830903, "loss": 0.0283, "step": 275 }, { "epoch": 0.41379310344827586, "grad_norm": 0.14080617976911697, "learning_rate": 0.00019914448613738106, "loss": 0.0244, "step": 276 }, { "epoch": 0.41529235382308843, "grad_norm": 0.10018538841227774, "learning_rate": 0.00019912155402515417, "loss": 0.0256, "step": 277 }, { "epoch": 0.41679160419790107, "grad_norm": 0.09288507466343968, "learning_rate": 0.00019909831997148362, "loss": 0.0205, "step": 278 }, { "epoch": 0.41829085457271364, "grad_norm": 0.10076396934289908, "learning_rate": 0.00019907478404714436, "loss": 0.0261, "step": 279 }, { "epoch": 0.4197901049475262, "grad_norm": 0.14235107824165918, "learning_rate": 0.0001990509463238309, "loss": 0.0317, "step": 280 }, { "epoch": 0.42128935532233885, "grad_norm": 0.06857782839954504, "learning_rate": 0.00019902680687415705, "loss": 0.016, "step": 281 }, { "epoch": 0.42278860569715143, "grad_norm": 0.21507070087913607, "learning_rate": 0.00019900236577165576, "loss": 0.0411, "step": 282 }, { "epoch": 0.424287856071964, "grad_norm": 0.0949166136382304, "learning_rate": 0.0001989776230907789, "loss": 0.0257, "step": 283 }, { "epoch": 0.4257871064467766, "grad_norm": 0.1934392939287711, "learning_rate": 0.00019895257890689696, "loss": 0.0381, "step": 284 }, { "epoch": 0.4272863568215892, "grad_norm": 0.17422202559565292, "learning_rate": 0.00019892723329629887, "loss": 0.0261, "step": 285 }, { "epoch": 0.4287856071964018, "grad_norm": 0.07223682667659033, "learning_rate": 0.0001989015863361917, "loss": 0.0255, "step": 286 }, { "epoch": 0.4302848575712144, "grad_norm": 0.12279145351334636, "learning_rate": 0.0001988756381047006, "loss": 0.046, "step": 287 }, { "epoch": 0.431784107946027, "grad_norm": 0.08991953597169815, "learning_rate": 0.00019884938868086835, "loss": 0.0315, "step": 288 }, { "epoch": 0.4332833583208396, "grad_norm": 0.11188039743306034, "learning_rate": 0.0001988228381446553, "loss": 0.0424, "step": 289 }, { "epoch": 0.43478260869565216, "grad_norm": 0.18905162896839114, "learning_rate": 0.00019879598657693891, "loss": 0.0427, "step": 290 }, { "epoch": 0.4362818590704648, "grad_norm": 0.12194462370002573, "learning_rate": 0.00019876883405951377, "loss": 0.0225, "step": 291 }, { "epoch": 0.43778110944527737, "grad_norm": 0.27514043276591943, "learning_rate": 0.00019874138067509117, "loss": 0.0556, "step": 292 }, { "epoch": 0.43928035982008995, "grad_norm": 0.09504669927857831, "learning_rate": 0.0001987136265072988, "loss": 0.0182, "step": 293 }, { "epoch": 0.4407796101949025, "grad_norm": 0.08844611007050139, "learning_rate": 0.00019868557164068074, "loss": 0.018, "step": 294 }, { "epoch": 0.44227886056971516, "grad_norm": 0.1369491212885427, "learning_rate": 0.00019865721616069696, "loss": 0.0492, "step": 295 }, { "epoch": 0.44377811094452774, "grad_norm": 0.1142377954005939, "learning_rate": 0.00019862856015372317, "loss": 0.0334, "step": 296 }, { "epoch": 0.4452773613193403, "grad_norm": 0.1961332653625071, "learning_rate": 0.0001985996037070505, "loss": 0.0481, "step": 297 }, { "epoch": 0.44677661169415295, "grad_norm": 0.050262046246270925, "learning_rate": 0.00019857034690888537, "loss": 0.0154, "step": 298 }, { "epoch": 0.4482758620689655, "grad_norm": 0.10957587831691169, "learning_rate": 0.00019854078984834903, "loss": 0.0302, "step": 299 }, { "epoch": 0.4497751124437781, "grad_norm": 0.18658220857685753, "learning_rate": 0.0001985109326154774, "loss": 0.0317, "step": 300 }, { "epoch": 0.4512743628185907, "grad_norm": 0.09056328641521297, "learning_rate": 0.00019848077530122083, "loss": 0.026, "step": 301 }, { "epoch": 0.4527736131934033, "grad_norm": 0.16097758286803418, "learning_rate": 0.00019845031799744367, "loss": 0.0457, "step": 302 }, { "epoch": 0.4542728635682159, "grad_norm": 0.07689789497731149, "learning_rate": 0.0001984195607969242, "loss": 0.0202, "step": 303 }, { "epoch": 0.45577211394302847, "grad_norm": 0.10663066866505838, "learning_rate": 0.00019838850379335417, "loss": 0.0288, "step": 304 }, { "epoch": 0.4572713643178411, "grad_norm": 0.10553280713408417, "learning_rate": 0.00019835714708133862, "loss": 0.0292, "step": 305 }, { "epoch": 0.4587706146926537, "grad_norm": 0.15579620357045054, "learning_rate": 0.0001983254907563955, "loss": 0.0236, "step": 306 }, { "epoch": 0.46026986506746626, "grad_norm": 0.09060539717469658, "learning_rate": 0.00019829353491495545, "loss": 0.0258, "step": 307 }, { "epoch": 0.4617691154422789, "grad_norm": 0.12084292750432825, "learning_rate": 0.00019826127965436152, "loss": 0.0292, "step": 308 }, { "epoch": 0.46326836581709147, "grad_norm": 0.20235716347011842, "learning_rate": 0.0001982287250728689, "loss": 0.0395, "step": 309 }, { "epoch": 0.46476761619190404, "grad_norm": 0.09144766914600604, "learning_rate": 0.00019819587126964437, "loss": 0.0243, "step": 310 }, { "epoch": 0.4662668665667166, "grad_norm": 0.1410307244411355, "learning_rate": 0.00019816271834476642, "loss": 0.0377, "step": 311 }, { "epoch": 0.46776611694152925, "grad_norm": 0.12151655600598343, "learning_rate": 0.0001981292663992245, "loss": 0.0228, "step": 312 }, { "epoch": 0.46926536731634183, "grad_norm": 0.13822001628609412, "learning_rate": 0.00019809551553491916, "loss": 0.031, "step": 313 }, { "epoch": 0.4707646176911544, "grad_norm": 0.19923329717115837, "learning_rate": 0.0001980614658546613, "loss": 0.0542, "step": 314 }, { "epoch": 0.47226386806596704, "grad_norm": 0.32163446242573723, "learning_rate": 0.00019802711746217218, "loss": 0.0399, "step": 315 }, { "epoch": 0.4737631184407796, "grad_norm": 0.16137241975788913, "learning_rate": 0.00019799247046208297, "loss": 0.0366, "step": 316 }, { "epoch": 0.4752623688155922, "grad_norm": 0.12091006379875421, "learning_rate": 0.0001979575249599344, "loss": 0.0246, "step": 317 }, { "epoch": 0.4767616191904048, "grad_norm": 0.12233306737992392, "learning_rate": 0.00019792228106217658, "loss": 0.0341, "step": 318 }, { "epoch": 0.4782608695652174, "grad_norm": 0.10780363270946153, "learning_rate": 0.0001978867388761685, "loss": 0.0259, "step": 319 }, { "epoch": 0.47976011994003, "grad_norm": 0.18307588018838652, "learning_rate": 0.00019785089851017787, "loss": 0.0296, "step": 320 }, { "epoch": 0.48125937031484256, "grad_norm": 0.2732266398896659, "learning_rate": 0.00019781476007338058, "loss": 0.0412, "step": 321 }, { "epoch": 0.4827586206896552, "grad_norm": 0.1607986328742017, "learning_rate": 0.00019777832367586063, "loss": 0.044, "step": 322 }, { "epoch": 0.48425787106446777, "grad_norm": 0.142980064491954, "learning_rate": 0.0001977415894286096, "loss": 0.0475, "step": 323 }, { "epoch": 0.48575712143928035, "grad_norm": 0.06557574574115624, "learning_rate": 0.0001977045574435264, "loss": 0.0198, "step": 324 }, { "epoch": 0.487256371814093, "grad_norm": 0.0839583522006225, "learning_rate": 0.0001976672278334168, "loss": 0.0232, "step": 325 }, { "epoch": 0.48875562218890556, "grad_norm": 0.14621694058585588, "learning_rate": 0.00019762960071199333, "loss": 0.0355, "step": 326 }, { "epoch": 0.49025487256371814, "grad_norm": 0.15402000439221136, "learning_rate": 0.00019759167619387476, "loss": 0.0274, "step": 327 }, { "epoch": 0.4917541229385307, "grad_norm": 0.08261626886873759, "learning_rate": 0.00019755345439458565, "loss": 0.0285, "step": 328 }, { "epoch": 0.49325337331334335, "grad_norm": 0.13478764449331881, "learning_rate": 0.00019751493543055632, "loss": 0.0311, "step": 329 }, { "epoch": 0.4947526236881559, "grad_norm": 0.14610763557960335, "learning_rate": 0.0001974761194191222, "loss": 0.035, "step": 330 }, { "epoch": 0.4962518740629685, "grad_norm": 0.1274425780126433, "learning_rate": 0.00019743700647852354, "loss": 0.0234, "step": 331 }, { "epoch": 0.49775112443778113, "grad_norm": 0.09253891406497454, "learning_rate": 0.0001973975967279052, "loss": 0.0265, "step": 332 }, { "epoch": 0.4992503748125937, "grad_norm": 0.07426081606342527, "learning_rate": 0.00019735789028731604, "loss": 0.0234, "step": 333 }, { "epoch": 0.5007496251874063, "grad_norm": 0.19068637883417455, "learning_rate": 0.00019731788727770885, "loss": 0.0382, "step": 334 }, { "epoch": 0.5022488755622189, "grad_norm": 0.11943906440034463, "learning_rate": 0.00019727758782093967, "loss": 0.0303, "step": 335 }, { "epoch": 0.5037481259370314, "grad_norm": 0.08313883841221202, "learning_rate": 0.00019723699203976766, "loss": 0.0197, "step": 336 }, { "epoch": 0.5052473763118441, "grad_norm": 0.1417916755841332, "learning_rate": 0.00019719610005785465, "loss": 0.0336, "step": 337 }, { "epoch": 0.5067466266866567, "grad_norm": 0.11137030534188308, "learning_rate": 0.0001971549119997646, "loss": 0.03, "step": 338 }, { "epoch": 0.5082458770614693, "grad_norm": 0.1474204159985949, "learning_rate": 0.00019711342799096361, "loss": 0.0305, "step": 339 }, { "epoch": 0.5097451274362819, "grad_norm": 0.6073752811706395, "learning_rate": 0.00019707164815781908, "loss": 0.0512, "step": 340 }, { "epoch": 0.5112443778110944, "grad_norm": 0.18317207850162331, "learning_rate": 0.00019702957262759965, "loss": 0.0244, "step": 341 }, { "epoch": 0.512743628185907, "grad_norm": 0.14617080172691455, "learning_rate": 0.00019698720152847468, "loss": 0.0293, "step": 342 }, { "epoch": 0.5142428785607196, "grad_norm": 0.21227866013123528, "learning_rate": 0.0001969445349895139, "loss": 0.0273, "step": 343 }, { "epoch": 0.5157421289355323, "grad_norm": 0.11079359835152516, "learning_rate": 0.00019690157314068696, "loss": 0.02, "step": 344 }, { "epoch": 0.5172413793103449, "grad_norm": 0.089040858251259, "learning_rate": 0.0001968583161128631, "loss": 0.0254, "step": 345 }, { "epoch": 0.5187406296851574, "grad_norm": 0.22413723696972337, "learning_rate": 0.0001968147640378108, "loss": 0.0298, "step": 346 }, { "epoch": 0.52023988005997, "grad_norm": 0.0875455049017176, "learning_rate": 0.00019677091704819715, "loss": 0.023, "step": 347 }, { "epoch": 0.5217391304347826, "grad_norm": 0.10256718521877016, "learning_rate": 0.0001967267752775877, "loss": 0.0277, "step": 348 }, { "epoch": 0.5232383808095952, "grad_norm": 0.17080254667328584, "learning_rate": 0.00019668233886044597, "loss": 0.0345, "step": 349 }, { "epoch": 0.5247376311844077, "grad_norm": 0.1475350728620058, "learning_rate": 0.00019663760793213296, "loss": 0.0338, "step": 350 }, { "epoch": 0.5262368815592204, "grad_norm": 0.10911209234569877, "learning_rate": 0.00019659258262890683, "loss": 0.0293, "step": 351 }, { "epoch": 0.527736131934033, "grad_norm": 0.10321559635673279, "learning_rate": 0.0001965472630879225, "loss": 0.0289, "step": 352 }, { "epoch": 0.5292353823088456, "grad_norm": 0.10787457610079623, "learning_rate": 0.00019650164944723115, "loss": 0.0361, "step": 353 }, { "epoch": 0.5307346326836582, "grad_norm": 0.0837429373957743, "learning_rate": 0.00019645574184577982, "loss": 0.0212, "step": 354 }, { "epoch": 0.5322338830584707, "grad_norm": 0.09945175724905725, "learning_rate": 0.00019640954042341103, "loss": 0.0172, "step": 355 }, { "epoch": 0.5337331334332833, "grad_norm": 0.059734968422014076, "learning_rate": 0.0001963630453208623, "loss": 0.0204, "step": 356 }, { "epoch": 0.5352323838080959, "grad_norm": 0.17974902185014147, "learning_rate": 0.00019631625667976583, "loss": 0.0358, "step": 357 }, { "epoch": 0.5367316341829086, "grad_norm": 0.09827234190096189, "learning_rate": 0.0001962691746426479, "loss": 0.0215, "step": 358 }, { "epoch": 0.5382308845577212, "grad_norm": 0.16005323730952734, "learning_rate": 0.00019622179935292855, "loss": 0.05, "step": 359 }, { "epoch": 0.5397301349325337, "grad_norm": 0.1349661412175, "learning_rate": 0.00019617413095492114, "loss": 0.0352, "step": 360 }, { "epoch": 0.5412293853073463, "grad_norm": 0.16328511581195324, "learning_rate": 0.0001961261695938319, "loss": 0.0397, "step": 361 }, { "epoch": 0.5427286356821589, "grad_norm": 0.1751158166212692, "learning_rate": 0.00019607791541575943, "loss": 0.039, "step": 362 }, { "epoch": 0.5442278860569715, "grad_norm": 0.14426941275516605, "learning_rate": 0.0001960293685676943, "loss": 0.0335, "step": 363 }, { "epoch": 0.545727136431784, "grad_norm": 0.13144454729693233, "learning_rate": 0.0001959805291975187, "loss": 0.0307, "step": 364 }, { "epoch": 0.5472263868065967, "grad_norm": 0.1025662618949345, "learning_rate": 0.00019593139745400576, "loss": 0.0257, "step": 365 }, { "epoch": 0.5487256371814093, "grad_norm": 0.0980815537212099, "learning_rate": 0.0001958819734868193, "loss": 0.0307, "step": 366 }, { "epoch": 0.5502248875562219, "grad_norm": 0.11890687937548734, "learning_rate": 0.00019583225744651333, "loss": 0.0217, "step": 367 }, { "epoch": 0.5517241379310345, "grad_norm": 0.05424144596101746, "learning_rate": 0.0001957822494845315, "loss": 0.0119, "step": 368 }, { "epoch": 0.553223388305847, "grad_norm": 0.35530516756748276, "learning_rate": 0.00019573194975320673, "loss": 0.0515, "step": 369 }, { "epoch": 0.5547226386806596, "grad_norm": 0.084039953250112, "learning_rate": 0.00019568135840576076, "loss": 0.0222, "step": 370 }, { "epoch": 0.5562218890554723, "grad_norm": 0.0719021205960788, "learning_rate": 0.00019563047559630357, "loss": 0.0129, "step": 371 }, { "epoch": 0.5577211394302849, "grad_norm": 0.0831894222429869, "learning_rate": 0.00019557930147983302, "loss": 0.0203, "step": 372 }, { "epoch": 0.5592203898050975, "grad_norm": 0.15329768574581998, "learning_rate": 0.00019552783621223436, "loss": 0.0395, "step": 373 }, { "epoch": 0.56071964017991, "grad_norm": 0.2358016986495495, "learning_rate": 0.00019547607995027978, "loss": 0.1109, "step": 374 }, { "epoch": 0.5622188905547226, "grad_norm": 0.12867019275964311, "learning_rate": 0.0001954240328516277, "loss": 0.0297, "step": 375 }, { "epoch": 0.5637181409295352, "grad_norm": 0.29639255369886375, "learning_rate": 0.0001953716950748227, "loss": 0.0372, "step": 376 }, { "epoch": 0.5652173913043478, "grad_norm": 0.11159538070632506, "learning_rate": 0.0001953190667792947, "loss": 0.0229, "step": 377 }, { "epoch": 0.5667166416791605, "grad_norm": 0.21855748359403565, "learning_rate": 0.00019526614812535864, "loss": 0.0608, "step": 378 }, { "epoch": 0.568215892053973, "grad_norm": 0.07792574697692431, "learning_rate": 0.00019521293927421388, "loss": 0.0178, "step": 379 }, { "epoch": 0.5697151424287856, "grad_norm": 0.19263126390415305, "learning_rate": 0.00019515944038794384, "loss": 0.0326, "step": 380 }, { "epoch": 0.5712143928035982, "grad_norm": 0.08384659637263117, "learning_rate": 0.00019510565162951537, "loss": 0.02, "step": 381 }, { "epoch": 0.5727136431784108, "grad_norm": 0.1666776601307308, "learning_rate": 0.00019505157316277837, "loss": 0.0371, "step": 382 }, { "epoch": 0.5742128935532234, "grad_norm": 0.07115309715230671, "learning_rate": 0.00019499720515246525, "loss": 0.0161, "step": 383 }, { "epoch": 0.5757121439280359, "grad_norm": 0.189506645403256, "learning_rate": 0.0001949425477641904, "loss": 0.0385, "step": 384 }, { "epoch": 0.5772113943028486, "grad_norm": 0.20463849444352333, "learning_rate": 0.00019488760116444966, "loss": 0.0334, "step": 385 }, { "epoch": 0.5787106446776612, "grad_norm": 0.14154099501140546, "learning_rate": 0.00019483236552061994, "loss": 0.047, "step": 386 }, { "epoch": 0.5802098950524738, "grad_norm": 0.14873654543331344, "learning_rate": 0.0001947768410009586, "loss": 0.0367, "step": 387 }, { "epoch": 0.5817091454272864, "grad_norm": 0.12840825492818217, "learning_rate": 0.0001947210277746029, "loss": 0.0292, "step": 388 }, { "epoch": 0.5832083958020989, "grad_norm": 0.23903676877323182, "learning_rate": 0.00019466492601156966, "loss": 0.0379, "step": 389 }, { "epoch": 0.5847076461769115, "grad_norm": 0.15724874073609016, "learning_rate": 0.00019460853588275454, "loss": 0.0282, "step": 390 }, { "epoch": 0.5862068965517241, "grad_norm": 0.13173093891442925, "learning_rate": 0.0001945518575599317, "loss": 0.0316, "step": 391 }, { "epoch": 0.5877061469265368, "grad_norm": 0.28666760943594705, "learning_rate": 0.0001944948912157531, "loss": 0.0197, "step": 392 }, { "epoch": 0.5892053973013494, "grad_norm": 0.1832605963942488, "learning_rate": 0.00019443763702374812, "loss": 0.0266, "step": 393 }, { "epoch": 0.5907046476761619, "grad_norm": 0.13319197003396382, "learning_rate": 0.00019438009515832297, "loss": 0.0374, "step": 394 }, { "epoch": 0.5922038980509745, "grad_norm": 0.0937432546642976, "learning_rate": 0.0001943222657947601, "loss": 0.0302, "step": 395 }, { "epoch": 0.5937031484257871, "grad_norm": 0.0791127172795629, "learning_rate": 0.00019426414910921787, "loss": 0.0258, "step": 396 }, { "epoch": 0.5952023988005997, "grad_norm": 0.11621842629982067, "learning_rate": 0.00019420574527872968, "loss": 0.0224, "step": 397 }, { "epoch": 0.5967016491754122, "grad_norm": 0.18111252384023568, "learning_rate": 0.0001941470544812038, "loss": 0.0343, "step": 398 }, { "epoch": 0.5982008995502249, "grad_norm": 0.12668476963938463, "learning_rate": 0.00019408807689542257, "loss": 0.0329, "step": 399 }, { "epoch": 0.5997001499250375, "grad_norm": 0.08125543588426189, "learning_rate": 0.0001940288127010419, "loss": 0.0187, "step": 400 }, { "epoch": 0.6011994002998501, "grad_norm": 0.12705282798340953, "learning_rate": 0.00019396926207859084, "loss": 0.0208, "step": 401 }, { "epoch": 0.6026986506746627, "grad_norm": 0.09903543853284616, "learning_rate": 0.0001939094252094709, "loss": 0.0203, "step": 402 }, { "epoch": 0.6041979010494752, "grad_norm": 0.09458490298044545, "learning_rate": 0.0001938493022759556, "loss": 0.0354, "step": 403 }, { "epoch": 0.6056971514242878, "grad_norm": 0.13799640097610966, "learning_rate": 0.0001937888934611898, "loss": 0.0305, "step": 404 }, { "epoch": 0.6071964017991005, "grad_norm": 0.10327912926042261, "learning_rate": 0.00019372819894918915, "loss": 0.0193, "step": 405 }, { "epoch": 0.6086956521739131, "grad_norm": 0.12134046776724058, "learning_rate": 0.00019366721892483978, "loss": 0.0289, "step": 406 }, { "epoch": 0.6101949025487257, "grad_norm": 0.3625741631498206, "learning_rate": 0.00019360595357389735, "loss": 0.0618, "step": 407 }, { "epoch": 0.6116941529235382, "grad_norm": 0.16276585017367157, "learning_rate": 0.00019354440308298675, "loss": 0.033, "step": 408 }, { "epoch": 0.6131934032983508, "grad_norm": 0.1337810168549044, "learning_rate": 0.00019348256763960145, "loss": 0.0334, "step": 409 }, { "epoch": 0.6146926536731634, "grad_norm": 0.46538099581190734, "learning_rate": 0.00019342044743210295, "loss": 0.063, "step": 410 }, { "epoch": 0.616191904047976, "grad_norm": 0.1528443799832324, "learning_rate": 0.00019335804264972018, "loss": 0.0419, "step": 411 }, { "epoch": 0.6176911544227887, "grad_norm": 0.11360241608381239, "learning_rate": 0.00019329535348254893, "loss": 0.0279, "step": 412 }, { "epoch": 0.6191904047976012, "grad_norm": 0.10731049141357688, "learning_rate": 0.00019323238012155123, "loss": 0.027, "step": 413 }, { "epoch": 0.6206896551724138, "grad_norm": 0.1197495553442816, "learning_rate": 0.0001931691227585549, "loss": 0.034, "step": 414 }, { "epoch": 0.6221889055472264, "grad_norm": 0.15775016023781924, "learning_rate": 0.00019310558158625285, "loss": 0.0377, "step": 415 }, { "epoch": 0.623688155922039, "grad_norm": 0.11263624033418991, "learning_rate": 0.00019304175679820247, "loss": 0.0185, "step": 416 }, { "epoch": 0.6251874062968515, "grad_norm": 0.1306077390581805, "learning_rate": 0.00019297764858882514, "loss": 0.0207, "step": 417 }, { "epoch": 0.6266866566716641, "grad_norm": 0.09670200550131532, "learning_rate": 0.00019291325715340563, "loss": 0.0219, "step": 418 }, { "epoch": 0.6281859070464768, "grad_norm": 0.08461352549162458, "learning_rate": 0.00019284858268809137, "loss": 0.0305, "step": 419 }, { "epoch": 0.6296851574212894, "grad_norm": 0.1169472728087451, "learning_rate": 0.000192783625389892, "loss": 0.0351, "step": 420 }, { "epoch": 0.631184407796102, "grad_norm": 0.3408290216906546, "learning_rate": 0.00019271838545667876, "loss": 0.0312, "step": 421 }, { "epoch": 0.6326836581709145, "grad_norm": 0.09856486201168368, "learning_rate": 0.00019265286308718372, "loss": 0.023, "step": 422 }, { "epoch": 0.6341829085457271, "grad_norm": 0.11853262613762018, "learning_rate": 0.0001925870584809995, "loss": 0.0219, "step": 423 }, { "epoch": 0.6356821589205397, "grad_norm": 0.12277530671201359, "learning_rate": 0.00019252097183857823, "loss": 0.026, "step": 424 }, { "epoch": 0.6371814092953523, "grad_norm": 0.2205369312743012, "learning_rate": 0.00019245460336123134, "loss": 0.046, "step": 425 }, { "epoch": 0.638680659670165, "grad_norm": 0.10222524463832733, "learning_rate": 0.0001923879532511287, "loss": 0.0272, "step": 426 }, { "epoch": 0.6401799100449775, "grad_norm": 0.1881232227964226, "learning_rate": 0.00019232102171129811, "loss": 0.0337, "step": 427 }, { "epoch": 0.6416791604197901, "grad_norm": 0.2542609797286416, "learning_rate": 0.00019225380894562463, "loss": 0.0462, "step": 428 }, { "epoch": 0.6431784107946027, "grad_norm": 0.15567631088633233, "learning_rate": 0.00019218631515885006, "loss": 0.032, "step": 429 }, { "epoch": 0.6446776611694153, "grad_norm": 0.2672438162396934, "learning_rate": 0.00019211854055657215, "loss": 0.0477, "step": 430 }, { "epoch": 0.6461769115442278, "grad_norm": 0.1926377684668028, "learning_rate": 0.00019205048534524406, "loss": 0.0419, "step": 431 }, { "epoch": 0.6476761619190404, "grad_norm": 0.15284392892312818, "learning_rate": 0.00019198214973217378, "loss": 0.0398, "step": 432 }, { "epoch": 0.6491754122938531, "grad_norm": 0.11044006384885063, "learning_rate": 0.00019191353392552344, "loss": 0.0288, "step": 433 }, { "epoch": 0.6506746626686657, "grad_norm": 0.06779580847579911, "learning_rate": 0.00019184463813430873, "loss": 0.0191, "step": 434 }, { "epoch": 0.6521739130434783, "grad_norm": 0.11624950451575107, "learning_rate": 0.00019177546256839812, "loss": 0.0312, "step": 435 }, { "epoch": 0.6536731634182908, "grad_norm": 0.09177369083814481, "learning_rate": 0.0001917060074385124, "loss": 0.0303, "step": 436 }, { "epoch": 0.6551724137931034, "grad_norm": 0.11556801968896689, "learning_rate": 0.00019163627295622397, "loss": 0.0223, "step": 437 }, { "epoch": 0.656671664167916, "grad_norm": 0.09772753082703028, "learning_rate": 0.00019156625933395614, "loss": 0.0262, "step": 438 }, { "epoch": 0.6581709145427287, "grad_norm": 0.05846912894146479, "learning_rate": 0.0001914959667849825, "loss": 0.0157, "step": 439 }, { "epoch": 0.6596701649175413, "grad_norm": 0.6528373118339207, "learning_rate": 0.00019142539552342638, "loss": 0.0269, "step": 440 }, { "epoch": 0.6611694152923538, "grad_norm": 0.0927370129280204, "learning_rate": 0.0001913545457642601, "loss": 0.0202, "step": 441 }, { "epoch": 0.6626686656671664, "grad_norm": 0.1825329305030154, "learning_rate": 0.0001912834177233043, "loss": 0.0403, "step": 442 }, { "epoch": 0.664167916041979, "grad_norm": 0.1740152075571954, "learning_rate": 0.0001912120116172273, "loss": 0.0491, "step": 443 }, { "epoch": 0.6656671664167916, "grad_norm": 0.14268776420650997, "learning_rate": 0.00019114032766354453, "loss": 0.0375, "step": 444 }, { "epoch": 0.6671664167916042, "grad_norm": 0.10780160091521133, "learning_rate": 0.00019106836608061772, "loss": 0.0327, "step": 445 }, { "epoch": 0.6686656671664168, "grad_norm": 0.11310647498952432, "learning_rate": 0.00019099612708765434, "loss": 0.0299, "step": 446 }, { "epoch": 0.6701649175412294, "grad_norm": 0.13740431187575763, "learning_rate": 0.00019092361090470688, "loss": 0.0358, "step": 447 }, { "epoch": 0.671664167916042, "grad_norm": 0.18068397863290914, "learning_rate": 0.0001908508177526722, "loss": 0.0457, "step": 448 }, { "epoch": 0.6731634182908546, "grad_norm": 0.10147402022124917, "learning_rate": 0.00019077774785329087, "loss": 0.0245, "step": 449 }, { "epoch": 0.6746626686656672, "grad_norm": 0.17659307041157327, "learning_rate": 0.0001907044014291465, "loss": 0.0373, "step": 450 }, { "epoch": 0.6761619190404797, "grad_norm": 0.21933608649614408, "learning_rate": 0.000190630778703665, "loss": 0.0724, "step": 451 }, { "epoch": 0.6776611694152923, "grad_norm": 0.18695425514273356, "learning_rate": 0.00019055687990111398, "loss": 0.0433, "step": 452 }, { "epoch": 0.679160419790105, "grad_norm": 0.09732659308557562, "learning_rate": 0.00019048270524660196, "loss": 0.0362, "step": 453 }, { "epoch": 0.6806596701649176, "grad_norm": 0.08889244620347342, "learning_rate": 0.00019040825496607786, "loss": 0.0268, "step": 454 }, { "epoch": 0.6821589205397302, "grad_norm": 0.10073399312054154, "learning_rate": 0.0001903335292863301, "loss": 0.0256, "step": 455 }, { "epoch": 0.6836581709145427, "grad_norm": 0.17468190137144324, "learning_rate": 0.00019025852843498607, "loss": 0.0445, "step": 456 }, { "epoch": 0.6851574212893553, "grad_norm": 0.09001197181665302, "learning_rate": 0.0001901832526405114, "loss": 0.0151, "step": 457 }, { "epoch": 0.6866566716641679, "grad_norm": 0.13748748434736058, "learning_rate": 0.00019010770213220916, "loss": 0.0221, "step": 458 }, { "epoch": 0.6881559220389805, "grad_norm": 0.13324313496496254, "learning_rate": 0.00019003187714021938, "loss": 0.0394, "step": 459 }, { "epoch": 0.6896551724137931, "grad_norm": 0.18122876203058136, "learning_rate": 0.00018995577789551803, "loss": 0.0438, "step": 460 }, { "epoch": 0.6911544227886057, "grad_norm": 0.13914037277459534, "learning_rate": 0.0001898794046299167, "loss": 0.0325, "step": 461 }, { "epoch": 0.6926536731634183, "grad_norm": 0.1250826760287831, "learning_rate": 0.00018980275757606157, "loss": 0.0324, "step": 462 }, { "epoch": 0.6941529235382309, "grad_norm": 0.11773650752390566, "learning_rate": 0.00018972583696743285, "loss": 0.0298, "step": 463 }, { "epoch": 0.6956521739130435, "grad_norm": 0.20782604564608717, "learning_rate": 0.00018964864303834406, "loss": 0.0281, "step": 464 }, { "epoch": 0.697151424287856, "grad_norm": 0.09525179474823385, "learning_rate": 0.0001895711760239413, "loss": 0.0252, "step": 465 }, { "epoch": 0.6986506746626686, "grad_norm": 0.1240832012888324, "learning_rate": 0.00018949343616020252, "loss": 0.0231, "step": 466 }, { "epoch": 0.7001499250374813, "grad_norm": 0.4265753911477756, "learning_rate": 0.0001894154236839368, "loss": 0.0669, "step": 467 }, { "epoch": 0.7016491754122939, "grad_norm": 0.12476523044968328, "learning_rate": 0.00018933713883278376, "loss": 0.0273, "step": 468 }, { "epoch": 0.7031484257871065, "grad_norm": 0.14830273846419462, "learning_rate": 0.00018925858184521256, "loss": 0.0303, "step": 469 }, { "epoch": 0.704647676161919, "grad_norm": 0.12142464965135223, "learning_rate": 0.00018917975296052142, "loss": 0.0387, "step": 470 }, { "epoch": 0.7061469265367316, "grad_norm": 0.12543262855683845, "learning_rate": 0.0001891006524188368, "loss": 0.0251, "step": 471 }, { "epoch": 0.7076461769115442, "grad_norm": 0.19339819413209244, "learning_rate": 0.00018902128046111266, "loss": 0.0257, "step": 472 }, { "epoch": 0.7091454272863568, "grad_norm": 0.15350744131356941, "learning_rate": 0.00018894163732912977, "loss": 0.0278, "step": 473 }, { "epoch": 0.7106446776611695, "grad_norm": 0.23391665071790627, "learning_rate": 0.0001888617232654949, "loss": 0.0448, "step": 474 }, { "epoch": 0.712143928035982, "grad_norm": 0.10994980391292404, "learning_rate": 0.00018878153851364013, "loss": 0.0227, "step": 475 }, { "epoch": 0.7136431784107946, "grad_norm": 0.10665227644215766, "learning_rate": 0.00018870108331782217, "loss": 0.0281, "step": 476 }, { "epoch": 0.7151424287856072, "grad_norm": 0.14531593463069997, "learning_rate": 0.00018862035792312147, "loss": 0.0326, "step": 477 }, { "epoch": 0.7166416791604198, "grad_norm": 0.1182913114959111, "learning_rate": 0.0001885393625754416, "loss": 0.0188, "step": 478 }, { "epoch": 0.7181409295352323, "grad_norm": 0.07106508169769966, "learning_rate": 0.0001884580975215084, "loss": 0.0138, "step": 479 }, { "epoch": 0.719640179910045, "grad_norm": 0.07950185634897355, "learning_rate": 0.00018837656300886937, "loss": 0.0165, "step": 480 }, { "epoch": 0.7211394302848576, "grad_norm": 0.09129244683295917, "learning_rate": 0.00018829475928589271, "loss": 0.0167, "step": 481 }, { "epoch": 0.7226386806596702, "grad_norm": 0.12771843853475826, "learning_rate": 0.00018821268660176678, "loss": 0.022, "step": 482 }, { "epoch": 0.7241379310344828, "grad_norm": 0.13265893909135285, "learning_rate": 0.0001881303452064992, "loss": 0.0345, "step": 483 }, { "epoch": 0.7256371814092953, "grad_norm": 0.12918348541457214, "learning_rate": 0.0001880477353509162, "loss": 0.0372, "step": 484 }, { "epoch": 0.7271364317841079, "grad_norm": 0.09841614043671974, "learning_rate": 0.00018796485728666165, "loss": 0.0214, "step": 485 }, { "epoch": 0.7286356821589205, "grad_norm": 0.12613750913388452, "learning_rate": 0.00018788171126619653, "loss": 0.0272, "step": 486 }, { "epoch": 0.7301349325337332, "grad_norm": 0.10042055414075088, "learning_rate": 0.00018779829754279805, "loss": 0.0286, "step": 487 }, { "epoch": 0.7316341829085458, "grad_norm": 0.1665493001660504, "learning_rate": 0.00018771461637055888, "loss": 0.033, "step": 488 }, { "epoch": 0.7331334332833583, "grad_norm": 0.1179355108298341, "learning_rate": 0.00018763066800438636, "loss": 0.0463, "step": 489 }, { "epoch": 0.7346326836581709, "grad_norm": 0.10791732823530058, "learning_rate": 0.0001875464527000018, "loss": 0.0217, "step": 490 }, { "epoch": 0.7361319340329835, "grad_norm": 0.3669927380669368, "learning_rate": 0.00018746197071393958, "loss": 0.0457, "step": 491 }, { "epoch": 0.7376311844077961, "grad_norm": 0.1987097980139362, "learning_rate": 0.00018737722230354655, "loss": 0.0423, "step": 492 }, { "epoch": 0.7391304347826086, "grad_norm": 0.09040003852732766, "learning_rate": 0.00018729220772698097, "loss": 0.0288, "step": 493 }, { "epoch": 0.7406296851574213, "grad_norm": 0.20269069942355697, "learning_rate": 0.00018720692724321207, "loss": 0.0455, "step": 494 }, { "epoch": 0.7421289355322339, "grad_norm": 0.1372255943190446, "learning_rate": 0.00018712138111201895, "loss": 0.0216, "step": 495 }, { "epoch": 0.7436281859070465, "grad_norm": 0.23882692775642725, "learning_rate": 0.00018703556959398998, "loss": 0.0487, "step": 496 }, { "epoch": 0.7451274362818591, "grad_norm": 0.10854063954832008, "learning_rate": 0.0001869494929505219, "loss": 0.0222, "step": 497 }, { "epoch": 0.7466266866566716, "grad_norm": 0.09480222600585039, "learning_rate": 0.00018686315144381913, "loss": 0.0257, "step": 498 }, { "epoch": 0.7481259370314842, "grad_norm": 0.16193331192920116, "learning_rate": 0.00018677654533689287, "loss": 0.0329, "step": 499 }, { "epoch": 0.7496251874062968, "grad_norm": 0.08005281597137838, "learning_rate": 0.00018668967489356028, "loss": 0.0169, "step": 500 }, { "epoch": 0.7511244377811095, "grad_norm": 0.12389056426209798, "learning_rate": 0.00018660254037844388, "loss": 0.0366, "step": 501 }, { "epoch": 0.7526236881559221, "grad_norm": 0.11869938873623041, "learning_rate": 0.00018651514205697046, "loss": 0.0321, "step": 502 }, { "epoch": 0.7541229385307346, "grad_norm": 0.0930246262254623, "learning_rate": 0.0001864274801953705, "loss": 0.022, "step": 503 }, { "epoch": 0.7556221889055472, "grad_norm": 0.27274941895288046, "learning_rate": 0.00018633955506067718, "loss": 0.045, "step": 504 }, { "epoch": 0.7571214392803598, "grad_norm": 0.0726365498076487, "learning_rate": 0.00018625136692072575, "loss": 0.0129, "step": 505 }, { "epoch": 0.7586206896551724, "grad_norm": 0.1638000705309204, "learning_rate": 0.00018616291604415258, "loss": 0.0507, "step": 506 }, { "epoch": 0.760119940029985, "grad_norm": 0.10985380636797697, "learning_rate": 0.0001860742027003944, "loss": 0.0172, "step": 507 }, { "epoch": 0.7616191904047976, "grad_norm": 0.11572389017878003, "learning_rate": 0.00018598522715968736, "loss": 0.0359, "step": 508 }, { "epoch": 0.7631184407796102, "grad_norm": 0.1957062355312362, "learning_rate": 0.00018589598969306645, "loss": 0.0286, "step": 509 }, { "epoch": 0.7646176911544228, "grad_norm": 0.1797311903894293, "learning_rate": 0.00018580649057236447, "loss": 0.0422, "step": 510 }, { "epoch": 0.7661169415292354, "grad_norm": 0.10855777727429075, "learning_rate": 0.00018571673007021123, "loss": 0.0348, "step": 511 }, { "epoch": 0.767616191904048, "grad_norm": 0.09640794818380023, "learning_rate": 0.00018562670846003284, "loss": 0.0123, "step": 512 }, { "epoch": 0.7691154422788605, "grad_norm": 0.21945784704110113, "learning_rate": 0.00018553642601605068, "loss": 0.0456, "step": 513 }, { "epoch": 0.7706146926536732, "grad_norm": 0.11478499434793961, "learning_rate": 0.00018544588301328075, "loss": 0.0225, "step": 514 }, { "epoch": 0.7721139430284858, "grad_norm": 0.10420495356687196, "learning_rate": 0.00018535507972753274, "loss": 0.0276, "step": 515 }, { "epoch": 0.7736131934032984, "grad_norm": 0.06375546661931226, "learning_rate": 0.00018526401643540922, "loss": 0.0143, "step": 516 }, { "epoch": 0.775112443778111, "grad_norm": 0.10013660059835842, "learning_rate": 0.00018517269341430476, "loss": 0.0316, "step": 517 }, { "epoch": 0.7766116941529235, "grad_norm": 0.1525040242862086, "learning_rate": 0.00018508111094240514, "loss": 0.0531, "step": 518 }, { "epoch": 0.7781109445277361, "grad_norm": 0.08246402484447012, "learning_rate": 0.00018498926929868642, "loss": 0.0232, "step": 519 }, { "epoch": 0.7796101949025487, "grad_norm": 0.1147034669265538, "learning_rate": 0.00018489716876291415, "loss": 0.0282, "step": 520 }, { "epoch": 0.7811094452773614, "grad_norm": 0.12214564867021378, "learning_rate": 0.0001848048096156426, "loss": 0.0247, "step": 521 }, { "epoch": 0.782608695652174, "grad_norm": 0.06751036662935833, "learning_rate": 0.00018471219213821375, "loss": 0.0152, "step": 522 }, { "epoch": 0.7841079460269865, "grad_norm": 0.07585774273233134, "learning_rate": 0.00018461931661275643, "loss": 0.0207, "step": 523 }, { "epoch": 0.7856071964017991, "grad_norm": 0.10748840649212606, "learning_rate": 0.00018452618332218563, "loss": 0.0254, "step": 524 }, { "epoch": 0.7871064467766117, "grad_norm": 0.10237085474160945, "learning_rate": 0.00018443279255020152, "loss": 0.0329, "step": 525 }, { "epoch": 0.7886056971514243, "grad_norm": 0.08837009279940057, "learning_rate": 0.0001843391445812886, "loss": 0.0199, "step": 526 }, { "epoch": 0.7901049475262368, "grad_norm": 0.06271891827693078, "learning_rate": 0.00018424523970071477, "loss": 0.0183, "step": 527 }, { "epoch": 0.7916041979010495, "grad_norm": 0.13599180635432398, "learning_rate": 0.00018415107819453062, "loss": 0.0447, "step": 528 }, { "epoch": 0.7931034482758621, "grad_norm": 0.0836743427473769, "learning_rate": 0.00018405666034956844, "loss": 0.0175, "step": 529 }, { "epoch": 0.7946026986506747, "grad_norm": 0.11593648054445677, "learning_rate": 0.00018396198645344135, "loss": 0.0246, "step": 530 }, { "epoch": 0.7961019490254873, "grad_norm": 0.07719300987275271, "learning_rate": 0.00018386705679454242, "loss": 0.0301, "step": 531 }, { "epoch": 0.7976011994002998, "grad_norm": 0.09778571203259609, "learning_rate": 0.0001837718716620439, "loss": 0.0212, "step": 532 }, { "epoch": 0.7991004497751124, "grad_norm": 0.12595595693675168, "learning_rate": 0.00018367643134589617, "loss": 0.0326, "step": 533 }, { "epoch": 0.800599700149925, "grad_norm": 0.07156209728774574, "learning_rate": 0.00018358073613682706, "loss": 0.0165, "step": 534 }, { "epoch": 0.8020989505247377, "grad_norm": 0.09505001360014041, "learning_rate": 0.00018348478632634066, "loss": 0.0187, "step": 535 }, { "epoch": 0.8035982008995503, "grad_norm": 0.12058029362246589, "learning_rate": 0.00018338858220671682, "loss": 0.0282, "step": 536 }, { "epoch": 0.8050974512743628, "grad_norm": 0.17693735284669823, "learning_rate": 0.00018329212407100994, "loss": 0.0324, "step": 537 }, { "epoch": 0.8065967016491754, "grad_norm": 0.13029106214616978, "learning_rate": 0.00018319541221304827, "loss": 0.0209, "step": 538 }, { "epoch": 0.808095952023988, "grad_norm": 0.1162879265464023, "learning_rate": 0.00018309844692743283, "loss": 0.0207, "step": 539 }, { "epoch": 0.8095952023988006, "grad_norm": 0.20317081990377658, "learning_rate": 0.00018300122850953675, "loss": 0.0227, "step": 540 }, { "epoch": 0.8110944527736131, "grad_norm": 0.19653799373191536, "learning_rate": 0.00018290375725550417, "loss": 0.0281, "step": 541 }, { "epoch": 0.8125937031484258, "grad_norm": 0.11236149669329604, "learning_rate": 0.00018280603346224945, "loss": 0.0216, "step": 542 }, { "epoch": 0.8140929535232384, "grad_norm": 0.1687101239767433, "learning_rate": 0.00018270805742745617, "loss": 0.0405, "step": 543 }, { "epoch": 0.815592203898051, "grad_norm": 0.16013841907480525, "learning_rate": 0.00018260982944957638, "loss": 0.0429, "step": 544 }, { "epoch": 0.8170914542728636, "grad_norm": 0.12453527639861811, "learning_rate": 0.00018251134982782952, "loss": 0.0177, "step": 545 }, { "epoch": 0.8185907046476761, "grad_norm": 0.0913439448193188, "learning_rate": 0.00018241261886220154, "loss": 0.0201, "step": 546 }, { "epoch": 0.8200899550224887, "grad_norm": 0.10694343936153222, "learning_rate": 0.0001823136368534442, "loss": 0.0168, "step": 547 }, { "epoch": 0.8215892053973014, "grad_norm": 0.12873854233244145, "learning_rate": 0.00018221440410307374, "loss": 0.0182, "step": 548 }, { "epoch": 0.823088455772114, "grad_norm": 0.12248324617863353, "learning_rate": 0.00018211492091337042, "loss": 0.0245, "step": 549 }, { "epoch": 0.8245877061469266, "grad_norm": 0.10118859098810772, "learning_rate": 0.00018201518758737724, "loss": 0.031, "step": 550 }, { "epoch": 0.8260869565217391, "grad_norm": 0.15858954636984837, "learning_rate": 0.0001819152044288992, "loss": 0.053, "step": 551 }, { "epoch": 0.8275862068965517, "grad_norm": 0.099734088957312, "learning_rate": 0.00018181497174250236, "loss": 0.0217, "step": 552 }, { "epoch": 0.8290854572713643, "grad_norm": 0.17676306995199245, "learning_rate": 0.00018171448983351284, "loss": 0.039, "step": 553 }, { "epoch": 0.8305847076461769, "grad_norm": 0.21198277345625827, "learning_rate": 0.00018161375900801604, "loss": 0.0233, "step": 554 }, { "epoch": 0.8320839580209896, "grad_norm": 0.10508077305912032, "learning_rate": 0.00018151277957285543, "loss": 0.0249, "step": 555 }, { "epoch": 0.8335832083958021, "grad_norm": 0.15839124476109767, "learning_rate": 0.00018141155183563193, "loss": 0.019, "step": 556 }, { "epoch": 0.8350824587706147, "grad_norm": 0.11063381503992227, "learning_rate": 0.00018131007610470276, "loss": 0.0284, "step": 557 }, { "epoch": 0.8365817091454273, "grad_norm": 0.09616652397576733, "learning_rate": 0.00018120835268918063, "loss": 0.0234, "step": 558 }, { "epoch": 0.8380809595202399, "grad_norm": 0.21337896061350242, "learning_rate": 0.00018110638189893267, "loss": 0.0221, "step": 559 }, { "epoch": 0.8395802098950524, "grad_norm": 0.11835080605666147, "learning_rate": 0.00018100416404457961, "loss": 0.0162, "step": 560 }, { "epoch": 0.841079460269865, "grad_norm": 0.0871110273154611, "learning_rate": 0.00018090169943749476, "loss": 0.0188, "step": 561 }, { "epoch": 0.8425787106446777, "grad_norm": 0.15991292108921587, "learning_rate": 0.00018079898838980305, "loss": 0.0293, "step": 562 }, { "epoch": 0.8440779610194903, "grad_norm": 0.10275028482175932, "learning_rate": 0.00018069603121438022, "loss": 0.0224, "step": 563 }, { "epoch": 0.8455772113943029, "grad_norm": 0.2277974036113205, "learning_rate": 0.00018059282822485158, "loss": 0.0258, "step": 564 }, { "epoch": 0.8470764617691154, "grad_norm": 0.10784532218963591, "learning_rate": 0.0001804893797355914, "loss": 0.0256, "step": 565 }, { "epoch": 0.848575712143928, "grad_norm": 0.13579456849315263, "learning_rate": 0.00018038568606172173, "loss": 0.0313, "step": 566 }, { "epoch": 0.8500749625187406, "grad_norm": 0.1139195289086778, "learning_rate": 0.00018028174751911146, "loss": 0.035, "step": 567 }, { "epoch": 0.8515742128935532, "grad_norm": 0.1437361790270228, "learning_rate": 0.0001801775644243754, "loss": 0.0246, "step": 568 }, { "epoch": 0.8530734632683659, "grad_norm": 0.1315744813308483, "learning_rate": 0.00018007313709487334, "loss": 0.0273, "step": 569 }, { "epoch": 0.8545727136431784, "grad_norm": 0.14726220554215622, "learning_rate": 0.00017996846584870908, "loss": 0.0169, "step": 570 }, { "epoch": 0.856071964017991, "grad_norm": 0.06770252618778136, "learning_rate": 0.00017986355100472928, "loss": 0.014, "step": 571 }, { "epoch": 0.8575712143928036, "grad_norm": 0.19654688986822602, "learning_rate": 0.00017975839288252287, "loss": 0.0524, "step": 572 }, { "epoch": 0.8590704647676162, "grad_norm": 0.11188521951103549, "learning_rate": 0.00017965299180241963, "loss": 0.0223, "step": 573 }, { "epoch": 0.8605697151424287, "grad_norm": 0.16155836256152575, "learning_rate": 0.00017954734808548958, "loss": 0.0214, "step": 574 }, { "epoch": 0.8620689655172413, "grad_norm": 0.09669657753074787, "learning_rate": 0.00017944146205354182, "loss": 0.0226, "step": 575 }, { "epoch": 0.863568215892054, "grad_norm": 0.07035277148350057, "learning_rate": 0.00017933533402912354, "loss": 0.0184, "step": 576 }, { "epoch": 0.8650674662668666, "grad_norm": 0.10734729487368178, "learning_rate": 0.00017922896433551907, "loss": 0.0227, "step": 577 }, { "epoch": 0.8665667166416792, "grad_norm": 0.12745410681963906, "learning_rate": 0.00017912235329674902, "loss": 0.0254, "step": 578 }, { "epoch": 0.8680659670164917, "grad_norm": 0.08587193689014985, "learning_rate": 0.00017901550123756906, "loss": 0.0201, "step": 579 }, { "epoch": 0.8695652173913043, "grad_norm": 0.08573667833277826, "learning_rate": 0.00017890840848346908, "loss": 0.0161, "step": 580 }, { "epoch": 0.8710644677661169, "grad_norm": 0.24128108163978015, "learning_rate": 0.00017880107536067218, "loss": 0.0317, "step": 581 }, { "epoch": 0.8725637181409296, "grad_norm": 0.11234634770019401, "learning_rate": 0.00017869350219613375, "loss": 0.0379, "step": 582 }, { "epoch": 0.8740629685157422, "grad_norm": 0.1786629143096894, "learning_rate": 0.0001785856893175402, "loss": 0.0265, "step": 583 }, { "epoch": 0.8755622188905547, "grad_norm": 0.17384457039894063, "learning_rate": 0.0001784776370533083, "loss": 0.0374, "step": 584 }, { "epoch": 0.8770614692653673, "grad_norm": 0.1507710793497251, "learning_rate": 0.000178369345732584, "loss": 0.027, "step": 585 }, { "epoch": 0.8785607196401799, "grad_norm": 0.09140520812025159, "learning_rate": 0.0001782608156852414, "loss": 0.0211, "step": 586 }, { "epoch": 0.8800599700149925, "grad_norm": 0.301767128796161, "learning_rate": 0.00017815204724188187, "loss": 0.0315, "step": 587 }, { "epoch": 0.881559220389805, "grad_norm": 0.16101936926022856, "learning_rate": 0.000178043040733833, "loss": 0.0306, "step": 588 }, { "epoch": 0.8830584707646177, "grad_norm": 0.1656349117966397, "learning_rate": 0.00017793379649314744, "loss": 0.045, "step": 589 }, { "epoch": 0.8845577211394303, "grad_norm": 0.2112962853210967, "learning_rate": 0.00017782431485260212, "loss": 0.0579, "step": 590 }, { "epoch": 0.8860569715142429, "grad_norm": 0.1372483465065148, "learning_rate": 0.0001777145961456971, "loss": 0.0376, "step": 591 }, { "epoch": 0.8875562218890555, "grad_norm": 0.10696584922742608, "learning_rate": 0.0001776046407066546, "loss": 0.0213, "step": 592 }, { "epoch": 0.889055472263868, "grad_norm": 0.1450728563008204, "learning_rate": 0.00017749444887041799, "loss": 0.0238, "step": 593 }, { "epoch": 0.8905547226386806, "grad_norm": 0.11601926896640119, "learning_rate": 0.00017738402097265064, "loss": 0.0333, "step": 594 }, { "epoch": 0.8920539730134932, "grad_norm": 0.10700704866001351, "learning_rate": 0.00017727335734973512, "loss": 0.0287, "step": 595 }, { "epoch": 0.8935532233883059, "grad_norm": 0.13363561466349677, "learning_rate": 0.00017716245833877201, "loss": 0.0256, "step": 596 }, { "epoch": 0.8950524737631185, "grad_norm": 0.0667796736223053, "learning_rate": 0.00017705132427757895, "loss": 0.0178, "step": 597 }, { "epoch": 0.896551724137931, "grad_norm": 0.10641527946736849, "learning_rate": 0.0001769399555046895, "loss": 0.0142, "step": 598 }, { "epoch": 0.8980509745127436, "grad_norm": 0.08884031549727572, "learning_rate": 0.00017682835235935236, "loss": 0.0217, "step": 599 }, { "epoch": 0.8995502248875562, "grad_norm": 0.30418982864519956, "learning_rate": 0.00017671651518153, "loss": 0.0576, "step": 600 }, { "epoch": 0.9010494752623688, "grad_norm": 0.13881219505976333, "learning_rate": 0.0001766044443118978, "loss": 0.0315, "step": 601 }, { "epoch": 0.9025487256371814, "grad_norm": 0.1076877239847274, "learning_rate": 0.0001764921400918432, "loss": 0.0275, "step": 602 }, { "epoch": 0.904047976011994, "grad_norm": 0.15154212814340556, "learning_rate": 0.00017637960286346425, "loss": 0.0334, "step": 603 }, { "epoch": 0.9055472263868066, "grad_norm": 0.07641353287413101, "learning_rate": 0.00017626683296956882, "loss": 0.0153, "step": 604 }, { "epoch": 0.9070464767616192, "grad_norm": 0.11340124769027278, "learning_rate": 0.0001761538307536737, "loss": 0.0252, "step": 605 }, { "epoch": 0.9085457271364318, "grad_norm": 0.08383585004456805, "learning_rate": 0.0001760405965600031, "loss": 0.0181, "step": 606 }, { "epoch": 0.9100449775112444, "grad_norm": 0.1960282699023016, "learning_rate": 0.00017592713073348807, "loss": 0.0236, "step": 607 }, { "epoch": 0.9115442278860569, "grad_norm": 0.12553047330249875, "learning_rate": 0.00017581343361976524, "loss": 0.0193, "step": 608 }, { "epoch": 0.9130434782608695, "grad_norm": 0.07993908623746622, "learning_rate": 0.00017569950556517566, "loss": 0.0118, "step": 609 }, { "epoch": 0.9145427286356822, "grad_norm": 0.07644624754596607, "learning_rate": 0.00017558534691676397, "loss": 0.0178, "step": 610 }, { "epoch": 0.9160419790104948, "grad_norm": 0.06474378879594389, "learning_rate": 0.00017547095802227723, "loss": 0.0134, "step": 611 }, { "epoch": 0.9175412293853074, "grad_norm": 0.24320103822426103, "learning_rate": 0.0001753563392301638, "loss": 0.0358, "step": 612 }, { "epoch": 0.9190404797601199, "grad_norm": 0.21606781623298388, "learning_rate": 0.00017524149088957245, "loss": 0.018, "step": 613 }, { "epoch": 0.9205397301349325, "grad_norm": 0.13724300632851888, "learning_rate": 0.00017512641335035113, "loss": 0.0239, "step": 614 }, { "epoch": 0.9220389805097451, "grad_norm": 0.24490694217869374, "learning_rate": 0.00017501110696304596, "loss": 0.0457, "step": 615 }, { "epoch": 0.9235382308845578, "grad_norm": 0.09583105854859192, "learning_rate": 0.00017489557207890023, "loss": 0.0344, "step": 616 }, { "epoch": 0.9250374812593704, "grad_norm": 0.11543373352251945, "learning_rate": 0.0001747798090498532, "loss": 0.0215, "step": 617 }, { "epoch": 0.9265367316341829, "grad_norm": 0.2650648107879051, "learning_rate": 0.00017466381822853915, "loss": 0.0417, "step": 618 }, { "epoch": 0.9280359820089955, "grad_norm": 0.15994404613414645, "learning_rate": 0.00017454759996828623, "loss": 0.0295, "step": 619 }, { "epoch": 0.9295352323838081, "grad_norm": 0.12513579686861107, "learning_rate": 0.0001744311546231154, "loss": 0.0348, "step": 620 }, { "epoch": 0.9310344827586207, "grad_norm": 0.09930949830951118, "learning_rate": 0.00017431448254773944, "loss": 0.0237, "step": 621 }, { "epoch": 0.9325337331334332, "grad_norm": 0.2202645779778369, "learning_rate": 0.00017419758409756164, "loss": 0.0199, "step": 622 }, { "epoch": 0.9340329835082459, "grad_norm": 0.12490401554820374, "learning_rate": 0.000174080459628675, "loss": 0.0367, "step": 623 }, { "epoch": 0.9355322338830585, "grad_norm": 0.1425795882080626, "learning_rate": 0.000173963109497861, "loss": 0.0256, "step": 624 }, { "epoch": 0.9370314842578711, "grad_norm": 0.12319971348321294, "learning_rate": 0.00017384553406258842, "loss": 0.029, "step": 625 }, { "epoch": 0.9385307346326837, "grad_norm": 0.19643269683415326, "learning_rate": 0.0001737277336810124, "loss": 0.052, "step": 626 }, { "epoch": 0.9400299850074962, "grad_norm": 0.12658722225852667, "learning_rate": 0.00017360970871197346, "loss": 0.0212, "step": 627 }, { "epoch": 0.9415292353823088, "grad_norm": 0.19280146331288842, "learning_rate": 0.000173491459514996, "loss": 0.0405, "step": 628 }, { "epoch": 0.9430284857571214, "grad_norm": 0.1201798355899327, "learning_rate": 0.00017337298645028764, "loss": 0.0266, "step": 629 }, { "epoch": 0.9445277361319341, "grad_norm": 0.06066065563687962, "learning_rate": 0.0001732542898787379, "loss": 0.0167, "step": 630 }, { "epoch": 0.9460269865067467, "grad_norm": 0.09652060635836711, "learning_rate": 0.00017313537016191706, "loss": 0.02, "step": 631 }, { "epoch": 0.9475262368815592, "grad_norm": 0.06863279756680676, "learning_rate": 0.00017301622766207527, "loss": 0.0155, "step": 632 }, { "epoch": 0.9490254872563718, "grad_norm": 0.12042412187593952, "learning_rate": 0.00017289686274214118, "loss": 0.0206, "step": 633 }, { "epoch": 0.9505247376311844, "grad_norm": 0.21939496082438123, "learning_rate": 0.00017277727576572107, "loss": 0.0213, "step": 634 }, { "epoch": 0.952023988005997, "grad_norm": 0.12191851364600712, "learning_rate": 0.0001726574670970976, "loss": 0.0305, "step": 635 }, { "epoch": 0.9535232383808095, "grad_norm": 0.27478912515600534, "learning_rate": 0.00017253743710122875, "loss": 0.0283, "step": 636 }, { "epoch": 0.9550224887556222, "grad_norm": 0.1281513047579704, "learning_rate": 0.00017241718614374678, "loss": 0.04, "step": 637 }, { "epoch": 0.9565217391304348, "grad_norm": 0.1266027647091267, "learning_rate": 0.00017229671459095683, "loss": 0.0376, "step": 638 }, { "epoch": 0.9580209895052474, "grad_norm": 0.16017298063734342, "learning_rate": 0.00017217602280983623, "loss": 0.0309, "step": 639 }, { "epoch": 0.95952023988006, "grad_norm": 0.12020816231782688, "learning_rate": 0.00017205511116803306, "loss": 0.0256, "step": 640 }, { "epoch": 0.9610194902548725, "grad_norm": 0.06023210487373309, "learning_rate": 0.0001719339800338651, "loss": 0.0159, "step": 641 }, { "epoch": 0.9625187406296851, "grad_norm": 0.12858383075069013, "learning_rate": 0.00017181262977631888, "loss": 0.0281, "step": 642 }, { "epoch": 0.9640179910044977, "grad_norm": 0.11429779029405647, "learning_rate": 0.0001716910607650483, "loss": 0.0284, "step": 643 }, { "epoch": 0.9655172413793104, "grad_norm": 0.1072273139514921, "learning_rate": 0.0001715692733703736, "loss": 0.027, "step": 644 }, { "epoch": 0.967016491754123, "grad_norm": 0.20115614173420412, "learning_rate": 0.00017144726796328034, "loss": 0.0259, "step": 645 }, { "epoch": 0.9685157421289355, "grad_norm": 0.06623043636689804, "learning_rate": 0.00017132504491541818, "loss": 0.0185, "step": 646 }, { "epoch": 0.9700149925037481, "grad_norm": 0.20340594306656323, "learning_rate": 0.00017120260459909967, "loss": 0.0286, "step": 647 }, { "epoch": 0.9715142428785607, "grad_norm": 0.0772642579410454, "learning_rate": 0.00017107994738729926, "loss": 0.0132, "step": 648 }, { "epoch": 0.9730134932533733, "grad_norm": 0.08015487685028685, "learning_rate": 0.0001709570736536521, "loss": 0.0233, "step": 649 }, { "epoch": 0.974512743628186, "grad_norm": 0.1661406803223815, "learning_rate": 0.0001708339837724529, "loss": 0.0302, "step": 650 }, { "epoch": 0.9760119940029985, "grad_norm": 0.1458374578095082, "learning_rate": 0.00017071067811865476, "loss": 0.033, "step": 651 }, { "epoch": 0.9775112443778111, "grad_norm": 0.07798257623963753, "learning_rate": 0.0001705871570678681, "loss": 0.0128, "step": 652 }, { "epoch": 0.9790104947526237, "grad_norm": 0.048418288434207646, "learning_rate": 0.00017046342099635948, "loss": 0.0123, "step": 653 }, { "epoch": 0.9805097451274363, "grad_norm": 0.13431687956769212, "learning_rate": 0.00017033947028105039, "loss": 0.0251, "step": 654 }, { "epoch": 0.9820089955022488, "grad_norm": 0.08226067771753236, "learning_rate": 0.00017021530529951625, "loss": 0.021, "step": 655 }, { "epoch": 0.9835082458770614, "grad_norm": 0.1462577267853189, "learning_rate": 0.0001700909264299851, "loss": 0.0421, "step": 656 }, { "epoch": 0.9850074962518741, "grad_norm": 0.3083259142195758, "learning_rate": 0.00016996633405133655, "loss": 0.0494, "step": 657 }, { "epoch": 0.9865067466266867, "grad_norm": 0.18522845664082035, "learning_rate": 0.0001698415285431006, "loss": 0.0363, "step": 658 }, { "epoch": 0.9880059970014993, "grad_norm": 0.09484537072137021, "learning_rate": 0.00016971651028545648, "loss": 0.0281, "step": 659 }, { "epoch": 0.9895052473763118, "grad_norm": 0.23884894163402523, "learning_rate": 0.00016959127965923142, "loss": 0.0456, "step": 660 }, { "epoch": 0.9910044977511244, "grad_norm": 0.1779125846943612, "learning_rate": 0.00016946583704589973, "loss": 0.0422, "step": 661 }, { "epoch": 0.992503748125937, "grad_norm": 0.10266626269337822, "learning_rate": 0.0001693401828275813, "loss": 0.0203, "step": 662 }, { "epoch": 0.9940029985007496, "grad_norm": 0.10277193051633117, "learning_rate": 0.0001692143173870407, "loss": 0.0213, "step": 663 }, { "epoch": 0.9955022488755623, "grad_norm": 0.11463949080930057, "learning_rate": 0.00016908824110768584, "loss": 0.0339, "step": 664 }, { "epoch": 0.9970014992503748, "grad_norm": 0.13859328494189713, "learning_rate": 0.000168961954373567, "loss": 0.023, "step": 665 }, { "epoch": 0.9985007496251874, "grad_norm": 0.10931098650164782, "learning_rate": 0.0001688354575693754, "loss": 0.022, "step": 666 }, { "epoch": 1.0, "grad_norm": 0.1521638053539932, "learning_rate": 0.0001687087510804423, "loss": 0.032, "step": 667 }, { "epoch": 1.0, "eval_loss": 0.03065001219511032, "eval_runtime": 483.0726, "eval_samples_per_second": 21.479, "eval_steps_per_second": 1.343, "step": 667 }, { "epoch": 1.0014992503748126, "grad_norm": 0.10472517028799198, "learning_rate": 0.00016858183529273765, "loss": 0.0194, "step": 668 }, { "epoch": 1.0029985007496252, "grad_norm": 0.07086065655028383, "learning_rate": 0.00016845471059286887, "loss": 0.0169, "step": 669 }, { "epoch": 1.0044977511244377, "grad_norm": 0.056470801319977317, "learning_rate": 0.00016832737736807994, "loss": 0.0178, "step": 670 }, { "epoch": 1.0059970014992503, "grad_norm": 0.21736541605778517, "learning_rate": 0.00016819983600624986, "loss": 0.0204, "step": 671 }, { "epoch": 1.0074962518740629, "grad_norm": 0.08048210993127422, "learning_rate": 0.0001680720868958918, "loss": 0.0152, "step": 672 }, { "epoch": 1.0089955022488755, "grad_norm": 0.13724751027412635, "learning_rate": 0.00016794413042615168, "loss": 0.025, "step": 673 }, { "epoch": 1.0104947526236883, "grad_norm": 0.1299998829933712, "learning_rate": 0.0001678159669868071, "loss": 0.0233, "step": 674 }, { "epoch": 1.0119940029985008, "grad_norm": 0.08324573169820197, "learning_rate": 0.00016768759696826608, "loss": 0.0161, "step": 675 }, { "epoch": 1.0134932533733134, "grad_norm": 0.24760800404844965, "learning_rate": 0.00016755902076156604, "loss": 0.0267, "step": 676 }, { "epoch": 1.014992503748126, "grad_norm": 0.057948619616265976, "learning_rate": 0.00016743023875837233, "loss": 0.0083, "step": 677 }, { "epoch": 1.0164917541229386, "grad_norm": 0.09704964114233473, "learning_rate": 0.00016730125135097735, "loss": 0.0111, "step": 678 }, { "epoch": 1.0179910044977512, "grad_norm": 0.11235791780825113, "learning_rate": 0.00016717205893229903, "loss": 0.0196, "step": 679 }, { "epoch": 1.0194902548725637, "grad_norm": 0.14212568674303996, "learning_rate": 0.0001670426618958799, "loss": 0.0179, "step": 680 }, { "epoch": 1.0209895052473763, "grad_norm": 0.13458409942428412, "learning_rate": 0.00016691306063588583, "loss": 0.015, "step": 681 }, { "epoch": 1.0224887556221889, "grad_norm": 0.10078342958992698, "learning_rate": 0.00016678325554710468, "loss": 0.0145, "step": 682 }, { "epoch": 1.0239880059970015, "grad_norm": 0.2163179066668823, "learning_rate": 0.00016665324702494524, "loss": 0.0352, "step": 683 }, { "epoch": 1.025487256371814, "grad_norm": 0.09314880164627648, "learning_rate": 0.00016652303546543608, "loss": 0.0118, "step": 684 }, { "epoch": 1.0269865067466266, "grad_norm": 0.12711682267879498, "learning_rate": 0.00016639262126522418, "loss": 0.0187, "step": 685 }, { "epoch": 1.0284857571214392, "grad_norm": 0.10899305518463998, "learning_rate": 0.00016626200482157378, "loss": 0.0295, "step": 686 }, { "epoch": 1.0299850074962518, "grad_norm": 0.08551960256639587, "learning_rate": 0.00016613118653236518, "loss": 0.0129, "step": 687 }, { "epoch": 1.0314842578710646, "grad_norm": 0.08800753901836143, "learning_rate": 0.0001660001667960937, "loss": 0.0138, "step": 688 }, { "epoch": 1.0329835082458771, "grad_norm": 0.052385346475073355, "learning_rate": 0.00016586894601186805, "loss": 0.0083, "step": 689 }, { "epoch": 1.0344827586206897, "grad_norm": 0.07054205702889728, "learning_rate": 0.0001657375245794096, "loss": 0.0124, "step": 690 }, { "epoch": 1.0359820089955023, "grad_norm": 0.09975501615825792, "learning_rate": 0.00016560590289905073, "loss": 0.0118, "step": 691 }, { "epoch": 1.0374812593703149, "grad_norm": 0.14047961341536686, "learning_rate": 0.00016547408137173396, "loss": 0.0099, "step": 692 }, { "epoch": 1.0389805097451275, "grad_norm": 0.1189003861776989, "learning_rate": 0.00016534206039901057, "loss": 0.0262, "step": 693 }, { "epoch": 1.04047976011994, "grad_norm": 0.11248270963982432, "learning_rate": 0.00016520984038303924, "loss": 0.0178, "step": 694 }, { "epoch": 1.0419790104947526, "grad_norm": 0.13888415150865596, "learning_rate": 0.0001650774217265851, "loss": 0.0225, "step": 695 }, { "epoch": 1.0434782608695652, "grad_norm": 0.11070223453464198, "learning_rate": 0.00016494480483301836, "loss": 0.0112, "step": 696 }, { "epoch": 1.0449775112443778, "grad_norm": 0.1551674425930212, "learning_rate": 0.0001648119901063131, "loss": 0.0193, "step": 697 }, { "epoch": 1.0464767616191903, "grad_norm": 0.13273313189669741, "learning_rate": 0.000164678977951046, "loss": 0.0283, "step": 698 }, { "epoch": 1.047976011994003, "grad_norm": 0.11544709997166336, "learning_rate": 0.00016454576877239507, "loss": 0.0174, "step": 699 }, { "epoch": 1.0494752623688155, "grad_norm": 0.1253059382700822, "learning_rate": 0.00016441236297613866, "loss": 0.0095, "step": 700 }, { "epoch": 1.050974512743628, "grad_norm": 0.231346802938159, "learning_rate": 0.00016427876096865394, "loss": 0.0127, "step": 701 }, { "epoch": 1.0524737631184409, "grad_norm": 0.09463184955976486, "learning_rate": 0.00016414496315691581, "loss": 0.0132, "step": 702 }, { "epoch": 1.0539730134932535, "grad_norm": 0.10175851797876952, "learning_rate": 0.00016401096994849557, "loss": 0.0165, "step": 703 }, { "epoch": 1.055472263868066, "grad_norm": 0.08269715474941393, "learning_rate": 0.00016387678175155978, "loss": 0.022, "step": 704 }, { "epoch": 1.0569715142428786, "grad_norm": 0.0859139526446948, "learning_rate": 0.000163742398974869, "loss": 0.0128, "step": 705 }, { "epoch": 1.0584707646176912, "grad_norm": 0.1135404099233105, "learning_rate": 0.0001636078220277764, "loss": 0.0183, "step": 706 }, { "epoch": 1.0599700149925038, "grad_norm": 0.09275250705549713, "learning_rate": 0.00016347305132022677, "loss": 0.0232, "step": 707 }, { "epoch": 1.0614692653673163, "grad_norm": 0.11572250326059627, "learning_rate": 0.000163338087262755, "loss": 0.0824, "step": 708 }, { "epoch": 1.062968515742129, "grad_norm": 0.15876584912162692, "learning_rate": 0.0001632029302664851, "loss": 0.0239, "step": 709 }, { "epoch": 1.0644677661169415, "grad_norm": 0.09255809526179942, "learning_rate": 0.00016306758074312864, "loss": 0.0154, "step": 710 }, { "epoch": 1.065967016491754, "grad_norm": 0.11390783895932384, "learning_rate": 0.00016293203910498376, "loss": 0.0143, "step": 711 }, { "epoch": 1.0674662668665666, "grad_norm": 0.23974840043852813, "learning_rate": 0.00016279630576493382, "loss": 0.0171, "step": 712 }, { "epoch": 1.0689655172413792, "grad_norm": 0.17086459332122525, "learning_rate": 0.00016266038113644607, "loss": 0.0244, "step": 713 }, { "epoch": 1.0704647676161918, "grad_norm": 0.05971167811590956, "learning_rate": 0.00016252426563357055, "loss": 0.0055, "step": 714 }, { "epoch": 1.0719640179910046, "grad_norm": 0.0964739917335777, "learning_rate": 0.00016238795967093864, "loss": 0.0185, "step": 715 }, { "epoch": 1.0734632683658172, "grad_norm": 0.10293996743254338, "learning_rate": 0.00016225146366376198, "loss": 0.0176, "step": 716 }, { "epoch": 1.0749625187406298, "grad_norm": 0.08690427481589705, "learning_rate": 0.00016211477802783103, "loss": 0.0123, "step": 717 }, { "epoch": 1.0764617691154423, "grad_norm": 0.08500732276952126, "learning_rate": 0.00016197790317951403, "loss": 0.0113, "step": 718 }, { "epoch": 1.077961019490255, "grad_norm": 0.11613707922898552, "learning_rate": 0.0001618408395357554, "loss": 0.0184, "step": 719 }, { "epoch": 1.0794602698650675, "grad_norm": 0.2060611324342325, "learning_rate": 0.00016170358751407487, "loss": 0.0442, "step": 720 }, { "epoch": 1.08095952023988, "grad_norm": 0.11248886768899795, "learning_rate": 0.0001615661475325658, "loss": 0.0215, "step": 721 }, { "epoch": 1.0824587706146926, "grad_norm": 0.17400298698519548, "learning_rate": 0.00016142852000989433, "loss": 0.0347, "step": 722 }, { "epoch": 1.0839580209895052, "grad_norm": 0.07815925376493392, "learning_rate": 0.00016129070536529766, "loss": 0.0106, "step": 723 }, { "epoch": 1.0854572713643178, "grad_norm": 0.11757645186998897, "learning_rate": 0.00016115270401858314, "loss": 0.0202, "step": 724 }, { "epoch": 1.0869565217391304, "grad_norm": 0.12060617886999422, "learning_rate": 0.0001610145163901268, "loss": 0.0133, "step": 725 }, { "epoch": 1.088455772113943, "grad_norm": 0.07666228061263336, "learning_rate": 0.00016087614290087208, "loss": 0.0114, "step": 726 }, { "epoch": 1.0899550224887555, "grad_norm": 0.15028068516880796, "learning_rate": 0.00016073758397232868, "loss": 0.0224, "step": 727 }, { "epoch": 1.0914542728635683, "grad_norm": 0.07317261491835224, "learning_rate": 0.0001605988400265711, "loss": 0.014, "step": 728 }, { "epoch": 1.092953523238381, "grad_norm": 0.08354275449747522, "learning_rate": 0.0001604599114862375, "loss": 0.0233, "step": 729 }, { "epoch": 1.0944527736131935, "grad_norm": 0.07542214763546543, "learning_rate": 0.00016032079877452825, "loss": 0.0121, "step": 730 }, { "epoch": 1.095952023988006, "grad_norm": 0.0866470368358821, "learning_rate": 0.00016018150231520486, "loss": 0.0142, "step": 731 }, { "epoch": 1.0974512743628186, "grad_norm": 0.10167617929252423, "learning_rate": 0.00016004202253258842, "loss": 0.0193, "step": 732 }, { "epoch": 1.0989505247376312, "grad_norm": 0.10275133468863129, "learning_rate": 0.0001599023598515586, "loss": 0.013, "step": 733 }, { "epoch": 1.1004497751124438, "grad_norm": 0.11247245630083348, "learning_rate": 0.00015976251469755214, "loss": 0.0171, "step": 734 }, { "epoch": 1.1019490254872564, "grad_norm": 0.16757428374213076, "learning_rate": 0.0001596224874965616, "loss": 0.0238, "step": 735 }, { "epoch": 1.103448275862069, "grad_norm": 0.09060490513425551, "learning_rate": 0.00015948227867513415, "loss": 0.0117, "step": 736 }, { "epoch": 1.1049475262368815, "grad_norm": 0.0958942229372783, "learning_rate": 0.00015934188866037016, "loss": 0.0118, "step": 737 }, { "epoch": 1.106446776611694, "grad_norm": 0.11408221833035843, "learning_rate": 0.00015920131787992197, "loss": 0.0181, "step": 738 }, { "epoch": 1.1079460269865067, "grad_norm": 0.08779211321518267, "learning_rate": 0.00015906056676199255, "loss": 0.0163, "step": 739 }, { "epoch": 1.1094452773613193, "grad_norm": 0.08339161981475392, "learning_rate": 0.0001589196357353342, "loss": 0.0132, "step": 740 }, { "epoch": 1.1109445277361318, "grad_norm": 0.1766144838591971, "learning_rate": 0.00015877852522924732, "loss": 0.0186, "step": 741 }, { "epoch": 1.1124437781109444, "grad_norm": 0.17475491667530327, "learning_rate": 0.00015863723567357892, "loss": 0.0464, "step": 742 }, { "epoch": 1.1139430284857572, "grad_norm": 0.11280224594809592, "learning_rate": 0.00015849576749872157, "loss": 0.0299, "step": 743 }, { "epoch": 1.1154422788605698, "grad_norm": 0.06496065765413261, "learning_rate": 0.00015835412113561175, "loss": 0.0086, "step": 744 }, { "epoch": 1.1169415292353824, "grad_norm": 0.0898308908216899, "learning_rate": 0.00015821229701572896, "loss": 0.014, "step": 745 }, { "epoch": 1.118440779610195, "grad_norm": 0.10886469850221588, "learning_rate": 0.00015807029557109398, "loss": 0.0195, "step": 746 }, { "epoch": 1.1199400299850075, "grad_norm": 0.14282651466132884, "learning_rate": 0.0001579281172342679, "loss": 0.0135, "step": 747 }, { "epoch": 1.12143928035982, "grad_norm": 0.10856367670748046, "learning_rate": 0.00015778576243835054, "loss": 0.0169, "step": 748 }, { "epoch": 1.1229385307346327, "grad_norm": 0.06823231902579771, "learning_rate": 0.00015764323161697935, "loss": 0.0109, "step": 749 }, { "epoch": 1.1244377811094453, "grad_norm": 0.08487293384058076, "learning_rate": 0.00015750052520432787, "loss": 0.0098, "step": 750 }, { "epoch": 1.1259370314842578, "grad_norm": 0.07963389915287164, "learning_rate": 0.0001573576436351046, "loss": 0.0104, "step": 751 }, { "epoch": 1.1274362818590704, "grad_norm": 0.159748695375408, "learning_rate": 0.00015721458734455163, "loss": 0.0232, "step": 752 }, { "epoch": 1.128935532233883, "grad_norm": 0.1269515763993384, "learning_rate": 0.0001570713567684432, "loss": 0.0171, "step": 753 }, { "epoch": 1.1304347826086956, "grad_norm": 0.11521833565751512, "learning_rate": 0.00015692795234308445, "loss": 0.0172, "step": 754 }, { "epoch": 1.1319340329835081, "grad_norm": 0.08114505514088168, "learning_rate": 0.00015678437450531013, "loss": 0.0167, "step": 755 }, { "epoch": 1.133433283358321, "grad_norm": 0.09194908508040527, "learning_rate": 0.00015664062369248328, "loss": 0.0129, "step": 756 }, { "epoch": 1.1349325337331335, "grad_norm": 0.07613583096910623, "learning_rate": 0.0001564967003424938, "loss": 0.0159, "step": 757 }, { "epoch": 1.136431784107946, "grad_norm": 0.10265920058076042, "learning_rate": 0.00015635260489375714, "loss": 0.0216, "step": 758 }, { "epoch": 1.1379310344827587, "grad_norm": 0.0631078940073356, "learning_rate": 0.00015620833778521307, "loss": 0.0091, "step": 759 }, { "epoch": 1.1394302848575713, "grad_norm": 0.11550928783152578, "learning_rate": 0.0001560638994563242, "loss": 0.0247, "step": 760 }, { "epoch": 1.1409295352323838, "grad_norm": 0.05701238781802943, "learning_rate": 0.0001559192903470747, "loss": 0.007, "step": 761 }, { "epoch": 1.1424287856071964, "grad_norm": 0.0863050001301802, "learning_rate": 0.00015577451089796905, "loss": 0.0117, "step": 762 }, { "epoch": 1.143928035982009, "grad_norm": 0.1095952763491608, "learning_rate": 0.0001556295615500305, "loss": 0.0139, "step": 763 }, { "epoch": 1.1454272863568216, "grad_norm": 0.10109899186437078, "learning_rate": 0.00015548444274479995, "loss": 0.0166, "step": 764 }, { "epoch": 1.1469265367316341, "grad_norm": 0.13502630641951327, "learning_rate": 0.00015533915492433443, "loss": 0.0315, "step": 765 }, { "epoch": 1.1484257871064467, "grad_norm": 0.13530559923309216, "learning_rate": 0.0001551936985312058, "loss": 0.0105, "step": 766 }, { "epoch": 1.1499250374812593, "grad_norm": 0.1087257528041358, "learning_rate": 0.00015504807400849958, "loss": 0.0162, "step": 767 }, { "epoch": 1.1514242878560719, "grad_norm": 0.1072652302854073, "learning_rate": 0.0001549022817998132, "loss": 0.0171, "step": 768 }, { "epoch": 1.1529235382308847, "grad_norm": 0.09402648254427196, "learning_rate": 0.00015475632234925504, "loss": 0.0132, "step": 769 }, { "epoch": 1.1544227886056972, "grad_norm": 0.10085528831110949, "learning_rate": 0.0001546101961014429, "loss": 0.0146, "step": 770 }, { "epoch": 1.1559220389805098, "grad_norm": 0.10445329841658932, "learning_rate": 0.00015446390350150273, "loss": 0.0217, "step": 771 }, { "epoch": 1.1574212893553224, "grad_norm": 0.07781814067350039, "learning_rate": 0.00015431744499506706, "loss": 0.0147, "step": 772 }, { "epoch": 1.158920539730135, "grad_norm": 0.10153183985035838, "learning_rate": 0.000154170821028274, "loss": 0.0129, "step": 773 }, { "epoch": 1.1604197901049476, "grad_norm": 0.1306523813414455, "learning_rate": 0.0001540240320477655, "loss": 0.0194, "step": 774 }, { "epoch": 1.1619190404797601, "grad_norm": 0.07875779988064713, "learning_rate": 0.0001538770785006863, "loss": 0.0126, "step": 775 }, { "epoch": 1.1634182908545727, "grad_norm": 0.06662289751984372, "learning_rate": 0.0001537299608346824, "loss": 0.0078, "step": 776 }, { "epoch": 1.1649175412293853, "grad_norm": 0.11948574312781496, "learning_rate": 0.00015358267949789966, "loss": 0.0121, "step": 777 }, { "epoch": 1.1664167916041979, "grad_norm": 0.0857036909342206, "learning_rate": 0.00015343523493898265, "loss": 0.0124, "step": 778 }, { "epoch": 1.1679160419790104, "grad_norm": 0.050863627021771815, "learning_rate": 0.000153287627607073, "loss": 0.0108, "step": 779 }, { "epoch": 1.169415292353823, "grad_norm": 0.07503552287179403, "learning_rate": 0.00015313985795180828, "loss": 0.0103, "step": 780 }, { "epoch": 1.1709145427286356, "grad_norm": 0.12348759956679127, "learning_rate": 0.0001529919264233205, "loss": 0.0183, "step": 781 }, { "epoch": 1.1724137931034484, "grad_norm": 0.16396605752248122, "learning_rate": 0.00015284383347223472, "loss": 0.0237, "step": 782 }, { "epoch": 1.1739130434782608, "grad_norm": 0.11710104871369872, "learning_rate": 0.00015269557954966778, "loss": 0.0116, "step": 783 }, { "epoch": 1.1754122938530736, "grad_norm": 0.05426089810325071, "learning_rate": 0.00015254716510722677, "loss": 0.0077, "step": 784 }, { "epoch": 1.1769115442278861, "grad_norm": 0.12476718035000098, "learning_rate": 0.00015239859059700794, "loss": 0.0148, "step": 785 }, { "epoch": 1.1784107946026987, "grad_norm": 0.11340575380427748, "learning_rate": 0.0001522498564715949, "loss": 0.0174, "step": 786 }, { "epoch": 1.1799100449775113, "grad_norm": 0.07460825828437444, "learning_rate": 0.00015210096318405767, "loss": 0.0107, "step": 787 }, { "epoch": 1.1814092953523239, "grad_norm": 0.11142555377576228, "learning_rate": 0.00015195191118795096, "loss": 0.0195, "step": 788 }, { "epoch": 1.1829085457271364, "grad_norm": 0.14977712416925076, "learning_rate": 0.00015180270093731303, "loss": 0.0223, "step": 789 }, { "epoch": 1.184407796101949, "grad_norm": 0.08700335413855392, "learning_rate": 0.0001516533328866642, "loss": 0.0123, "step": 790 }, { "epoch": 1.1859070464767616, "grad_norm": 0.0724645987757823, "learning_rate": 0.00015150380749100545, "loss": 0.0086, "step": 791 }, { "epoch": 1.1874062968515742, "grad_norm": 0.14991112523341207, "learning_rate": 0.00015135412520581702, "loss": 0.0116, "step": 792 }, { "epoch": 1.1889055472263867, "grad_norm": 0.14228737821653165, "learning_rate": 0.00015120428648705717, "loss": 0.0189, "step": 793 }, { "epoch": 1.1904047976011993, "grad_norm": 0.24633364927454285, "learning_rate": 0.0001510542917911606, "loss": 0.0327, "step": 794 }, { "epoch": 1.191904047976012, "grad_norm": 0.09327854611652424, "learning_rate": 0.00015090414157503714, "loss": 0.0122, "step": 795 }, { "epoch": 1.1934032983508245, "grad_norm": 0.23468339856981874, "learning_rate": 0.00015075383629607042, "loss": 0.0176, "step": 796 }, { "epoch": 1.1949025487256373, "grad_norm": 0.05100781753856239, "learning_rate": 0.00015060337641211637, "loss": 0.0104, "step": 797 }, { "epoch": 1.1964017991004499, "grad_norm": 0.19141044396688225, "learning_rate": 0.00015045276238150192, "loss": 0.0124, "step": 798 }, { "epoch": 1.1979010494752624, "grad_norm": 0.08338970503080669, "learning_rate": 0.00015030199466302353, "loss": 0.0124, "step": 799 }, { "epoch": 1.199400299850075, "grad_norm": 0.15672763842519447, "learning_rate": 0.00015015107371594573, "loss": 0.0146, "step": 800 }, { "epoch": 1.2008995502248876, "grad_norm": 0.0879518026205462, "learning_rate": 0.00015000000000000001, "loss": 0.0105, "step": 801 }, { "epoch": 1.2023988005997002, "grad_norm": 0.049258422423769856, "learning_rate": 0.00014984877397538303, "loss": 0.0072, "step": 802 }, { "epoch": 1.2038980509745127, "grad_norm": 0.10542578394208428, "learning_rate": 0.00014969739610275556, "loss": 0.0124, "step": 803 }, { "epoch": 1.2053973013493253, "grad_norm": 0.09901336472239698, "learning_rate": 0.00014954586684324078, "loss": 0.0177, "step": 804 }, { "epoch": 1.206896551724138, "grad_norm": 0.13921198140006943, "learning_rate": 0.0001493941866584231, "loss": 0.0203, "step": 805 }, { "epoch": 1.2083958020989505, "grad_norm": 0.17504557436505902, "learning_rate": 0.00014924235601034672, "loss": 0.0272, "step": 806 }, { "epoch": 1.209895052473763, "grad_norm": 0.1707709517487399, "learning_rate": 0.00014909037536151409, "loss": 0.0203, "step": 807 }, { "epoch": 1.2113943028485756, "grad_norm": 0.12956420487730475, "learning_rate": 0.00014893824517488464, "loss": 0.0185, "step": 808 }, { "epoch": 1.2128935532233882, "grad_norm": 0.16545558068001448, "learning_rate": 0.0001487859659138733, "loss": 0.017, "step": 809 }, { "epoch": 1.214392803598201, "grad_norm": 0.10015980215539548, "learning_rate": 0.00014863353804234905, "loss": 0.0144, "step": 810 }, { "epoch": 1.2158920539730136, "grad_norm": 0.07422865061562015, "learning_rate": 0.00014848096202463372, "loss": 0.0141, "step": 811 }, { "epoch": 1.2173913043478262, "grad_norm": 0.16658923532851774, "learning_rate": 0.00014832823832550024, "loss": 0.0171, "step": 812 }, { "epoch": 1.2188905547226387, "grad_norm": 0.2952339023440661, "learning_rate": 0.00014817536741017152, "loss": 0.0249, "step": 813 }, { "epoch": 1.2203898050974513, "grad_norm": 0.343627117743269, "learning_rate": 0.0001480223497443189, "loss": 0.0203, "step": 814 }, { "epoch": 1.221889055472264, "grad_norm": 0.15860364996490584, "learning_rate": 0.0001478691857940607, "loss": 0.0173, "step": 815 }, { "epoch": 1.2233883058470765, "grad_norm": 0.07397939793708072, "learning_rate": 0.00014771587602596084, "loss": 0.0165, "step": 816 }, { "epoch": 1.224887556221889, "grad_norm": 0.09124884086427787, "learning_rate": 0.00014756242090702756, "loss": 0.0189, "step": 817 }, { "epoch": 1.2263868065967016, "grad_norm": 0.06925619214926025, "learning_rate": 0.0001474088209047116, "loss": 0.0198, "step": 818 }, { "epoch": 1.2278860569715142, "grad_norm": 0.11626344421046819, "learning_rate": 0.00014725507648690543, "loss": 0.0219, "step": 819 }, { "epoch": 1.2293853073463268, "grad_norm": 0.18165886590704167, "learning_rate": 0.000147101188121941, "loss": 0.0289, "step": 820 }, { "epoch": 1.2308845577211394, "grad_norm": 0.08731841646113968, "learning_rate": 0.00014694715627858908, "loss": 0.014, "step": 821 }, { "epoch": 1.232383808095952, "grad_norm": 0.06959854220083778, "learning_rate": 0.00014679298142605734, "loss": 0.0114, "step": 822 }, { "epoch": 1.2338830584707647, "grad_norm": 0.36508436863746124, "learning_rate": 0.00014663866403398913, "loss": 0.0244, "step": 823 }, { "epoch": 1.235382308845577, "grad_norm": 0.09368142163812859, "learning_rate": 0.000146484204572462, "loss": 0.0174, "step": 824 }, { "epoch": 1.23688155922039, "grad_norm": 0.08193058500216474, "learning_rate": 0.00014632960351198618, "loss": 0.0132, "step": 825 }, { "epoch": 1.2383808095952025, "grad_norm": 0.15614447953538005, "learning_rate": 0.00014617486132350343, "loss": 0.0195, "step": 826 }, { "epoch": 1.239880059970015, "grad_norm": 0.0779958529833662, "learning_rate": 0.00014601997847838518, "loss": 0.0108, "step": 827 }, { "epoch": 1.2413793103448276, "grad_norm": 0.0888114278475023, "learning_rate": 0.00014586495544843152, "loss": 0.0147, "step": 828 }, { "epoch": 1.2428785607196402, "grad_norm": 0.11955268016013557, "learning_rate": 0.00014570979270586945, "loss": 0.0201, "step": 829 }, { "epoch": 1.2443778110944528, "grad_norm": 0.202954135904049, "learning_rate": 0.00014555449072335157, "loss": 0.0301, "step": 830 }, { "epoch": 1.2458770614692654, "grad_norm": 0.18587709894174873, "learning_rate": 0.00014539904997395468, "loss": 0.0368, "step": 831 }, { "epoch": 1.247376311844078, "grad_norm": 0.07571382735258467, "learning_rate": 0.00014524347093117828, "loss": 0.0131, "step": 832 }, { "epoch": 1.2488755622188905, "grad_norm": 0.1422916655610059, "learning_rate": 0.00014508775406894307, "loss": 0.021, "step": 833 }, { "epoch": 1.250374812593703, "grad_norm": 0.13587129669974282, "learning_rate": 0.00014493189986158965, "loss": 0.0162, "step": 834 }, { "epoch": 1.2518740629685157, "grad_norm": 0.08417093744852887, "learning_rate": 0.00014477590878387696, "loss": 0.0167, "step": 835 }, { "epoch": 1.2533733133433285, "grad_norm": 0.18400438316412573, "learning_rate": 0.00014461978131098088, "loss": 0.0235, "step": 836 }, { "epoch": 1.2548725637181408, "grad_norm": 0.07333689889670233, "learning_rate": 0.00014446351791849276, "loss": 0.0097, "step": 837 }, { "epoch": 1.2563718140929536, "grad_norm": 0.16562856351156585, "learning_rate": 0.00014430711908241798, "loss": 0.0481, "step": 838 }, { "epoch": 1.2578710644677662, "grad_norm": 0.11118124090310846, "learning_rate": 0.00014415058527917452, "loss": 0.0214, "step": 839 }, { "epoch": 1.2593703148425788, "grad_norm": 0.18915002411038295, "learning_rate": 0.00014399391698559152, "loss": 0.0272, "step": 840 }, { "epoch": 1.2608695652173914, "grad_norm": 0.054608561460708256, "learning_rate": 0.00014383711467890774, "loss": 0.0103, "step": 841 }, { "epoch": 1.262368815592204, "grad_norm": 0.11913677742871757, "learning_rate": 0.00014368017883677024, "loss": 0.0264, "step": 842 }, { "epoch": 1.2638680659670165, "grad_norm": 0.14267753773889416, "learning_rate": 0.00014352310993723277, "loss": 0.0257, "step": 843 }, { "epoch": 1.265367316341829, "grad_norm": 0.20047662055314494, "learning_rate": 0.00014336590845875446, "loss": 0.0277, "step": 844 }, { "epoch": 1.2668665667166417, "grad_norm": 0.13109076906839104, "learning_rate": 0.00014320857488019824, "loss": 0.0227, "step": 845 }, { "epoch": 1.2683658170914542, "grad_norm": 0.12458461929595395, "learning_rate": 0.00014305110968082952, "loss": 0.0164, "step": 846 }, { "epoch": 1.2698650674662668, "grad_norm": 0.09746106663140872, "learning_rate": 0.0001428935133403146, "loss": 0.0124, "step": 847 }, { "epoch": 1.2713643178410794, "grad_norm": 0.12378914670534663, "learning_rate": 0.00014273578633871927, "loss": 0.0233, "step": 848 }, { "epoch": 1.272863568215892, "grad_norm": 0.06854962001087733, "learning_rate": 0.00014257792915650728, "loss": 0.011, "step": 849 }, { "epoch": 1.2743628185907045, "grad_norm": 0.10385595925660064, "learning_rate": 0.00014241994227453901, "loss": 0.0102, "step": 850 }, { "epoch": 1.2758620689655173, "grad_norm": 0.11662093941235979, "learning_rate": 0.00014226182617406996, "loss": 0.0128, "step": 851 }, { "epoch": 1.2773613193403297, "grad_norm": 0.17930864492543278, "learning_rate": 0.00014210358133674912, "loss": 0.0126, "step": 852 }, { "epoch": 1.2788605697151425, "grad_norm": 0.1372485077515871, "learning_rate": 0.00014194520824461771, "loss": 0.012, "step": 853 }, { "epoch": 1.280359820089955, "grad_norm": 0.16345335022814164, "learning_rate": 0.0001417867073801077, "loss": 0.0249, "step": 854 }, { "epoch": 1.2818590704647677, "grad_norm": 0.0997906139397832, "learning_rate": 0.00014162807922604012, "loss": 0.012, "step": 855 }, { "epoch": 1.2833583208395802, "grad_norm": 0.09163084280524973, "learning_rate": 0.00014146932426562392, "loss": 0.0147, "step": 856 }, { "epoch": 1.2848575712143928, "grad_norm": 0.15802403576676588, "learning_rate": 0.0001413104429824542, "loss": 0.0116, "step": 857 }, { "epoch": 1.2863568215892054, "grad_norm": 0.0854686973988408, "learning_rate": 0.00014115143586051088, "loss": 0.0128, "step": 858 }, { "epoch": 1.287856071964018, "grad_norm": 0.06726116883117537, "learning_rate": 0.00014099230338415728, "loss": 0.0121, "step": 859 }, { "epoch": 1.2893553223388305, "grad_norm": 0.11037334160421672, "learning_rate": 0.00014083304603813848, "loss": 0.0103, "step": 860 }, { "epoch": 1.2908545727136431, "grad_norm": 0.12332181597901376, "learning_rate": 0.00014067366430758004, "loss": 0.0224, "step": 861 }, { "epoch": 1.2923538230884557, "grad_norm": 0.18194326760827853, "learning_rate": 0.00014051415867798628, "loss": 0.0352, "step": 862 }, { "epoch": 1.2938530734632683, "grad_norm": 0.08430952227790076, "learning_rate": 0.00014035452963523902, "loss": 0.0142, "step": 863 }, { "epoch": 1.295352323838081, "grad_norm": 0.07124275034999716, "learning_rate": 0.00014019477766559604, "loss": 0.007, "step": 864 }, { "epoch": 1.2968515742128934, "grad_norm": 0.09273754187553625, "learning_rate": 0.00014003490325568954, "loss": 0.0149, "step": 865 }, { "epoch": 1.2983508245877062, "grad_norm": 0.24506349645358916, "learning_rate": 0.00013987490689252463, "loss": 0.035, "step": 866 }, { "epoch": 1.2998500749625188, "grad_norm": 0.12312219048398249, "learning_rate": 0.00013971478906347806, "loss": 0.0211, "step": 867 }, { "epoch": 1.3013493253373314, "grad_norm": 0.09854369818513668, "learning_rate": 0.00013955455025629651, "loss": 0.0168, "step": 868 }, { "epoch": 1.302848575712144, "grad_norm": 0.18077070256023667, "learning_rate": 0.00013939419095909512, "loss": 0.0265, "step": 869 }, { "epoch": 1.3043478260869565, "grad_norm": 0.08104201839225528, "learning_rate": 0.00013923371166035616, "loss": 0.0141, "step": 870 }, { "epoch": 1.3058470764617691, "grad_norm": 0.05669225042597822, "learning_rate": 0.00013907311284892736, "loss": 0.0067, "step": 871 }, { "epoch": 1.3073463268365817, "grad_norm": 0.170396059391146, "learning_rate": 0.0001389123950140206, "loss": 0.0144, "step": 872 }, { "epoch": 1.3088455772113943, "grad_norm": 0.09404710563168223, "learning_rate": 0.0001387515586452103, "loss": 0.015, "step": 873 }, { "epoch": 1.3103448275862069, "grad_norm": 0.07336649457841225, "learning_rate": 0.00013859060423243187, "loss": 0.0134, "step": 874 }, { "epoch": 1.3118440779610194, "grad_norm": 0.15102499216151205, "learning_rate": 0.00013842953226598037, "loss": 0.031, "step": 875 }, { "epoch": 1.313343328335832, "grad_norm": 0.11710688194401644, "learning_rate": 0.000138268343236509, "loss": 0.0228, "step": 876 }, { "epoch": 1.3148425787106448, "grad_norm": 0.08364534811984702, "learning_rate": 0.00013810703763502744, "loss": 0.0129, "step": 877 }, { "epoch": 1.3163418290854572, "grad_norm": 0.08023322549168056, "learning_rate": 0.00013794561595290052, "loss": 0.0198, "step": 878 }, { "epoch": 1.31784107946027, "grad_norm": 0.10058428662726825, "learning_rate": 0.00013778407868184672, "loss": 0.0288, "step": 879 }, { "epoch": 1.3193403298350825, "grad_norm": 0.10611806737953573, "learning_rate": 0.00013762242631393655, "loss": 0.0154, "step": 880 }, { "epoch": 1.3208395802098951, "grad_norm": 0.07758538286432358, "learning_rate": 0.00013746065934159123, "loss": 0.0133, "step": 881 }, { "epoch": 1.3223388305847077, "grad_norm": 0.11661318831041984, "learning_rate": 0.0001372987782575809, "loss": 0.0147, "step": 882 }, { "epoch": 1.3238380809595203, "grad_norm": 0.10345149158644346, "learning_rate": 0.00013713678355502351, "loss": 0.0175, "step": 883 }, { "epoch": 1.3253373313343328, "grad_norm": 0.17899088086579465, "learning_rate": 0.00013697467572738295, "loss": 0.0159, "step": 884 }, { "epoch": 1.3268365817091454, "grad_norm": 0.10366071968769505, "learning_rate": 0.00013681245526846783, "loss": 0.0185, "step": 885 }, { "epoch": 1.328335832083958, "grad_norm": 0.2320430785565242, "learning_rate": 0.00013665012267242974, "loss": 0.0109, "step": 886 }, { "epoch": 1.3298350824587706, "grad_norm": 0.2165729911004686, "learning_rate": 0.00013648767843376196, "loss": 0.0353, "step": 887 }, { "epoch": 1.3313343328335832, "grad_norm": 0.090948134296178, "learning_rate": 0.00013632512304729785, "loss": 0.0117, "step": 888 }, { "epoch": 1.3328335832083957, "grad_norm": 0.1723357033638432, "learning_rate": 0.00013616245700820922, "loss": 0.0211, "step": 889 }, { "epoch": 1.3343328335832085, "grad_norm": 0.12328155582579832, "learning_rate": 0.00013599968081200514, "loss": 0.0222, "step": 890 }, { "epoch": 1.3358320839580209, "grad_norm": 0.1148672322347492, "learning_rate": 0.00013583679495453, "loss": 0.017, "step": 891 }, { "epoch": 1.3373313343328337, "grad_norm": 0.08259323816085176, "learning_rate": 0.00013567379993196252, "loss": 0.0109, "step": 892 }, { "epoch": 1.338830584707646, "grad_norm": 0.1149875760609526, "learning_rate": 0.0001355106962408137, "loss": 0.0202, "step": 893 }, { "epoch": 1.3403298350824588, "grad_norm": 0.08902107605424671, "learning_rate": 0.00013534748437792573, "loss": 0.0146, "step": 894 }, { "epoch": 1.3418290854572714, "grad_norm": 0.2630218492094372, "learning_rate": 0.00013518416484047018, "loss": 0.0327, "step": 895 }, { "epoch": 1.343328335832084, "grad_norm": 0.09214551614879145, "learning_rate": 0.00013502073812594675, "loss": 0.0149, "step": 896 }, { "epoch": 1.3448275862068966, "grad_norm": 0.08385569860592369, "learning_rate": 0.00013485720473218154, "loss": 0.0182, "step": 897 }, { "epoch": 1.3463268365817092, "grad_norm": 0.10306630082691921, "learning_rate": 0.00013469356515732558, "loss": 0.0153, "step": 898 }, { "epoch": 1.3478260869565217, "grad_norm": 0.07604456448727595, "learning_rate": 0.00013452981989985348, "loss": 0.0129, "step": 899 }, { "epoch": 1.3493253373313343, "grad_norm": 0.08738991305136598, "learning_rate": 0.00013436596945856164, "loss": 0.0159, "step": 900 }, { "epoch": 1.3508245877061469, "grad_norm": 0.15222706121553015, "learning_rate": 0.00013420201433256689, "loss": 0.021, "step": 901 }, { "epoch": 1.3523238380809595, "grad_norm": 0.1306031031322151, "learning_rate": 0.00013403795502130503, "loss": 0.0167, "step": 902 }, { "epoch": 1.353823088455772, "grad_norm": 0.08808517355752062, "learning_rate": 0.00013387379202452917, "loss": 0.0139, "step": 903 }, { "epoch": 1.3553223388305846, "grad_norm": 0.11211685787913292, "learning_rate": 0.00013370952584230823, "loss": 0.0137, "step": 904 }, { "epoch": 1.3568215892053974, "grad_norm": 0.08778023618067164, "learning_rate": 0.00013354515697502553, "loss": 0.0107, "step": 905 }, { "epoch": 1.3583208395802098, "grad_norm": 0.06388026832690835, "learning_rate": 0.0001333806859233771, "loss": 0.0098, "step": 906 }, { "epoch": 1.3598200899550226, "grad_norm": 0.1148047545564487, "learning_rate": 0.00013321611318837032, "loss": 0.0105, "step": 907 }, { "epoch": 1.3613193403298351, "grad_norm": 0.10499549294742563, "learning_rate": 0.0001330514392713223, "loss": 0.0186, "step": 908 }, { "epoch": 1.3628185907046477, "grad_norm": 0.07183000371069133, "learning_rate": 0.00013288666467385833, "loss": 0.0053, "step": 909 }, { "epoch": 1.3643178410794603, "grad_norm": 0.09527252116065761, "learning_rate": 0.0001327217898979104, "loss": 0.0144, "step": 910 }, { "epoch": 1.3658170914542729, "grad_norm": 0.17631678204685763, "learning_rate": 0.00013255681544571568, "loss": 0.0232, "step": 911 }, { "epoch": 1.3673163418290855, "grad_norm": 0.09351491513727911, "learning_rate": 0.00013239174181981495, "loss": 0.0167, "step": 912 }, { "epoch": 1.368815592203898, "grad_norm": 0.13425537805814766, "learning_rate": 0.00013222656952305113, "loss": 0.0191, "step": 913 }, { "epoch": 1.3703148425787106, "grad_norm": 0.14230677646930492, "learning_rate": 0.00013206129905856765, "loss": 0.0229, "step": 914 }, { "epoch": 1.3718140929535232, "grad_norm": 0.09153923973277442, "learning_rate": 0.00013189593092980702, "loss": 0.0131, "step": 915 }, { "epoch": 1.3733133433283358, "grad_norm": 0.0578622169127209, "learning_rate": 0.00013173046564050924, "loss": 0.0113, "step": 916 }, { "epoch": 1.3748125937031483, "grad_norm": 0.13891677776816042, "learning_rate": 0.00013156490369471027, "loss": 0.0137, "step": 917 }, { "epoch": 1.3763118440779611, "grad_norm": 0.04903272701124958, "learning_rate": 0.00013139924559674052, "loss": 0.0062, "step": 918 }, { "epoch": 1.3778110944527735, "grad_norm": 0.08762005054912386, "learning_rate": 0.00013123349185122327, "loss": 0.0183, "step": 919 }, { "epoch": 1.3793103448275863, "grad_norm": 0.08308919330791038, "learning_rate": 0.0001310676429630732, "loss": 0.0134, "step": 920 }, { "epoch": 1.3808095952023989, "grad_norm": 0.05992308927118659, "learning_rate": 0.00013090169943749476, "loss": 0.0128, "step": 921 }, { "epoch": 1.3823088455772115, "grad_norm": 0.06850172645375537, "learning_rate": 0.00013073566177998074, "loss": 0.0078, "step": 922 }, { "epoch": 1.383808095952024, "grad_norm": 0.11509963104785531, "learning_rate": 0.00013056953049631057, "loss": 0.0143, "step": 923 }, { "epoch": 1.3853073463268366, "grad_norm": 0.24649040007973805, "learning_rate": 0.00013040330609254903, "loss": 0.0124, "step": 924 }, { "epoch": 1.3868065967016492, "grad_norm": 0.07928186572736103, "learning_rate": 0.00013023698907504446, "loss": 0.0159, "step": 925 }, { "epoch": 1.3883058470764618, "grad_norm": 0.11963166407145638, "learning_rate": 0.00013007057995042732, "loss": 0.033, "step": 926 }, { "epoch": 1.3898050974512743, "grad_norm": 0.09256628228543629, "learning_rate": 0.00012990407922560868, "loss": 0.0168, "step": 927 }, { "epoch": 1.391304347826087, "grad_norm": 0.12551820161013388, "learning_rate": 0.0001297374874077786, "loss": 0.0128, "step": 928 }, { "epoch": 1.3928035982008995, "grad_norm": 0.09396953539790104, "learning_rate": 0.00012957080500440468, "loss": 0.0126, "step": 929 }, { "epoch": 1.394302848575712, "grad_norm": 0.0814330557561346, "learning_rate": 0.0001294040325232304, "loss": 0.0109, "step": 930 }, { "epoch": 1.3958020989505249, "grad_norm": 0.09170961671586021, "learning_rate": 0.00012923717047227368, "loss": 0.0123, "step": 931 }, { "epoch": 1.3973013493253372, "grad_norm": 0.12429481547170583, "learning_rate": 0.00012907021935982526, "loss": 0.017, "step": 932 }, { "epoch": 1.39880059970015, "grad_norm": 0.07547138071359946, "learning_rate": 0.00012890317969444716, "loss": 0.0101, "step": 933 }, { "epoch": 1.4002998500749624, "grad_norm": 0.10947780038801162, "learning_rate": 0.00012873605198497124, "loss": 0.0122, "step": 934 }, { "epoch": 1.4017991004497752, "grad_norm": 0.14974745222283475, "learning_rate": 0.00012856883674049736, "loss": 0.021, "step": 935 }, { "epoch": 1.4032983508245878, "grad_norm": 0.06311397898354858, "learning_rate": 0.00012840153447039228, "loss": 0.0097, "step": 936 }, { "epoch": 1.4047976011994003, "grad_norm": 0.18221914938913733, "learning_rate": 0.00012823414568428768, "loss": 0.031, "step": 937 }, { "epoch": 1.406296851574213, "grad_norm": 0.08321813040032747, "learning_rate": 0.0001280666708920788, "loss": 0.0096, "step": 938 }, { "epoch": 1.4077961019490255, "grad_norm": 0.08421041313135452, "learning_rate": 0.00012789911060392294, "loss": 0.0148, "step": 939 }, { "epoch": 1.409295352323838, "grad_norm": 0.06717497940090396, "learning_rate": 0.00012773146533023782, "loss": 0.0148, "step": 940 }, { "epoch": 1.4107946026986506, "grad_norm": 0.09168806606704279, "learning_rate": 0.0001275637355816999, "loss": 0.0171, "step": 941 }, { "epoch": 1.4122938530734632, "grad_norm": 0.10477491574939178, "learning_rate": 0.00012739592186924328, "loss": 0.0134, "step": 942 }, { "epoch": 1.4137931034482758, "grad_norm": 0.12092733743202747, "learning_rate": 0.00012722802470405744, "loss": 0.0167, "step": 943 }, { "epoch": 1.4152923538230884, "grad_norm": 0.14068944332954328, "learning_rate": 0.00012706004459758636, "loss": 0.0296, "step": 944 }, { "epoch": 1.416791604197901, "grad_norm": 0.14137864202064285, "learning_rate": 0.00012689198206152657, "loss": 0.0134, "step": 945 }, { "epoch": 1.4182908545727138, "grad_norm": 0.0860572351607591, "learning_rate": 0.00012672383760782568, "loss": 0.0178, "step": 946 }, { "epoch": 1.419790104947526, "grad_norm": 0.07404000519398271, "learning_rate": 0.00012655561174868088, "loss": 0.0102, "step": 947 }, { "epoch": 1.421289355322339, "grad_norm": 0.11582588947336643, "learning_rate": 0.0001263873049965373, "loss": 0.0186, "step": 948 }, { "epoch": 1.4227886056971515, "grad_norm": 0.13400349083924593, "learning_rate": 0.00012621891786408648, "loss": 0.011, "step": 949 }, { "epoch": 1.424287856071964, "grad_norm": 0.20636194375809655, "learning_rate": 0.00012605045086426487, "loss": 0.0199, "step": 950 }, { "epoch": 1.4257871064467766, "grad_norm": 0.03725195400505915, "learning_rate": 0.00012588190451025207, "loss": 0.0046, "step": 951 }, { "epoch": 1.4272863568215892, "grad_norm": 0.05892090161081139, "learning_rate": 0.00012571327931546963, "loss": 0.0097, "step": 952 }, { "epoch": 1.4287856071964018, "grad_norm": 0.12490487802223602, "learning_rate": 0.00012554457579357905, "loss": 0.009, "step": 953 }, { "epoch": 1.4302848575712144, "grad_norm": 0.11179141701581286, "learning_rate": 0.00012537579445848058, "loss": 0.0183, "step": 954 }, { "epoch": 1.431784107946027, "grad_norm": 0.06972546035600884, "learning_rate": 0.0001252069358243114, "loss": 0.0122, "step": 955 }, { "epoch": 1.4332833583208395, "grad_norm": 0.09131261727248277, "learning_rate": 0.00012503800040544416, "loss": 0.0095, "step": 956 }, { "epoch": 1.434782608695652, "grad_norm": 0.1109811036679139, "learning_rate": 0.0001248689887164855, "loss": 0.0105, "step": 957 }, { "epoch": 1.4362818590704647, "grad_norm": 0.07591084965117433, "learning_rate": 0.00012469990127227432, "loss": 0.0138, "step": 958 }, { "epoch": 1.4377811094452775, "grad_norm": 0.20788394817699887, "learning_rate": 0.00012453073858788026, "loss": 0.0384, "step": 959 }, { "epoch": 1.4392803598200898, "grad_norm": 0.12391628024007592, "learning_rate": 0.00012436150117860225, "loss": 0.0136, "step": 960 }, { "epoch": 1.4407796101949026, "grad_norm": 0.12604999899644895, "learning_rate": 0.00012419218955996676, "loss": 0.0211, "step": 961 }, { "epoch": 1.4422788605697152, "grad_norm": 0.09347292600591614, "learning_rate": 0.00012402280424772639, "loss": 0.0056, "step": 962 }, { "epoch": 1.4437781109445278, "grad_norm": 0.09362580278945488, "learning_rate": 0.0001238533457578581, "loss": 0.0149, "step": 963 }, { "epoch": 1.4452773613193404, "grad_norm": 0.13450314184024556, "learning_rate": 0.0001236838146065619, "loss": 0.0183, "step": 964 }, { "epoch": 1.446776611694153, "grad_norm": 0.06683370298885613, "learning_rate": 0.000123514211310259, "loss": 0.0114, "step": 965 }, { "epoch": 1.4482758620689655, "grad_norm": 0.07533136377894585, "learning_rate": 0.00012334453638559057, "loss": 0.0088, "step": 966 }, { "epoch": 1.449775112443778, "grad_norm": 0.11385337434889103, "learning_rate": 0.00012317479034941573, "loss": 0.0132, "step": 967 }, { "epoch": 1.4512743628185907, "grad_norm": 0.08660001131021804, "learning_rate": 0.00012300497371881046, "loss": 0.0183, "step": 968 }, { "epoch": 1.4527736131934033, "grad_norm": 0.21035359781149526, "learning_rate": 0.00012283508701106557, "loss": 0.015, "step": 969 }, { "epoch": 1.4542728635682158, "grad_norm": 0.14422586173075505, "learning_rate": 0.0001226651307436855, "loss": 0.0205, "step": 970 }, { "epoch": 1.4557721139430284, "grad_norm": 0.147239584775246, "learning_rate": 0.0001224951054343865, "loss": 0.0103, "step": 971 }, { "epoch": 1.4572713643178412, "grad_norm": 0.11512591488947908, "learning_rate": 0.00012232501160109514, "loss": 0.0154, "step": 972 }, { "epoch": 1.4587706146926536, "grad_norm": 0.16628564022136066, "learning_rate": 0.00012215484976194676, "loss": 0.0208, "step": 973 }, { "epoch": 1.4602698650674664, "grad_norm": 0.1705959708232762, "learning_rate": 0.00012198462043528376, "loss": 0.0147, "step": 974 }, { "epoch": 1.461769115442279, "grad_norm": 0.1385993826378694, "learning_rate": 0.00012181432413965428, "loss": 0.0122, "step": 975 }, { "epoch": 1.4632683658170915, "grad_norm": 0.14300288286624654, "learning_rate": 0.00012164396139381029, "loss": 0.0196, "step": 976 }, { "epoch": 1.464767616191904, "grad_norm": 0.080459269082652, "learning_rate": 0.00012147353271670634, "loss": 0.0085, "step": 977 }, { "epoch": 1.4662668665667167, "grad_norm": 0.14683345902187586, "learning_rate": 0.00012130303862749767, "loss": 0.0337, "step": 978 }, { "epoch": 1.4677661169415293, "grad_norm": 0.08820586327672093, "learning_rate": 0.00012113247964553888, "loss": 0.0091, "step": 979 }, { "epoch": 1.4692653673163418, "grad_norm": 0.11691373222525352, "learning_rate": 0.0001209618562903822, "loss": 0.0138, "step": 980 }, { "epoch": 1.4707646176911544, "grad_norm": 0.13161051562985052, "learning_rate": 0.00012079116908177593, "loss": 0.0087, "step": 981 }, { "epoch": 1.472263868065967, "grad_norm": 0.09613499888195008, "learning_rate": 0.00012062041853966298, "loss": 0.0088, "step": 982 }, { "epoch": 1.4737631184407796, "grad_norm": 0.09195368296362892, "learning_rate": 0.00012044960518417903, "loss": 0.0136, "step": 983 }, { "epoch": 1.4752623688155921, "grad_norm": 0.17465872281858535, "learning_rate": 0.00012027872953565125, "loss": 0.0206, "step": 984 }, { "epoch": 1.4767616191904047, "grad_norm": 0.07836679172273597, "learning_rate": 0.00012010779211459648, "loss": 0.011, "step": 985 }, { "epoch": 1.4782608695652173, "grad_norm": 0.09365663057692926, "learning_rate": 0.00011993679344171973, "loss": 0.0182, "step": 986 }, { "epoch": 1.47976011994003, "grad_norm": 0.12682030561564137, "learning_rate": 0.00011976573403791262, "loss": 0.0089, "step": 987 }, { "epoch": 1.4812593703148424, "grad_norm": 0.08393483257842721, "learning_rate": 0.00011959461442425177, "loss": 0.0114, "step": 988 }, { "epoch": 1.4827586206896552, "grad_norm": 0.1050372238110733, "learning_rate": 0.0001194234351219972, "loss": 0.0203, "step": 989 }, { "epoch": 1.4842578710644678, "grad_norm": 0.11112289986216253, "learning_rate": 0.00011925219665259075, "loss": 0.016, "step": 990 }, { "epoch": 1.4857571214392804, "grad_norm": 0.10020695139140749, "learning_rate": 0.00011908089953765449, "loss": 0.0081, "step": 991 }, { "epoch": 1.487256371814093, "grad_norm": 0.11236596316087173, "learning_rate": 0.00011890954429898912, "loss": 0.0198, "step": 992 }, { "epoch": 1.4887556221889056, "grad_norm": 0.07348295688717253, "learning_rate": 0.00011873813145857249, "loss": 0.0103, "step": 993 }, { "epoch": 1.4902548725637181, "grad_norm": 0.07744821371483936, "learning_rate": 0.00011856666153855776, "loss": 0.0096, "step": 994 }, { "epoch": 1.4917541229385307, "grad_norm": 0.09069561623382377, "learning_rate": 0.00011839513506127203, "loss": 0.0104, "step": 995 }, { "epoch": 1.4932533733133433, "grad_norm": 0.05709624694520552, "learning_rate": 0.00011822355254921478, "loss": 0.0072, "step": 996 }, { "epoch": 1.4947526236881559, "grad_norm": 0.11572106025581866, "learning_rate": 0.00011805191452505602, "loss": 0.0242, "step": 997 }, { "epoch": 1.4962518740629684, "grad_norm": 0.09230690991056163, "learning_rate": 0.00011788022151163495, "loss": 0.017, "step": 998 }, { "epoch": 1.497751124437781, "grad_norm": 0.10875800562076483, "learning_rate": 0.00011770847403195834, "loss": 0.0132, "step": 999 }, { "epoch": 1.4992503748125938, "grad_norm": 0.13231704496690558, "learning_rate": 0.00011753667260919872, "loss": 0.0195, "step": 1000 }, { "epoch": 1.5007496251874062, "grad_norm": 0.1281968401653238, "learning_rate": 0.00011736481776669306, "loss": 0.012, "step": 1001 }, { "epoch": 1.502248875562219, "grad_norm": 0.10072623997610383, "learning_rate": 0.00011719291002794096, "loss": 0.0118, "step": 1002 }, { "epoch": 1.5037481259370313, "grad_norm": 0.20173093622273117, "learning_rate": 0.00011702094991660326, "loss": 0.0302, "step": 1003 }, { "epoch": 1.5052473763118441, "grad_norm": 0.24868628674117718, "learning_rate": 0.00011684893795650027, "loss": 0.026, "step": 1004 }, { "epoch": 1.5067466266866567, "grad_norm": 0.20149856701683075, "learning_rate": 0.00011667687467161024, "loss": 0.0315, "step": 1005 }, { "epoch": 1.5082458770614693, "grad_norm": 0.09025687720729765, "learning_rate": 0.00011650476058606777, "loss": 0.0128, "step": 1006 }, { "epoch": 1.5097451274362819, "grad_norm": 0.07593193449503319, "learning_rate": 0.00011633259622416224, "loss": 0.0077, "step": 1007 }, { "epoch": 1.5112443778110944, "grad_norm": 0.14453173196748317, "learning_rate": 0.00011616038211033613, "loss": 0.0206, "step": 1008 }, { "epoch": 1.512743628185907, "grad_norm": 0.12215216306236011, "learning_rate": 0.0001159881187691835, "loss": 0.0138, "step": 1009 }, { "epoch": 1.5142428785607196, "grad_norm": 0.1512560808287919, "learning_rate": 0.00011581580672544838, "loss": 0.0251, "step": 1010 }, { "epoch": 1.5157421289355324, "grad_norm": 0.08198991344869536, "learning_rate": 0.0001156434465040231, "loss": 0.0116, "step": 1011 }, { "epoch": 1.5172413793103448, "grad_norm": 0.07771015418992609, "learning_rate": 0.00011547103862994684, "loss": 0.0176, "step": 1012 }, { "epoch": 1.5187406296851576, "grad_norm": 0.1523429461936693, "learning_rate": 0.00011529858362840382, "loss": 0.0189, "step": 1013 }, { "epoch": 1.52023988005997, "grad_norm": 0.11328394560469286, "learning_rate": 0.00011512608202472194, "loss": 0.0161, "step": 1014 }, { "epoch": 1.5217391304347827, "grad_norm": 0.11980943303869283, "learning_rate": 0.00011495353434437098, "loss": 0.0259, "step": 1015 }, { "epoch": 1.523238380809595, "grad_norm": 0.1359052987718006, "learning_rate": 0.00011478094111296109, "loss": 0.0167, "step": 1016 }, { "epoch": 1.5247376311844079, "grad_norm": 0.06578587863029016, "learning_rate": 0.00011460830285624118, "loss": 0.0104, "step": 1017 }, { "epoch": 1.5262368815592204, "grad_norm": 0.09597998810321515, "learning_rate": 0.00011443562010009731, "loss": 0.0082, "step": 1018 }, { "epoch": 1.527736131934033, "grad_norm": 0.12322964223446362, "learning_rate": 0.00011426289337055119, "loss": 0.009, "step": 1019 }, { "epoch": 1.5292353823088456, "grad_norm": 0.0719193270536759, "learning_rate": 0.00011409012319375827, "loss": 0.0106, "step": 1020 }, { "epoch": 1.5307346326836582, "grad_norm": 0.09940735793245362, "learning_rate": 0.00011391731009600654, "loss": 0.0092, "step": 1021 }, { "epoch": 1.5322338830584707, "grad_norm": 0.07250731895990385, "learning_rate": 0.00011374445460371466, "loss": 0.0105, "step": 1022 }, { "epoch": 1.5337331334332833, "grad_norm": 0.05856277224745514, "learning_rate": 0.00011357155724343045, "loss": 0.0101, "step": 1023 }, { "epoch": 1.535232383808096, "grad_norm": 0.07803522533264978, "learning_rate": 0.00011339861854182922, "loss": 0.0157, "step": 1024 }, { "epoch": 1.5367316341829085, "grad_norm": 0.23493948718254026, "learning_rate": 0.00011322563902571226, "loss": 0.027, "step": 1025 }, { "epoch": 1.5382308845577213, "grad_norm": 0.044799835235369186, "learning_rate": 0.00011305261922200519, "loss": 0.0093, "step": 1026 }, { "epoch": 1.5397301349325336, "grad_norm": 0.09126298677383313, "learning_rate": 0.0001128795596577563, "loss": 0.0162, "step": 1027 }, { "epoch": 1.5412293853073464, "grad_norm": 0.1038456479026185, "learning_rate": 0.00011270646086013505, "loss": 0.0131, "step": 1028 }, { "epoch": 1.5427286356821588, "grad_norm": 0.07105696856441536, "learning_rate": 0.00011253332335643043, "loss": 0.013, "step": 1029 }, { "epoch": 1.5442278860569716, "grad_norm": 0.07978824589108958, "learning_rate": 0.00011236014767404927, "loss": 0.0119, "step": 1030 }, { "epoch": 1.545727136431784, "grad_norm": 0.10620185121687949, "learning_rate": 0.00011218693434051475, "loss": 0.0194, "step": 1031 }, { "epoch": 1.5472263868065967, "grad_norm": 0.06597817870162546, "learning_rate": 0.00011201368388346471, "loss": 0.0122, "step": 1032 }, { "epoch": 1.5487256371814093, "grad_norm": 0.1391579923371684, "learning_rate": 0.00011184039683065013, "loss": 0.0159, "step": 1033 }, { "epoch": 1.550224887556222, "grad_norm": 0.0583734987893039, "learning_rate": 0.00011166707370993333, "loss": 0.0089, "step": 1034 }, { "epoch": 1.5517241379310345, "grad_norm": 0.05835705798534313, "learning_rate": 0.00011149371504928668, "loss": 0.0095, "step": 1035 }, { "epoch": 1.553223388305847, "grad_norm": 0.1162274585062899, "learning_rate": 0.0001113203213767907, "loss": 0.0093, "step": 1036 }, { "epoch": 1.5547226386806596, "grad_norm": 0.12624340143523036, "learning_rate": 0.00011114689322063255, "loss": 0.0133, "step": 1037 }, { "epoch": 1.5562218890554722, "grad_norm": 0.124541965655317, "learning_rate": 0.00011097343110910452, "loss": 0.0128, "step": 1038 }, { "epoch": 1.557721139430285, "grad_norm": 0.1032713740550231, "learning_rate": 0.0001107999355706023, "loss": 0.0137, "step": 1039 }, { "epoch": 1.5592203898050974, "grad_norm": 0.11800022175627153, "learning_rate": 0.00011062640713362333, "loss": 0.0158, "step": 1040 }, { "epoch": 1.5607196401799102, "grad_norm": 0.1302341773557835, "learning_rate": 0.00011045284632676536, "loss": 0.0146, "step": 1041 }, { "epoch": 1.5622188905547225, "grad_norm": 0.1546164674351709, "learning_rate": 0.00011027925367872469, "loss": 0.0388, "step": 1042 }, { "epoch": 1.5637181409295353, "grad_norm": 0.05168069492010779, "learning_rate": 0.00011010562971829463, "loss": 0.0084, "step": 1043 }, { "epoch": 1.5652173913043477, "grad_norm": 0.13440988075774132, "learning_rate": 0.00010993197497436391, "loss": 0.0192, "step": 1044 }, { "epoch": 1.5667166416791605, "grad_norm": 0.1296790212762601, "learning_rate": 0.00010975828997591495, "loss": 0.0157, "step": 1045 }, { "epoch": 1.568215892053973, "grad_norm": 0.076210180946536, "learning_rate": 0.00010958457525202241, "loss": 0.0173, "step": 1046 }, { "epoch": 1.5697151424287856, "grad_norm": 0.07970386766287546, "learning_rate": 0.00010941083133185146, "loss": 0.01, "step": 1047 }, { "epoch": 1.5712143928035982, "grad_norm": 0.17171859330243994, "learning_rate": 0.00010923705874465618, "loss": 0.0361, "step": 1048 }, { "epoch": 1.5727136431784108, "grad_norm": 0.18708680882057213, "learning_rate": 0.00010906325801977804, "loss": 0.0139, "step": 1049 }, { "epoch": 1.5742128935532234, "grad_norm": 0.11969812919320348, "learning_rate": 0.00010888942968664417, "loss": 0.0175, "step": 1050 }, { "epoch": 1.575712143928036, "grad_norm": 0.15840014306422295, "learning_rate": 0.00010871557427476583, "loss": 0.0169, "step": 1051 }, { "epoch": 1.5772113943028487, "grad_norm": 0.13291084407135878, "learning_rate": 0.00010854169231373676, "loss": 0.0142, "step": 1052 }, { "epoch": 1.578710644677661, "grad_norm": 0.09889951255469409, "learning_rate": 0.00010836778433323158, "loss": 0.0168, "step": 1053 }, { "epoch": 1.580209895052474, "grad_norm": 0.1018450276304327, "learning_rate": 0.0001081938508630041, "loss": 0.0191, "step": 1054 }, { "epoch": 1.5817091454272862, "grad_norm": 0.11305829451392016, "learning_rate": 0.00010801989243288589, "loss": 0.0183, "step": 1055 }, { "epoch": 1.583208395802099, "grad_norm": 0.0819216552478812, "learning_rate": 0.0001078459095727845, "loss": 0.0131, "step": 1056 }, { "epoch": 1.5847076461769114, "grad_norm": 0.12687192279885026, "learning_rate": 0.00010767190281268187, "loss": 0.0222, "step": 1057 }, { "epoch": 1.5862068965517242, "grad_norm": 0.07362844017008341, "learning_rate": 0.00010749787268263279, "loss": 0.0101, "step": 1058 }, { "epoch": 1.5877061469265368, "grad_norm": 0.06079600135413754, "learning_rate": 0.00010732381971276318, "loss": 0.0097, "step": 1059 }, { "epoch": 1.5892053973013494, "grad_norm": 0.06661318973072765, "learning_rate": 0.0001071497444332686, "loss": 0.0126, "step": 1060 }, { "epoch": 1.590704647676162, "grad_norm": 0.15022551284140978, "learning_rate": 0.00010697564737441252, "loss": 0.0157, "step": 1061 }, { "epoch": 1.5922038980509745, "grad_norm": 0.12196412330052128, "learning_rate": 0.00010680152906652483, "loss": 0.0216, "step": 1062 }, { "epoch": 1.593703148425787, "grad_norm": 0.10057060869798758, "learning_rate": 0.00010662739004000005, "loss": 0.0146, "step": 1063 }, { "epoch": 1.5952023988005997, "grad_norm": 0.10678495246345038, "learning_rate": 0.00010645323082529581, "loss": 0.02, "step": 1064 }, { "epoch": 1.5967016491754122, "grad_norm": 0.07484074937474987, "learning_rate": 0.00010627905195293135, "loss": 0.0115, "step": 1065 }, { "epoch": 1.5982008995502248, "grad_norm": 0.07512912626075581, "learning_rate": 0.00010610485395348571, "loss": 0.0123, "step": 1066 }, { "epoch": 1.5997001499250376, "grad_norm": 0.08792742242003913, "learning_rate": 0.00010593063735759618, "loss": 0.0129, "step": 1067 }, { "epoch": 1.60119940029985, "grad_norm": 0.1622803893742513, "learning_rate": 0.00010575640269595675, "loss": 0.0238, "step": 1068 }, { "epoch": 1.6026986506746628, "grad_norm": 0.13411386012680196, "learning_rate": 0.00010558215049931638, "loss": 0.0255, "step": 1069 }, { "epoch": 1.6041979010494751, "grad_norm": 0.11582998216511321, "learning_rate": 0.00010540788129847756, "loss": 0.0089, "step": 1070 }, { "epoch": 1.605697151424288, "grad_norm": 0.08034952754248964, "learning_rate": 0.0001052335956242944, "loss": 0.0142, "step": 1071 }, { "epoch": 1.6071964017991005, "grad_norm": 0.0763174800207671, "learning_rate": 0.00010505929400767134, "loss": 0.0113, "step": 1072 }, { "epoch": 1.608695652173913, "grad_norm": 0.11861926773103704, "learning_rate": 0.00010488497697956135, "loss": 0.0199, "step": 1073 }, { "epoch": 1.6101949025487257, "grad_norm": 0.10574532949429762, "learning_rate": 0.00010471064507096426, "loss": 0.0187, "step": 1074 }, { "epoch": 1.6116941529235382, "grad_norm": 0.13423987932488682, "learning_rate": 0.00010453629881292538, "loss": 0.0157, "step": 1075 }, { "epoch": 1.6131934032983508, "grad_norm": 0.07764456401541044, "learning_rate": 0.00010436193873653361, "loss": 0.0088, "step": 1076 }, { "epoch": 1.6146926536731634, "grad_norm": 0.0659718362577642, "learning_rate": 0.00010418756537291996, "loss": 0.0125, "step": 1077 }, { "epoch": 1.616191904047976, "grad_norm": 0.0851649381893994, "learning_rate": 0.00010401317925325598, "loss": 0.0116, "step": 1078 }, { "epoch": 1.6176911544227885, "grad_norm": 0.11170558677294636, "learning_rate": 0.00010383878090875201, "loss": 0.0148, "step": 1079 }, { "epoch": 1.6191904047976013, "grad_norm": 0.2357508518298938, "learning_rate": 0.00010366437087065564, "loss": 0.0266, "step": 1080 }, { "epoch": 1.6206896551724137, "grad_norm": 0.15444209087426583, "learning_rate": 0.00010348994967025012, "loss": 0.0215, "step": 1081 }, { "epoch": 1.6221889055472265, "grad_norm": 0.20019270471396725, "learning_rate": 0.00010331551783885263, "loss": 0.0283, "step": 1082 }, { "epoch": 1.6236881559220389, "grad_norm": 0.06468184788830145, "learning_rate": 0.00010314107590781284, "loss": 0.0103, "step": 1083 }, { "epoch": 1.6251874062968517, "grad_norm": 0.06501689028923738, "learning_rate": 0.00010296662440851108, "loss": 0.0089, "step": 1084 }, { "epoch": 1.626686656671664, "grad_norm": 0.07867494972789672, "learning_rate": 0.0001027921638723569, "loss": 0.012, "step": 1085 }, { "epoch": 1.6281859070464768, "grad_norm": 0.053082944965822235, "learning_rate": 0.00010261769483078733, "loss": 0.0088, "step": 1086 }, { "epoch": 1.6296851574212894, "grad_norm": 0.10268869576541034, "learning_rate": 0.00010244321781526533, "loss": 0.0107, "step": 1087 }, { "epoch": 1.631184407796102, "grad_norm": 0.09311046814429251, "learning_rate": 0.00010226873335727816, "loss": 0.0145, "step": 1088 }, { "epoch": 1.6326836581709145, "grad_norm": 0.05118955528054772, "learning_rate": 0.0001020942419883357, "loss": 0.0066, "step": 1089 }, { "epoch": 1.6341829085457271, "grad_norm": 0.06776547119993927, "learning_rate": 0.000101919744239969, "loss": 0.0094, "step": 1090 }, { "epoch": 1.6356821589205397, "grad_norm": 0.0735948110395489, "learning_rate": 0.00010174524064372837, "loss": 0.0083, "step": 1091 }, { "epoch": 1.6371814092953523, "grad_norm": 0.16353843455702943, "learning_rate": 0.00010157073173118208, "loss": 0.0191, "step": 1092 }, { "epoch": 1.638680659670165, "grad_norm": 0.07729308076766099, "learning_rate": 0.00010139621803391455, "loss": 0.0113, "step": 1093 }, { "epoch": 1.6401799100449774, "grad_norm": 0.07065440168839313, "learning_rate": 0.00010122170008352472, "loss": 0.0111, "step": 1094 }, { "epoch": 1.6416791604197902, "grad_norm": 0.08052200609350296, "learning_rate": 0.00010104717841162458, "loss": 0.0086, "step": 1095 }, { "epoch": 1.6431784107946026, "grad_norm": 0.11666520112605404, "learning_rate": 0.0001008726535498374, "loss": 0.0144, "step": 1096 }, { "epoch": 1.6446776611694154, "grad_norm": 0.0883241856124164, "learning_rate": 0.00010069812602979615, "loss": 0.0148, "step": 1097 }, { "epoch": 1.6461769115442277, "grad_norm": 0.05690807601176237, "learning_rate": 0.00010052359638314195, "loss": 0.01, "step": 1098 }, { "epoch": 1.6476761619190405, "grad_norm": 0.11692485410485566, "learning_rate": 0.00010034906514152238, "loss": 0.0204, "step": 1099 }, { "epoch": 1.6491754122938531, "grad_norm": 0.04741744619163783, "learning_rate": 0.00010017453283658984, "loss": 0.0094, "step": 1100 }, { "epoch": 1.6506746626686657, "grad_norm": 0.07678788315580089, "learning_rate": 0.0001, "loss": 0.0135, "step": 1101 }, { "epoch": 1.6521739130434783, "grad_norm": 0.08617460350136832, "learning_rate": 9.98254671634102e-05, "loss": 0.0204, "step": 1102 }, { "epoch": 1.6536731634182908, "grad_norm": 0.03884006559581419, "learning_rate": 9.965093485847767e-05, "loss": 0.0065, "step": 1103 }, { "epoch": 1.6551724137931034, "grad_norm": 0.08381647185517602, "learning_rate": 9.947640361685804e-05, "loss": 0.0135, "step": 1104 }, { "epoch": 1.656671664167916, "grad_norm": 0.08865105073519507, "learning_rate": 9.930187397020386e-05, "loss": 0.0195, "step": 1105 }, { "epoch": 1.6581709145427288, "grad_norm": 0.04051930233354598, "learning_rate": 9.912734645016263e-05, "loss": 0.0059, "step": 1106 }, { "epoch": 1.6596701649175412, "grad_norm": 0.08542703043188983, "learning_rate": 9.895282158837545e-05, "loss": 0.009, "step": 1107 }, { "epoch": 1.661169415292354, "grad_norm": 0.09553707719347451, "learning_rate": 9.877829991647528e-05, "loss": 0.0069, "step": 1108 }, { "epoch": 1.6626686656671663, "grad_norm": 0.17415049236905383, "learning_rate": 9.860378196608549e-05, "loss": 0.0121, "step": 1109 }, { "epoch": 1.6641679160419791, "grad_norm": 0.12950093216787284, "learning_rate": 9.842926826881796e-05, "loss": 0.018, "step": 1110 }, { "epoch": 1.6656671664167915, "grad_norm": 0.06414792177091079, "learning_rate": 9.825475935627165e-05, "loss": 0.0069, "step": 1111 }, { "epoch": 1.6671664167916043, "grad_norm": 0.1994947874307293, "learning_rate": 9.808025576003104e-05, "loss": 0.0318, "step": 1112 }, { "epoch": 1.6686656671664168, "grad_norm": 0.08133150905555013, "learning_rate": 9.790575801166432e-05, "loss": 0.0143, "step": 1113 }, { "epoch": 1.6701649175412294, "grad_norm": 0.16043323092034265, "learning_rate": 9.773126664272186e-05, "loss": 0.0229, "step": 1114 }, { "epoch": 1.671664167916042, "grad_norm": 0.13438362563029066, "learning_rate": 9.755678218473469e-05, "loss": 0.0136, "step": 1115 }, { "epoch": 1.6731634182908546, "grad_norm": 0.07606777859064674, "learning_rate": 9.73823051692127e-05, "loss": 0.0082, "step": 1116 }, { "epoch": 1.6746626686656672, "grad_norm": 0.04420437383237939, "learning_rate": 9.720783612764314e-05, "loss": 0.0085, "step": 1117 }, { "epoch": 1.6761619190404797, "grad_norm": 0.06837921063159548, "learning_rate": 9.703337559148892e-05, "loss": 0.0058, "step": 1118 }, { "epoch": 1.6776611694152923, "grad_norm": 0.10263259805785357, "learning_rate": 9.685892409218717e-05, "loss": 0.0146, "step": 1119 }, { "epoch": 1.6791604197901049, "grad_norm": 0.06638028522741139, "learning_rate": 9.668448216114739e-05, "loss": 0.0056, "step": 1120 }, { "epoch": 1.6806596701649177, "grad_norm": 0.07281657562169337, "learning_rate": 9.651005032974994e-05, "loss": 0.0089, "step": 1121 }, { "epoch": 1.68215892053973, "grad_norm": 0.07288356305521924, "learning_rate": 9.633562912934436e-05, "loss": 0.0055, "step": 1122 }, { "epoch": 1.6836581709145428, "grad_norm": 0.08467351587105704, "learning_rate": 9.616121909124801e-05, "loss": 0.0201, "step": 1123 }, { "epoch": 1.6851574212893552, "grad_norm": 0.08335449765159114, "learning_rate": 9.598682074674405e-05, "loss": 0.0105, "step": 1124 }, { "epoch": 1.686656671664168, "grad_norm": 0.21044561607887044, "learning_rate": 9.581243462708006e-05, "loss": 0.025, "step": 1125 }, { "epoch": 1.6881559220389803, "grad_norm": 0.10557438306733595, "learning_rate": 9.563806126346642e-05, "loss": 0.0165, "step": 1126 }, { "epoch": 1.6896551724137931, "grad_norm": 0.05735360112575192, "learning_rate": 9.546370118707463e-05, "loss": 0.0085, "step": 1127 }, { "epoch": 1.6911544227886057, "grad_norm": 0.08984041269285588, "learning_rate": 9.528935492903575e-05, "loss": 0.0148, "step": 1128 }, { "epoch": 1.6926536731634183, "grad_norm": 0.05724955939727679, "learning_rate": 9.511502302043868e-05, "loss": 0.0065, "step": 1129 }, { "epoch": 1.6941529235382309, "grad_norm": 0.0879374177037415, "learning_rate": 9.494070599232868e-05, "loss": 0.0152, "step": 1130 }, { "epoch": 1.6956521739130435, "grad_norm": 0.10563935529155026, "learning_rate": 9.476640437570562e-05, "loss": 0.0184, "step": 1131 }, { "epoch": 1.697151424287856, "grad_norm": 0.12099493793103591, "learning_rate": 9.459211870152245e-05, "loss": 0.0114, "step": 1132 }, { "epoch": 1.6986506746626686, "grad_norm": 0.1493726682077126, "learning_rate": 9.441784950068362e-05, "loss": 0.0179, "step": 1133 }, { "epoch": 1.7001499250374814, "grad_norm": 0.06855320485306697, "learning_rate": 9.424359730404329e-05, "loss": 0.0125, "step": 1134 }, { "epoch": 1.7016491754122938, "grad_norm": 0.14082186154079793, "learning_rate": 9.406936264240386e-05, "loss": 0.0117, "step": 1135 }, { "epoch": 1.7031484257871066, "grad_norm": 0.12243397983876568, "learning_rate": 9.38951460465143e-05, "loss": 0.0163, "step": 1136 }, { "epoch": 1.704647676161919, "grad_norm": 0.07295698279479494, "learning_rate": 9.372094804706867e-05, "loss": 0.0142, "step": 1137 }, { "epoch": 1.7061469265367317, "grad_norm": 0.07864559697537851, "learning_rate": 9.354676917470422e-05, "loss": 0.0121, "step": 1138 }, { "epoch": 1.707646176911544, "grad_norm": 0.14262164347217002, "learning_rate": 9.337260996000002e-05, "loss": 0.0205, "step": 1139 }, { "epoch": 1.7091454272863569, "grad_norm": 0.1533650338506401, "learning_rate": 9.319847093347522e-05, "loss": 0.0154, "step": 1140 }, { "epoch": 1.7106446776611695, "grad_norm": 0.07096001994126981, "learning_rate": 9.302435262558747e-05, "loss": 0.0114, "step": 1141 }, { "epoch": 1.712143928035982, "grad_norm": 0.087774881508882, "learning_rate": 9.285025556673141e-05, "loss": 0.0043, "step": 1142 }, { "epoch": 1.7136431784107946, "grad_norm": 0.12753925625272527, "learning_rate": 9.267618028723686e-05, "loss": 0.0108, "step": 1143 }, { "epoch": 1.7151424287856072, "grad_norm": 0.10939037079915889, "learning_rate": 9.250212731736726e-05, "loss": 0.0153, "step": 1144 }, { "epoch": 1.7166416791604198, "grad_norm": 0.10489461806582333, "learning_rate": 9.232809718731814e-05, "loss": 0.0161, "step": 1145 }, { "epoch": 1.7181409295352323, "grad_norm": 0.07579596628726051, "learning_rate": 9.215409042721552e-05, "loss": 0.0115, "step": 1146 }, { "epoch": 1.7196401799100451, "grad_norm": 0.06387563602644145, "learning_rate": 9.198010756711412e-05, "loss": 0.0111, "step": 1147 }, { "epoch": 1.7211394302848575, "grad_norm": 0.08515512442384568, "learning_rate": 9.180614913699592e-05, "loss": 0.0115, "step": 1148 }, { "epoch": 1.7226386806596703, "grad_norm": 0.08561417637376921, "learning_rate": 9.163221566676847e-05, "loss": 0.009, "step": 1149 }, { "epoch": 1.7241379310344827, "grad_norm": 0.10607108021738235, "learning_rate": 9.145830768626327e-05, "loss": 0.0181, "step": 1150 }, { "epoch": 1.7256371814092955, "grad_norm": 0.1611719445615411, "learning_rate": 9.128442572523417e-05, "loss": 0.0087, "step": 1151 }, { "epoch": 1.7271364317841078, "grad_norm": 0.04705216120726678, "learning_rate": 9.111057031335585e-05, "loss": 0.0045, "step": 1152 }, { "epoch": 1.7286356821589206, "grad_norm": 0.07828025749540468, "learning_rate": 9.093674198022201e-05, "loss": 0.012, "step": 1153 }, { "epoch": 1.7301349325337332, "grad_norm": 0.08108251835589797, "learning_rate": 9.076294125534383e-05, "loss": 0.0049, "step": 1154 }, { "epoch": 1.7316341829085458, "grad_norm": 0.07381251899456927, "learning_rate": 9.058916866814858e-05, "loss": 0.0079, "step": 1155 }, { "epoch": 1.7331334332833583, "grad_norm": 0.05566501298258808, "learning_rate": 9.04154247479776e-05, "loss": 0.0074, "step": 1156 }, { "epoch": 1.734632683658171, "grad_norm": 0.07840451984596064, "learning_rate": 9.024171002408506e-05, "loss": 0.0105, "step": 1157 }, { "epoch": 1.7361319340329835, "grad_norm": 0.06648110796795341, "learning_rate": 9.006802502563612e-05, "loss": 0.0077, "step": 1158 }, { "epoch": 1.737631184407796, "grad_norm": 0.04739088697599641, "learning_rate": 8.989437028170537e-05, "loss": 0.0034, "step": 1159 }, { "epoch": 1.7391304347826086, "grad_norm": 0.1271185732601886, "learning_rate": 8.972074632127533e-05, "loss": 0.0206, "step": 1160 }, { "epoch": 1.7406296851574212, "grad_norm": 0.15903655821127385, "learning_rate": 8.954715367323468e-05, "loss": 0.0426, "step": 1161 }, { "epoch": 1.742128935532234, "grad_norm": 0.10094008224842381, "learning_rate": 8.93735928663767e-05, "loss": 0.0199, "step": 1162 }, { "epoch": 1.7436281859070464, "grad_norm": 0.07079249377199838, "learning_rate": 8.920006442939772e-05, "loss": 0.01, "step": 1163 }, { "epoch": 1.7451274362818592, "grad_norm": 0.05078697634845653, "learning_rate": 8.902656889089548e-05, "loss": 0.0062, "step": 1164 }, { "epoch": 1.7466266866566715, "grad_norm": 0.05448053623416766, "learning_rate": 8.885310677936746e-05, "loss": 0.0053, "step": 1165 }, { "epoch": 1.7481259370314843, "grad_norm": 0.05217380091305068, "learning_rate": 8.867967862320934e-05, "loss": 0.0067, "step": 1166 }, { "epoch": 1.7496251874062967, "grad_norm": 0.08488127237102319, "learning_rate": 8.850628495071336e-05, "loss": 0.0104, "step": 1167 }, { "epoch": 1.7511244377811095, "grad_norm": 0.12524978803940454, "learning_rate": 8.833292629006668e-05, "loss": 0.0108, "step": 1168 }, { "epoch": 1.752623688155922, "grad_norm": 0.09170930081644955, "learning_rate": 8.81596031693499e-05, "loss": 0.0124, "step": 1169 }, { "epoch": 1.7541229385307346, "grad_norm": 0.0646182404385978, "learning_rate": 8.79863161165353e-05, "loss": 0.0099, "step": 1170 }, { "epoch": 1.7556221889055472, "grad_norm": 0.060318759018513406, "learning_rate": 8.781306565948528e-05, "loss": 0.0082, "step": 1171 }, { "epoch": 1.7571214392803598, "grad_norm": 0.07161845397887222, "learning_rate": 8.763985232595075e-05, "loss": 0.007, "step": 1172 }, { "epoch": 1.7586206896551724, "grad_norm": 0.103642095909335, "learning_rate": 8.746667664356956e-05, "loss": 0.0161, "step": 1173 }, { "epoch": 1.760119940029985, "grad_norm": 0.08550791524475224, "learning_rate": 8.729353913986496e-05, "loss": 0.0147, "step": 1174 }, { "epoch": 1.7616191904047978, "grad_norm": 0.06894588907543088, "learning_rate": 8.712044034224374e-05, "loss": 0.0101, "step": 1175 }, { "epoch": 1.76311844077961, "grad_norm": 0.09780901112803982, "learning_rate": 8.694738077799488e-05, "loss": 0.0181, "step": 1176 }, { "epoch": 1.764617691154423, "grad_norm": 0.07230873928187732, "learning_rate": 8.677436097428775e-05, "loss": 0.0061, "step": 1177 }, { "epoch": 1.7661169415292353, "grad_norm": 0.13013689151465993, "learning_rate": 8.660138145817079e-05, "loss": 0.0113, "step": 1178 }, { "epoch": 1.767616191904048, "grad_norm": 0.06939024763422029, "learning_rate": 8.642844275656957e-05, "loss": 0.0105, "step": 1179 }, { "epoch": 1.7691154422788604, "grad_norm": 0.15077017041092736, "learning_rate": 8.625554539628535e-05, "loss": 0.031, "step": 1180 }, { "epoch": 1.7706146926536732, "grad_norm": 0.12499087340016703, "learning_rate": 8.608268990399349e-05, "loss": 0.0277, "step": 1181 }, { "epoch": 1.7721139430284858, "grad_norm": 0.11591487119439382, "learning_rate": 8.590987680624174e-05, "loss": 0.0157, "step": 1182 }, { "epoch": 1.7736131934032984, "grad_norm": 0.06002291691845019, "learning_rate": 8.573710662944885e-05, "loss": 0.0082, "step": 1183 }, { "epoch": 1.775112443778111, "grad_norm": 0.1038340304012269, "learning_rate": 8.55643798999027e-05, "loss": 0.0119, "step": 1184 }, { "epoch": 1.7766116941529235, "grad_norm": 0.09039360537283966, "learning_rate": 8.539169714375885e-05, "loss": 0.0138, "step": 1185 }, { "epoch": 1.778110944527736, "grad_norm": 0.08968147616663716, "learning_rate": 8.521905888703893e-05, "loss": 0.0076, "step": 1186 }, { "epoch": 1.7796101949025487, "grad_norm": 0.06865223310681502, "learning_rate": 8.504646565562906e-05, "loss": 0.0119, "step": 1187 }, { "epoch": 1.7811094452773615, "grad_norm": 0.1473984760272428, "learning_rate": 8.487391797527808e-05, "loss": 0.0165, "step": 1188 }, { "epoch": 1.7826086956521738, "grad_norm": 0.06834986338018911, "learning_rate": 8.47014163715962e-05, "loss": 0.0072, "step": 1189 }, { "epoch": 1.7841079460269866, "grad_norm": 0.06542032785241195, "learning_rate": 8.452896137005321e-05, "loss": 0.0091, "step": 1190 }, { "epoch": 1.785607196401799, "grad_norm": 0.15656986967497505, "learning_rate": 8.435655349597689e-05, "loss": 0.0153, "step": 1191 }, { "epoch": 1.7871064467766118, "grad_norm": 0.05047228873977786, "learning_rate": 8.418419327455164e-05, "loss": 0.0066, "step": 1192 }, { "epoch": 1.7886056971514241, "grad_norm": 0.05650765639240601, "learning_rate": 8.401188123081653e-05, "loss": 0.0074, "step": 1193 }, { "epoch": 1.790104947526237, "grad_norm": 0.09070303775142517, "learning_rate": 8.383961788966391e-05, "loss": 0.0138, "step": 1194 }, { "epoch": 1.7916041979010495, "grad_norm": 0.12727798124455994, "learning_rate": 8.366740377583781e-05, "loss": 0.0164, "step": 1195 }, { "epoch": 1.793103448275862, "grad_norm": 0.08435936567425999, "learning_rate": 8.349523941393224e-05, "loss": 0.0142, "step": 1196 }, { "epoch": 1.7946026986506747, "grad_norm": 0.081825272446604, "learning_rate": 8.332312532838978e-05, "loss": 0.0228, "step": 1197 }, { "epoch": 1.7961019490254873, "grad_norm": 0.048776431818418335, "learning_rate": 8.315106204349976e-05, "loss": 0.0071, "step": 1198 }, { "epoch": 1.7976011994002998, "grad_norm": 0.10574381543468121, "learning_rate": 8.297905008339677e-05, "loss": 0.0132, "step": 1199 }, { "epoch": 1.7991004497751124, "grad_norm": 0.17066204868006726, "learning_rate": 8.280708997205904e-05, "loss": 0.013, "step": 1200 }, { "epoch": 1.800599700149925, "grad_norm": 0.10668783815163481, "learning_rate": 8.263518223330697e-05, "loss": 0.0095, "step": 1201 }, { "epoch": 1.8020989505247376, "grad_norm": 0.16567996473627022, "learning_rate": 8.24633273908013e-05, "loss": 0.0211, "step": 1202 }, { "epoch": 1.8035982008995504, "grad_norm": 0.08443187319668671, "learning_rate": 8.229152596804168e-05, "loss": 0.0149, "step": 1203 }, { "epoch": 1.8050974512743627, "grad_norm": 0.08806920008710213, "learning_rate": 8.211977848836506e-05, "loss": 0.0088, "step": 1204 }, { "epoch": 1.8065967016491755, "grad_norm": 0.11415903061944807, "learning_rate": 8.194808547494401e-05, "loss": 0.0199, "step": 1205 }, { "epoch": 1.8080959520239879, "grad_norm": 0.09961276004274798, "learning_rate": 8.177644745078526e-05, "loss": 0.0074, "step": 1206 }, { "epoch": 1.8095952023988007, "grad_norm": 0.0891187100998618, "learning_rate": 8.160486493872798e-05, "loss": 0.0104, "step": 1207 }, { "epoch": 1.811094452773613, "grad_norm": 0.12230491937645052, "learning_rate": 8.14333384614423e-05, "loss": 0.0144, "step": 1208 }, { "epoch": 1.8125937031484258, "grad_norm": 0.05271611218754871, "learning_rate": 8.126186854142752e-05, "loss": 0.0113, "step": 1209 }, { "epoch": 1.8140929535232384, "grad_norm": 0.12062545658852573, "learning_rate": 8.109045570101086e-05, "loss": 0.0184, "step": 1210 }, { "epoch": 1.815592203898051, "grad_norm": 0.13604210998453206, "learning_rate": 8.091910046234552e-05, "loss": 0.023, "step": 1211 }, { "epoch": 1.8170914542728636, "grad_norm": 0.10981765208088426, "learning_rate": 8.074780334740928e-05, "loss": 0.0097, "step": 1212 }, { "epoch": 1.8185907046476761, "grad_norm": 0.13794783867457727, "learning_rate": 8.057656487800282e-05, "loss": 0.0267, "step": 1213 }, { "epoch": 1.8200899550224887, "grad_norm": 0.1645787389155614, "learning_rate": 8.040538557574822e-05, "loss": 0.0257, "step": 1214 }, { "epoch": 1.8215892053973013, "grad_norm": 0.1349840201186409, "learning_rate": 8.023426596208739e-05, "loss": 0.0135, "step": 1215 }, { "epoch": 1.823088455772114, "grad_norm": 0.09749351855460471, "learning_rate": 8.00632065582803e-05, "loss": 0.0118, "step": 1216 }, { "epoch": 1.8245877061469264, "grad_norm": 0.07799943350266138, "learning_rate": 7.989220788540355e-05, "loss": 0.0081, "step": 1217 }, { "epoch": 1.8260869565217392, "grad_norm": 0.09827289301737527, "learning_rate": 7.972127046434878e-05, "loss": 0.0093, "step": 1218 }, { "epoch": 1.8275862068965516, "grad_norm": 0.09805007129974536, "learning_rate": 7.955039481582097e-05, "loss": 0.015, "step": 1219 }, { "epoch": 1.8290854572713644, "grad_norm": 0.09422086692048044, "learning_rate": 7.937958146033705e-05, "loss": 0.0087, "step": 1220 }, { "epoch": 1.8305847076461768, "grad_norm": 0.0712763271172262, "learning_rate": 7.920883091822408e-05, "loss": 0.0131, "step": 1221 }, { "epoch": 1.8320839580209896, "grad_norm": 0.11306696361521645, "learning_rate": 7.903814370961784e-05, "loss": 0.0125, "step": 1222 }, { "epoch": 1.8335832083958021, "grad_norm": 0.08672321400900379, "learning_rate": 7.886752035446114e-05, "loss": 0.0133, "step": 1223 }, { "epoch": 1.8350824587706147, "grad_norm": 0.1672675013757812, "learning_rate": 7.869696137250235e-05, "loss": 0.0309, "step": 1224 }, { "epoch": 1.8365817091454273, "grad_norm": 0.10355505496804655, "learning_rate": 7.852646728329368e-05, "loss": 0.0104, "step": 1225 }, { "epoch": 1.8380809595202399, "grad_norm": 0.11892893681660466, "learning_rate": 7.835603860618972e-05, "loss": 0.0142, "step": 1226 }, { "epoch": 1.8395802098950524, "grad_norm": 0.14739440025777525, "learning_rate": 7.818567586034577e-05, "loss": 0.0204, "step": 1227 }, { "epoch": 1.841079460269865, "grad_norm": 0.08697589155480914, "learning_rate": 7.801537956471624e-05, "loss": 0.0095, "step": 1228 }, { "epoch": 1.8425787106446778, "grad_norm": 0.0581146762167273, "learning_rate": 7.784515023805328e-05, "loss": 0.0078, "step": 1229 }, { "epoch": 1.8440779610194902, "grad_norm": 0.08434253900090809, "learning_rate": 7.767498839890488e-05, "loss": 0.0075, "step": 1230 }, { "epoch": 1.845577211394303, "grad_norm": 0.10900161640713057, "learning_rate": 7.750489456561352e-05, "loss": 0.0191, "step": 1231 }, { "epoch": 1.8470764617691153, "grad_norm": 0.10994173678131566, "learning_rate": 7.733486925631447e-05, "loss": 0.0115, "step": 1232 }, { "epoch": 1.8485757121439281, "grad_norm": 0.1171064895868367, "learning_rate": 7.716491298893442e-05, "loss": 0.014, "step": 1233 }, { "epoch": 1.8500749625187405, "grad_norm": 0.10594334669279865, "learning_rate": 7.699502628118958e-05, "loss": 0.0144, "step": 1234 }, { "epoch": 1.8515742128935533, "grad_norm": 0.054006355912338475, "learning_rate": 7.682520965058428e-05, "loss": 0.0065, "step": 1235 }, { "epoch": 1.8530734632683659, "grad_norm": 0.05698555933294013, "learning_rate": 7.66554636144095e-05, "loss": 0.0099, "step": 1236 }, { "epoch": 1.8545727136431784, "grad_norm": 0.15549806710918523, "learning_rate": 7.6485788689741e-05, "loss": 0.0202, "step": 1237 }, { "epoch": 1.856071964017991, "grad_norm": 0.1409960439001676, "learning_rate": 7.631618539343814e-05, "loss": 0.0123, "step": 1238 }, { "epoch": 1.8575712143928036, "grad_norm": 0.10436639288234584, "learning_rate": 7.614665424214193e-05, "loss": 0.0102, "step": 1239 }, { "epoch": 1.8590704647676162, "grad_norm": 0.3067562050568254, "learning_rate": 7.597719575227364e-05, "loss": 0.0202, "step": 1240 }, { "epoch": 1.8605697151424287, "grad_norm": 0.09367926022511773, "learning_rate": 7.580781044003324e-05, "loss": 0.0173, "step": 1241 }, { "epoch": 1.8620689655172413, "grad_norm": 0.09079090097092132, "learning_rate": 7.563849882139776e-05, "loss": 0.0101, "step": 1242 }, { "epoch": 1.863568215892054, "grad_norm": 0.10042340395367601, "learning_rate": 7.546926141211974e-05, "loss": 0.0101, "step": 1243 }, { "epoch": 1.8650674662668667, "grad_norm": 0.05249586616928009, "learning_rate": 7.530009872772572e-05, "loss": 0.0073, "step": 1244 }, { "epoch": 1.866566716641679, "grad_norm": 0.06628730417621559, "learning_rate": 7.513101128351454e-05, "loss": 0.006, "step": 1245 }, { "epoch": 1.8680659670164919, "grad_norm": 0.09985779934896404, "learning_rate": 7.496199959455584e-05, "loss": 0.009, "step": 1246 }, { "epoch": 1.8695652173913042, "grad_norm": 0.07189435277325107, "learning_rate": 7.479306417568864e-05, "loss": 0.0079, "step": 1247 }, { "epoch": 1.871064467766117, "grad_norm": 0.04129148635536298, "learning_rate": 7.462420554151944e-05, "loss": 0.006, "step": 1248 }, { "epoch": 1.8725637181409296, "grad_norm": 0.08500137505977622, "learning_rate": 7.445542420642097e-05, "loss": 0.0147, "step": 1249 }, { "epoch": 1.8740629685157422, "grad_norm": 0.09066062675805478, "learning_rate": 7.42867206845304e-05, "loss": 0.0127, "step": 1250 }, { "epoch": 1.8755622188905547, "grad_norm": 0.10120598113171132, "learning_rate": 7.411809548974792e-05, "loss": 0.0131, "step": 1251 }, { "epoch": 1.8770614692653673, "grad_norm": 0.050739274353256675, "learning_rate": 7.394954913573517e-05, "loss": 0.0056, "step": 1252 }, { "epoch": 1.87856071964018, "grad_norm": 0.07245322597365843, "learning_rate": 7.378108213591355e-05, "loss": 0.0059, "step": 1253 }, { "epoch": 1.8800599700149925, "grad_norm": 0.15359008958070705, "learning_rate": 7.361269500346274e-05, "loss": 0.0164, "step": 1254 }, { "epoch": 1.881559220389805, "grad_norm": 0.19317251107868752, "learning_rate": 7.344438825131911e-05, "loss": 0.0189, "step": 1255 }, { "epoch": 1.8830584707646176, "grad_norm": 0.061858025534464414, "learning_rate": 7.327616239217431e-05, "loss": 0.0043, "step": 1256 }, { "epoch": 1.8845577211394304, "grad_norm": 0.13091196479249864, "learning_rate": 7.310801793847344e-05, "loss": 0.0159, "step": 1257 }, { "epoch": 1.8860569715142428, "grad_norm": 0.0544033205013549, "learning_rate": 7.293995540241366e-05, "loss": 0.0056, "step": 1258 }, { "epoch": 1.8875562218890556, "grad_norm": 0.0783279329482906, "learning_rate": 7.277197529594257e-05, "loss": 0.0085, "step": 1259 }, { "epoch": 1.889055472263868, "grad_norm": 0.060875813189406686, "learning_rate": 7.260407813075676e-05, "loss": 0.01, "step": 1260 }, { "epoch": 1.8905547226386807, "grad_norm": 0.08212593368295347, "learning_rate": 7.243626441830009e-05, "loss": 0.0106, "step": 1261 }, { "epoch": 1.892053973013493, "grad_norm": 0.14448153096893138, "learning_rate": 7.226853466976222e-05, "loss": 0.0316, "step": 1262 }, { "epoch": 1.893553223388306, "grad_norm": 0.11583432656692919, "learning_rate": 7.210088939607708e-05, "loss": 0.0103, "step": 1263 }, { "epoch": 1.8950524737631185, "grad_norm": 0.192579385165103, "learning_rate": 7.193332910792124e-05, "loss": 0.0236, "step": 1264 }, { "epoch": 1.896551724137931, "grad_norm": 0.06822480918750315, "learning_rate": 7.176585431571235e-05, "loss": 0.0165, "step": 1265 }, { "epoch": 1.8980509745127436, "grad_norm": 0.08350988856289665, "learning_rate": 7.159846552960774e-05, "loss": 0.0149, "step": 1266 }, { "epoch": 1.8995502248875562, "grad_norm": 0.09708205671430116, "learning_rate": 7.143116325950265e-05, "loss": 0.0196, "step": 1267 }, { "epoch": 1.9010494752623688, "grad_norm": 0.06975686550376206, "learning_rate": 7.126394801502882e-05, "loss": 0.008, "step": 1268 }, { "epoch": 1.9025487256371814, "grad_norm": 0.0937817538752956, "learning_rate": 7.109682030555283e-05, "loss": 0.009, "step": 1269 }, { "epoch": 1.9040479760119942, "grad_norm": 0.09751887136446406, "learning_rate": 7.092978064017475e-05, "loss": 0.0155, "step": 1270 }, { "epoch": 1.9055472263868065, "grad_norm": 0.12135514343914165, "learning_rate": 7.076282952772633e-05, "loss": 0.0131, "step": 1271 }, { "epoch": 1.9070464767616193, "grad_norm": 0.14474648955209451, "learning_rate": 7.059596747676962e-05, "loss": 0.0186, "step": 1272 }, { "epoch": 1.9085457271364317, "grad_norm": 0.09456198132154989, "learning_rate": 7.042919499559537e-05, "loss": 0.0109, "step": 1273 }, { "epoch": 1.9100449775112445, "grad_norm": 0.06368532914601598, "learning_rate": 7.026251259222141e-05, "loss": 0.011, "step": 1274 }, { "epoch": 1.9115442278860568, "grad_norm": 0.07815558315112578, "learning_rate": 7.009592077439134e-05, "loss": 0.0096, "step": 1275 }, { "epoch": 1.9130434782608696, "grad_norm": 0.07587411820029523, "learning_rate": 6.992942004957271e-05, "loss": 0.0094, "step": 1276 }, { "epoch": 1.9145427286356822, "grad_norm": 0.14301760261232488, "learning_rate": 6.976301092495556e-05, "loss": 0.0195, "step": 1277 }, { "epoch": 1.9160419790104948, "grad_norm": 0.05601598197369331, "learning_rate": 6.959669390745097e-05, "loss": 0.0077, "step": 1278 }, { "epoch": 1.9175412293853074, "grad_norm": 0.17913703313927748, "learning_rate": 6.943046950368944e-05, "loss": 0.0207, "step": 1279 }, { "epoch": 1.91904047976012, "grad_norm": 0.04621976604774044, "learning_rate": 6.92643382200193e-05, "loss": 0.0051, "step": 1280 }, { "epoch": 1.9205397301349325, "grad_norm": 0.059822571193335644, "learning_rate": 6.909830056250527e-05, "loss": 0.0073, "step": 1281 }, { "epoch": 1.922038980509745, "grad_norm": 0.050398816706857336, "learning_rate": 6.893235703692685e-05, "loss": 0.006, "step": 1282 }, { "epoch": 1.9235382308845579, "grad_norm": 0.15403140307959207, "learning_rate": 6.876650814877674e-05, "loss": 0.0233, "step": 1283 }, { "epoch": 1.9250374812593702, "grad_norm": 0.08002270556903888, "learning_rate": 6.860075440325951e-05, "loss": 0.0127, "step": 1284 }, { "epoch": 1.926536731634183, "grad_norm": 0.09217416653554218, "learning_rate": 6.843509630528977e-05, "loss": 0.0078, "step": 1285 }, { "epoch": 1.9280359820089954, "grad_norm": 0.08217684285186015, "learning_rate": 6.826953435949081e-05, "loss": 0.0087, "step": 1286 }, { "epoch": 1.9295352323838082, "grad_norm": 0.1135626679946753, "learning_rate": 6.8104069070193e-05, "loss": 0.0106, "step": 1287 }, { "epoch": 1.9310344827586206, "grad_norm": 0.11547720907926104, "learning_rate": 6.793870094143238e-05, "loss": 0.0104, "step": 1288 }, { "epoch": 1.9325337331334334, "grad_norm": 0.06982269585114415, "learning_rate": 6.77734304769489e-05, "loss": 0.0078, "step": 1289 }, { "epoch": 1.934032983508246, "grad_norm": 0.15410842004892794, "learning_rate": 6.760825818018508e-05, "loss": 0.0275, "step": 1290 }, { "epoch": 1.9355322338830585, "grad_norm": 0.12670821336659963, "learning_rate": 6.744318455428436e-05, "loss": 0.0159, "step": 1291 }, { "epoch": 1.937031484257871, "grad_norm": 0.0885086070508493, "learning_rate": 6.727821010208962e-05, "loss": 0.0111, "step": 1292 }, { "epoch": 1.9385307346326837, "grad_norm": 0.0632145088319827, "learning_rate": 6.711333532614168e-05, "loss": 0.0098, "step": 1293 }, { "epoch": 1.9400299850074962, "grad_norm": 0.10005119826138181, "learning_rate": 6.69485607286777e-05, "loss": 0.017, "step": 1294 }, { "epoch": 1.9415292353823088, "grad_norm": 0.0719843361060488, "learning_rate": 6.67838868116297e-05, "loss": 0.0099, "step": 1295 }, { "epoch": 1.9430284857571214, "grad_norm": 0.09905966753940625, "learning_rate": 6.661931407662292e-05, "loss": 0.0118, "step": 1296 }, { "epoch": 1.944527736131934, "grad_norm": 0.07485985195447457, "learning_rate": 6.64548430249745e-05, "loss": 0.0095, "step": 1297 }, { "epoch": 1.9460269865067468, "grad_norm": 0.10312354721673073, "learning_rate": 6.62904741576918e-05, "loss": 0.0075, "step": 1298 }, { "epoch": 1.9475262368815591, "grad_norm": 0.0926401569025068, "learning_rate": 6.612620797547087e-05, "loss": 0.0106, "step": 1299 }, { "epoch": 1.949025487256372, "grad_norm": 0.08554451492013519, "learning_rate": 6.5962044978695e-05, "loss": 0.0071, "step": 1300 }, { "epoch": 1.9505247376311843, "grad_norm": 0.07530310677555721, "learning_rate": 6.579798566743314e-05, "loss": 0.0059, "step": 1301 }, { "epoch": 1.952023988005997, "grad_norm": 0.09012179453580464, "learning_rate": 6.56340305414384e-05, "loss": 0.007, "step": 1302 }, { "epoch": 1.9535232383808094, "grad_norm": 0.0732412409351987, "learning_rate": 6.547018010014654e-05, "loss": 0.009, "step": 1303 }, { "epoch": 1.9550224887556222, "grad_norm": 0.1122090001398675, "learning_rate": 6.530643484267443e-05, "loss": 0.0123, "step": 1304 }, { "epoch": 1.9565217391304348, "grad_norm": 0.05854864784619732, "learning_rate": 6.51427952678185e-05, "loss": 0.0088, "step": 1305 }, { "epoch": 1.9580209895052474, "grad_norm": 0.17221127456654123, "learning_rate": 6.497926187405326e-05, "loss": 0.0265, "step": 1306 }, { "epoch": 1.95952023988006, "grad_norm": 0.11007336458310958, "learning_rate": 6.481583515952983e-05, "loss": 0.0195, "step": 1307 }, { "epoch": 1.9610194902548725, "grad_norm": 0.1603764107496845, "learning_rate": 6.465251562207431e-05, "loss": 0.0127, "step": 1308 }, { "epoch": 1.9625187406296851, "grad_norm": 0.10303600158564385, "learning_rate": 6.448930375918631e-05, "loss": 0.0091, "step": 1309 }, { "epoch": 1.9640179910044977, "grad_norm": 0.1059759977454014, "learning_rate": 6.432620006803746e-05, "loss": 0.0147, "step": 1310 }, { "epoch": 1.9655172413793105, "grad_norm": 0.13503577362942953, "learning_rate": 6.416320504546997e-05, "loss": 0.0247, "step": 1311 }, { "epoch": 1.9670164917541229, "grad_norm": 0.06892530634050763, "learning_rate": 6.40003191879949e-05, "loss": 0.01, "step": 1312 }, { "epoch": 1.9685157421289357, "grad_norm": 0.07291043016423708, "learning_rate": 6.383754299179079e-05, "loss": 0.0097, "step": 1313 }, { "epoch": 1.970014992503748, "grad_norm": 0.13914268368623947, "learning_rate": 6.367487695270217e-05, "loss": 0.0164, "step": 1314 }, { "epoch": 1.9715142428785608, "grad_norm": 0.0714584621260185, "learning_rate": 6.351232156623803e-05, "loss": 0.01, "step": 1315 }, { "epoch": 1.9730134932533732, "grad_norm": 0.12977132489990004, "learning_rate": 6.334987732757029e-05, "loss": 0.0124, "step": 1316 }, { "epoch": 1.974512743628186, "grad_norm": 0.11389052086228867, "learning_rate": 6.318754473153221e-05, "loss": 0.0191, "step": 1317 }, { "epoch": 1.9760119940029985, "grad_norm": 0.13954486292599097, "learning_rate": 6.302532427261709e-05, "loss": 0.0178, "step": 1318 }, { "epoch": 1.9775112443778111, "grad_norm": 0.2039430304512413, "learning_rate": 6.286321644497655e-05, "loss": 0.0188, "step": 1319 }, { "epoch": 1.9790104947526237, "grad_norm": 0.04606291429944488, "learning_rate": 6.27012217424191e-05, "loss": 0.003, "step": 1320 }, { "epoch": 1.9805097451274363, "grad_norm": 0.07084029408482302, "learning_rate": 6.25393406584088e-05, "loss": 0.0082, "step": 1321 }, { "epoch": 1.9820089955022488, "grad_norm": 0.06195579957195022, "learning_rate": 6.237757368606345e-05, "loss": 0.0056, "step": 1322 }, { "epoch": 1.9835082458770614, "grad_norm": 0.10572839742430142, "learning_rate": 6.22159213181533e-05, "loss": 0.0227, "step": 1323 }, { "epoch": 1.9850074962518742, "grad_norm": 0.0953707827939962, "learning_rate": 6.205438404709947e-05, "loss": 0.0135, "step": 1324 }, { "epoch": 1.9865067466266866, "grad_norm": 0.10133284635764458, "learning_rate": 6.18929623649726e-05, "loss": 0.0167, "step": 1325 }, { "epoch": 1.9880059970014994, "grad_norm": 0.1046927216676533, "learning_rate": 6.173165676349103e-05, "loss": 0.0153, "step": 1326 }, { "epoch": 1.9895052473763117, "grad_norm": 0.09125433955424463, "learning_rate": 6.157046773401964e-05, "loss": 0.0098, "step": 1327 }, { "epoch": 1.9910044977511245, "grad_norm": 0.07206163283122839, "learning_rate": 6.140939576756817e-05, "loss": 0.0088, "step": 1328 }, { "epoch": 1.992503748125937, "grad_norm": 0.09624780783173856, "learning_rate": 6.12484413547897e-05, "loss": 0.0097, "step": 1329 }, { "epoch": 1.9940029985007497, "grad_norm": 0.08193537478772897, "learning_rate": 6.108760498597938e-05, "loss": 0.0078, "step": 1330 }, { "epoch": 1.9955022488755623, "grad_norm": 0.10581177346811316, "learning_rate": 6.092688715107264e-05, "loss": 0.0201, "step": 1331 }, { "epoch": 1.9970014992503748, "grad_norm": 0.08582699378434917, "learning_rate": 6.076628833964388e-05, "loss": 0.0087, "step": 1332 }, { "epoch": 1.9985007496251874, "grad_norm": 0.07360485364749729, "learning_rate": 6.0605809040904894e-05, "loss": 0.0089, "step": 1333 }, { "epoch": 2.0, "grad_norm": 0.10066888348562322, "learning_rate": 6.0445449743703516e-05, "loss": 0.0169, "step": 1334 }, { "epoch": 2.0, "eval_loss": 0.02858872152864933, "eval_runtime": 481.6247, "eval_samples_per_second": 21.544, "eval_steps_per_second": 1.348, "step": 1334 }, { "epoch": 2.001499250374813, "grad_norm": 0.05691507076831063, "learning_rate": 6.0285210936521955e-05, "loss": 0.0081, "step": 1335 }, { "epoch": 2.002998500749625, "grad_norm": 0.047619108961378546, "learning_rate": 6.012509310747538e-05, "loss": 0.0059, "step": 1336 }, { "epoch": 2.004497751124438, "grad_norm": 0.12728571721122714, "learning_rate": 5.9965096744310526e-05, "loss": 0.0128, "step": 1337 }, { "epoch": 2.0059970014992503, "grad_norm": 0.05533519939522902, "learning_rate": 5.9805222334404e-05, "loss": 0.0063, "step": 1338 }, { "epoch": 2.007496251874063, "grad_norm": 0.058973041277131814, "learning_rate": 5.964547036476099e-05, "loss": 0.0078, "step": 1339 }, { "epoch": 2.0089955022488755, "grad_norm": 0.053801378638280425, "learning_rate": 5.9485841322013755e-05, "loss": 0.0045, "step": 1340 }, { "epoch": 2.0104947526236883, "grad_norm": 0.08346927650771466, "learning_rate": 5.9326335692419995e-05, "loss": 0.0112, "step": 1341 }, { "epoch": 2.0119940029985006, "grad_norm": 0.06796501482552034, "learning_rate": 5.916695396186154e-05, "loss": 0.0059, "step": 1342 }, { "epoch": 2.0134932533733134, "grad_norm": 0.02845322851169814, "learning_rate": 5.900769661584272e-05, "loss": 0.0051, "step": 1343 }, { "epoch": 2.0149925037481258, "grad_norm": 0.09502613453475849, "learning_rate": 5.884856413948913e-05, "loss": 0.0157, "step": 1344 }, { "epoch": 2.0164917541229386, "grad_norm": 0.033394665506260045, "learning_rate": 5.868955701754584e-05, "loss": 0.0027, "step": 1345 }, { "epoch": 2.017991004497751, "grad_norm": 0.08324949873495568, "learning_rate": 5.853067573437612e-05, "loss": 0.0101, "step": 1346 }, { "epoch": 2.0194902548725637, "grad_norm": 0.06530137085651484, "learning_rate": 5.83719207739599e-05, "loss": 0.0083, "step": 1347 }, { "epoch": 2.0209895052473765, "grad_norm": 0.057351896571536574, "learning_rate": 5.82132926198923e-05, "loss": 0.0057, "step": 1348 }, { "epoch": 2.022488755622189, "grad_norm": 0.07266435520449091, "learning_rate": 5.805479175538229e-05, "loss": 0.004, "step": 1349 }, { "epoch": 2.0239880059970017, "grad_norm": 0.046130085433988374, "learning_rate": 5.789641866325091e-05, "loss": 0.0049, "step": 1350 }, { "epoch": 2.025487256371814, "grad_norm": 0.06091418273733409, "learning_rate": 5.773817382593008e-05, "loss": 0.005, "step": 1351 }, { "epoch": 2.026986506746627, "grad_norm": 0.05620982339093061, "learning_rate": 5.758005772546097e-05, "loss": 0.0074, "step": 1352 }, { "epoch": 2.028485757121439, "grad_norm": 0.07166290846746312, "learning_rate": 5.7422070843492734e-05, "loss": 0.0075, "step": 1353 }, { "epoch": 2.029985007496252, "grad_norm": 0.05351548956206307, "learning_rate": 5.7264213661280765e-05, "loss": 0.0046, "step": 1354 }, { "epoch": 2.0314842578710643, "grad_norm": 0.16771028867831342, "learning_rate": 5.710648665968543e-05, "loss": 0.0061, "step": 1355 }, { "epoch": 2.032983508245877, "grad_norm": 0.049188943209597075, "learning_rate": 5.694889031917047e-05, "loss": 0.0041, "step": 1356 }, { "epoch": 2.0344827586206895, "grad_norm": 0.04547779735623479, "learning_rate": 5.679142511980175e-05, "loss": 0.0027, "step": 1357 }, { "epoch": 2.0359820089955023, "grad_norm": 0.10403833741318744, "learning_rate": 5.663409154124557e-05, "loss": 0.0049, "step": 1358 }, { "epoch": 2.0374812593703147, "grad_norm": 0.04247535677566507, "learning_rate": 5.647689006276726e-05, "loss": 0.0018, "step": 1359 }, { "epoch": 2.0389805097451275, "grad_norm": 0.07701782209515198, "learning_rate": 5.63198211632298e-05, "loss": 0.0061, "step": 1360 }, { "epoch": 2.04047976011994, "grad_norm": 0.08747035535017979, "learning_rate": 5.616288532109225e-05, "loss": 0.0047, "step": 1361 }, { "epoch": 2.0419790104947526, "grad_norm": 0.1850652979487481, "learning_rate": 5.6006083014408484e-05, "loss": 0.0081, "step": 1362 }, { "epoch": 2.0434782608695654, "grad_norm": 0.03442503949688273, "learning_rate": 5.584941472082549e-05, "loss": 0.0028, "step": 1363 }, { "epoch": 2.0449775112443778, "grad_norm": 0.10724209405230377, "learning_rate": 5.5692880917582046e-05, "loss": 0.0092, "step": 1364 }, { "epoch": 2.0464767616191906, "grad_norm": 0.06723361064783695, "learning_rate": 5.553648208150728e-05, "loss": 0.0049, "step": 1365 }, { "epoch": 2.047976011994003, "grad_norm": 0.06341433310483063, "learning_rate": 5.5380218689019125e-05, "loss": 0.0043, "step": 1366 }, { "epoch": 2.0494752623688157, "grad_norm": 0.06092073707056924, "learning_rate": 5.522409121612304e-05, "loss": 0.0046, "step": 1367 }, { "epoch": 2.050974512743628, "grad_norm": 0.21552813309206953, "learning_rate": 5.506810013841036e-05, "loss": 0.0138, "step": 1368 }, { "epoch": 2.052473763118441, "grad_norm": 0.0655401729299929, "learning_rate": 5.491224593105695e-05, "loss": 0.0032, "step": 1369 }, { "epoch": 2.0539730134932532, "grad_norm": 0.07021885435075632, "learning_rate": 5.475652906882173e-05, "loss": 0.0077, "step": 1370 }, { "epoch": 2.055472263868066, "grad_norm": 0.11480398420184994, "learning_rate": 5.4600950026045326e-05, "loss": 0.0122, "step": 1371 }, { "epoch": 2.0569715142428784, "grad_norm": 0.1415004918550936, "learning_rate": 5.4445509276648466e-05, "loss": 0.0055, "step": 1372 }, { "epoch": 2.058470764617691, "grad_norm": 0.08525010803571845, "learning_rate": 5.4290207294130615e-05, "loss": 0.0043, "step": 1373 }, { "epoch": 2.0599700149925035, "grad_norm": 0.039178570486034396, "learning_rate": 5.4135044551568546e-05, "loss": 0.0019, "step": 1374 }, { "epoch": 2.0614692653673163, "grad_norm": 0.0817349771881642, "learning_rate": 5.398002152161484e-05, "loss": 0.0027, "step": 1375 }, { "epoch": 2.062968515742129, "grad_norm": 0.12797938737383258, "learning_rate": 5.382513867649663e-05, "loss": 0.007, "step": 1376 }, { "epoch": 2.0644677661169415, "grad_norm": 0.14457775493372024, "learning_rate": 5.3670396488013854e-05, "loss": 0.0166, "step": 1377 }, { "epoch": 2.0659670164917543, "grad_norm": 0.09943983374773288, "learning_rate": 5.351579542753807e-05, "loss": 0.01, "step": 1378 }, { "epoch": 2.0674662668665666, "grad_norm": 0.04088894150509561, "learning_rate": 5.33613359660109e-05, "loss": 0.0026, "step": 1379 }, { "epoch": 2.0689655172413794, "grad_norm": 0.043089887974945645, "learning_rate": 5.320701857394268e-05, "loss": 0.0052, "step": 1380 }, { "epoch": 2.070464767616192, "grad_norm": 0.017301891563744867, "learning_rate": 5.305284372141095e-05, "loss": 0.0012, "step": 1381 }, { "epoch": 2.0719640179910046, "grad_norm": 0.05740234754230655, "learning_rate": 5.2898811878059e-05, "loss": 0.0027, "step": 1382 }, { "epoch": 2.073463268365817, "grad_norm": 0.05104654315941495, "learning_rate": 5.274492351309461e-05, "loss": 0.0039, "step": 1383 }, { "epoch": 2.0749625187406298, "grad_norm": 0.060428329511087744, "learning_rate": 5.259117909528839e-05, "loss": 0.0099, "step": 1384 }, { "epoch": 2.076461769115442, "grad_norm": 0.05507881001018212, "learning_rate": 5.243757909297247e-05, "loss": 0.0059, "step": 1385 }, { "epoch": 2.077961019490255, "grad_norm": 0.08630299613505883, "learning_rate": 5.2284123974039154e-05, "loss": 0.0041, "step": 1386 }, { "epoch": 2.0794602698650673, "grad_norm": 0.14844088217359913, "learning_rate": 5.213081420593933e-05, "loss": 0.0085, "step": 1387 }, { "epoch": 2.08095952023988, "grad_norm": 0.11946133785518649, "learning_rate": 5.197765025568109e-05, "loss": 0.0112, "step": 1388 }, { "epoch": 2.082458770614693, "grad_norm": 0.08999841950189766, "learning_rate": 5.182463258982846e-05, "loss": 0.0068, "step": 1389 }, { "epoch": 2.0839580209895052, "grad_norm": 0.08234997169720312, "learning_rate": 5.1671761674499765e-05, "loss": 0.0067, "step": 1390 }, { "epoch": 2.085457271364318, "grad_norm": 0.05543198018151636, "learning_rate": 5.15190379753663e-05, "loss": 0.0037, "step": 1391 }, { "epoch": 2.0869565217391304, "grad_norm": 0.1728213920668675, "learning_rate": 5.1366461957650954e-05, "loss": 0.014, "step": 1392 }, { "epoch": 2.088455772113943, "grad_norm": 0.059050509744897936, "learning_rate": 5.121403408612672e-05, "loss": 0.0061, "step": 1393 }, { "epoch": 2.0899550224887555, "grad_norm": 0.103095043846112, "learning_rate": 5.1061754825115374e-05, "loss": 0.018, "step": 1394 }, { "epoch": 2.0914542728635683, "grad_norm": 0.03936790723429008, "learning_rate": 5.090962463848592e-05, "loss": 0.0025, "step": 1395 }, { "epoch": 2.0929535232383807, "grad_norm": 0.15600675536862346, "learning_rate": 5.07576439896533e-05, "loss": 0.0123, "step": 1396 }, { "epoch": 2.0944527736131935, "grad_norm": 0.13895637920809223, "learning_rate": 5.0605813341576924e-05, "loss": 0.0083, "step": 1397 }, { "epoch": 2.095952023988006, "grad_norm": 0.07600005087241365, "learning_rate": 5.045413315675924e-05, "loss": 0.0057, "step": 1398 }, { "epoch": 2.0974512743628186, "grad_norm": 0.07128253832259326, "learning_rate": 5.0302603897244474e-05, "loss": 0.0062, "step": 1399 }, { "epoch": 2.098950524737631, "grad_norm": 0.03157517477798966, "learning_rate": 5.015122602461698e-05, "loss": 0.002, "step": 1400 }, { "epoch": 2.100449775112444, "grad_norm": 0.063283575874792, "learning_rate": 5.000000000000002e-05, "loss": 0.0065, "step": 1401 }, { "epoch": 2.101949025487256, "grad_norm": 0.048367470551601933, "learning_rate": 4.9848926284054255e-05, "loss": 0.0032, "step": 1402 }, { "epoch": 2.103448275862069, "grad_norm": 0.03119576646559798, "learning_rate": 4.969800533697649e-05, "loss": 0.003, "step": 1403 }, { "epoch": 2.1049475262368817, "grad_norm": 0.08189145443012984, "learning_rate": 4.9547237618498085e-05, "loss": 0.0016, "step": 1404 }, { "epoch": 2.106446776611694, "grad_norm": 0.04950617409138366, "learning_rate": 4.939662358788364e-05, "loss": 0.003, "step": 1405 }, { "epoch": 2.107946026986507, "grad_norm": 0.07768299028368211, "learning_rate": 4.924616370392961e-05, "loss": 0.0069, "step": 1406 }, { "epoch": 2.1094452773613193, "grad_norm": 0.03281358107954231, "learning_rate": 4.909585842496287e-05, "loss": 0.0032, "step": 1407 }, { "epoch": 2.110944527736132, "grad_norm": 0.11945720424334462, "learning_rate": 4.894570820883944e-05, "loss": 0.0158, "step": 1408 }, { "epoch": 2.1124437781109444, "grad_norm": 0.07932882476750797, "learning_rate": 4.8795713512942865e-05, "loss": 0.0067, "step": 1409 }, { "epoch": 2.113943028485757, "grad_norm": 0.026045115469265436, "learning_rate": 4.864587479418302e-05, "loss": 0.0017, "step": 1410 }, { "epoch": 2.1154422788605696, "grad_norm": 0.08432479899132923, "learning_rate": 4.8496192508994576e-05, "loss": 0.0068, "step": 1411 }, { "epoch": 2.1169415292353824, "grad_norm": 0.08090217442628098, "learning_rate": 4.834666711333582e-05, "loss": 0.0075, "step": 1412 }, { "epoch": 2.1184407796101947, "grad_norm": 0.0623686180681052, "learning_rate": 4.8197299062686995e-05, "loss": 0.0066, "step": 1413 }, { "epoch": 2.1199400299850075, "grad_norm": 0.053337961961056765, "learning_rate": 4.8048088812049096e-05, "loss": 0.0035, "step": 1414 }, { "epoch": 2.12143928035982, "grad_norm": 0.022375264029270973, "learning_rate": 4.78990368159424e-05, "loss": 0.0021, "step": 1415 }, { "epoch": 2.1229385307346327, "grad_norm": 0.11775003892209943, "learning_rate": 4.7750143528405126e-05, "loss": 0.01, "step": 1416 }, { "epoch": 2.1244377811094455, "grad_norm": 0.04122303945871826, "learning_rate": 4.7601409402992106e-05, "loss": 0.0027, "step": 1417 }, { "epoch": 2.125937031484258, "grad_norm": 0.047835533341228295, "learning_rate": 4.745283489277325e-05, "loss": 0.0022, "step": 1418 }, { "epoch": 2.1274362818590706, "grad_norm": 0.05086786601082947, "learning_rate": 4.7304420450332244e-05, "loss": 0.003, "step": 1419 }, { "epoch": 2.128935532233883, "grad_norm": 0.051518725132630144, "learning_rate": 4.71561665277653e-05, "loss": 0.0054, "step": 1420 }, { "epoch": 2.130434782608696, "grad_norm": 0.05190216880471178, "learning_rate": 4.700807357667952e-05, "loss": 0.0023, "step": 1421 }, { "epoch": 2.131934032983508, "grad_norm": 0.10337376557969855, "learning_rate": 4.68601420481917e-05, "loss": 0.0058, "step": 1422 }, { "epoch": 2.133433283358321, "grad_norm": 0.03507599171549857, "learning_rate": 4.6712372392927e-05, "loss": 0.0018, "step": 1423 }, { "epoch": 2.1349325337331333, "grad_norm": 0.0504070606518784, "learning_rate": 4.656476506101737e-05, "loss": 0.0041, "step": 1424 }, { "epoch": 2.136431784107946, "grad_norm": 0.04550825149065688, "learning_rate": 4.6417320502100316e-05, "loss": 0.0066, "step": 1425 }, { "epoch": 2.1379310344827585, "grad_norm": 0.08210457443940872, "learning_rate": 4.6270039165317605e-05, "loss": 0.0054, "step": 1426 }, { "epoch": 2.1394302848575713, "grad_norm": 0.09426410394325767, "learning_rate": 4.612292149931369e-05, "loss": 0.0045, "step": 1427 }, { "epoch": 2.1409295352323836, "grad_norm": 0.05591749148306491, "learning_rate": 4.59759679522345e-05, "loss": 0.0115, "step": 1428 }, { "epoch": 2.1424287856071964, "grad_norm": 0.051607173257308396, "learning_rate": 4.582917897172603e-05, "loss": 0.0039, "step": 1429 }, { "epoch": 2.143928035982009, "grad_norm": 0.08203477167290794, "learning_rate": 4.568255500493292e-05, "loss": 0.0019, "step": 1430 }, { "epoch": 2.1454272863568216, "grad_norm": 0.0702400046217267, "learning_rate": 4.5536096498497295e-05, "loss": 0.005, "step": 1431 }, { "epoch": 2.1469265367316344, "grad_norm": 0.04253399244012623, "learning_rate": 4.5389803898557106e-05, "loss": 0.0021, "step": 1432 }, { "epoch": 2.1484257871064467, "grad_norm": 0.10867953782868013, "learning_rate": 4.524367765074499e-05, "loss": 0.0103, "step": 1433 }, { "epoch": 2.1499250374812595, "grad_norm": 0.0672757026971409, "learning_rate": 4.5097718200186814e-05, "loss": 0.0046, "step": 1434 }, { "epoch": 2.151424287856072, "grad_norm": 0.06915524666361397, "learning_rate": 4.495192599150044e-05, "loss": 0.0047, "step": 1435 }, { "epoch": 2.1529235382308847, "grad_norm": 0.20209814933952153, "learning_rate": 4.480630146879419e-05, "loss": 0.0133, "step": 1436 }, { "epoch": 2.154422788605697, "grad_norm": 0.05989447472439885, "learning_rate": 4.46608450756656e-05, "loss": 0.0048, "step": 1437 }, { "epoch": 2.15592203898051, "grad_norm": 0.06402998012232496, "learning_rate": 4.451555725520009e-05, "loss": 0.0051, "step": 1438 }, { "epoch": 2.157421289355322, "grad_norm": 0.054064533624966084, "learning_rate": 4.437043844996952e-05, "loss": 0.0037, "step": 1439 }, { "epoch": 2.158920539730135, "grad_norm": 0.03250489649463816, "learning_rate": 4.4225489102030995e-05, "loss": 0.0027, "step": 1440 }, { "epoch": 2.1604197901049473, "grad_norm": 0.07946031963668018, "learning_rate": 4.4080709652925336e-05, "loss": 0.0049, "step": 1441 }, { "epoch": 2.16191904047976, "grad_norm": 0.12942691934162556, "learning_rate": 4.393610054367585e-05, "loss": 0.0092, "step": 1442 }, { "epoch": 2.1634182908545725, "grad_norm": 0.10645765981635241, "learning_rate": 4.379166221478697e-05, "loss": 0.0102, "step": 1443 }, { "epoch": 2.1649175412293853, "grad_norm": 0.06232769787707574, "learning_rate": 4.364739510624286e-05, "loss": 0.0031, "step": 1444 }, { "epoch": 2.166416791604198, "grad_norm": 0.06283478350375783, "learning_rate": 4.350329965750621e-05, "loss": 0.0027, "step": 1445 }, { "epoch": 2.1679160419790104, "grad_norm": 0.0458686612766849, "learning_rate": 4.335937630751674e-05, "loss": 0.0037, "step": 1446 }, { "epoch": 2.1694152923538232, "grad_norm": 0.08933097544850213, "learning_rate": 4.32156254946899e-05, "loss": 0.0069, "step": 1447 }, { "epoch": 2.1709145427286356, "grad_norm": 0.08001891739481207, "learning_rate": 4.307204765691558e-05, "loss": 0.0052, "step": 1448 }, { "epoch": 2.1724137931034484, "grad_norm": 0.19361680061202965, "learning_rate": 4.2928643231556844e-05, "loss": 0.0076, "step": 1449 }, { "epoch": 2.1739130434782608, "grad_norm": 0.06531644648753111, "learning_rate": 4.27854126554484e-05, "loss": 0.0069, "step": 1450 }, { "epoch": 2.1754122938530736, "grad_norm": 0.11015887039069647, "learning_rate": 4.264235636489542e-05, "loss": 0.0055, "step": 1451 }, { "epoch": 2.176911544227886, "grad_norm": 0.057749351044845736, "learning_rate": 4.249947479567218e-05, "loss": 0.003, "step": 1452 }, { "epoch": 2.1784107946026987, "grad_norm": 0.07640922995670499, "learning_rate": 4.235676838302068e-05, "loss": 0.0056, "step": 1453 }, { "epoch": 2.179910044977511, "grad_norm": 0.05338471185436684, "learning_rate": 4.221423756164948e-05, "loss": 0.0034, "step": 1454 }, { "epoch": 2.181409295352324, "grad_norm": 0.11526166319637095, "learning_rate": 4.207188276573214e-05, "loss": 0.009, "step": 1455 }, { "epoch": 2.1829085457271367, "grad_norm": 0.0758586393121374, "learning_rate": 4.1929704428906026e-05, "loss": 0.004, "step": 1456 }, { "epoch": 2.184407796101949, "grad_norm": 0.08116550249079785, "learning_rate": 4.1787702984271074e-05, "loss": 0.0161, "step": 1457 }, { "epoch": 2.185907046476762, "grad_norm": 0.08050820007995133, "learning_rate": 4.1645878864388266e-05, "loss": 0.009, "step": 1458 }, { "epoch": 2.187406296851574, "grad_norm": 0.08238424313644067, "learning_rate": 4.150423250127845e-05, "loss": 0.0049, "step": 1459 }, { "epoch": 2.188905547226387, "grad_norm": 0.08089595012768978, "learning_rate": 4.1362764326421064e-05, "loss": 0.0033, "step": 1460 }, { "epoch": 2.1904047976011993, "grad_norm": 0.03361021628627364, "learning_rate": 4.12214747707527e-05, "loss": 0.0031, "step": 1461 }, { "epoch": 2.191904047976012, "grad_norm": 0.08619764958760784, "learning_rate": 4.1080364264665774e-05, "loss": 0.0115, "step": 1462 }, { "epoch": 2.1934032983508245, "grad_norm": 0.18569051233692943, "learning_rate": 4.093943323800745e-05, "loss": 0.0127, "step": 1463 }, { "epoch": 2.1949025487256373, "grad_norm": 0.019196872401471405, "learning_rate": 4.0798682120078044e-05, "loss": 0.0007, "step": 1464 }, { "epoch": 2.1964017991004496, "grad_norm": 0.09065601441816457, "learning_rate": 4.065811133962987e-05, "loss": 0.0065, "step": 1465 }, { "epoch": 2.1979010494752624, "grad_norm": 0.04523517478416136, "learning_rate": 4.0517721324865884e-05, "loss": 0.0012, "step": 1466 }, { "epoch": 2.199400299850075, "grad_norm": 0.06034653817480533, "learning_rate": 4.037751250343841e-05, "loss": 0.0042, "step": 1467 }, { "epoch": 2.2008995502248876, "grad_norm": 0.025739476261626783, "learning_rate": 4.023748530244789e-05, "loss": 0.002, "step": 1468 }, { "epoch": 2.2023988005997, "grad_norm": 0.09451823682114383, "learning_rate": 4.009764014844143e-05, "loss": 0.0032, "step": 1469 }, { "epoch": 2.2038980509745127, "grad_norm": 0.14766234026107614, "learning_rate": 3.9957977467411615e-05, "loss": 0.0096, "step": 1470 }, { "epoch": 2.2053973013493255, "grad_norm": 0.249925133129023, "learning_rate": 3.981849768479517e-05, "loss": 0.0056, "step": 1471 }, { "epoch": 2.206896551724138, "grad_norm": 0.05670171893846676, "learning_rate": 3.967920122547175e-05, "loss": 0.0058, "step": 1472 }, { "epoch": 2.2083958020989507, "grad_norm": 0.06840035500620087, "learning_rate": 3.954008851376252e-05, "loss": 0.0042, "step": 1473 }, { "epoch": 2.209895052473763, "grad_norm": 0.045333811302793486, "learning_rate": 3.940115997342891e-05, "loss": 0.0029, "step": 1474 }, { "epoch": 2.211394302848576, "grad_norm": 0.04822527848446775, "learning_rate": 3.9262416027671356e-05, "loss": 0.0032, "step": 1475 }, { "epoch": 2.212893553223388, "grad_norm": 0.09769165771156645, "learning_rate": 3.9123857099127936e-05, "loss": 0.0066, "step": 1476 }, { "epoch": 2.214392803598201, "grad_norm": 0.07720924648038567, "learning_rate": 3.8985483609873244e-05, "loss": 0.0014, "step": 1477 }, { "epoch": 2.2158920539730134, "grad_norm": 0.04125133249769171, "learning_rate": 3.8847295981416896e-05, "loss": 0.003, "step": 1478 }, { "epoch": 2.217391304347826, "grad_norm": 0.047023293986731454, "learning_rate": 3.8709294634702376e-05, "loss": 0.0027, "step": 1479 }, { "epoch": 2.2188905547226385, "grad_norm": 0.03636072085533822, "learning_rate": 3.857147999010567e-05, "loss": 0.0028, "step": 1480 }, { "epoch": 2.2203898050974513, "grad_norm": 0.03091756783151599, "learning_rate": 3.843385246743417e-05, "loss": 0.0018, "step": 1481 }, { "epoch": 2.2218890554722637, "grad_norm": 0.06714013114185455, "learning_rate": 3.829641248592515e-05, "loss": 0.0049, "step": 1482 }, { "epoch": 2.2233883058470765, "grad_norm": 0.10511179919131185, "learning_rate": 3.8159160464244606e-05, "loss": 0.0046, "step": 1483 }, { "epoch": 2.224887556221889, "grad_norm": 0.03003926401709271, "learning_rate": 3.802209682048602e-05, "loss": 0.0023, "step": 1484 }, { "epoch": 2.2263868065967016, "grad_norm": 0.03524357503089674, "learning_rate": 3.788522197216897e-05, "loss": 0.0027, "step": 1485 }, { "epoch": 2.2278860569715144, "grad_norm": 0.11632249193591344, "learning_rate": 3.774853633623806e-05, "loss": 0.0047, "step": 1486 }, { "epoch": 2.229385307346327, "grad_norm": 0.06473791327123747, "learning_rate": 3.7612040329061405e-05, "loss": 0.0054, "step": 1487 }, { "epoch": 2.2308845577211396, "grad_norm": 0.049416266463761196, "learning_rate": 3.747573436642951e-05, "loss": 0.0034, "step": 1488 }, { "epoch": 2.232383808095952, "grad_norm": 0.0388053166882485, "learning_rate": 3.733961886355398e-05, "loss": 0.004, "step": 1489 }, { "epoch": 2.2338830584707647, "grad_norm": 0.0510239158161139, "learning_rate": 3.720369423506622e-05, "loss": 0.0035, "step": 1490 }, { "epoch": 2.235382308845577, "grad_norm": 0.06027442038507764, "learning_rate": 3.7067960895016275e-05, "loss": 0.0038, "step": 1491 }, { "epoch": 2.23688155922039, "grad_norm": 0.11520121661098545, "learning_rate": 3.6932419256871406e-05, "loss": 0.0062, "step": 1492 }, { "epoch": 2.2383808095952022, "grad_norm": 0.06456017036528984, "learning_rate": 3.679706973351491e-05, "loss": 0.0027, "step": 1493 }, { "epoch": 2.239880059970015, "grad_norm": 0.09253651512540281, "learning_rate": 3.6661912737245e-05, "loss": 0.0076, "step": 1494 }, { "epoch": 2.2413793103448274, "grad_norm": 0.09671661159708796, "learning_rate": 3.6526948679773257e-05, "loss": 0.0033, "step": 1495 }, { "epoch": 2.24287856071964, "grad_norm": 0.06702412437333265, "learning_rate": 3.6392177972223594e-05, "loss": 0.0036, "step": 1496 }, { "epoch": 2.244377811094453, "grad_norm": 0.09090522248170291, "learning_rate": 3.6257601025131026e-05, "loss": 0.0062, "step": 1497 }, { "epoch": 2.2458770614692654, "grad_norm": 0.058717896724903795, "learning_rate": 3.612321824844024e-05, "loss": 0.008, "step": 1498 }, { "epoch": 2.247376311844078, "grad_norm": 0.05083383150547819, "learning_rate": 3.5989030051504434e-05, "loss": 0.0035, "step": 1499 }, { "epoch": 2.2488755622188905, "grad_norm": 0.0698571999573116, "learning_rate": 3.585503684308421e-05, "loss": 0.004, "step": 1500 }, { "epoch": 2.2503748125937033, "grad_norm": 0.03210517089158349, "learning_rate": 3.5721239031346066e-05, "loss": 0.0012, "step": 1501 }, { "epoch": 2.2518740629685157, "grad_norm": 0.0508811347970621, "learning_rate": 3.558763702386135e-05, "loss": 0.0019, "step": 1502 }, { "epoch": 2.2533733133433285, "grad_norm": 0.055664797337053686, "learning_rate": 3.545423122760493e-05, "loss": 0.0038, "step": 1503 }, { "epoch": 2.254872563718141, "grad_norm": 0.08996644981999107, "learning_rate": 3.5321022048954035e-05, "loss": 0.0077, "step": 1504 }, { "epoch": 2.2563718140929536, "grad_norm": 0.13609555101705403, "learning_rate": 3.518800989368691e-05, "loss": 0.0065, "step": 1505 }, { "epoch": 2.257871064467766, "grad_norm": 0.050562321159216154, "learning_rate": 3.5055195166981645e-05, "loss": 0.0057, "step": 1506 }, { "epoch": 2.2593703148425788, "grad_norm": 0.0728087625749308, "learning_rate": 3.492257827341492e-05, "loss": 0.0045, "step": 1507 }, { "epoch": 2.260869565217391, "grad_norm": 0.09718734094804558, "learning_rate": 3.479015961696077e-05, "loss": 0.0071, "step": 1508 }, { "epoch": 2.262368815592204, "grad_norm": 0.17375393173330367, "learning_rate": 3.465793960098945e-05, "loss": 0.002, "step": 1509 }, { "epoch": 2.2638680659670163, "grad_norm": 0.021400919519601855, "learning_rate": 3.452591862826603e-05, "loss": 0.0018, "step": 1510 }, { "epoch": 2.265367316341829, "grad_norm": 0.06996342276647459, "learning_rate": 3.439409710094929e-05, "loss": 0.0089, "step": 1511 }, { "epoch": 2.266866566716642, "grad_norm": 0.061074064657634694, "learning_rate": 3.426247542059041e-05, "loss": 0.0011, "step": 1512 }, { "epoch": 2.2683658170914542, "grad_norm": 0.06945067478625049, "learning_rate": 3.413105398813195e-05, "loss": 0.0051, "step": 1513 }, { "epoch": 2.269865067466267, "grad_norm": 0.042079105603492865, "learning_rate": 3.3999833203906326e-05, "loss": 0.004, "step": 1514 }, { "epoch": 2.2713643178410794, "grad_norm": 0.1636319307235689, "learning_rate": 3.386881346763483e-05, "loss": 0.0099, "step": 1515 }, { "epoch": 2.272863568215892, "grad_norm": 0.06664106371082085, "learning_rate": 3.373799517842627e-05, "loss": 0.0044, "step": 1516 }, { "epoch": 2.2743628185907045, "grad_norm": 0.08483429638899491, "learning_rate": 3.360737873477584e-05, "loss": 0.0034, "step": 1517 }, { "epoch": 2.2758620689655173, "grad_norm": 0.06849123537105387, "learning_rate": 3.347696453456393e-05, "loss": 0.0076, "step": 1518 }, { "epoch": 2.2773613193403297, "grad_norm": 0.18864254560141094, "learning_rate": 3.334675297505476e-05, "loss": 0.0068, "step": 1519 }, { "epoch": 2.2788605697151425, "grad_norm": 0.02900241632175253, "learning_rate": 3.3216744452895354e-05, "loss": 0.0014, "step": 1520 }, { "epoch": 2.280359820089955, "grad_norm": 0.06275154975100304, "learning_rate": 3.308693936411421e-05, "loss": 0.0055, "step": 1521 }, { "epoch": 2.2818590704647677, "grad_norm": 0.1528582473234831, "learning_rate": 3.2957338104120096e-05, "loss": 0.0047, "step": 1522 }, { "epoch": 2.2833583208395805, "grad_norm": 0.07980744631055499, "learning_rate": 3.2827941067700996e-05, "loss": 0.0056, "step": 1523 }, { "epoch": 2.284857571214393, "grad_norm": 0.11944836679739665, "learning_rate": 3.269874864902269e-05, "loss": 0.0046, "step": 1524 }, { "epoch": 2.286356821589205, "grad_norm": 0.023772227427949762, "learning_rate": 3.2569761241627696e-05, "loss": 0.0016, "step": 1525 }, { "epoch": 2.287856071964018, "grad_norm": 0.14329800784900046, "learning_rate": 3.244097923843398e-05, "loss": 0.0148, "step": 1526 }, { "epoch": 2.2893553223388308, "grad_norm": 0.1195966321417723, "learning_rate": 3.231240303173394e-05, "loss": 0.0057, "step": 1527 }, { "epoch": 2.290854572713643, "grad_norm": 0.10708809110295565, "learning_rate": 3.218403301319296e-05, "loss": 0.0065, "step": 1528 }, { "epoch": 2.292353823088456, "grad_norm": 0.08568289613574692, "learning_rate": 3.205586957384838e-05, "loss": 0.0032, "step": 1529 }, { "epoch": 2.2938530734632683, "grad_norm": 0.035756635002319404, "learning_rate": 3.192791310410822e-05, "loss": 0.002, "step": 1530 }, { "epoch": 2.295352323838081, "grad_norm": 0.06053419631054795, "learning_rate": 3.1800163993750166e-05, "loss": 0.0056, "step": 1531 }, { "epoch": 2.2968515742128934, "grad_norm": 0.1039350364479123, "learning_rate": 3.16726226319201e-05, "loss": 0.0081, "step": 1532 }, { "epoch": 2.2983508245877062, "grad_norm": 0.07906247819210281, "learning_rate": 3.154528940713113e-05, "loss": 0.0067, "step": 1533 }, { "epoch": 2.2998500749625186, "grad_norm": 0.08407334961019255, "learning_rate": 3.141816470726238e-05, "loss": 0.0113, "step": 1534 }, { "epoch": 2.3013493253373314, "grad_norm": 0.08815120231572529, "learning_rate": 3.129124891955771e-05, "loss": 0.0052, "step": 1535 }, { "epoch": 2.3028485757121437, "grad_norm": 0.0666277487037083, "learning_rate": 3.116454243062459e-05, "loss": 0.0032, "step": 1536 }, { "epoch": 2.3043478260869565, "grad_norm": 0.04918000761524814, "learning_rate": 3.103804562643302e-05, "loss": 0.0043, "step": 1537 }, { "epoch": 2.3058470764617693, "grad_norm": 0.060187697186195384, "learning_rate": 3.0911758892314166e-05, "loss": 0.0035, "step": 1538 }, { "epoch": 2.3073463268365817, "grad_norm": 0.16836993548608092, "learning_rate": 3.078568261295933e-05, "loss": 0.0052, "step": 1539 }, { "epoch": 2.3088455772113945, "grad_norm": 0.06729086558094731, "learning_rate": 3.0659817172418693e-05, "loss": 0.0049, "step": 1540 }, { "epoch": 2.310344827586207, "grad_norm": 0.07617128616033088, "learning_rate": 3.053416295410026e-05, "loss": 0.0056, "step": 1541 }, { "epoch": 2.3118440779610197, "grad_norm": 0.114464765928607, "learning_rate": 3.0408720340768572e-05, "loss": 0.0029, "step": 1542 }, { "epoch": 2.313343328335832, "grad_norm": 0.063347013144408, "learning_rate": 3.0283489714543556e-05, "loss": 0.0045, "step": 1543 }, { "epoch": 2.314842578710645, "grad_norm": 0.06774768128325656, "learning_rate": 3.0158471456899428e-05, "loss": 0.0022, "step": 1544 }, { "epoch": 2.316341829085457, "grad_norm": 0.048376078472133446, "learning_rate": 3.0033665948663448e-05, "loss": 0.0028, "step": 1545 }, { "epoch": 2.31784107946027, "grad_norm": 0.0385182292711496, "learning_rate": 2.9909073570014912e-05, "loss": 0.0014, "step": 1546 }, { "epoch": 2.3193403298350823, "grad_norm": 0.03475187689682614, "learning_rate": 2.9784694700483762e-05, "loss": 0.0017, "step": 1547 }, { "epoch": 2.320839580209895, "grad_norm": 0.07347410055145828, "learning_rate": 2.9660529718949627e-05, "loss": 0.0049, "step": 1548 }, { "epoch": 2.3223388305847075, "grad_norm": 0.06424491061862778, "learning_rate": 2.953657900364053e-05, "loss": 0.004, "step": 1549 }, { "epoch": 2.3238380809595203, "grad_norm": 0.09419900753622726, "learning_rate": 2.94128429321319e-05, "loss": 0.007, "step": 1550 }, { "epoch": 2.3253373313343326, "grad_norm": 0.08715139140623997, "learning_rate": 2.9289321881345254e-05, "loss": 0.0045, "step": 1551 }, { "epoch": 2.3268365817091454, "grad_norm": 0.06268684955631437, "learning_rate": 2.9166016227547133e-05, "loss": 0.0041, "step": 1552 }, { "epoch": 2.3283358320839582, "grad_norm": 0.1379363777149399, "learning_rate": 2.904292634634793e-05, "loss": 0.0034, "step": 1553 }, { "epoch": 2.3298350824587706, "grad_norm": 0.10089329635825812, "learning_rate": 2.8920052612700754e-05, "loss": 0.0062, "step": 1554 }, { "epoch": 2.3313343328335834, "grad_norm": 0.05387050027903078, "learning_rate": 2.879739540090036e-05, "loss": 0.0066, "step": 1555 }, { "epoch": 2.3328335832083957, "grad_norm": 0.05660842464190973, "learning_rate": 2.8674955084581857e-05, "loss": 0.0042, "step": 1556 }, { "epoch": 2.3343328335832085, "grad_norm": 0.05686811860047845, "learning_rate": 2.8552732036719687e-05, "loss": 0.0051, "step": 1557 }, { "epoch": 2.335832083958021, "grad_norm": 0.04938712425283094, "learning_rate": 2.8430726629626413e-05, "loss": 0.0018, "step": 1558 }, { "epoch": 2.3373313343328337, "grad_norm": 0.13710965858068258, "learning_rate": 2.8308939234951726e-05, "loss": 0.0084, "step": 1559 }, { "epoch": 2.338830584707646, "grad_norm": 0.09065138517655111, "learning_rate": 2.8187370223681132e-05, "loss": 0.0052, "step": 1560 }, { "epoch": 2.340329835082459, "grad_norm": 0.04322879635848861, "learning_rate": 2.8066019966134904e-05, "loss": 0.0054, "step": 1561 }, { "epoch": 2.341829085457271, "grad_norm": 0.14140024545462096, "learning_rate": 2.7944888831966987e-05, "loss": 0.0069, "step": 1562 }, { "epoch": 2.343328335832084, "grad_norm": 0.05855557490735462, "learning_rate": 2.7823977190163786e-05, "loss": 0.0038, "step": 1563 }, { "epoch": 2.344827586206897, "grad_norm": 0.06774189505154501, "learning_rate": 2.770328540904319e-05, "loss": 0.0055, "step": 1564 }, { "epoch": 2.346326836581709, "grad_norm": 0.06685740409439345, "learning_rate": 2.7582813856253275e-05, "loss": 0.0073, "step": 1565 }, { "epoch": 2.3478260869565215, "grad_norm": 0.06597092919168687, "learning_rate": 2.746256289877126e-05, "loss": 0.0049, "step": 1566 }, { "epoch": 2.3493253373313343, "grad_norm": 0.06984959414446605, "learning_rate": 2.734253290290242e-05, "loss": 0.0029, "step": 1567 }, { "epoch": 2.350824587706147, "grad_norm": 0.08794480426989766, "learning_rate": 2.722272423427896e-05, "loss": 0.0046, "step": 1568 }, { "epoch": 2.3523238380809595, "grad_norm": 0.08294773284079503, "learning_rate": 2.7103137257858868e-05, "loss": 0.0053, "step": 1569 }, { "epoch": 2.3538230884557723, "grad_norm": 0.05311205720612156, "learning_rate": 2.698377233792476e-05, "loss": 0.0031, "step": 1570 }, { "epoch": 2.3553223388305846, "grad_norm": 0.044514573918505385, "learning_rate": 2.6864629838082956e-05, "loss": 0.0045, "step": 1571 }, { "epoch": 2.3568215892053974, "grad_norm": 0.08026506351168762, "learning_rate": 2.6745710121262136e-05, "loss": 0.0049, "step": 1572 }, { "epoch": 2.3583208395802098, "grad_norm": 0.030283279059830456, "learning_rate": 2.6627013549712355e-05, "loss": 0.0017, "step": 1573 }, { "epoch": 2.3598200899550226, "grad_norm": 0.056873656422552585, "learning_rate": 2.6508540485004006e-05, "loss": 0.0031, "step": 1574 }, { "epoch": 2.361319340329835, "grad_norm": 0.0739899949993969, "learning_rate": 2.639029128802657e-05, "loss": 0.0048, "step": 1575 }, { "epoch": 2.3628185907046477, "grad_norm": 0.07078189352862575, "learning_rate": 2.6272266318987603e-05, "loss": 0.0038, "step": 1576 }, { "epoch": 2.36431784107946, "grad_norm": 0.09589625252399156, "learning_rate": 2.615446593741161e-05, "loss": 0.0061, "step": 1577 }, { "epoch": 2.365817091454273, "grad_norm": 0.07845365529561737, "learning_rate": 2.603689050213902e-05, "loss": 0.0039, "step": 1578 }, { "epoch": 2.3673163418290857, "grad_norm": 0.09914753466937914, "learning_rate": 2.5919540371325e-05, "loss": 0.0052, "step": 1579 }, { "epoch": 2.368815592203898, "grad_norm": 0.12930439990410358, "learning_rate": 2.580241590243837e-05, "loss": 0.0081, "step": 1580 }, { "epoch": 2.370314842578711, "grad_norm": 0.023463536942587522, "learning_rate": 2.5685517452260567e-05, "loss": 0.0022, "step": 1581 }, { "epoch": 2.371814092953523, "grad_norm": 0.07218813844672958, "learning_rate": 2.5568845376884587e-05, "loss": 0.0027, "step": 1582 }, { "epoch": 2.373313343328336, "grad_norm": 0.06641391796530317, "learning_rate": 2.5452400031713785e-05, "loss": 0.0064, "step": 1583 }, { "epoch": 2.3748125937031483, "grad_norm": 0.03574796789731248, "learning_rate": 2.5336181771460876e-05, "loss": 0.0019, "step": 1584 }, { "epoch": 2.376311844077961, "grad_norm": 0.04934069132677985, "learning_rate": 2.5220190950146827e-05, "loss": 0.0026, "step": 1585 }, { "epoch": 2.3778110944527735, "grad_norm": 0.053344136812610306, "learning_rate": 2.5104427921099782e-05, "loss": 0.0035, "step": 1586 }, { "epoch": 2.3793103448275863, "grad_norm": 0.05934960469965204, "learning_rate": 2.4988893036954043e-05, "loss": 0.0052, "step": 1587 }, { "epoch": 2.3808095952023987, "grad_norm": 0.07363174118581005, "learning_rate": 2.4873586649648894e-05, "loss": 0.0021, "step": 1588 }, { "epoch": 2.3823088455772115, "grad_norm": 0.06507057844090482, "learning_rate": 2.4758509110427575e-05, "loss": 0.0064, "step": 1589 }, { "epoch": 2.383808095952024, "grad_norm": 0.04011004024330606, "learning_rate": 2.464366076983623e-05, "loss": 0.0016, "step": 1590 }, { "epoch": 2.3853073463268366, "grad_norm": 0.029963463107639907, "learning_rate": 2.45290419777228e-05, "loss": 0.0012, "step": 1591 }, { "epoch": 2.386806596701649, "grad_norm": 0.11065244961492637, "learning_rate": 2.441465308323605e-05, "loss": 0.0063, "step": 1592 }, { "epoch": 2.3883058470764618, "grad_norm": 0.07476696771495371, "learning_rate": 2.4300494434824373e-05, "loss": 0.003, "step": 1593 }, { "epoch": 2.3898050974512746, "grad_norm": 0.10626192717766343, "learning_rate": 2.4186566380234798e-05, "loss": 0.0101, "step": 1594 }, { "epoch": 2.391304347826087, "grad_norm": 0.06898129289247987, "learning_rate": 2.407286926651192e-05, "loss": 0.0037, "step": 1595 }, { "epoch": 2.3928035982008997, "grad_norm": 0.030965851115252685, "learning_rate": 2.3959403439996907e-05, "loss": 0.0023, "step": 1596 }, { "epoch": 2.394302848575712, "grad_norm": 0.12448248521634331, "learning_rate": 2.3846169246326343e-05, "loss": 0.004, "step": 1597 }, { "epoch": 2.395802098950525, "grad_norm": 0.0496245094173845, "learning_rate": 2.373316703043119e-05, "loss": 0.0039, "step": 1598 }, { "epoch": 2.3973013493253372, "grad_norm": 0.06468143373332225, "learning_rate": 2.362039713653581e-05, "loss": 0.0039, "step": 1599 }, { "epoch": 2.39880059970015, "grad_norm": 0.02745569957950435, "learning_rate": 2.3507859908156827e-05, "loss": 0.0035, "step": 1600 }, { "epoch": 2.4002998500749624, "grad_norm": 0.0884132237063567, "learning_rate": 2.339555568810221e-05, "loss": 0.0073, "step": 1601 }, { "epoch": 2.401799100449775, "grad_norm": 0.09407281067698077, "learning_rate": 2.328348481847006e-05, "loss": 0.0043, "step": 1602 }, { "epoch": 2.4032983508245875, "grad_norm": 0.023867483923462753, "learning_rate": 2.3171647640647687e-05, "loss": 0.0009, "step": 1603 }, { "epoch": 2.4047976011994003, "grad_norm": 0.05163847859209614, "learning_rate": 2.3060044495310505e-05, "loss": 0.0023, "step": 1604 }, { "epoch": 2.406296851574213, "grad_norm": 0.047121807189413514, "learning_rate": 2.2948675722421086e-05, "loss": 0.003, "step": 1605 }, { "epoch": 2.4077961019490255, "grad_norm": 0.07119567947811782, "learning_rate": 2.2837541661228025e-05, "loss": 0.003, "step": 1606 }, { "epoch": 2.409295352323838, "grad_norm": 0.053849133740421475, "learning_rate": 2.2726642650264895e-05, "loss": 0.0019, "step": 1607 }, { "epoch": 2.4107946026986506, "grad_norm": 0.07662771818389892, "learning_rate": 2.2615979027349387e-05, "loss": 0.0051, "step": 1608 }, { "epoch": 2.4122938530734634, "grad_norm": 0.14925873267020182, "learning_rate": 2.2505551129582047e-05, "loss": 0.0066, "step": 1609 }, { "epoch": 2.413793103448276, "grad_norm": 0.09592241042779556, "learning_rate": 2.2395359293345396e-05, "loss": 0.0061, "step": 1610 }, { "epoch": 2.4152923538230886, "grad_norm": 0.127006071375077, "learning_rate": 2.2285403854302912e-05, "loss": 0.0055, "step": 1611 }, { "epoch": 2.416791604197901, "grad_norm": 0.06229520326818543, "learning_rate": 2.2175685147397906e-05, "loss": 0.0054, "step": 1612 }, { "epoch": 2.4182908545727138, "grad_norm": 0.04966340726882021, "learning_rate": 2.2066203506852566e-05, "loss": 0.0025, "step": 1613 }, { "epoch": 2.419790104947526, "grad_norm": 0.052824372953982054, "learning_rate": 2.195695926616702e-05, "loss": 0.0024, "step": 1614 }, { "epoch": 2.421289355322339, "grad_norm": 0.0693320471877942, "learning_rate": 2.1847952758118117e-05, "loss": 0.0057, "step": 1615 }, { "epoch": 2.4227886056971513, "grad_norm": 0.1585165428374079, "learning_rate": 2.173918431475861e-05, "loss": 0.0079, "step": 1616 }, { "epoch": 2.424287856071964, "grad_norm": 0.12183346017418488, "learning_rate": 2.163065426741603e-05, "loss": 0.0064, "step": 1617 }, { "epoch": 2.4257871064467764, "grad_norm": 0.045242084358939956, "learning_rate": 2.1522362946691698e-05, "loss": 0.0054, "step": 1618 }, { "epoch": 2.427286356821589, "grad_norm": 0.0722336789715876, "learning_rate": 2.1414310682459802e-05, "loss": 0.0031, "step": 1619 }, { "epoch": 2.428785607196402, "grad_norm": 0.07131568496067937, "learning_rate": 2.1306497803866277e-05, "loss": 0.0058, "step": 1620 }, { "epoch": 2.4302848575712144, "grad_norm": 0.037029961293887234, "learning_rate": 2.119892463932781e-05, "loss": 0.0028, "step": 1621 }, { "epoch": 2.431784107946027, "grad_norm": 0.05980642170587757, "learning_rate": 2.1091591516530952e-05, "loss": 0.0089, "step": 1622 }, { "epoch": 2.4332833583208395, "grad_norm": 0.06090138836468679, "learning_rate": 2.098449876243096e-05, "loss": 0.0025, "step": 1623 }, { "epoch": 2.4347826086956523, "grad_norm": 0.05351342549456341, "learning_rate": 2.0877646703250996e-05, "loss": 0.0057, "step": 1624 }, { "epoch": 2.4362818590704647, "grad_norm": 0.03433223890330924, "learning_rate": 2.0771035664480942e-05, "loss": 0.0013, "step": 1625 }, { "epoch": 2.4377811094452775, "grad_norm": 0.047845862718465054, "learning_rate": 2.0664665970876496e-05, "loss": 0.0015, "step": 1626 }, { "epoch": 2.43928035982009, "grad_norm": 0.02633163873788185, "learning_rate": 2.0558537946458177e-05, "loss": 0.0013, "step": 1627 }, { "epoch": 2.4407796101949026, "grad_norm": 0.07332498941147272, "learning_rate": 2.045265191451041e-05, "loss": 0.0044, "step": 1628 }, { "epoch": 2.442278860569715, "grad_norm": 0.16264733297325074, "learning_rate": 2.0347008197580374e-05, "loss": 0.0213, "step": 1629 }, { "epoch": 2.443778110944528, "grad_norm": 0.03432376223511998, "learning_rate": 2.024160711747717e-05, "loss": 0.0027, "step": 1630 }, { "epoch": 2.44527736131934, "grad_norm": 0.08444598432390558, "learning_rate": 2.013644899527074e-05, "loss": 0.0086, "step": 1631 }, { "epoch": 2.446776611694153, "grad_norm": 0.09785171406288519, "learning_rate": 2.0031534151290943e-05, "loss": 0.0145, "step": 1632 }, { "epoch": 2.4482758620689653, "grad_norm": 0.10686482838519329, "learning_rate": 1.9926862905126665e-05, "loss": 0.0106, "step": 1633 }, { "epoch": 2.449775112443778, "grad_norm": 0.07512971821946084, "learning_rate": 1.9822435575624608e-05, "loss": 0.0045, "step": 1634 }, { "epoch": 2.451274362818591, "grad_norm": 0.10390776644311271, "learning_rate": 1.9718252480888566e-05, "loss": 0.0067, "step": 1635 }, { "epoch": 2.4527736131934033, "grad_norm": 0.054546501477964644, "learning_rate": 1.9614313938278272e-05, "loss": 0.0041, "step": 1636 }, { "epoch": 2.454272863568216, "grad_norm": 0.1236643546327681, "learning_rate": 1.9510620264408596e-05, "loss": 0.0076, "step": 1637 }, { "epoch": 2.4557721139430284, "grad_norm": 0.06368717471421793, "learning_rate": 1.9407171775148436e-05, "loss": 0.0046, "step": 1638 }, { "epoch": 2.457271364317841, "grad_norm": 0.058824393187371, "learning_rate": 1.930396878561983e-05, "loss": 0.0056, "step": 1639 }, { "epoch": 2.4587706146926536, "grad_norm": 0.052763276928937744, "learning_rate": 1.9201011610196973e-05, "loss": 0.0042, "step": 1640 }, { "epoch": 2.4602698650674664, "grad_norm": 0.04366071201528183, "learning_rate": 1.9098300562505266e-05, "loss": 0.0038, "step": 1641 }, { "epoch": 2.4617691154422787, "grad_norm": 0.060492485188534235, "learning_rate": 1.8995835955420416e-05, "loss": 0.0027, "step": 1642 }, { "epoch": 2.4632683658170915, "grad_norm": 0.12378841704182127, "learning_rate": 1.8893618101067355e-05, "loss": 0.0098, "step": 1643 }, { "epoch": 2.464767616191904, "grad_norm": 0.06572768630271339, "learning_rate": 1.879164731081937e-05, "loss": 0.0026, "step": 1644 }, { "epoch": 2.4662668665667167, "grad_norm": 0.09310343642883542, "learning_rate": 1.8689923895297245e-05, "loss": 0.0047, "step": 1645 }, { "epoch": 2.4677661169415295, "grad_norm": 0.044289718611579314, "learning_rate": 1.858844816436809e-05, "loss": 0.0016, "step": 1646 }, { "epoch": 2.469265367316342, "grad_norm": 0.04572702054619638, "learning_rate": 1.848722042714457e-05, "loss": 0.0018, "step": 1647 }, { "epoch": 2.470764617691154, "grad_norm": 0.04476649599527431, "learning_rate": 1.838624099198397e-05, "loss": 0.0048, "step": 1648 }, { "epoch": 2.472263868065967, "grad_norm": 0.07032354780767394, "learning_rate": 1.8285510166487152e-05, "loss": 0.0078, "step": 1649 }, { "epoch": 2.47376311844078, "grad_norm": 0.09134832895870344, "learning_rate": 1.818502825749764e-05, "loss": 0.0084, "step": 1650 }, { "epoch": 2.475262368815592, "grad_norm": 0.10898437749116632, "learning_rate": 1.808479557110081e-05, "loss": 0.0026, "step": 1651 }, { "epoch": 2.476761619190405, "grad_norm": 0.09805488182454494, "learning_rate": 1.7984812412622787e-05, "loss": 0.0023, "step": 1652 }, { "epoch": 2.4782608695652173, "grad_norm": 0.08765732874500178, "learning_rate": 1.78850790866296e-05, "loss": 0.0035, "step": 1653 }, { "epoch": 2.47976011994003, "grad_norm": 0.062345434719609814, "learning_rate": 1.7785595896926265e-05, "loss": 0.0037, "step": 1654 }, { "epoch": 2.4812593703148424, "grad_norm": 0.05004851068742012, "learning_rate": 1.7686363146555805e-05, "loss": 0.0019, "step": 1655 }, { "epoch": 2.4827586206896552, "grad_norm": 0.12233324261355612, "learning_rate": 1.7587381137798432e-05, "loss": 0.0084, "step": 1656 }, { "epoch": 2.4842578710644676, "grad_norm": 0.25709397112046917, "learning_rate": 1.7488650172170496e-05, "loss": 0.0133, "step": 1657 }, { "epoch": 2.4857571214392804, "grad_norm": 0.1112658667316051, "learning_rate": 1.7390170550423625e-05, "loss": 0.0071, "step": 1658 }, { "epoch": 2.4872563718140928, "grad_norm": 0.033739803171623846, "learning_rate": 1.7291942572543807e-05, "loss": 0.0031, "step": 1659 }, { "epoch": 2.4887556221889056, "grad_norm": 0.05031209731885149, "learning_rate": 1.719396653775056e-05, "loss": 0.0033, "step": 1660 }, { "epoch": 2.4902548725637184, "grad_norm": 0.10642934597922178, "learning_rate": 1.7096242744495837e-05, "loss": 0.0055, "step": 1661 }, { "epoch": 2.4917541229385307, "grad_norm": 0.07260588589200684, "learning_rate": 1.6998771490463262e-05, "loss": 0.0084, "step": 1662 }, { "epoch": 2.4932533733133435, "grad_norm": 0.0583410528108744, "learning_rate": 1.690155307256719e-05, "loss": 0.0048, "step": 1663 }, { "epoch": 2.494752623688156, "grad_norm": 0.09011124093388319, "learning_rate": 1.680458778695174e-05, "loss": 0.0043, "step": 1664 }, { "epoch": 2.4962518740629687, "grad_norm": 0.21013483131124727, "learning_rate": 1.6707875928990058e-05, "loss": 0.0126, "step": 1665 }, { "epoch": 2.497751124437781, "grad_norm": 0.044788060768008446, "learning_rate": 1.661141779328319e-05, "loss": 0.0045, "step": 1666 }, { "epoch": 2.499250374812594, "grad_norm": 0.09719586856323512, "learning_rate": 1.6515213673659357e-05, "loss": 0.0081, "step": 1667 }, { "epoch": 2.500749625187406, "grad_norm": 0.044720423379557714, "learning_rate": 1.6419263863172997e-05, "loss": 0.004, "step": 1668 }, { "epoch": 2.502248875562219, "grad_norm": 0.052589040164308765, "learning_rate": 1.632356865410384e-05, "loss": 0.006, "step": 1669 }, { "epoch": 2.5037481259370313, "grad_norm": 0.1143440138081822, "learning_rate": 1.622812833795613e-05, "loss": 0.008, "step": 1670 }, { "epoch": 2.505247376311844, "grad_norm": 0.09238843455164125, "learning_rate": 1.6132943205457606e-05, "loss": 0.0082, "step": 1671 }, { "epoch": 2.506746626686657, "grad_norm": 0.056736669939192, "learning_rate": 1.6038013546558696e-05, "loss": 0.0026, "step": 1672 }, { "epoch": 2.5082458770614693, "grad_norm": 0.12807829910070073, "learning_rate": 1.5943339650431576e-05, "loss": 0.0086, "step": 1673 }, { "epoch": 2.5097451274362816, "grad_norm": 0.050687004065841715, "learning_rate": 1.5848921805469397e-05, "loss": 0.0041, "step": 1674 }, { "epoch": 2.5112443778110944, "grad_norm": 0.07876565827972132, "learning_rate": 1.5754760299285252e-05, "loss": 0.005, "step": 1675 }, { "epoch": 2.5127436281859072, "grad_norm": 0.04259095718764495, "learning_rate": 1.566085541871145e-05, "loss": 0.0014, "step": 1676 }, { "epoch": 2.5142428785607196, "grad_norm": 0.1209625050394713, "learning_rate": 1.5567207449798515e-05, "loss": 0.0108, "step": 1677 }, { "epoch": 2.5157421289355324, "grad_norm": 0.037696859736511955, "learning_rate": 1.547381667781439e-05, "loss": 0.0028, "step": 1678 }, { "epoch": 2.5172413793103448, "grad_norm": 0.03303993337115884, "learning_rate": 1.538068338724361e-05, "loss": 0.0024, "step": 1679 }, { "epoch": 2.5187406296851576, "grad_norm": 0.04950974456854536, "learning_rate": 1.528780786178631e-05, "loss": 0.0059, "step": 1680 }, { "epoch": 2.52023988005997, "grad_norm": 0.07067979674428795, "learning_rate": 1.5195190384357404e-05, "loss": 0.0055, "step": 1681 }, { "epoch": 2.5217391304347827, "grad_norm": 0.052179735514149166, "learning_rate": 1.5102831237085857e-05, "loss": 0.0043, "step": 1682 }, { "epoch": 2.523238380809595, "grad_norm": 0.08034175791414389, "learning_rate": 1.5010730701313625e-05, "loss": 0.0082, "step": 1683 }, { "epoch": 2.524737631184408, "grad_norm": 0.06320598136407608, "learning_rate": 1.4918889057594876e-05, "loss": 0.0025, "step": 1684 }, { "epoch": 2.52623688155922, "grad_norm": 0.1498719753089046, "learning_rate": 1.4827306585695234e-05, "loss": 0.0032, "step": 1685 }, { "epoch": 2.527736131934033, "grad_norm": 0.07584597146514462, "learning_rate": 1.4735983564590783e-05, "loss": 0.003, "step": 1686 }, { "epoch": 2.529235382308846, "grad_norm": 0.04125809843734668, "learning_rate": 1.4644920272467244e-05, "loss": 0.0015, "step": 1687 }, { "epoch": 2.530734632683658, "grad_norm": 0.05314353996550423, "learning_rate": 1.4554116986719257e-05, "loss": 0.0073, "step": 1688 }, { "epoch": 2.5322338830584705, "grad_norm": 0.010199439979138412, "learning_rate": 1.4463573983949341e-05, "loss": 0.0004, "step": 1689 }, { "epoch": 2.5337331334332833, "grad_norm": 0.13381608061216652, "learning_rate": 1.4373291539967182e-05, "loss": 0.0059, "step": 1690 }, { "epoch": 2.535232383808096, "grad_norm": 0.09355504087054788, "learning_rate": 1.4283269929788779e-05, "loss": 0.005, "step": 1691 }, { "epoch": 2.5367316341829085, "grad_norm": 0.02939896423318598, "learning_rate": 1.4193509427635543e-05, "loss": 0.0027, "step": 1692 }, { "epoch": 2.5382308845577213, "grad_norm": 0.0400275347334245, "learning_rate": 1.4104010306933557e-05, "loss": 0.0025, "step": 1693 }, { "epoch": 2.5397301349325336, "grad_norm": 0.07400393800455013, "learning_rate": 1.4014772840312663e-05, "loss": 0.0036, "step": 1694 }, { "epoch": 2.5412293853073464, "grad_norm": 0.07901281731584704, "learning_rate": 1.3925797299605647e-05, "loss": 0.0027, "step": 1695 }, { "epoch": 2.542728635682159, "grad_norm": 0.05405597758510519, "learning_rate": 1.3837083955847418e-05, "loss": 0.0033, "step": 1696 }, { "epoch": 2.5442278860569716, "grad_norm": 0.05375559377296624, "learning_rate": 1.3748633079274253e-05, "loss": 0.0032, "step": 1697 }, { "epoch": 2.545727136431784, "grad_norm": 0.05039178419782183, "learning_rate": 1.3660444939322836e-05, "loss": 0.0035, "step": 1698 }, { "epoch": 2.5472263868065967, "grad_norm": 0.07167457099538735, "learning_rate": 1.3572519804629536e-05, "loss": 0.0056, "step": 1699 }, { "epoch": 2.548725637181409, "grad_norm": 0.07372779209688211, "learning_rate": 1.3484857943029572e-05, "loss": 0.0083, "step": 1700 }, { "epoch": 2.550224887556222, "grad_norm": 0.040476404349508645, "learning_rate": 1.339745962155613e-05, "loss": 0.0026, "step": 1701 }, { "epoch": 2.5517241379310347, "grad_norm": 0.08291264682769349, "learning_rate": 1.3310325106439726e-05, "loss": 0.0032, "step": 1702 }, { "epoch": 2.553223388305847, "grad_norm": 0.04934727510500086, "learning_rate": 1.3223454663107172e-05, "loss": 0.0039, "step": 1703 }, { "epoch": 2.5547226386806594, "grad_norm": 0.03721964515275045, "learning_rate": 1.3136848556180892e-05, "loss": 0.0027, "step": 1704 }, { "epoch": 2.556221889055472, "grad_norm": 0.06989750045733815, "learning_rate": 1.30505070494781e-05, "loss": 0.0098, "step": 1705 }, { "epoch": 2.557721139430285, "grad_norm": 0.038253858696665755, "learning_rate": 1.296443040601003e-05, "loss": 0.0023, "step": 1706 }, { "epoch": 2.5592203898050974, "grad_norm": 0.04073764123862176, "learning_rate": 1.2878618887981064e-05, "loss": 0.0017, "step": 1707 }, { "epoch": 2.56071964017991, "grad_norm": 0.08121678653741697, "learning_rate": 1.279307275678795e-05, "loss": 0.0042, "step": 1708 }, { "epoch": 2.5622188905547225, "grad_norm": 0.06896884513801788, "learning_rate": 1.2707792273019048e-05, "loss": 0.0041, "step": 1709 }, { "epoch": 2.5637181409295353, "grad_norm": 0.07423572460387555, "learning_rate": 1.262277769645348e-05, "loss": 0.004, "step": 1710 }, { "epoch": 2.5652173913043477, "grad_norm": 0.032450974009567375, "learning_rate": 1.2538029286060426e-05, "loss": 0.0018, "step": 1711 }, { "epoch": 2.5667166416791605, "grad_norm": 0.04898128979522446, "learning_rate": 1.2453547299998225e-05, "loss": 0.004, "step": 1712 }, { "epoch": 2.5682158920539733, "grad_norm": 0.043498324705939055, "learning_rate": 1.2369331995613665e-05, "loss": 0.0017, "step": 1713 }, { "epoch": 2.5697151424287856, "grad_norm": 0.08309623767839033, "learning_rate": 1.228538362944115e-05, "loss": 0.0041, "step": 1714 }, { "epoch": 2.571214392803598, "grad_norm": 0.1282714955750836, "learning_rate": 1.2201702457201947e-05, "loss": 0.0044, "step": 1715 }, { "epoch": 2.572713643178411, "grad_norm": 0.03853257098388854, "learning_rate": 1.2118288733803473e-05, "loss": 0.0054, "step": 1716 }, { "epoch": 2.5742128935532236, "grad_norm": 0.13020391770943424, "learning_rate": 1.2035142713338366e-05, "loss": 0.0115, "step": 1717 }, { "epoch": 2.575712143928036, "grad_norm": 0.07576542607588402, "learning_rate": 1.19522646490838e-05, "loss": 0.0038, "step": 1718 }, { "epoch": 2.5772113943028487, "grad_norm": 0.025413308174999216, "learning_rate": 1.1869654793500784e-05, "loss": 0.0035, "step": 1719 }, { "epoch": 2.578710644677661, "grad_norm": 0.022381286165155924, "learning_rate": 1.1787313398233235e-05, "loss": 0.0008, "step": 1720 }, { "epoch": 2.580209895052474, "grad_norm": 0.07623231370859024, "learning_rate": 1.1705240714107302e-05, "loss": 0.0041, "step": 1721 }, { "epoch": 2.5817091454272862, "grad_norm": 0.0827073481443242, "learning_rate": 1.1623436991130654e-05, "loss": 0.0063, "step": 1722 }, { "epoch": 2.583208395802099, "grad_norm": 0.042460507585494335, "learning_rate": 1.1541902478491606e-05, "loss": 0.002, "step": 1723 }, { "epoch": 2.5847076461769114, "grad_norm": 0.04491600293936866, "learning_rate": 1.1460637424558407e-05, "loss": 0.0037, "step": 1724 }, { "epoch": 2.586206896551724, "grad_norm": 0.06266018554204954, "learning_rate": 1.1379642076878527e-05, "loss": 0.0078, "step": 1725 }, { "epoch": 2.5877061469265366, "grad_norm": 0.1519755677743987, "learning_rate": 1.129891668217783e-05, "loss": 0.0092, "step": 1726 }, { "epoch": 2.5892053973013494, "grad_norm": 0.01931708277491169, "learning_rate": 1.1218461486359877e-05, "loss": 0.0011, "step": 1727 }, { "epoch": 2.590704647676162, "grad_norm": 0.13331809055097277, "learning_rate": 1.1138276734505104e-05, "loss": 0.0084, "step": 1728 }, { "epoch": 2.5922038980509745, "grad_norm": 0.08602979516663713, "learning_rate": 1.1058362670870249e-05, "loss": 0.0047, "step": 1729 }, { "epoch": 2.593703148425787, "grad_norm": 0.20475185102113513, "learning_rate": 1.0978719538887349e-05, "loss": 0.005, "step": 1730 }, { "epoch": 2.5952023988005997, "grad_norm": 0.05396284422100616, "learning_rate": 1.0899347581163221e-05, "loss": 0.0033, "step": 1731 }, { "epoch": 2.5967016491754125, "grad_norm": 0.04632926541173506, "learning_rate": 1.0820247039478604e-05, "loss": 0.0031, "step": 1732 }, { "epoch": 2.598200899550225, "grad_norm": 0.06330786884284814, "learning_rate": 1.0741418154787442e-05, "loss": 0.0056, "step": 1733 }, { "epoch": 2.5997001499250376, "grad_norm": 0.049581430029072754, "learning_rate": 1.0662861167216243e-05, "loss": 0.0048, "step": 1734 }, { "epoch": 2.60119940029985, "grad_norm": 0.014809634178959625, "learning_rate": 1.0584576316063188e-05, "loss": 0.0011, "step": 1735 }, { "epoch": 2.6026986506746628, "grad_norm": 0.06952939363160453, "learning_rate": 1.0506563839797501e-05, "loss": 0.0046, "step": 1736 }, { "epoch": 2.604197901049475, "grad_norm": 0.025305092673321244, "learning_rate": 1.042882397605871e-05, "loss": 0.0011, "step": 1737 }, { "epoch": 2.605697151424288, "grad_norm": 0.06048091669387347, "learning_rate": 1.0351356961655945e-05, "loss": 0.0037, "step": 1738 }, { "epoch": 2.6071964017991007, "grad_norm": 0.0928376653466004, "learning_rate": 1.0274163032567163e-05, "loss": 0.0038, "step": 1739 }, { "epoch": 2.608695652173913, "grad_norm": 0.05539979196115197, "learning_rate": 1.0197242423938446e-05, "loss": 0.0076, "step": 1740 }, { "epoch": 2.6101949025487254, "grad_norm": 0.07917966581807787, "learning_rate": 1.0120595370083318e-05, "loss": 0.0042, "step": 1741 }, { "epoch": 2.6116941529235382, "grad_norm": 0.060022482722549786, "learning_rate": 1.0044222104481971e-05, "loss": 0.0031, "step": 1742 }, { "epoch": 2.613193403298351, "grad_norm": 0.17585147443485435, "learning_rate": 9.968122859780648e-06, "loss": 0.0082, "step": 1743 }, { "epoch": 2.6146926536731634, "grad_norm": 0.09870264943142629, "learning_rate": 9.892297867790845e-06, "loss": 0.0036, "step": 1744 }, { "epoch": 2.6161919040479757, "grad_norm": 0.06095280132546412, "learning_rate": 9.816747359488632e-06, "loss": 0.002, "step": 1745 }, { "epoch": 2.6176911544227885, "grad_norm": 0.10967913707388899, "learning_rate": 9.74147156501396e-06, "loss": 0.0035, "step": 1746 }, { "epoch": 2.6191904047976013, "grad_norm": 0.33902675973415286, "learning_rate": 9.666470713669918e-06, "loss": 0.045, "step": 1747 }, { "epoch": 2.6206896551724137, "grad_norm": 0.04253165972530954, "learning_rate": 9.591745033922173e-06, "loss": 0.0024, "step": 1748 }, { "epoch": 2.6221889055472265, "grad_norm": 0.04818383699809805, "learning_rate": 9.517294753398064e-06, "loss": 0.0024, "step": 1749 }, { "epoch": 2.623688155922039, "grad_norm": 0.03809410049607696, "learning_rate": 9.443120098886061e-06, "loss": 0.0032, "step": 1750 }, { "epoch": 2.6251874062968517, "grad_norm": 0.05909250477315833, "learning_rate": 9.369221296335006e-06, "loss": 0.0034, "step": 1751 }, { "epoch": 2.626686656671664, "grad_norm": 0.041243274161096855, "learning_rate": 9.295598570853514e-06, "loss": 0.004, "step": 1752 }, { "epoch": 2.628185907046477, "grad_norm": 0.05659045575181391, "learning_rate": 9.222252146709142e-06, "loss": 0.0032, "step": 1753 }, { "epoch": 2.6296851574212896, "grad_norm": 0.0812942025512981, "learning_rate": 9.149182247327837e-06, "loss": 0.0047, "step": 1754 }, { "epoch": 2.631184407796102, "grad_norm": 0.0662188120787313, "learning_rate": 9.076389095293148e-06, "loss": 0.0058, "step": 1755 }, { "epoch": 2.6326836581709143, "grad_norm": 0.09566246288727459, "learning_rate": 9.00387291234569e-06, "loss": 0.0051, "step": 1756 }, { "epoch": 2.634182908545727, "grad_norm": 0.08790520476527616, "learning_rate": 8.931633919382298e-06, "loss": 0.0092, "step": 1757 }, { "epoch": 2.63568215892054, "grad_norm": 0.09603411271828961, "learning_rate": 8.85967233645547e-06, "loss": 0.0035, "step": 1758 }, { "epoch": 2.6371814092953523, "grad_norm": 0.07679943096437322, "learning_rate": 8.787988382772705e-06, "loss": 0.0052, "step": 1759 }, { "epoch": 2.638680659670165, "grad_norm": 0.06355610491952349, "learning_rate": 8.716582276695728e-06, "loss": 0.0033, "step": 1760 }, { "epoch": 2.6401799100449774, "grad_norm": 0.11913399959459053, "learning_rate": 8.645454235739903e-06, "loss": 0.0043, "step": 1761 }, { "epoch": 2.6416791604197902, "grad_norm": 0.05791114392247628, "learning_rate": 8.574604476573621e-06, "loss": 0.0047, "step": 1762 }, { "epoch": 2.6431784107946026, "grad_norm": 0.10461299194881798, "learning_rate": 8.504033215017527e-06, "loss": 0.0062, "step": 1763 }, { "epoch": 2.6446776611694154, "grad_norm": 0.10373987138158351, "learning_rate": 8.433740666043898e-06, "loss": 0.0096, "step": 1764 }, { "epoch": 2.6461769115442277, "grad_norm": 0.1713331811915574, "learning_rate": 8.363727043776038e-06, "loss": 0.004, "step": 1765 }, { "epoch": 2.6476761619190405, "grad_norm": 0.06786862150547672, "learning_rate": 8.293992561487596e-06, "loss": 0.0044, "step": 1766 }, { "epoch": 2.649175412293853, "grad_norm": 0.07532039055497501, "learning_rate": 8.224537431601886e-06, "loss": 0.0033, "step": 1767 }, { "epoch": 2.6506746626686657, "grad_norm": 0.026496825220480674, "learning_rate": 8.15536186569129e-06, "loss": 0.0025, "step": 1768 }, { "epoch": 2.6521739130434785, "grad_norm": 0.04795260350131189, "learning_rate": 8.086466074476563e-06, "loss": 0.0043, "step": 1769 }, { "epoch": 2.653673163418291, "grad_norm": 0.04733553100664191, "learning_rate": 8.017850267826232e-06, "loss": 0.0024, "step": 1770 }, { "epoch": 2.655172413793103, "grad_norm": 0.05951843530460022, "learning_rate": 7.949514654755962e-06, "loss": 0.0035, "step": 1771 }, { "epoch": 2.656671664167916, "grad_norm": 0.04050730299729002, "learning_rate": 7.881459443427886e-06, "loss": 0.0037, "step": 1772 }, { "epoch": 2.658170914542729, "grad_norm": 0.06302619474576679, "learning_rate": 7.81368484114996e-06, "loss": 0.0035, "step": 1773 }, { "epoch": 2.659670164917541, "grad_norm": 0.03824673473843314, "learning_rate": 7.746191054375362e-06, "loss": 0.0037, "step": 1774 }, { "epoch": 2.661169415292354, "grad_norm": 0.06707823730001641, "learning_rate": 7.67897828870191e-06, "loss": 0.007, "step": 1775 }, { "epoch": 2.6626686656671663, "grad_norm": 0.07806204973737949, "learning_rate": 7.612046748871327e-06, "loss": 0.0056, "step": 1776 }, { "epoch": 2.664167916041979, "grad_norm": 0.043370114908618695, "learning_rate": 7.545396638768698e-06, "loss": 0.0035, "step": 1777 }, { "epoch": 2.6656671664167915, "grad_norm": 0.05273280042907484, "learning_rate": 7.479028161421797e-06, "loss": 0.0019, "step": 1778 }, { "epoch": 2.6671664167916043, "grad_norm": 0.059897018536616276, "learning_rate": 7.412941519000527e-06, "loss": 0.003, "step": 1779 }, { "epoch": 2.668665667166417, "grad_norm": 0.07985855599457413, "learning_rate": 7.347136912816277e-06, "loss": 0.0051, "step": 1780 }, { "epoch": 2.6701649175412294, "grad_norm": 0.03581758546197187, "learning_rate": 7.281614543321269e-06, "loss": 0.0042, "step": 1781 }, { "epoch": 2.6716641679160418, "grad_norm": 0.04288618656047732, "learning_rate": 7.216374610108012e-06, "loss": 0.0035, "step": 1782 }, { "epoch": 2.6731634182908546, "grad_norm": 0.0663922133344574, "learning_rate": 7.151417311908648e-06, "loss": 0.0029, "step": 1783 }, { "epoch": 2.6746626686656674, "grad_norm": 0.1918430476451125, "learning_rate": 7.086742846594385e-06, "loss": 0.0135, "step": 1784 }, { "epoch": 2.6761619190404797, "grad_norm": 0.04484775879018914, "learning_rate": 7.022351411174866e-06, "loss": 0.0037, "step": 1785 }, { "epoch": 2.677661169415292, "grad_norm": 0.062248407404835907, "learning_rate": 6.958243201797554e-06, "loss": 0.004, "step": 1786 }, { "epoch": 2.679160419790105, "grad_norm": 0.10446631857945263, "learning_rate": 6.894418413747183e-06, "loss": 0.0045, "step": 1787 }, { "epoch": 2.6806596701649177, "grad_norm": 0.03529468151449674, "learning_rate": 6.830877241445111e-06, "loss": 0.0021, "step": 1788 }, { "epoch": 2.68215892053973, "grad_norm": 0.03798589090052098, "learning_rate": 6.767619878448783e-06, "loss": 0.0026, "step": 1789 }, { "epoch": 2.683658170914543, "grad_norm": 0.08669293863873424, "learning_rate": 6.704646517451107e-06, "loss": 0.0056, "step": 1790 }, { "epoch": 2.685157421289355, "grad_norm": 0.046908300538182957, "learning_rate": 6.6419573502798374e-06, "loss": 0.0025, "step": 1791 }, { "epoch": 2.686656671664168, "grad_norm": 0.03637779784955495, "learning_rate": 6.579552567897051e-06, "loss": 0.0016, "step": 1792 }, { "epoch": 2.6881559220389803, "grad_norm": 0.04687808627810223, "learning_rate": 6.517432360398556e-06, "loss": 0.0053, "step": 1793 }, { "epoch": 2.689655172413793, "grad_norm": 0.037364738444089864, "learning_rate": 6.455596917013273e-06, "loss": 0.0014, "step": 1794 }, { "epoch": 2.691154422788606, "grad_norm": 0.1021041733000301, "learning_rate": 6.394046426102674e-06, "loss": 0.0083, "step": 1795 }, { "epoch": 2.6926536731634183, "grad_norm": 0.018456788471741113, "learning_rate": 6.332781075160243e-06, "loss": 0.0009, "step": 1796 }, { "epoch": 2.6941529235382307, "grad_norm": 0.14926815053881523, "learning_rate": 6.2718010508108545e-06, "loss": 0.0102, "step": 1797 }, { "epoch": 2.6956521739130435, "grad_norm": 0.04352809658591788, "learning_rate": 6.21110653881023e-06, "loss": 0.0028, "step": 1798 }, { "epoch": 2.6971514242878563, "grad_norm": 0.05279283988150011, "learning_rate": 6.1506977240444074e-06, "loss": 0.0017, "step": 1799 }, { "epoch": 2.6986506746626686, "grad_norm": 0.019605598342381503, "learning_rate": 6.090574790529091e-06, "loss": 0.0006, "step": 1800 }, { "epoch": 2.7001499250374814, "grad_norm": 0.0777744775067246, "learning_rate": 6.030737921409169e-06, "loss": 0.0029, "step": 1801 }, { "epoch": 2.7016491754122938, "grad_norm": 0.05414552236249822, "learning_rate": 5.971187298958103e-06, "loss": 0.0056, "step": 1802 }, { "epoch": 2.7031484257871066, "grad_norm": 0.04962408711837054, "learning_rate": 5.911923104577455e-06, "loss": 0.0039, "step": 1803 }, { "epoch": 2.704647676161919, "grad_norm": 0.0696464944244991, "learning_rate": 5.852945518796205e-06, "loss": 0.0079, "step": 1804 }, { "epoch": 2.7061469265367317, "grad_norm": 0.03714539340218805, "learning_rate": 5.7942547212703315e-06, "loss": 0.0019, "step": 1805 }, { "epoch": 2.707646176911544, "grad_norm": 0.05223972716082169, "learning_rate": 5.735850890782157e-06, "loss": 0.0054, "step": 1806 }, { "epoch": 2.709145427286357, "grad_norm": 0.07430860604160193, "learning_rate": 5.6777342052399045e-06, "loss": 0.0035, "step": 1807 }, { "epoch": 2.7106446776611692, "grad_norm": 0.060445175053381824, "learning_rate": 5.619904841677059e-06, "loss": 0.0023, "step": 1808 }, { "epoch": 2.712143928035982, "grad_norm": 0.05925366881856817, "learning_rate": 5.562362976251901e-06, "loss": 0.0047, "step": 1809 }, { "epoch": 2.713643178410795, "grad_norm": 0.0516545752906272, "learning_rate": 5.505108784246926e-06, "loss": 0.0022, "step": 1810 }, { "epoch": 2.715142428785607, "grad_norm": 0.10602896675857203, "learning_rate": 5.448142440068316e-06, "loss": 0.0042, "step": 1811 }, { "epoch": 2.7166416791604195, "grad_norm": 0.32243857508421625, "learning_rate": 5.39146411724547e-06, "loss": 0.0072, "step": 1812 }, { "epoch": 2.7181409295352323, "grad_norm": 0.05465551687606462, "learning_rate": 5.335073988430372e-06, "loss": 0.0036, "step": 1813 }, { "epoch": 2.719640179910045, "grad_norm": 0.08368985646892235, "learning_rate": 5.278972225397127e-06, "loss": 0.0119, "step": 1814 }, { "epoch": 2.7211394302848575, "grad_norm": 0.051321818493470575, "learning_rate": 5.223158999041444e-06, "loss": 0.0048, "step": 1815 }, { "epoch": 2.7226386806596703, "grad_norm": 0.04408943737843986, "learning_rate": 5.167634479380068e-06, "loss": 0.0024, "step": 1816 }, { "epoch": 2.7241379310344827, "grad_norm": 0.1202534156928614, "learning_rate": 5.1123988355503475e-06, "loss": 0.0078, "step": 1817 }, { "epoch": 2.7256371814092955, "grad_norm": 0.052112172067874475, "learning_rate": 5.057452235809624e-06, "loss": 0.0058, "step": 1818 }, { "epoch": 2.727136431784108, "grad_norm": 0.1431485997481884, "learning_rate": 5.002794847534764e-06, "loss": 0.0097, "step": 1819 }, { "epoch": 2.7286356821589206, "grad_norm": 0.051277101130307555, "learning_rate": 4.948426837221631e-06, "loss": 0.0055, "step": 1820 }, { "epoch": 2.7301349325337334, "grad_norm": 0.053350618119373285, "learning_rate": 4.8943483704846475e-06, "loss": 0.0038, "step": 1821 }, { "epoch": 2.7316341829085458, "grad_norm": 0.030903608998343296, "learning_rate": 4.840559612056183e-06, "loss": 0.0014, "step": 1822 }, { "epoch": 2.733133433283358, "grad_norm": 0.047055897536594594, "learning_rate": 4.7870607257861415e-06, "loss": 0.0051, "step": 1823 }, { "epoch": 2.734632683658171, "grad_norm": 0.02818509921299283, "learning_rate": 4.733851874641382e-06, "loss": 0.0008, "step": 1824 }, { "epoch": 2.7361319340329837, "grad_norm": 0.04570475809580388, "learning_rate": 4.680933220705308e-06, "loss": 0.0018, "step": 1825 }, { "epoch": 2.737631184407796, "grad_norm": 0.0538111788348984, "learning_rate": 4.628304925177318e-06, "loss": 0.0037, "step": 1826 }, { "epoch": 2.7391304347826084, "grad_norm": 0.13116424213383152, "learning_rate": 4.575967148372317e-06, "loss": 0.0136, "step": 1827 }, { "epoch": 2.7406296851574212, "grad_norm": 0.04885765875397696, "learning_rate": 4.523920049720265e-06, "loss": 0.0025, "step": 1828 }, { "epoch": 2.742128935532234, "grad_norm": 0.10154857801433739, "learning_rate": 4.4721637877656375e-06, "loss": 0.0066, "step": 1829 }, { "epoch": 2.7436281859070464, "grad_norm": 0.047254869802745965, "learning_rate": 4.420698520166988e-06, "loss": 0.0034, "step": 1830 }, { "epoch": 2.745127436281859, "grad_norm": 0.08641710628596849, "learning_rate": 4.369524403696457e-06, "loss": 0.0084, "step": 1831 }, { "epoch": 2.7466266866566715, "grad_norm": 0.08831269449702958, "learning_rate": 4.318641594239259e-06, "loss": 0.0067, "step": 1832 }, { "epoch": 2.7481259370314843, "grad_norm": 0.07518037530585348, "learning_rate": 4.268050246793276e-06, "loss": 0.0067, "step": 1833 }, { "epoch": 2.7496251874062967, "grad_norm": 0.08886601880182082, "learning_rate": 4.217750515468522e-06, "loss": 0.0056, "step": 1834 }, { "epoch": 2.7511244377811095, "grad_norm": 0.047343986827075496, "learning_rate": 4.167742553486675e-06, "loss": 0.0037, "step": 1835 }, { "epoch": 2.7526236881559223, "grad_norm": 0.061256674999085256, "learning_rate": 4.118026513180695e-06, "loss": 0.0032, "step": 1836 }, { "epoch": 2.7541229385307346, "grad_norm": 0.09671352746387149, "learning_rate": 4.068602545994249e-06, "loss": 0.0105, "step": 1837 }, { "epoch": 2.755622188905547, "grad_norm": 0.12504578715825013, "learning_rate": 4.019470802481307e-06, "loss": 0.012, "step": 1838 }, { "epoch": 2.75712143928036, "grad_norm": 0.05061259533408433, "learning_rate": 3.970631432305694e-06, "loss": 0.004, "step": 1839 }, { "epoch": 2.7586206896551726, "grad_norm": 0.029406937261118465, "learning_rate": 3.922084584240582e-06, "loss": 0.002, "step": 1840 }, { "epoch": 2.760119940029985, "grad_norm": 0.02532563135738694, "learning_rate": 3.873830406168111e-06, "loss": 0.002, "step": 1841 }, { "epoch": 2.7616191904047978, "grad_norm": 0.10231518549891974, "learning_rate": 3.825869045078867e-06, "loss": 0.0048, "step": 1842 }, { "epoch": 2.76311844077961, "grad_norm": 0.04405091545119458, "learning_rate": 3.7782006470714616e-06, "loss": 0.0034, "step": 1843 }, { "epoch": 2.764617691154423, "grad_norm": 0.04214896707196821, "learning_rate": 3.730825357352119e-06, "loss": 0.0016, "step": 1844 }, { "epoch": 2.7661169415292353, "grad_norm": 0.04381641722204364, "learning_rate": 3.68374332023419e-06, "loss": 0.0015, "step": 1845 }, { "epoch": 2.767616191904048, "grad_norm": 0.1714819540968367, "learning_rate": 3.6369546791377052e-06, "loss": 0.0053, "step": 1846 }, { "epoch": 2.7691154422788604, "grad_norm": 0.030777510297242894, "learning_rate": 3.5904595765890005e-06, "loss": 0.0012, "step": 1847 }, { "epoch": 2.770614692653673, "grad_norm": 0.06278492467131958, "learning_rate": 3.5442581542201923e-06, "loss": 0.0038, "step": 1848 }, { "epoch": 2.7721139430284856, "grad_norm": 0.04525524756217157, "learning_rate": 3.4983505527688586e-06, "loss": 0.0032, "step": 1849 }, { "epoch": 2.7736131934032984, "grad_norm": 0.040158798347074746, "learning_rate": 3.452736912077503e-06, "loss": 0.0022, "step": 1850 }, { "epoch": 2.775112443778111, "grad_norm": 0.05351156811338634, "learning_rate": 3.40741737109318e-06, "loss": 0.0022, "step": 1851 }, { "epoch": 2.7766116941529235, "grad_norm": 0.08841883689534029, "learning_rate": 3.3623920678670597e-06, "loss": 0.003, "step": 1852 }, { "epoch": 2.778110944527736, "grad_norm": 0.07130261183515092, "learning_rate": 3.3176611395540626e-06, "loss": 0.0056, "step": 1853 }, { "epoch": 2.7796101949025487, "grad_norm": 0.09438468992037789, "learning_rate": 3.273224722412327e-06, "loss": 0.0046, "step": 1854 }, { "epoch": 2.7811094452773615, "grad_norm": 0.06383128845407085, "learning_rate": 3.2290829518028862e-06, "loss": 0.0022, "step": 1855 }, { "epoch": 2.782608695652174, "grad_norm": 0.07593396064006204, "learning_rate": 3.1852359621892367e-06, "loss": 0.0034, "step": 1856 }, { "epoch": 2.7841079460269866, "grad_norm": 0.04086990619865844, "learning_rate": 3.1416838871368924e-06, "loss": 0.0023, "step": 1857 }, { "epoch": 2.785607196401799, "grad_norm": 0.10911709604280903, "learning_rate": 3.0984268593130528e-06, "loss": 0.008, "step": 1858 }, { "epoch": 2.787106446776612, "grad_norm": 0.06393379283366997, "learning_rate": 3.0554650104861136e-06, "loss": 0.002, "step": 1859 }, { "epoch": 2.788605697151424, "grad_norm": 0.08071123917158006, "learning_rate": 3.012798471525324e-06, "loss": 0.0043, "step": 1860 }, { "epoch": 2.790104947526237, "grad_norm": 0.03866299410166067, "learning_rate": 2.970427372400353e-06, "loss": 0.0023, "step": 1861 }, { "epoch": 2.7916041979010497, "grad_norm": 0.12924488448937485, "learning_rate": 2.928351842180921e-06, "loss": 0.0083, "step": 1862 }, { "epoch": 2.793103448275862, "grad_norm": 0.044056078109747175, "learning_rate": 2.8865720090364034e-06, "loss": 0.002, "step": 1863 }, { "epoch": 2.7946026986506745, "grad_norm": 0.0932750403527346, "learning_rate": 2.845088000235396e-06, "loss": 0.0074, "step": 1864 }, { "epoch": 2.7961019490254873, "grad_norm": 0.21546183050930776, "learning_rate": 2.8038999421453826e-06, "loss": 0.0067, "step": 1865 }, { "epoch": 2.7976011994003, "grad_norm": 0.04083158743885419, "learning_rate": 2.7630079602323442e-06, "loss": 0.0032, "step": 1866 }, { "epoch": 2.7991004497751124, "grad_norm": 0.026451998528825394, "learning_rate": 2.7224121790603517e-06, "loss": 0.0014, "step": 1867 }, { "epoch": 2.8005997001499248, "grad_norm": 0.026543988690067504, "learning_rate": 2.6821127222911857e-06, "loss": 0.0012, "step": 1868 }, { "epoch": 2.8020989505247376, "grad_norm": 0.11502358287074994, "learning_rate": 2.6421097126839712e-06, "loss": 0.005, "step": 1869 }, { "epoch": 2.8035982008995504, "grad_norm": 0.06937276307112113, "learning_rate": 2.6024032720948443e-06, "loss": 0.0035, "step": 1870 }, { "epoch": 2.8050974512743627, "grad_norm": 0.04889190826037233, "learning_rate": 2.5629935214764865e-06, "loss": 0.0053, "step": 1871 }, { "epoch": 2.8065967016491755, "grad_norm": 0.01964804857509374, "learning_rate": 2.5238805808778242e-06, "loss": 0.0011, "step": 1872 }, { "epoch": 2.808095952023988, "grad_norm": 0.08944021499669409, "learning_rate": 2.4850645694436736e-06, "loss": 0.0043, "step": 1873 }, { "epoch": 2.8095952023988007, "grad_norm": 0.09138194545450047, "learning_rate": 2.446545605414341e-06, "loss": 0.0188, "step": 1874 }, { "epoch": 2.811094452773613, "grad_norm": 0.06082928618151328, "learning_rate": 2.4083238061252567e-06, "loss": 0.0034, "step": 1875 }, { "epoch": 2.812593703148426, "grad_norm": 0.0518451376073689, "learning_rate": 2.3703992880066638e-06, "loss": 0.0025, "step": 1876 }, { "epoch": 2.8140929535232386, "grad_norm": 0.08943695133377194, "learning_rate": 2.332772166583208e-06, "loss": 0.0025, "step": 1877 }, { "epoch": 2.815592203898051, "grad_norm": 0.054040088180756354, "learning_rate": 2.295442556473637e-06, "loss": 0.0051, "step": 1878 }, { "epoch": 2.8170914542728633, "grad_norm": 0.07247814988589195, "learning_rate": 2.2584105713904125e-06, "loss": 0.0068, "step": 1879 }, { "epoch": 2.818590704647676, "grad_norm": 0.06849834319985831, "learning_rate": 2.2216763241393767e-06, "loss": 0.0036, "step": 1880 }, { "epoch": 2.820089955022489, "grad_norm": 0.07362282563904983, "learning_rate": 2.1852399266194314e-06, "loss": 0.0073, "step": 1881 }, { "epoch": 2.8215892053973013, "grad_norm": 0.06944920144902611, "learning_rate": 2.1491014898221582e-06, "loss": 0.0024, "step": 1882 }, { "epoch": 2.823088455772114, "grad_norm": 0.04248072490573655, "learning_rate": 2.1132611238315003e-06, "loss": 0.0035, "step": 1883 }, { "epoch": 2.8245877061469264, "grad_norm": 0.0817432276248758, "learning_rate": 2.0777189378234143e-06, "loss": 0.0017, "step": 1884 }, { "epoch": 2.8260869565217392, "grad_norm": 0.07240802576223083, "learning_rate": 2.0424750400655947e-06, "loss": 0.0098, "step": 1885 }, { "epoch": 2.8275862068965516, "grad_norm": 0.0385971978938863, "learning_rate": 2.0075295379170412e-06, "loss": 0.0033, "step": 1886 }, { "epoch": 2.8290854572713644, "grad_norm": 0.053910895946008223, "learning_rate": 1.9728825378278246e-06, "loss": 0.0032, "step": 1887 }, { "epoch": 2.8305847076461768, "grad_norm": 0.09526927822276723, "learning_rate": 1.9385341453386995e-06, "loss": 0.0048, "step": 1888 }, { "epoch": 2.8320839580209896, "grad_norm": 0.06975698097659441, "learning_rate": 1.904484465080847e-06, "loss": 0.0082, "step": 1889 }, { "epoch": 2.833583208395802, "grad_norm": 0.17772085747797686, "learning_rate": 1.8707336007754873e-06, "loss": 0.0086, "step": 1890 }, { "epoch": 2.8350824587706147, "grad_norm": 0.084062140467934, "learning_rate": 1.8372816552336026e-06, "loss": 0.0035, "step": 1891 }, { "epoch": 2.8365817091454275, "grad_norm": 0.031985655169939474, "learning_rate": 1.8041287303556364e-06, "loss": 0.002, "step": 1892 }, { "epoch": 2.83808095952024, "grad_norm": 0.048848294887181724, "learning_rate": 1.771274927131139e-06, "loss": 0.0044, "step": 1893 }, { "epoch": 2.839580209895052, "grad_norm": 0.160521076075769, "learning_rate": 1.7387203456384782e-06, "loss": 0.0063, "step": 1894 }, { "epoch": 2.841079460269865, "grad_norm": 0.0284880955110527, "learning_rate": 1.706465085044584e-06, "loss": 0.0011, "step": 1895 }, { "epoch": 2.842578710644678, "grad_norm": 0.04788444160974722, "learning_rate": 1.6745092436045494e-06, "loss": 0.0035, "step": 1896 }, { "epoch": 2.84407796101949, "grad_norm": 0.017049447472496787, "learning_rate": 1.6428529186614195e-06, "loss": 0.0005, "step": 1897 }, { "epoch": 2.845577211394303, "grad_norm": 0.05683597326229523, "learning_rate": 1.611496206645835e-06, "loss": 0.0035, "step": 1898 }, { "epoch": 2.8470764617691153, "grad_norm": 0.08486491218156729, "learning_rate": 1.580439203075812e-06, "loss": 0.0025, "step": 1899 }, { "epoch": 2.848575712143928, "grad_norm": 0.03387217796464905, "learning_rate": 1.5496820025563409e-06, "loss": 0.0013, "step": 1900 }, { "epoch": 2.8500749625187405, "grad_norm": 0.04816811922646314, "learning_rate": 1.5192246987791981e-06, "loss": 0.004, "step": 1901 }, { "epoch": 2.8515742128935533, "grad_norm": 0.07910064579652644, "learning_rate": 1.4890673845226133e-06, "loss": 0.0039, "step": 1902 }, { "epoch": 2.853073463268366, "grad_norm": 0.032934113339092096, "learning_rate": 1.4592101516509914e-06, "loss": 0.0016, "step": 1903 }, { "epoch": 2.8545727136431784, "grad_norm": 0.0449861007395873, "learning_rate": 1.4296530911146466e-06, "loss": 0.0033, "step": 1904 }, { "epoch": 2.856071964017991, "grad_norm": 0.0526419044525919, "learning_rate": 1.400396292949513e-06, "loss": 0.0025, "step": 1905 }, { "epoch": 2.8575712143928036, "grad_norm": 0.06178628171991531, "learning_rate": 1.3714398462768563e-06, "loss": 0.0026, "step": 1906 }, { "epoch": 2.8590704647676164, "grad_norm": 0.10844669704867088, "learning_rate": 1.3427838393030633e-06, "loss": 0.0093, "step": 1907 }, { "epoch": 2.8605697151424287, "grad_norm": 0.08539087115108733, "learning_rate": 1.3144283593192752e-06, "loss": 0.0072, "step": 1908 }, { "epoch": 2.862068965517241, "grad_norm": 0.09803222726799544, "learning_rate": 1.2863734927012095e-06, "loss": 0.0103, "step": 1909 }, { "epoch": 2.863568215892054, "grad_norm": 0.05577463106669858, "learning_rate": 1.2586193249088608e-06, "loss": 0.0028, "step": 1910 }, { "epoch": 2.8650674662668667, "grad_norm": 0.2114427332970624, "learning_rate": 1.231165940486234e-06, "loss": 0.0313, "step": 1911 }, { "epoch": 2.866566716641679, "grad_norm": 0.14063204905909968, "learning_rate": 1.20401342306109e-06, "loss": 0.006, "step": 1912 }, { "epoch": 2.868065967016492, "grad_norm": 0.08238020394722131, "learning_rate": 1.1771618553447216e-06, "loss": 0.0076, "step": 1913 }, { "epoch": 2.869565217391304, "grad_norm": 0.02690643600372609, "learning_rate": 1.1506113191316447e-06, "loss": 0.0012, "step": 1914 }, { "epoch": 2.871064467766117, "grad_norm": 0.08444372271644372, "learning_rate": 1.1243618952994195e-06, "loss": 0.0052, "step": 1915 }, { "epoch": 2.8725637181409294, "grad_norm": 0.042791866109907366, "learning_rate": 1.0984136638083177e-06, "loss": 0.0018, "step": 1916 }, { "epoch": 2.874062968515742, "grad_norm": 0.0822631678321723, "learning_rate": 1.0727667037011668e-06, "loss": 0.007, "step": 1917 }, { "epoch": 2.875562218890555, "grad_norm": 0.06524093103957149, "learning_rate": 1.0474210931030514e-06, "loss": 0.0059, "step": 1918 }, { "epoch": 2.8770614692653673, "grad_norm": 0.03448809415597592, "learning_rate": 1.0223769092211012e-06, "loss": 0.0026, "step": 1919 }, { "epoch": 2.8785607196401797, "grad_norm": 0.046141321115156876, "learning_rate": 9.976342283442463e-07, "loss": 0.0022, "step": 1920 }, { "epoch": 2.8800599700149925, "grad_norm": 0.07243493966503523, "learning_rate": 9.731931258429638e-07, "loss": 0.0024, "step": 1921 }, { "epoch": 2.8815592203898053, "grad_norm": 0.056477163143498946, "learning_rate": 9.490536761691204e-07, "loss": 0.0033, "step": 1922 }, { "epoch": 2.8830584707646176, "grad_norm": 0.04470477657428264, "learning_rate": 9.252159528556403e-07, "loss": 0.0024, "step": 1923 }, { "epoch": 2.8845577211394304, "grad_norm": 0.027848123681753996, "learning_rate": 9.016800285163718e-07, "loss": 0.0024, "step": 1924 }, { "epoch": 2.886056971514243, "grad_norm": 0.05260745768799866, "learning_rate": 8.784459748458318e-07, "loss": 0.0046, "step": 1925 }, { "epoch": 2.8875562218890556, "grad_norm": 0.04282082202332062, "learning_rate": 8.555138626189618e-07, "loss": 0.0017, "step": 1926 }, { "epoch": 2.889055472263868, "grad_norm": 0.02613070330147928, "learning_rate": 8.328837616909613e-07, "loss": 0.0019, "step": 1927 }, { "epoch": 2.8905547226386807, "grad_norm": 0.10700496393861944, "learning_rate": 8.105557409970432e-07, "loss": 0.009, "step": 1928 }, { "epoch": 2.892053973013493, "grad_norm": 0.0802537427420226, "learning_rate": 7.885298685522235e-07, "loss": 0.0038, "step": 1929 }, { "epoch": 2.893553223388306, "grad_norm": 0.0505236671377372, "learning_rate": 7.668062114511321e-07, "loss": 0.0019, "step": 1930 }, { "epoch": 2.8950524737631183, "grad_norm": 0.033520163658288964, "learning_rate": 7.453848358678017e-07, "loss": 0.0017, "step": 1931 }, { "epoch": 2.896551724137931, "grad_norm": 0.07259240298956116, "learning_rate": 7.242658070554464e-07, "loss": 0.0062, "step": 1932 }, { "epoch": 2.898050974512744, "grad_norm": 0.06383788331881589, "learning_rate": 7.034491893463058e-07, "loss": 0.0036, "step": 1933 }, { "epoch": 2.899550224887556, "grad_norm": 0.05929371421132908, "learning_rate": 6.829350461514006e-07, "loss": 0.0038, "step": 1934 }, { "epoch": 2.9010494752623686, "grad_norm": 0.10791565065380813, "learning_rate": 6.627234399603555e-07, "loss": 0.0076, "step": 1935 }, { "epoch": 2.9025487256371814, "grad_norm": 0.03199906382024088, "learning_rate": 6.428144323412544e-07, "loss": 0.002, "step": 1936 }, { "epoch": 2.904047976011994, "grad_norm": 0.044793332598965306, "learning_rate": 6.232080839403631e-07, "loss": 0.0031, "step": 1937 }, { "epoch": 2.9055472263868065, "grad_norm": 0.034949113643794064, "learning_rate": 6.039044544820404e-07, "loss": 0.0017, "step": 1938 }, { "epoch": 2.9070464767616193, "grad_norm": 0.061169476787731235, "learning_rate": 5.849036027684606e-07, "loss": 0.003, "step": 1939 }, { "epoch": 2.9085457271364317, "grad_norm": 0.09671098321705973, "learning_rate": 5.662055866795357e-07, "loss": 0.0048, "step": 1940 }, { "epoch": 2.9100449775112445, "grad_norm": 0.09483559527940648, "learning_rate": 5.478104631726711e-07, "loss": 0.0027, "step": 1941 }, { "epoch": 2.911544227886057, "grad_norm": 0.03107820245195978, "learning_rate": 5.297182882825879e-07, "loss": 0.0015, "step": 1942 }, { "epoch": 2.9130434782608696, "grad_norm": 0.07421783376655451, "learning_rate": 5.119291171211793e-07, "loss": 0.0094, "step": 1943 }, { "epoch": 2.9145427286356824, "grad_norm": 0.09218285026776223, "learning_rate": 4.944430038773762e-07, "loss": 0.0085, "step": 1944 }, { "epoch": 2.9160419790104948, "grad_norm": 0.07360416997364519, "learning_rate": 4.772600018168816e-07, "loss": 0.0033, "step": 1945 }, { "epoch": 2.917541229385307, "grad_norm": 0.043634796881273585, "learning_rate": 4.6038016328211476e-07, "loss": 0.0034, "step": 1946 }, { "epoch": 2.91904047976012, "grad_norm": 0.07685943518323182, "learning_rate": 4.438035396920004e-07, "loss": 0.005, "step": 1947 }, { "epoch": 2.9205397301349327, "grad_norm": 0.07806672517785551, "learning_rate": 4.275301815417909e-07, "loss": 0.002, "step": 1948 }, { "epoch": 2.922038980509745, "grad_norm": 0.04375018889513186, "learning_rate": 4.115601384029666e-07, "loss": 0.0045, "step": 1949 }, { "epoch": 2.923538230884558, "grad_norm": 0.1430345786843898, "learning_rate": 3.958934589230467e-07, "loss": 0.0085, "step": 1950 }, { "epoch": 2.9250374812593702, "grad_norm": 0.10516629111264399, "learning_rate": 3.805301908254455e-07, "loss": 0.0124, "step": 1951 }, { "epoch": 2.926536731634183, "grad_norm": 0.04196769665260615, "learning_rate": 3.654703809093607e-07, "loss": 0.0032, "step": 1952 }, { "epoch": 2.9280359820089954, "grad_norm": 0.12229056029404549, "learning_rate": 3.50714075049563e-07, "loss": 0.0044, "step": 1953 }, { "epoch": 2.929535232383808, "grad_norm": 0.05573736381133186, "learning_rate": 3.362613181963403e-07, "loss": 0.0029, "step": 1954 }, { "epoch": 2.9310344827586206, "grad_norm": 0.040779137055409795, "learning_rate": 3.2211215437528694e-07, "loss": 0.0021, "step": 1955 }, { "epoch": 2.9325337331334334, "grad_norm": 0.053397472075800284, "learning_rate": 3.0826662668720364e-07, "loss": 0.0069, "step": 1956 }, { "epoch": 2.9340329835082457, "grad_norm": 0.07113095453573431, "learning_rate": 2.947247773079753e-07, "loss": 0.0076, "step": 1957 }, { "epoch": 2.9355322338830585, "grad_norm": 0.04310065622239369, "learning_rate": 2.81486647488427e-07, "loss": 0.0038, "step": 1958 }, { "epoch": 2.9370314842578713, "grad_norm": 0.12131577298406716, "learning_rate": 2.685522775541904e-07, "loss": 0.0058, "step": 1959 }, { "epoch": 2.9385307346326837, "grad_norm": 0.06607798391376911, "learning_rate": 2.5592170690560414e-07, "loss": 0.0033, "step": 1960 }, { "epoch": 2.940029985007496, "grad_norm": 0.07686025540345201, "learning_rate": 2.4359497401758024e-07, "loss": 0.0067, "step": 1961 }, { "epoch": 2.941529235382309, "grad_norm": 0.12480781751450243, "learning_rate": 2.315721164394713e-07, "loss": 0.0092, "step": 1962 }, { "epoch": 2.9430284857571216, "grad_norm": 0.06710923415617372, "learning_rate": 2.1985317079500356e-07, "loss": 0.0039, "step": 1963 }, { "epoch": 2.944527736131934, "grad_norm": 0.05328518709101643, "learning_rate": 2.0843817278209942e-07, "loss": 0.0041, "step": 1964 }, { "epoch": 2.9460269865067468, "grad_norm": 0.03205720007508785, "learning_rate": 1.973271571728441e-07, "loss": 0.0021, "step": 1965 }, { "epoch": 2.947526236881559, "grad_norm": 0.08713150735817307, "learning_rate": 1.86520157813308e-07, "loss": 0.005, "step": 1966 }, { "epoch": 2.949025487256372, "grad_norm": 0.0897607603221516, "learning_rate": 1.7601720762346897e-07, "loss": 0.0109, "step": 1967 }, { "epoch": 2.9505247376311843, "grad_norm": 0.09456767421388704, "learning_rate": 1.6581833859716788e-07, "loss": 0.0025, "step": 1968 }, { "epoch": 2.952023988005997, "grad_norm": 0.02654611858782301, "learning_rate": 1.5592358180189782e-07, "loss": 0.0011, "step": 1969 }, { "epoch": 2.9535232383808094, "grad_norm": 0.057986054790013425, "learning_rate": 1.4633296737882607e-07, "loss": 0.0051, "step": 1970 }, { "epoch": 2.9550224887556222, "grad_norm": 0.05325409118977268, "learning_rate": 1.3704652454261668e-07, "loss": 0.0026, "step": 1971 }, { "epoch": 2.9565217391304346, "grad_norm": 0.05995378335721647, "learning_rate": 1.2806428158138596e-07, "loss": 0.0039, "step": 1972 }, { "epoch": 2.9580209895052474, "grad_norm": 0.11321346048305181, "learning_rate": 1.193862658566025e-07, "loss": 0.0046, "step": 1973 }, { "epoch": 2.95952023988006, "grad_norm": 0.07230973808230436, "learning_rate": 1.1101250380300965e-07, "loss": 0.0055, "step": 1974 }, { "epoch": 2.9610194902548725, "grad_norm": 0.08029645889349672, "learning_rate": 1.0294302092853647e-07, "loss": 0.0035, "step": 1975 }, { "epoch": 2.962518740629685, "grad_norm": 0.041158986213349215, "learning_rate": 9.517784181422019e-08, "loss": 0.0047, "step": 1976 }, { "epoch": 2.9640179910044977, "grad_norm": 0.05686175756545626, "learning_rate": 8.771699011416168e-08, "loss": 0.0038, "step": 1977 }, { "epoch": 2.9655172413793105, "grad_norm": 0.04893183911640975, "learning_rate": 8.056048855540343e-08, "loss": 0.0346, "step": 1978 }, { "epoch": 2.967016491754123, "grad_norm": 0.1379441273961598, "learning_rate": 7.370835893788508e-08, "loss": 0.0091, "step": 1979 }, { "epoch": 2.9685157421289357, "grad_norm": 0.06399841406714543, "learning_rate": 6.71606221343768e-08, "loss": 0.0043, "step": 1980 }, { "epoch": 2.970014992503748, "grad_norm": 0.11199174671151793, "learning_rate": 6.09172980904238e-08, "loss": 0.0122, "step": 1981 }, { "epoch": 2.971514242878561, "grad_norm": 0.04482360637940532, "learning_rate": 5.497840582429081e-08, "loss": 0.0034, "step": 1982 }, { "epoch": 2.973013493253373, "grad_norm": 0.17153408725973948, "learning_rate": 4.934396342684e-08, "loss": 0.0052, "step": 1983 }, { "epoch": 2.974512743628186, "grad_norm": 0.04980670314050741, "learning_rate": 4.401398806159751e-08, "loss": 0.0029, "step": 1984 }, { "epoch": 2.9760119940029988, "grad_norm": 0.08613536118025303, "learning_rate": 3.898849596456478e-08, "loss": 0.0064, "step": 1985 }, { "epoch": 2.977511244377811, "grad_norm": 0.03045405966760291, "learning_rate": 3.4267502444274015e-08, "loss": 0.0016, "step": 1986 }, { "epoch": 2.9790104947526235, "grad_norm": 0.036884946821497074, "learning_rate": 2.985102188168831e-08, "loss": 0.0019, "step": 1987 }, { "epoch": 2.9805097451274363, "grad_norm": 0.0524860686083844, "learning_rate": 2.573906773016832e-08, "loss": 0.0034, "step": 1988 }, { "epoch": 2.982008995502249, "grad_norm": 0.06829182858407432, "learning_rate": 2.193165251545004e-08, "loss": 0.0077, "step": 1989 }, { "epoch": 2.9835082458770614, "grad_norm": 0.05804353038563701, "learning_rate": 1.842878783557822e-08, "loss": 0.0024, "step": 1990 }, { "epoch": 2.9850074962518742, "grad_norm": 0.023737341328751025, "learning_rate": 1.5230484360873044e-08, "loss": 0.0012, "step": 1991 }, { "epoch": 2.9865067466266866, "grad_norm": 0.04638140957693581, "learning_rate": 1.2336751833941229e-08, "loss": 0.0019, "step": 1992 }, { "epoch": 2.9880059970014994, "grad_norm": 0.10493790831990461, "learning_rate": 9.747599069576119e-09, "loss": 0.008, "step": 1993 }, { "epoch": 2.9895052473763117, "grad_norm": 0.06846796686090073, "learning_rate": 7.463033954802078e-09, "loss": 0.0064, "step": 1994 }, { "epoch": 2.9910044977511245, "grad_norm": 0.11136491254887901, "learning_rate": 5.483063448785686e-09, "loss": 0.0059, "step": 1995 }, { "epoch": 2.992503748125937, "grad_norm": 0.10745241960262616, "learning_rate": 3.807693582869032e-09, "loss": 0.0033, "step": 1996 }, { "epoch": 2.9940029985007497, "grad_norm": 0.08617751521895838, "learning_rate": 2.4369294605253166e-09, "loss": 0.0041, "step": 1997 }, { "epoch": 2.995502248875562, "grad_norm": 0.11414137883273419, "learning_rate": 1.3707752573255405e-09, "loss": 0.0066, "step": 1998 }, { "epoch": 2.997001499250375, "grad_norm": 0.07137774757587402, "learning_rate": 6.092342209607083e-10, "loss": 0.0047, "step": 1999 }, { "epoch": 2.9985007496251876, "grad_norm": 0.04455773434098678, "learning_rate": 1.5230867123072756e-10, "loss": 0.0032, "step": 2000 }, { "epoch": 3.0, "grad_norm": 0.05174225537101476, "learning_rate": 0.0, "loss": 0.0019, "step": 2001 }, { "epoch": 3.0, "eval_loss": 0.03369831293821335, "eval_runtime": 481.6669, "eval_samples_per_second": 21.542, "eval_steps_per_second": 1.347, "step": 2001 }, { "epoch": 3.0, "step": 2001, "total_flos": 2.2064687673350554e+17, "train_loss": 0.016641332297180983, "train_runtime": 13115.3958, "train_samples_per_second": 4.882, "train_steps_per_second": 0.153 } ], "logging_steps": 1, "max_steps": 2001, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2.2064687673350554e+17, "train_batch_size": 2, "trial_name": null, "trial_params": null }