{ "best_metric": 0.015713628381490707, "best_model_checkpoint": "/home/paperspace/Data/models/asianpaints/llm3br256-v1.5/checkpoint-1700", "epoch": 9.178743961352657, "eval_steps": 25, "global_step": 1900, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.004830917874396135, "grad_norm": 0.5997226238250732, "learning_rate": 1.9305019305019306e-07, "loss": 0.1739, "step": 1 }, { "epoch": 0.00966183574879227, "grad_norm": 0.586310863494873, "learning_rate": 3.8610038610038613e-07, "loss": 0.1804, "step": 2 }, { "epoch": 0.014492753623188406, "grad_norm": 0.6426325440406799, "learning_rate": 5.791505791505791e-07, "loss": 0.1612, "step": 3 }, { "epoch": 0.01932367149758454, "grad_norm": 0.5951049327850342, "learning_rate": 7.722007722007723e-07, "loss": 0.1851, "step": 4 }, { "epoch": 0.024154589371980676, "grad_norm": 0.574536144733429, "learning_rate": 9.652509652509653e-07, "loss": 0.1837, "step": 5 }, { "epoch": 0.028985507246376812, "grad_norm": 0.5631237030029297, "learning_rate": 1.1583011583011583e-06, "loss": 0.1905, "step": 6 }, { "epoch": 0.033816425120772944, "grad_norm": 0.5657174587249756, "learning_rate": 1.3513513513513515e-06, "loss": 0.1719, "step": 7 }, { "epoch": 0.03864734299516908, "grad_norm": 0.5333519577980042, "learning_rate": 1.5444015444015445e-06, "loss": 0.1662, "step": 8 }, { "epoch": 0.043478260869565216, "grad_norm": 0.5621788501739502, "learning_rate": 1.7374517374517375e-06, "loss": 0.1862, "step": 9 }, { "epoch": 0.04830917874396135, "grad_norm": 0.5728170275688171, "learning_rate": 1.9305019305019305e-06, "loss": 0.1817, "step": 10 }, { "epoch": 0.05314009661835749, "grad_norm": 0.5463429689407349, "learning_rate": 2.1235521235521236e-06, "loss": 0.1787, "step": 11 }, { "epoch": 0.057971014492753624, "grad_norm": 0.5315811038017273, "learning_rate": 2.3166023166023166e-06, "loss": 0.1844, "step": 12 }, { "epoch": 0.06280193236714976, "grad_norm": 0.4727146625518799, "learning_rate": 2.5096525096525096e-06, "loss": 0.1704, "step": 13 }, { "epoch": 0.06763285024154589, "grad_norm": 0.4733094573020935, "learning_rate": 2.702702702702703e-06, "loss": 0.1517, "step": 14 }, { "epoch": 0.07246376811594203, "grad_norm": 0.42338845133781433, "learning_rate": 2.895752895752896e-06, "loss": 0.175, "step": 15 }, { "epoch": 0.07729468599033816, "grad_norm": 0.3983526825904846, "learning_rate": 3.088803088803089e-06, "loss": 0.1692, "step": 16 }, { "epoch": 0.0821256038647343, "grad_norm": 0.3729923963546753, "learning_rate": 3.2818532818532816e-06, "loss": 0.1597, "step": 17 }, { "epoch": 0.08695652173913043, "grad_norm": 0.3619980216026306, "learning_rate": 3.474903474903475e-06, "loss": 0.1753, "step": 18 }, { "epoch": 0.09178743961352658, "grad_norm": 0.3523077070713043, "learning_rate": 3.6679536679536685e-06, "loss": 0.1581, "step": 19 }, { "epoch": 0.0966183574879227, "grad_norm": 0.33768779039382935, "learning_rate": 3.861003861003861e-06, "loss": 0.1625, "step": 20 }, { "epoch": 0.10144927536231885, "grad_norm": 0.31936874985694885, "learning_rate": 4.0540540540540545e-06, "loss": 0.1505, "step": 21 }, { "epoch": 0.10628019323671498, "grad_norm": 0.324440598487854, "learning_rate": 4.247104247104247e-06, "loss": 0.1468, "step": 22 }, { "epoch": 0.1111111111111111, "grad_norm": 0.3510577380657196, "learning_rate": 4.4401544401544405e-06, "loss": 0.1703, "step": 23 }, { "epoch": 0.11594202898550725, "grad_norm": 0.295952707529068, "learning_rate": 4.633204633204633e-06, "loss": 0.1359, "step": 24 }, { "epoch": 0.12077294685990338, "grad_norm": 0.284309059381485, "learning_rate": 4.8262548262548266e-06, "loss": 0.1302, "step": 25 }, { "epoch": 0.12077294685990338, "eval_loss": 0.14291973412036896, "eval_runtime": 23.4942, "eval_samples_per_second": 4.256, "eval_steps_per_second": 0.128, "step": 25 }, { "epoch": 0.12560386473429952, "grad_norm": 0.2970901131629944, "learning_rate": 5.019305019305019e-06, "loss": 0.131, "step": 26 }, { "epoch": 0.13043478260869565, "grad_norm": 0.2529473304748535, "learning_rate": 5.212355212355213e-06, "loss": 0.1251, "step": 27 }, { "epoch": 0.13526570048309178, "grad_norm": 0.2643989026546478, "learning_rate": 5.405405405405406e-06, "loss": 0.1462, "step": 28 }, { "epoch": 0.14009661835748793, "grad_norm": 0.2184019535779953, "learning_rate": 5.598455598455599e-06, "loss": 0.1332, "step": 29 }, { "epoch": 0.14492753623188406, "grad_norm": 0.19679449498653412, "learning_rate": 5.791505791505792e-06, "loss": 0.1343, "step": 30 }, { "epoch": 0.1497584541062802, "grad_norm": 0.19991901516914368, "learning_rate": 5.984555984555985e-06, "loss": 0.1373, "step": 31 }, { "epoch": 0.15458937198067632, "grad_norm": 0.23242047429084778, "learning_rate": 6.177606177606178e-06, "loss": 0.1496, "step": 32 }, { "epoch": 0.15942028985507245, "grad_norm": 0.21723514795303345, "learning_rate": 6.370656370656371e-06, "loss": 0.1167, "step": 33 }, { "epoch": 0.1642512077294686, "grad_norm": 0.1968252956867218, "learning_rate": 6.563706563706563e-06, "loss": 0.125, "step": 34 }, { "epoch": 0.16908212560386474, "grad_norm": 0.1917056292295456, "learning_rate": 6.7567567567567575e-06, "loss": 0.1311, "step": 35 }, { "epoch": 0.17391304347826086, "grad_norm": 0.18152444064617157, "learning_rate": 6.94980694980695e-06, "loss": 0.1158, "step": 36 }, { "epoch": 0.178743961352657, "grad_norm": 0.20013165473937988, "learning_rate": 7.142857142857143e-06, "loss": 0.1153, "step": 37 }, { "epoch": 0.18357487922705315, "grad_norm": 0.190288707613945, "learning_rate": 7.335907335907337e-06, "loss": 0.1039, "step": 38 }, { "epoch": 0.18840579710144928, "grad_norm": 0.19227007031440735, "learning_rate": 7.52895752895753e-06, "loss": 0.1261, "step": 39 }, { "epoch": 0.1932367149758454, "grad_norm": 0.19719913601875305, "learning_rate": 7.722007722007722e-06, "loss": 0.1233, "step": 40 }, { "epoch": 0.19806763285024154, "grad_norm": 0.1750050187110901, "learning_rate": 7.915057915057915e-06, "loss": 0.1132, "step": 41 }, { "epoch": 0.2028985507246377, "grad_norm": 0.16355034708976746, "learning_rate": 8.108108108108109e-06, "loss": 0.119, "step": 42 }, { "epoch": 0.20772946859903382, "grad_norm": 0.16448749601840973, "learning_rate": 8.301158301158302e-06, "loss": 0.1124, "step": 43 }, { "epoch": 0.21256038647342995, "grad_norm": 0.16512104868888855, "learning_rate": 8.494208494208494e-06, "loss": 0.0985, "step": 44 }, { "epoch": 0.21739130434782608, "grad_norm": 0.18671417236328125, "learning_rate": 8.687258687258689e-06, "loss": 0.1016, "step": 45 }, { "epoch": 0.2222222222222222, "grad_norm": 0.18079911172389984, "learning_rate": 8.880308880308881e-06, "loss": 0.1053, "step": 46 }, { "epoch": 0.22705314009661837, "grad_norm": 0.16191694140434265, "learning_rate": 9.073359073359074e-06, "loss": 0.1205, "step": 47 }, { "epoch": 0.2318840579710145, "grad_norm": 0.15231011807918549, "learning_rate": 9.266409266409266e-06, "loss": 0.098, "step": 48 }, { "epoch": 0.23671497584541062, "grad_norm": 0.14875712990760803, "learning_rate": 9.45945945945946e-06, "loss": 0.089, "step": 49 }, { "epoch": 0.24154589371980675, "grad_norm": 0.13682980835437775, "learning_rate": 9.652509652509653e-06, "loss": 0.0909, "step": 50 }, { "epoch": 0.24154589371980675, "eval_loss": 0.09681461751461029, "eval_runtime": 20.6089, "eval_samples_per_second": 4.852, "eval_steps_per_second": 0.146, "step": 50 }, { "epoch": 0.2463768115942029, "grad_norm": 0.14475570619106293, "learning_rate": 9.845559845559846e-06, "loss": 0.0881, "step": 51 }, { "epoch": 0.25120772946859904, "grad_norm": 0.13333068788051605, "learning_rate": 1.0038610038610038e-05, "loss": 0.0858, "step": 52 }, { "epoch": 0.2560386473429952, "grad_norm": 0.15850719809532166, "learning_rate": 1.0231660231660233e-05, "loss": 0.1189, "step": 53 }, { "epoch": 0.2608695652173913, "grad_norm": 0.1379750370979309, "learning_rate": 1.0424710424710425e-05, "loss": 0.0839, "step": 54 }, { "epoch": 0.26570048309178745, "grad_norm": 0.1474492847919464, "learning_rate": 1.0617760617760618e-05, "loss": 0.0921, "step": 55 }, { "epoch": 0.27053140096618356, "grad_norm": 0.12119033187627792, "learning_rate": 1.0810810810810812e-05, "loss": 0.0814, "step": 56 }, { "epoch": 0.2753623188405797, "grad_norm": 0.1305706650018692, "learning_rate": 1.1003861003861005e-05, "loss": 0.0896, "step": 57 }, { "epoch": 0.28019323671497587, "grad_norm": 0.127845898270607, "learning_rate": 1.1196911196911197e-05, "loss": 0.0744, "step": 58 }, { "epoch": 0.28502415458937197, "grad_norm": 0.1521938443183899, "learning_rate": 1.138996138996139e-05, "loss": 0.087, "step": 59 }, { "epoch": 0.2898550724637681, "grad_norm": 0.12970903515815735, "learning_rate": 1.1583011583011584e-05, "loss": 0.0721, "step": 60 }, { "epoch": 0.2946859903381642, "grad_norm": 0.13914500176906586, "learning_rate": 1.1776061776061777e-05, "loss": 0.0765, "step": 61 }, { "epoch": 0.2995169082125604, "grad_norm": 0.1305667757987976, "learning_rate": 1.196911196911197e-05, "loss": 0.0822, "step": 62 }, { "epoch": 0.30434782608695654, "grad_norm": 0.15017420053482056, "learning_rate": 1.2162162162162164e-05, "loss": 0.1075, "step": 63 }, { "epoch": 0.30917874396135264, "grad_norm": 0.15811654925346375, "learning_rate": 1.2355212355212356e-05, "loss": 0.0992, "step": 64 }, { "epoch": 0.3140096618357488, "grad_norm": 0.17078033089637756, "learning_rate": 1.2548262548262549e-05, "loss": 0.108, "step": 65 }, { "epoch": 0.3188405797101449, "grad_norm": 0.14823614060878754, "learning_rate": 1.2741312741312741e-05, "loss": 0.0707, "step": 66 }, { "epoch": 0.32367149758454106, "grad_norm": 0.11846248060464859, "learning_rate": 1.2934362934362934e-05, "loss": 0.0836, "step": 67 }, { "epoch": 0.3285024154589372, "grad_norm": 0.1331644058227539, "learning_rate": 1.3127413127413127e-05, "loss": 0.0847, "step": 68 }, { "epoch": 0.3333333333333333, "grad_norm": 0.12533916532993317, "learning_rate": 1.3320463320463322e-05, "loss": 0.0616, "step": 69 }, { "epoch": 0.33816425120772947, "grad_norm": 0.13879841566085815, "learning_rate": 1.3513513513513515e-05, "loss": 0.0886, "step": 70 }, { "epoch": 0.34299516908212563, "grad_norm": 0.13229133188724518, "learning_rate": 1.3706563706563708e-05, "loss": 0.081, "step": 71 }, { "epoch": 0.34782608695652173, "grad_norm": 0.12827053666114807, "learning_rate": 1.38996138996139e-05, "loss": 0.0778, "step": 72 }, { "epoch": 0.3526570048309179, "grad_norm": 0.11902689933776855, "learning_rate": 1.4092664092664093e-05, "loss": 0.0695, "step": 73 }, { "epoch": 0.357487922705314, "grad_norm": 0.14065556228160858, "learning_rate": 1.4285714285714285e-05, "loss": 0.0702, "step": 74 }, { "epoch": 0.36231884057971014, "grad_norm": 0.11380023509263992, "learning_rate": 1.4478764478764478e-05, "loss": 0.062, "step": 75 }, { "epoch": 0.36231884057971014, "eval_loss": 0.07478907704353333, "eval_runtime": 20.612, "eval_samples_per_second": 4.852, "eval_steps_per_second": 0.146, "step": 75 }, { "epoch": 0.3671497584541063, "grad_norm": 0.13626235723495483, "learning_rate": 1.4671814671814674e-05, "loss": 0.0897, "step": 76 }, { "epoch": 0.3719806763285024, "grad_norm": 0.11516878753900528, "learning_rate": 1.4864864864864867e-05, "loss": 0.0613, "step": 77 }, { "epoch": 0.37681159420289856, "grad_norm": 0.11802902072668076, "learning_rate": 1.505791505791506e-05, "loss": 0.0661, "step": 78 }, { "epoch": 0.38164251207729466, "grad_norm": 0.14430475234985352, "learning_rate": 1.5250965250965252e-05, "loss": 0.0765, "step": 79 }, { "epoch": 0.3864734299516908, "grad_norm": 0.11286959052085876, "learning_rate": 1.5444015444015444e-05, "loss": 0.0699, "step": 80 }, { "epoch": 0.391304347826087, "grad_norm": 0.12507657706737518, "learning_rate": 1.5637065637065637e-05, "loss": 0.0734, "step": 81 }, { "epoch": 0.3961352657004831, "grad_norm": 0.12890633940696716, "learning_rate": 1.583011583011583e-05, "loss": 0.0715, "step": 82 }, { "epoch": 0.40096618357487923, "grad_norm": 0.12751871347427368, "learning_rate": 1.6023166023166026e-05, "loss": 0.0699, "step": 83 }, { "epoch": 0.4057971014492754, "grad_norm": 0.1366698145866394, "learning_rate": 1.6216216216216218e-05, "loss": 0.0765, "step": 84 }, { "epoch": 0.4106280193236715, "grad_norm": 0.10429048538208008, "learning_rate": 1.640926640926641e-05, "loss": 0.057, "step": 85 }, { "epoch": 0.41545893719806765, "grad_norm": 0.10773543268442154, "learning_rate": 1.6602316602316603e-05, "loss": 0.0666, "step": 86 }, { "epoch": 0.42028985507246375, "grad_norm": 0.110495425760746, "learning_rate": 1.6795366795366796e-05, "loss": 0.0635, "step": 87 }, { "epoch": 0.4251207729468599, "grad_norm": 0.139565110206604, "learning_rate": 1.698841698841699e-05, "loss": 0.0874, "step": 88 }, { "epoch": 0.42995169082125606, "grad_norm": 0.1371033936738968, "learning_rate": 1.718146718146718e-05, "loss": 0.0669, "step": 89 }, { "epoch": 0.43478260869565216, "grad_norm": 0.14201398193836212, "learning_rate": 1.7374517374517377e-05, "loss": 0.0638, "step": 90 }, { "epoch": 0.4396135265700483, "grad_norm": 0.13561639189720154, "learning_rate": 1.756756756756757e-05, "loss": 0.0654, "step": 91 }, { "epoch": 0.4444444444444444, "grad_norm": 0.11701712012290955, "learning_rate": 1.7760617760617762e-05, "loss": 0.0589, "step": 92 }, { "epoch": 0.4492753623188406, "grad_norm": 0.11068889498710632, "learning_rate": 1.7953667953667955e-05, "loss": 0.0707, "step": 93 }, { "epoch": 0.45410628019323673, "grad_norm": 0.11085647344589233, "learning_rate": 1.8146718146718147e-05, "loss": 0.0604, "step": 94 }, { "epoch": 0.45893719806763283, "grad_norm": 0.11306110769510269, "learning_rate": 1.833976833976834e-05, "loss": 0.0526, "step": 95 }, { "epoch": 0.463768115942029, "grad_norm": 0.12925806641578674, "learning_rate": 1.8532818532818533e-05, "loss": 0.0541, "step": 96 }, { "epoch": 0.46859903381642515, "grad_norm": 0.11406856775283813, "learning_rate": 1.8725868725868725e-05, "loss": 0.0575, "step": 97 }, { "epoch": 0.47342995169082125, "grad_norm": 0.11861402541399002, "learning_rate": 1.891891891891892e-05, "loss": 0.0593, "step": 98 }, { "epoch": 0.4782608695652174, "grad_norm": 0.15513893961906433, "learning_rate": 1.9111969111969114e-05, "loss": 0.0565, "step": 99 }, { "epoch": 0.4830917874396135, "grad_norm": 0.1187349483370781, "learning_rate": 1.9305019305019306e-05, "loss": 0.0602, "step": 100 }, { "epoch": 0.4830917874396135, "eval_loss": 0.06083356589078903, "eval_runtime": 20.6118, "eval_samples_per_second": 4.852, "eval_steps_per_second": 0.146, "step": 100 }, { "epoch": 0.48792270531400966, "grad_norm": 0.11506832391023636, "learning_rate": 1.94980694980695e-05, "loss": 0.0572, "step": 101 }, { "epoch": 0.4927536231884058, "grad_norm": 0.12325363606214523, "learning_rate": 1.969111969111969e-05, "loss": 0.0615, "step": 102 }, { "epoch": 0.4975845410628019, "grad_norm": 0.13748331367969513, "learning_rate": 1.9884169884169884e-05, "loss": 0.0644, "step": 103 }, { "epoch": 0.5024154589371981, "grad_norm": 0.12451450526714325, "learning_rate": 2.0077220077220077e-05, "loss": 0.0532, "step": 104 }, { "epoch": 0.5072463768115942, "grad_norm": 0.13660487532615662, "learning_rate": 2.0270270270270273e-05, "loss": 0.0505, "step": 105 }, { "epoch": 0.5120772946859904, "grad_norm": 0.11977998912334442, "learning_rate": 2.0463320463320465e-05, "loss": 0.0587, "step": 106 }, { "epoch": 0.5169082125603864, "grad_norm": 0.11822707206010818, "learning_rate": 2.0656370656370658e-05, "loss": 0.0465, "step": 107 }, { "epoch": 0.5217391304347826, "grad_norm": 0.13716371357440948, "learning_rate": 2.084942084942085e-05, "loss": 0.0452, "step": 108 }, { "epoch": 0.5265700483091788, "grad_norm": 0.1545065939426422, "learning_rate": 2.1042471042471043e-05, "loss": 0.0589, "step": 109 }, { "epoch": 0.5314009661835749, "grad_norm": 0.12922868132591248, "learning_rate": 2.1235521235521236e-05, "loss": 0.0758, "step": 110 }, { "epoch": 0.5362318840579711, "grad_norm": 0.17041181027889252, "learning_rate": 2.1428571428571428e-05, "loss": 0.0653, "step": 111 }, { "epoch": 0.5410628019323671, "grad_norm": 0.14169234037399292, "learning_rate": 2.1621621621621624e-05, "loss": 0.0567, "step": 112 }, { "epoch": 0.5458937198067633, "grad_norm": 0.158858060836792, "learning_rate": 2.1814671814671817e-05, "loss": 0.0492, "step": 113 }, { "epoch": 0.5507246376811594, "grad_norm": 0.17282356321811676, "learning_rate": 2.200772200772201e-05, "loss": 0.0561, "step": 114 }, { "epoch": 0.5555555555555556, "grad_norm": 0.12390092015266418, "learning_rate": 2.2200772200772202e-05, "loss": 0.0671, "step": 115 }, { "epoch": 0.5603864734299517, "grad_norm": 0.1567428857088089, "learning_rate": 2.2393822393822394e-05, "loss": 0.0556, "step": 116 }, { "epoch": 0.5652173913043478, "grad_norm": 0.16888275742530823, "learning_rate": 2.2586872586872587e-05, "loss": 0.0521, "step": 117 }, { "epoch": 0.5700483091787439, "grad_norm": 0.1285606473684311, "learning_rate": 2.277992277992278e-05, "loss": 0.0558, "step": 118 }, { "epoch": 0.5748792270531401, "grad_norm": 0.12908753752708435, "learning_rate": 2.2972972972972976e-05, "loss": 0.0872, "step": 119 }, { "epoch": 0.5797101449275363, "grad_norm": 0.13818618655204773, "learning_rate": 2.3166023166023168e-05, "loss": 0.0486, "step": 120 }, { "epoch": 0.5845410628019324, "grad_norm": 0.14711114764213562, "learning_rate": 2.335907335907336e-05, "loss": 0.0608, "step": 121 }, { "epoch": 0.5893719806763285, "grad_norm": 0.1265874058008194, "learning_rate": 2.3552123552123553e-05, "loss": 0.0816, "step": 122 }, { "epoch": 0.5942028985507246, "grad_norm": 0.1758117824792862, "learning_rate": 2.3745173745173746e-05, "loss": 0.0599, "step": 123 }, { "epoch": 0.5990338164251208, "grad_norm": 0.17134052515029907, "learning_rate": 2.393822393822394e-05, "loss": 0.0512, "step": 124 }, { "epoch": 0.6038647342995169, "grad_norm": 0.12125252187252045, "learning_rate": 2.413127413127413e-05, "loss": 0.057, "step": 125 }, { "epoch": 0.6038647342995169, "eval_loss": 0.05518662929534912, "eval_runtime": 20.6274, "eval_samples_per_second": 4.848, "eval_steps_per_second": 0.145, "step": 125 }, { "epoch": 0.6086956521739131, "grad_norm": 0.16893912851810455, "learning_rate": 2.4324324324324327e-05, "loss": 0.0477, "step": 126 }, { "epoch": 0.6135265700483091, "grad_norm": 0.16074183583259583, "learning_rate": 2.451737451737452e-05, "loss": 0.0544, "step": 127 }, { "epoch": 0.6183574879227053, "grad_norm": 0.14464056491851807, "learning_rate": 2.4710424710424712e-05, "loss": 0.0578, "step": 128 }, { "epoch": 0.6231884057971014, "grad_norm": 0.14202292263507843, "learning_rate": 2.4903474903474905e-05, "loss": 0.0398, "step": 129 }, { "epoch": 0.6280193236714976, "grad_norm": 0.16005530953407288, "learning_rate": 2.5096525096525097e-05, "loss": 0.0527, "step": 130 }, { "epoch": 0.6328502415458938, "grad_norm": 0.10633774846792221, "learning_rate": 2.528957528957529e-05, "loss": 0.0532, "step": 131 }, { "epoch": 0.6376811594202898, "grad_norm": 0.13469672203063965, "learning_rate": 2.5482625482625483e-05, "loss": 0.0477, "step": 132 }, { "epoch": 0.642512077294686, "grad_norm": 0.13366642594337463, "learning_rate": 2.5675675675675675e-05, "loss": 0.0473, "step": 133 }, { "epoch": 0.6473429951690821, "grad_norm": 0.13137054443359375, "learning_rate": 2.5868725868725868e-05, "loss": 0.0829, "step": 134 }, { "epoch": 0.6521739130434783, "grad_norm": 0.14475256204605103, "learning_rate": 2.606177606177606e-05, "loss": 0.0612, "step": 135 }, { "epoch": 0.6570048309178744, "grad_norm": 0.16254907846450806, "learning_rate": 2.6254826254826253e-05, "loss": 0.0557, "step": 136 }, { "epoch": 0.6618357487922706, "grad_norm": 0.11103875190019608, "learning_rate": 2.6447876447876452e-05, "loss": 0.0495, "step": 137 }, { "epoch": 0.6666666666666666, "grad_norm": 0.15056666731834412, "learning_rate": 2.6640926640926645e-05, "loss": 0.0443, "step": 138 }, { "epoch": 0.6714975845410628, "grad_norm": 0.1170201525092125, "learning_rate": 2.6833976833976838e-05, "loss": 0.0515, "step": 139 }, { "epoch": 0.6763285024154589, "grad_norm": 0.13160322606563568, "learning_rate": 2.702702702702703e-05, "loss": 0.0513, "step": 140 }, { "epoch": 0.6811594202898551, "grad_norm": 0.11938180029392242, "learning_rate": 2.7220077220077223e-05, "loss": 0.0473, "step": 141 }, { "epoch": 0.6859903381642513, "grad_norm": 0.11169520020484924, "learning_rate": 2.7413127413127415e-05, "loss": 0.0668, "step": 142 }, { "epoch": 0.6908212560386473, "grad_norm": 0.12496872991323471, "learning_rate": 2.7606177606177608e-05, "loss": 0.0609, "step": 143 }, { "epoch": 0.6956521739130435, "grad_norm": 0.150443434715271, "learning_rate": 2.77992277992278e-05, "loss": 0.0782, "step": 144 }, { "epoch": 0.7004830917874396, "grad_norm": 0.11996805667877197, "learning_rate": 2.7992277992277993e-05, "loss": 0.0475, "step": 145 }, { "epoch": 0.7053140096618358, "grad_norm": 0.11406496167182922, "learning_rate": 2.8185328185328186e-05, "loss": 0.0663, "step": 146 }, { "epoch": 0.7101449275362319, "grad_norm": 0.10466250032186508, "learning_rate": 2.8378378378378378e-05, "loss": 0.0483, "step": 147 }, { "epoch": 0.714975845410628, "grad_norm": 0.11609887331724167, "learning_rate": 2.857142857142857e-05, "loss": 0.0452, "step": 148 }, { "epoch": 0.7198067632850241, "grad_norm": 0.12703728675842285, "learning_rate": 2.8764478764478763e-05, "loss": 0.0504, "step": 149 }, { "epoch": 0.7246376811594203, "grad_norm": 0.11724966764450073, "learning_rate": 2.8957528957528956e-05, "loss": 0.0456, "step": 150 }, { "epoch": 0.7246376811594203, "eval_loss": 0.05008581280708313, "eval_runtime": 20.7274, "eval_samples_per_second": 4.825, "eval_steps_per_second": 0.145, "step": 150 }, { "epoch": 0.7294685990338164, "grad_norm": 0.11703478544950485, "learning_rate": 2.915057915057915e-05, "loss": 0.0445, "step": 151 }, { "epoch": 0.7342995169082126, "grad_norm": 0.1328444480895996, "learning_rate": 2.9343629343629348e-05, "loss": 0.0534, "step": 152 }, { "epoch": 0.7391304347826086, "grad_norm": 0.1250855177640915, "learning_rate": 2.953667953667954e-05, "loss": 0.0655, "step": 153 }, { "epoch": 0.7439613526570048, "grad_norm": 0.0979803130030632, "learning_rate": 2.9729729729729733e-05, "loss": 0.0484, "step": 154 }, { "epoch": 0.748792270531401, "grad_norm": 0.10639014095067978, "learning_rate": 2.9922779922779926e-05, "loss": 0.0458, "step": 155 }, { "epoch": 0.7536231884057971, "grad_norm": 0.10299219191074371, "learning_rate": 3.011583011583012e-05, "loss": 0.0343, "step": 156 }, { "epoch": 0.7584541062801933, "grad_norm": 0.12237931042909622, "learning_rate": 3.030888030888031e-05, "loss": 0.0493, "step": 157 }, { "epoch": 0.7632850241545893, "grad_norm": 0.14812174439430237, "learning_rate": 3.0501930501930504e-05, "loss": 0.0526, "step": 158 }, { "epoch": 0.7681159420289855, "grad_norm": 0.1112482100725174, "learning_rate": 3.0694980694980696e-05, "loss": 0.0427, "step": 159 }, { "epoch": 0.7729468599033816, "grad_norm": 0.13047181069850922, "learning_rate": 3.088803088803089e-05, "loss": 0.0584, "step": 160 }, { "epoch": 0.7777777777777778, "grad_norm": 0.1432395875453949, "learning_rate": 3.108108108108108e-05, "loss": 0.0499, "step": 161 }, { "epoch": 0.782608695652174, "grad_norm": 0.1461172252893448, "learning_rate": 3.1274131274131274e-05, "loss": 0.0636, "step": 162 }, { "epoch": 0.7874396135265701, "grad_norm": 0.11835489422082901, "learning_rate": 3.1467181467181466e-05, "loss": 0.0412, "step": 163 }, { "epoch": 0.7922705314009661, "grad_norm": 0.18038594722747803, "learning_rate": 3.166023166023166e-05, "loss": 0.0517, "step": 164 }, { "epoch": 0.7971014492753623, "grad_norm": 0.11978429555892944, "learning_rate": 3.185328185328185e-05, "loss": 0.037, "step": 165 }, { "epoch": 0.8019323671497585, "grad_norm": 0.13728344440460205, "learning_rate": 3.204633204633205e-05, "loss": 0.0346, "step": 166 }, { "epoch": 0.8067632850241546, "grad_norm": 0.173124760389328, "learning_rate": 3.2239382239382244e-05, "loss": 0.0512, "step": 167 }, { "epoch": 0.8115942028985508, "grad_norm": 0.09541524946689606, "learning_rate": 3.2432432432432436e-05, "loss": 0.0369, "step": 168 }, { "epoch": 0.8164251207729468, "grad_norm": 0.14173609018325806, "learning_rate": 3.262548262548263e-05, "loss": 0.0536, "step": 169 }, { "epoch": 0.821256038647343, "grad_norm": 0.14616850018501282, "learning_rate": 3.281853281853282e-05, "loss": 0.039, "step": 170 }, { "epoch": 0.8260869565217391, "grad_norm": 0.12013251334428787, "learning_rate": 3.3011583011583014e-05, "loss": 0.0488, "step": 171 }, { "epoch": 0.8309178743961353, "grad_norm": 0.1221630722284317, "learning_rate": 3.3204633204633207e-05, "loss": 0.0486, "step": 172 }, { "epoch": 0.8357487922705314, "grad_norm": 0.1088973730802536, "learning_rate": 3.33976833976834e-05, "loss": 0.043, "step": 173 }, { "epoch": 0.8405797101449275, "grad_norm": 0.11308977752923965, "learning_rate": 3.359073359073359e-05, "loss": 0.0478, "step": 174 }, { "epoch": 0.8454106280193237, "grad_norm": 0.10720162838697433, "learning_rate": 3.3783783783783784e-05, "loss": 0.0432, "step": 175 }, { "epoch": 0.8454106280193237, "eval_loss": 0.04698673263192177, "eval_runtime": 20.6051, "eval_samples_per_second": 4.853, "eval_steps_per_second": 0.146, "step": 175 }, { "epoch": 0.8502415458937198, "grad_norm": 0.1255665421485901, "learning_rate": 3.397683397683398e-05, "loss": 0.0444, "step": 176 }, { "epoch": 0.855072463768116, "grad_norm": 0.108990378677845, "learning_rate": 3.416988416988417e-05, "loss": 0.0469, "step": 177 }, { "epoch": 0.8599033816425121, "grad_norm": 0.10082966834306717, "learning_rate": 3.436293436293436e-05, "loss": 0.0508, "step": 178 }, { "epoch": 0.8647342995169082, "grad_norm": 0.11663074791431427, "learning_rate": 3.4555984555984555e-05, "loss": 0.04, "step": 179 }, { "epoch": 0.8695652173913043, "grad_norm": 0.11602170765399933, "learning_rate": 3.4749034749034754e-05, "loss": 0.0382, "step": 180 }, { "epoch": 0.8743961352657005, "grad_norm": 0.10239856690168381, "learning_rate": 3.4942084942084947e-05, "loss": 0.0384, "step": 181 }, { "epoch": 0.8792270531400966, "grad_norm": 0.1251702755689621, "learning_rate": 3.513513513513514e-05, "loss": 0.0557, "step": 182 }, { "epoch": 0.8840579710144928, "grad_norm": 0.1245378628373146, "learning_rate": 3.532818532818533e-05, "loss": 0.0564, "step": 183 }, { "epoch": 0.8888888888888888, "grad_norm": 0.11449820548295975, "learning_rate": 3.5521235521235524e-05, "loss": 0.0459, "step": 184 }, { "epoch": 0.893719806763285, "grad_norm": 0.11923693865537643, "learning_rate": 3.571428571428572e-05, "loss": 0.0382, "step": 185 }, { "epoch": 0.8985507246376812, "grad_norm": 0.132978156208992, "learning_rate": 3.590733590733591e-05, "loss": 0.0572, "step": 186 }, { "epoch": 0.9033816425120773, "grad_norm": 0.1146927997469902, "learning_rate": 3.61003861003861e-05, "loss": 0.034, "step": 187 }, { "epoch": 0.9082125603864735, "grad_norm": 0.1315777748823166, "learning_rate": 3.6293436293436295e-05, "loss": 0.0522, "step": 188 }, { "epoch": 0.9130434782608695, "grad_norm": 0.10078331083059311, "learning_rate": 3.648648648648649e-05, "loss": 0.0361, "step": 189 }, { "epoch": 0.9178743961352657, "grad_norm": 0.12355741858482361, "learning_rate": 3.667953667953668e-05, "loss": 0.0568, "step": 190 }, { "epoch": 0.9227053140096618, "grad_norm": 0.11051063239574432, "learning_rate": 3.687258687258687e-05, "loss": 0.053, "step": 191 }, { "epoch": 0.927536231884058, "grad_norm": 0.12470270693302155, "learning_rate": 3.7065637065637065e-05, "loss": 0.0517, "step": 192 }, { "epoch": 0.9323671497584541, "grad_norm": 0.10947062075138092, "learning_rate": 3.725868725868726e-05, "loss": 0.037, "step": 193 }, { "epoch": 0.9371980676328503, "grad_norm": 0.12240796536207199, "learning_rate": 3.745173745173745e-05, "loss": 0.0536, "step": 194 }, { "epoch": 0.9420289855072463, "grad_norm": 0.12782225012779236, "learning_rate": 3.764478764478765e-05, "loss": 0.0519, "step": 195 }, { "epoch": 0.9468599033816425, "grad_norm": 0.11264186352491379, "learning_rate": 3.783783783783784e-05, "loss": 0.043, "step": 196 }, { "epoch": 0.9516908212560387, "grad_norm": 0.10567909479141235, "learning_rate": 3.8030888030888035e-05, "loss": 0.0528, "step": 197 }, { "epoch": 0.9565217391304348, "grad_norm": 0.11025624722242355, "learning_rate": 3.822393822393823e-05, "loss": 0.0423, "step": 198 }, { "epoch": 0.961352657004831, "grad_norm": 0.12501060962677002, "learning_rate": 3.841698841698842e-05, "loss": 0.0466, "step": 199 }, { "epoch": 0.966183574879227, "grad_norm": 0.11633525043725967, "learning_rate": 3.861003861003861e-05, "loss": 0.0416, "step": 200 }, { "epoch": 0.966183574879227, "eval_loss": 0.04470054805278778, "eval_runtime": 21.3093, "eval_samples_per_second": 4.693, "eval_steps_per_second": 0.141, "step": 200 }, { "epoch": 0.9710144927536232, "grad_norm": 0.09972456842660904, "learning_rate": 3.8803088803088805e-05, "loss": 0.043, "step": 201 }, { "epoch": 0.9758454106280193, "grad_norm": 0.12026359885931015, "learning_rate": 3.8996138996139e-05, "loss": 0.0388, "step": 202 }, { "epoch": 0.9806763285024155, "grad_norm": 0.12035425007343292, "learning_rate": 3.918918918918919e-05, "loss": 0.0609, "step": 203 }, { "epoch": 0.9855072463768116, "grad_norm": 0.11870287358760834, "learning_rate": 3.938223938223938e-05, "loss": 0.0478, "step": 204 }, { "epoch": 0.9903381642512077, "grad_norm": 0.13712666928768158, "learning_rate": 3.9575289575289576e-05, "loss": 0.0497, "step": 205 }, { "epoch": 0.9951690821256038, "grad_norm": 0.1284405142068863, "learning_rate": 3.976833976833977e-05, "loss": 0.0457, "step": 206 }, { "epoch": 1.0, "grad_norm": 0.16791677474975586, "learning_rate": 3.996138996138996e-05, "loss": 0.0323, "step": 207 }, { "epoch": 1.0048309178743962, "grad_norm": 0.15935100615024567, "learning_rate": 4.015444015444015e-05, "loss": 0.04, "step": 208 }, { "epoch": 1.0096618357487923, "grad_norm": 0.09461764991283417, "learning_rate": 4.034749034749035e-05, "loss": 0.0431, "step": 209 }, { "epoch": 1.0144927536231885, "grad_norm": 0.15307606756687164, "learning_rate": 4.0540540540540545e-05, "loss": 0.0393, "step": 210 }, { "epoch": 1.0193236714975846, "grad_norm": 0.1187703013420105, "learning_rate": 4.073359073359074e-05, "loss": 0.0495, "step": 211 }, { "epoch": 1.0241545893719808, "grad_norm": 0.14451833069324493, "learning_rate": 4.092664092664093e-05, "loss": 0.0375, "step": 212 }, { "epoch": 1.0289855072463767, "grad_norm": 0.11559619009494781, "learning_rate": 4.111969111969112e-05, "loss": 0.0275, "step": 213 }, { "epoch": 1.0338164251207729, "grad_norm": 0.1330314427614212, "learning_rate": 4.1312741312741316e-05, "loss": 0.0515, "step": 214 }, { "epoch": 1.038647342995169, "grad_norm": 0.10810398310422897, "learning_rate": 4.150579150579151e-05, "loss": 0.0377, "step": 215 }, { "epoch": 1.0434782608695652, "grad_norm": 0.11305640637874603, "learning_rate": 4.16988416988417e-05, "loss": 0.0362, "step": 216 }, { "epoch": 1.0483091787439613, "grad_norm": 0.10515987873077393, "learning_rate": 4.189189189189189e-05, "loss": 0.0379, "step": 217 }, { "epoch": 1.0531400966183575, "grad_norm": 0.11391875147819519, "learning_rate": 4.2084942084942086e-05, "loss": 0.0441, "step": 218 }, { "epoch": 1.0579710144927537, "grad_norm": 0.1047784611582756, "learning_rate": 4.227799227799228e-05, "loss": 0.046, "step": 219 }, { "epoch": 1.0628019323671498, "grad_norm": 0.13552096486091614, "learning_rate": 4.247104247104247e-05, "loss": 0.0638, "step": 220 }, { "epoch": 1.067632850241546, "grad_norm": 0.09388095885515213, "learning_rate": 4.2664092664092664e-05, "loss": 0.0558, "step": 221 }, { "epoch": 1.0724637681159421, "grad_norm": 0.14217819273471832, "learning_rate": 4.2857142857142856e-05, "loss": 0.0377, "step": 222 }, { "epoch": 1.077294685990338, "grad_norm": 0.0974343866109848, "learning_rate": 4.305019305019305e-05, "loss": 0.0531, "step": 223 }, { "epoch": 1.0821256038647342, "grad_norm": 0.10802186280488968, "learning_rate": 4.324324324324325e-05, "loss": 0.0441, "step": 224 }, { "epoch": 1.0869565217391304, "grad_norm": 0.11783885210752487, "learning_rate": 4.343629343629344e-05, "loss": 0.0441, "step": 225 }, { "epoch": 1.0869565217391304, "eval_loss": 0.042829480022192, "eval_runtime": 20.6021, "eval_samples_per_second": 4.854, "eval_steps_per_second": 0.146, "step": 225 }, { "epoch": 1.0917874396135265, "grad_norm": 0.09211745113134384, "learning_rate": 4.3629343629343633e-05, "loss": 0.0354, "step": 226 }, { "epoch": 1.0966183574879227, "grad_norm": 0.10802285373210907, "learning_rate": 4.3822393822393826e-05, "loss": 0.0463, "step": 227 }, { "epoch": 1.1014492753623188, "grad_norm": 0.11975615471601486, "learning_rate": 4.401544401544402e-05, "loss": 0.0358, "step": 228 }, { "epoch": 1.106280193236715, "grad_norm": 0.10383973270654678, "learning_rate": 4.420849420849421e-05, "loss": 0.0513, "step": 229 }, { "epoch": 1.1111111111111112, "grad_norm": 0.09274217486381531, "learning_rate": 4.4401544401544404e-05, "loss": 0.0391, "step": 230 }, { "epoch": 1.1159420289855073, "grad_norm": 0.13271774351596832, "learning_rate": 4.4594594594594596e-05, "loss": 0.0373, "step": 231 }, { "epoch": 1.1207729468599035, "grad_norm": 0.12627987563610077, "learning_rate": 4.478764478764479e-05, "loss": 0.0436, "step": 232 }, { "epoch": 1.1256038647342996, "grad_norm": 0.1234574168920517, "learning_rate": 4.498069498069498e-05, "loss": 0.0389, "step": 233 }, { "epoch": 1.1304347826086956, "grad_norm": 0.1091989129781723, "learning_rate": 4.5173745173745174e-05, "loss": 0.0409, "step": 234 }, { "epoch": 1.1352657004830917, "grad_norm": 0.10395840555429459, "learning_rate": 4.536679536679537e-05, "loss": 0.0357, "step": 235 }, { "epoch": 1.1400966183574879, "grad_norm": 0.10336270928382874, "learning_rate": 4.555984555984556e-05, "loss": 0.0431, "step": 236 }, { "epoch": 1.144927536231884, "grad_norm": 0.09441915899515152, "learning_rate": 4.575289575289575e-05, "loss": 0.0346, "step": 237 }, { "epoch": 1.1497584541062802, "grad_norm": 0.11620728671550751, "learning_rate": 4.594594594594595e-05, "loss": 0.0436, "step": 238 }, { "epoch": 1.1545893719806763, "grad_norm": 0.09966114163398743, "learning_rate": 4.6138996138996144e-05, "loss": 0.0377, "step": 239 }, { "epoch": 1.1594202898550725, "grad_norm": 0.07768991589546204, "learning_rate": 4.6332046332046336e-05, "loss": 0.0328, "step": 240 }, { "epoch": 1.1642512077294687, "grad_norm": 0.09339828789234161, "learning_rate": 4.652509652509653e-05, "loss": 0.0363, "step": 241 }, { "epoch": 1.1690821256038648, "grad_norm": 0.11105544120073318, "learning_rate": 4.671814671814672e-05, "loss": 0.0392, "step": 242 }, { "epoch": 1.1739130434782608, "grad_norm": 0.09607312828302383, "learning_rate": 4.6911196911196914e-05, "loss": 0.0404, "step": 243 }, { "epoch": 1.178743961352657, "grad_norm": 0.1227753683924675, "learning_rate": 4.710424710424711e-05, "loss": 0.035, "step": 244 }, { "epoch": 1.183574879227053, "grad_norm": 0.09502430260181427, "learning_rate": 4.72972972972973e-05, "loss": 0.0381, "step": 245 }, { "epoch": 1.1884057971014492, "grad_norm": 0.10500189661979675, "learning_rate": 4.749034749034749e-05, "loss": 0.0441, "step": 246 }, { "epoch": 1.1932367149758454, "grad_norm": 0.10394671559333801, "learning_rate": 4.7683397683397685e-05, "loss": 0.0434, "step": 247 }, { "epoch": 1.1980676328502415, "grad_norm": 0.14034156501293182, "learning_rate": 4.787644787644788e-05, "loss": 0.067, "step": 248 }, { "epoch": 1.2028985507246377, "grad_norm": 0.11108700186014175, "learning_rate": 4.806949806949807e-05, "loss": 0.0399, "step": 249 }, { "epoch": 1.2077294685990339, "grad_norm": 0.09620919078588486, "learning_rate": 4.826254826254826e-05, "loss": 0.0348, "step": 250 }, { "epoch": 1.2077294685990339, "eval_loss": 0.040457550436258316, "eval_runtime": 20.5825, "eval_samples_per_second": 4.859, "eval_steps_per_second": 0.146, "step": 250 }, { "epoch": 1.21256038647343, "grad_norm": 0.09701947867870331, "learning_rate": 4.8455598455598455e-05, "loss": 0.0364, "step": 251 }, { "epoch": 1.2173913043478262, "grad_norm": 0.09098837524652481, "learning_rate": 4.8648648648648654e-05, "loss": 0.0315, "step": 252 }, { "epoch": 1.2222222222222223, "grad_norm": 0.11047874391078949, "learning_rate": 4.884169884169885e-05, "loss": 0.0293, "step": 253 }, { "epoch": 1.2270531400966185, "grad_norm": 0.12440991401672363, "learning_rate": 4.903474903474904e-05, "loss": 0.0399, "step": 254 }, { "epoch": 1.2318840579710144, "grad_norm": 0.11009038239717484, "learning_rate": 4.922779922779923e-05, "loss": 0.0435, "step": 255 }, { "epoch": 1.2367149758454106, "grad_norm": 0.09897538274526596, "learning_rate": 4.9420849420849425e-05, "loss": 0.0402, "step": 256 }, { "epoch": 1.2415458937198067, "grad_norm": 0.10943715274333954, "learning_rate": 4.961389961389962e-05, "loss": 0.0365, "step": 257 }, { "epoch": 1.2463768115942029, "grad_norm": 0.11430627107620239, "learning_rate": 4.980694980694981e-05, "loss": 0.0395, "step": 258 }, { "epoch": 1.251207729468599, "grad_norm": 0.1264277547597885, "learning_rate": 5e-05, "loss": 0.0361, "step": 259 }, { "epoch": 1.2560386473429952, "grad_norm": 0.10544534027576447, "learning_rate": 5.0193050193050195e-05, "loss": 0.0316, "step": 260 }, { "epoch": 1.2608695652173914, "grad_norm": 0.08700928092002869, "learning_rate": 5.038610038610039e-05, "loss": 0.0335, "step": 261 }, { "epoch": 1.2657004830917875, "grad_norm": 0.10566674917936325, "learning_rate": 5.057915057915058e-05, "loss": 0.0362, "step": 262 }, { "epoch": 1.2705314009661834, "grad_norm": 0.10313791781663895, "learning_rate": 5.077220077220077e-05, "loss": 0.0457, "step": 263 }, { "epoch": 1.2753623188405796, "grad_norm": 0.10589325428009033, "learning_rate": 5.0965250965250965e-05, "loss": 0.033, "step": 264 }, { "epoch": 1.2801932367149758, "grad_norm": 0.1091860756278038, "learning_rate": 5.115830115830116e-05, "loss": 0.0408, "step": 265 }, { "epoch": 1.285024154589372, "grad_norm": 0.11298917233943939, "learning_rate": 5.135135135135135e-05, "loss": 0.0354, "step": 266 }, { "epoch": 1.289855072463768, "grad_norm": 0.09899977594614029, "learning_rate": 5.154440154440154e-05, "loss": 0.0441, "step": 267 }, { "epoch": 1.2946859903381642, "grad_norm": 0.09906386584043503, "learning_rate": 5.1737451737451736e-05, "loss": 0.043, "step": 268 }, { "epoch": 1.2995169082125604, "grad_norm": 0.09934026002883911, "learning_rate": 5.193050193050193e-05, "loss": 0.0307, "step": 269 }, { "epoch": 1.3043478260869565, "grad_norm": 0.09088680893182755, "learning_rate": 5.212355212355212e-05, "loss": 0.036, "step": 270 }, { "epoch": 1.3091787439613527, "grad_norm": 0.10093280673027039, "learning_rate": 5.2316602316602313e-05, "loss": 0.0361, "step": 271 }, { "epoch": 1.3140096618357489, "grad_norm": 0.12925419211387634, "learning_rate": 5.2509652509652506e-05, "loss": 0.0415, "step": 272 }, { "epoch": 1.318840579710145, "grad_norm": 0.10835471749305725, "learning_rate": 5.27027027027027e-05, "loss": 0.0412, "step": 273 }, { "epoch": 1.3236714975845412, "grad_norm": 0.12303876876831055, "learning_rate": 5.2895752895752905e-05, "loss": 0.0489, "step": 274 }, { "epoch": 1.3285024154589373, "grad_norm": 0.09909823536872864, "learning_rate": 5.30888030888031e-05, "loss": 0.0355, "step": 275 }, { "epoch": 1.3285024154589373, "eval_loss": 0.03857467696070671, "eval_runtime": 20.5983, "eval_samples_per_second": 4.855, "eval_steps_per_second": 0.146, "step": 275 }, { "epoch": 1.3333333333333333, "grad_norm": 0.12137026339769363, "learning_rate": 5.328185328185329e-05, "loss": 0.0376, "step": 276 }, { "epoch": 1.3381642512077294, "grad_norm": 0.10434843599796295, "learning_rate": 5.347490347490348e-05, "loss": 0.0453, "step": 277 }, { "epoch": 1.3429951690821256, "grad_norm": 0.07658998668193817, "learning_rate": 5.3667953667953675e-05, "loss": 0.0256, "step": 278 }, { "epoch": 1.3478260869565217, "grad_norm": 0.09582667052745819, "learning_rate": 5.386100386100387e-05, "loss": 0.0308, "step": 279 }, { "epoch": 1.3526570048309179, "grad_norm": 0.09777297079563141, "learning_rate": 5.405405405405406e-05, "loss": 0.0343, "step": 280 }, { "epoch": 1.357487922705314, "grad_norm": 0.09060278534889221, "learning_rate": 5.424710424710425e-05, "loss": 0.0385, "step": 281 }, { "epoch": 1.3623188405797102, "grad_norm": 0.09172627329826355, "learning_rate": 5.4440154440154445e-05, "loss": 0.037, "step": 282 }, { "epoch": 1.3671497584541064, "grad_norm": 0.10294190049171448, "learning_rate": 5.463320463320464e-05, "loss": 0.0386, "step": 283 }, { "epoch": 1.3719806763285023, "grad_norm": 0.10812503099441528, "learning_rate": 5.482625482625483e-05, "loss": 0.045, "step": 284 }, { "epoch": 1.3768115942028984, "grad_norm": 0.07604746520519257, "learning_rate": 5.501930501930502e-05, "loss": 0.0308, "step": 285 }, { "epoch": 1.3816425120772946, "grad_norm": 0.10711754858493805, "learning_rate": 5.5212355212355216e-05, "loss": 0.039, "step": 286 }, { "epoch": 1.3864734299516908, "grad_norm": 0.14930835366249084, "learning_rate": 5.540540540540541e-05, "loss": 0.0309, "step": 287 }, { "epoch": 1.391304347826087, "grad_norm": 0.10966020077466965, "learning_rate": 5.55984555984556e-05, "loss": 0.0384, "step": 288 }, { "epoch": 1.396135265700483, "grad_norm": 0.10118502378463745, "learning_rate": 5.5791505791505794e-05, "loss": 0.0387, "step": 289 }, { "epoch": 1.4009661835748792, "grad_norm": 0.10067062824964523, "learning_rate": 5.5984555984555986e-05, "loss": 0.0401, "step": 290 }, { "epoch": 1.4057971014492754, "grad_norm": 0.08439848572015762, "learning_rate": 5.617760617760618e-05, "loss": 0.0307, "step": 291 }, { "epoch": 1.4106280193236715, "grad_norm": 0.09646830707788467, "learning_rate": 5.637065637065637e-05, "loss": 0.0391, "step": 292 }, { "epoch": 1.4154589371980677, "grad_norm": 0.09401887655258179, "learning_rate": 5.6563706563706564e-05, "loss": 0.0384, "step": 293 }, { "epoch": 1.4202898550724639, "grad_norm": 0.10123006999492645, "learning_rate": 5.6756756756756757e-05, "loss": 0.038, "step": 294 }, { "epoch": 1.42512077294686, "grad_norm": 0.09332787990570068, "learning_rate": 5.694980694980695e-05, "loss": 0.0349, "step": 295 }, { "epoch": 1.4299516908212562, "grad_norm": 0.07685629278421402, "learning_rate": 5.714285714285714e-05, "loss": 0.0322, "step": 296 }, { "epoch": 1.434782608695652, "grad_norm": 0.10600823909044266, "learning_rate": 5.7335907335907334e-05, "loss": 0.0452, "step": 297 }, { "epoch": 1.4396135265700483, "grad_norm": 0.12247564643621445, "learning_rate": 5.752895752895753e-05, "loss": 0.0398, "step": 298 }, { "epoch": 1.4444444444444444, "grad_norm": 0.09121886640787125, "learning_rate": 5.772200772200772e-05, "loss": 0.0388, "step": 299 }, { "epoch": 1.4492753623188406, "grad_norm": 0.13584840297698975, "learning_rate": 5.791505791505791e-05, "loss": 0.0379, "step": 300 }, { "epoch": 1.4492753623188406, "eval_loss": 0.03581848368048668, "eval_runtime": 20.6035, "eval_samples_per_second": 4.854, "eval_steps_per_second": 0.146, "step": 300 }, { "epoch": 1.4541062801932367, "grad_norm": 0.09803396463394165, "learning_rate": 5.8108108108108105e-05, "loss": 0.0353, "step": 301 }, { "epoch": 1.458937198067633, "grad_norm": 0.10005538165569305, "learning_rate": 5.83011583011583e-05, "loss": 0.0377, "step": 302 }, { "epoch": 1.463768115942029, "grad_norm": 0.08617591857910156, "learning_rate": 5.8494208494208503e-05, "loss": 0.0277, "step": 303 }, { "epoch": 1.4685990338164252, "grad_norm": 0.10733260959386826, "learning_rate": 5.8687258687258696e-05, "loss": 0.0413, "step": 304 }, { "epoch": 1.4734299516908211, "grad_norm": 0.08133935183286667, "learning_rate": 5.888030888030889e-05, "loss": 0.0386, "step": 305 }, { "epoch": 1.4782608695652173, "grad_norm": 0.10636088252067566, "learning_rate": 5.907335907335908e-05, "loss": 0.0358, "step": 306 }, { "epoch": 1.4830917874396135, "grad_norm": 0.10020960122346878, "learning_rate": 5.9266409266409274e-05, "loss": 0.0301, "step": 307 }, { "epoch": 1.4879227053140096, "grad_norm": 0.08966588973999023, "learning_rate": 5.9459459459459466e-05, "loss": 0.0418, "step": 308 }, { "epoch": 1.4927536231884058, "grad_norm": 0.12198576331138611, "learning_rate": 5.965250965250966e-05, "loss": 0.0417, "step": 309 }, { "epoch": 1.497584541062802, "grad_norm": 0.11839515715837479, "learning_rate": 5.984555984555985e-05, "loss": 0.0402, "step": 310 }, { "epoch": 1.502415458937198, "grad_norm": 0.10999340564012527, "learning_rate": 6.0038610038610044e-05, "loss": 0.048, "step": 311 }, { "epoch": 1.5072463768115942, "grad_norm": 0.10006649047136307, "learning_rate": 6.023166023166024e-05, "loss": 0.0349, "step": 312 }, { "epoch": 1.5120772946859904, "grad_norm": 0.10847444832324982, "learning_rate": 6.042471042471043e-05, "loss": 0.0378, "step": 313 }, { "epoch": 1.5169082125603865, "grad_norm": 0.08905550837516785, "learning_rate": 6.061776061776062e-05, "loss": 0.0338, "step": 314 }, { "epoch": 1.5217391304347827, "grad_norm": 0.1286388784646988, "learning_rate": 6.0810810810810814e-05, "loss": 0.0286, "step": 315 }, { "epoch": 1.5265700483091789, "grad_norm": 0.0705726221203804, "learning_rate": 6.100386100386101e-05, "loss": 0.0254, "step": 316 }, { "epoch": 1.531400966183575, "grad_norm": 0.08916336297988892, "learning_rate": 6.11969111969112e-05, "loss": 0.0344, "step": 317 }, { "epoch": 1.5362318840579712, "grad_norm": 0.08876073360443115, "learning_rate": 6.138996138996139e-05, "loss": 0.0313, "step": 318 }, { "epoch": 1.541062801932367, "grad_norm": 0.07882793992757797, "learning_rate": 6.158301158301159e-05, "loss": 0.0475, "step": 319 }, { "epoch": 1.5458937198067633, "grad_norm": 0.09577280282974243, "learning_rate": 6.177606177606178e-05, "loss": 0.0373, "step": 320 }, { "epoch": 1.5507246376811594, "grad_norm": 0.08283651620149612, "learning_rate": 6.196911196911198e-05, "loss": 0.0398, "step": 321 }, { "epoch": 1.5555555555555556, "grad_norm": 0.08609731495380402, "learning_rate": 6.216216216216216e-05, "loss": 0.0333, "step": 322 }, { "epoch": 1.5603864734299517, "grad_norm": 0.1029728502035141, "learning_rate": 6.235521235521236e-05, "loss": 0.0378, "step": 323 }, { "epoch": 1.5652173913043477, "grad_norm": 0.08994361758232117, "learning_rate": 6.254826254826255e-05, "loss": 0.0385, "step": 324 }, { "epoch": 1.5700483091787438, "grad_norm": 0.09228026866912842, "learning_rate": 6.274131274131275e-05, "loss": 0.032, "step": 325 }, { "epoch": 1.5700483091787438, "eval_loss": 0.03536108508706093, "eval_runtime": 20.5807, "eval_samples_per_second": 4.859, "eval_steps_per_second": 0.146, "step": 325 }, { "epoch": 1.57487922705314, "grad_norm": 0.10328676551580429, "learning_rate": 6.293436293436293e-05, "loss": 0.0358, "step": 326 }, { "epoch": 1.5797101449275361, "grad_norm": 0.09661010652780533, "learning_rate": 6.312741312741313e-05, "loss": 0.0346, "step": 327 }, { "epoch": 1.5845410628019323, "grad_norm": 0.09876400232315063, "learning_rate": 6.332046332046332e-05, "loss": 0.0348, "step": 328 }, { "epoch": 1.5893719806763285, "grad_norm": 0.10440344363451004, "learning_rate": 6.351351351351352e-05, "loss": 0.0345, "step": 329 }, { "epoch": 1.5942028985507246, "grad_norm": 0.09511513262987137, "learning_rate": 6.37065637065637e-05, "loss": 0.037, "step": 330 }, { "epoch": 1.5990338164251208, "grad_norm": 0.1074519231915474, "learning_rate": 6.38996138996139e-05, "loss": 0.0469, "step": 331 }, { "epoch": 1.603864734299517, "grad_norm": 0.10466543585062027, "learning_rate": 6.40926640926641e-05, "loss": 0.0305, "step": 332 }, { "epoch": 1.608695652173913, "grad_norm": 0.08107554912567139, "learning_rate": 6.428571428571429e-05, "loss": 0.0345, "step": 333 }, { "epoch": 1.6135265700483092, "grad_norm": 0.08928783237934113, "learning_rate": 6.447876447876449e-05, "loss": 0.0332, "step": 334 }, { "epoch": 1.6183574879227054, "grad_norm": 0.10237609595060349, "learning_rate": 6.467181467181467e-05, "loss": 0.0372, "step": 335 }, { "epoch": 1.6231884057971016, "grad_norm": 0.08093827217817307, "learning_rate": 6.486486486486487e-05, "loss": 0.0365, "step": 336 }, { "epoch": 1.6280193236714977, "grad_norm": 0.08602860569953918, "learning_rate": 6.505791505791506e-05, "loss": 0.035, "step": 337 }, { "epoch": 1.6328502415458939, "grad_norm": 0.09471640735864639, "learning_rate": 6.525096525096526e-05, "loss": 0.0422, "step": 338 }, { "epoch": 1.6376811594202898, "grad_norm": 0.10974764078855515, "learning_rate": 6.544401544401544e-05, "loss": 0.0369, "step": 339 }, { "epoch": 1.642512077294686, "grad_norm": 0.08305064588785172, "learning_rate": 6.563706563706564e-05, "loss": 0.0508, "step": 340 }, { "epoch": 1.6473429951690821, "grad_norm": 0.09225677698850632, "learning_rate": 6.583011583011583e-05, "loss": 0.0392, "step": 341 }, { "epoch": 1.6521739130434783, "grad_norm": 0.10002700239419937, "learning_rate": 6.602316602316603e-05, "loss": 0.0376, "step": 342 }, { "epoch": 1.6570048309178744, "grad_norm": 0.064641572535038, "learning_rate": 6.621621621621621e-05, "loss": 0.0284, "step": 343 }, { "epoch": 1.6618357487922706, "grad_norm": 0.08822232484817505, "learning_rate": 6.640926640926641e-05, "loss": 0.0411, "step": 344 }, { "epoch": 1.6666666666666665, "grad_norm": 0.07955625653266907, "learning_rate": 6.66023166023166e-05, "loss": 0.0286, "step": 345 }, { "epoch": 1.6714975845410627, "grad_norm": 0.08300414681434631, "learning_rate": 6.67953667953668e-05, "loss": 0.0248, "step": 346 }, { "epoch": 1.6763285024154588, "grad_norm": 0.08104908466339111, "learning_rate": 6.6988416988417e-05, "loss": 0.0342, "step": 347 }, { "epoch": 1.681159420289855, "grad_norm": 0.0685056522488594, "learning_rate": 6.718146718146718e-05, "loss": 0.0277, "step": 348 }, { "epoch": 1.6859903381642511, "grad_norm": 0.09631809592247009, "learning_rate": 6.737451737451738e-05, "loss": 0.0342, "step": 349 }, { "epoch": 1.6908212560386473, "grad_norm": 0.09193519502878189, "learning_rate": 6.756756756756757e-05, "loss": 0.0342, "step": 350 }, { "epoch": 1.6908212560386473, "eval_loss": 0.033477578312158585, "eval_runtime": 20.5968, "eval_samples_per_second": 4.855, "eval_steps_per_second": 0.146, "step": 350 }, { "epoch": 1.6956521739130435, "grad_norm": 0.08512197434902191, "learning_rate": 6.776061776061777e-05, "loss": 0.0344, "step": 351 }, { "epoch": 1.7004830917874396, "grad_norm": 0.11075153201818466, "learning_rate": 6.795366795366795e-05, "loss": 0.041, "step": 352 }, { "epoch": 1.7053140096618358, "grad_norm": 0.0982983410358429, "learning_rate": 6.814671814671815e-05, "loss": 0.0319, "step": 353 }, { "epoch": 1.710144927536232, "grad_norm": 0.0894618108868599, "learning_rate": 6.833976833976834e-05, "loss": 0.0422, "step": 354 }, { "epoch": 1.714975845410628, "grad_norm": 0.11245858669281006, "learning_rate": 6.853281853281854e-05, "loss": 0.0443, "step": 355 }, { "epoch": 1.7198067632850242, "grad_norm": 0.0929783433675766, "learning_rate": 6.872586872586872e-05, "loss": 0.0341, "step": 356 }, { "epoch": 1.7246376811594204, "grad_norm": 0.07707154005765915, "learning_rate": 6.891891891891892e-05, "loss": 0.0301, "step": 357 }, { "epoch": 1.7294685990338166, "grad_norm": 0.08752428740262985, "learning_rate": 6.911196911196911e-05, "loss": 0.0341, "step": 358 }, { "epoch": 1.7342995169082127, "grad_norm": 0.11009437590837479, "learning_rate": 6.930501930501931e-05, "loss": 0.0393, "step": 359 }, { "epoch": 1.7391304347826086, "grad_norm": 0.09310301393270493, "learning_rate": 6.949806949806951e-05, "loss": 0.0334, "step": 360 }, { "epoch": 1.7439613526570048, "grad_norm": 0.10008851438760757, "learning_rate": 6.96911196911197e-05, "loss": 0.0332, "step": 361 }, { "epoch": 1.748792270531401, "grad_norm": 0.09315995872020721, "learning_rate": 6.988416988416989e-05, "loss": 0.043, "step": 362 }, { "epoch": 1.7536231884057971, "grad_norm": 0.09456358104944229, "learning_rate": 7.007722007722008e-05, "loss": 0.0498, "step": 363 }, { "epoch": 1.7584541062801933, "grad_norm": 0.07672219723463058, "learning_rate": 7.027027027027028e-05, "loss": 0.0352, "step": 364 }, { "epoch": 1.7632850241545892, "grad_norm": 0.07560185343027115, "learning_rate": 7.046332046332046e-05, "loss": 0.0346, "step": 365 }, { "epoch": 1.7681159420289854, "grad_norm": 0.07019049674272537, "learning_rate": 7.065637065637066e-05, "loss": 0.0322, "step": 366 }, { "epoch": 1.7729468599033815, "grad_norm": 0.0801636353135109, "learning_rate": 7.084942084942085e-05, "loss": 0.0277, "step": 367 }, { "epoch": 1.7777777777777777, "grad_norm": 0.08249375224113464, "learning_rate": 7.104247104247105e-05, "loss": 0.0523, "step": 368 }, { "epoch": 1.7826086956521738, "grad_norm": 0.08103843778371811, "learning_rate": 7.123552123552123e-05, "loss": 0.0343, "step": 369 }, { "epoch": 1.78743961352657, "grad_norm": 0.08398215472698212, "learning_rate": 7.142857142857143e-05, "loss": 0.032, "step": 370 }, { "epoch": 1.7922705314009661, "grad_norm": 0.07273257523775101, "learning_rate": 7.162162162162162e-05, "loss": 0.0315, "step": 371 }, { "epoch": 1.7971014492753623, "grad_norm": 0.0843791663646698, "learning_rate": 7.181467181467182e-05, "loss": 0.0343, "step": 372 }, { "epoch": 1.8019323671497585, "grad_norm": 0.06816914677619934, "learning_rate": 7.2007722007722e-05, "loss": 0.0323, "step": 373 }, { "epoch": 1.8067632850241546, "grad_norm": 0.0995233952999115, "learning_rate": 7.22007722007722e-05, "loss": 0.0252, "step": 374 }, { "epoch": 1.8115942028985508, "grad_norm": 0.07620427757501602, "learning_rate": 7.23938223938224e-05, "loss": 0.0318, "step": 375 }, { "epoch": 1.8115942028985508, "eval_loss": 0.03241335600614548, "eval_runtime": 20.5896, "eval_samples_per_second": 4.857, "eval_steps_per_second": 0.146, "step": 375 }, { "epoch": 1.816425120772947, "grad_norm": 0.08072451502084732, "learning_rate": 7.258687258687259e-05, "loss": 0.0306, "step": 376 }, { "epoch": 1.821256038647343, "grad_norm": 0.08463121950626373, "learning_rate": 7.277992277992279e-05, "loss": 0.0246, "step": 377 }, { "epoch": 1.8260869565217392, "grad_norm": 0.08696448057889938, "learning_rate": 7.297297297297297e-05, "loss": 0.0391, "step": 378 }, { "epoch": 1.8309178743961354, "grad_norm": 0.08394154906272888, "learning_rate": 7.316602316602317e-05, "loss": 0.0503, "step": 379 }, { "epoch": 1.8357487922705316, "grad_norm": 0.09307282418012619, "learning_rate": 7.335907335907336e-05, "loss": 0.0338, "step": 380 }, { "epoch": 1.8405797101449275, "grad_norm": 0.08928454667329788, "learning_rate": 7.355212355212356e-05, "loss": 0.0414, "step": 381 }, { "epoch": 1.8454106280193237, "grad_norm": 0.06912971287965775, "learning_rate": 7.374517374517374e-05, "loss": 0.0299, "step": 382 }, { "epoch": 1.8502415458937198, "grad_norm": 0.07190634310245514, "learning_rate": 7.393822393822394e-05, "loss": 0.0293, "step": 383 }, { "epoch": 1.855072463768116, "grad_norm": 0.08201923221349716, "learning_rate": 7.413127413127413e-05, "loss": 0.0378, "step": 384 }, { "epoch": 1.8599033816425121, "grad_norm": 0.09282411634922028, "learning_rate": 7.432432432432433e-05, "loss": 0.0318, "step": 385 }, { "epoch": 1.864734299516908, "grad_norm": 0.09017174690961838, "learning_rate": 7.451737451737452e-05, "loss": 0.0373, "step": 386 }, { "epoch": 1.8695652173913042, "grad_norm": 0.09520550072193146, "learning_rate": 7.471042471042471e-05, "loss": 0.0374, "step": 387 }, { "epoch": 1.8743961352657004, "grad_norm": 0.08649712800979614, "learning_rate": 7.49034749034749e-05, "loss": 0.038, "step": 388 }, { "epoch": 1.8792270531400965, "grad_norm": 0.08308789879083633, "learning_rate": 7.50965250965251e-05, "loss": 0.0318, "step": 389 }, { "epoch": 1.8840579710144927, "grad_norm": 0.07252487540245056, "learning_rate": 7.52895752895753e-05, "loss": 0.0266, "step": 390 }, { "epoch": 1.8888888888888888, "grad_norm": 0.07101837545633316, "learning_rate": 7.548262548262549e-05, "loss": 0.0334, "step": 391 }, { "epoch": 1.893719806763285, "grad_norm": 0.07671766728162766, "learning_rate": 7.567567567567568e-05, "loss": 0.037, "step": 392 }, { "epoch": 1.8985507246376812, "grad_norm": 0.0921880453824997, "learning_rate": 7.586872586872587e-05, "loss": 0.0349, "step": 393 }, { "epoch": 1.9033816425120773, "grad_norm": 0.1014850065112114, "learning_rate": 7.606177606177607e-05, "loss": 0.0505, "step": 394 }, { "epoch": 1.9082125603864735, "grad_norm": 0.07078011333942413, "learning_rate": 7.625482625482626e-05, "loss": 0.0276, "step": 395 }, { "epoch": 1.9130434782608696, "grad_norm": 0.07615656405687332, "learning_rate": 7.644787644787645e-05, "loss": 0.0291, "step": 396 }, { "epoch": 1.9178743961352658, "grad_norm": 0.06996695697307587, "learning_rate": 7.664092664092664e-05, "loss": 0.026, "step": 397 }, { "epoch": 1.922705314009662, "grad_norm": 0.06688771396875381, "learning_rate": 7.683397683397684e-05, "loss": 0.0321, "step": 398 }, { "epoch": 1.927536231884058, "grad_norm": 0.06577587872743607, "learning_rate": 7.702702702702703e-05, "loss": 0.0288, "step": 399 }, { "epoch": 1.9323671497584543, "grad_norm": 0.07597734034061432, "learning_rate": 7.722007722007723e-05, "loss": 0.031, "step": 400 }, { "epoch": 1.9323671497584543, "eval_loss": 0.031759362667798996, "eval_runtime": 20.6062, "eval_samples_per_second": 4.853, "eval_steps_per_second": 0.146, "step": 400 }, { "epoch": 1.9371980676328504, "grad_norm": 0.08483527600765228, "learning_rate": 7.741312741312741e-05, "loss": 0.0339, "step": 401 }, { "epoch": 1.9420289855072463, "grad_norm": 0.06974739581346512, "learning_rate": 7.760617760617761e-05, "loss": 0.0313, "step": 402 }, { "epoch": 1.9468599033816425, "grad_norm": 0.093897745013237, "learning_rate": 7.779922779922781e-05, "loss": 0.0332, "step": 403 }, { "epoch": 1.9516908212560387, "grad_norm": 0.06541071832180023, "learning_rate": 7.7992277992278e-05, "loss": 0.0245, "step": 404 }, { "epoch": 1.9565217391304348, "grad_norm": 0.07860124856233597, "learning_rate": 7.81853281853282e-05, "loss": 0.0288, "step": 405 }, { "epoch": 1.961352657004831, "grad_norm": 0.09602763503789902, "learning_rate": 7.837837837837838e-05, "loss": 0.0374, "step": 406 }, { "epoch": 1.966183574879227, "grad_norm": 0.07784169167280197, "learning_rate": 7.857142857142858e-05, "loss": 0.0284, "step": 407 }, { "epoch": 1.971014492753623, "grad_norm": 0.08984600007534027, "learning_rate": 7.876447876447877e-05, "loss": 0.0325, "step": 408 }, { "epoch": 1.9758454106280192, "grad_norm": 0.06867984682321548, "learning_rate": 7.895752895752897e-05, "loss": 0.028, "step": 409 }, { "epoch": 1.9806763285024154, "grad_norm": 0.09378314763307571, "learning_rate": 7.915057915057915e-05, "loss": 0.0453, "step": 410 }, { "epoch": 1.9855072463768115, "grad_norm": 0.07811439037322998, "learning_rate": 7.934362934362935e-05, "loss": 0.0309, "step": 411 }, { "epoch": 1.9903381642512077, "grad_norm": 0.059358660131692886, "learning_rate": 7.953667953667954e-05, "loss": 0.0226, "step": 412 }, { "epoch": 1.9951690821256038, "grad_norm": 0.0751197412610054, "learning_rate": 7.972972972972974e-05, "loss": 0.0271, "step": 413 }, { "epoch": 2.0, "grad_norm": 0.14637447893619537, "learning_rate": 7.992277992277992e-05, "loss": 0.0192, "step": 414 }, { "epoch": 2.004830917874396, "grad_norm": 0.08730384707450867, "learning_rate": 8.011583011583012e-05, "loss": 0.0383, "step": 415 }, { "epoch": 2.0096618357487923, "grad_norm": 0.0848565399646759, "learning_rate": 8.03088803088803e-05, "loss": 0.0304, "step": 416 }, { "epoch": 2.0144927536231885, "grad_norm": 0.0779394656419754, "learning_rate": 8.05019305019305e-05, "loss": 0.0281, "step": 417 }, { "epoch": 2.0193236714975846, "grad_norm": 0.09085428714752197, "learning_rate": 8.06949806949807e-05, "loss": 0.0282, "step": 418 }, { "epoch": 2.024154589371981, "grad_norm": 0.07310831546783447, "learning_rate": 8.088803088803089e-05, "loss": 0.0259, "step": 419 }, { "epoch": 2.028985507246377, "grad_norm": 0.07102567702531815, "learning_rate": 8.108108108108109e-05, "loss": 0.0217, "step": 420 }, { "epoch": 2.033816425120773, "grad_norm": 0.0762273445725441, "learning_rate": 8.127413127413128e-05, "loss": 0.0287, "step": 421 }, { "epoch": 2.0386473429951693, "grad_norm": 0.0700920894742012, "learning_rate": 8.146718146718148e-05, "loss": 0.0251, "step": 422 }, { "epoch": 2.0434782608695654, "grad_norm": 0.08108256012201309, "learning_rate": 8.166023166023166e-05, "loss": 0.0274, "step": 423 }, { "epoch": 2.0483091787439616, "grad_norm": 0.07796541601419449, "learning_rate": 8.185328185328186e-05, "loss": 0.0222, "step": 424 }, { "epoch": 2.0531400966183573, "grad_norm": 0.08583205938339233, "learning_rate": 8.204633204633205e-05, "loss": 0.0283, "step": 425 }, { "epoch": 2.0531400966183573, "eval_loss": 0.03210844099521637, "eval_runtime": 21.3092, "eval_samples_per_second": 4.693, "eval_steps_per_second": 0.141, "step": 425 }, { "epoch": 2.0579710144927534, "grad_norm": 0.0870027244091034, "learning_rate": 8.223938223938225e-05, "loss": 0.0316, "step": 426 }, { "epoch": 2.0628019323671496, "grad_norm": 0.11122583597898483, "learning_rate": 8.243243243243243e-05, "loss": 0.0345, "step": 427 }, { "epoch": 2.0676328502415457, "grad_norm": 0.07087469846010208, "learning_rate": 8.262548262548263e-05, "loss": 0.0256, "step": 428 }, { "epoch": 2.072463768115942, "grad_norm": 0.07167962193489075, "learning_rate": 8.281853281853282e-05, "loss": 0.029, "step": 429 }, { "epoch": 2.077294685990338, "grad_norm": 0.0908481627702713, "learning_rate": 8.301158301158302e-05, "loss": 0.0337, "step": 430 }, { "epoch": 2.082125603864734, "grad_norm": 0.14270740747451782, "learning_rate": 8.32046332046332e-05, "loss": 0.0605, "step": 431 }, { "epoch": 2.0869565217391304, "grad_norm": 0.08899253606796265, "learning_rate": 8.33976833976834e-05, "loss": 0.0232, "step": 432 }, { "epoch": 2.0917874396135265, "grad_norm": 0.08228383958339691, "learning_rate": 8.35907335907336e-05, "loss": 0.022, "step": 433 }, { "epoch": 2.0966183574879227, "grad_norm": 0.08265294134616852, "learning_rate": 8.378378378378379e-05, "loss": 0.026, "step": 434 }, { "epoch": 2.101449275362319, "grad_norm": 0.08507855236530304, "learning_rate": 8.397683397683399e-05, "loss": 0.0286, "step": 435 }, { "epoch": 2.106280193236715, "grad_norm": 0.09821953624486923, "learning_rate": 8.416988416988417e-05, "loss": 0.0233, "step": 436 }, { "epoch": 2.111111111111111, "grad_norm": 0.09826794266700745, "learning_rate": 8.436293436293437e-05, "loss": 0.0245, "step": 437 }, { "epoch": 2.1159420289855073, "grad_norm": 0.06491506844758987, "learning_rate": 8.455598455598456e-05, "loss": 0.0238, "step": 438 }, { "epoch": 2.1207729468599035, "grad_norm": 0.09309545159339905, "learning_rate": 8.474903474903476e-05, "loss": 0.0244, "step": 439 }, { "epoch": 2.1256038647342996, "grad_norm": 0.07788459956645966, "learning_rate": 8.494208494208494e-05, "loss": 0.0316, "step": 440 }, { "epoch": 2.130434782608696, "grad_norm": 0.08980804681777954, "learning_rate": 8.513513513513514e-05, "loss": 0.035, "step": 441 }, { "epoch": 2.135265700483092, "grad_norm": 0.11862210929393768, "learning_rate": 8.532818532818533e-05, "loss": 0.0355, "step": 442 }, { "epoch": 2.140096618357488, "grad_norm": 0.09856075048446655, "learning_rate": 8.552123552123553e-05, "loss": 0.042, "step": 443 }, { "epoch": 2.1449275362318843, "grad_norm": 0.06776445358991623, "learning_rate": 8.571428571428571e-05, "loss": 0.0211, "step": 444 }, { "epoch": 2.14975845410628, "grad_norm": 0.0670533999800682, "learning_rate": 8.590733590733591e-05, "loss": 0.0216, "step": 445 }, { "epoch": 2.154589371980676, "grad_norm": 0.07002390176057816, "learning_rate": 8.61003861003861e-05, "loss": 0.0252, "step": 446 }, { "epoch": 2.1594202898550723, "grad_norm": 0.07812980562448502, "learning_rate": 8.62934362934363e-05, "loss": 0.0363, "step": 447 }, { "epoch": 2.1642512077294684, "grad_norm": 0.08270443975925446, "learning_rate": 8.64864864864865e-05, "loss": 0.0257, "step": 448 }, { "epoch": 2.1690821256038646, "grad_norm": 0.08274964988231659, "learning_rate": 8.667953667953668e-05, "loss": 0.033, "step": 449 }, { "epoch": 2.1739130434782608, "grad_norm": 0.0919279232621193, "learning_rate": 8.687258687258688e-05, "loss": 0.0275, "step": 450 }, { "epoch": 2.1739130434782608, "eval_loss": 0.03367118537425995, "eval_runtime": 20.6662, "eval_samples_per_second": 4.839, "eval_steps_per_second": 0.145, "step": 450 }, { "epoch": 2.178743961352657, "grad_norm": 0.0769813284277916, "learning_rate": 8.706563706563707e-05, "loss": 0.0219, "step": 451 }, { "epoch": 2.183574879227053, "grad_norm": 0.0803806260228157, "learning_rate": 8.725868725868727e-05, "loss": 0.0304, "step": 452 }, { "epoch": 2.1884057971014492, "grad_norm": 0.07121909409761429, "learning_rate": 8.745173745173745e-05, "loss": 0.0269, "step": 453 }, { "epoch": 2.1932367149758454, "grad_norm": 0.08506215363740921, "learning_rate": 8.764478764478765e-05, "loss": 0.0285, "step": 454 }, { "epoch": 2.1980676328502415, "grad_norm": 0.07936449348926544, "learning_rate": 8.783783783783784e-05, "loss": 0.0235, "step": 455 }, { "epoch": 2.2028985507246377, "grad_norm": 0.084862120449543, "learning_rate": 8.803088803088804e-05, "loss": 0.0307, "step": 456 }, { "epoch": 2.207729468599034, "grad_norm": 0.08003883808851242, "learning_rate": 8.822393822393822e-05, "loss": 0.025, "step": 457 }, { "epoch": 2.21256038647343, "grad_norm": 0.07105076313018799, "learning_rate": 8.841698841698842e-05, "loss": 0.0225, "step": 458 }, { "epoch": 2.217391304347826, "grad_norm": 0.07181154191493988, "learning_rate": 8.861003861003861e-05, "loss": 0.0207, "step": 459 }, { "epoch": 2.2222222222222223, "grad_norm": 0.09786058962345123, "learning_rate": 8.880308880308881e-05, "loss": 0.0314, "step": 460 }, { "epoch": 2.2270531400966185, "grad_norm": 0.08781251311302185, "learning_rate": 8.899613899613901e-05, "loss": 0.0248, "step": 461 }, { "epoch": 2.2318840579710146, "grad_norm": 0.08720511198043823, "learning_rate": 8.918918918918919e-05, "loss": 0.029, "step": 462 }, { "epoch": 2.236714975845411, "grad_norm": 0.08457627147436142, "learning_rate": 8.938223938223939e-05, "loss": 0.0286, "step": 463 }, { "epoch": 2.241545893719807, "grad_norm": 0.07502002269029617, "learning_rate": 8.957528957528958e-05, "loss": 0.0265, "step": 464 }, { "epoch": 2.246376811594203, "grad_norm": 0.06273361295461655, "learning_rate": 8.976833976833978e-05, "loss": 0.021, "step": 465 }, { "epoch": 2.2512077294685993, "grad_norm": 0.0748235285282135, "learning_rate": 8.996138996138996e-05, "loss": 0.0291, "step": 466 }, { "epoch": 2.2560386473429954, "grad_norm": 0.06298107653856277, "learning_rate": 9.015444015444016e-05, "loss": 0.02, "step": 467 }, { "epoch": 2.260869565217391, "grad_norm": 0.08730265498161316, "learning_rate": 9.034749034749035e-05, "loss": 0.0311, "step": 468 }, { "epoch": 2.2657004830917873, "grad_norm": 0.07982222735881805, "learning_rate": 9.054054054054055e-05, "loss": 0.031, "step": 469 }, { "epoch": 2.2705314009661834, "grad_norm": 0.09329807758331299, "learning_rate": 9.073359073359073e-05, "loss": 0.0309, "step": 470 }, { "epoch": 2.2753623188405796, "grad_norm": 0.08422908186912537, "learning_rate": 9.092664092664093e-05, "loss": 0.0253, "step": 471 }, { "epoch": 2.2801932367149758, "grad_norm": 0.08790479600429535, "learning_rate": 9.111969111969112e-05, "loss": 0.0236, "step": 472 }, { "epoch": 2.285024154589372, "grad_norm": 0.09036684781312943, "learning_rate": 9.131274131274132e-05, "loss": 0.0273, "step": 473 }, { "epoch": 2.289855072463768, "grad_norm": 0.08374612033367157, "learning_rate": 9.15057915057915e-05, "loss": 0.0325, "step": 474 }, { "epoch": 2.2946859903381642, "grad_norm": 0.07355117052793503, "learning_rate": 9.16988416988417e-05, "loss": 0.026, "step": 475 }, { "epoch": 2.2946859903381642, "eval_loss": 0.03135251998901367, "eval_runtime": 20.6353, "eval_samples_per_second": 4.846, "eval_steps_per_second": 0.145, "step": 475 }, { "epoch": 2.2995169082125604, "grad_norm": 0.13940849900245667, "learning_rate": 9.18918918918919e-05, "loss": 0.0299, "step": 476 }, { "epoch": 2.3043478260869565, "grad_norm": 0.09694752097129822, "learning_rate": 9.208494208494209e-05, "loss": 0.0342, "step": 477 }, { "epoch": 2.3091787439613527, "grad_norm": 0.07969164848327637, "learning_rate": 9.227799227799229e-05, "loss": 0.0265, "step": 478 }, { "epoch": 2.314009661835749, "grad_norm": 0.07230686396360397, "learning_rate": 9.247104247104247e-05, "loss": 0.0255, "step": 479 }, { "epoch": 2.318840579710145, "grad_norm": 0.08269919455051422, "learning_rate": 9.266409266409267e-05, "loss": 0.0327, "step": 480 }, { "epoch": 2.323671497584541, "grad_norm": 0.07120717316865921, "learning_rate": 9.285714285714286e-05, "loss": 0.0229, "step": 481 }, { "epoch": 2.3285024154589373, "grad_norm": 0.0936795324087143, "learning_rate": 9.305019305019306e-05, "loss": 0.0261, "step": 482 }, { "epoch": 2.3333333333333335, "grad_norm": 0.07093029469251633, "learning_rate": 9.324324324324324e-05, "loss": 0.0215, "step": 483 }, { "epoch": 2.3381642512077296, "grad_norm": 0.0930793434381485, "learning_rate": 9.343629343629344e-05, "loss": 0.0285, "step": 484 }, { "epoch": 2.342995169082126, "grad_norm": 0.05888902395963669, "learning_rate": 9.362934362934363e-05, "loss": 0.0186, "step": 485 }, { "epoch": 2.3478260869565215, "grad_norm": 0.07734473794698715, "learning_rate": 9.382239382239383e-05, "loss": 0.0326, "step": 486 }, { "epoch": 2.3526570048309177, "grad_norm": 0.10177651047706604, "learning_rate": 9.401544401544401e-05, "loss": 0.0305, "step": 487 }, { "epoch": 2.357487922705314, "grad_norm": 0.08325226604938507, "learning_rate": 9.420849420849421e-05, "loss": 0.0372, "step": 488 }, { "epoch": 2.36231884057971, "grad_norm": 0.068690724670887, "learning_rate": 9.44015444015444e-05, "loss": 0.0352, "step": 489 }, { "epoch": 2.367149758454106, "grad_norm": 0.08754605799913406, "learning_rate": 9.45945945945946e-05, "loss": 0.0269, "step": 490 }, { "epoch": 2.3719806763285023, "grad_norm": 0.06412243843078613, "learning_rate": 9.47876447876448e-05, "loss": 0.0248, "step": 491 }, { "epoch": 2.3768115942028984, "grad_norm": 0.07275252044200897, "learning_rate": 9.498069498069498e-05, "loss": 0.0279, "step": 492 }, { "epoch": 2.3816425120772946, "grad_norm": 0.07856742292642593, "learning_rate": 9.517374517374518e-05, "loss": 0.0386, "step": 493 }, { "epoch": 2.3864734299516908, "grad_norm": 0.08813096582889557, "learning_rate": 9.536679536679537e-05, "loss": 0.0283, "step": 494 }, { "epoch": 2.391304347826087, "grad_norm": 0.07463379204273224, "learning_rate": 9.555984555984557e-05, "loss": 0.035, "step": 495 }, { "epoch": 2.396135265700483, "grad_norm": 0.0677594393491745, "learning_rate": 9.575289575289575e-05, "loss": 0.0245, "step": 496 }, { "epoch": 2.4009661835748792, "grad_norm": 0.0777975544333458, "learning_rate": 9.594594594594595e-05, "loss": 0.0222, "step": 497 }, { "epoch": 2.4057971014492754, "grad_norm": 0.0731094554066658, "learning_rate": 9.613899613899614e-05, "loss": 0.0259, "step": 498 }, { "epoch": 2.4106280193236715, "grad_norm": 0.07815240323543549, "learning_rate": 9.633204633204634e-05, "loss": 0.0246, "step": 499 }, { "epoch": 2.4154589371980677, "grad_norm": 0.07070033997297287, "learning_rate": 9.652509652509652e-05, "loss": 0.0244, "step": 500 }, { "epoch": 2.4154589371980677, "eval_loss": 0.02849414572119713, "eval_runtime": 20.5961, "eval_samples_per_second": 4.855, "eval_steps_per_second": 0.146, "step": 500 }, { "epoch": 2.420289855072464, "grad_norm": 0.07829859107732773, "learning_rate": 9.671814671814672e-05, "loss": 0.0208, "step": 501 }, { "epoch": 2.42512077294686, "grad_norm": 0.10270782560110092, "learning_rate": 9.691119691119691e-05, "loss": 0.0294, "step": 502 }, { "epoch": 2.429951690821256, "grad_norm": 0.08430244773626328, "learning_rate": 9.710424710424711e-05, "loss": 0.0221, "step": 503 }, { "epoch": 2.4347826086956523, "grad_norm": 0.07175758481025696, "learning_rate": 9.729729729729731e-05, "loss": 0.0251, "step": 504 }, { "epoch": 2.4396135265700485, "grad_norm": 0.06120215356349945, "learning_rate": 9.74903474903475e-05, "loss": 0.0208, "step": 505 }, { "epoch": 2.4444444444444446, "grad_norm": 0.07098649442195892, "learning_rate": 9.76833976833977e-05, "loss": 0.0225, "step": 506 }, { "epoch": 2.449275362318841, "grad_norm": 0.07696057856082916, "learning_rate": 9.787644787644788e-05, "loss": 0.028, "step": 507 }, { "epoch": 2.454106280193237, "grad_norm": 0.09887323528528214, "learning_rate": 9.806949806949808e-05, "loss": 0.0377, "step": 508 }, { "epoch": 2.4589371980676327, "grad_norm": 0.06533748656511307, "learning_rate": 9.826254826254826e-05, "loss": 0.0252, "step": 509 }, { "epoch": 2.463768115942029, "grad_norm": 0.06457691639661789, "learning_rate": 9.845559845559846e-05, "loss": 0.0316, "step": 510 }, { "epoch": 2.468599033816425, "grad_norm": 0.07615062594413757, "learning_rate": 9.864864864864865e-05, "loss": 0.0262, "step": 511 }, { "epoch": 2.473429951690821, "grad_norm": 0.05695481225848198, "learning_rate": 9.884169884169885e-05, "loss": 0.0232, "step": 512 }, { "epoch": 2.4782608695652173, "grad_norm": 0.06162761151790619, "learning_rate": 9.903474903474904e-05, "loss": 0.0279, "step": 513 }, { "epoch": 2.4830917874396135, "grad_norm": 0.07245969772338867, "learning_rate": 9.922779922779923e-05, "loss": 0.0242, "step": 514 }, { "epoch": 2.4879227053140096, "grad_norm": 0.06011466681957245, "learning_rate": 9.942084942084942e-05, "loss": 0.0195, "step": 515 }, { "epoch": 2.4927536231884058, "grad_norm": 0.07303909212350845, "learning_rate": 9.961389961389962e-05, "loss": 0.0221, "step": 516 }, { "epoch": 2.497584541062802, "grad_norm": 0.05043727904558182, "learning_rate": 9.98069498069498e-05, "loss": 0.0134, "step": 517 }, { "epoch": 2.502415458937198, "grad_norm": 0.06988222897052765, "learning_rate": 0.0001, "loss": 0.0215, "step": 518 }, { "epoch": 2.5072463768115942, "grad_norm": 0.0786987692117691, "learning_rate": 9.9999988623013e-05, "loss": 0.0276, "step": 519 }, { "epoch": 2.5120772946859904, "grad_norm": 0.11683060973882675, "learning_rate": 9.999995449205719e-05, "loss": 0.0261, "step": 520 }, { "epoch": 2.5169082125603865, "grad_norm": 0.08037126064300537, "learning_rate": 9.999989760714809e-05, "loss": 0.0255, "step": 521 }, { "epoch": 2.5217391304347827, "grad_norm": 0.05731765180826187, "learning_rate": 9.999981796831159e-05, "loss": 0.0201, "step": 522 }, { "epoch": 2.526570048309179, "grad_norm": 0.07567796856164932, "learning_rate": 9.999971557558395e-05, "loss": 0.0262, "step": 523 }, { "epoch": 2.531400966183575, "grad_norm": 0.07090573757886887, "learning_rate": 9.999959042901174e-05, "loss": 0.0248, "step": 524 }, { "epoch": 2.536231884057971, "grad_norm": 0.08650020509958267, "learning_rate": 9.999944252865192e-05, "loss": 0.0281, "step": 525 }, { "epoch": 2.536231884057971, "eval_loss": 0.02850218303501606, "eval_runtime": 20.6152, "eval_samples_per_second": 4.851, "eval_steps_per_second": 0.146, "step": 525 }, { "epoch": 2.541062801932367, "grad_norm": 0.0966835767030716, "learning_rate": 9.999927187457181e-05, "loss": 0.0385, "step": 526 }, { "epoch": 2.545893719806763, "grad_norm": 0.06122938171029091, "learning_rate": 9.999907846684906e-05, "loss": 0.024, "step": 527 }, { "epoch": 2.550724637681159, "grad_norm": 0.06990386545658112, "learning_rate": 9.999886230557167e-05, "loss": 0.0259, "step": 528 }, { "epoch": 2.5555555555555554, "grad_norm": 0.07949528098106384, "learning_rate": 9.999862339083804e-05, "loss": 0.0266, "step": 529 }, { "epoch": 2.5603864734299515, "grad_norm": 0.06846433132886887, "learning_rate": 9.999836172275688e-05, "loss": 0.0254, "step": 530 }, { "epoch": 2.5652173913043477, "grad_norm": 0.06829697638750076, "learning_rate": 9.999807730144728e-05, "loss": 0.0257, "step": 531 }, { "epoch": 2.570048309178744, "grad_norm": 0.08349194377660751, "learning_rate": 9.999777012703866e-05, "loss": 0.0256, "step": 532 }, { "epoch": 2.57487922705314, "grad_norm": 0.09388142079114914, "learning_rate": 9.999744019967081e-05, "loss": 0.0202, "step": 533 }, { "epoch": 2.579710144927536, "grad_norm": 0.06379712373018265, "learning_rate": 9.999708751949389e-05, "loss": 0.0293, "step": 534 }, { "epoch": 2.5845410628019323, "grad_norm": 0.07250544428825378, "learning_rate": 9.999671208666838e-05, "loss": 0.0234, "step": 535 }, { "epoch": 2.5893719806763285, "grad_norm": 0.073805071413517, "learning_rate": 9.999631390136513e-05, "loss": 0.0235, "step": 536 }, { "epoch": 2.5942028985507246, "grad_norm": 0.06304319947957993, "learning_rate": 9.999589296376537e-05, "loss": 0.0221, "step": 537 }, { "epoch": 2.5990338164251208, "grad_norm": 0.07283611595630646, "learning_rate": 9.999544927406063e-05, "loss": 0.0233, "step": 538 }, { "epoch": 2.603864734299517, "grad_norm": 0.08080217987298965, "learning_rate": 9.999498283245284e-05, "loss": 0.0256, "step": 539 }, { "epoch": 2.608695652173913, "grad_norm": 0.07075383514165878, "learning_rate": 9.999449363915427e-05, "loss": 0.0231, "step": 540 }, { "epoch": 2.6135265700483092, "grad_norm": 0.07120423018932343, "learning_rate": 9.999398169438754e-05, "loss": 0.0308, "step": 541 }, { "epoch": 2.6183574879227054, "grad_norm": 0.09204115718603134, "learning_rate": 9.999344699838562e-05, "loss": 0.0204, "step": 542 }, { "epoch": 2.6231884057971016, "grad_norm": 0.07161355018615723, "learning_rate": 9.999288955139183e-05, "loss": 0.0264, "step": 543 }, { "epoch": 2.6280193236714977, "grad_norm": 0.07486795634031296, "learning_rate": 9.999230935365989e-05, "loss": 0.0272, "step": 544 }, { "epoch": 2.632850241545894, "grad_norm": 0.06936871260404587, "learning_rate": 9.999170640545378e-05, "loss": 0.0252, "step": 545 }, { "epoch": 2.63768115942029, "grad_norm": 0.07348743826150894, "learning_rate": 9.999108070704795e-05, "loss": 0.0301, "step": 546 }, { "epoch": 2.642512077294686, "grad_norm": 0.07645001262426376, "learning_rate": 9.99904322587271e-05, "loss": 0.0271, "step": 547 }, { "epoch": 2.6473429951690823, "grad_norm": 0.0847916230559349, "learning_rate": 9.998976106078634e-05, "loss": 0.0321, "step": 548 }, { "epoch": 2.6521739130434785, "grad_norm": 0.0603368803858757, "learning_rate": 9.99890671135311e-05, "loss": 0.0188, "step": 549 }, { "epoch": 2.6570048309178746, "grad_norm": 0.05718471109867096, "learning_rate": 9.998835041727723e-05, "loss": 0.0212, "step": 550 }, { "epoch": 2.6570048309178746, "eval_loss": 0.02679239958524704, "eval_runtime": 20.6052, "eval_samples_per_second": 4.853, "eval_steps_per_second": 0.146, "step": 550 }, { "epoch": 2.661835748792271, "grad_norm": 0.06785376369953156, "learning_rate": 9.998761097235083e-05, "loss": 0.0313, "step": 551 }, { "epoch": 2.6666666666666665, "grad_norm": 0.11118228733539581, "learning_rate": 9.998684877908844e-05, "loss": 0.0486, "step": 552 }, { "epoch": 2.6714975845410627, "grad_norm": 0.07705774903297424, "learning_rate": 9.998606383783691e-05, "loss": 0.0294, "step": 553 }, { "epoch": 2.676328502415459, "grad_norm": 0.06260982900857925, "learning_rate": 9.998525614895343e-05, "loss": 0.019, "step": 554 }, { "epoch": 2.681159420289855, "grad_norm": 0.07005689293146133, "learning_rate": 9.99844257128056e-05, "loss": 0.0253, "step": 555 }, { "epoch": 2.685990338164251, "grad_norm": 0.06078100576996803, "learning_rate": 9.99835725297713e-05, "loss": 0.0209, "step": 556 }, { "epoch": 2.6908212560386473, "grad_norm": 0.059335701167583466, "learning_rate": 9.998269660023882e-05, "loss": 0.0266, "step": 557 }, { "epoch": 2.6956521739130435, "grad_norm": 0.06634645164012909, "learning_rate": 9.998179792460676e-05, "loss": 0.0301, "step": 558 }, { "epoch": 2.7004830917874396, "grad_norm": 0.062110912054777145, "learning_rate": 9.99808765032841e-05, "loss": 0.0262, "step": 559 }, { "epoch": 2.7053140096618358, "grad_norm": 0.07377223670482635, "learning_rate": 9.997993233669014e-05, "loss": 0.0225, "step": 560 }, { "epoch": 2.710144927536232, "grad_norm": 0.05780097842216492, "learning_rate": 9.997896542525459e-05, "loss": 0.0184, "step": 561 }, { "epoch": 2.714975845410628, "grad_norm": 0.050899870693683624, "learning_rate": 9.997797576941744e-05, "loss": 0.0182, "step": 562 }, { "epoch": 2.7198067632850242, "grad_norm": 0.07264772057533264, "learning_rate": 9.997696336962907e-05, "loss": 0.0235, "step": 563 }, { "epoch": 2.7246376811594204, "grad_norm": 0.0789547711610794, "learning_rate": 9.997592822635021e-05, "loss": 0.0294, "step": 564 }, { "epoch": 2.7294685990338166, "grad_norm": 0.10101927071809769, "learning_rate": 9.997487034005193e-05, "loss": 0.0264, "step": 565 }, { "epoch": 2.7342995169082127, "grad_norm": 0.059210047125816345, "learning_rate": 9.997378971121564e-05, "loss": 0.0182, "step": 566 }, { "epoch": 2.7391304347826084, "grad_norm": 0.05791271850466728, "learning_rate": 9.997268634033312e-05, "loss": 0.0183, "step": 567 }, { "epoch": 2.7439613526570046, "grad_norm": 0.07589870691299438, "learning_rate": 9.99715602279065e-05, "loss": 0.0256, "step": 568 }, { "epoch": 2.7487922705314007, "grad_norm": 0.0665014311671257, "learning_rate": 9.997041137444823e-05, "loss": 0.0205, "step": 569 }, { "epoch": 2.753623188405797, "grad_norm": 0.05872458219528198, "learning_rate": 9.996923978048115e-05, "loss": 0.0182, "step": 570 }, { "epoch": 2.758454106280193, "grad_norm": 0.07963749021291733, "learning_rate": 9.996804544653842e-05, "loss": 0.0224, "step": 571 }, { "epoch": 2.763285024154589, "grad_norm": 0.06978920102119446, "learning_rate": 9.996682837316356e-05, "loss": 0.0222, "step": 572 }, { "epoch": 2.7681159420289854, "grad_norm": 0.06920437514781952, "learning_rate": 9.996558856091043e-05, "loss": 0.0233, "step": 573 }, { "epoch": 2.7729468599033815, "grad_norm": 0.06745556741952896, "learning_rate": 9.996432601034324e-05, "loss": 0.0203, "step": 574 }, { "epoch": 2.7777777777777777, "grad_norm": 0.06809904426336288, "learning_rate": 9.996304072203657e-05, "loss": 0.0221, "step": 575 }, { "epoch": 2.7777777777777777, "eval_loss": 0.02665121853351593, "eval_runtime": 20.6302, "eval_samples_per_second": 4.847, "eval_steps_per_second": 0.145, "step": 575 }, { "epoch": 2.782608695652174, "grad_norm": 0.05648452416062355, "learning_rate": 9.99617326965753e-05, "loss": 0.019, "step": 576 }, { "epoch": 2.78743961352657, "grad_norm": 0.06754475831985474, "learning_rate": 9.996040193455472e-05, "loss": 0.019, "step": 577 }, { "epoch": 2.792270531400966, "grad_norm": 0.07781781256198883, "learning_rate": 9.99590484365804e-05, "loss": 0.0279, "step": 578 }, { "epoch": 2.7971014492753623, "grad_norm": 0.06808023899793625, "learning_rate": 9.995767220326829e-05, "loss": 0.0188, "step": 579 }, { "epoch": 2.8019323671497585, "grad_norm": 0.08858218789100647, "learning_rate": 9.99562732352447e-05, "loss": 0.0254, "step": 580 }, { "epoch": 2.8067632850241546, "grad_norm": 0.0699530616402626, "learning_rate": 9.995485153314628e-05, "loss": 0.0229, "step": 581 }, { "epoch": 2.8115942028985508, "grad_norm": 0.067716583609581, "learning_rate": 9.995340709762002e-05, "loss": 0.0248, "step": 582 }, { "epoch": 2.816425120772947, "grad_norm": 0.06093629077076912, "learning_rate": 9.995193992932321e-05, "loss": 0.0201, "step": 583 }, { "epoch": 2.821256038647343, "grad_norm": 0.061484821140766144, "learning_rate": 9.995045002892358e-05, "loss": 0.0279, "step": 584 }, { "epoch": 2.8260869565217392, "grad_norm": 0.05432453379034996, "learning_rate": 9.994893739709912e-05, "loss": 0.022, "step": 585 }, { "epoch": 2.8309178743961354, "grad_norm": 0.06907227635383606, "learning_rate": 9.994740203453821e-05, "loss": 0.0229, "step": 586 }, { "epoch": 2.8357487922705316, "grad_norm": 0.08959613740444183, "learning_rate": 9.994584394193957e-05, "loss": 0.0293, "step": 587 }, { "epoch": 2.8405797101449277, "grad_norm": 0.06915111094713211, "learning_rate": 9.994426312001223e-05, "loss": 0.0195, "step": 588 }, { "epoch": 2.845410628019324, "grad_norm": 0.06990355998277664, "learning_rate": 9.994265956947563e-05, "loss": 0.0204, "step": 589 }, { "epoch": 2.85024154589372, "grad_norm": 0.08454722166061401, "learning_rate": 9.994103329105947e-05, "loss": 0.024, "step": 590 }, { "epoch": 2.855072463768116, "grad_norm": 0.11382115632295609, "learning_rate": 9.993938428550387e-05, "loss": 0.0297, "step": 591 }, { "epoch": 2.8599033816425123, "grad_norm": 0.07049544900655746, "learning_rate": 9.993771255355921e-05, "loss": 0.0237, "step": 592 }, { "epoch": 2.864734299516908, "grad_norm": 0.06032729148864746, "learning_rate": 9.993601809598634e-05, "loss": 0.0192, "step": 593 }, { "epoch": 2.869565217391304, "grad_norm": 0.0696573555469513, "learning_rate": 9.99343009135563e-05, "loss": 0.0233, "step": 594 }, { "epoch": 2.8743961352657004, "grad_norm": 0.06839869916439056, "learning_rate": 9.993256100705058e-05, "loss": 0.0233, "step": 595 }, { "epoch": 2.8792270531400965, "grad_norm": 0.061628565192222595, "learning_rate": 9.993079837726096e-05, "loss": 0.0228, "step": 596 }, { "epoch": 2.8840579710144927, "grad_norm": 0.06356792151927948, "learning_rate": 9.992901302498959e-05, "loss": 0.0239, "step": 597 }, { "epoch": 2.888888888888889, "grad_norm": 0.06923532485961914, "learning_rate": 9.992720495104895e-05, "loss": 0.0264, "step": 598 }, { "epoch": 2.893719806763285, "grad_norm": 0.07253517955541611, "learning_rate": 9.992537415626183e-05, "loss": 0.0218, "step": 599 }, { "epoch": 2.898550724637681, "grad_norm": 0.07643329352140427, "learning_rate": 9.992352064146142e-05, "loss": 0.0225, "step": 600 }, { "epoch": 2.898550724637681, "eval_loss": 0.026624999940395355, "eval_runtime": 20.5987, "eval_samples_per_second": 4.855, "eval_steps_per_second": 0.146, "step": 600 }, { "epoch": 2.9033816425120773, "grad_norm": 0.055665720254182816, "learning_rate": 9.992164440749119e-05, "loss": 0.0198, "step": 601 }, { "epoch": 2.9082125603864735, "grad_norm": 0.0848175436258316, "learning_rate": 9.9919745455205e-05, "loss": 0.0293, "step": 602 }, { "epoch": 2.9130434782608696, "grad_norm": 0.06863771378993988, "learning_rate": 9.991782378546702e-05, "loss": 0.0237, "step": 603 }, { "epoch": 2.917874396135266, "grad_norm": 0.07147162407636642, "learning_rate": 9.991587939915173e-05, "loss": 0.0246, "step": 604 }, { "epoch": 2.922705314009662, "grad_norm": 0.07691524177789688, "learning_rate": 9.991391229714401e-05, "loss": 0.0251, "step": 605 }, { "epoch": 2.927536231884058, "grad_norm": 0.06954370439052582, "learning_rate": 9.991192248033908e-05, "loss": 0.0251, "step": 606 }, { "epoch": 2.9323671497584543, "grad_norm": 0.09142833948135376, "learning_rate": 9.990990994964239e-05, "loss": 0.0339, "step": 607 }, { "epoch": 2.9371980676328504, "grad_norm": 0.09188190847635269, "learning_rate": 9.990787470596985e-05, "loss": 0.0228, "step": 608 }, { "epoch": 2.942028985507246, "grad_norm": 0.07027522474527359, "learning_rate": 9.990581675024763e-05, "loss": 0.0232, "step": 609 }, { "epoch": 2.9468599033816423, "grad_norm": 0.0655936598777771, "learning_rate": 9.99037360834123e-05, "loss": 0.0216, "step": 610 }, { "epoch": 2.9516908212560384, "grad_norm": 0.062273845076560974, "learning_rate": 9.99016327064107e-05, "loss": 0.0189, "step": 611 }, { "epoch": 2.9565217391304346, "grad_norm": 0.08527228981256485, "learning_rate": 9.989950662020007e-05, "loss": 0.0286, "step": 612 }, { "epoch": 2.9613526570048307, "grad_norm": 0.06256254762411118, "learning_rate": 9.98973578257479e-05, "loss": 0.0204, "step": 613 }, { "epoch": 2.966183574879227, "grad_norm": 0.08768539875745773, "learning_rate": 9.989518632403208e-05, "loss": 0.0325, "step": 614 }, { "epoch": 2.971014492753623, "grad_norm": 0.07069282233715057, "learning_rate": 9.989299211604082e-05, "loss": 0.021, "step": 615 }, { "epoch": 2.975845410628019, "grad_norm": 0.06434057652950287, "learning_rate": 9.989077520277264e-05, "loss": 0.0195, "step": 616 }, { "epoch": 2.9806763285024154, "grad_norm": 0.07296837121248245, "learning_rate": 9.988853558523646e-05, "loss": 0.0254, "step": 617 }, { "epoch": 2.9855072463768115, "grad_norm": 0.08000582456588745, "learning_rate": 9.988627326445143e-05, "loss": 0.0261, "step": 618 }, { "epoch": 2.9903381642512077, "grad_norm": 0.08119505643844604, "learning_rate": 9.988398824144714e-05, "loss": 0.0316, "step": 619 }, { "epoch": 2.995169082125604, "grad_norm": 0.053310543298721313, "learning_rate": 9.98816805172634e-05, "loss": 0.0216, "step": 620 }, { "epoch": 3.0, "grad_norm": 0.12208092957735062, "learning_rate": 9.987935009295044e-05, "loss": 0.016, "step": 621 }, { "epoch": 3.004830917874396, "grad_norm": 0.06508877128362656, "learning_rate": 9.987699696956878e-05, "loss": 0.0244, "step": 622 }, { "epoch": 3.0096618357487923, "grad_norm": 0.049330223351716995, "learning_rate": 9.987462114818928e-05, "loss": 0.0159, "step": 623 }, { "epoch": 3.0144927536231885, "grad_norm": 0.06686580181121826, "learning_rate": 9.987222262989315e-05, "loss": 0.018, "step": 624 }, { "epoch": 3.0193236714975846, "grad_norm": 0.061486225575208664, "learning_rate": 9.986980141577187e-05, "loss": 0.0264, "step": 625 }, { "epoch": 3.0193236714975846, "eval_loss": 0.02915319800376892, "eval_runtime": 20.6012, "eval_samples_per_second": 4.854, "eval_steps_per_second": 0.146, "step": 625 }, { "epoch": 3.024154589371981, "grad_norm": 0.06663984060287476, "learning_rate": 9.98673575069273e-05, "loss": 0.0181, "step": 626 }, { "epoch": 3.028985507246377, "grad_norm": 0.06651222705841064, "learning_rate": 9.98648909044716e-05, "loss": 0.0183, "step": 627 }, { "epoch": 3.033816425120773, "grad_norm": 0.06698988378047943, "learning_rate": 9.986240160952732e-05, "loss": 0.018, "step": 628 }, { "epoch": 3.0386473429951693, "grad_norm": 0.08998807519674301, "learning_rate": 9.985988962322721e-05, "loss": 0.0198, "step": 629 }, { "epoch": 3.0434782608695654, "grad_norm": 0.08781043440103531, "learning_rate": 9.985735494671448e-05, "loss": 0.0241, "step": 630 }, { "epoch": 3.0483091787439616, "grad_norm": 0.06608462333679199, "learning_rate": 9.985479758114259e-05, "loss": 0.0139, "step": 631 }, { "epoch": 3.0531400966183573, "grad_norm": 0.06002631038427353, "learning_rate": 9.985221752767535e-05, "loss": 0.0162, "step": 632 }, { "epoch": 3.0579710144927534, "grad_norm": 0.05958976596593857, "learning_rate": 9.984961478748688e-05, "loss": 0.0185, "step": 633 }, { "epoch": 3.0628019323671496, "grad_norm": 0.06394585967063904, "learning_rate": 9.984698936176164e-05, "loss": 0.0235, "step": 634 }, { "epoch": 3.0676328502415457, "grad_norm": 0.07614199072122574, "learning_rate": 9.984434125169441e-05, "loss": 0.0202, "step": 635 }, { "epoch": 3.072463768115942, "grad_norm": 0.06564124673604965, "learning_rate": 9.98416704584903e-05, "loss": 0.0165, "step": 636 }, { "epoch": 3.077294685990338, "grad_norm": 0.06295037269592285, "learning_rate": 9.983897698336471e-05, "loss": 0.0193, "step": 637 }, { "epoch": 3.082125603864734, "grad_norm": 0.06796164065599442, "learning_rate": 9.98362608275434e-05, "loss": 0.0185, "step": 638 }, { "epoch": 3.0869565217391304, "grad_norm": 0.08391165733337402, "learning_rate": 9.983352199226243e-05, "loss": 0.0181, "step": 639 }, { "epoch": 3.0917874396135265, "grad_norm": 0.062315862625837326, "learning_rate": 9.98307604787682e-05, "loss": 0.0164, "step": 640 }, { "epoch": 3.0966183574879227, "grad_norm": 0.05803072825074196, "learning_rate": 9.982797628831739e-05, "loss": 0.0144, "step": 641 }, { "epoch": 3.101449275362319, "grad_norm": 0.06700735539197922, "learning_rate": 9.982516942217705e-05, "loss": 0.0175, "step": 642 }, { "epoch": 3.106280193236715, "grad_norm": 0.08924155682325363, "learning_rate": 9.982233988162455e-05, "loss": 0.0225, "step": 643 }, { "epoch": 3.111111111111111, "grad_norm": 0.08006013929843903, "learning_rate": 9.981948766794752e-05, "loss": 0.016, "step": 644 }, { "epoch": 3.1159420289855073, "grad_norm": 0.06664010137319565, "learning_rate": 9.981661278244394e-05, "loss": 0.0139, "step": 645 }, { "epoch": 3.1207729468599035, "grad_norm": 0.062214843928813934, "learning_rate": 9.981371522642212e-05, "loss": 0.0143, "step": 646 }, { "epoch": 3.1256038647342996, "grad_norm": 0.06204639747738838, "learning_rate": 9.98107950012007e-05, "loss": 0.0167, "step": 647 }, { "epoch": 3.130434782608696, "grad_norm": 0.0846833735704422, "learning_rate": 9.980785210810859e-05, "loss": 0.0325, "step": 648 }, { "epoch": 3.135265700483092, "grad_norm": 0.06417378038167953, "learning_rate": 9.980488654848505e-05, "loss": 0.0182, "step": 649 }, { "epoch": 3.140096618357488, "grad_norm": 0.06934668123722076, "learning_rate": 9.980189832367966e-05, "loss": 0.0196, "step": 650 }, { "epoch": 3.140096618357488, "eval_loss": 0.027954233810305595, "eval_runtime": 21.3091, "eval_samples_per_second": 4.693, "eval_steps_per_second": 0.141, "step": 650 }, { "epoch": 3.1449275362318843, "grad_norm": 0.13086830079555511, "learning_rate": 9.979888743505225e-05, "loss": 0.0146, "step": 651 }, { "epoch": 3.14975845410628, "grad_norm": 0.06644698232412338, "learning_rate": 9.979585388397308e-05, "loss": 0.017, "step": 652 }, { "epoch": 3.154589371980676, "grad_norm": 0.06060745194554329, "learning_rate": 9.979279767182262e-05, "loss": 0.0189, "step": 653 }, { "epoch": 3.1594202898550723, "grad_norm": 0.05614025145769119, "learning_rate": 9.978971879999169e-05, "loss": 0.0154, "step": 654 }, { "epoch": 3.1642512077294684, "grad_norm": 0.0934065654873848, "learning_rate": 9.97866172698814e-05, "loss": 0.0234, "step": 655 }, { "epoch": 3.1690821256038646, "grad_norm": 0.08930619060993195, "learning_rate": 9.978349308290325e-05, "loss": 0.0178, "step": 656 }, { "epoch": 3.1739130434782608, "grad_norm": 0.07199080288410187, "learning_rate": 9.978034624047895e-05, "loss": 0.0189, "step": 657 }, { "epoch": 3.178743961352657, "grad_norm": 0.06786323338747025, "learning_rate": 9.977717674404056e-05, "loss": 0.0203, "step": 658 }, { "epoch": 3.183574879227053, "grad_norm": 0.05707789584994316, "learning_rate": 9.977398459503049e-05, "loss": 0.0168, "step": 659 }, { "epoch": 3.1884057971014492, "grad_norm": 0.05864753946661949, "learning_rate": 9.977076979490138e-05, "loss": 0.0145, "step": 660 }, { "epoch": 3.1932367149758454, "grad_norm": 0.07442363351583481, "learning_rate": 9.976753234511627e-05, "loss": 0.0224, "step": 661 }, { "epoch": 3.1980676328502415, "grad_norm": 0.05928374454379082, "learning_rate": 9.97642722471484e-05, "loss": 0.0151, "step": 662 }, { "epoch": 3.2028985507246377, "grad_norm": 0.05867256596684456, "learning_rate": 9.976098950248141e-05, "loss": 0.0153, "step": 663 }, { "epoch": 3.207729468599034, "grad_norm": 0.05609915778040886, "learning_rate": 9.975768411260917e-05, "loss": 0.017, "step": 664 }, { "epoch": 3.21256038647343, "grad_norm": 0.08102008700370789, "learning_rate": 9.975435607903596e-05, "loss": 0.0224, "step": 665 }, { "epoch": 3.217391304347826, "grad_norm": 0.04532697796821594, "learning_rate": 9.975100540327624e-05, "loss": 0.0127, "step": 666 }, { "epoch": 3.2222222222222223, "grad_norm": 0.07046009600162506, "learning_rate": 9.974763208685487e-05, "loss": 0.0233, "step": 667 }, { "epoch": 3.2270531400966185, "grad_norm": 0.059037089347839355, "learning_rate": 9.974423613130697e-05, "loss": 0.0155, "step": 668 }, { "epoch": 3.2318840579710146, "grad_norm": 0.06592457741498947, "learning_rate": 9.974081753817795e-05, "loss": 0.0187, "step": 669 }, { "epoch": 3.236714975845411, "grad_norm": 0.058792632073163986, "learning_rate": 9.973737630902356e-05, "loss": 0.0162, "step": 670 }, { "epoch": 3.241545893719807, "grad_norm": 0.07621821016073227, "learning_rate": 9.973391244540983e-05, "loss": 0.0223, "step": 671 }, { "epoch": 3.246376811594203, "grad_norm": 0.09599140286445618, "learning_rate": 9.973042594891309e-05, "loss": 0.0231, "step": 672 }, { "epoch": 3.2512077294685993, "grad_norm": 0.06377045810222626, "learning_rate": 9.972691682111997e-05, "loss": 0.0178, "step": 673 }, { "epoch": 3.2560386473429954, "grad_norm": 0.08155626058578491, "learning_rate": 9.972338506362742e-05, "loss": 0.0228, "step": 674 }, { "epoch": 3.260869565217391, "grad_norm": 0.07435385137796402, "learning_rate": 9.971983067804265e-05, "loss": 0.0185, "step": 675 }, { "epoch": 3.260869565217391, "eval_loss": 0.02637997642159462, "eval_runtime": 20.6021, "eval_samples_per_second": 4.854, "eval_steps_per_second": 0.146, "step": 675 }, { "epoch": 3.2657004830917873, "grad_norm": 0.07467380911111832, "learning_rate": 9.971625366598319e-05, "loss": 0.0192, "step": 676 }, { "epoch": 3.2705314009661834, "grad_norm": 0.07276886701583862, "learning_rate": 9.971265402907688e-05, "loss": 0.0203, "step": 677 }, { "epoch": 3.2753623188405796, "grad_norm": 0.06594165414571762, "learning_rate": 9.970903176896183e-05, "loss": 0.0159, "step": 678 }, { "epoch": 3.2801932367149758, "grad_norm": 0.06967869400978088, "learning_rate": 9.970538688728644e-05, "loss": 0.0171, "step": 679 }, { "epoch": 3.285024154589372, "grad_norm": 0.07510808855295181, "learning_rate": 9.970171938570946e-05, "loss": 0.0174, "step": 680 }, { "epoch": 3.289855072463768, "grad_norm": 0.06190778687596321, "learning_rate": 9.969802926589986e-05, "loss": 0.0171, "step": 681 }, { "epoch": 3.2946859903381642, "grad_norm": 0.07371743023395538, "learning_rate": 9.969431652953695e-05, "loss": 0.0193, "step": 682 }, { "epoch": 3.2995169082125604, "grad_norm": 0.14643917977809906, "learning_rate": 9.969058117831034e-05, "loss": 0.0229, "step": 683 }, { "epoch": 3.3043478260869565, "grad_norm": 0.0634489580988884, "learning_rate": 9.968682321391986e-05, "loss": 0.0183, "step": 684 }, { "epoch": 3.3091787439613527, "grad_norm": 0.07199045270681381, "learning_rate": 9.968304263807574e-05, "loss": 0.0167, "step": 685 }, { "epoch": 3.314009661835749, "grad_norm": 0.05730137601494789, "learning_rate": 9.96792394524984e-05, "loss": 0.0165, "step": 686 }, { "epoch": 3.318840579710145, "grad_norm": 0.07792101055383682, "learning_rate": 9.967541365891863e-05, "loss": 0.0243, "step": 687 }, { "epoch": 3.323671497584541, "grad_norm": 0.06612062454223633, "learning_rate": 9.967156525907743e-05, "loss": 0.0177, "step": 688 }, { "epoch": 3.3285024154589373, "grad_norm": 0.08418621122837067, "learning_rate": 9.966769425472616e-05, "loss": 0.0174, "step": 689 }, { "epoch": 3.3333333333333335, "grad_norm": 0.08468875288963318, "learning_rate": 9.966380064762642e-05, "loss": 0.0152, "step": 690 }, { "epoch": 3.3381642512077296, "grad_norm": 0.05754866451025009, "learning_rate": 9.96598844395501e-05, "loss": 0.0132, "step": 691 }, { "epoch": 3.342995169082126, "grad_norm": 0.07585375756025314, "learning_rate": 9.96559456322794e-05, "loss": 0.0158, "step": 692 }, { "epoch": 3.3478260869565215, "grad_norm": 0.08202559500932693, "learning_rate": 9.96519842276068e-05, "loss": 0.0208, "step": 693 }, { "epoch": 3.3526570048309177, "grad_norm": 0.0768493190407753, "learning_rate": 9.964800022733504e-05, "loss": 0.024, "step": 694 }, { "epoch": 3.357487922705314, "grad_norm": 0.09011416882276535, "learning_rate": 9.964399363327716e-05, "loss": 0.0229, "step": 695 }, { "epoch": 3.36231884057971, "grad_norm": 0.06828708201646805, "learning_rate": 9.963996444725647e-05, "loss": 0.0216, "step": 696 }, { "epoch": 3.367149758454106, "grad_norm": 0.09616364538669586, "learning_rate": 9.96359126711066e-05, "loss": 0.0213, "step": 697 }, { "epoch": 3.3719806763285023, "grad_norm": 0.056025609374046326, "learning_rate": 9.963183830667138e-05, "loss": 0.0138, "step": 698 }, { "epoch": 3.3768115942028984, "grad_norm": 0.052819810807704926, "learning_rate": 9.9627741355805e-05, "loss": 0.0145, "step": 699 }, { "epoch": 3.3816425120772946, "grad_norm": 0.05790337547659874, "learning_rate": 9.96236218203719e-05, "loss": 0.0161, "step": 700 }, { "epoch": 3.3816425120772946, "eval_loss": 0.024786872789263725, "eval_runtime": 20.6335, "eval_samples_per_second": 4.846, "eval_steps_per_second": 0.145, "step": 700 }, { "epoch": 3.3864734299516908, "grad_norm": 0.053379714488983154, "learning_rate": 9.96194797022468e-05, "loss": 0.0141, "step": 701 }, { "epoch": 3.391304347826087, "grad_norm": 0.059687525033950806, "learning_rate": 9.961531500331469e-05, "loss": 0.0166, "step": 702 }, { "epoch": 3.396135265700483, "grad_norm": 0.05600206181406975, "learning_rate": 9.961112772547083e-05, "loss": 0.0164, "step": 703 }, { "epoch": 3.4009661835748792, "grad_norm": 0.07950199395418167, "learning_rate": 9.960691787062076e-05, "loss": 0.0169, "step": 704 }, { "epoch": 3.4057971014492754, "grad_norm": 0.08069173246622086, "learning_rate": 9.960268544068032e-05, "loss": 0.025, "step": 705 }, { "epoch": 3.4106280193236715, "grad_norm": 0.06130361557006836, "learning_rate": 9.959843043757557e-05, "loss": 0.0155, "step": 706 }, { "epoch": 3.4154589371980677, "grad_norm": 0.06326994299888611, "learning_rate": 9.959415286324289e-05, "loss": 0.0141, "step": 707 }, { "epoch": 3.420289855072464, "grad_norm": 0.07043007761240005, "learning_rate": 9.958985271962896e-05, "loss": 0.0143, "step": 708 }, { "epoch": 3.42512077294686, "grad_norm": 0.09066811949014664, "learning_rate": 9.958553000869061e-05, "loss": 0.026, "step": 709 }, { "epoch": 3.429951690821256, "grad_norm": 0.06408656388521194, "learning_rate": 9.958118473239507e-05, "loss": 0.0139, "step": 710 }, { "epoch": 3.4347826086956523, "grad_norm": 0.06362169981002808, "learning_rate": 9.957681689271977e-05, "loss": 0.0166, "step": 711 }, { "epoch": 3.4396135265700485, "grad_norm": 0.07092583179473877, "learning_rate": 9.957242649165241e-05, "loss": 0.0171, "step": 712 }, { "epoch": 3.4444444444444446, "grad_norm": 0.07049217075109482, "learning_rate": 9.956801353119099e-05, "loss": 0.0176, "step": 713 }, { "epoch": 3.449275362318841, "grad_norm": 0.06697899103164673, "learning_rate": 9.956357801334375e-05, "loss": 0.0182, "step": 714 }, { "epoch": 3.454106280193237, "grad_norm": 0.06497643142938614, "learning_rate": 9.955911994012923e-05, "loss": 0.016, "step": 715 }, { "epoch": 3.4589371980676327, "grad_norm": 0.06678501516580582, "learning_rate": 9.955463931357616e-05, "loss": 0.0229, "step": 716 }, { "epoch": 3.463768115942029, "grad_norm": 0.04944721981883049, "learning_rate": 9.955013613572362e-05, "loss": 0.0177, "step": 717 }, { "epoch": 3.468599033816425, "grad_norm": 0.054069582372903824, "learning_rate": 9.954561040862088e-05, "loss": 0.0165, "step": 718 }, { "epoch": 3.473429951690821, "grad_norm": 0.07551441341638565, "learning_rate": 9.954106213432755e-05, "loss": 0.0286, "step": 719 }, { "epoch": 3.4782608695652173, "grad_norm": 0.061375074088573456, "learning_rate": 9.95364913149134e-05, "loss": 0.0185, "step": 720 }, { "epoch": 3.4830917874396135, "grad_norm": 0.060356441885232925, "learning_rate": 9.953189795245857e-05, "loss": 0.0168, "step": 721 }, { "epoch": 3.4879227053140096, "grad_norm": 0.07686608284711838, "learning_rate": 9.952728204905338e-05, "loss": 0.0148, "step": 722 }, { "epoch": 3.4927536231884058, "grad_norm": 0.05384129658341408, "learning_rate": 9.952264360679844e-05, "loss": 0.0171, "step": 723 }, { "epoch": 3.497584541062802, "grad_norm": 0.05942288041114807, "learning_rate": 9.951798262780458e-05, "loss": 0.0172, "step": 724 }, { "epoch": 3.502415458937198, "grad_norm": 0.07827403396368027, "learning_rate": 9.951329911419298e-05, "loss": 0.0186, "step": 725 }, { "epoch": 3.502415458937198, "eval_loss": 0.02256963960826397, "eval_runtime": 20.6118, "eval_samples_per_second": 4.852, "eval_steps_per_second": 0.146, "step": 725 }, { "epoch": 3.5072463768115942, "grad_norm": 0.04993787035346031, "learning_rate": 9.950859306809494e-05, "loss": 0.0139, "step": 726 }, { "epoch": 3.5120772946859904, "grad_norm": 0.07030709087848663, "learning_rate": 9.950386449165212e-05, "loss": 0.0166, "step": 727 }, { "epoch": 3.5169082125603865, "grad_norm": 0.04969039931893349, "learning_rate": 9.94991133870164e-05, "loss": 0.012, "step": 728 }, { "epoch": 3.5217391304347827, "grad_norm": 0.06086138263344765, "learning_rate": 9.949433975634992e-05, "loss": 0.019, "step": 729 }, { "epoch": 3.526570048309179, "grad_norm": 0.060854848474264145, "learning_rate": 9.948954360182503e-05, "loss": 0.0144, "step": 730 }, { "epoch": 3.531400966183575, "grad_norm": 0.06262096762657166, "learning_rate": 9.948472492562438e-05, "loss": 0.0164, "step": 731 }, { "epoch": 3.536231884057971, "grad_norm": 0.093068428337574, "learning_rate": 9.947988372994086e-05, "loss": 0.0327, "step": 732 }, { "epoch": 3.541062801932367, "grad_norm": 0.07284866273403168, "learning_rate": 9.947502001697757e-05, "loss": 0.0171, "step": 733 }, { "epoch": 3.545893719806763, "grad_norm": 0.06965293735265732, "learning_rate": 9.947013378894792e-05, "loss": 0.0212, "step": 734 }, { "epoch": 3.550724637681159, "grad_norm": 0.049892608076334, "learning_rate": 9.946522504807551e-05, "loss": 0.0135, "step": 735 }, { "epoch": 3.5555555555555554, "grad_norm": 0.09122290462255478, "learning_rate": 9.94602937965942e-05, "loss": 0.0222, "step": 736 }, { "epoch": 3.5603864734299515, "grad_norm": 0.055852945894002914, "learning_rate": 9.945534003674812e-05, "loss": 0.0201, "step": 737 }, { "epoch": 3.5652173913043477, "grad_norm": 0.056373223662376404, "learning_rate": 9.945036377079164e-05, "loss": 0.0149, "step": 738 }, { "epoch": 3.570048309178744, "grad_norm": 0.07455029338598251, "learning_rate": 9.94453650009893e-05, "loss": 0.0166, "step": 739 }, { "epoch": 3.57487922705314, "grad_norm": 0.07053821533918381, "learning_rate": 9.9440343729616e-05, "loss": 0.0246, "step": 740 }, { "epoch": 3.579710144927536, "grad_norm": 0.056464601308107376, "learning_rate": 9.943529995895679e-05, "loss": 0.0151, "step": 741 }, { "epoch": 3.5845410628019323, "grad_norm": 0.0730137899518013, "learning_rate": 9.943023369130698e-05, "loss": 0.0172, "step": 742 }, { "epoch": 3.5893719806763285, "grad_norm": 0.06213295832276344, "learning_rate": 9.942514492897212e-05, "loss": 0.0166, "step": 743 }, { "epoch": 3.5942028985507246, "grad_norm": 0.07693560421466827, "learning_rate": 9.942003367426803e-05, "loss": 0.0184, "step": 744 }, { "epoch": 3.5990338164251208, "grad_norm": 0.06450362503528595, "learning_rate": 9.941489992952071e-05, "loss": 0.0188, "step": 745 }, { "epoch": 3.603864734299517, "grad_norm": 0.06144733726978302, "learning_rate": 9.940974369706642e-05, "loss": 0.0177, "step": 746 }, { "epoch": 3.608695652173913, "grad_norm": 0.061011552810668945, "learning_rate": 9.940456497925168e-05, "loss": 0.0135, "step": 747 }, { "epoch": 3.6135265700483092, "grad_norm": 0.05580676719546318, "learning_rate": 9.939936377843321e-05, "loss": 0.0131, "step": 748 }, { "epoch": 3.6183574879227054, "grad_norm": 0.07480030506849289, "learning_rate": 9.939414009697795e-05, "loss": 0.018, "step": 749 }, { "epoch": 3.6231884057971016, "grad_norm": 0.1004641205072403, "learning_rate": 9.938889393726314e-05, "loss": 0.0166, "step": 750 }, { "epoch": 3.6231884057971016, "eval_loss": 0.021343432366847992, "eval_runtime": 20.6071, "eval_samples_per_second": 4.853, "eval_steps_per_second": 0.146, "step": 750 }, { "epoch": 3.6280193236714977, "grad_norm": 0.07337497919797897, "learning_rate": 9.938362530167613e-05, "loss": 0.0187, "step": 751 }, { "epoch": 3.632850241545894, "grad_norm": 0.06580359488725662, "learning_rate": 9.937833419261462e-05, "loss": 0.0131, "step": 752 }, { "epoch": 3.63768115942029, "grad_norm": 0.07170310616493225, "learning_rate": 9.937302061248646e-05, "loss": 0.0182, "step": 753 }, { "epoch": 3.642512077294686, "grad_norm": 0.07878907769918442, "learning_rate": 9.936768456370977e-05, "loss": 0.0165, "step": 754 }, { "epoch": 3.6473429951690823, "grad_norm": 0.06542734801769257, "learning_rate": 9.936232604871285e-05, "loss": 0.0126, "step": 755 }, { "epoch": 3.6521739130434785, "grad_norm": 0.057411663234233856, "learning_rate": 9.935694506993427e-05, "loss": 0.0136, "step": 756 }, { "epoch": 3.6570048309178746, "grad_norm": 0.05752653628587723, "learning_rate": 9.935154162982281e-05, "loss": 0.0175, "step": 757 }, { "epoch": 3.661835748792271, "grad_norm": 0.0549551360309124, "learning_rate": 9.934611573083744e-05, "loss": 0.0144, "step": 758 }, { "epoch": 3.6666666666666665, "grad_norm": 0.06211842969059944, "learning_rate": 9.934066737544741e-05, "loss": 0.0159, "step": 759 }, { "epoch": 3.6714975845410627, "grad_norm": 0.08960019052028656, "learning_rate": 9.93351965661321e-05, "loss": 0.0232, "step": 760 }, { "epoch": 3.676328502415459, "grad_norm": 0.08052323758602142, "learning_rate": 9.932970330538123e-05, "loss": 0.0194, "step": 761 }, { "epoch": 3.681159420289855, "grad_norm": 0.07267940789461136, "learning_rate": 9.932418759569462e-05, "loss": 0.022, "step": 762 }, { "epoch": 3.685990338164251, "grad_norm": 0.04540833458304405, "learning_rate": 9.931864943958238e-05, "loss": 0.0146, "step": 763 }, { "epoch": 3.6908212560386473, "grad_norm": 0.06606751680374146, "learning_rate": 9.931308883956479e-05, "loss": 0.0188, "step": 764 }, { "epoch": 3.6956521739130435, "grad_norm": 0.0551118440926075, "learning_rate": 9.930750579817239e-05, "loss": 0.0155, "step": 765 }, { "epoch": 3.7004830917874396, "grad_norm": 0.06126049533486366, "learning_rate": 9.93019003179459e-05, "loss": 0.0138, "step": 766 }, { "epoch": 3.7053140096618358, "grad_norm": 0.06616178900003433, "learning_rate": 9.929627240143625e-05, "loss": 0.017, "step": 767 }, { "epoch": 3.710144927536232, "grad_norm": 0.04692595824599266, "learning_rate": 9.92906220512046e-05, "loss": 0.0133, "step": 768 }, { "epoch": 3.714975845410628, "grad_norm": 0.06590303778648376, "learning_rate": 9.92849492698223e-05, "loss": 0.0157, "step": 769 }, { "epoch": 3.7198067632850242, "grad_norm": 0.0660870224237442, "learning_rate": 9.927925405987093e-05, "loss": 0.0165, "step": 770 }, { "epoch": 3.7246376811594204, "grad_norm": 0.06835846602916718, "learning_rate": 9.927353642394224e-05, "loss": 0.0185, "step": 771 }, { "epoch": 3.7294685990338166, "grad_norm": 0.06855834275484085, "learning_rate": 9.926779636463824e-05, "loss": 0.0157, "step": 772 }, { "epoch": 3.7342995169082127, "grad_norm": 0.07278718799352646, "learning_rate": 9.926203388457107e-05, "loss": 0.0207, "step": 773 }, { "epoch": 3.7391304347826084, "grad_norm": 0.0773162767291069, "learning_rate": 9.925624898636317e-05, "loss": 0.0186, "step": 774 }, { "epoch": 3.7439613526570046, "grad_norm": 0.0571766160428524, "learning_rate": 9.925044167264708e-05, "loss": 0.0141, "step": 775 }, { "epoch": 3.7439613526570046, "eval_loss": 0.021501345559954643, "eval_runtime": 21.3287, "eval_samples_per_second": 4.689, "eval_steps_per_second": 0.141, "step": 775 }, { "epoch": 3.7487922705314007, "grad_norm": 0.08599105477333069, "learning_rate": 9.924461194606561e-05, "loss": 0.0195, "step": 776 }, { "epoch": 3.753623188405797, "grad_norm": 0.050466280430555344, "learning_rate": 9.923875980927175e-05, "loss": 0.0146, "step": 777 }, { "epoch": 3.758454106280193, "grad_norm": 0.061095286160707474, "learning_rate": 9.923288526492869e-05, "loss": 0.0166, "step": 778 }, { "epoch": 3.763285024154589, "grad_norm": 0.05484727770090103, "learning_rate": 9.922698831570982e-05, "loss": 0.0166, "step": 779 }, { "epoch": 3.7681159420289854, "grad_norm": 0.06217224523425102, "learning_rate": 9.92210689642987e-05, "loss": 0.0218, "step": 780 }, { "epoch": 3.7729468599033815, "grad_norm": 0.07372289896011353, "learning_rate": 9.921512721338912e-05, "loss": 0.019, "step": 781 }, { "epoch": 3.7777777777777777, "grad_norm": 0.07227632403373718, "learning_rate": 9.920916306568504e-05, "loss": 0.0154, "step": 782 }, { "epoch": 3.782608695652174, "grad_norm": 0.05255334824323654, "learning_rate": 9.920317652390063e-05, "loss": 0.0128, "step": 783 }, { "epoch": 3.78743961352657, "grad_norm": 0.052771199494600296, "learning_rate": 9.919716759076025e-05, "loss": 0.0131, "step": 784 }, { "epoch": 3.792270531400966, "grad_norm": 0.05544090270996094, "learning_rate": 9.919113626899841e-05, "loss": 0.0152, "step": 785 }, { "epoch": 3.7971014492753623, "grad_norm": 0.058198217302560806, "learning_rate": 9.918508256135988e-05, "loss": 0.015, "step": 786 }, { "epoch": 3.8019323671497585, "grad_norm": 0.058879077434539795, "learning_rate": 9.917900647059955e-05, "loss": 0.0156, "step": 787 }, { "epoch": 3.8067632850241546, "grad_norm": 0.05986057594418526, "learning_rate": 9.917290799948253e-05, "loss": 0.0137, "step": 788 }, { "epoch": 3.8115942028985508, "grad_norm": 0.07952949404716492, "learning_rate": 9.916678715078411e-05, "loss": 0.0173, "step": 789 }, { "epoch": 3.816425120772947, "grad_norm": 0.07398118823766708, "learning_rate": 9.916064392728979e-05, "loss": 0.0186, "step": 790 }, { "epoch": 3.821256038647343, "grad_norm": 0.0736398994922638, "learning_rate": 9.915447833179519e-05, "loss": 0.014, "step": 791 }, { "epoch": 3.8260869565217392, "grad_norm": 0.08044742792844772, "learning_rate": 9.914829036710614e-05, "loss": 0.0162, "step": 792 }, { "epoch": 3.8309178743961354, "grad_norm": 0.05634943023324013, "learning_rate": 9.914208003603869e-05, "loss": 0.0166, "step": 793 }, { "epoch": 3.8357487922705316, "grad_norm": 0.06842391192913055, "learning_rate": 9.913584734141901e-05, "loss": 0.0125, "step": 794 }, { "epoch": 3.8405797101449277, "grad_norm": 0.05727720633149147, "learning_rate": 9.912959228608348e-05, "loss": 0.0157, "step": 795 }, { "epoch": 3.845410628019324, "grad_norm": 0.0616447813808918, "learning_rate": 9.912331487287864e-05, "loss": 0.0204, "step": 796 }, { "epoch": 3.85024154589372, "grad_norm": 0.08196991682052612, "learning_rate": 9.911701510466124e-05, "loss": 0.0246, "step": 797 }, { "epoch": 3.855072463768116, "grad_norm": 0.06222192198038101, "learning_rate": 9.911069298429814e-05, "loss": 0.0144, "step": 798 }, { "epoch": 3.8599033816425123, "grad_norm": 0.1069980263710022, "learning_rate": 9.910434851466642e-05, "loss": 0.0194, "step": 799 }, { "epoch": 3.864734299516908, "grad_norm": 0.06227678433060646, "learning_rate": 9.90979816986533e-05, "loss": 0.0186, "step": 800 }, { "epoch": 3.864734299516908, "eval_loss": 0.021084513515233994, "eval_runtime": 20.6111, "eval_samples_per_second": 4.852, "eval_steps_per_second": 0.146, "step": 800 }, { "epoch": 3.869565217391304, "grad_norm": 0.06213577091693878, "learning_rate": 9.909159253915623e-05, "loss": 0.016, "step": 801 }, { "epoch": 3.8743961352657004, "grad_norm": 0.061741989105939865, "learning_rate": 9.908518103908274e-05, "loss": 0.015, "step": 802 }, { "epoch": 3.8792270531400965, "grad_norm": 0.05534172058105469, "learning_rate": 9.907874720135061e-05, "loss": 0.0138, "step": 803 }, { "epoch": 3.8840579710144927, "grad_norm": 0.06317410618066788, "learning_rate": 9.907229102888772e-05, "loss": 0.0156, "step": 804 }, { "epoch": 3.888888888888889, "grad_norm": 0.06512167304754257, "learning_rate": 9.906581252463216e-05, "loss": 0.0202, "step": 805 }, { "epoch": 3.893719806763285, "grad_norm": 0.08261583000421524, "learning_rate": 9.905931169153215e-05, "loss": 0.0151, "step": 806 }, { "epoch": 3.898550724637681, "grad_norm": 0.0779329165816307, "learning_rate": 9.905278853254609e-05, "loss": 0.0253, "step": 807 }, { "epoch": 3.9033816425120773, "grad_norm": 0.07737044990062714, "learning_rate": 9.904624305064255e-05, "loss": 0.0151, "step": 808 }, { "epoch": 3.9082125603864735, "grad_norm": 0.07593659311532974, "learning_rate": 9.903967524880022e-05, "loss": 0.0204, "step": 809 }, { "epoch": 3.9130434782608696, "grad_norm": 0.05290926992893219, "learning_rate": 9.903308513000798e-05, "loss": 0.0175, "step": 810 }, { "epoch": 3.917874396135266, "grad_norm": 0.05973507836461067, "learning_rate": 9.902647269726489e-05, "loss": 0.0144, "step": 811 }, { "epoch": 3.922705314009662, "grad_norm": 0.0560474619269371, "learning_rate": 9.901983795358008e-05, "loss": 0.0127, "step": 812 }, { "epoch": 3.927536231884058, "grad_norm": 0.06078074499964714, "learning_rate": 9.901318090197291e-05, "loss": 0.0146, "step": 813 }, { "epoch": 3.9323671497584543, "grad_norm": 0.06304194033145905, "learning_rate": 9.900650154547286e-05, "loss": 0.0179, "step": 814 }, { "epoch": 3.9371980676328504, "grad_norm": 0.07986494153738022, "learning_rate": 9.89997998871196e-05, "loss": 0.0228, "step": 815 }, { "epoch": 3.942028985507246, "grad_norm": 0.06479338556528091, "learning_rate": 9.899307592996287e-05, "loss": 0.0142, "step": 816 }, { "epoch": 3.9468599033816423, "grad_norm": 0.06285912543535233, "learning_rate": 9.898632967706264e-05, "loss": 0.0166, "step": 817 }, { "epoch": 3.9516908212560384, "grad_norm": 0.06384378671646118, "learning_rate": 9.897956113148899e-05, "loss": 0.0174, "step": 818 }, { "epoch": 3.9565217391304346, "grad_norm": 0.06164379417896271, "learning_rate": 9.897277029632212e-05, "loss": 0.0143, "step": 819 }, { "epoch": 3.9613526570048307, "grad_norm": 0.08341162651777267, "learning_rate": 9.896595717465243e-05, "loss": 0.0162, "step": 820 }, { "epoch": 3.966183574879227, "grad_norm": 0.08315421640872955, "learning_rate": 9.89591217695804e-05, "loss": 0.0201, "step": 821 }, { "epoch": 3.971014492753623, "grad_norm": 0.0733807235956192, "learning_rate": 9.895226408421669e-05, "loss": 0.0213, "step": 822 }, { "epoch": 3.975845410628019, "grad_norm": 0.05566437914967537, "learning_rate": 9.894538412168213e-05, "loss": 0.0173, "step": 823 }, { "epoch": 3.9806763285024154, "grad_norm": 0.0825580433011055, "learning_rate": 9.89384818851076e-05, "loss": 0.0231, "step": 824 }, { "epoch": 3.9855072463768115, "grad_norm": 0.05818888917565346, "learning_rate": 9.893155737763419e-05, "loss": 0.0119, "step": 825 }, { "epoch": 3.9855072463768115, "eval_loss": 0.020353205502033234, "eval_runtime": 21.3607, "eval_samples_per_second": 4.681, "eval_steps_per_second": 0.14, "step": 825 }, { "epoch": 3.9903381642512077, "grad_norm": 0.055032528936862946, "learning_rate": 9.892461060241312e-05, "loss": 0.0158, "step": 826 }, { "epoch": 3.995169082125604, "grad_norm": 0.0609041228890419, "learning_rate": 9.891764156260568e-05, "loss": 0.0172, "step": 827 }, { "epoch": 4.0, "grad_norm": 0.07889176905155182, "learning_rate": 9.891065026138338e-05, "loss": 0.0125, "step": 828 }, { "epoch": 4.004830917874396, "grad_norm": 0.04410771280527115, "learning_rate": 9.890363670192776e-05, "loss": 0.0131, "step": 829 }, { "epoch": 4.009661835748792, "grad_norm": 0.05885075405240059, "learning_rate": 9.889660088743063e-05, "loss": 0.0163, "step": 830 }, { "epoch": 4.0144927536231885, "grad_norm": 0.042296018451452255, "learning_rate": 9.888954282109377e-05, "loss": 0.0104, "step": 831 }, { "epoch": 4.019323671497585, "grad_norm": 0.06296995282173157, "learning_rate": 9.888246250612921e-05, "loss": 0.0141, "step": 832 }, { "epoch": 4.024154589371981, "grad_norm": 0.06166675686836243, "learning_rate": 9.887535994575902e-05, "loss": 0.0108, "step": 833 }, { "epoch": 4.028985507246377, "grad_norm": 0.06275904923677444, "learning_rate": 9.886823514321548e-05, "loss": 0.0102, "step": 834 }, { "epoch": 4.033816425120773, "grad_norm": 0.07666383683681488, "learning_rate": 9.886108810174088e-05, "loss": 0.0161, "step": 835 }, { "epoch": 4.038647342995169, "grad_norm": 0.07572819292545319, "learning_rate": 9.885391882458773e-05, "loss": 0.0149, "step": 836 }, { "epoch": 4.043478260869565, "grad_norm": 0.05219224467873573, "learning_rate": 9.88467273150186e-05, "loss": 0.0065, "step": 837 }, { "epoch": 4.048309178743962, "grad_norm": 0.05272039771080017, "learning_rate": 9.883951357630622e-05, "loss": 0.009, "step": 838 }, { "epoch": 4.053140096618358, "grad_norm": 0.058741066604852676, "learning_rate": 9.88322776117334e-05, "loss": 0.0121, "step": 839 }, { "epoch": 4.057971014492754, "grad_norm": 0.0712701827287674, "learning_rate": 9.882501942459308e-05, "loss": 0.0107, "step": 840 }, { "epoch": 4.06280193236715, "grad_norm": 0.05130418390035629, "learning_rate": 9.881773901818832e-05, "loss": 0.0102, "step": 841 }, { "epoch": 4.067632850241546, "grad_norm": 0.07418301701545715, "learning_rate": 9.881043639583227e-05, "loss": 0.0114, "step": 842 }, { "epoch": 4.072463768115942, "grad_norm": 0.0563097782433033, "learning_rate": 9.880311156084823e-05, "loss": 0.0101, "step": 843 }, { "epoch": 4.0772946859903385, "grad_norm": 0.11274167895317078, "learning_rate": 9.879576451656955e-05, "loss": 0.0311, "step": 844 }, { "epoch": 4.082125603864735, "grad_norm": 0.06734360754489899, "learning_rate": 9.878839526633974e-05, "loss": 0.0105, "step": 845 }, { "epoch": 4.086956521739131, "grad_norm": 0.04483459144830704, "learning_rate": 9.878100381351239e-05, "loss": 0.0095, "step": 846 }, { "epoch": 4.091787439613527, "grad_norm": 0.06002177298069, "learning_rate": 9.877359016145117e-05, "loss": 0.0105, "step": 847 }, { "epoch": 4.096618357487923, "grad_norm": 0.059757571667432785, "learning_rate": 9.876615431352994e-05, "loss": 0.0116, "step": 848 }, { "epoch": 4.101449275362318, "grad_norm": 0.07245704531669617, "learning_rate": 9.875869627313255e-05, "loss": 0.0111, "step": 849 }, { "epoch": 4.106280193236715, "grad_norm": 0.05853939801454544, "learning_rate": 9.875121604365302e-05, "loss": 0.0097, "step": 850 }, { "epoch": 4.106280193236715, "eval_loss": 0.02096910960972309, "eval_runtime": 20.6071, "eval_samples_per_second": 4.853, "eval_steps_per_second": 0.146, "step": 850 }, { "epoch": 4.111111111111111, "grad_norm": 0.07421394437551498, "learning_rate": 9.874371362849548e-05, "loss": 0.0117, "step": 851 }, { "epoch": 4.115942028985507, "grad_norm": 0.05691864714026451, "learning_rate": 9.873618903107406e-05, "loss": 0.0111, "step": 852 }, { "epoch": 4.120772946859903, "grad_norm": 0.06150828301906586, "learning_rate": 9.87286422548131e-05, "loss": 0.0124, "step": 853 }, { "epoch": 4.125603864734299, "grad_norm": 0.06017156317830086, "learning_rate": 9.872107330314696e-05, "loss": 0.0116, "step": 854 }, { "epoch": 4.130434782608695, "grad_norm": 0.0666033923625946, "learning_rate": 9.871348217952012e-05, "loss": 0.0132, "step": 855 }, { "epoch": 4.1352657004830915, "grad_norm": 0.07256283611059189, "learning_rate": 9.870586888738715e-05, "loss": 0.011, "step": 856 }, { "epoch": 4.140096618357488, "grad_norm": 0.06847047805786133, "learning_rate": 9.869823343021271e-05, "loss": 0.0122, "step": 857 }, { "epoch": 4.144927536231884, "grad_norm": 0.0449073351919651, "learning_rate": 9.869057581147152e-05, "loss": 0.0085, "step": 858 }, { "epoch": 4.14975845410628, "grad_norm": 0.06358068436384201, "learning_rate": 9.868289603464842e-05, "loss": 0.0111, "step": 859 }, { "epoch": 4.154589371980676, "grad_norm": 0.0689551904797554, "learning_rate": 9.86751941032383e-05, "loss": 0.0142, "step": 860 }, { "epoch": 4.159420289855072, "grad_norm": 0.08959997445344925, "learning_rate": 9.866747002074617e-05, "loss": 0.015, "step": 861 }, { "epoch": 4.164251207729468, "grad_norm": 0.059535712003707886, "learning_rate": 9.865972379068711e-05, "loss": 0.0132, "step": 862 }, { "epoch": 4.169082125603865, "grad_norm": 0.047409236431121826, "learning_rate": 9.865195541658623e-05, "loss": 0.0098, "step": 863 }, { "epoch": 4.173913043478261, "grad_norm": 0.05755442753434181, "learning_rate": 9.86441649019788e-05, "loss": 0.0143, "step": 864 }, { "epoch": 4.178743961352657, "grad_norm": 0.05657627061009407, "learning_rate": 9.86363522504101e-05, "loss": 0.0114, "step": 865 }, { "epoch": 4.183574879227053, "grad_norm": 0.05874384567141533, "learning_rate": 9.862851746543554e-05, "loss": 0.0112, "step": 866 }, { "epoch": 4.188405797101449, "grad_norm": 0.057612430304288864, "learning_rate": 9.862066055062051e-05, "loss": 0.0102, "step": 867 }, { "epoch": 4.193236714975845, "grad_norm": 0.07835967093706131, "learning_rate": 9.861278150954059e-05, "loss": 0.013, "step": 868 }, { "epoch": 4.1980676328502415, "grad_norm": 0.07963374257087708, "learning_rate": 9.860488034578132e-05, "loss": 0.0118, "step": 869 }, { "epoch": 4.202898550724638, "grad_norm": 0.06441879272460938, "learning_rate": 9.85969570629384e-05, "loss": 0.0139, "step": 870 }, { "epoch": 4.207729468599034, "grad_norm": 0.061957623809576035, "learning_rate": 9.858901166461754e-05, "loss": 0.0115, "step": 871 }, { "epoch": 4.21256038647343, "grad_norm": 0.1095745861530304, "learning_rate": 9.85810441544345e-05, "loss": 0.011, "step": 872 }, { "epoch": 4.217391304347826, "grad_norm": 0.06752188503742218, "learning_rate": 9.857305453601517e-05, "loss": 0.0167, "step": 873 }, { "epoch": 4.222222222222222, "grad_norm": 0.08148062229156494, "learning_rate": 9.856504281299546e-05, "loss": 0.0157, "step": 874 }, { "epoch": 4.2270531400966185, "grad_norm": 0.050344184041023254, "learning_rate": 9.85570089890213e-05, "loss": 0.0095, "step": 875 }, { "epoch": 4.2270531400966185, "eval_loss": 0.020448356866836548, "eval_runtime": 20.6294, "eval_samples_per_second": 4.847, "eval_steps_per_second": 0.145, "step": 875 }, { "epoch": 4.231884057971015, "grad_norm": 0.056328706443309784, "learning_rate": 9.854895306774876e-05, "loss": 0.0135, "step": 876 }, { "epoch": 4.236714975845411, "grad_norm": 0.04534146934747696, "learning_rate": 9.854087505284391e-05, "loss": 0.0097, "step": 877 }, { "epoch": 4.241545893719807, "grad_norm": 0.08734095841646194, "learning_rate": 9.853277494798287e-05, "loss": 0.0116, "step": 878 }, { "epoch": 4.246376811594203, "grad_norm": 0.07698874920606613, "learning_rate": 9.852465275685187e-05, "loss": 0.0115, "step": 879 }, { "epoch": 4.251207729468599, "grad_norm": 0.04827764257788658, "learning_rate": 9.851650848314713e-05, "loss": 0.0112, "step": 880 }, { "epoch": 4.256038647342995, "grad_norm": 0.07417704910039902, "learning_rate": 9.850834213057494e-05, "loss": 0.0141, "step": 881 }, { "epoch": 4.260869565217392, "grad_norm": 0.06426126509904861, "learning_rate": 9.850015370285164e-05, "loss": 0.0141, "step": 882 }, { "epoch": 4.265700483091788, "grad_norm": 0.05702616274356842, "learning_rate": 9.84919432037036e-05, "loss": 0.0094, "step": 883 }, { "epoch": 4.270531400966184, "grad_norm": 0.050962239503860474, "learning_rate": 9.84837106368673e-05, "loss": 0.0123, "step": 884 }, { "epoch": 4.27536231884058, "grad_norm": 0.05947406217455864, "learning_rate": 9.847545600608917e-05, "loss": 0.0106, "step": 885 }, { "epoch": 4.280193236714976, "grad_norm": 0.0490386001765728, "learning_rate": 9.846717931512573e-05, "loss": 0.0137, "step": 886 }, { "epoch": 4.285024154589372, "grad_norm": 0.05706576630473137, "learning_rate": 9.845888056774354e-05, "loss": 0.0119, "step": 887 }, { "epoch": 4.2898550724637685, "grad_norm": 0.056531455367803574, "learning_rate": 9.845055976771919e-05, "loss": 0.0118, "step": 888 }, { "epoch": 4.294685990338165, "grad_norm": 0.060446158051490784, "learning_rate": 9.844221691883929e-05, "loss": 0.0138, "step": 889 }, { "epoch": 4.29951690821256, "grad_norm": 0.056813742965459824, "learning_rate": 9.843385202490051e-05, "loss": 0.0111, "step": 890 }, { "epoch": 4.304347826086957, "grad_norm": 0.07757464796304703, "learning_rate": 9.842546508970955e-05, "loss": 0.0117, "step": 891 }, { "epoch": 4.309178743961352, "grad_norm": 0.07340128719806671, "learning_rate": 9.841705611708311e-05, "loss": 0.0188, "step": 892 }, { "epoch": 4.314009661835748, "grad_norm": 0.07141716778278351, "learning_rate": 9.840862511084798e-05, "loss": 0.0136, "step": 893 }, { "epoch": 4.318840579710145, "grad_norm": 0.07496558129787445, "learning_rate": 9.840017207484089e-05, "loss": 0.0135, "step": 894 }, { "epoch": 4.323671497584541, "grad_norm": 0.11215018481016159, "learning_rate": 9.839169701290868e-05, "loss": 0.0123, "step": 895 }, { "epoch": 4.328502415458937, "grad_norm": 0.07307403534650803, "learning_rate": 9.838319992890816e-05, "loss": 0.0121, "step": 896 }, { "epoch": 4.333333333333333, "grad_norm": 0.06442231684923172, "learning_rate": 9.837468082670617e-05, "loss": 0.0122, "step": 897 }, { "epoch": 4.338164251207729, "grad_norm": 0.07327039539813995, "learning_rate": 9.836613971017961e-05, "loss": 0.0131, "step": 898 }, { "epoch": 4.342995169082125, "grad_norm": 0.06935934722423553, "learning_rate": 9.835757658321533e-05, "loss": 0.0121, "step": 899 }, { "epoch": 4.3478260869565215, "grad_norm": 0.05545048043131828, "learning_rate": 9.834899144971025e-05, "loss": 0.0119, "step": 900 }, { "epoch": 4.3478260869565215, "eval_loss": 0.020673103630542755, "eval_runtime": 20.6013, "eval_samples_per_second": 4.854, "eval_steps_per_second": 0.146, "step": 900 }, { "epoch": 4.352657004830918, "grad_norm": 0.0609741285443306, "learning_rate": 9.834038431357129e-05, "loss": 0.0115, "step": 901 }, { "epoch": 4.357487922705314, "grad_norm": 0.04639492928981781, "learning_rate": 9.833175517871538e-05, "loss": 0.0086, "step": 902 }, { "epoch": 4.36231884057971, "grad_norm": 0.06929849088191986, "learning_rate": 9.832310404906946e-05, "loss": 0.0126, "step": 903 }, { "epoch": 4.367149758454106, "grad_norm": 0.07134636491537094, "learning_rate": 9.831443092857049e-05, "loss": 0.0147, "step": 904 }, { "epoch": 4.371980676328502, "grad_norm": 0.05646871030330658, "learning_rate": 9.830573582116542e-05, "loss": 0.0126, "step": 905 }, { "epoch": 4.3768115942028984, "grad_norm": 0.05775350332260132, "learning_rate": 9.829701873081122e-05, "loss": 0.0114, "step": 906 }, { "epoch": 4.381642512077295, "grad_norm": 0.07218906283378601, "learning_rate": 9.828827966147485e-05, "loss": 0.0173, "step": 907 }, { "epoch": 4.386473429951691, "grad_norm": 0.04605533555150032, "learning_rate": 9.827951861713329e-05, "loss": 0.0127, "step": 908 }, { "epoch": 4.391304347826087, "grad_norm": 1.5717793703079224, "learning_rate": 9.827073560177351e-05, "loss": 0.0215, "step": 909 }, { "epoch": 4.396135265700483, "grad_norm": 0.05545201897621155, "learning_rate": 9.826193061939249e-05, "loss": 0.0182, "step": 910 }, { "epoch": 4.400966183574879, "grad_norm": 0.07933639734983444, "learning_rate": 9.825310367399716e-05, "loss": 0.017, "step": 911 }, { "epoch": 4.405797101449275, "grad_norm": 0.23300503194332123, "learning_rate": 9.824425476960453e-05, "loss": 0.0348, "step": 912 }, { "epoch": 4.4106280193236715, "grad_norm": 0.08851808309555054, "learning_rate": 9.823538391024151e-05, "loss": 0.0244, "step": 913 }, { "epoch": 4.415458937198068, "grad_norm": 0.23018939793109894, "learning_rate": 9.822649109994508e-05, "loss": 0.0213, "step": 914 }, { "epoch": 4.420289855072464, "grad_norm": 0.06539275497198105, "learning_rate": 9.821757634276217e-05, "loss": 0.0216, "step": 915 }, { "epoch": 4.42512077294686, "grad_norm": 0.08064073324203491, "learning_rate": 9.820863964274969e-05, "loss": 0.0235, "step": 916 }, { "epoch": 4.429951690821256, "grad_norm": 0.09684203565120697, "learning_rate": 9.819968100397455e-05, "loss": 0.02, "step": 917 }, { "epoch": 4.434782608695652, "grad_norm": 0.0828334391117096, "learning_rate": 9.819070043051366e-05, "loss": 0.0179, "step": 918 }, { "epoch": 4.4396135265700485, "grad_norm": 0.08924815058708191, "learning_rate": 9.818169792645388e-05, "loss": 0.0188, "step": 919 }, { "epoch": 4.444444444444445, "grad_norm": 0.08673924207687378, "learning_rate": 9.817267349589205e-05, "loss": 0.0159, "step": 920 }, { "epoch": 4.449275362318841, "grad_norm": 0.08189379423856735, "learning_rate": 9.816362714293504e-05, "loss": 0.0125, "step": 921 }, { "epoch": 4.454106280193237, "grad_norm": 0.09734530746936798, "learning_rate": 9.815455887169965e-05, "loss": 0.0206, "step": 922 }, { "epoch": 4.458937198067633, "grad_norm": 0.12248886376619339, "learning_rate": 9.814546868631264e-05, "loss": 0.0171, "step": 923 }, { "epoch": 4.463768115942029, "grad_norm": 0.15595419704914093, "learning_rate": 9.813635659091078e-05, "loss": 0.0158, "step": 924 }, { "epoch": 4.468599033816425, "grad_norm": 0.07428774982690811, "learning_rate": 9.81272225896408e-05, "loss": 0.0131, "step": 925 }, { "epoch": 4.468599033816425, "eval_loss": 0.025713801383972168, "eval_runtime": 21.3609, "eval_samples_per_second": 4.681, "eval_steps_per_second": 0.14, "step": 925 }, { "epoch": 4.473429951690822, "grad_norm": 0.09882636368274689, "learning_rate": 9.81180666866594e-05, "loss": 0.0176, "step": 926 }, { "epoch": 4.478260869565218, "grad_norm": 0.07712045311927795, "learning_rate": 9.810888888613323e-05, "loss": 0.014, "step": 927 }, { "epoch": 4.483091787439614, "grad_norm": 0.10221463441848755, "learning_rate": 9.809968919223893e-05, "loss": 0.0204, "step": 928 }, { "epoch": 4.48792270531401, "grad_norm": 0.07028807699680328, "learning_rate": 9.809046760916309e-05, "loss": 0.0123, "step": 929 }, { "epoch": 4.492753623188406, "grad_norm": 0.08684271574020386, "learning_rate": 9.808122414110228e-05, "loss": 0.0147, "step": 930 }, { "epoch": 4.4975845410628015, "grad_norm": 0.08106634020805359, "learning_rate": 9.807195879226298e-05, "loss": 0.0167, "step": 931 }, { "epoch": 4.5024154589371985, "grad_norm": 0.10349582135677338, "learning_rate": 9.806267156686169e-05, "loss": 0.0139, "step": 932 }, { "epoch": 4.507246376811594, "grad_norm": 0.08159704506397247, "learning_rate": 9.80533624691248e-05, "loss": 0.0131, "step": 933 }, { "epoch": 4.512077294685991, "grad_norm": 0.07411141693592072, "learning_rate": 9.804403150328872e-05, "loss": 0.0145, "step": 934 }, { "epoch": 4.516908212560386, "grad_norm": 0.08534754812717438, "learning_rate": 9.803467867359977e-05, "loss": 0.0164, "step": 935 }, { "epoch": 4.521739130434782, "grad_norm": 0.07628322392702103, "learning_rate": 9.802530398431424e-05, "loss": 0.0154, "step": 936 }, { "epoch": 4.526570048309178, "grad_norm": 0.06755129992961884, "learning_rate": 9.801590743969835e-05, "loss": 0.013, "step": 937 }, { "epoch": 4.531400966183575, "grad_norm": 0.07023568451404572, "learning_rate": 9.800648904402827e-05, "loss": 0.0143, "step": 938 }, { "epoch": 4.536231884057971, "grad_norm": 0.11462641507387161, "learning_rate": 9.799704880159013e-05, "loss": 0.0223, "step": 939 }, { "epoch": 4.541062801932367, "grad_norm": 0.07971464842557907, "learning_rate": 9.798758671668e-05, "loss": 0.0144, "step": 940 }, { "epoch": 4.545893719806763, "grad_norm": 0.07792872935533524, "learning_rate": 9.797810279360385e-05, "loss": 0.0141, "step": 941 }, { "epoch": 4.550724637681159, "grad_norm": 0.06341885030269623, "learning_rate": 9.796859703667762e-05, "loss": 0.0141, "step": 942 }, { "epoch": 4.555555555555555, "grad_norm": 0.0954955667257309, "learning_rate": 9.795906945022721e-05, "loss": 0.0171, "step": 943 }, { "epoch": 4.5603864734299515, "grad_norm": 0.08577444404363632, "learning_rate": 9.794952003858844e-05, "loss": 0.0171, "step": 944 }, { "epoch": 4.565217391304348, "grad_norm": 0.07980582863092422, "learning_rate": 9.793994880610702e-05, "loss": 0.0163, "step": 945 }, { "epoch": 4.570048309178744, "grad_norm": 0.0770900696516037, "learning_rate": 9.793035575713862e-05, "loss": 0.0153, "step": 946 }, { "epoch": 4.57487922705314, "grad_norm": 0.07175513356924057, "learning_rate": 9.792074089604886e-05, "loss": 0.0132, "step": 947 }, { "epoch": 4.579710144927536, "grad_norm": 0.07475357502698898, "learning_rate": 9.791110422721326e-05, "loss": 0.0116, "step": 948 }, { "epoch": 4.584541062801932, "grad_norm": 0.07414525747299194, "learning_rate": 9.790144575501724e-05, "loss": 0.0175, "step": 949 }, { "epoch": 4.5893719806763285, "grad_norm": 0.058295838534832, "learning_rate": 9.789176548385624e-05, "loss": 0.0123, "step": 950 }, { "epoch": 4.5893719806763285, "eval_loss": 0.02279841899871826, "eval_runtime": 20.6131, "eval_samples_per_second": 4.851, "eval_steps_per_second": 0.146, "step": 950 }, { "epoch": 4.594202898550725, "grad_norm": 0.061069000512361526, "learning_rate": 9.78820634181355e-05, "loss": 0.0099, "step": 951 }, { "epoch": 4.599033816425121, "grad_norm": 0.0968296080827713, "learning_rate": 9.787233956227023e-05, "loss": 0.0177, "step": 952 }, { "epoch": 4.603864734299517, "grad_norm": 0.07276859134435654, "learning_rate": 9.786259392068559e-05, "loss": 0.0124, "step": 953 }, { "epoch": 4.608695652173913, "grad_norm": 0.08032967150211334, "learning_rate": 9.78528264978166e-05, "loss": 0.0121, "step": 954 }, { "epoch": 4.613526570048309, "grad_norm": 0.09521745890378952, "learning_rate": 9.78430372981082e-05, "loss": 0.0122, "step": 955 }, { "epoch": 4.618357487922705, "grad_norm": 0.08671263605356216, "learning_rate": 9.783322632601529e-05, "loss": 0.0132, "step": 956 }, { "epoch": 4.6231884057971016, "grad_norm": 0.09144391119480133, "learning_rate": 9.78233935860026e-05, "loss": 0.0214, "step": 957 }, { "epoch": 4.628019323671498, "grad_norm": 0.06860186159610748, "learning_rate": 9.781353908254485e-05, "loss": 0.0141, "step": 958 }, { "epoch": 4.632850241545894, "grad_norm": 0.05559476092457771, "learning_rate": 9.78036628201266e-05, "loss": 0.0123, "step": 959 }, { "epoch": 4.63768115942029, "grad_norm": 0.06625663489103317, "learning_rate": 9.779376480324233e-05, "loss": 0.0134, "step": 960 }, { "epoch": 4.642512077294686, "grad_norm": 0.10225360840559006, "learning_rate": 9.778384503639644e-05, "loss": 0.0237, "step": 961 }, { "epoch": 4.647342995169082, "grad_norm": 0.06415294110774994, "learning_rate": 9.777390352410319e-05, "loss": 0.0184, "step": 962 }, { "epoch": 4.6521739130434785, "grad_norm": 0.057260725647211075, "learning_rate": 9.776394027088678e-05, "loss": 0.0131, "step": 963 }, { "epoch": 4.657004830917875, "grad_norm": 0.07303762435913086, "learning_rate": 9.775395528128128e-05, "loss": 0.0139, "step": 964 }, { "epoch": 4.661835748792271, "grad_norm": 0.059439219534397125, "learning_rate": 9.774394855983065e-05, "loss": 0.0125, "step": 965 }, { "epoch": 4.666666666666667, "grad_norm": 0.07815828174352646, "learning_rate": 9.773392011108873e-05, "loss": 0.015, "step": 966 }, { "epoch": 4.671497584541063, "grad_norm": 0.0615699328482151, "learning_rate": 9.772386993961928e-05, "loss": 0.013, "step": 967 }, { "epoch": 4.676328502415459, "grad_norm": 0.05981515347957611, "learning_rate": 9.771379804999592e-05, "loss": 0.0123, "step": 968 }, { "epoch": 4.681159420289855, "grad_norm": 0.0661519393324852, "learning_rate": 9.770370444680217e-05, "loss": 0.0127, "step": 969 }, { "epoch": 4.685990338164252, "grad_norm": 0.05894894897937775, "learning_rate": 9.769358913463141e-05, "loss": 0.0098, "step": 970 }, { "epoch": 4.690821256038648, "grad_norm": 0.06429290771484375, "learning_rate": 9.76834521180869e-05, "loss": 0.0125, "step": 971 }, { "epoch": 4.695652173913043, "grad_norm": 0.06996926665306091, "learning_rate": 9.767329340178179e-05, "loss": 0.0141, "step": 972 }, { "epoch": 4.70048309178744, "grad_norm": 0.060817550867795944, "learning_rate": 9.766311299033913e-05, "loss": 0.0127, "step": 973 }, { "epoch": 4.705314009661835, "grad_norm": 0.06457247585058212, "learning_rate": 9.76529108883918e-05, "loss": 0.0131, "step": 974 }, { "epoch": 4.710144927536232, "grad_norm": 0.07117529958486557, "learning_rate": 9.764268710058258e-05, "loss": 0.0133, "step": 975 }, { "epoch": 4.710144927536232, "eval_loss": 0.020405687391757965, "eval_runtime": 20.6298, "eval_samples_per_second": 4.847, "eval_steps_per_second": 0.145, "step": 975 }, { "epoch": 4.714975845410628, "grad_norm": 0.06411249190568924, "learning_rate": 9.763244163156407e-05, "loss": 0.0125, "step": 976 }, { "epoch": 4.719806763285024, "grad_norm": 0.0879925787448883, "learning_rate": 9.762217448599882e-05, "loss": 0.0187, "step": 977 }, { "epoch": 4.72463768115942, "grad_norm": 0.05720191076397896, "learning_rate": 9.761188566855916e-05, "loss": 0.0106, "step": 978 }, { "epoch": 4.729468599033816, "grad_norm": 0.06640984117984772, "learning_rate": 9.760157518392735e-05, "loss": 0.0113, "step": 979 }, { "epoch": 4.734299516908212, "grad_norm": 0.05869889631867409, "learning_rate": 9.759124303679544e-05, "loss": 0.011, "step": 980 }, { "epoch": 4.739130434782608, "grad_norm": 0.052954427897930145, "learning_rate": 9.758088923186541e-05, "loss": 0.0115, "step": 981 }, { "epoch": 4.743961352657005, "grad_norm": 0.049915559589862823, "learning_rate": 9.757051377384906e-05, "loss": 0.0114, "step": 982 }, { "epoch": 4.748792270531401, "grad_norm": 0.07832711189985275, "learning_rate": 9.756011666746805e-05, "loss": 0.015, "step": 983 }, { "epoch": 4.753623188405797, "grad_norm": 0.06998534500598907, "learning_rate": 9.754969791745387e-05, "loss": 0.0132, "step": 984 }, { "epoch": 4.758454106280193, "grad_norm": 0.06319990009069443, "learning_rate": 9.75392575285479e-05, "loss": 0.0117, "step": 985 }, { "epoch": 4.763285024154589, "grad_norm": 0.04931075870990753, "learning_rate": 9.752879550550133e-05, "loss": 0.0108, "step": 986 }, { "epoch": 4.768115942028985, "grad_norm": 0.04907230660319328, "learning_rate": 9.751831185307523e-05, "loss": 0.0098, "step": 987 }, { "epoch": 4.7729468599033815, "grad_norm": 0.06535334140062332, "learning_rate": 9.750780657604051e-05, "loss": 0.0144, "step": 988 }, { "epoch": 4.777777777777778, "grad_norm": 0.08152740448713303, "learning_rate": 9.749727967917785e-05, "loss": 0.0126, "step": 989 }, { "epoch": 4.782608695652174, "grad_norm": 0.055403295904397964, "learning_rate": 9.748673116727787e-05, "loss": 0.0134, "step": 990 }, { "epoch": 4.78743961352657, "grad_norm": 0.0632065013051033, "learning_rate": 9.747616104514097e-05, "loss": 0.0123, "step": 991 }, { "epoch": 4.792270531400966, "grad_norm": 0.058326445519924164, "learning_rate": 9.74655693175774e-05, "loss": 0.0119, "step": 992 }, { "epoch": 4.797101449275362, "grad_norm": 0.04671332985162735, "learning_rate": 9.745495598940722e-05, "loss": 0.0094, "step": 993 }, { "epoch": 4.8019323671497585, "grad_norm": 0.058310024440288544, "learning_rate": 9.744432106546035e-05, "loss": 0.0103, "step": 994 }, { "epoch": 4.806763285024155, "grad_norm": 0.061691973358392715, "learning_rate": 9.743366455057654e-05, "loss": 0.0129, "step": 995 }, { "epoch": 4.811594202898551, "grad_norm": 0.0760740339756012, "learning_rate": 9.742298644960533e-05, "loss": 0.0154, "step": 996 }, { "epoch": 4.816425120772947, "grad_norm": 0.06041797623038292, "learning_rate": 9.741228676740611e-05, "loss": 0.013, "step": 997 }, { "epoch": 4.821256038647343, "grad_norm": 0.06114066019654274, "learning_rate": 9.74015655088481e-05, "loss": 0.0069, "step": 998 }, { "epoch": 4.826086956521739, "grad_norm": 0.11095196008682251, "learning_rate": 9.739082267881029e-05, "loss": 0.0128, "step": 999 }, { "epoch": 4.830917874396135, "grad_norm": 0.058319080621004105, "learning_rate": 9.738005828218155e-05, "loss": 0.0115, "step": 1000 }, { "epoch": 4.830917874396135, "eval_loss": 0.019139988347887993, "eval_runtime": 20.6148, "eval_samples_per_second": 4.851, "eval_steps_per_second": 0.146, "step": 1000 }, { "epoch": 4.835748792270532, "grad_norm": 0.06817466765642166, "learning_rate": 9.736927232386055e-05, "loss": 0.0156, "step": 1001 }, { "epoch": 4.840579710144928, "grad_norm": 0.05912366136908531, "learning_rate": 9.735846480875572e-05, "loss": 0.0144, "step": 1002 }, { "epoch": 4.845410628019324, "grad_norm": 0.08707557618618011, "learning_rate": 9.734763574178536e-05, "loss": 0.0128, "step": 1003 }, { "epoch": 4.85024154589372, "grad_norm": 0.08672689646482468, "learning_rate": 9.733678512787757e-05, "loss": 0.0122, "step": 1004 }, { "epoch": 4.855072463768116, "grad_norm": 0.06592126190662384, "learning_rate": 9.732591297197021e-05, "loss": 0.0163, "step": 1005 }, { "epoch": 4.859903381642512, "grad_norm": 0.0610274113714695, "learning_rate": 9.731501927901096e-05, "loss": 0.0146, "step": 1006 }, { "epoch": 4.8647342995169085, "grad_norm": 0.05984511226415634, "learning_rate": 9.730410405395739e-05, "loss": 0.0112, "step": 1007 }, { "epoch": 4.869565217391305, "grad_norm": 0.04920889064669609, "learning_rate": 9.729316730177674e-05, "loss": 0.0091, "step": 1008 }, { "epoch": 4.874396135265701, "grad_norm": 0.07654332369565964, "learning_rate": 9.72822090274461e-05, "loss": 0.012, "step": 1009 }, { "epoch": 4.879227053140097, "grad_norm": 0.04907603561878204, "learning_rate": 9.727122923595238e-05, "loss": 0.0112, "step": 1010 }, { "epoch": 4.884057971014493, "grad_norm": 0.04914016276597977, "learning_rate": 9.726022793229224e-05, "loss": 0.0105, "step": 1011 }, { "epoch": 4.888888888888889, "grad_norm": 0.055787090212106705, "learning_rate": 9.724920512147215e-05, "loss": 0.0138, "step": 1012 }, { "epoch": 4.8937198067632846, "grad_norm": 0.06109369918704033, "learning_rate": 9.723816080850839e-05, "loss": 0.0113, "step": 1013 }, { "epoch": 4.898550724637682, "grad_norm": 0.0676606222987175, "learning_rate": 9.722709499842695e-05, "loss": 0.0123, "step": 1014 }, { "epoch": 4.903381642512077, "grad_norm": 0.05014638230204582, "learning_rate": 9.721600769626368e-05, "loss": 0.013, "step": 1015 }, { "epoch": 4.908212560386474, "grad_norm": 0.07281427085399628, "learning_rate": 9.720489890706421e-05, "loss": 0.0152, "step": 1016 }, { "epoch": 4.913043478260869, "grad_norm": 0.08129395544528961, "learning_rate": 9.71937686358839e-05, "loss": 0.0194, "step": 1017 }, { "epoch": 4.917874396135265, "grad_norm": 0.07277390360832214, "learning_rate": 9.718261688778789e-05, "loss": 0.0134, "step": 1018 }, { "epoch": 4.9227053140096615, "grad_norm": 0.06831222772598267, "learning_rate": 9.717144366785112e-05, "loss": 0.0126, "step": 1019 }, { "epoch": 4.927536231884058, "grad_norm": 0.059769704937934875, "learning_rate": 9.716024898115832e-05, "loss": 0.015, "step": 1020 }, { "epoch": 4.932367149758454, "grad_norm": 0.07310938090085983, "learning_rate": 9.714903283280393e-05, "loss": 0.0125, "step": 1021 }, { "epoch": 4.93719806763285, "grad_norm": 0.06537635624408722, "learning_rate": 9.71377952278922e-05, "loss": 0.0113, "step": 1022 }, { "epoch": 4.942028985507246, "grad_norm": 0.07284176349639893, "learning_rate": 9.712653617153715e-05, "loss": 0.0131, "step": 1023 }, { "epoch": 4.946859903381642, "grad_norm": 0.0574897825717926, "learning_rate": 9.71152556688625e-05, "loss": 0.0109, "step": 1024 }, { "epoch": 4.951690821256038, "grad_norm": 0.07543810456991196, "learning_rate": 9.710395372500183e-05, "loss": 0.0152, "step": 1025 }, { "epoch": 4.951690821256038, "eval_loss": 0.020091358572244644, "eval_runtime": 20.6004, "eval_samples_per_second": 4.854, "eval_steps_per_second": 0.146, "step": 1025 }, { "epoch": 4.956521739130435, "grad_norm": 0.08204186707735062, "learning_rate": 9.70926303450984e-05, "loss": 0.0118, "step": 1026 }, { "epoch": 4.961352657004831, "grad_norm": 0.06970630586147308, "learning_rate": 9.708128553430524e-05, "loss": 0.0109, "step": 1027 }, { "epoch": 4.966183574879227, "grad_norm": 0.06959115713834763, "learning_rate": 9.706991929778516e-05, "loss": 0.0133, "step": 1028 }, { "epoch": 4.971014492753623, "grad_norm": 0.05290956422686577, "learning_rate": 9.705853164071065e-05, "loss": 0.0113, "step": 1029 }, { "epoch": 4.975845410628019, "grad_norm": 0.08411956578493118, "learning_rate": 9.704712256826407e-05, "loss": 0.0173, "step": 1030 }, { "epoch": 4.980676328502415, "grad_norm": 0.05763190984725952, "learning_rate": 9.703569208563741e-05, "loss": 0.0105, "step": 1031 }, { "epoch": 4.9855072463768115, "grad_norm": 0.07256191968917847, "learning_rate": 9.702424019803247e-05, "loss": 0.0163, "step": 1032 }, { "epoch": 4.990338164251208, "grad_norm": 0.052556559443473816, "learning_rate": 9.701276691066075e-05, "loss": 0.0122, "step": 1033 }, { "epoch": 4.995169082125604, "grad_norm": 0.06209128350019455, "learning_rate": 9.700127222874351e-05, "loss": 0.0155, "step": 1034 }, { "epoch": 5.0, "grad_norm": 0.07931708544492722, "learning_rate": 9.698975615751176e-05, "loss": 0.0092, "step": 1035 }, { "epoch": 5.004830917874396, "grad_norm": 0.06323495507240295, "learning_rate": 9.697821870220623e-05, "loss": 0.0101, "step": 1036 }, { "epoch": 5.009661835748792, "grad_norm": 0.057949360460042953, "learning_rate": 9.696665986807734e-05, "loss": 0.0108, "step": 1037 }, { "epoch": 5.0144927536231885, "grad_norm": 0.04627874121069908, "learning_rate": 9.695507966038533e-05, "loss": 0.0079, "step": 1038 }, { "epoch": 5.019323671497585, "grad_norm": 0.05395060405135155, "learning_rate": 9.694347808440007e-05, "loss": 0.0145, "step": 1039 }, { "epoch": 5.024154589371981, "grad_norm": 0.0515243224799633, "learning_rate": 9.693185514540122e-05, "loss": 0.0089, "step": 1040 }, { "epoch": 5.028985507246377, "grad_norm": 0.07171349972486496, "learning_rate": 9.692021084867814e-05, "loss": 0.0079, "step": 1041 }, { "epoch": 5.033816425120773, "grad_norm": 0.0560278482735157, "learning_rate": 9.690854519952995e-05, "loss": 0.0081, "step": 1042 }, { "epoch": 5.038647342995169, "grad_norm": 0.059097371995449066, "learning_rate": 9.689685820326537e-05, "loss": 0.0107, "step": 1043 }, { "epoch": 5.043478260869565, "grad_norm": 0.055744510143995285, "learning_rate": 9.688514986520296e-05, "loss": 0.0087, "step": 1044 }, { "epoch": 5.048309178743962, "grad_norm": 0.0558600053191185, "learning_rate": 9.687342019067093e-05, "loss": 0.0098, "step": 1045 }, { "epoch": 5.053140096618358, "grad_norm": 0.06191682815551758, "learning_rate": 9.686166918500723e-05, "loss": 0.0116, "step": 1046 }, { "epoch": 5.057971014492754, "grad_norm": 0.06881178915500641, "learning_rate": 9.684989685355947e-05, "loss": 0.0075, "step": 1047 }, { "epoch": 5.06280193236715, "grad_norm": 0.05575813725590706, "learning_rate": 9.683810320168504e-05, "loss": 0.0069, "step": 1048 }, { "epoch": 5.067632850241546, "grad_norm": 0.08796427398920059, "learning_rate": 9.682628823475095e-05, "loss": 0.0084, "step": 1049 }, { "epoch": 5.072463768115942, "grad_norm": 0.04683411493897438, "learning_rate": 9.681445195813399e-05, "loss": 0.0075, "step": 1050 }, { "epoch": 5.072463768115942, "eval_loss": 0.018816815689206123, "eval_runtime": 20.6451, "eval_samples_per_second": 4.844, "eval_steps_per_second": 0.145, "step": 1050 }, { "epoch": 5.0772946859903385, "grad_norm": 0.07573356479406357, "learning_rate": 9.680259437722055e-05, "loss": 0.0077, "step": 1051 }, { "epoch": 5.082125603864735, "grad_norm": 0.056269437074661255, "learning_rate": 9.679071549740682e-05, "loss": 0.0093, "step": 1052 }, { "epoch": 5.086956521739131, "grad_norm": 0.06449293345212936, "learning_rate": 9.677881532409863e-05, "loss": 0.0102, "step": 1053 }, { "epoch": 5.091787439613527, "grad_norm": 0.047550760209560394, "learning_rate": 9.676689386271147e-05, "loss": 0.0076, "step": 1054 }, { "epoch": 5.096618357487923, "grad_norm": 0.04935392364859581, "learning_rate": 9.675495111867059e-05, "loss": 0.0061, "step": 1055 }, { "epoch": 5.101449275362318, "grad_norm": 0.05520419031381607, "learning_rate": 9.674298709741087e-05, "loss": 0.01, "step": 1056 }, { "epoch": 5.106280193236715, "grad_norm": 0.06270143389701843, "learning_rate": 9.673100180437689e-05, "loss": 0.0083, "step": 1057 }, { "epoch": 5.111111111111111, "grad_norm": 0.06183341518044472, "learning_rate": 9.671899524502291e-05, "loss": 0.006, "step": 1058 }, { "epoch": 5.115942028985507, "grad_norm": 0.06538157910108566, "learning_rate": 9.670696742481289e-05, "loss": 0.0087, "step": 1059 }, { "epoch": 5.120772946859903, "grad_norm": 0.06029185280203819, "learning_rate": 9.66949183492204e-05, "loss": 0.0111, "step": 1060 }, { "epoch": 5.125603864734299, "grad_norm": 0.06412973999977112, "learning_rate": 9.668284802372877e-05, "loss": 0.0089, "step": 1061 }, { "epoch": 5.130434782608695, "grad_norm": 0.07017388939857483, "learning_rate": 9.667075645383095e-05, "loss": 0.0088, "step": 1062 }, { "epoch": 5.1352657004830915, "grad_norm": 0.07803435623645782, "learning_rate": 9.665864364502955e-05, "loss": 0.0095, "step": 1063 }, { "epoch": 5.140096618357488, "grad_norm": 0.05609812214970589, "learning_rate": 9.664650960283688e-05, "loss": 0.0058, "step": 1064 }, { "epoch": 5.144927536231884, "grad_norm": 0.052468396723270416, "learning_rate": 9.663435433277486e-05, "loss": 0.0087, "step": 1065 }, { "epoch": 5.14975845410628, "grad_norm": 0.06423533707857132, "learning_rate": 9.662217784037513e-05, "loss": 0.008, "step": 1066 }, { "epoch": 5.154589371980676, "grad_norm": 0.061567891389131546, "learning_rate": 9.660998013117897e-05, "loss": 0.0126, "step": 1067 }, { "epoch": 5.159420289855072, "grad_norm": 0.053131893277168274, "learning_rate": 9.659776121073729e-05, "loss": 0.0077, "step": 1068 }, { "epoch": 5.164251207729468, "grad_norm": 0.07207377254962921, "learning_rate": 9.658552108461068e-05, "loss": 0.0106, "step": 1069 }, { "epoch": 5.169082125603865, "grad_norm": 0.0715598464012146, "learning_rate": 9.657325975836935e-05, "loss": 0.0113, "step": 1070 }, { "epoch": 5.173913043478261, "grad_norm": 0.050884295254945755, "learning_rate": 9.656097723759319e-05, "loss": 0.0075, "step": 1071 }, { "epoch": 5.178743961352657, "grad_norm": 0.049650318920612335, "learning_rate": 9.654867352787174e-05, "loss": 0.0092, "step": 1072 }, { "epoch": 5.183574879227053, "grad_norm": 0.06503750383853912, "learning_rate": 9.653634863480414e-05, "loss": 0.0107, "step": 1073 }, { "epoch": 5.188405797101449, "grad_norm": 0.05913107097148895, "learning_rate": 9.652400256399922e-05, "loss": 0.0088, "step": 1074 }, { "epoch": 5.193236714975845, "grad_norm": 0.05056409910321236, "learning_rate": 9.651163532107541e-05, "loss": 0.0069, "step": 1075 }, { "epoch": 5.193236714975845, "eval_loss": 0.016926271840929985, "eval_runtime": 20.5906, "eval_samples_per_second": 4.857, "eval_steps_per_second": 0.146, "step": 1075 }, { "epoch": 5.1980676328502415, "grad_norm": 0.056902553886175156, "learning_rate": 9.649924691166078e-05, "loss": 0.0087, "step": 1076 }, { "epoch": 5.202898550724638, "grad_norm": 0.09051110595464706, "learning_rate": 9.648683734139305e-05, "loss": 0.0108, "step": 1077 }, { "epoch": 5.207729468599034, "grad_norm": 0.06023527309298515, "learning_rate": 9.647440661591958e-05, "loss": 0.01, "step": 1078 }, { "epoch": 5.21256038647343, "grad_norm": 0.10421982407569885, "learning_rate": 9.64619547408973e-05, "loss": 0.0136, "step": 1079 }, { "epoch": 5.217391304347826, "grad_norm": 0.09369756281375885, "learning_rate": 9.644948172199283e-05, "loss": 0.0095, "step": 1080 }, { "epoch": 5.222222222222222, "grad_norm": 0.062193822115659714, "learning_rate": 9.643698756488238e-05, "loss": 0.009, "step": 1081 }, { "epoch": 5.2270531400966185, "grad_norm": 0.04095260426402092, "learning_rate": 9.642447227525178e-05, "loss": 0.0072, "step": 1082 }, { "epoch": 5.231884057971015, "grad_norm": 0.06756602972745895, "learning_rate": 9.641193585879646e-05, "loss": 0.0097, "step": 1083 }, { "epoch": 5.236714975845411, "grad_norm": 0.05496843904256821, "learning_rate": 9.639937832122154e-05, "loss": 0.0076, "step": 1084 }, { "epoch": 5.241545893719807, "grad_norm": 0.055532056838274, "learning_rate": 9.638679966824165e-05, "loss": 0.0076, "step": 1085 }, { "epoch": 5.246376811594203, "grad_norm": 0.04633563756942749, "learning_rate": 9.637419990558111e-05, "loss": 0.0066, "step": 1086 }, { "epoch": 5.251207729468599, "grad_norm": 0.06235886737704277, "learning_rate": 9.636157903897379e-05, "loss": 0.0123, "step": 1087 }, { "epoch": 5.256038647342995, "grad_norm": 0.06465509533882141, "learning_rate": 9.63489370741632e-05, "loss": 0.0127, "step": 1088 }, { "epoch": 5.260869565217392, "grad_norm": 0.08732569962739944, "learning_rate": 9.633627401690241e-05, "loss": 0.013, "step": 1089 }, { "epoch": 5.265700483091788, "grad_norm": 0.059913862496614456, "learning_rate": 9.632358987295416e-05, "loss": 0.009, "step": 1090 }, { "epoch": 5.270531400966184, "grad_norm": 0.05865868553519249, "learning_rate": 9.631088464809072e-05, "loss": 0.008, "step": 1091 }, { "epoch": 5.27536231884058, "grad_norm": 0.07305967807769775, "learning_rate": 9.629815834809397e-05, "loss": 0.0099, "step": 1092 }, { "epoch": 5.280193236714976, "grad_norm": 0.060723256319761276, "learning_rate": 9.628541097875541e-05, "loss": 0.0105, "step": 1093 }, { "epoch": 5.285024154589372, "grad_norm": 0.042938102036714554, "learning_rate": 9.62726425458761e-05, "loss": 0.0086, "step": 1094 }, { "epoch": 5.2898550724637685, "grad_norm": 0.04853806272149086, "learning_rate": 9.625985305526669e-05, "loss": 0.0085, "step": 1095 }, { "epoch": 5.294685990338165, "grad_norm": 0.056957464665174484, "learning_rate": 9.62470425127474e-05, "loss": 0.0077, "step": 1096 }, { "epoch": 5.29951690821256, "grad_norm": 0.05829184129834175, "learning_rate": 9.623421092414806e-05, "loss": 0.0073, "step": 1097 }, { "epoch": 5.304347826086957, "grad_norm": 0.08711043000221252, "learning_rate": 9.622135829530807e-05, "loss": 0.0073, "step": 1098 }, { "epoch": 5.309178743961352, "grad_norm": 0.06256822496652603, "learning_rate": 9.620848463207637e-05, "loss": 0.0084, "step": 1099 }, { "epoch": 5.314009661835748, "grad_norm": 0.04850585386157036, "learning_rate": 9.619558994031152e-05, "loss": 0.0073, "step": 1100 }, { "epoch": 5.314009661835748, "eval_loss": 0.018239127472043037, "eval_runtime": 20.6235, "eval_samples_per_second": 4.849, "eval_steps_per_second": 0.145, "step": 1100 }, { "epoch": 5.318840579710145, "grad_norm": 0.05552690103650093, "learning_rate": 9.618267422588164e-05, "loss": 0.0093, "step": 1101 }, { "epoch": 5.323671497584541, "grad_norm": 0.07265713810920715, "learning_rate": 9.616973749466438e-05, "loss": 0.0089, "step": 1102 }, { "epoch": 5.328502415458937, "grad_norm": 0.06223799288272858, "learning_rate": 9.615677975254701e-05, "loss": 0.0079, "step": 1103 }, { "epoch": 5.333333333333333, "grad_norm": 0.06156628578901291, "learning_rate": 9.614380100542631e-05, "loss": 0.0098, "step": 1104 }, { "epoch": 5.338164251207729, "grad_norm": 0.06282912194728851, "learning_rate": 9.613080125920866e-05, "loss": 0.0083, "step": 1105 }, { "epoch": 5.342995169082125, "grad_norm": 0.052787791937589645, "learning_rate": 9.611778051980995e-05, "loss": 0.0089, "step": 1106 }, { "epoch": 5.3478260869565215, "grad_norm": 0.05192113295197487, "learning_rate": 9.610473879315567e-05, "loss": 0.0076, "step": 1107 }, { "epoch": 5.352657004830918, "grad_norm": 0.04756897687911987, "learning_rate": 9.609167608518084e-05, "loss": 0.0082, "step": 1108 }, { "epoch": 5.357487922705314, "grad_norm": 0.060673873871564865, "learning_rate": 9.607859240183003e-05, "loss": 0.0087, "step": 1109 }, { "epoch": 5.36231884057971, "grad_norm": 0.05874047428369522, "learning_rate": 9.606548774905737e-05, "loss": 0.0099, "step": 1110 }, { "epoch": 5.367149758454106, "grad_norm": 0.05346541106700897, "learning_rate": 9.605236213282649e-05, "loss": 0.0078, "step": 1111 }, { "epoch": 5.371980676328502, "grad_norm": 0.042168211191892624, "learning_rate": 9.603921555911061e-05, "loss": 0.0067, "step": 1112 }, { "epoch": 5.3768115942028984, "grad_norm": 0.0657646432518959, "learning_rate": 9.602604803389246e-05, "loss": 0.0119, "step": 1113 }, { "epoch": 5.381642512077295, "grad_norm": 0.060943853110075, "learning_rate": 9.601285956316429e-05, "loss": 0.0096, "step": 1114 }, { "epoch": 5.386473429951691, "grad_norm": 0.07080808281898499, "learning_rate": 9.599965015292794e-05, "loss": 0.0101, "step": 1115 }, { "epoch": 5.391304347826087, "grad_norm": 0.049159590154886246, "learning_rate": 9.598641980919472e-05, "loss": 0.0077, "step": 1116 }, { "epoch": 5.396135265700483, "grad_norm": 0.04659260809421539, "learning_rate": 9.59731685379855e-05, "loss": 0.0067, "step": 1117 }, { "epoch": 5.400966183574879, "grad_norm": 0.06035299971699715, "learning_rate": 9.595989634533065e-05, "loss": 0.0084, "step": 1118 }, { "epoch": 5.405797101449275, "grad_norm": 0.0531546026468277, "learning_rate": 9.594660323727006e-05, "loss": 0.0092, "step": 1119 }, { "epoch": 5.4106280193236715, "grad_norm": 0.06887521594762802, "learning_rate": 9.593328921985319e-05, "loss": 0.0108, "step": 1120 }, { "epoch": 5.415458937198068, "grad_norm": 0.050380952656269073, "learning_rate": 9.591995429913894e-05, "loss": 0.0094, "step": 1121 }, { "epoch": 5.420289855072464, "grad_norm": 0.06734389066696167, "learning_rate": 9.590659848119577e-05, "loss": 0.0073, "step": 1122 }, { "epoch": 5.42512077294686, "grad_norm": 0.06262669712305069, "learning_rate": 9.589322177210163e-05, "loss": 0.0117, "step": 1123 }, { "epoch": 5.429951690821256, "grad_norm": 0.05138638615608215, "learning_rate": 9.587982417794401e-05, "loss": 0.0067, "step": 1124 }, { "epoch": 5.434782608695652, "grad_norm": 0.05120410770177841, "learning_rate": 9.586640570481986e-05, "loss": 0.0076, "step": 1125 }, { "epoch": 5.434782608695652, "eval_loss": 0.01660805754363537, "eval_runtime": 21.4666, "eval_samples_per_second": 4.658, "eval_steps_per_second": 0.14, "step": 1125 }, { "epoch": 5.4396135265700485, "grad_norm": 0.04313327744603157, "learning_rate": 9.585296635883567e-05, "loss": 0.008, "step": 1126 }, { "epoch": 5.444444444444445, "grad_norm": 0.07868213951587677, "learning_rate": 9.583950614610738e-05, "loss": 0.0076, "step": 1127 }, { "epoch": 5.449275362318841, "grad_norm": 0.04285581782460213, "learning_rate": 9.582602507276048e-05, "loss": 0.0085, "step": 1128 }, { "epoch": 5.454106280193237, "grad_norm": 0.062434256076812744, "learning_rate": 9.581252314492995e-05, "loss": 0.0094, "step": 1129 }, { "epoch": 5.458937198067633, "grad_norm": 0.05328233167529106, "learning_rate": 9.579900036876018e-05, "loss": 0.0078, "step": 1130 }, { "epoch": 5.463768115942029, "grad_norm": 0.08031158894300461, "learning_rate": 9.578545675040516e-05, "loss": 0.0084, "step": 1131 }, { "epoch": 5.468599033816425, "grad_norm": 0.06782043725252151, "learning_rate": 9.57718922960283e-05, "loss": 0.0116, "step": 1132 }, { "epoch": 5.473429951690822, "grad_norm": 0.06157178059220314, "learning_rate": 9.575830701180248e-05, "loss": 0.0107, "step": 1133 }, { "epoch": 5.478260869565218, "grad_norm": 0.05720106512308121, "learning_rate": 9.574470090391012e-05, "loss": 0.0074, "step": 1134 }, { "epoch": 5.483091787439614, "grad_norm": 0.06687933206558228, "learning_rate": 9.573107397854307e-05, "loss": 0.0081, "step": 1135 }, { "epoch": 5.48792270531401, "grad_norm": 0.0551472008228302, "learning_rate": 9.571742624190265e-05, "loss": 0.0103, "step": 1136 }, { "epoch": 5.492753623188406, "grad_norm": 0.05003214254975319, "learning_rate": 9.570375770019968e-05, "loss": 0.0072, "step": 1137 }, { "epoch": 5.4975845410628015, "grad_norm": 0.047905758023262024, "learning_rate": 9.569006835965443e-05, "loss": 0.0072, "step": 1138 }, { "epoch": 5.5024154589371985, "grad_norm": 0.04713456332683563, "learning_rate": 9.567635822649663e-05, "loss": 0.0089, "step": 1139 }, { "epoch": 5.507246376811594, "grad_norm": 0.04646996036171913, "learning_rate": 9.566262730696548e-05, "loss": 0.007, "step": 1140 }, { "epoch": 5.512077294685991, "grad_norm": 0.044961195439100266, "learning_rate": 9.564887560730966e-05, "loss": 0.0079, "step": 1141 }, { "epoch": 5.516908212560386, "grad_norm": 0.051607292145490646, "learning_rate": 9.563510313378725e-05, "loss": 0.0089, "step": 1142 }, { "epoch": 5.521739130434782, "grad_norm": 0.06743261218070984, "learning_rate": 9.562130989266586e-05, "loss": 0.0122, "step": 1143 }, { "epoch": 5.526570048309178, "grad_norm": 0.08333702385425568, "learning_rate": 9.560749589022249e-05, "loss": 0.0095, "step": 1144 }, { "epoch": 5.531400966183575, "grad_norm": 0.08595249801874161, "learning_rate": 9.55936611327436e-05, "loss": 0.01, "step": 1145 }, { "epoch": 5.536231884057971, "grad_norm": 0.06153739243745804, "learning_rate": 9.557980562652513e-05, "loss": 0.0094, "step": 1146 }, { "epoch": 5.541062801932367, "grad_norm": 0.08083317428827286, "learning_rate": 9.556592937787241e-05, "loss": 0.011, "step": 1147 }, { "epoch": 5.545893719806763, "grad_norm": 0.07135039567947388, "learning_rate": 9.555203239310024e-05, "loss": 0.0094, "step": 1148 }, { "epoch": 5.550724637681159, "grad_norm": 0.06862451136112213, "learning_rate": 9.553811467853288e-05, "loss": 0.009, "step": 1149 }, { "epoch": 5.555555555555555, "grad_norm": 0.052251044660806656, "learning_rate": 9.552417624050398e-05, "loss": 0.0084, "step": 1150 }, { "epoch": 5.555555555555555, "eval_loss": 0.01725536584854126, "eval_runtime": 20.6146, "eval_samples_per_second": 4.851, "eval_steps_per_second": 0.146, "step": 1150 }, { "epoch": 5.5603864734299515, "grad_norm": 0.05517849326133728, "learning_rate": 9.55102170853566e-05, "loss": 0.0097, "step": 1151 }, { "epoch": 5.565217391304348, "grad_norm": 0.04879816621541977, "learning_rate": 9.549623721944334e-05, "loss": 0.0097, "step": 1152 }, { "epoch": 5.570048309178744, "grad_norm": 0.05694446712732315, "learning_rate": 9.54822366491261e-05, "loss": 0.0081, "step": 1153 }, { "epoch": 5.57487922705314, "grad_norm": 0.04808945581316948, "learning_rate": 9.546821538077626e-05, "loss": 0.0085, "step": 1154 }, { "epoch": 5.579710144927536, "grad_norm": 0.05540106073021889, "learning_rate": 9.54541734207746e-05, "loss": 0.0102, "step": 1155 }, { "epoch": 5.584541062801932, "grad_norm": 0.052780650556087494, "learning_rate": 9.544011077551136e-05, "loss": 0.0088, "step": 1156 }, { "epoch": 5.5893719806763285, "grad_norm": 0.06642550975084305, "learning_rate": 9.542602745138614e-05, "loss": 0.0087, "step": 1157 }, { "epoch": 5.594202898550725, "grad_norm": 0.06861164420843124, "learning_rate": 9.541192345480795e-05, "loss": 0.0083, "step": 1158 }, { "epoch": 5.599033816425121, "grad_norm": 0.05863761156797409, "learning_rate": 9.539779879219528e-05, "loss": 0.0101, "step": 1159 }, { "epoch": 5.603864734299517, "grad_norm": 0.04983343183994293, "learning_rate": 9.538365346997594e-05, "loss": 0.0077, "step": 1160 }, { "epoch": 5.608695652173913, "grad_norm": 0.06960771232843399, "learning_rate": 9.536948749458718e-05, "loss": 0.0156, "step": 1161 }, { "epoch": 5.613526570048309, "grad_norm": 0.05016974359750748, "learning_rate": 9.535530087247567e-05, "loss": 0.0089, "step": 1162 }, { "epoch": 5.618357487922705, "grad_norm": 0.04743446037173271, "learning_rate": 9.53410936100974e-05, "loss": 0.0062, "step": 1163 }, { "epoch": 5.6231884057971016, "grad_norm": 0.05176220461726189, "learning_rate": 9.532686571391784e-05, "loss": 0.0075, "step": 1164 }, { "epoch": 5.628019323671498, "grad_norm": 0.05748028680682182, "learning_rate": 9.53126171904118e-05, "loss": 0.0099, "step": 1165 }, { "epoch": 5.632850241545894, "grad_norm": 0.05922771245241165, "learning_rate": 9.529834804606351e-05, "loss": 0.009, "step": 1166 }, { "epoch": 5.63768115942029, "grad_norm": 0.0787665918469429, "learning_rate": 9.528405828736656e-05, "loss": 0.015, "step": 1167 }, { "epoch": 5.642512077294686, "grad_norm": 0.05979447066783905, "learning_rate": 9.52697479208239e-05, "loss": 0.0092, "step": 1168 }, { "epoch": 5.647342995169082, "grad_norm": 0.04697132483124733, "learning_rate": 9.525541695294791e-05, "loss": 0.008, "step": 1169 }, { "epoch": 5.6521739130434785, "grad_norm": 0.06020793691277504, "learning_rate": 9.52410653902603e-05, "loss": 0.0133, "step": 1170 }, { "epoch": 5.657004830917875, "grad_norm": 0.0417938232421875, "learning_rate": 9.52266932392922e-05, "loss": 0.0073, "step": 1171 }, { "epoch": 5.661835748792271, "grad_norm": 0.041864678263664246, "learning_rate": 9.521230050658405e-05, "loss": 0.0061, "step": 1172 }, { "epoch": 5.666666666666667, "grad_norm": 0.05600718781352043, "learning_rate": 9.51978871986857e-05, "loss": 0.0104, "step": 1173 }, { "epoch": 5.671497584541063, "grad_norm": 0.09407051652669907, "learning_rate": 9.518345332215637e-05, "loss": 0.0089, "step": 1174 }, { "epoch": 5.676328502415459, "grad_norm": 0.07216166704893112, "learning_rate": 9.516899888356459e-05, "loss": 0.0091, "step": 1175 }, { "epoch": 5.676328502415459, "eval_loss": 0.017482729628682137, "eval_runtime": 20.5935, "eval_samples_per_second": 4.856, "eval_steps_per_second": 0.146, "step": 1175 }, { "epoch": 5.681159420289855, "grad_norm": 0.0422208234667778, "learning_rate": 9.515452388948829e-05, "loss": 0.0066, "step": 1176 }, { "epoch": 5.685990338164252, "grad_norm": 0.06615504622459412, "learning_rate": 9.514002834651474e-05, "loss": 0.0099, "step": 1177 }, { "epoch": 5.690821256038648, "grad_norm": 0.05418713763356209, "learning_rate": 9.512551226124058e-05, "loss": 0.0089, "step": 1178 }, { "epoch": 5.695652173913043, "grad_norm": 0.060746192932128906, "learning_rate": 9.511097564027175e-05, "loss": 0.0132, "step": 1179 }, { "epoch": 5.70048309178744, "grad_norm": 0.05127759650349617, "learning_rate": 9.509641849022362e-05, "loss": 0.0097, "step": 1180 }, { "epoch": 5.705314009661835, "grad_norm": 0.05560661479830742, "learning_rate": 9.508184081772081e-05, "loss": 0.0098, "step": 1181 }, { "epoch": 5.710144927536232, "grad_norm": 0.062068238854408264, "learning_rate": 9.506724262939732e-05, "loss": 0.0075, "step": 1182 }, { "epoch": 5.714975845410628, "grad_norm": 0.061363585293293, "learning_rate": 9.505262393189649e-05, "loss": 0.0071, "step": 1183 }, { "epoch": 5.719806763285024, "grad_norm": 0.05139882117509842, "learning_rate": 9.5037984731871e-05, "loss": 0.009, "step": 1184 }, { "epoch": 5.72463768115942, "grad_norm": 0.05671796575188637, "learning_rate": 9.502332503598283e-05, "loss": 0.0091, "step": 1185 }, { "epoch": 5.729468599033816, "grad_norm": 0.0627102181315422, "learning_rate": 9.500864485090334e-05, "loss": 0.0094, "step": 1186 }, { "epoch": 5.734299516908212, "grad_norm": 0.0828758180141449, "learning_rate": 9.499394418331315e-05, "loss": 0.0132, "step": 1187 }, { "epoch": 5.739130434782608, "grad_norm": 0.06312569975852966, "learning_rate": 9.497922303990225e-05, "loss": 0.0086, "step": 1188 }, { "epoch": 5.743961352657005, "grad_norm": 0.04896373301744461, "learning_rate": 9.496448142736991e-05, "loss": 0.0064, "step": 1189 }, { "epoch": 5.748792270531401, "grad_norm": 0.10162996500730515, "learning_rate": 9.494971935242477e-05, "loss": 0.0175, "step": 1190 }, { "epoch": 5.753623188405797, "grad_norm": 0.055514682084321976, "learning_rate": 9.493493682178472e-05, "loss": 0.0101, "step": 1191 }, { "epoch": 5.758454106280193, "grad_norm": 0.05568890646100044, "learning_rate": 9.492013384217699e-05, "loss": 0.0096, "step": 1192 }, { "epoch": 5.763285024154589, "grad_norm": 0.05123963579535484, "learning_rate": 9.49053104203381e-05, "loss": 0.0106, "step": 1193 }, { "epoch": 5.768115942028985, "grad_norm": 0.05242984741926193, "learning_rate": 9.489046656301393e-05, "loss": 0.0101, "step": 1194 }, { "epoch": 5.7729468599033815, "grad_norm": 0.052701305598020554, "learning_rate": 9.487560227695955e-05, "loss": 0.0131, "step": 1195 }, { "epoch": 5.777777777777778, "grad_norm": 0.0495404452085495, "learning_rate": 9.486071756893944e-05, "loss": 0.0075, "step": 1196 }, { "epoch": 5.782608695652174, "grad_norm": 0.0440664142370224, "learning_rate": 9.484581244572732e-05, "loss": 0.0081, "step": 1197 }, { "epoch": 5.78743961352657, "grad_norm": 0.04695248231291771, "learning_rate": 9.48308869141062e-05, "loss": 0.0084, "step": 1198 }, { "epoch": 5.792270531400966, "grad_norm": 0.059293605387210846, "learning_rate": 9.481594098086837e-05, "loss": 0.011, "step": 1199 }, { "epoch": 5.797101449275362, "grad_norm": 0.06034954637289047, "learning_rate": 9.480097465281543e-05, "loss": 0.0081, "step": 1200 }, { "epoch": 5.797101449275362, "eval_loss": 0.01758287288248539, "eval_runtime": 20.7195, "eval_samples_per_second": 4.826, "eval_steps_per_second": 0.145, "step": 1200 }, { "epoch": 5.8019323671497585, "grad_norm": 0.05583087354898453, "learning_rate": 9.478598793675825e-05, "loss": 0.0082, "step": 1201 }, { "epoch": 5.806763285024155, "grad_norm": 0.05717151612043381, "learning_rate": 9.477098083951696e-05, "loss": 0.0096, "step": 1202 }, { "epoch": 5.811594202898551, "grad_norm": 0.06151145324110985, "learning_rate": 9.475595336792102e-05, "loss": 0.0072, "step": 1203 }, { "epoch": 5.816425120772947, "grad_norm": 0.06331399083137512, "learning_rate": 9.474090552880907e-05, "loss": 0.0111, "step": 1204 }, { "epoch": 5.821256038647343, "grad_norm": 0.05892903357744217, "learning_rate": 9.472583732902914e-05, "loss": 0.0086, "step": 1205 }, { "epoch": 5.826086956521739, "grad_norm": 0.05928612872958183, "learning_rate": 9.47107487754384e-05, "loss": 0.0088, "step": 1206 }, { "epoch": 5.830917874396135, "grad_norm": 0.06218522787094116, "learning_rate": 9.469563987490335e-05, "loss": 0.0072, "step": 1207 }, { "epoch": 5.835748792270532, "grad_norm": 0.07384292781352997, "learning_rate": 9.468051063429977e-05, "loss": 0.0093, "step": 1208 }, { "epoch": 5.840579710144928, "grad_norm": 0.05546915531158447, "learning_rate": 9.466536106051265e-05, "loss": 0.01, "step": 1209 }, { "epoch": 5.845410628019324, "grad_norm": 0.05841489136219025, "learning_rate": 9.465019116043625e-05, "loss": 0.01, "step": 1210 }, { "epoch": 5.85024154589372, "grad_norm": 0.049476198852062225, "learning_rate": 9.463500094097406e-05, "loss": 0.0094, "step": 1211 }, { "epoch": 5.855072463768116, "grad_norm": 0.04447973519563675, "learning_rate": 9.461979040903888e-05, "loss": 0.0067, "step": 1212 }, { "epoch": 5.859903381642512, "grad_norm": 0.048743024468421936, "learning_rate": 9.460455957155267e-05, "loss": 0.0075, "step": 1213 }, { "epoch": 5.8647342995169085, "grad_norm": 0.05664145573973656, "learning_rate": 9.458930843544671e-05, "loss": 0.0104, "step": 1214 }, { "epoch": 5.869565217391305, "grad_norm": 0.06544919312000275, "learning_rate": 9.457403700766144e-05, "loss": 0.0119, "step": 1215 }, { "epoch": 5.874396135265701, "grad_norm": 0.06714371591806412, "learning_rate": 9.455874529514661e-05, "loss": 0.0121, "step": 1216 }, { "epoch": 5.879227053140097, "grad_norm": 0.04535643011331558, "learning_rate": 9.454343330486113e-05, "loss": 0.0083, "step": 1217 }, { "epoch": 5.884057971014493, "grad_norm": 0.05146961286664009, "learning_rate": 9.45281010437732e-05, "loss": 0.0084, "step": 1218 }, { "epoch": 5.888888888888889, "grad_norm": 0.0556597076356411, "learning_rate": 9.451274851886021e-05, "loss": 0.0098, "step": 1219 }, { "epoch": 5.8937198067632846, "grad_norm": 0.053466394543647766, "learning_rate": 9.449737573710876e-05, "loss": 0.0079, "step": 1220 }, { "epoch": 5.898550724637682, "grad_norm": 0.06228194385766983, "learning_rate": 9.448198270551472e-05, "loss": 0.0085, "step": 1221 }, { "epoch": 5.903381642512077, "grad_norm": 0.05365237966179848, "learning_rate": 9.446656943108312e-05, "loss": 0.0079, "step": 1222 }, { "epoch": 5.908212560386474, "grad_norm": 0.04714881628751755, "learning_rate": 9.445113592082823e-05, "loss": 0.0069, "step": 1223 }, { "epoch": 5.913043478260869, "grad_norm": 0.047249406576156616, "learning_rate": 9.443568218177353e-05, "loss": 0.0076, "step": 1224 }, { "epoch": 5.917874396135265, "grad_norm": 0.06869522482156754, "learning_rate": 9.442020822095168e-05, "loss": 0.0071, "step": 1225 }, { "epoch": 5.917874396135265, "eval_loss": 0.01747715473175049, "eval_runtime": 21.3272, "eval_samples_per_second": 4.689, "eval_steps_per_second": 0.141, "step": 1225 }, { "epoch": 5.9227053140096615, "grad_norm": 0.04264363646507263, "learning_rate": 9.440471404540459e-05, "loss": 0.0063, "step": 1226 }, { "epoch": 5.927536231884058, "grad_norm": 0.06179254874587059, "learning_rate": 9.438919966218332e-05, "loss": 0.0068, "step": 1227 }, { "epoch": 5.932367149758454, "grad_norm": 0.05508408322930336, "learning_rate": 9.437366507834816e-05, "loss": 0.0082, "step": 1228 }, { "epoch": 5.93719806763285, "grad_norm": 0.06304654479026794, "learning_rate": 9.435811030096857e-05, "loss": 0.0105, "step": 1229 }, { "epoch": 5.942028985507246, "grad_norm": 0.06207441911101341, "learning_rate": 9.434253533712322e-05, "loss": 0.0105, "step": 1230 }, { "epoch": 5.946859903381642, "grad_norm": 0.06103399395942688, "learning_rate": 9.432694019389997e-05, "loss": 0.0084, "step": 1231 }, { "epoch": 5.951690821256038, "grad_norm": 0.05756254121661186, "learning_rate": 9.43113248783958e-05, "loss": 0.0097, "step": 1232 }, { "epoch": 5.956521739130435, "grad_norm": 0.055301081389188766, "learning_rate": 9.429568939771698e-05, "loss": 0.0081, "step": 1233 }, { "epoch": 5.961352657004831, "grad_norm": 0.05765131488442421, "learning_rate": 9.428003375897885e-05, "loss": 0.0096, "step": 1234 }, { "epoch": 5.966183574879227, "grad_norm": 0.0458630807697773, "learning_rate": 9.426435796930599e-05, "loss": 0.0089, "step": 1235 }, { "epoch": 5.971014492753623, "grad_norm": 0.053680840879678726, "learning_rate": 9.424866203583213e-05, "loss": 0.0096, "step": 1236 }, { "epoch": 5.975845410628019, "grad_norm": 0.04151618480682373, "learning_rate": 9.423294596570016e-05, "loss": 0.0088, "step": 1237 }, { "epoch": 5.980676328502415, "grad_norm": 0.04098724573850632, "learning_rate": 9.421720976606217e-05, "loss": 0.0079, "step": 1238 }, { "epoch": 5.9855072463768115, "grad_norm": 0.03668149933218956, "learning_rate": 9.420145344407932e-05, "loss": 0.0072, "step": 1239 }, { "epoch": 5.990338164251208, "grad_norm": 0.044391755014657974, "learning_rate": 9.418567700692205e-05, "loss": 0.0105, "step": 1240 }, { "epoch": 5.995169082125604, "grad_norm": 0.05425712838768959, "learning_rate": 9.416988046176987e-05, "loss": 0.0116, "step": 1241 }, { "epoch": 6.0, "grad_norm": 0.09103017300367355, "learning_rate": 9.415406381581146e-05, "loss": 0.006, "step": 1242 }, { "epoch": 6.004830917874396, "grad_norm": 0.043098460882902145, "learning_rate": 9.413822707624465e-05, "loss": 0.006, "step": 1243 }, { "epoch": 6.009661835748792, "grad_norm": 0.041810572147369385, "learning_rate": 9.412237025027642e-05, "loss": 0.0067, "step": 1244 }, { "epoch": 6.0144927536231885, "grad_norm": 0.03212606534361839, "learning_rate": 9.410649334512289e-05, "loss": 0.0046, "step": 1245 }, { "epoch": 6.019323671497585, "grad_norm": 0.04054751247167587, "learning_rate": 9.40905963680093e-05, "loss": 0.0068, "step": 1246 }, { "epoch": 6.024154589371981, "grad_norm": 0.0570932999253273, "learning_rate": 9.407467932617006e-05, "loss": 0.006, "step": 1247 }, { "epoch": 6.028985507246377, "grad_norm": 0.03890130668878555, "learning_rate": 9.405874222684867e-05, "loss": 0.0053, "step": 1248 }, { "epoch": 6.033816425120773, "grad_norm": 0.07312671095132828, "learning_rate": 9.404278507729777e-05, "loss": 0.0088, "step": 1249 }, { "epoch": 6.038647342995169, "grad_norm": 0.047085270285606384, "learning_rate": 9.402680788477916e-05, "loss": 0.0058, "step": 1250 }, { "epoch": 6.038647342995169, "eval_loss": 0.018731381744146347, "eval_runtime": 20.6014, "eval_samples_per_second": 4.854, "eval_steps_per_second": 0.146, "step": 1250 }, { "epoch": 6.043478260869565, "grad_norm": 0.04745680093765259, "learning_rate": 9.401081065656372e-05, "loss": 0.0056, "step": 1251 }, { "epoch": 6.048309178743962, "grad_norm": 0.05071539804339409, "learning_rate": 9.399479339993144e-05, "loss": 0.0091, "step": 1252 }, { "epoch": 6.053140096618358, "grad_norm": 0.05788218230009079, "learning_rate": 9.397875612217148e-05, "loss": 0.0053, "step": 1253 }, { "epoch": 6.057971014492754, "grad_norm": 0.04517432302236557, "learning_rate": 9.396269883058206e-05, "loss": 0.0055, "step": 1254 }, { "epoch": 6.06280193236715, "grad_norm": 0.04658951237797737, "learning_rate": 9.394662153247051e-05, "loss": 0.0064, "step": 1255 }, { "epoch": 6.067632850241546, "grad_norm": 0.06327016651630402, "learning_rate": 9.393052423515328e-05, "loss": 0.0072, "step": 1256 }, { "epoch": 6.072463768115942, "grad_norm": 0.04202742502093315, "learning_rate": 9.391440694595595e-05, "loss": 0.0054, "step": 1257 }, { "epoch": 6.0772946859903385, "grad_norm": 0.04662294313311577, "learning_rate": 9.389826967221314e-05, "loss": 0.0049, "step": 1258 }, { "epoch": 6.082125603864735, "grad_norm": 0.06586906313896179, "learning_rate": 9.38821124212686e-05, "loss": 0.0054, "step": 1259 }, { "epoch": 6.086956521739131, "grad_norm": 0.03706370294094086, "learning_rate": 9.386593520047516e-05, "loss": 0.0041, "step": 1260 }, { "epoch": 6.091787439613527, "grad_norm": 0.03640469163656235, "learning_rate": 9.384973801719475e-05, "loss": 0.0053, "step": 1261 }, { "epoch": 6.096618357487923, "grad_norm": 0.05960666388273239, "learning_rate": 9.383352087879838e-05, "loss": 0.0062, "step": 1262 }, { "epoch": 6.101449275362318, "grad_norm": 0.08039825409650803, "learning_rate": 9.38172837926661e-05, "loss": 0.0122, "step": 1263 }, { "epoch": 6.106280193236715, "grad_norm": 0.061855610460042953, "learning_rate": 9.380102676618712e-05, "loss": 0.0072, "step": 1264 }, { "epoch": 6.111111111111111, "grad_norm": 0.0521203950047493, "learning_rate": 9.378474980675966e-05, "loss": 0.0068, "step": 1265 }, { "epoch": 6.115942028985507, "grad_norm": 0.056785743683576584, "learning_rate": 9.376845292179103e-05, "loss": 0.0065, "step": 1266 }, { "epoch": 6.120772946859903, "grad_norm": 0.050204869359731674, "learning_rate": 9.37521361186976e-05, "loss": 0.0045, "step": 1267 }, { "epoch": 6.125603864734299, "grad_norm": 0.06700023263692856, "learning_rate": 9.373579940490484e-05, "loss": 0.0095, "step": 1268 }, { "epoch": 6.130434782608695, "grad_norm": 0.043674830347299576, "learning_rate": 9.371944278784723e-05, "loss": 0.0054, "step": 1269 }, { "epoch": 6.1352657004830915, "grad_norm": 0.05406811088323593, "learning_rate": 9.370306627496832e-05, "loss": 0.0079, "step": 1270 }, { "epoch": 6.140096618357488, "grad_norm": 0.04308624565601349, "learning_rate": 9.368666987372074e-05, "loss": 0.0057, "step": 1271 }, { "epoch": 6.144927536231884, "grad_norm": 0.03334224224090576, "learning_rate": 9.367025359156618e-05, "loss": 0.0051, "step": 1272 }, { "epoch": 6.14975845410628, "grad_norm": 0.0489167757332325, "learning_rate": 9.36538174359753e-05, "loss": 0.005, "step": 1273 }, { "epoch": 6.154589371980676, "grad_norm": 0.06193888559937477, "learning_rate": 9.363736141442791e-05, "loss": 0.0074, "step": 1274 }, { "epoch": 6.159420289855072, "grad_norm": 0.05088794603943825, "learning_rate": 9.362088553441277e-05, "loss": 0.0081, "step": 1275 }, { "epoch": 6.159420289855072, "eval_loss": 0.016518954187631607, "eval_runtime": 20.607, "eval_samples_per_second": 4.853, "eval_steps_per_second": 0.146, "step": 1275 }, { "epoch": 6.164251207729468, "grad_norm": 0.04955621808767319, "learning_rate": 9.360438980342774e-05, "loss": 0.006, "step": 1276 }, { "epoch": 6.169082125603865, "grad_norm": 0.0462108850479126, "learning_rate": 9.358787422897968e-05, "loss": 0.0055, "step": 1277 }, { "epoch": 6.173913043478261, "grad_norm": 0.05635419487953186, "learning_rate": 9.357133881858448e-05, "loss": 0.0064, "step": 1278 }, { "epoch": 6.178743961352657, "grad_norm": 0.046862367540597916, "learning_rate": 9.355478357976706e-05, "loss": 0.0059, "step": 1279 }, { "epoch": 6.183574879227053, "grad_norm": 0.04464318975806236, "learning_rate": 9.35382085200614e-05, "loss": 0.0053, "step": 1280 }, { "epoch": 6.188405797101449, "grad_norm": 0.04266749694943428, "learning_rate": 9.352161364701044e-05, "loss": 0.0057, "step": 1281 }, { "epoch": 6.193236714975845, "grad_norm": 0.04351837933063507, "learning_rate": 9.35049989681662e-05, "loss": 0.0065, "step": 1282 }, { "epoch": 6.1980676328502415, "grad_norm": 0.03657788783311844, "learning_rate": 9.348836449108962e-05, "loss": 0.0043, "step": 1283 }, { "epoch": 6.202898550724638, "grad_norm": 0.04788932949304581, "learning_rate": 9.347171022335077e-05, "loss": 0.0051, "step": 1284 }, { "epoch": 6.207729468599034, "grad_norm": 0.06068427860736847, "learning_rate": 9.345503617252864e-05, "loss": 0.0061, "step": 1285 }, { "epoch": 6.21256038647343, "grad_norm": 0.05637691169977188, "learning_rate": 9.343834234621124e-05, "loss": 0.0072, "step": 1286 }, { "epoch": 6.217391304347826, "grad_norm": 0.043992042541503906, "learning_rate": 9.342162875199562e-05, "loss": 0.0057, "step": 1287 }, { "epoch": 6.222222222222222, "grad_norm": 0.04882000759243965, "learning_rate": 9.340489539748775e-05, "loss": 0.0069, "step": 1288 }, { "epoch": 6.2270531400966185, "grad_norm": 0.047943197190761566, "learning_rate": 9.338814229030268e-05, "loss": 0.0054, "step": 1289 }, { "epoch": 6.231884057971015, "grad_norm": 0.05657848343253136, "learning_rate": 9.337136943806437e-05, "loss": 0.0072, "step": 1290 }, { "epoch": 6.236714975845411, "grad_norm": 0.049666404724121094, "learning_rate": 9.335457684840581e-05, "loss": 0.0058, "step": 1291 }, { "epoch": 6.241545893719807, "grad_norm": 0.0505865253508091, "learning_rate": 9.333776452896897e-05, "loss": 0.0064, "step": 1292 }, { "epoch": 6.246376811594203, "grad_norm": 0.07791588455438614, "learning_rate": 9.332093248740479e-05, "loss": 0.0058, "step": 1293 }, { "epoch": 6.251207729468599, "grad_norm": 0.03757460042834282, "learning_rate": 9.330408073137319e-05, "loss": 0.0039, "step": 1294 }, { "epoch": 6.256038647342995, "grad_norm": 0.05717682093381882, "learning_rate": 9.328720926854305e-05, "loss": 0.0052, "step": 1295 }, { "epoch": 6.260869565217392, "grad_norm": 0.05694163963198662, "learning_rate": 9.327031810659224e-05, "loss": 0.008, "step": 1296 }, { "epoch": 6.265700483091788, "grad_norm": 0.06574101746082306, "learning_rate": 9.325340725320755e-05, "loss": 0.0061, "step": 1297 }, { "epoch": 6.270531400966184, "grad_norm": 0.03355651721358299, "learning_rate": 9.323647671608479e-05, "loss": 0.0041, "step": 1298 }, { "epoch": 6.27536231884058, "grad_norm": 0.047428570687770844, "learning_rate": 9.321952650292871e-05, "loss": 0.0058, "step": 1299 }, { "epoch": 6.280193236714976, "grad_norm": 0.046640850603580475, "learning_rate": 9.320255662145298e-05, "loss": 0.0057, "step": 1300 }, { "epoch": 6.280193236714976, "eval_loss": 0.017070595175027847, "eval_runtime": 20.6176, "eval_samples_per_second": 4.85, "eval_steps_per_second": 0.146, "step": 1300 }, { "epoch": 6.285024154589372, "grad_norm": 0.06039797514677048, "learning_rate": 9.318556707938025e-05, "loss": 0.0087, "step": 1301 }, { "epoch": 6.2898550724637685, "grad_norm": 0.04693787172436714, "learning_rate": 9.316855788444212e-05, "loss": 0.0063, "step": 1302 }, { "epoch": 6.294685990338165, "grad_norm": 0.048959773033857346, "learning_rate": 9.315152904437911e-05, "loss": 0.0067, "step": 1303 }, { "epoch": 6.29951690821256, "grad_norm": 0.09099757671356201, "learning_rate": 9.313448056694071e-05, "loss": 0.0063, "step": 1304 }, { "epoch": 6.304347826086957, "grad_norm": 0.04881768301129341, "learning_rate": 9.311741245988535e-05, "loss": 0.007, "step": 1305 }, { "epoch": 6.309178743961352, "grad_norm": 0.044358160346746445, "learning_rate": 9.310032473098033e-05, "loss": 0.0058, "step": 1306 }, { "epoch": 6.314009661835748, "grad_norm": 0.04900408908724785, "learning_rate": 9.308321738800196e-05, "loss": 0.0049, "step": 1307 }, { "epoch": 6.318840579710145, "grad_norm": 0.03707882761955261, "learning_rate": 9.306609043873542e-05, "loss": 0.0047, "step": 1308 }, { "epoch": 6.323671497584541, "grad_norm": 0.0487239807844162, "learning_rate": 9.304894389097486e-05, "loss": 0.0071, "step": 1309 }, { "epoch": 6.328502415458937, "grad_norm": 0.06111222505569458, "learning_rate": 9.303177775252327e-05, "loss": 0.0087, "step": 1310 }, { "epoch": 6.333333333333333, "grad_norm": 0.03556101396679878, "learning_rate": 9.301459203119267e-05, "loss": 0.0052, "step": 1311 }, { "epoch": 6.338164251207729, "grad_norm": 0.06353302299976349, "learning_rate": 9.299738673480391e-05, "loss": 0.0063, "step": 1312 }, { "epoch": 6.342995169082125, "grad_norm": 0.09204978495836258, "learning_rate": 9.298016187118675e-05, "loss": 0.0081, "step": 1313 }, { "epoch": 6.3478260869565215, "grad_norm": 0.0807780846953392, "learning_rate": 9.296291744817987e-05, "loss": 0.0065, "step": 1314 }, { "epoch": 6.352657004830918, "grad_norm": 0.06295431405305862, "learning_rate": 9.294565347363085e-05, "loss": 0.0062, "step": 1315 }, { "epoch": 6.357487922705314, "grad_norm": 0.046747080981731415, "learning_rate": 9.29283699553962e-05, "loss": 0.006, "step": 1316 }, { "epoch": 6.36231884057971, "grad_norm": 0.04669469594955444, "learning_rate": 9.291106690134128e-05, "loss": 0.0068, "step": 1317 }, { "epoch": 6.367149758454106, "grad_norm": 0.049420833587646484, "learning_rate": 9.289374431934035e-05, "loss": 0.0063, "step": 1318 }, { "epoch": 6.371980676328502, "grad_norm": 0.07315026968717575, "learning_rate": 9.287640221727658e-05, "loss": 0.0079, "step": 1319 }, { "epoch": 6.3768115942028984, "grad_norm": 0.044721487909555435, "learning_rate": 9.285904060304198e-05, "loss": 0.0064, "step": 1320 }, { "epoch": 6.381642512077295, "grad_norm": 0.045167844742536545, "learning_rate": 9.284165948453747e-05, "loss": 0.006, "step": 1321 }, { "epoch": 6.386473429951691, "grad_norm": 0.045468948781490326, "learning_rate": 9.282425886967284e-05, "loss": 0.0066, "step": 1322 }, { "epoch": 6.391304347826087, "grad_norm": 0.03726964443922043, "learning_rate": 9.280683876636677e-05, "loss": 0.0057, "step": 1323 }, { "epoch": 6.396135265700483, "grad_norm": 0.054976094514131546, "learning_rate": 9.278939918254677e-05, "loss": 0.0064, "step": 1324 }, { "epoch": 6.400966183574879, "grad_norm": 0.059892453253269196, "learning_rate": 9.277194012614925e-05, "loss": 0.0068, "step": 1325 }, { "epoch": 6.400966183574879, "eval_loss": 0.016453834250569344, "eval_runtime": 21.3612, "eval_samples_per_second": 4.681, "eval_steps_per_second": 0.14, "step": 1325 }, { "epoch": 6.405797101449275, "grad_norm": 0.048259880393743515, "learning_rate": 9.275446160511946e-05, "loss": 0.0078, "step": 1326 }, { "epoch": 6.4106280193236715, "grad_norm": 0.06093078851699829, "learning_rate": 9.273696362741151e-05, "loss": 0.0078, "step": 1327 }, { "epoch": 6.415458937198068, "grad_norm": 0.048800837248563766, "learning_rate": 9.271944620098841e-05, "loss": 0.0071, "step": 1328 }, { "epoch": 6.420289855072464, "grad_norm": 0.056117165833711624, "learning_rate": 9.270190933382192e-05, "loss": 0.0071, "step": 1329 }, { "epoch": 6.42512077294686, "grad_norm": 0.06087968125939369, "learning_rate": 9.268435303389275e-05, "loss": 0.0081, "step": 1330 }, { "epoch": 6.429951690821256, "grad_norm": 0.040126699954271317, "learning_rate": 9.266677730919043e-05, "loss": 0.0048, "step": 1331 }, { "epoch": 6.434782608695652, "grad_norm": 0.04188450053334236, "learning_rate": 9.264918216771325e-05, "loss": 0.0055, "step": 1332 }, { "epoch": 6.4396135265700485, "grad_norm": 0.04577482119202614, "learning_rate": 9.263156761746844e-05, "loss": 0.0061, "step": 1333 }, { "epoch": 6.444444444444445, "grad_norm": 0.05482233315706253, "learning_rate": 9.261393366647202e-05, "loss": 0.0055, "step": 1334 }, { "epoch": 6.449275362318841, "grad_norm": 0.04913961887359619, "learning_rate": 9.259628032274882e-05, "loss": 0.0055, "step": 1335 }, { "epoch": 6.454106280193237, "grad_norm": 0.0711643248796463, "learning_rate": 9.257860759433252e-05, "loss": 0.0063, "step": 1336 }, { "epoch": 6.458937198067633, "grad_norm": 0.040503520518541336, "learning_rate": 9.256091548926565e-05, "loss": 0.0049, "step": 1337 }, { "epoch": 6.463768115942029, "grad_norm": 0.04341737926006317, "learning_rate": 9.254320401559947e-05, "loss": 0.005, "step": 1338 }, { "epoch": 6.468599033816425, "grad_norm": 0.04081616923213005, "learning_rate": 9.252547318139415e-05, "loss": 0.0054, "step": 1339 }, { "epoch": 6.473429951690822, "grad_norm": 0.059148360043764114, "learning_rate": 9.250772299471859e-05, "loss": 0.0062, "step": 1340 }, { "epoch": 6.478260869565218, "grad_norm": 0.07044517248868942, "learning_rate": 9.248995346365058e-05, "loss": 0.0076, "step": 1341 }, { "epoch": 6.483091787439614, "grad_norm": 0.0680868849158287, "learning_rate": 9.247216459627665e-05, "loss": 0.0059, "step": 1342 }, { "epoch": 6.48792270531401, "grad_norm": 0.04491333290934563, "learning_rate": 9.245435640069213e-05, "loss": 0.0058, "step": 1343 }, { "epoch": 6.492753623188406, "grad_norm": 0.06955177336931229, "learning_rate": 9.243652888500119e-05, "loss": 0.011, "step": 1344 }, { "epoch": 6.4975845410628015, "grad_norm": 0.054043661803007126, "learning_rate": 9.241868205731676e-05, "loss": 0.0053, "step": 1345 }, { "epoch": 6.5024154589371985, "grad_norm": 0.04067247360944748, "learning_rate": 9.240081592576056e-05, "loss": 0.0052, "step": 1346 }, { "epoch": 6.507246376811594, "grad_norm": 0.04347148910164833, "learning_rate": 9.23829304984631e-05, "loss": 0.0064, "step": 1347 }, { "epoch": 6.512077294685991, "grad_norm": 0.046369269490242004, "learning_rate": 9.236502578356368e-05, "loss": 0.0063, "step": 1348 }, { "epoch": 6.516908212560386, "grad_norm": 0.048870883882045746, "learning_rate": 9.234710178921035e-05, "loss": 0.0059, "step": 1349 }, { "epoch": 6.521739130434782, "grad_norm": 0.04518553987145424, "learning_rate": 9.232915852355996e-05, "loss": 0.0059, "step": 1350 }, { "epoch": 6.521739130434782, "eval_loss": 0.01632525771856308, "eval_runtime": 20.6147, "eval_samples_per_second": 4.851, "eval_steps_per_second": 0.146, "step": 1350 }, { "epoch": 6.526570048309178, "grad_norm": 0.050386469811201096, "learning_rate": 9.231119599477814e-05, "loss": 0.0071, "step": 1351 }, { "epoch": 6.531400966183575, "grad_norm": 0.03906089812517166, "learning_rate": 9.229321421103924e-05, "loss": 0.0053, "step": 1352 }, { "epoch": 6.536231884057971, "grad_norm": 0.056843534111976624, "learning_rate": 9.227521318052641e-05, "loss": 0.0087, "step": 1353 }, { "epoch": 6.541062801932367, "grad_norm": 0.05382772162556648, "learning_rate": 9.225719291143156e-05, "loss": 0.0061, "step": 1354 }, { "epoch": 6.545893719806763, "grad_norm": 0.04173978790640831, "learning_rate": 9.223915341195535e-05, "loss": 0.0054, "step": 1355 }, { "epoch": 6.550724637681159, "grad_norm": 0.055689819157123566, "learning_rate": 9.222109469030716e-05, "loss": 0.0059, "step": 1356 }, { "epoch": 6.555555555555555, "grad_norm": 0.04744328185915947, "learning_rate": 9.220301675470518e-05, "loss": 0.0076, "step": 1357 }, { "epoch": 6.5603864734299515, "grad_norm": 0.0510600246489048, "learning_rate": 9.218491961337626e-05, "loss": 0.0062, "step": 1358 }, { "epoch": 6.565217391304348, "grad_norm": 0.041862208396196365, "learning_rate": 9.216680327455608e-05, "loss": 0.0056, "step": 1359 }, { "epoch": 6.570048309178744, "grad_norm": 0.0443810373544693, "learning_rate": 9.214866774648899e-05, "loss": 0.0055, "step": 1360 }, { "epoch": 6.57487922705314, "grad_norm": 0.05620473250746727, "learning_rate": 9.21305130374281e-05, "loss": 0.0073, "step": 1361 }, { "epoch": 6.579710144927536, "grad_norm": 0.04563259333372116, "learning_rate": 9.211233915563526e-05, "loss": 0.0064, "step": 1362 }, { "epoch": 6.584541062801932, "grad_norm": 0.05491187795996666, "learning_rate": 9.209414610938101e-05, "loss": 0.0077, "step": 1363 }, { "epoch": 6.5893719806763285, "grad_norm": 0.055255644023418427, "learning_rate": 9.207593390694466e-05, "loss": 0.0071, "step": 1364 }, { "epoch": 6.594202898550725, "grad_norm": 0.05295202136039734, "learning_rate": 9.205770255661417e-05, "loss": 0.0067, "step": 1365 }, { "epoch": 6.599033816425121, "grad_norm": 0.07129408419132233, "learning_rate": 9.20394520666863e-05, "loss": 0.0099, "step": 1366 }, { "epoch": 6.603864734299517, "grad_norm": 0.04434022307395935, "learning_rate": 9.202118244546643e-05, "loss": 0.0057, "step": 1367 }, { "epoch": 6.608695652173913, "grad_norm": 0.050070665776729584, "learning_rate": 9.200289370126871e-05, "loss": 0.0057, "step": 1368 }, { "epoch": 6.613526570048309, "grad_norm": 0.050253286957740784, "learning_rate": 9.198458584241597e-05, "loss": 0.0065, "step": 1369 }, { "epoch": 6.618357487922705, "grad_norm": 0.05804944410920143, "learning_rate": 9.196625887723975e-05, "loss": 0.0058, "step": 1370 }, { "epoch": 6.6231884057971016, "grad_norm": 0.05418354645371437, "learning_rate": 9.194791281408026e-05, "loss": 0.0065, "step": 1371 }, { "epoch": 6.628019323671498, "grad_norm": 0.03469512239098549, "learning_rate": 9.192954766128643e-05, "loss": 0.0044, "step": 1372 }, { "epoch": 6.632850241545894, "grad_norm": 0.044930633157491684, "learning_rate": 9.191116342721586e-05, "loss": 0.0055, "step": 1373 }, { "epoch": 6.63768115942029, "grad_norm": 0.05017385631799698, "learning_rate": 9.189276012023483e-05, "loss": 0.006, "step": 1374 }, { "epoch": 6.642512077294686, "grad_norm": 0.05084605887532234, "learning_rate": 9.187433774871832e-05, "loss": 0.0057, "step": 1375 }, { "epoch": 6.642512077294686, "eval_loss": 0.01513008400797844, "eval_runtime": 20.6668, "eval_samples_per_second": 4.839, "eval_steps_per_second": 0.145, "step": 1375 }, { "epoch": 6.647342995169082, "grad_norm": 0.04787183925509453, "learning_rate": 9.185589632104997e-05, "loss": 0.0063, "step": 1376 }, { "epoch": 6.6521739130434785, "grad_norm": 0.05482997000217438, "learning_rate": 9.183743584562208e-05, "loss": 0.0067, "step": 1377 }, { "epoch": 6.657004830917875, "grad_norm": 0.0503399483859539, "learning_rate": 9.181895633083564e-05, "loss": 0.0068, "step": 1378 }, { "epoch": 6.661835748792271, "grad_norm": 0.052327338606119156, "learning_rate": 9.180045778510031e-05, "loss": 0.0063, "step": 1379 }, { "epoch": 6.666666666666667, "grad_norm": 0.05098320543766022, "learning_rate": 9.178194021683441e-05, "loss": 0.005, "step": 1380 }, { "epoch": 6.671497584541063, "grad_norm": 0.060967762023210526, "learning_rate": 9.176340363446488e-05, "loss": 0.0053, "step": 1381 }, { "epoch": 6.676328502415459, "grad_norm": 0.04176538065075874, "learning_rate": 9.174484804642733e-05, "loss": 0.005, "step": 1382 }, { "epoch": 6.681159420289855, "grad_norm": 0.04020035266876221, "learning_rate": 9.172627346116606e-05, "loss": 0.0071, "step": 1383 }, { "epoch": 6.685990338164252, "grad_norm": 0.0703640952706337, "learning_rate": 9.170767988713396e-05, "loss": 0.008, "step": 1384 }, { "epoch": 6.690821256038648, "grad_norm": 0.049804870039224625, "learning_rate": 9.168906733279259e-05, "loss": 0.0075, "step": 1385 }, { "epoch": 6.695652173913043, "grad_norm": 0.04536246880888939, "learning_rate": 9.167043580661215e-05, "loss": 0.0054, "step": 1386 }, { "epoch": 6.70048309178744, "grad_norm": 0.03730194270610809, "learning_rate": 9.165178531707144e-05, "loss": 0.005, "step": 1387 }, { "epoch": 6.705314009661835, "grad_norm": 0.045735448598861694, "learning_rate": 9.163311587265793e-05, "loss": 0.0061, "step": 1388 }, { "epoch": 6.710144927536232, "grad_norm": 0.04150770232081413, "learning_rate": 9.161442748186771e-05, "loss": 0.0051, "step": 1389 }, { "epoch": 6.714975845410628, "grad_norm": 0.04654907062649727, "learning_rate": 9.159572015320548e-05, "loss": 0.0061, "step": 1390 }, { "epoch": 6.719806763285024, "grad_norm": 0.06273826211690903, "learning_rate": 9.157699389518456e-05, "loss": 0.0083, "step": 1391 }, { "epoch": 6.72463768115942, "grad_norm": 0.050027329474687576, "learning_rate": 9.155824871632688e-05, "loss": 0.0061, "step": 1392 }, { "epoch": 6.729468599033816, "grad_norm": 0.04607383534312248, "learning_rate": 9.153948462516299e-05, "loss": 0.0053, "step": 1393 }, { "epoch": 6.734299516908212, "grad_norm": 0.044204093515872955, "learning_rate": 9.152070163023203e-05, "loss": 0.0047, "step": 1394 }, { "epoch": 6.739130434782608, "grad_norm": 0.043753284960985184, "learning_rate": 9.150189974008179e-05, "loss": 0.0067, "step": 1395 }, { "epoch": 6.743961352657005, "grad_norm": 0.052429623901844025, "learning_rate": 9.148307896326857e-05, "loss": 0.007, "step": 1396 }, { "epoch": 6.748792270531401, "grad_norm": 0.06474307924509048, "learning_rate": 9.14642393083574e-05, "loss": 0.0086, "step": 1397 }, { "epoch": 6.753623188405797, "grad_norm": 0.0789802297949791, "learning_rate": 9.144538078392173e-05, "loss": 0.0082, "step": 1398 }, { "epoch": 6.758454106280193, "grad_norm": 0.0541478730738163, "learning_rate": 9.142650339854374e-05, "loss": 0.0058, "step": 1399 }, { "epoch": 6.763285024154589, "grad_norm": 0.04348256438970566, "learning_rate": 9.140760716081414e-05, "loss": 0.0061, "step": 1400 }, { "epoch": 6.763285024154589, "eval_loss": 0.016424350440502167, "eval_runtime": 20.611, "eval_samples_per_second": 4.852, "eval_steps_per_second": 0.146, "step": 1400 }, { "epoch": 6.768115942028985, "grad_norm": 0.06769420206546783, "learning_rate": 9.13886920793322e-05, "loss": 0.0064, "step": 1401 }, { "epoch": 6.7729468599033815, "grad_norm": 0.03817087411880493, "learning_rate": 9.136975816270579e-05, "loss": 0.005, "step": 1402 }, { "epoch": 6.777777777777778, "grad_norm": 0.03812692314386368, "learning_rate": 9.135080541955136e-05, "loss": 0.0057, "step": 1403 }, { "epoch": 6.782608695652174, "grad_norm": 0.050985757261514664, "learning_rate": 9.133183385849392e-05, "loss": 0.0066, "step": 1404 }, { "epoch": 6.78743961352657, "grad_norm": 0.06294248253107071, "learning_rate": 9.1312843488167e-05, "loss": 0.0068, "step": 1405 }, { "epoch": 6.792270531400966, "grad_norm": 0.05425696820020676, "learning_rate": 9.129383431721276e-05, "loss": 0.0064, "step": 1406 }, { "epoch": 6.797101449275362, "grad_norm": 0.05892189219594002, "learning_rate": 9.12748063542819e-05, "loss": 0.0063, "step": 1407 }, { "epoch": 6.8019323671497585, "grad_norm": 0.04670066386461258, "learning_rate": 9.125575960803361e-05, "loss": 0.0061, "step": 1408 }, { "epoch": 6.806763285024155, "grad_norm": 0.0500483438372612, "learning_rate": 9.123669408713569e-05, "loss": 0.0057, "step": 1409 }, { "epoch": 6.811594202898551, "grad_norm": 0.06111598014831543, "learning_rate": 9.121760980026449e-05, "loss": 0.0059, "step": 1410 }, { "epoch": 6.816425120772947, "grad_norm": 0.040104810148477554, "learning_rate": 9.119850675610486e-05, "loss": 0.005, "step": 1411 }, { "epoch": 6.821256038647343, "grad_norm": 0.058813292533159256, "learning_rate": 9.11793849633502e-05, "loss": 0.0076, "step": 1412 }, { "epoch": 6.826086956521739, "grad_norm": 0.04925285279750824, "learning_rate": 9.116024443070243e-05, "loss": 0.0059, "step": 1413 }, { "epoch": 6.830917874396135, "grad_norm": 0.05609273910522461, "learning_rate": 9.114108516687205e-05, "loss": 0.0068, "step": 1414 }, { "epoch": 6.835748792270532, "grad_norm": 0.06939279288053513, "learning_rate": 9.112190718057802e-05, "loss": 0.0071, "step": 1415 }, { "epoch": 6.840579710144928, "grad_norm": 0.057253647595644, "learning_rate": 9.110271048054787e-05, "loss": 0.0063, "step": 1416 }, { "epoch": 6.845410628019324, "grad_norm": 0.0675768330693245, "learning_rate": 9.10834950755176e-05, "loss": 0.0079, "step": 1417 }, { "epoch": 6.85024154589372, "grad_norm": 0.06414294987916946, "learning_rate": 9.106426097423178e-05, "loss": 0.0076, "step": 1418 }, { "epoch": 6.855072463768116, "grad_norm": 0.05850673466920853, "learning_rate": 9.104500818544342e-05, "loss": 0.0088, "step": 1419 }, { "epoch": 6.859903381642512, "grad_norm": 0.08548006415367126, "learning_rate": 9.10257367179141e-05, "loss": 0.0106, "step": 1420 }, { "epoch": 6.8647342995169085, "grad_norm": 0.04931001737713814, "learning_rate": 9.100644658041382e-05, "loss": 0.0078, "step": 1421 }, { "epoch": 6.869565217391305, "grad_norm": 0.0717255249619484, "learning_rate": 9.098713778172119e-05, "loss": 0.0074, "step": 1422 }, { "epoch": 6.874396135265701, "grad_norm": 0.045794110745191574, "learning_rate": 9.09678103306232e-05, "loss": 0.0067, "step": 1423 }, { "epoch": 6.879227053140097, "grad_norm": 0.04047313332557678, "learning_rate": 9.094846423591539e-05, "loss": 0.006, "step": 1424 }, { "epoch": 6.884057971014493, "grad_norm": 0.0553482323884964, "learning_rate": 9.092909950640179e-05, "loss": 0.006, "step": 1425 }, { "epoch": 6.884057971014493, "eval_loss": 0.015602950006723404, "eval_runtime": 21.3308, "eval_samples_per_second": 4.688, "eval_steps_per_second": 0.141, "step": 1425 }, { "epoch": 6.888888888888889, "grad_norm": 0.044621117413043976, "learning_rate": 9.090971615089486e-05, "loss": 0.0064, "step": 1426 }, { "epoch": 6.8937198067632846, "grad_norm": 0.05033578351140022, "learning_rate": 9.089031417821558e-05, "loss": 0.0066, "step": 1427 }, { "epoch": 6.898550724637682, "grad_norm": 0.03769382834434509, "learning_rate": 9.08708935971934e-05, "loss": 0.005, "step": 1428 }, { "epoch": 6.903381642512077, "grad_norm": 0.04533964768052101, "learning_rate": 9.085145441666622e-05, "loss": 0.0064, "step": 1429 }, { "epoch": 6.908212560386474, "grad_norm": 0.04200972989201546, "learning_rate": 9.08319966454804e-05, "loss": 0.0052, "step": 1430 }, { "epoch": 6.913043478260869, "grad_norm": 0.05959639698266983, "learning_rate": 9.081252029249078e-05, "loss": 0.0069, "step": 1431 }, { "epoch": 6.917874396135265, "grad_norm": 0.05127401277422905, "learning_rate": 9.079302536656068e-05, "loss": 0.0065, "step": 1432 }, { "epoch": 6.9227053140096615, "grad_norm": 0.05631054937839508, "learning_rate": 9.07735118765618e-05, "loss": 0.0089, "step": 1433 }, { "epoch": 6.927536231884058, "grad_norm": 0.0575186163187027, "learning_rate": 9.075397983137434e-05, "loss": 0.0087, "step": 1434 }, { "epoch": 6.932367149758454, "grad_norm": 0.055179186165332794, "learning_rate": 9.073442923988694e-05, "loss": 0.0091, "step": 1435 }, { "epoch": 6.93719806763285, "grad_norm": 0.04829880967736244, "learning_rate": 9.071486011099665e-05, "loss": 0.006, "step": 1436 }, { "epoch": 6.942028985507246, "grad_norm": 0.05416601151227951, "learning_rate": 9.069527245360902e-05, "loss": 0.0055, "step": 1437 }, { "epoch": 6.946859903381642, "grad_norm": 0.045596893876791, "learning_rate": 9.067566627663796e-05, "loss": 0.0068, "step": 1438 }, { "epoch": 6.951690821256038, "grad_norm": 0.04718053713440895, "learning_rate": 9.065604158900585e-05, "loss": 0.0061, "step": 1439 }, { "epoch": 6.956521739130435, "grad_norm": 0.04946717619895935, "learning_rate": 9.063639839964347e-05, "loss": 0.0061, "step": 1440 }, { "epoch": 6.961352657004831, "grad_norm": 0.06101173162460327, "learning_rate": 9.061673671749005e-05, "loss": 0.0083, "step": 1441 }, { "epoch": 6.966183574879227, "grad_norm": 0.05813223123550415, "learning_rate": 9.05970565514932e-05, "loss": 0.0077, "step": 1442 }, { "epoch": 6.971014492753623, "grad_norm": 0.04281511530280113, "learning_rate": 9.057735791060897e-05, "loss": 0.0043, "step": 1443 }, { "epoch": 6.975845410628019, "grad_norm": 0.04492162540555, "learning_rate": 9.055764080380182e-05, "loss": 0.006, "step": 1444 }, { "epoch": 6.980676328502415, "grad_norm": 0.03525848686695099, "learning_rate": 9.053790524004459e-05, "loss": 0.0055, "step": 1445 }, { "epoch": 6.9855072463768115, "grad_norm": 0.0547606386244297, "learning_rate": 9.051815122831851e-05, "loss": 0.0055, "step": 1446 }, { "epoch": 6.990338164251208, "grad_norm": 0.05137596279382706, "learning_rate": 9.049837877761325e-05, "loss": 0.0072, "step": 1447 }, { "epoch": 6.995169082125604, "grad_norm": 0.08006022125482559, "learning_rate": 9.047858789692684e-05, "loss": 0.0068, "step": 1448 }, { "epoch": 7.0, "grad_norm": 0.0994105115532875, "learning_rate": 9.045877859526573e-05, "loss": 0.0079, "step": 1449 }, { "epoch": 7.004830917874396, "grad_norm": 0.049069393426179886, "learning_rate": 9.043895088164467e-05, "loss": 0.0062, "step": 1450 }, { "epoch": 7.004830917874396, "eval_loss": 0.016072051599621773, "eval_runtime": 20.6059, "eval_samples_per_second": 4.853, "eval_steps_per_second": 0.146, "step": 1450 }, { "epoch": 7.009661835748792, "grad_norm": 0.03314821794629097, "learning_rate": 9.041910476508688e-05, "loss": 0.0038, "step": 1451 }, { "epoch": 7.0144927536231885, "grad_norm": 0.047244373708963394, "learning_rate": 9.039924025462392e-05, "loss": 0.0046, "step": 1452 }, { "epoch": 7.019323671497585, "grad_norm": 0.0837038904428482, "learning_rate": 9.037935735929572e-05, "loss": 0.0075, "step": 1453 }, { "epoch": 7.024154589371981, "grad_norm": 0.031166687607765198, "learning_rate": 9.035945608815056e-05, "loss": 0.0045, "step": 1454 }, { "epoch": 7.028985507246377, "grad_norm": 0.04692283272743225, "learning_rate": 9.033953645024514e-05, "loss": 0.0049, "step": 1455 }, { "epoch": 7.033816425120773, "grad_norm": 0.057650938630104065, "learning_rate": 9.031959845464443e-05, "loss": 0.0042, "step": 1456 }, { "epoch": 7.038647342995169, "grad_norm": 0.04346553981304169, "learning_rate": 9.029964211042184e-05, "loss": 0.0045, "step": 1457 }, { "epoch": 7.043478260869565, "grad_norm": 0.04347117245197296, "learning_rate": 9.027966742665906e-05, "loss": 0.004, "step": 1458 }, { "epoch": 7.048309178743962, "grad_norm": 0.07791293412446976, "learning_rate": 9.025967441244621e-05, "loss": 0.0046, "step": 1459 }, { "epoch": 7.053140096618358, "grad_norm": 0.05497191846370697, "learning_rate": 9.023966307688164e-05, "loss": 0.0046, "step": 1460 }, { "epoch": 7.057971014492754, "grad_norm": 0.07032640278339386, "learning_rate": 9.021963342907213e-05, "loss": 0.0066, "step": 1461 }, { "epoch": 7.06280193236715, "grad_norm": 0.055110178887844086, "learning_rate": 9.019958547813277e-05, "loss": 0.0071, "step": 1462 }, { "epoch": 7.067632850241546, "grad_norm": 0.05101725459098816, "learning_rate": 9.017951923318694e-05, "loss": 0.0056, "step": 1463 }, { "epoch": 7.072463768115942, "grad_norm": 0.041964564472436905, "learning_rate": 9.015943470336643e-05, "loss": 0.0053, "step": 1464 }, { "epoch": 7.0772946859903385, "grad_norm": 0.07983453571796417, "learning_rate": 9.013933189781124e-05, "loss": 0.0059, "step": 1465 }, { "epoch": 7.082125603864735, "grad_norm": 0.05337269604206085, "learning_rate": 9.011921082566977e-05, "loss": 0.006, "step": 1466 }, { "epoch": 7.086956521739131, "grad_norm": 0.05293073132634163, "learning_rate": 9.009907149609869e-05, "loss": 0.0059, "step": 1467 }, { "epoch": 7.091787439613527, "grad_norm": 0.04248853400349617, "learning_rate": 9.007891391826304e-05, "loss": 0.0045, "step": 1468 }, { "epoch": 7.096618357487923, "grad_norm": 0.04298196732997894, "learning_rate": 9.005873810133606e-05, "loss": 0.0046, "step": 1469 }, { "epoch": 7.101449275362318, "grad_norm": 0.03962076082825661, "learning_rate": 9.003854405449939e-05, "loss": 0.0049, "step": 1470 }, { "epoch": 7.106280193236715, "grad_norm": 0.04526727646589279, "learning_rate": 9.001833178694292e-05, "loss": 0.0046, "step": 1471 }, { "epoch": 7.111111111111111, "grad_norm": 0.04316214099526405, "learning_rate": 8.999810130786484e-05, "loss": 0.0063, "step": 1472 }, { "epoch": 7.115942028985507, "grad_norm": 0.044104162603616714, "learning_rate": 8.99778526264716e-05, "loss": 0.0066, "step": 1473 }, { "epoch": 7.120772946859903, "grad_norm": 0.08153866231441498, "learning_rate": 8.9957585751978e-05, "loss": 0.0049, "step": 1474 }, { "epoch": 7.125603864734299, "grad_norm": 0.04352913051843643, "learning_rate": 8.993730069360706e-05, "loss": 0.006, "step": 1475 }, { "epoch": 7.125603864734299, "eval_loss": 0.017760971561074257, "eval_runtime": 20.5955, "eval_samples_per_second": 4.855, "eval_steps_per_second": 0.146, "step": 1475 }, { "epoch": 7.130434782608695, "grad_norm": 0.04243316128849983, "learning_rate": 8.991699746059007e-05, "loss": 0.0046, "step": 1476 }, { "epoch": 7.1352657004830915, "grad_norm": 0.04733670502901077, "learning_rate": 8.989667606216668e-05, "loss": 0.0049, "step": 1477 }, { "epoch": 7.140096618357488, "grad_norm": 0.04068618267774582, "learning_rate": 8.987633650758466e-05, "loss": 0.0048, "step": 1478 }, { "epoch": 7.144927536231884, "grad_norm": 0.04422817751765251, "learning_rate": 8.985597880610018e-05, "loss": 0.0039, "step": 1479 }, { "epoch": 7.14975845410628, "grad_norm": 0.04335469380021095, "learning_rate": 8.983560296697757e-05, "loss": 0.0045, "step": 1480 }, { "epoch": 7.154589371980676, "grad_norm": 0.03592665493488312, "learning_rate": 8.981520899948952e-05, "loss": 0.0035, "step": 1481 }, { "epoch": 7.159420289855072, "grad_norm": 0.056261006742715836, "learning_rate": 8.979479691291683e-05, "loss": 0.0045, "step": 1482 }, { "epoch": 7.164251207729468, "grad_norm": 0.06332986056804657, "learning_rate": 8.977436671654869e-05, "loss": 0.0066, "step": 1483 }, { "epoch": 7.169082125603865, "grad_norm": 0.038734789937734604, "learning_rate": 8.97539184196824e-05, "loss": 0.0032, "step": 1484 }, { "epoch": 7.173913043478261, "grad_norm": 0.05872436612844467, "learning_rate": 8.973345203162362e-05, "loss": 0.0059, "step": 1485 }, { "epoch": 7.178743961352657, "grad_norm": 0.06064346432685852, "learning_rate": 8.971296756168615e-05, "loss": 0.0067, "step": 1486 }, { "epoch": 7.183574879227053, "grad_norm": 0.04277418926358223, "learning_rate": 8.969246501919204e-05, "loss": 0.0057, "step": 1487 }, { "epoch": 7.188405797101449, "grad_norm": 0.027813758701086044, "learning_rate": 8.967194441347161e-05, "loss": 0.0028, "step": 1488 }, { "epoch": 7.193236714975845, "grad_norm": 0.0485023558139801, "learning_rate": 8.965140575386336e-05, "loss": 0.0044, "step": 1489 }, { "epoch": 7.1980676328502415, "grad_norm": 0.04239838570356369, "learning_rate": 8.963084904971397e-05, "loss": 0.0043, "step": 1490 }, { "epoch": 7.202898550724638, "grad_norm": 0.043228596448898315, "learning_rate": 8.961027431037843e-05, "loss": 0.0047, "step": 1491 }, { "epoch": 7.207729468599034, "grad_norm": 0.06608793139457703, "learning_rate": 8.958968154521985e-05, "loss": 0.0053, "step": 1492 }, { "epoch": 7.21256038647343, "grad_norm": 0.045232873409986496, "learning_rate": 8.956907076360958e-05, "loss": 0.0043, "step": 1493 }, { "epoch": 7.217391304347826, "grad_norm": 0.057870227843523026, "learning_rate": 8.954844197492719e-05, "loss": 0.0047, "step": 1494 }, { "epoch": 7.222222222222222, "grad_norm": 0.03591402992606163, "learning_rate": 8.952779518856038e-05, "loss": 0.0038, "step": 1495 }, { "epoch": 7.2270531400966185, "grad_norm": 0.05448876693844795, "learning_rate": 8.950713041390511e-05, "loss": 0.0053, "step": 1496 }, { "epoch": 7.231884057971015, "grad_norm": 0.044634390622377396, "learning_rate": 8.948644766036546e-05, "loss": 0.0042, "step": 1497 }, { "epoch": 7.236714975845411, "grad_norm": 0.04010568931698799, "learning_rate": 8.946574693735375e-05, "loss": 0.0048, "step": 1498 }, { "epoch": 7.241545893719807, "grad_norm": 0.07772307097911835, "learning_rate": 8.944502825429045e-05, "loss": 0.0058, "step": 1499 }, { "epoch": 7.246376811594203, "grad_norm": 0.06972262263298035, "learning_rate": 8.942429162060421e-05, "loss": 0.0059, "step": 1500 }, { "epoch": 7.246376811594203, "eval_loss": 0.01691284403204918, "eval_runtime": 20.613, "eval_samples_per_second": 4.851, "eval_steps_per_second": 0.146, "step": 1500 }, { "epoch": 7.251207729468599, "grad_norm": 0.04732871428132057, "learning_rate": 8.940353704573186e-05, "loss": 0.0053, "step": 1501 }, { "epoch": 7.256038647342995, "grad_norm": 0.03620045259594917, "learning_rate": 8.938276453911834e-05, "loss": 0.0042, "step": 1502 }, { "epoch": 7.260869565217392, "grad_norm": 0.06282836198806763, "learning_rate": 8.936197411021683e-05, "loss": 0.0065, "step": 1503 }, { "epoch": 7.265700483091788, "grad_norm": 0.057784438133239746, "learning_rate": 8.934116576848861e-05, "loss": 0.0049, "step": 1504 }, { "epoch": 7.270531400966184, "grad_norm": 0.04168645292520523, "learning_rate": 8.932033952340314e-05, "loss": 0.0062, "step": 1505 }, { "epoch": 7.27536231884058, "grad_norm": 0.06200752034783363, "learning_rate": 8.9299495384438e-05, "loss": 0.0062, "step": 1506 }, { "epoch": 7.280193236714976, "grad_norm": 0.03737403079867363, "learning_rate": 8.927863336107895e-05, "loss": 0.004, "step": 1507 }, { "epoch": 7.285024154589372, "grad_norm": 0.0492960587143898, "learning_rate": 8.925775346281988e-05, "loss": 0.0051, "step": 1508 }, { "epoch": 7.2898550724637685, "grad_norm": 0.046553079038858414, "learning_rate": 8.923685569916276e-05, "loss": 0.0061, "step": 1509 }, { "epoch": 7.294685990338165, "grad_norm": 0.0506129264831543, "learning_rate": 8.921594007961774e-05, "loss": 0.0049, "step": 1510 }, { "epoch": 7.29951690821256, "grad_norm": 0.05896652489900589, "learning_rate": 8.919500661370313e-05, "loss": 0.0046, "step": 1511 }, { "epoch": 7.304347826086957, "grad_norm": 0.058990173041820526, "learning_rate": 8.917405531094529e-05, "loss": 0.0052, "step": 1512 }, { "epoch": 7.309178743961352, "grad_norm": 0.05548791214823723, "learning_rate": 8.915308618087874e-05, "loss": 0.0058, "step": 1513 }, { "epoch": 7.314009661835748, "grad_norm": 0.03932204470038414, "learning_rate": 8.913209923304608e-05, "loss": 0.0053, "step": 1514 }, { "epoch": 7.318840579710145, "grad_norm": 0.05363031104207039, "learning_rate": 8.911109447699806e-05, "loss": 0.0055, "step": 1515 }, { "epoch": 7.323671497584541, "grad_norm": 0.030770711600780487, "learning_rate": 8.909007192229352e-05, "loss": 0.0036, "step": 1516 }, { "epoch": 7.328502415458937, "grad_norm": 0.06688834726810455, "learning_rate": 8.906903157849937e-05, "loss": 0.0066, "step": 1517 }, { "epoch": 7.333333333333333, "grad_norm": 0.04870985448360443, "learning_rate": 8.904797345519065e-05, "loss": 0.0047, "step": 1518 }, { "epoch": 7.338164251207729, "grad_norm": 0.037944402545690536, "learning_rate": 8.902689756195049e-05, "loss": 0.004, "step": 1519 }, { "epoch": 7.342995169082125, "grad_norm": 0.05561332404613495, "learning_rate": 8.900580390837007e-05, "loss": 0.007, "step": 1520 }, { "epoch": 7.3478260869565215, "grad_norm": 0.05865646153688431, "learning_rate": 8.89846925040487e-05, "loss": 0.0051, "step": 1521 }, { "epoch": 7.352657004830918, "grad_norm": 0.03414580225944519, "learning_rate": 8.896356335859375e-05, "loss": 0.004, "step": 1522 }, { "epoch": 7.357487922705314, "grad_norm": 0.055078282952308655, "learning_rate": 8.894241648162064e-05, "loss": 0.0045, "step": 1523 }, { "epoch": 7.36231884057971, "grad_norm": 0.044683605432510376, "learning_rate": 8.89212518827529e-05, "loss": 0.0048, "step": 1524 }, { "epoch": 7.367149758454106, "grad_norm": 0.04428521916270256, "learning_rate": 8.89000695716221e-05, "loss": 0.0043, "step": 1525 }, { "epoch": 7.367149758454106, "eval_loss": 0.01745455525815487, "eval_runtime": 20.6235, "eval_samples_per_second": 4.849, "eval_steps_per_second": 0.145, "step": 1525 }, { "epoch": 7.371980676328502, "grad_norm": 0.06754156947135925, "learning_rate": 8.887886955786786e-05, "loss": 0.0048, "step": 1526 }, { "epoch": 7.3768115942028984, "grad_norm": 0.05913345888257027, "learning_rate": 8.885765185113789e-05, "loss": 0.0066, "step": 1527 }, { "epoch": 7.381642512077295, "grad_norm": 0.06382418423891068, "learning_rate": 8.883641646108793e-05, "loss": 0.0069, "step": 1528 }, { "epoch": 7.386473429951691, "grad_norm": 0.06488420814275742, "learning_rate": 8.881516339738176e-05, "loss": 0.0067, "step": 1529 }, { "epoch": 7.391304347826087, "grad_norm": 0.04919671267271042, "learning_rate": 8.879389266969122e-05, "loss": 0.0044, "step": 1530 }, { "epoch": 7.396135265700483, "grad_norm": 0.040717657655477524, "learning_rate": 8.877260428769619e-05, "loss": 0.0038, "step": 1531 }, { "epoch": 7.400966183574879, "grad_norm": 0.06025019288063049, "learning_rate": 8.875129826108456e-05, "loss": 0.0049, "step": 1532 }, { "epoch": 7.405797101449275, "grad_norm": 0.06267979741096497, "learning_rate": 8.872997459955226e-05, "loss": 0.0056, "step": 1533 }, { "epoch": 7.4106280193236715, "grad_norm": 0.05493423342704773, "learning_rate": 8.870863331280327e-05, "loss": 0.0056, "step": 1534 }, { "epoch": 7.415458937198068, "grad_norm": 0.046555474400520325, "learning_rate": 8.868727441054958e-05, "loss": 0.0049, "step": 1535 }, { "epoch": 7.420289855072464, "grad_norm": 0.06376277655363083, "learning_rate": 8.866589790251118e-05, "loss": 0.0059, "step": 1536 }, { "epoch": 7.42512077294686, "grad_norm": 0.05157245695590973, "learning_rate": 8.864450379841604e-05, "loss": 0.006, "step": 1537 }, { "epoch": 7.429951690821256, "grad_norm": 0.054448019713163376, "learning_rate": 8.862309210800024e-05, "loss": 0.0057, "step": 1538 }, { "epoch": 7.434782608695652, "grad_norm": 0.05123317614197731, "learning_rate": 8.860166284100776e-05, "loss": 0.0057, "step": 1539 }, { "epoch": 7.4396135265700485, "grad_norm": 0.04118344932794571, "learning_rate": 8.858021600719065e-05, "loss": 0.0046, "step": 1540 }, { "epoch": 7.444444444444445, "grad_norm": 0.05112331360578537, "learning_rate": 8.855875161630887e-05, "loss": 0.0051, "step": 1541 }, { "epoch": 7.449275362318841, "grad_norm": 0.07118396461009979, "learning_rate": 8.853726967813048e-05, "loss": 0.0064, "step": 1542 }, { "epoch": 7.454106280193237, "grad_norm": 0.04440603032708168, "learning_rate": 8.851577020243144e-05, "loss": 0.005, "step": 1543 }, { "epoch": 7.458937198067633, "grad_norm": 0.033633433282375336, "learning_rate": 8.849425319899574e-05, "loss": 0.0039, "step": 1544 }, { "epoch": 7.463768115942029, "grad_norm": 0.05634831264615059, "learning_rate": 8.847271867761532e-05, "loss": 0.0051, "step": 1545 }, { "epoch": 7.468599033816425, "grad_norm": 0.05170667916536331, "learning_rate": 8.845116664809007e-05, "loss": 0.0057, "step": 1546 }, { "epoch": 7.473429951690822, "grad_norm": 0.05650138854980469, "learning_rate": 8.842959712022794e-05, "loss": 0.0051, "step": 1547 }, { "epoch": 7.478260869565218, "grad_norm": 0.05898016318678856, "learning_rate": 8.84080101038447e-05, "loss": 0.0057, "step": 1548 }, { "epoch": 7.483091787439614, "grad_norm": 0.0635242760181427, "learning_rate": 8.838640560876422e-05, "loss": 0.0046, "step": 1549 }, { "epoch": 7.48792270531401, "grad_norm": 0.08860024064779282, "learning_rate": 8.836478364481824e-05, "loss": 0.0049, "step": 1550 }, { "epoch": 7.48792270531401, "eval_loss": 0.017847729846835136, "eval_runtime": 20.6294, "eval_samples_per_second": 4.847, "eval_steps_per_second": 0.145, "step": 1550 }, { "epoch": 7.492753623188406, "grad_norm": 0.06238357722759247, "learning_rate": 8.834314422184648e-05, "loss": 0.008, "step": 1551 }, { "epoch": 7.4975845410628015, "grad_norm": 0.03494458273053169, "learning_rate": 8.832148734969657e-05, "loss": 0.0052, "step": 1552 }, { "epoch": 7.5024154589371985, "grad_norm": 0.044305965304374695, "learning_rate": 8.829981303822415e-05, "loss": 0.0047, "step": 1553 }, { "epoch": 7.507246376811594, "grad_norm": 0.0723017156124115, "learning_rate": 8.827812129729271e-05, "loss": 0.0069, "step": 1554 }, { "epoch": 7.512077294685991, "grad_norm": 0.049968816339969635, "learning_rate": 8.825641213677375e-05, "loss": 0.0042, "step": 1555 }, { "epoch": 7.516908212560386, "grad_norm": 0.05750703811645508, "learning_rate": 8.823468556654666e-05, "loss": 0.0067, "step": 1556 }, { "epoch": 7.521739130434782, "grad_norm": 0.07065118849277496, "learning_rate": 8.821294159649874e-05, "loss": 0.0067, "step": 1557 }, { "epoch": 7.526570048309178, "grad_norm": 0.046665143221616745, "learning_rate": 8.819118023652525e-05, "loss": 0.0042, "step": 1558 }, { "epoch": 7.531400966183575, "grad_norm": 0.06773227453231812, "learning_rate": 8.81694014965293e-05, "loss": 0.0058, "step": 1559 }, { "epoch": 7.536231884057971, "grad_norm": 0.04640137404203415, "learning_rate": 8.814760538642198e-05, "loss": 0.0056, "step": 1560 }, { "epoch": 7.541062801932367, "grad_norm": 0.04420126974582672, "learning_rate": 8.812579191612224e-05, "loss": 0.0052, "step": 1561 }, { "epoch": 7.545893719806763, "grad_norm": 0.05039765685796738, "learning_rate": 8.810396109555695e-05, "loss": 0.0054, "step": 1562 }, { "epoch": 7.550724637681159, "grad_norm": 0.050329990684986115, "learning_rate": 8.808211293466084e-05, "loss": 0.0051, "step": 1563 }, { "epoch": 7.555555555555555, "grad_norm": 0.05938250944018364, "learning_rate": 8.80602474433766e-05, "loss": 0.0063, "step": 1564 }, { "epoch": 7.5603864734299515, "grad_norm": 0.041535988450050354, "learning_rate": 8.803836463165476e-05, "loss": 0.0053, "step": 1565 }, { "epoch": 7.565217391304348, "grad_norm": 0.051995351910591125, "learning_rate": 8.801646450945371e-05, "loss": 0.0043, "step": 1566 }, { "epoch": 7.570048309178744, "grad_norm": 0.04369938373565674, "learning_rate": 8.799454708673977e-05, "loss": 0.0048, "step": 1567 }, { "epoch": 7.57487922705314, "grad_norm": 0.04908826947212219, "learning_rate": 8.79726123734871e-05, "loss": 0.0053, "step": 1568 }, { "epoch": 7.579710144927536, "grad_norm": 0.05574946478009224, "learning_rate": 8.795066037967775e-05, "loss": 0.0067, "step": 1569 }, { "epoch": 7.584541062801932, "grad_norm": 0.04000121355056763, "learning_rate": 8.792869111530161e-05, "loss": 0.0047, "step": 1570 }, { "epoch": 7.5893719806763285, "grad_norm": 0.060283225029706955, "learning_rate": 8.790670459035645e-05, "loss": 0.0072, "step": 1571 }, { "epoch": 7.594202898550725, "grad_norm": 0.06155461445450783, "learning_rate": 8.788470081484787e-05, "loss": 0.0069, "step": 1572 }, { "epoch": 7.599033816425121, "grad_norm": 0.05512505769729614, "learning_rate": 8.786267979878934e-05, "loss": 0.0057, "step": 1573 }, { "epoch": 7.603864734299517, "grad_norm": 0.037634339183568954, "learning_rate": 8.78406415522022e-05, "loss": 0.0042, "step": 1574 }, { "epoch": 7.608695652173913, "grad_norm": 0.05426057055592537, "learning_rate": 8.78185860851156e-05, "loss": 0.0058, "step": 1575 }, { "epoch": 7.608695652173913, "eval_loss": 0.015573016367852688, "eval_runtime": 20.5607, "eval_samples_per_second": 4.864, "eval_steps_per_second": 0.146, "step": 1575 }, { "epoch": 7.613526570048309, "grad_norm": 0.07868023961782455, "learning_rate": 8.779651340756647e-05, "loss": 0.0053, "step": 1576 }, { "epoch": 7.618357487922705, "grad_norm": 0.06837618350982666, "learning_rate": 8.77744235295997e-05, "loss": 0.0044, "step": 1577 }, { "epoch": 7.6231884057971016, "grad_norm": 0.05811638385057449, "learning_rate": 8.775231646126791e-05, "loss": 0.0053, "step": 1578 }, { "epoch": 7.628019323671498, "grad_norm": 0.03706006333231926, "learning_rate": 8.77301922126316e-05, "loss": 0.0047, "step": 1579 }, { "epoch": 7.632850241545894, "grad_norm": 0.044062525033950806, "learning_rate": 8.770805079375902e-05, "loss": 0.0052, "step": 1580 }, { "epoch": 7.63768115942029, "grad_norm": 0.05137691646814346, "learning_rate": 8.768589221472632e-05, "loss": 0.0051, "step": 1581 }, { "epoch": 7.642512077294686, "grad_norm": 0.04805343598127365, "learning_rate": 8.766371648561738e-05, "loss": 0.0053, "step": 1582 }, { "epoch": 7.647342995169082, "grad_norm": 0.046713635325431824, "learning_rate": 8.764152361652392e-05, "loss": 0.0054, "step": 1583 }, { "epoch": 7.6521739130434785, "grad_norm": 0.04009406641125679, "learning_rate": 8.761931361754547e-05, "loss": 0.005, "step": 1584 }, { "epoch": 7.657004830917875, "grad_norm": 0.055948395282030106, "learning_rate": 8.759708649878935e-05, "loss": 0.005, "step": 1585 }, { "epoch": 7.661835748792271, "grad_norm": 0.052176788449287415, "learning_rate": 8.757484227037068e-05, "loss": 0.0061, "step": 1586 }, { "epoch": 7.666666666666667, "grad_norm": 0.05058382824063301, "learning_rate": 8.75525809424123e-05, "loss": 0.0047, "step": 1587 }, { "epoch": 7.671497584541063, "grad_norm": 0.03293626755475998, "learning_rate": 8.753030252504494e-05, "loss": 0.0045, "step": 1588 }, { "epoch": 7.676328502415459, "grad_norm": 0.038541220128536224, "learning_rate": 8.750800702840702e-05, "loss": 0.0045, "step": 1589 }, { "epoch": 7.681159420289855, "grad_norm": 0.04410076513886452, "learning_rate": 8.748569446264479e-05, "loss": 0.0047, "step": 1590 }, { "epoch": 7.685990338164252, "grad_norm": 0.050878047943115234, "learning_rate": 8.74633648379122e-05, "loss": 0.0051, "step": 1591 }, { "epoch": 7.690821256038648, "grad_norm": 0.0634491965174675, "learning_rate": 8.744101816437103e-05, "loss": 0.0058, "step": 1592 }, { "epoch": 7.695652173913043, "grad_norm": 0.06506568938493729, "learning_rate": 8.741865445219077e-05, "loss": 0.0043, "step": 1593 }, { "epoch": 7.70048309178744, "grad_norm": 0.043520476669073105, "learning_rate": 8.739627371154872e-05, "loss": 0.0044, "step": 1594 }, { "epoch": 7.705314009661835, "grad_norm": 0.04149644076824188, "learning_rate": 8.73738759526299e-05, "loss": 0.0045, "step": 1595 }, { "epoch": 7.710144927536232, "grad_norm": 0.04817346855998039, "learning_rate": 8.735146118562703e-05, "loss": 0.0054, "step": 1596 }, { "epoch": 7.714975845410628, "grad_norm": 0.07699626684188843, "learning_rate": 8.732902942074061e-05, "loss": 0.0074, "step": 1597 }, { "epoch": 7.719806763285024, "grad_norm": 0.06528923660516739, "learning_rate": 8.730658066817893e-05, "loss": 0.0059, "step": 1598 }, { "epoch": 7.72463768115942, "grad_norm": 0.052303560078144073, "learning_rate": 8.728411493815791e-05, "loss": 0.0058, "step": 1599 }, { "epoch": 7.729468599033816, "grad_norm": 0.05582994595170021, "learning_rate": 8.726163224090125e-05, "loss": 0.0062, "step": 1600 }, { "epoch": 7.729468599033816, "eval_loss": 0.015776341781020164, "eval_runtime": 20.5969, "eval_samples_per_second": 4.855, "eval_steps_per_second": 0.146, "step": 1600 }, { "epoch": 7.734299516908212, "grad_norm": 0.0504007451236248, "learning_rate": 8.723913258664038e-05, "loss": 0.0061, "step": 1601 }, { "epoch": 7.739130434782608, "grad_norm": 0.04513921961188316, "learning_rate": 8.721661598561442e-05, "loss": 0.0048, "step": 1602 }, { "epoch": 7.743961352657005, "grad_norm": 0.050863321870565414, "learning_rate": 8.719408244807021e-05, "loss": 0.0058, "step": 1603 }, { "epoch": 7.748792270531401, "grad_norm": 0.034483492374420166, "learning_rate": 8.717153198426232e-05, "loss": 0.0046, "step": 1604 }, { "epoch": 7.753623188405797, "grad_norm": 0.05862283334136009, "learning_rate": 8.714896460445297e-05, "loss": 0.0061, "step": 1605 }, { "epoch": 7.758454106280193, "grad_norm": 0.041233327239751816, "learning_rate": 8.712638031891212e-05, "loss": 0.0052, "step": 1606 }, { "epoch": 7.763285024154589, "grad_norm": 0.051079899072647095, "learning_rate": 8.710377913791746e-05, "loss": 0.0062, "step": 1607 }, { "epoch": 7.768115942028985, "grad_norm": 0.05466492846608162, "learning_rate": 8.708116107175426e-05, "loss": 0.0047, "step": 1608 }, { "epoch": 7.7729468599033815, "grad_norm": 0.03811094909906387, "learning_rate": 8.705852613071557e-05, "loss": 0.0044, "step": 1609 }, { "epoch": 7.777777777777778, "grad_norm": 0.06326041370630264, "learning_rate": 8.703587432510207e-05, "loss": 0.0061, "step": 1610 }, { "epoch": 7.782608695652174, "grad_norm": 0.04274899885058403, "learning_rate": 8.701320566522217e-05, "loss": 0.0038, "step": 1611 }, { "epoch": 7.78743961352657, "grad_norm": 0.06019265204668045, "learning_rate": 8.699052016139188e-05, "loss": 0.007, "step": 1612 }, { "epoch": 7.792270531400966, "grad_norm": 0.04409782588481903, "learning_rate": 8.696781782393491e-05, "loss": 0.0044, "step": 1613 }, { "epoch": 7.797101449275362, "grad_norm": 0.035827215760946274, "learning_rate": 8.694509866318263e-05, "loss": 0.004, "step": 1614 }, { "epoch": 7.8019323671497585, "grad_norm": 0.04697452858090401, "learning_rate": 8.692236268947408e-05, "loss": 0.0043, "step": 1615 }, { "epoch": 7.806763285024155, "grad_norm": 0.056868743151426315, "learning_rate": 8.68996099131559e-05, "loss": 0.0059, "step": 1616 }, { "epoch": 7.811594202898551, "grad_norm": 0.04871121421456337, "learning_rate": 8.687684034458245e-05, "loss": 0.0068, "step": 1617 }, { "epoch": 7.816425120772947, "grad_norm": 0.04334891214966774, "learning_rate": 8.685405399411568e-05, "loss": 0.0048, "step": 1618 }, { "epoch": 7.821256038647343, "grad_norm": 0.08146171271800995, "learning_rate": 8.683125087212518e-05, "loss": 0.0048, "step": 1619 }, { "epoch": 7.826086956521739, "grad_norm": 0.04689481481909752, "learning_rate": 8.680843098898819e-05, "loss": 0.0055, "step": 1620 }, { "epoch": 7.830917874396135, "grad_norm": 0.0372406505048275, "learning_rate": 8.678559435508958e-05, "loss": 0.0054, "step": 1621 }, { "epoch": 7.835748792270532, "grad_norm": 0.045953404158353806, "learning_rate": 8.67627409808218e-05, "loss": 0.0042, "step": 1622 }, { "epoch": 7.840579710144928, "grad_norm": 0.04336479306221008, "learning_rate": 8.6739870876585e-05, "loss": 0.0056, "step": 1623 }, { "epoch": 7.845410628019324, "grad_norm": 0.055914826691150665, "learning_rate": 8.671698405278685e-05, "loss": 0.0058, "step": 1624 }, { "epoch": 7.85024154589372, "grad_norm": 0.05467041954398155, "learning_rate": 8.66940805198427e-05, "loss": 0.0045, "step": 1625 }, { "epoch": 7.85024154589372, "eval_loss": 0.015103131532669067, "eval_runtime": 21.329, "eval_samples_per_second": 4.688, "eval_steps_per_second": 0.141, "step": 1625 }, { "epoch": 7.855072463768116, "grad_norm": 0.042717594653367996, "learning_rate": 8.667116028817548e-05, "loss": 0.0056, "step": 1626 }, { "epoch": 7.859903381642512, "grad_norm": 0.0534902848303318, "learning_rate": 8.66482233682157e-05, "loss": 0.0055, "step": 1627 }, { "epoch": 7.8647342995169085, "grad_norm": 0.040888525545597076, "learning_rate": 8.662526977040148e-05, "loss": 0.0046, "step": 1628 }, { "epoch": 7.869565217391305, "grad_norm": 0.04883519187569618, "learning_rate": 8.660229950517857e-05, "loss": 0.005, "step": 1629 }, { "epoch": 7.874396135265701, "grad_norm": 0.039643142372369766, "learning_rate": 8.657931258300023e-05, "loss": 0.0041, "step": 1630 }, { "epoch": 7.879227053140097, "grad_norm": 0.039371222257614136, "learning_rate": 8.655630901432734e-05, "loss": 0.0048, "step": 1631 }, { "epoch": 7.884057971014493, "grad_norm": 0.04320335388183594, "learning_rate": 8.653328880962836e-05, "loss": 0.0038, "step": 1632 }, { "epoch": 7.888888888888889, "grad_norm": 0.06078319251537323, "learning_rate": 8.651025197937931e-05, "loss": 0.0051, "step": 1633 }, { "epoch": 7.8937198067632846, "grad_norm": 0.05101476237177849, "learning_rate": 8.648719853406377e-05, "loss": 0.0055, "step": 1634 }, { "epoch": 7.898550724637682, "grad_norm": 0.06241091340780258, "learning_rate": 8.646412848417292e-05, "loss": 0.0052, "step": 1635 }, { "epoch": 7.903381642512077, "grad_norm": 0.0694604367017746, "learning_rate": 8.644104184020542e-05, "loss": 0.007, "step": 1636 }, { "epoch": 7.908212560386474, "grad_norm": 0.04248025640845299, "learning_rate": 8.641793861266757e-05, "loss": 0.0054, "step": 1637 }, { "epoch": 7.913043478260869, "grad_norm": 0.04426901042461395, "learning_rate": 8.639481881207314e-05, "loss": 0.0056, "step": 1638 }, { "epoch": 7.917874396135265, "grad_norm": 0.05379689857363701, "learning_rate": 8.637168244894351e-05, "loss": 0.0052, "step": 1639 }, { "epoch": 7.9227053140096615, "grad_norm": 0.05331415683031082, "learning_rate": 8.634852953380756e-05, "loss": 0.0063, "step": 1640 }, { "epoch": 7.927536231884058, "grad_norm": 0.05386782065033913, "learning_rate": 8.632536007720167e-05, "loss": 0.0059, "step": 1641 }, { "epoch": 7.932367149758454, "grad_norm": 0.07617385685443878, "learning_rate": 8.630217408966983e-05, "loss": 0.0062, "step": 1642 }, { "epoch": 7.93719806763285, "grad_norm": 0.05009270831942558, "learning_rate": 8.627897158176346e-05, "loss": 0.0044, "step": 1643 }, { "epoch": 7.942028985507246, "grad_norm": 0.043669749051332474, "learning_rate": 8.62557525640416e-05, "loss": 0.0054, "step": 1644 }, { "epoch": 7.946859903381642, "grad_norm": 0.04212132841348648, "learning_rate": 8.623251704707071e-05, "loss": 0.0041, "step": 1645 }, { "epoch": 7.951690821256038, "grad_norm": 0.04619299992918968, "learning_rate": 8.62092650414248e-05, "loss": 0.0049, "step": 1646 }, { "epoch": 7.956521739130435, "grad_norm": 0.04855784773826599, "learning_rate": 8.61859965576854e-05, "loss": 0.0051, "step": 1647 }, { "epoch": 7.961352657004831, "grad_norm": 0.051754970103502274, "learning_rate": 8.616271160644149e-05, "loss": 0.0061, "step": 1648 }, { "epoch": 7.966183574879227, "grad_norm": 0.04847295209765434, "learning_rate": 8.613941019828961e-05, "loss": 0.0053, "step": 1649 }, { "epoch": 7.971014492753623, "grad_norm": 0.04961397871375084, "learning_rate": 8.611609234383374e-05, "loss": 0.0054, "step": 1650 }, { "epoch": 7.971014492753623, "eval_loss": 0.014968699775636196, "eval_runtime": 20.6206, "eval_samples_per_second": 4.85, "eval_steps_per_second": 0.145, "step": 1650 }, { "epoch": 7.975845410628019, "grad_norm": 0.09579728543758392, "learning_rate": 8.609275805368532e-05, "loss": 0.0063, "step": 1651 }, { "epoch": 7.980676328502415, "grad_norm": 0.03453390672802925, "learning_rate": 8.606940733846335e-05, "loss": 0.0037, "step": 1652 }, { "epoch": 7.9855072463768115, "grad_norm": 0.03772159665822983, "learning_rate": 8.604604020879427e-05, "loss": 0.004, "step": 1653 }, { "epoch": 7.990338164251208, "grad_norm": 0.04900532588362694, "learning_rate": 8.602265667531193e-05, "loss": 0.0064, "step": 1654 }, { "epoch": 7.995169082125604, "grad_norm": 0.049145378172397614, "learning_rate": 8.599925674865774e-05, "loss": 0.0057, "step": 1655 }, { "epoch": 8.0, "grad_norm": 0.08164945989847183, "learning_rate": 8.597584043948053e-05, "loss": 0.0057, "step": 1656 }, { "epoch": 8.004830917874395, "grad_norm": 0.05199301242828369, "learning_rate": 8.595240775843653e-05, "loss": 0.004, "step": 1657 }, { "epoch": 8.009661835748792, "grad_norm": 0.0596538744866848, "learning_rate": 8.592895871618953e-05, "loss": 0.0046, "step": 1658 }, { "epoch": 8.014492753623188, "grad_norm": 0.0621766597032547, "learning_rate": 8.590549332341068e-05, "loss": 0.0041, "step": 1659 }, { "epoch": 8.019323671497585, "grad_norm": 0.03637773171067238, "learning_rate": 8.58820115907786e-05, "loss": 0.0032, "step": 1660 }, { "epoch": 8.02415458937198, "grad_norm": 0.04378081485629082, "learning_rate": 8.585851352897935e-05, "loss": 0.0041, "step": 1661 }, { "epoch": 8.028985507246377, "grad_norm": 0.04154336825013161, "learning_rate": 8.58349991487064e-05, "loss": 0.0036, "step": 1662 }, { "epoch": 8.033816425120772, "grad_norm": 0.04366371035575867, "learning_rate": 8.581146846066071e-05, "loss": 0.0044, "step": 1663 }, { "epoch": 8.03864734299517, "grad_norm": 0.04134861007332802, "learning_rate": 8.578792147555055e-05, "loss": 0.0047, "step": 1664 }, { "epoch": 8.043478260869565, "grad_norm": 0.051076147705316544, "learning_rate": 8.576435820409171e-05, "loss": 0.0039, "step": 1665 }, { "epoch": 8.048309178743962, "grad_norm": 0.04088986665010452, "learning_rate": 8.574077865700734e-05, "loss": 0.0044, "step": 1666 }, { "epoch": 8.053140096618357, "grad_norm": 0.056600384414196014, "learning_rate": 8.5717182845028e-05, "loss": 0.0036, "step": 1667 }, { "epoch": 8.057971014492754, "grad_norm": 0.05603436380624771, "learning_rate": 8.569357077889167e-05, "loss": 0.0043, "step": 1668 }, { "epoch": 8.06280193236715, "grad_norm": 0.04432399570941925, "learning_rate": 8.566994246934372e-05, "loss": 0.0041, "step": 1669 }, { "epoch": 8.067632850241546, "grad_norm": 0.04413706064224243, "learning_rate": 8.564629792713691e-05, "loss": 0.0039, "step": 1670 }, { "epoch": 8.072463768115941, "grad_norm": 0.04078343138098717, "learning_rate": 8.562263716303138e-05, "loss": 0.0038, "step": 1671 }, { "epoch": 8.077294685990339, "grad_norm": 0.05788585543632507, "learning_rate": 8.559896018779463e-05, "loss": 0.0044, "step": 1672 }, { "epoch": 8.082125603864734, "grad_norm": 0.04225607216358185, "learning_rate": 8.557526701220162e-05, "loss": 0.0043, "step": 1673 }, { "epoch": 8.08695652173913, "grad_norm": 0.0805094763636589, "learning_rate": 8.55515576470346e-05, "loss": 0.0049, "step": 1674 }, { "epoch": 8.091787439613526, "grad_norm": 0.03907567262649536, "learning_rate": 8.552783210308321e-05, "loss": 0.0042, "step": 1675 }, { "epoch": 8.091787439613526, "eval_loss": 0.01567777618765831, "eval_runtime": 20.6168, "eval_samples_per_second": 4.85, "eval_steps_per_second": 0.146, "step": 1675 }, { "epoch": 8.096618357487923, "grad_norm": 0.0445052869617939, "learning_rate": 8.550409039114447e-05, "loss": 0.0036, "step": 1676 }, { "epoch": 8.101449275362318, "grad_norm": 0.043486401438713074, "learning_rate": 8.548033252202274e-05, "loss": 0.0043, "step": 1677 }, { "epoch": 8.106280193236715, "grad_norm": 0.03254947438836098, "learning_rate": 8.545655850652972e-05, "loss": 0.0038, "step": 1678 }, { "epoch": 8.11111111111111, "grad_norm": 0.04751577973365784, "learning_rate": 8.543276835548452e-05, "loss": 0.0036, "step": 1679 }, { "epoch": 8.115942028985508, "grad_norm": 0.042737722396850586, "learning_rate": 8.540896207971352e-05, "loss": 0.0042, "step": 1680 }, { "epoch": 8.120772946859903, "grad_norm": 0.03660351410508156, "learning_rate": 8.538513969005047e-05, "loss": 0.0038, "step": 1681 }, { "epoch": 8.1256038647343, "grad_norm": 0.0444486103951931, "learning_rate": 8.536130119733647e-05, "loss": 0.0042, "step": 1682 }, { "epoch": 8.130434782608695, "grad_norm": 0.036316633224487305, "learning_rate": 8.53374466124199e-05, "loss": 0.003, "step": 1683 }, { "epoch": 8.135265700483092, "grad_norm": 0.0428357869386673, "learning_rate": 8.53135759461565e-05, "loss": 0.0036, "step": 1684 }, { "epoch": 8.140096618357488, "grad_norm": 0.03688013181090355, "learning_rate": 8.528968920940933e-05, "loss": 0.0038, "step": 1685 }, { "epoch": 8.144927536231885, "grad_norm": 0.041358936578035355, "learning_rate": 8.526578641304874e-05, "loss": 0.0037, "step": 1686 }, { "epoch": 8.14975845410628, "grad_norm": 0.047944411635398865, "learning_rate": 8.524186756795242e-05, "loss": 0.0036, "step": 1687 }, { "epoch": 8.154589371980677, "grad_norm": 0.09513163566589355, "learning_rate": 8.521793268500535e-05, "loss": 0.0061, "step": 1688 }, { "epoch": 8.159420289855072, "grad_norm": 0.025504419580101967, "learning_rate": 8.519398177509978e-05, "loss": 0.0029, "step": 1689 }, { "epoch": 8.16425120772947, "grad_norm": 0.05210709571838379, "learning_rate": 8.51700148491353e-05, "loss": 0.0034, "step": 1690 }, { "epoch": 8.169082125603865, "grad_norm": 0.044690266251564026, "learning_rate": 8.514603191801873e-05, "loss": 0.0038, "step": 1691 }, { "epoch": 8.173913043478262, "grad_norm": 0.04786112904548645, "learning_rate": 8.512203299266425e-05, "loss": 0.0046, "step": 1692 }, { "epoch": 8.178743961352657, "grad_norm": 0.03577842935919762, "learning_rate": 8.509801808399326e-05, "loss": 0.0038, "step": 1693 }, { "epoch": 8.183574879227054, "grad_norm": 0.07160219550132751, "learning_rate": 8.507398720293447e-05, "loss": 0.0064, "step": 1694 }, { "epoch": 8.18840579710145, "grad_norm": 0.07394543290138245, "learning_rate": 8.504994036042381e-05, "loss": 0.0054, "step": 1695 }, { "epoch": 8.193236714975846, "grad_norm": 0.034880660474300385, "learning_rate": 8.502587756740452e-05, "loss": 0.0034, "step": 1696 }, { "epoch": 8.198067632850242, "grad_norm": 0.0353291779756546, "learning_rate": 8.500179883482709e-05, "loss": 0.0041, "step": 1697 }, { "epoch": 8.202898550724637, "grad_norm": 0.03237657994031906, "learning_rate": 8.497770417364925e-05, "loss": 0.0035, "step": 1698 }, { "epoch": 8.207729468599034, "grad_norm": 0.04432433471083641, "learning_rate": 8.495359359483599e-05, "loss": 0.0057, "step": 1699 }, { "epoch": 8.21256038647343, "grad_norm": 0.037772729992866516, "learning_rate": 8.492946710935953e-05, "loss": 0.0039, "step": 1700 }, { "epoch": 8.21256038647343, "eval_loss": 0.015713628381490707, "eval_runtime": 20.6015, "eval_samples_per_second": 4.854, "eval_steps_per_second": 0.146, "step": 1700 }, { "epoch": 8.217391304347826, "grad_norm": 0.03376033529639244, "learning_rate": 8.490532472819937e-05, "loss": 0.0036, "step": 1701 }, { "epoch": 8.222222222222221, "grad_norm": 0.03267105296254158, "learning_rate": 8.488116646234215e-05, "loss": 0.003, "step": 1702 }, { "epoch": 8.227053140096618, "grad_norm": 0.029169436544179916, "learning_rate": 8.485699232278186e-05, "loss": 0.0028, "step": 1703 }, { "epoch": 8.231884057971014, "grad_norm": 0.04096829146146774, "learning_rate": 8.483280232051962e-05, "loss": 0.0048, "step": 1704 }, { "epoch": 8.23671497584541, "grad_norm": 0.08265417814254761, "learning_rate": 8.480859646656382e-05, "loss": 0.0051, "step": 1705 }, { "epoch": 8.241545893719806, "grad_norm": 0.043714456260204315, "learning_rate": 8.478437477193006e-05, "loss": 0.0045, "step": 1706 }, { "epoch": 8.246376811594203, "grad_norm": 0.03147174045443535, "learning_rate": 8.476013724764112e-05, "loss": 0.0033, "step": 1707 }, { "epoch": 8.251207729468598, "grad_norm": 0.04156283661723137, "learning_rate": 8.4735883904727e-05, "loss": 0.0042, "step": 1708 }, { "epoch": 8.256038647342995, "grad_norm": 0.05636567994952202, "learning_rate": 8.471161475422487e-05, "loss": 0.0046, "step": 1709 }, { "epoch": 8.26086956521739, "grad_norm": 0.04402732104063034, "learning_rate": 8.468732980717918e-05, "loss": 0.0042, "step": 1710 }, { "epoch": 8.265700483091788, "grad_norm": 0.03477048873901367, "learning_rate": 8.466302907464147e-05, "loss": 0.0035, "step": 1711 }, { "epoch": 8.270531400966183, "grad_norm": 0.0782352164387703, "learning_rate": 8.463871256767053e-05, "loss": 0.0044, "step": 1712 }, { "epoch": 8.27536231884058, "grad_norm": 0.04401532560586929, "learning_rate": 8.461438029733228e-05, "loss": 0.004, "step": 1713 }, { "epoch": 8.280193236714975, "grad_norm": 0.04359148442745209, "learning_rate": 8.459003227469985e-05, "loss": 0.0057, "step": 1714 }, { "epoch": 8.285024154589372, "grad_norm": 0.03307943418622017, "learning_rate": 8.456566851085354e-05, "loss": 0.003, "step": 1715 }, { "epoch": 8.289855072463768, "grad_norm": 0.04660947248339653, "learning_rate": 8.454128901688077e-05, "loss": 0.0036, "step": 1716 }, { "epoch": 8.294685990338165, "grad_norm": 0.05288444831967354, "learning_rate": 8.451689380387616e-05, "loss": 0.0047, "step": 1717 }, { "epoch": 8.29951690821256, "grad_norm": 0.03572872653603554, "learning_rate": 8.449248288294145e-05, "loss": 0.004, "step": 1718 }, { "epoch": 8.304347826086957, "grad_norm": 0.04837818816304207, "learning_rate": 8.446805626518559e-05, "loss": 0.0047, "step": 1719 }, { "epoch": 8.309178743961352, "grad_norm": 0.05197962373495102, "learning_rate": 8.444361396172462e-05, "loss": 0.0037, "step": 1720 }, { "epoch": 8.31400966183575, "grad_norm": 0.04145919159054756, "learning_rate": 8.44191559836817e-05, "loss": 0.0043, "step": 1721 }, { "epoch": 8.318840579710145, "grad_norm": 0.03252958506345749, "learning_rate": 8.439468234218721e-05, "loss": 0.0038, "step": 1722 }, { "epoch": 8.323671497584542, "grad_norm": 0.058449309319257736, "learning_rate": 8.437019304837855e-05, "loss": 0.0066, "step": 1723 }, { "epoch": 8.328502415458937, "grad_norm": 0.041896555572748184, "learning_rate": 8.43456881134003e-05, "loss": 0.004, "step": 1724 }, { "epoch": 8.333333333333334, "grad_norm": 0.047328148037195206, "learning_rate": 8.43211675484042e-05, "loss": 0.0046, "step": 1725 }, { "epoch": 8.333333333333334, "eval_loss": 0.016981441527605057, "eval_runtime": 20.6049, "eval_samples_per_second": 4.853, "eval_steps_per_second": 0.146, "step": 1725 }, { "epoch": 8.33816425120773, "grad_norm": 0.02978409081697464, "learning_rate": 8.4296631364549e-05, "loss": 0.0035, "step": 1726 }, { "epoch": 8.342995169082126, "grad_norm": 0.043455787003040314, "learning_rate": 8.427207957300062e-05, "loss": 0.0046, "step": 1727 }, { "epoch": 8.347826086956522, "grad_norm": 0.03001406230032444, "learning_rate": 8.424751218493213e-05, "loss": 0.0029, "step": 1728 }, { "epoch": 8.352657004830919, "grad_norm": 0.03560645505785942, "learning_rate": 8.422292921152358e-05, "loss": 0.0043, "step": 1729 }, { "epoch": 8.357487922705314, "grad_norm": 0.052797071635723114, "learning_rate": 8.419833066396222e-05, "loss": 0.0053, "step": 1730 }, { "epoch": 8.36231884057971, "grad_norm": 0.06399412453174591, "learning_rate": 8.417371655344232e-05, "loss": 0.0062, "step": 1731 }, { "epoch": 8.367149758454106, "grad_norm": 0.039564114063978195, "learning_rate": 8.414908689116528e-05, "loss": 0.0058, "step": 1732 }, { "epoch": 8.371980676328503, "grad_norm": 0.04812881350517273, "learning_rate": 8.412444168833953e-05, "loss": 0.0046, "step": 1733 }, { "epoch": 8.376811594202898, "grad_norm": 0.03506116196513176, "learning_rate": 8.40997809561806e-05, "loss": 0.0032, "step": 1734 }, { "epoch": 8.381642512077295, "grad_norm": 0.03764081746339798, "learning_rate": 8.40751047059111e-05, "loss": 0.0033, "step": 1735 }, { "epoch": 8.38647342995169, "grad_norm": 0.03246073052287102, "learning_rate": 8.405041294876066e-05, "loss": 0.0043, "step": 1736 }, { "epoch": 8.391304347826088, "grad_norm": 0.03924914821982384, "learning_rate": 8.402570569596601e-05, "loss": 0.0043, "step": 1737 }, { "epoch": 8.396135265700483, "grad_norm": 0.05545099079608917, "learning_rate": 8.400098295877092e-05, "loss": 0.0045, "step": 1738 }, { "epoch": 8.40096618357488, "grad_norm": 0.04152775928378105, "learning_rate": 8.397624474842617e-05, "loss": 0.0036, "step": 1739 }, { "epoch": 8.405797101449275, "grad_norm": 0.05080555006861687, "learning_rate": 8.395149107618965e-05, "loss": 0.0031, "step": 1740 }, { "epoch": 8.41062801932367, "grad_norm": 0.03303583338856697, "learning_rate": 8.392672195332622e-05, "loss": 0.0042, "step": 1741 }, { "epoch": 8.415458937198068, "grad_norm": 0.0483396053314209, "learning_rate": 8.39019373911078e-05, "loss": 0.004, "step": 1742 }, { "epoch": 8.420289855072463, "grad_norm": 0.06342915445566177, "learning_rate": 8.387713740081335e-05, "loss": 0.0048, "step": 1743 }, { "epoch": 8.42512077294686, "grad_norm": 0.034104350954294205, "learning_rate": 8.385232199372885e-05, "loss": 0.0036, "step": 1744 }, { "epoch": 8.429951690821255, "grad_norm": 0.06683149188756943, "learning_rate": 8.382749118114725e-05, "loss": 0.0051, "step": 1745 }, { "epoch": 8.434782608695652, "grad_norm": 0.041379477828741074, "learning_rate": 8.380264497436856e-05, "loss": 0.0041, "step": 1746 }, { "epoch": 8.439613526570048, "grad_norm": 0.044812314212322235, "learning_rate": 8.377778338469977e-05, "loss": 0.0042, "step": 1747 }, { "epoch": 8.444444444444445, "grad_norm": 0.04295221343636513, "learning_rate": 8.37529064234549e-05, "loss": 0.0035, "step": 1748 }, { "epoch": 8.44927536231884, "grad_norm": 0.03582591563463211, "learning_rate": 8.372801410195492e-05, "loss": 0.0031, "step": 1749 }, { "epoch": 8.454106280193237, "grad_norm": 0.030500952154397964, "learning_rate": 8.370310643152782e-05, "loss": 0.0025, "step": 1750 }, { "epoch": 8.454106280193237, "eval_loss": 0.015351679176092148, "eval_runtime": 20.6217, "eval_samples_per_second": 4.849, "eval_steps_per_second": 0.145, "step": 1750 }, { "epoch": 8.458937198067632, "grad_norm": 0.08864534646272659, "learning_rate": 8.36781834235086e-05, "loss": 0.0089, "step": 1751 }, { "epoch": 8.46376811594203, "grad_norm": 0.049155447632074356, "learning_rate": 8.365324508923915e-05, "loss": 0.0038, "step": 1752 }, { "epoch": 8.468599033816425, "grad_norm": 0.039812907576560974, "learning_rate": 8.362829144006846e-05, "loss": 0.0044, "step": 1753 }, { "epoch": 8.473429951690822, "grad_norm": 0.038564134389162064, "learning_rate": 8.360332248735237e-05, "loss": 0.0036, "step": 1754 }, { "epoch": 8.478260869565217, "grad_norm": 0.05134737864136696, "learning_rate": 8.357833824245376e-05, "loss": 0.0045, "step": 1755 }, { "epoch": 8.483091787439614, "grad_norm": 0.032198935747146606, "learning_rate": 8.355333871674246e-05, "loss": 0.0031, "step": 1756 }, { "epoch": 8.48792270531401, "grad_norm": 0.03891000896692276, "learning_rate": 8.352832392159523e-05, "loss": 0.0035, "step": 1757 }, { "epoch": 8.492753623188406, "grad_norm": 0.03442859277129173, "learning_rate": 8.350329386839577e-05, "loss": 0.0037, "step": 1758 }, { "epoch": 8.497584541062801, "grad_norm": 0.04633500054478645, "learning_rate": 8.347824856853476e-05, "loss": 0.0044, "step": 1759 }, { "epoch": 8.502415458937199, "grad_norm": 0.08129475265741348, "learning_rate": 8.345318803340982e-05, "loss": 0.0078, "step": 1760 }, { "epoch": 8.507246376811594, "grad_norm": 0.05236010625958443, "learning_rate": 8.342811227442548e-05, "loss": 0.0068, "step": 1761 }, { "epoch": 8.51207729468599, "grad_norm": 0.04416520893573761, "learning_rate": 8.340302130299318e-05, "loss": 0.0036, "step": 1762 }, { "epoch": 8.516908212560386, "grad_norm": 0.04746413603425026, "learning_rate": 8.33779151305313e-05, "loss": 0.0038, "step": 1763 }, { "epoch": 8.521739130434783, "grad_norm": 0.05367453023791313, "learning_rate": 8.33527937684652e-05, "loss": 0.0044, "step": 1764 }, { "epoch": 8.526570048309178, "grad_norm": 0.04992407187819481, "learning_rate": 8.332765722822702e-05, "loss": 0.0046, "step": 1765 }, { "epoch": 8.531400966183575, "grad_norm": 0.04729165509343147, "learning_rate": 8.330250552125594e-05, "loss": 0.0058, "step": 1766 }, { "epoch": 8.53623188405797, "grad_norm": 0.03023175336420536, "learning_rate": 8.327733865899799e-05, "loss": 0.0035, "step": 1767 }, { "epoch": 8.541062801932368, "grad_norm": 0.05886807292699814, "learning_rate": 8.325215665290604e-05, "loss": 0.0051, "step": 1768 }, { "epoch": 8.545893719806763, "grad_norm": 0.055516134947538376, "learning_rate": 8.322695951443995e-05, "loss": 0.0043, "step": 1769 }, { "epoch": 8.55072463768116, "grad_norm": 0.05037660151720047, "learning_rate": 8.32017472550664e-05, "loss": 0.0043, "step": 1770 }, { "epoch": 8.555555555555555, "grad_norm": 0.0338997058570385, "learning_rate": 8.317651988625898e-05, "loss": 0.0026, "step": 1771 }, { "epoch": 8.560386473429952, "grad_norm": 0.04957183822989464, "learning_rate": 8.315127741949816e-05, "loss": 0.0052, "step": 1772 }, { "epoch": 8.565217391304348, "grad_norm": 0.03905091434717178, "learning_rate": 8.312601986627124e-05, "loss": 0.0034, "step": 1773 }, { "epoch": 8.570048309178745, "grad_norm": 0.04605364799499512, "learning_rate": 8.310074723807242e-05, "loss": 0.005, "step": 1774 }, { "epoch": 8.57487922705314, "grad_norm": 0.04077398031949997, "learning_rate": 8.307545954640278e-05, "loss": 0.0047, "step": 1775 }, { "epoch": 8.57487922705314, "eval_loss": 0.015577097423374653, "eval_runtime": 20.6393, "eval_samples_per_second": 4.845, "eval_steps_per_second": 0.145, "step": 1775 }, { "epoch": 8.579710144927537, "grad_norm": 0.04953567311167717, "learning_rate": 8.30501568027702e-05, "loss": 0.0045, "step": 1776 }, { "epoch": 8.584541062801932, "grad_norm": 0.0420534722507, "learning_rate": 8.302483901868947e-05, "loss": 0.0038, "step": 1777 }, { "epoch": 8.58937198067633, "grad_norm": 0.03566531836986542, "learning_rate": 8.299950620568216e-05, "loss": 0.0044, "step": 1778 }, { "epoch": 8.594202898550725, "grad_norm": 0.03538971394300461, "learning_rate": 8.297415837527673e-05, "loss": 0.0036, "step": 1779 }, { "epoch": 8.59903381642512, "grad_norm": 0.05101785808801651, "learning_rate": 8.294879553900847e-05, "loss": 0.0045, "step": 1780 }, { "epoch": 8.603864734299517, "grad_norm": 0.03426271304488182, "learning_rate": 8.292341770841945e-05, "loss": 0.003, "step": 1781 }, { "epoch": 8.608695652173914, "grad_norm": 0.06078261137008667, "learning_rate": 8.289802489505865e-05, "loss": 0.0058, "step": 1782 }, { "epoch": 8.61352657004831, "grad_norm": 0.04777718707919121, "learning_rate": 8.287261711048178e-05, "loss": 0.0047, "step": 1783 }, { "epoch": 8.618357487922705, "grad_norm": 0.036562662571668625, "learning_rate": 8.284719436625141e-05, "loss": 0.0032, "step": 1784 }, { "epoch": 8.623188405797102, "grad_norm": 0.037714749574661255, "learning_rate": 8.282175667393691e-05, "loss": 0.0044, "step": 1785 }, { "epoch": 8.628019323671497, "grad_norm": 0.03402454033493996, "learning_rate": 8.279630404511447e-05, "loss": 0.003, "step": 1786 }, { "epoch": 8.632850241545894, "grad_norm": 0.03453662991523743, "learning_rate": 8.277083649136703e-05, "loss": 0.0039, "step": 1787 }, { "epoch": 8.63768115942029, "grad_norm": 0.03423767164349556, "learning_rate": 8.274535402428436e-05, "loss": 0.0041, "step": 1788 }, { "epoch": 8.642512077294686, "grad_norm": 0.0426657572388649, "learning_rate": 8.271985665546303e-05, "loss": 0.004, "step": 1789 }, { "epoch": 8.647342995169081, "grad_norm": 0.05961960554122925, "learning_rate": 8.269434439650634e-05, "loss": 0.0051, "step": 1790 }, { "epoch": 8.652173913043478, "grad_norm": 0.055078715085983276, "learning_rate": 8.266881725902439e-05, "loss": 0.0054, "step": 1791 }, { "epoch": 8.657004830917874, "grad_norm": 0.03182747960090637, "learning_rate": 8.264327525463408e-05, "loss": 0.0038, "step": 1792 }, { "epoch": 8.66183574879227, "grad_norm": 0.028650034219026566, "learning_rate": 8.261771839495907e-05, "loss": 0.0037, "step": 1793 }, { "epoch": 8.666666666666666, "grad_norm": 0.028739208355545998, "learning_rate": 8.259214669162972e-05, "loss": 0.003, "step": 1794 }, { "epoch": 8.671497584541063, "grad_norm": 0.04689854383468628, "learning_rate": 8.256656015628321e-05, "loss": 0.0036, "step": 1795 }, { "epoch": 8.676328502415458, "grad_norm": 0.04869701340794563, "learning_rate": 8.254095880056345e-05, "loss": 0.0037, "step": 1796 }, { "epoch": 8.681159420289855, "grad_norm": 0.04003773257136345, "learning_rate": 8.251534263612106e-05, "loss": 0.0031, "step": 1797 }, { "epoch": 8.68599033816425, "grad_norm": 0.0333455391228199, "learning_rate": 8.248971167461348e-05, "loss": 0.0037, "step": 1798 }, { "epoch": 8.690821256038648, "grad_norm": 0.04959764704108238, "learning_rate": 8.24640659277048e-05, "loss": 0.0035, "step": 1799 }, { "epoch": 8.695652173913043, "grad_norm": 0.046579279005527496, "learning_rate": 8.243840540706588e-05, "loss": 0.0044, "step": 1800 }, { "epoch": 8.695652173913043, "eval_loss": 0.0165646243840456, "eval_runtime": 20.6036, "eval_samples_per_second": 4.854, "eval_steps_per_second": 0.146, "step": 1800 }, { "epoch": 8.70048309178744, "grad_norm": 0.038630325347185135, "learning_rate": 8.24127301243743e-05, "loss": 0.0033, "step": 1801 }, { "epoch": 8.705314009661835, "grad_norm": 0.04081215336918831, "learning_rate": 8.238704009131437e-05, "loss": 0.0033, "step": 1802 }, { "epoch": 8.710144927536232, "grad_norm": 0.05585318058729172, "learning_rate": 8.236133531957706e-05, "loss": 0.0042, "step": 1803 }, { "epoch": 8.714975845410628, "grad_norm": 0.05531182512640953, "learning_rate": 8.233561582086012e-05, "loss": 0.0051, "step": 1804 }, { "epoch": 8.719806763285025, "grad_norm": 0.04835759848356247, "learning_rate": 8.230988160686793e-05, "loss": 0.0032, "step": 1805 }, { "epoch": 8.72463768115942, "grad_norm": 0.07098240405321121, "learning_rate": 8.228413268931162e-05, "loss": 0.0028, "step": 1806 }, { "epoch": 8.729468599033817, "grad_norm": 0.06070776656270027, "learning_rate": 8.225836907990902e-05, "loss": 0.0055, "step": 1807 }, { "epoch": 8.734299516908212, "grad_norm": 0.04384101182222366, "learning_rate": 8.223259079038459e-05, "loss": 0.0038, "step": 1808 }, { "epoch": 8.73913043478261, "grad_norm": 0.06258191913366318, "learning_rate": 8.22067978324695e-05, "loss": 0.0043, "step": 1809 }, { "epoch": 8.743961352657005, "grad_norm": 0.07775823026895523, "learning_rate": 8.218099021790162e-05, "loss": 0.0038, "step": 1810 }, { "epoch": 8.748792270531402, "grad_norm": 0.05379994586110115, "learning_rate": 8.215516795842542e-05, "loss": 0.004, "step": 1811 }, { "epoch": 8.753623188405797, "grad_norm": 0.04109995439648628, "learning_rate": 8.212933106579213e-05, "loss": 0.004, "step": 1812 }, { "epoch": 8.758454106280194, "grad_norm": 0.06280244141817093, "learning_rate": 8.210347955175956e-05, "loss": 0.0042, "step": 1813 }, { "epoch": 8.76328502415459, "grad_norm": 0.04878249764442444, "learning_rate": 8.207761342809224e-05, "loss": 0.0045, "step": 1814 }, { "epoch": 8.768115942028986, "grad_norm": 0.03682022541761398, "learning_rate": 8.205173270656125e-05, "loss": 0.0041, "step": 1815 }, { "epoch": 8.772946859903382, "grad_norm": 0.058330096304416656, "learning_rate": 8.202583739894443e-05, "loss": 0.0055, "step": 1816 }, { "epoch": 8.777777777777779, "grad_norm": 0.051801662892103195, "learning_rate": 8.199992751702618e-05, "loss": 0.0048, "step": 1817 }, { "epoch": 8.782608695652174, "grad_norm": 0.04357374832034111, "learning_rate": 8.197400307259757e-05, "loss": 0.0048, "step": 1818 }, { "epoch": 8.78743961352657, "grad_norm": 0.04432510957121849, "learning_rate": 8.194806407745625e-05, "loss": 0.0043, "step": 1819 }, { "epoch": 8.792270531400966, "grad_norm": 0.05476441606879234, "learning_rate": 8.192211054340657e-05, "loss": 0.0048, "step": 1820 }, { "epoch": 8.797101449275363, "grad_norm": 0.05567046254873276, "learning_rate": 8.189614248225942e-05, "loss": 0.0055, "step": 1821 }, { "epoch": 8.801932367149758, "grad_norm": 0.05372453108429909, "learning_rate": 8.187015990583234e-05, "loss": 0.0043, "step": 1822 }, { "epoch": 8.806763285024154, "grad_norm": 0.03940294682979584, "learning_rate": 8.184416282594946e-05, "loss": 0.0038, "step": 1823 }, { "epoch": 8.81159420289855, "grad_norm": 0.04525710269808769, "learning_rate": 8.181815125444152e-05, "loss": 0.0047, "step": 1824 }, { "epoch": 8.816425120772946, "grad_norm": 0.03286072239279747, "learning_rate": 8.179212520314586e-05, "loss": 0.0031, "step": 1825 }, { "epoch": 8.816425120772946, "eval_loss": 0.017208324745297432, "eval_runtime": 20.6517, "eval_samples_per_second": 4.842, "eval_steps_per_second": 0.145, "step": 1825 }, { "epoch": 8.821256038647343, "grad_norm": 0.1259443759918213, "learning_rate": 8.17660846839064e-05, "loss": 0.0043, "step": 1826 }, { "epoch": 8.826086956521738, "grad_norm": 0.0417412668466568, "learning_rate": 8.174002970857364e-05, "loss": 0.004, "step": 1827 }, { "epoch": 8.830917874396135, "grad_norm": 0.03266125172376633, "learning_rate": 8.171396028900468e-05, "loss": 0.0045, "step": 1828 }, { "epoch": 8.83574879227053, "grad_norm": 0.05389874801039696, "learning_rate": 8.168787643706315e-05, "loss": 0.0046, "step": 1829 }, { "epoch": 8.840579710144928, "grad_norm": 0.06979716569185257, "learning_rate": 8.16617781646193e-05, "loss": 0.0036, "step": 1830 }, { "epoch": 8.845410628019323, "grad_norm": 0.049693189561367035, "learning_rate": 8.163566548354991e-05, "loss": 0.0049, "step": 1831 }, { "epoch": 8.85024154589372, "grad_norm": 0.05725079029798508, "learning_rate": 8.160953840573834e-05, "loss": 0.0048, "step": 1832 }, { "epoch": 8.855072463768115, "grad_norm": 0.044530972838401794, "learning_rate": 8.158339694307446e-05, "loss": 0.0044, "step": 1833 }, { "epoch": 8.859903381642512, "grad_norm": 0.03541207313537598, "learning_rate": 8.155724110745473e-05, "loss": 0.0051, "step": 1834 }, { "epoch": 8.864734299516908, "grad_norm": 0.058717984706163406, "learning_rate": 8.153107091078212e-05, "loss": 0.0045, "step": 1835 }, { "epoch": 8.869565217391305, "grad_norm": 0.061354346573352814, "learning_rate": 8.150488636496618e-05, "loss": 0.0047, "step": 1836 }, { "epoch": 8.8743961352657, "grad_norm": 0.037550557404756546, "learning_rate": 8.14786874819229e-05, "loss": 0.0052, "step": 1837 }, { "epoch": 8.879227053140097, "grad_norm": 0.045513689517974854, "learning_rate": 8.145247427357494e-05, "loss": 0.0041, "step": 1838 }, { "epoch": 8.884057971014492, "grad_norm": 0.05298766493797302, "learning_rate": 8.142624675185134e-05, "loss": 0.005, "step": 1839 }, { "epoch": 8.88888888888889, "grad_norm": 0.03655264526605606, "learning_rate": 8.140000492868768e-05, "loss": 0.0048, "step": 1840 }, { "epoch": 8.893719806763285, "grad_norm": 0.044772639870643616, "learning_rate": 8.137374881602614e-05, "loss": 0.0035, "step": 1841 }, { "epoch": 8.898550724637682, "grad_norm": 0.04357172176241875, "learning_rate": 8.134747842581529e-05, "loss": 0.0057, "step": 1842 }, { "epoch": 8.903381642512077, "grad_norm": 0.07336288690567017, "learning_rate": 8.132119377001027e-05, "loss": 0.0057, "step": 1843 }, { "epoch": 8.908212560386474, "grad_norm": 0.04212936386466026, "learning_rate": 8.129489486057267e-05, "loss": 0.0036, "step": 1844 }, { "epoch": 8.91304347826087, "grad_norm": 0.04403834789991379, "learning_rate": 8.12685817094706e-05, "loss": 0.0034, "step": 1845 }, { "epoch": 8.917874396135266, "grad_norm": 0.062063757330179214, "learning_rate": 8.124225432867863e-05, "loss": 0.0041, "step": 1846 }, { "epoch": 8.922705314009661, "grad_norm": 0.05694074556231499, "learning_rate": 8.121591273017782e-05, "loss": 0.0044, "step": 1847 }, { "epoch": 8.927536231884059, "grad_norm": 0.06622301787137985, "learning_rate": 8.118955692595567e-05, "loss": 0.0048, "step": 1848 }, { "epoch": 8.932367149758454, "grad_norm": 0.04630272090435028, "learning_rate": 8.116318692800618e-05, "loss": 0.0044, "step": 1849 }, { "epoch": 8.93719806763285, "grad_norm": 0.040268830955028534, "learning_rate": 8.113680274832978e-05, "loss": 0.0029, "step": 1850 }, { "epoch": 8.93719806763285, "eval_loss": 0.016717983409762383, "eval_runtime": 20.5962, "eval_samples_per_second": 4.855, "eval_steps_per_second": 0.146, "step": 1850 }, { "epoch": 8.942028985507246, "grad_norm": 0.038945265114307404, "learning_rate": 8.111040439893338e-05, "loss": 0.0038, "step": 1851 }, { "epoch": 8.946859903381643, "grad_norm": 0.05721553415060043, "learning_rate": 8.108399189183034e-05, "loss": 0.0038, "step": 1852 }, { "epoch": 8.951690821256038, "grad_norm": 0.03646910563111305, "learning_rate": 8.105756523904045e-05, "loss": 0.0029, "step": 1853 }, { "epoch": 8.956521739130435, "grad_norm": 0.04455795884132385, "learning_rate": 8.10311244525899e-05, "loss": 0.0038, "step": 1854 }, { "epoch": 8.96135265700483, "grad_norm": 0.054944783449172974, "learning_rate": 8.100466954451139e-05, "loss": 0.0044, "step": 1855 }, { "epoch": 8.966183574879228, "grad_norm": 0.0795806273818016, "learning_rate": 8.097820052684397e-05, "loss": 0.005, "step": 1856 }, { "epoch": 8.971014492753623, "grad_norm": 0.08476272970438004, "learning_rate": 8.095171741163318e-05, "loss": 0.0043, "step": 1857 }, { "epoch": 8.97584541062802, "grad_norm": 0.06524492055177689, "learning_rate": 8.092522021093091e-05, "loss": 0.0051, "step": 1858 }, { "epoch": 8.980676328502415, "grad_norm": 0.03843354806303978, "learning_rate": 8.089870893679554e-05, "loss": 0.0036, "step": 1859 }, { "epoch": 8.985507246376812, "grad_norm": 0.04924914613366127, "learning_rate": 8.087218360129176e-05, "loss": 0.0047, "step": 1860 }, { "epoch": 8.990338164251208, "grad_norm": 0.06344044208526611, "learning_rate": 8.084564421649073e-05, "loss": 0.0052, "step": 1861 }, { "epoch": 8.995169082125603, "grad_norm": 0.05186523124575615, "learning_rate": 8.081909079446998e-05, "loss": 0.0059, "step": 1862 }, { "epoch": 9.0, "grad_norm": 0.04302655905485153, "learning_rate": 8.079252334731342e-05, "loss": 0.0028, "step": 1863 }, { "epoch": 9.004830917874395, "grad_norm": 0.04055224359035492, "learning_rate": 8.076594188711135e-05, "loss": 0.0032, "step": 1864 }, { "epoch": 9.009661835748792, "grad_norm": 0.036405082792043686, "learning_rate": 8.073934642596044e-05, "loss": 0.0038, "step": 1865 }, { "epoch": 9.014492753623188, "grad_norm": 0.043045781552791595, "learning_rate": 8.071273697596374e-05, "loss": 0.0035, "step": 1866 }, { "epoch": 9.019323671497585, "grad_norm": 0.036786459386348724, "learning_rate": 8.068611354923068e-05, "loss": 0.0036, "step": 1867 }, { "epoch": 9.02415458937198, "grad_norm": 0.03571231663227081, "learning_rate": 8.065947615787703e-05, "loss": 0.0035, "step": 1868 }, { "epoch": 9.028985507246377, "grad_norm": 0.043411120772361755, "learning_rate": 8.063282481402491e-05, "loss": 0.0038, "step": 1869 }, { "epoch": 9.033816425120772, "grad_norm": 0.03864371031522751, "learning_rate": 8.060615952980281e-05, "loss": 0.0044, "step": 1870 }, { "epoch": 9.03864734299517, "grad_norm": 0.06767242401838303, "learning_rate": 8.057948031734555e-05, "loss": 0.0049, "step": 1871 }, { "epoch": 9.043478260869565, "grad_norm": 0.03778842091560364, "learning_rate": 8.055278718879429e-05, "loss": 0.0032, "step": 1872 }, { "epoch": 9.048309178743962, "grad_norm": 0.0641743466258049, "learning_rate": 8.052608015629654e-05, "loss": 0.0072, "step": 1873 }, { "epoch": 9.053140096618357, "grad_norm": 0.03386566415429115, "learning_rate": 8.049935923200608e-05, "loss": 0.0031, "step": 1874 }, { "epoch": 9.057971014492754, "grad_norm": 0.03833288326859474, "learning_rate": 8.04726244280831e-05, "loss": 0.0032, "step": 1875 }, { "epoch": 9.057971014492754, "eval_loss": 0.016934776678681374, "eval_runtime": 20.5938, "eval_samples_per_second": 4.856, "eval_steps_per_second": 0.146, "step": 1875 }, { "epoch": 9.06280193236715, "grad_norm": 0.03243975341320038, "learning_rate": 8.044587575669406e-05, "loss": 0.0031, "step": 1876 }, { "epoch": 9.067632850241546, "grad_norm": 0.04720361903309822, "learning_rate": 8.04191132300117e-05, "loss": 0.0043, "step": 1877 }, { "epoch": 9.072463768115941, "grad_norm": 0.04073271527886391, "learning_rate": 8.039233686021512e-05, "loss": 0.0037, "step": 1878 }, { "epoch": 9.077294685990339, "grad_norm": 0.03657795116305351, "learning_rate": 8.036554665948968e-05, "loss": 0.0025, "step": 1879 }, { "epoch": 9.082125603864734, "grad_norm": 0.05789383128285408, "learning_rate": 8.033874264002707e-05, "loss": 0.0037, "step": 1880 }, { "epoch": 9.08695652173913, "grad_norm": 0.08801286667585373, "learning_rate": 8.031192481402523e-05, "loss": 0.0031, "step": 1881 }, { "epoch": 9.091787439613526, "grad_norm": 0.039658673107624054, "learning_rate": 8.028509319368842e-05, "loss": 0.0029, "step": 1882 }, { "epoch": 9.096618357487923, "grad_norm": 0.03780438005924225, "learning_rate": 8.025824779122714e-05, "loss": 0.0032, "step": 1883 }, { "epoch": 9.101449275362318, "grad_norm": 0.029846245422959328, "learning_rate": 8.02313886188582e-05, "loss": 0.0025, "step": 1884 }, { "epoch": 9.106280193236715, "grad_norm": 0.03359315171837807, "learning_rate": 8.020451568880465e-05, "loss": 0.0025, "step": 1885 }, { "epoch": 9.11111111111111, "grad_norm": 0.06273707747459412, "learning_rate": 8.017762901329582e-05, "loss": 0.0037, "step": 1886 }, { "epoch": 9.115942028985508, "grad_norm": 0.05208655446767807, "learning_rate": 8.015072860456727e-05, "loss": 0.004, "step": 1887 }, { "epoch": 9.120772946859903, "grad_norm": 0.05353975296020508, "learning_rate": 8.012381447486081e-05, "loss": 0.0034, "step": 1888 }, { "epoch": 9.1256038647343, "grad_norm": 0.05432965233922005, "learning_rate": 8.009688663642454e-05, "loss": 0.0031, "step": 1889 }, { "epoch": 9.130434782608695, "grad_norm": 0.04310327395796776, "learning_rate": 8.006994510151275e-05, "loss": 0.0028, "step": 1890 }, { "epoch": 9.135265700483092, "grad_norm": 0.044074833393096924, "learning_rate": 8.004298988238598e-05, "loss": 0.0031, "step": 1891 }, { "epoch": 9.140096618357488, "grad_norm": 0.044513918459415436, "learning_rate": 8.001602099131103e-05, "loss": 0.0032, "step": 1892 }, { "epoch": 9.144927536231885, "grad_norm": 0.051814496517181396, "learning_rate": 7.998903844056083e-05, "loss": 0.0053, "step": 1893 }, { "epoch": 9.14975845410628, "grad_norm": 0.02605430781841278, "learning_rate": 7.996204224241462e-05, "loss": 0.0028, "step": 1894 }, { "epoch": 9.154589371980677, "grad_norm": 0.033382847905159, "learning_rate": 7.993503240915781e-05, "loss": 0.0029, "step": 1895 }, { "epoch": 9.159420289855072, "grad_norm": 0.05876319855451584, "learning_rate": 7.990800895308202e-05, "loss": 0.0045, "step": 1896 }, { "epoch": 9.16425120772947, "grad_norm": 0.061334945261478424, "learning_rate": 7.988097188648507e-05, "loss": 0.0028, "step": 1897 }, { "epoch": 9.169082125603865, "grad_norm": 0.06805699318647385, "learning_rate": 7.985392122167098e-05, "loss": 0.0044, "step": 1898 }, { "epoch": 9.173913043478262, "grad_norm": 0.02976112812757492, "learning_rate": 7.982685697094995e-05, "loss": 0.003, "step": 1899 }, { "epoch": 9.178743961352657, "grad_norm": 0.0250865388661623, "learning_rate": 7.979977914663836e-05, "loss": 0.0036, "step": 1900 }, { "epoch": 9.178743961352657, "eval_loss": 0.01666688360273838, "eval_runtime": 20.6487, "eval_samples_per_second": 4.843, "eval_steps_per_second": 0.145, "step": 1900 }, { "epoch": 9.178743961352657, "step": 1900, "total_flos": 7.164017972028703e+18, "train_loss": 0.021690809104444556, "train_runtime": 50941.7483, "train_samples_per_second": 4.858, "train_steps_per_second": 0.102 } ], "logging_steps": 1, "max_steps": 5175, "num_input_tokens_seen": 0, "num_train_epochs": 25, "save_steps": 100, "stateful_callbacks": { "EarlyStoppingCallback": { "args": { "early_stopping_patience": 5, "early_stopping_threshold": 0.0 }, "attributes": { "early_stopping_patience_counter": 5 } }, "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 7.164017972028703e+18, "train_batch_size": 48, "trial_name": null, "trial_params": null }