{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 110, "global_step": 438, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.00228310502283105, "grad_norm": 49763.3828125, "learning_rate": 1e-05, "loss": 1.458, "step": 1 }, { "epoch": 0.00228310502283105, "eval_loss": 1.3469293117523193, "eval_runtime": 5.899, "eval_samples_per_second": 16.952, "eval_steps_per_second": 8.476, "step": 1 }, { "epoch": 0.0045662100456621, "grad_norm": 47645.4765625, "learning_rate": 2e-05, "loss": 1.663, "step": 2 }, { "epoch": 0.00684931506849315, "grad_norm": 53930.67578125, "learning_rate": 3e-05, "loss": 1.2375, "step": 3 }, { "epoch": 0.0091324200913242, "grad_norm": 79495.359375, "learning_rate": 4e-05, "loss": 1.4099, "step": 4 }, { "epoch": 0.01141552511415525, "grad_norm": 51649.53125, "learning_rate": 5e-05, "loss": 1.3848, "step": 5 }, { "epoch": 0.0136986301369863, "grad_norm": 62501.76171875, "learning_rate": 6e-05, "loss": 1.3113, "step": 6 }, { "epoch": 0.01598173515981735, "grad_norm": 56830.796875, "learning_rate": 7e-05, "loss": 1.394, "step": 7 }, { "epoch": 0.0182648401826484, "grad_norm": 69029.3671875, "learning_rate": 8e-05, "loss": 1.1999, "step": 8 }, { "epoch": 0.02054794520547945, "grad_norm": 77872.7421875, "learning_rate": 9e-05, "loss": 1.4121, "step": 9 }, { "epoch": 0.0228310502283105, "grad_norm": 67018.5234375, "learning_rate": 0.0001, "loss": 1.5292, "step": 10 }, { "epoch": 0.02511415525114155, "grad_norm": 52883.80078125, "learning_rate": 0.00011000000000000002, "loss": 1.1904, "step": 11 }, { "epoch": 0.0273972602739726, "grad_norm": 89700.34375, "learning_rate": 0.00012, "loss": 1.6428, "step": 12 }, { "epoch": 0.02968036529680365, "grad_norm": 70959.6875, "learning_rate": 0.00013000000000000002, "loss": 1.8412, "step": 13 }, { "epoch": 0.0319634703196347, "grad_norm": 123958.625, "learning_rate": 0.00014, "loss": 1.3549, "step": 14 }, { "epoch": 0.03424657534246575, "grad_norm": 55201.86328125, "learning_rate": 0.00015000000000000001, "loss": 0.717, "step": 15 }, { "epoch": 0.0365296803652968, "grad_norm": 79495.921875, "learning_rate": 0.00016, "loss": 2.5554, "step": 16 }, { "epoch": 0.03881278538812785, "grad_norm": 94823.359375, "learning_rate": 0.00017, "loss": 1.3731, "step": 17 }, { "epoch": 0.0410958904109589, "grad_norm": 78739.7890625, "learning_rate": 0.00018, "loss": 0.9813, "step": 18 }, { "epoch": 0.04337899543378995, "grad_norm": 70018.5234375, "learning_rate": 0.00019, "loss": 1.3944, "step": 19 }, { "epoch": 0.045662100456621, "grad_norm": 93789.5625, "learning_rate": 0.0002, "loss": 1.3604, "step": 20 }, { "epoch": 0.04794520547945205, "grad_norm": 69996.328125, "learning_rate": 0.0001999971756719333, "loss": 0.7975, "step": 21 }, { "epoch": 0.0502283105022831, "grad_norm": 62440.68359375, "learning_rate": 0.00019998870284726968, "loss": 1.3632, "step": 22 }, { "epoch": 0.05251141552511415, "grad_norm": 57081.87890625, "learning_rate": 0.00019997458200460993, "loss": 1.181, "step": 23 }, { "epoch": 0.0547945205479452, "grad_norm": 68026.21875, "learning_rate": 0.00019995481394159188, "loss": 1.2883, "step": 24 }, { "epoch": 0.05707762557077625, "grad_norm": 68773.109375, "learning_rate": 0.0001999293997748454, "loss": 1.209, "step": 25 }, { "epoch": 0.0593607305936073, "grad_norm": 54953.078125, "learning_rate": 0.00019989834093992945, "loss": 1.2207, "step": 26 }, { "epoch": 0.06164383561643835, "grad_norm": 68915.734375, "learning_rate": 0.00019986163919125075, "loss": 1.1395, "step": 27 }, { "epoch": 0.0639269406392694, "grad_norm": 77723.328125, "learning_rate": 0.00019981929660196492, "loss": 1.4178, "step": 28 }, { "epoch": 0.06621004566210045, "grad_norm": 57558.64453125, "learning_rate": 0.0001997713155638592, "loss": 1.3991, "step": 29 }, { "epoch": 0.0684931506849315, "grad_norm": 57487.20703125, "learning_rate": 0.00019971769878721743, "loss": 1.2361, "step": 30 }, { "epoch": 0.07077625570776255, "grad_norm": 56958.12109375, "learning_rate": 0.000199658449300667, "loss": 1.0886, "step": 31 }, { "epoch": 0.0730593607305936, "grad_norm": 69826.890625, "learning_rate": 0.00019959357045100764, "loss": 1.2462, "step": 32 }, { "epoch": 0.07534246575342465, "grad_norm": 83606.59375, "learning_rate": 0.00019952306590302247, "loss": 1.5512, "step": 33 }, { "epoch": 0.0776255707762557, "grad_norm": 96047.1171875, "learning_rate": 0.00019944693963927092, "loss": 1.2642, "step": 34 }, { "epoch": 0.07990867579908675, "grad_norm": 75340.171875, "learning_rate": 0.00019936519595986394, "loss": 1.3945, "step": 35 }, { "epoch": 0.0821917808219178, "grad_norm": 53603.67578125, "learning_rate": 0.00019927783948222084, "loss": 0.8141, "step": 36 }, { "epoch": 0.08447488584474885, "grad_norm": 60950.3359375, "learning_rate": 0.00019918487514080865, "loss": 1.1856, "step": 37 }, { "epoch": 0.0867579908675799, "grad_norm": 57694.49609375, "learning_rate": 0.00019908630818686338, "loss": 0.5208, "step": 38 }, { "epoch": 0.08904109589041095, "grad_norm": 83846.890625, "learning_rate": 0.0001989821441880933, "loss": 1.2401, "step": 39 }, { "epoch": 0.091324200913242, "grad_norm": 60294.91796875, "learning_rate": 0.00019887238902836448, "loss": 1.3362, "step": 40 }, { "epoch": 0.09360730593607305, "grad_norm": 55908.76953125, "learning_rate": 0.00019875704890736853, "loss": 1.0295, "step": 41 }, { "epoch": 0.0958904109589041, "grad_norm": 79842.359375, "learning_rate": 0.00019863613034027224, "loss": 1.3764, "step": 42 }, { "epoch": 0.09817351598173515, "grad_norm": 53915.5703125, "learning_rate": 0.0001985096401573497, "loss": 1.1399, "step": 43 }, { "epoch": 0.1004566210045662, "grad_norm": 56798.44921875, "learning_rate": 0.00019837758550359636, "loss": 0.9945, "step": 44 }, { "epoch": 0.10273972602739725, "grad_norm": 72487.9921875, "learning_rate": 0.0001982399738383255, "loss": 1.4328, "step": 45 }, { "epoch": 0.1050228310502283, "grad_norm": 56317.5546875, "learning_rate": 0.00019809681293474693, "loss": 1.2217, "step": 46 }, { "epoch": 0.10730593607305935, "grad_norm": 108634.3203125, "learning_rate": 0.0001979481108795278, "loss": 1.5159, "step": 47 }, { "epoch": 0.1095890410958904, "grad_norm": 77436.7578125, "learning_rate": 0.00019779387607233586, "loss": 1.051, "step": 48 }, { "epoch": 0.11187214611872145, "grad_norm": 61779.69921875, "learning_rate": 0.00019763411722536502, "loss": 1.2487, "step": 49 }, { "epoch": 0.1141552511415525, "grad_norm": 74456.3359375, "learning_rate": 0.00019746884336284317, "loss": 1.5362, "step": 50 }, { "epoch": 0.11643835616438356, "grad_norm": 65273.90625, "learning_rate": 0.00019729806382052248, "loss": 1.125, "step": 51 }, { "epoch": 0.1187214611872146, "grad_norm": 74326.5390625, "learning_rate": 0.00019712178824515212, "loss": 1.4872, "step": 52 }, { "epoch": 0.12100456621004566, "grad_norm": 66311.375, "learning_rate": 0.00019694002659393305, "loss": 1.3129, "step": 53 }, { "epoch": 0.1232876712328767, "grad_norm": 93956.140625, "learning_rate": 0.00019675278913395606, "loss": 1.1963, "step": 54 }, { "epoch": 0.12557077625570776, "grad_norm": 54108.7421875, "learning_rate": 0.0001965600864416213, "loss": 1.3549, "step": 55 }, { "epoch": 0.1278538812785388, "grad_norm": 82672.5234375, "learning_rate": 0.00019636192940204134, "loss": 1.3415, "step": 56 }, { "epoch": 0.13013698630136986, "grad_norm": 79705.0234375, "learning_rate": 0.00019615832920842586, "loss": 1.0711, "step": 57 }, { "epoch": 0.1324200913242009, "grad_norm": 66569.171875, "learning_rate": 0.00019594929736144976, "loss": 1.5752, "step": 58 }, { "epoch": 0.13470319634703196, "grad_norm": 54818.6953125, "learning_rate": 0.0001957348456686032, "loss": 1.2086, "step": 59 }, { "epoch": 0.136986301369863, "grad_norm": 84023.5625, "learning_rate": 0.00019551498624352496, "loss": 1.2713, "step": 60 }, { "epoch": 0.13926940639269406, "grad_norm": 58488.77734375, "learning_rate": 0.00019528973150531787, "loss": 1.1957, "step": 61 }, { "epoch": 0.1415525114155251, "grad_norm": 59256.1328125, "learning_rate": 0.00019505909417784754, "loss": 1.1863, "step": 62 }, { "epoch": 0.14383561643835616, "grad_norm": 58009.8359375, "learning_rate": 0.00019482308728902356, "loss": 1.0046, "step": 63 }, { "epoch": 0.1461187214611872, "grad_norm": 58970.30859375, "learning_rate": 0.00019458172417006347, "loss": 1.3599, "step": 64 }, { "epoch": 0.14840182648401826, "grad_norm": 75271.78125, "learning_rate": 0.00019433501845473995, "loss": 1.175, "step": 65 }, { "epoch": 0.1506849315068493, "grad_norm": 60125.44140625, "learning_rate": 0.00019408298407861042, "loss": 1.2096, "step": 66 }, { "epoch": 0.15296803652968036, "grad_norm": 62565.88671875, "learning_rate": 0.00019382563527823026, "loss": 1.0284, "step": 67 }, { "epoch": 0.1552511415525114, "grad_norm": 64562.3359375, "learning_rate": 0.00019356298659034817, "loss": 1.1955, "step": 68 }, { "epoch": 0.15753424657534246, "grad_norm": 61627.109375, "learning_rate": 0.00019329505285108542, "loss": 1.1498, "step": 69 }, { "epoch": 0.1598173515981735, "grad_norm": 65598.3515625, "learning_rate": 0.00019302184919509755, "loss": 1.046, "step": 70 }, { "epoch": 0.16210045662100456, "grad_norm": 56694.41015625, "learning_rate": 0.00019274339105471971, "loss": 0.6779, "step": 71 }, { "epoch": 0.1643835616438356, "grad_norm": 82800.1015625, "learning_rate": 0.00019245969415909465, "loss": 1.2381, "step": 72 }, { "epoch": 0.16666666666666666, "grad_norm": 44540.94140625, "learning_rate": 0.00019217077453328449, "loss": 1.0871, "step": 73 }, { "epoch": 0.1689497716894977, "grad_norm": 62427.3125, "learning_rate": 0.0001918766484973654, "loss": 1.4182, "step": 74 }, { "epoch": 0.17123287671232876, "grad_norm": 62395.83203125, "learning_rate": 0.00019157733266550575, "loss": 1.1391, "step": 75 }, { "epoch": 0.1735159817351598, "grad_norm": 86479.0546875, "learning_rate": 0.0001912728439450276, "loss": 1.4769, "step": 76 }, { "epoch": 0.17579908675799086, "grad_norm": 52077.796875, "learning_rate": 0.00019096319953545185, "loss": 0.9951, "step": 77 }, { "epoch": 0.1780821917808219, "grad_norm": 65181.8671875, "learning_rate": 0.0001906484169275263, "loss": 1.1435, "step": 78 }, { "epoch": 0.18036529680365296, "grad_norm": 61927.88671875, "learning_rate": 0.00019032851390223812, "loss": 1.3638, "step": 79 }, { "epoch": 0.182648401826484, "grad_norm": 65557.1796875, "learning_rate": 0.00019000350852980909, "loss": 1.0751, "step": 80 }, { "epoch": 0.18493150684931506, "grad_norm": 67026.625, "learning_rate": 0.00018967341916867518, "loss": 1.4392, "step": 81 }, { "epoch": 0.1872146118721461, "grad_norm": 65608.7578125, "learning_rate": 0.00018933826446444933, "loss": 1.4204, "step": 82 }, { "epoch": 0.18949771689497716, "grad_norm": 62439.32421875, "learning_rate": 0.0001889980633488683, "loss": 1.6256, "step": 83 }, { "epoch": 0.1917808219178082, "grad_norm": 58315.05078125, "learning_rate": 0.00018865283503872324, "loss": 1.3327, "step": 84 }, { "epoch": 0.19406392694063926, "grad_norm": 63276.6953125, "learning_rate": 0.00018830259903477426, "loss": 1.3005, "step": 85 }, { "epoch": 0.1963470319634703, "grad_norm": 60068.3125, "learning_rate": 0.0001879473751206489, "loss": 1.3073, "step": 86 }, { "epoch": 0.19863013698630136, "grad_norm": 60654.34375, "learning_rate": 0.0001875871833617246, "loss": 1.1668, "step": 87 }, { "epoch": 0.2009132420091324, "grad_norm": 49492.6875, "learning_rate": 0.0001872220441039952, "loss": 0.8938, "step": 88 }, { "epoch": 0.20319634703196346, "grad_norm": 82807.6953125, "learning_rate": 0.0001868519779729218, "loss": 1.014, "step": 89 }, { "epoch": 0.2054794520547945, "grad_norm": 50824.89453125, "learning_rate": 0.0001864770058722676, "loss": 1.0941, "step": 90 }, { "epoch": 0.20776255707762556, "grad_norm": 53105.984375, "learning_rate": 0.00018609714898291718, "loss": 0.7681, "step": 91 }, { "epoch": 0.2100456621004566, "grad_norm": 57766.73046875, "learning_rate": 0.00018571242876167996, "loss": 1.0531, "step": 92 }, { "epoch": 0.21232876712328766, "grad_norm": 66334.5625, "learning_rate": 0.0001853228669400784, "loss": 1.2699, "step": 93 }, { "epoch": 0.2146118721461187, "grad_norm": 54520.6015625, "learning_rate": 0.00018492848552312014, "loss": 1.4723, "step": 94 }, { "epoch": 0.21689497716894976, "grad_norm": 75962.671875, "learning_rate": 0.00018452930678805536, "loss": 1.379, "step": 95 }, { "epoch": 0.2191780821917808, "grad_norm": 57191.44921875, "learning_rate": 0.00018412535328311814, "loss": 1.3189, "step": 96 }, { "epoch": 0.22146118721461186, "grad_norm": 63262.0625, "learning_rate": 0.00018371664782625287, "loss": 0.9332, "step": 97 }, { "epoch": 0.2237442922374429, "grad_norm": 55938.12890625, "learning_rate": 0.00018330321350382544, "loss": 1.3675, "step": 98 }, { "epoch": 0.22602739726027396, "grad_norm": 48929.921875, "learning_rate": 0.00018288507366931905, "loss": 1.0751, "step": 99 }, { "epoch": 0.228310502283105, "grad_norm": 60707.87890625, "learning_rate": 0.00018246225194201517, "loss": 1.2708, "step": 100 }, { "epoch": 0.23059360730593606, "grad_norm": 59498.01171875, "learning_rate": 0.00018203477220565912, "loss": 1.3661, "step": 101 }, { "epoch": 0.2328767123287671, "grad_norm": 52212.64453125, "learning_rate": 0.00018160265860711134, "loss": 0.9827, "step": 102 }, { "epoch": 0.23515981735159816, "grad_norm": 47361.78515625, "learning_rate": 0.00018116593555498307, "loss": 1.0684, "step": 103 }, { "epoch": 0.2374429223744292, "grad_norm": 63993.5078125, "learning_rate": 0.0001807246277182578, "loss": 1.2669, "step": 104 }, { "epoch": 0.23972602739726026, "grad_norm": 55152.2578125, "learning_rate": 0.0001802787600248977, "loss": 0.9088, "step": 105 }, { "epoch": 0.2420091324200913, "grad_norm": 49352.40625, "learning_rate": 0.0001798283576604356, "loss": 1.0416, "step": 106 }, { "epoch": 0.24429223744292236, "grad_norm": 60283.234375, "learning_rate": 0.0001793734460665523, "loss": 1.3194, "step": 107 }, { "epoch": 0.2465753424657534, "grad_norm": 46655.90234375, "learning_rate": 0.00017891405093963938, "loss": 0.8581, "step": 108 }, { "epoch": 0.24885844748858446, "grad_norm": 65966.4609375, "learning_rate": 0.0001784501982293479, "loss": 0.9858, "step": 109 }, { "epoch": 0.2511415525114155, "grad_norm": 72021.5703125, "learning_rate": 0.00017798191413712243, "loss": 1.0411, "step": 110 }, { "epoch": 0.2511415525114155, "eval_loss": 1.1379607915878296, "eval_runtime": 6.0407, "eval_samples_per_second": 16.554, "eval_steps_per_second": 8.277, "step": 110 }, { "epoch": 0.2534246575342466, "grad_norm": 60144.625, "learning_rate": 0.0001775092251147211, "loss": 1.1175, "step": 111 }, { "epoch": 0.2557077625570776, "grad_norm": 55304.96484375, "learning_rate": 0.0001770321578627213, "loss": 0.7432, "step": 112 }, { "epoch": 0.2579908675799087, "grad_norm": 50044.91015625, "learning_rate": 0.00017655073932901168, "loss": 0.8335, "step": 113 }, { "epoch": 0.2602739726027397, "grad_norm": 63828.10546875, "learning_rate": 0.0001760649967072697, "loss": 1.1305, "step": 114 }, { "epoch": 0.2625570776255708, "grad_norm": 56665.87890625, "learning_rate": 0.00017557495743542585, "loss": 1.1719, "step": 115 }, { "epoch": 0.2648401826484018, "grad_norm": 63755.87109375, "learning_rate": 0.00017508064919411344, "loss": 1.0547, "step": 116 }, { "epoch": 0.2671232876712329, "grad_norm": 55144.74609375, "learning_rate": 0.00017458209990510527, "loss": 0.8684, "step": 117 }, { "epoch": 0.2694063926940639, "grad_norm": 68380.3125, "learning_rate": 0.00017407933772973637, "loss": 1.1672, "step": 118 }, { "epoch": 0.271689497716895, "grad_norm": 67942.5, "learning_rate": 0.00017357239106731317, "loss": 1.3715, "step": 119 }, { "epoch": 0.273972602739726, "grad_norm": 50505.95703125, "learning_rate": 0.00017306128855350942, "loss": 0.9512, "step": 120 }, { "epoch": 0.2762557077625571, "grad_norm": 56973.859375, "learning_rate": 0.0001725460590587486, "loss": 0.9059, "step": 121 }, { "epoch": 0.2785388127853881, "grad_norm": 47352.71484375, "learning_rate": 0.00017202673168657318, "loss": 0.9492, "step": 122 }, { "epoch": 0.2808219178082192, "grad_norm": 60938.4921875, "learning_rate": 0.0001715033357720006, "loss": 1.1549, "step": 123 }, { "epoch": 0.2831050228310502, "grad_norm": 50557.2265625, "learning_rate": 0.00017097590087986633, "loss": 1.1143, "step": 124 }, { "epoch": 0.2853881278538813, "grad_norm": 61211.25390625, "learning_rate": 0.00017044445680315372, "loss": 1.1829, "step": 125 }, { "epoch": 0.2876712328767123, "grad_norm": 58093.75, "learning_rate": 0.00016990903356131124, "loss": 1.1329, "step": 126 }, { "epoch": 0.2899543378995434, "grad_norm": 62758.921875, "learning_rate": 0.00016936966139855663, "loss": 0.8082, "step": 127 }, { "epoch": 0.2922374429223744, "grad_norm": 55444.03515625, "learning_rate": 0.00016882637078216868, "loss": 1.0763, "step": 128 }, { "epoch": 0.2945205479452055, "grad_norm": 45902.125, "learning_rate": 0.0001682791924007661, "loss": 0.5558, "step": 129 }, { "epoch": 0.2968036529680365, "grad_norm": 67659.0546875, "learning_rate": 0.00016772815716257412, "loss": 0.8364, "step": 130 }, { "epoch": 0.2990867579908676, "grad_norm": 58708.359375, "learning_rate": 0.0001671732961936785, "loss": 1.1031, "step": 131 }, { "epoch": 0.3013698630136986, "grad_norm": 44826.8671875, "learning_rate": 0.00016661464083626734, "loss": 0.6981, "step": 132 }, { "epoch": 0.3036529680365297, "grad_norm": 49845.3125, "learning_rate": 0.00016605222264686086, "loss": 0.932, "step": 133 }, { "epoch": 0.3059360730593607, "grad_norm": 68260.5703125, "learning_rate": 0.00016548607339452853, "loss": 1.1092, "step": 134 }, { "epoch": 0.3082191780821918, "grad_norm": 51835.33203125, "learning_rate": 0.00016491622505909482, "loss": 1.175, "step": 135 }, { "epoch": 0.3105022831050228, "grad_norm": 47168.08203125, "learning_rate": 0.00016434270982933273, "loss": 1.1088, "step": 136 }, { "epoch": 0.3127853881278539, "grad_norm": 65501.48828125, "learning_rate": 0.0001637655601011454, "loss": 1.0026, "step": 137 }, { "epoch": 0.3150684931506849, "grad_norm": 58316.41015625, "learning_rate": 0.00016318480847573642, "loss": 1.073, "step": 138 }, { "epoch": 0.317351598173516, "grad_norm": 47404.8125, "learning_rate": 0.00016260048775776804, "loss": 0.9182, "step": 139 }, { "epoch": 0.319634703196347, "grad_norm": 67868.96875, "learning_rate": 0.00016201263095350833, "loss": 1.2406, "step": 140 }, { "epoch": 0.3219178082191781, "grad_norm": 62392.62109375, "learning_rate": 0.0001614212712689668, "loss": 0.6007, "step": 141 }, { "epoch": 0.3242009132420091, "grad_norm": 44980.81640625, "learning_rate": 0.00016082644210801844, "loss": 0.9601, "step": 142 }, { "epoch": 0.3264840182648402, "grad_norm": 53141.99609375, "learning_rate": 0.00016022817707051724, "loss": 0.7017, "step": 143 }, { "epoch": 0.3287671232876712, "grad_norm": 54526.64453125, "learning_rate": 0.00015962650995039783, "loss": 1.1726, "step": 144 }, { "epoch": 0.3310502283105023, "grad_norm": 60199.36328125, "learning_rate": 0.00015902147473376694, "loss": 1.0703, "step": 145 }, { "epoch": 0.3333333333333333, "grad_norm": 62003.74609375, "learning_rate": 0.00015841310559698343, "loss": 1.384, "step": 146 }, { "epoch": 0.3356164383561644, "grad_norm": 74924.28125, "learning_rate": 0.0001578014369047279, "loss": 1.4119, "step": 147 }, { "epoch": 0.3378995433789954, "grad_norm": 53506.546875, "learning_rate": 0.00015718650320806142, "loss": 1.1047, "step": 148 }, { "epoch": 0.3401826484018265, "grad_norm": 95636.2578125, "learning_rate": 0.00015656833924247398, "loss": 1.2457, "step": 149 }, { "epoch": 0.3424657534246575, "grad_norm": 84311.3671875, "learning_rate": 0.00015594697992592232, "loss": 1.8571, "step": 150 }, { "epoch": 0.3447488584474886, "grad_norm": 49623.90625, "learning_rate": 0.00015532246035685756, "loss": 0.8779, "step": 151 }, { "epoch": 0.3470319634703196, "grad_norm": 75495.09375, "learning_rate": 0.00015469481581224272, "loss": 1.1597, "step": 152 }, { "epoch": 0.3493150684931507, "grad_norm": 61961.66796875, "learning_rate": 0.00015406408174555976, "loss": 1.1428, "step": 153 }, { "epoch": 0.3515981735159817, "grad_norm": 63387.83203125, "learning_rate": 0.0001534302937848073, "loss": 1.2391, "step": 154 }, { "epoch": 0.3538812785388128, "grad_norm": 59753.67578125, "learning_rate": 0.00015279348773048786, "loss": 1.6021, "step": 155 }, { "epoch": 0.3561643835616438, "grad_norm": 61863.0078125, "learning_rate": 0.00015215369955358566, "loss": 1.241, "step": 156 }, { "epoch": 0.3584474885844749, "grad_norm": 57081.796875, "learning_rate": 0.0001515109653935348, "loss": 1.1971, "step": 157 }, { "epoch": 0.3607305936073059, "grad_norm": 71862.75, "learning_rate": 0.00015086532155617784, "loss": 1.1196, "step": 158 }, { "epoch": 0.363013698630137, "grad_norm": 53156.65234375, "learning_rate": 0.00015021680451171498, "loss": 1.1128, "step": 159 }, { "epoch": 0.365296803652968, "grad_norm": 65439.45703125, "learning_rate": 0.00014956545089264407, "loss": 1.1221, "step": 160 }, { "epoch": 0.3675799086757991, "grad_norm": 61590.7734375, "learning_rate": 0.0001489112974916912, "loss": 1.1785, "step": 161 }, { "epoch": 0.3698630136986301, "grad_norm": 44939.07421875, "learning_rate": 0.00014825438125973264, "loss": 1.1794, "step": 162 }, { "epoch": 0.3721461187214612, "grad_norm": 52118.67578125, "learning_rate": 0.00014759473930370736, "loss": 0.9673, "step": 163 }, { "epoch": 0.3744292237442922, "grad_norm": 56795.2734375, "learning_rate": 0.0001469324088845212, "loss": 1.0183, "step": 164 }, { "epoch": 0.3767123287671233, "grad_norm": 54740.015625, "learning_rate": 0.00014626742741494206, "loss": 1.2797, "step": 165 }, { "epoch": 0.3789954337899543, "grad_norm": 54072.75, "learning_rate": 0.00014559983245748638, "loss": 1.1121, "step": 166 }, { "epoch": 0.3812785388127854, "grad_norm": 63319.68359375, "learning_rate": 0.00014492966172229777, "loss": 1.1224, "step": 167 }, { "epoch": 0.3835616438356164, "grad_norm": 70814.8828125, "learning_rate": 0.00014425695306501658, "loss": 1.5157, "step": 168 }, { "epoch": 0.3858447488584475, "grad_norm": 57420.37890625, "learning_rate": 0.00014358174448464154, "loss": 1.213, "step": 169 }, { "epoch": 0.3881278538812785, "grad_norm": 53243.4375, "learning_rate": 0.00014290407412138366, "loss": 1.3539, "step": 170 }, { "epoch": 0.3904109589041096, "grad_norm": 55964.484375, "learning_rate": 0.00014222398025451135, "loss": 1.0541, "step": 171 }, { "epoch": 0.3926940639269406, "grad_norm": 60372.61328125, "learning_rate": 0.00014154150130018866, "loss": 0.9876, "step": 172 }, { "epoch": 0.3949771689497717, "grad_norm": 78811.0546875, "learning_rate": 0.0001408566758093048, "loss": 1.1479, "step": 173 }, { "epoch": 0.3972602739726027, "grad_norm": 68956.203125, "learning_rate": 0.00014016954246529696, "loss": 1.3174, "step": 174 }, { "epoch": 0.3995433789954338, "grad_norm": 54806.1328125, "learning_rate": 0.00013948014008196487, "loss": 1.1019, "step": 175 }, { "epoch": 0.4018264840182648, "grad_norm": 62154.96875, "learning_rate": 0.0001387885076012785, "loss": 1.2393, "step": 176 }, { "epoch": 0.4041095890410959, "grad_norm": 53926.55078125, "learning_rate": 0.00013809468409117846, "loss": 1.1682, "step": 177 }, { "epoch": 0.4063926940639269, "grad_norm": 55390.63671875, "learning_rate": 0.00013739870874336898, "loss": 1.0232, "step": 178 }, { "epoch": 0.408675799086758, "grad_norm": 60079.51953125, "learning_rate": 0.00013670062087110422, "loss": 1.2492, "step": 179 }, { "epoch": 0.410958904109589, "grad_norm": 68658.15625, "learning_rate": 0.00013600045990696762, "loss": 1.432, "step": 180 }, { "epoch": 0.4132420091324201, "grad_norm": 79412.75, "learning_rate": 0.0001352982654006444, "loss": 1.3679, "step": 181 }, { "epoch": 0.4155251141552511, "grad_norm": 76066.828125, "learning_rate": 0.00013459407701668763, "loss": 1.3032, "step": 182 }, { "epoch": 0.4178082191780822, "grad_norm": 74717.1328125, "learning_rate": 0.00013388793453227767, "loss": 1.1265, "step": 183 }, { "epoch": 0.4200913242009132, "grad_norm": 83329.0859375, "learning_rate": 0.0001331798778349752, "loss": 1.5738, "step": 184 }, { "epoch": 0.4223744292237443, "grad_norm": 56585.3203125, "learning_rate": 0.00013246994692046836, "loss": 1.1222, "step": 185 }, { "epoch": 0.4246575342465753, "grad_norm": 66546.9140625, "learning_rate": 0.00013175818189031327, "loss": 1.1622, "step": 186 }, { "epoch": 0.4269406392694064, "grad_norm": 89381.2734375, "learning_rate": 0.00013104462294966896, "loss": 1.0249, "step": 187 }, { "epoch": 0.4292237442922374, "grad_norm": 61278.7734375, "learning_rate": 0.00013032931040502627, "loss": 0.9255, "step": 188 }, { "epoch": 0.4315068493150685, "grad_norm": 73469.1484375, "learning_rate": 0.00012961228466193116, "loss": 1.2164, "step": 189 }, { "epoch": 0.4337899543378995, "grad_norm": 56180.0546875, "learning_rate": 0.00012889358622270223, "loss": 0.8783, "step": 190 }, { "epoch": 0.4360730593607306, "grad_norm": 55383.29296875, "learning_rate": 0.00012817325568414297, "loss": 1.0513, "step": 191 }, { "epoch": 0.4383561643835616, "grad_norm": 67318.65625, "learning_rate": 0.00012745133373524853, "loss": 1.2075, "step": 192 }, { "epoch": 0.4406392694063927, "grad_norm": 41490.49609375, "learning_rate": 0.0001267278611549073, "loss": 0.4758, "step": 193 }, { "epoch": 0.4429223744292237, "grad_norm": 56853.50390625, "learning_rate": 0.00012600287880959763, "loss": 1.2679, "step": 194 }, { "epoch": 0.4452054794520548, "grad_norm": 34868.66796875, "learning_rate": 0.0001252764276510792, "loss": 0.5454, "step": 195 }, { "epoch": 0.4474885844748858, "grad_norm": 53731.953125, "learning_rate": 0.00012454854871407994, "loss": 1.2682, "step": 196 }, { "epoch": 0.4497716894977169, "grad_norm": 57530.828125, "learning_rate": 0.00012381928311397806, "loss": 1.1872, "step": 197 }, { "epoch": 0.4520547945205479, "grad_norm": 59639.96875, "learning_rate": 0.0001230886720444796, "loss": 1.2913, "step": 198 }, { "epoch": 0.454337899543379, "grad_norm": 76171.09375, "learning_rate": 0.00012235675677529158, "loss": 1.314, "step": 199 }, { "epoch": 0.45662100456621, "grad_norm": 73176.8984375, "learning_rate": 0.00012162357864979072, "loss": 0.9921, "step": 200 }, { "epoch": 0.4589041095890411, "grad_norm": 44535.90625, "learning_rate": 0.00012088917908268821, "loss": 0.9701, "step": 201 }, { "epoch": 0.4611872146118721, "grad_norm": 60352.7734375, "learning_rate": 0.00012015359955769021, "loss": 1.0526, "step": 202 }, { "epoch": 0.4634703196347032, "grad_norm": 60158.63671875, "learning_rate": 0.00011941688162515467, "loss": 0.9069, "step": 203 }, { "epoch": 0.4657534246575342, "grad_norm": 56008.59375, "learning_rate": 0.00011867906689974428, "loss": 1.1581, "step": 204 }, { "epoch": 0.4680365296803653, "grad_norm": 60848.1328125, "learning_rate": 0.00011794019705807584, "loss": 1.4732, "step": 205 }, { "epoch": 0.4703196347031963, "grad_norm": 73643.15625, "learning_rate": 0.00011720031383636585, "loss": 1.4934, "step": 206 }, { "epoch": 0.4726027397260274, "grad_norm": 60684.5625, "learning_rate": 0.00011645945902807341, "loss": 0.9241, "step": 207 }, { "epoch": 0.4748858447488584, "grad_norm": 56336.984375, "learning_rate": 0.00011571767448153901, "loss": 0.7535, "step": 208 }, { "epoch": 0.4771689497716895, "grad_norm": 57490.16015625, "learning_rate": 0.00011497500209762102, "loss": 1.0042, "step": 209 }, { "epoch": 0.4794520547945205, "grad_norm": 58042.80859375, "learning_rate": 0.00011423148382732853, "loss": 1.0596, "step": 210 }, { "epoch": 0.4817351598173516, "grad_norm": 72169.4375, "learning_rate": 0.00011348716166945195, "loss": 1.259, "step": 211 }, { "epoch": 0.4840182648401826, "grad_norm": 60079.5859375, "learning_rate": 0.0001127420776681905, "loss": 1.189, "step": 212 }, { "epoch": 0.4863013698630137, "grad_norm": 67140.265625, "learning_rate": 0.00011199627391077732, "loss": 1.4022, "step": 213 }, { "epoch": 0.4885844748858447, "grad_norm": 52046.58984375, "learning_rate": 0.00011124979252510208, "loss": 1.3088, "step": 214 }, { "epoch": 0.4908675799086758, "grad_norm": 61540.34375, "learning_rate": 0.0001105026756773314, "loss": 1.1, "step": 215 }, { "epoch": 0.4931506849315068, "grad_norm": 54976.57421875, "learning_rate": 0.00010975496556952682, "loss": 0.9573, "step": 216 }, { "epoch": 0.4954337899543379, "grad_norm": 47444.56640625, "learning_rate": 0.00010900670443726135, "loss": 0.7793, "step": 217 }, { "epoch": 0.4977168949771689, "grad_norm": 114338.7734375, "learning_rate": 0.00010825793454723325, "loss": 1.4479, "step": 218 }, { "epoch": 0.5, "grad_norm": 55353.40625, "learning_rate": 0.00010750869819487883, "loss": 1.1219, "step": 219 }, { "epoch": 0.502283105022831, "grad_norm": 56450.48046875, "learning_rate": 0.00010675903770198333, "loss": 1.143, "step": 220 }, { "epoch": 0.502283105022831, "eval_loss": 1.120017647743225, "eval_runtime": 6.072, "eval_samples_per_second": 16.469, "eval_steps_per_second": 8.235, "step": 220 }, { "epoch": 0.5045662100456622, "grad_norm": 60988.2421875, "learning_rate": 0.00010600899541429004, "loss": 1.1982, "step": 221 }, { "epoch": 0.5068493150684932, "grad_norm": 66918.2578125, "learning_rate": 0.00010525861369910877, "loss": 1.2181, "step": 222 }, { "epoch": 0.5091324200913242, "grad_norm": 79952.6640625, "learning_rate": 0.00010450793494292224, "loss": 1.1986, "step": 223 }, { "epoch": 0.5114155251141552, "grad_norm": 66536.671875, "learning_rate": 0.00010375700154899208, "loss": 1.5032, "step": 224 }, { "epoch": 0.5136986301369864, "grad_norm": 71625.8984375, "learning_rate": 0.00010300585593496348, "loss": 1.3998, "step": 225 }, { "epoch": 0.5159817351598174, "grad_norm": 66669.5703125, "learning_rate": 0.00010225454053046921, "loss": 0.8312, "step": 226 }, { "epoch": 0.5182648401826484, "grad_norm": 39306.3359375, "learning_rate": 0.00010150309777473306, "loss": 0.6047, "step": 227 }, { "epoch": 0.5205479452054794, "grad_norm": 53523.4609375, "learning_rate": 0.0001007515701141722, "loss": 1.1629, "step": 228 }, { "epoch": 0.5228310502283106, "grad_norm": 74079.578125, "learning_rate": 0.0001, "loss": 1.0456, "step": 229 }, { "epoch": 0.5251141552511416, "grad_norm": 62511.48046875, "learning_rate": 9.924842988582782e-05, "loss": 1.0753, "step": 230 }, { "epoch": 0.5273972602739726, "grad_norm": 53424.6328125, "learning_rate": 9.849690222526698e-05, "loss": 1.0496, "step": 231 }, { "epoch": 0.5296803652968036, "grad_norm": 76170.8671875, "learning_rate": 9.77454594695308e-05, "loss": 1.291, "step": 232 }, { "epoch": 0.5319634703196348, "grad_norm": 54918.74609375, "learning_rate": 9.699414406503654e-05, "loss": 1.0644, "step": 233 }, { "epoch": 0.5342465753424658, "grad_norm": 90057.484375, "learning_rate": 9.624299845100795e-05, "loss": 1.6448, "step": 234 }, { "epoch": 0.5365296803652968, "grad_norm": 50608.3125, "learning_rate": 9.549206505707777e-05, "loss": 1.0954, "step": 235 }, { "epoch": 0.5388127853881278, "grad_norm": 75716.2109375, "learning_rate": 9.474138630089124e-05, "loss": 1.2342, "step": 236 }, { "epoch": 0.541095890410959, "grad_norm": 43616.71875, "learning_rate": 9.399100458570997e-05, "loss": 0.8373, "step": 237 }, { "epoch": 0.54337899543379, "grad_norm": 51082.984375, "learning_rate": 9.324096229801674e-05, "loss": 1.0579, "step": 238 }, { "epoch": 0.545662100456621, "grad_norm": 64326.8359375, "learning_rate": 9.249130180512118e-05, "loss": 1.1863, "step": 239 }, { "epoch": 0.547945205479452, "grad_norm": 63323.40234375, "learning_rate": 9.174206545276677e-05, "loss": 1.1812, "step": 240 }, { "epoch": 0.5502283105022832, "grad_norm": 57428.61328125, "learning_rate": 9.099329556273866e-05, "loss": 1.0751, "step": 241 }, { "epoch": 0.5525114155251142, "grad_norm": 78367.109375, "learning_rate": 9.024503443047319e-05, "loss": 1.0871, "step": 242 }, { "epoch": 0.5547945205479452, "grad_norm": 56003.375, "learning_rate": 8.949732432266866e-05, "loss": 1.2125, "step": 243 }, { "epoch": 0.5570776255707762, "grad_norm": 50056.81640625, "learning_rate": 8.875020747489794e-05, "loss": 1.1738, "step": 244 }, { "epoch": 0.5593607305936074, "grad_norm": 54685.7265625, "learning_rate": 8.800372608922271e-05, "loss": 1.2343, "step": 245 }, { "epoch": 0.5616438356164384, "grad_norm": 75873.0, "learning_rate": 8.72579223318095e-05, "loss": 1.4155, "step": 246 }, { "epoch": 0.5639269406392694, "grad_norm": 73813.1171875, "learning_rate": 8.651283833054809e-05, "loss": 1.3641, "step": 247 }, { "epoch": 0.5662100456621004, "grad_norm": 52799.9140625, "learning_rate": 8.57685161726715e-05, "loss": 1.0217, "step": 248 }, { "epoch": 0.5684931506849316, "grad_norm": 53224.50390625, "learning_rate": 8.5024997902379e-05, "loss": 1.1589, "step": 249 }, { "epoch": 0.5707762557077626, "grad_norm": 101657.953125, "learning_rate": 8.428232551846101e-05, "loss": 1.5682, "step": 250 }, { "epoch": 0.5730593607305936, "grad_norm": 61659.9453125, "learning_rate": 8.35405409719266e-05, "loss": 1.0378, "step": 251 }, { "epoch": 0.5753424657534246, "grad_norm": 56497.0859375, "learning_rate": 8.279968616363418e-05, "loss": 1.2885, "step": 252 }, { "epoch": 0.5776255707762558, "grad_norm": 64090.8359375, "learning_rate": 8.205980294192421e-05, "loss": 1.3739, "step": 253 }, { "epoch": 0.5799086757990868, "grad_norm": 56467.49609375, "learning_rate": 8.132093310025571e-05, "loss": 1.0083, "step": 254 }, { "epoch": 0.5821917808219178, "grad_norm": 66332.6328125, "learning_rate": 8.058311837484535e-05, "loss": 1.2237, "step": 255 }, { "epoch": 0.5844748858447488, "grad_norm": 68149.140625, "learning_rate": 7.984640044230983e-05, "loss": 0.8762, "step": 256 }, { "epoch": 0.58675799086758, "grad_norm": 56869.44140625, "learning_rate": 7.911082091731181e-05, "loss": 1.4265, "step": 257 }, { "epoch": 0.589041095890411, "grad_norm": 66520.6640625, "learning_rate": 7.837642135020929e-05, "loss": 1.3615, "step": 258 }, { "epoch": 0.591324200913242, "grad_norm": 50649.890625, "learning_rate": 7.764324322470841e-05, "loss": 1.1493, "step": 259 }, { "epoch": 0.593607305936073, "grad_norm": 55484.89453125, "learning_rate": 7.691132795552043e-05, "loss": 1.125, "step": 260 }, { "epoch": 0.5958904109589042, "grad_norm": 69880.4765625, "learning_rate": 7.618071688602199e-05, "loss": 1.0974, "step": 261 }, { "epoch": 0.5981735159817352, "grad_norm": 69534.0234375, "learning_rate": 7.54514512859201e-05, "loss": 1.4658, "step": 262 }, { "epoch": 0.6004566210045662, "grad_norm": 63003.56640625, "learning_rate": 7.472357234892082e-05, "loss": 1.3326, "step": 263 }, { "epoch": 0.6027397260273972, "grad_norm": 48729.6796875, "learning_rate": 7.399712119040238e-05, "loss": 0.9631, "step": 264 }, { "epoch": 0.6050228310502284, "grad_norm": 53327.13671875, "learning_rate": 7.327213884509272e-05, "loss": 0.8053, "step": 265 }, { "epoch": 0.6073059360730594, "grad_norm": 56525.19921875, "learning_rate": 7.254866626475152e-05, "loss": 0.9795, "step": 266 }, { "epoch": 0.6095890410958904, "grad_norm": 54734.92578125, "learning_rate": 7.182674431585704e-05, "loss": 1.0737, "step": 267 }, { "epoch": 0.6118721461187214, "grad_norm": 62297.49609375, "learning_rate": 7.110641377729778e-05, "loss": 1.3582, "step": 268 }, { "epoch": 0.6141552511415526, "grad_norm": 69470.6484375, "learning_rate": 7.038771533806884e-05, "loss": 1.261, "step": 269 }, { "epoch": 0.6164383561643836, "grad_norm": 75963.8046875, "learning_rate": 6.967068959497376e-05, "loss": 1.2154, "step": 270 }, { "epoch": 0.6187214611872146, "grad_norm": 66193.421875, "learning_rate": 6.895537705033108e-05, "loss": 1.2711, "step": 271 }, { "epoch": 0.6210045662100456, "grad_norm": 53985.34765625, "learning_rate": 6.824181810968675e-05, "loss": 0.9666, "step": 272 }, { "epoch": 0.6232876712328768, "grad_norm": 54472.45703125, "learning_rate": 6.753005307953167e-05, "loss": 0.7533, "step": 273 }, { "epoch": 0.6255707762557078, "grad_norm": 56472.04296875, "learning_rate": 6.682012216502484e-05, "loss": 1.1015, "step": 274 }, { "epoch": 0.6278538812785388, "grad_norm": 59576.98046875, "learning_rate": 6.611206546772237e-05, "loss": 1.243, "step": 275 }, { "epoch": 0.6301369863013698, "grad_norm": 46372.140625, "learning_rate": 6.54059229833124e-05, "loss": 0.919, "step": 276 }, { "epoch": 0.632420091324201, "grad_norm": 67258.171875, "learning_rate": 6.47017345993556e-05, "loss": 1.2923, "step": 277 }, { "epoch": 0.634703196347032, "grad_norm": 97251.375, "learning_rate": 6.39995400930324e-05, "loss": 1.3966, "step": 278 }, { "epoch": 0.636986301369863, "grad_norm": 55997.97265625, "learning_rate": 6.329937912889582e-05, "loss": 0.7084, "step": 279 }, { "epoch": 0.639269406392694, "grad_norm": 52094.359375, "learning_rate": 6.260129125663106e-05, "loss": 0.9639, "step": 280 }, { "epoch": 0.6415525114155252, "grad_norm": 59095.65625, "learning_rate": 6.190531590882159e-05, "loss": 1.2343, "step": 281 }, { "epoch": 0.6438356164383562, "grad_norm": 50110.4375, "learning_rate": 6.121149239872151e-05, "loss": 1.1458, "step": 282 }, { "epoch": 0.6461187214611872, "grad_norm": 61831.3359375, "learning_rate": 6.051985991803517e-05, "loss": 1.1047, "step": 283 }, { "epoch": 0.6484018264840182, "grad_norm": 68382.28125, "learning_rate": 5.983045753470308e-05, "loss": 1.0395, "step": 284 }, { "epoch": 0.6506849315068494, "grad_norm": 52874.1796875, "learning_rate": 5.9143324190695196e-05, "loss": 1.1445, "step": 285 }, { "epoch": 0.6529680365296804, "grad_norm": 53148.92578125, "learning_rate": 5.845849869981137e-05, "loss": 1.1164, "step": 286 }, { "epoch": 0.6552511415525114, "grad_norm": 49849.4765625, "learning_rate": 5.777601974548866e-05, "loss": 0.8629, "step": 287 }, { "epoch": 0.6575342465753424, "grad_norm": 55761.171875, "learning_rate": 5.709592587861637e-05, "loss": 1.0324, "step": 288 }, { "epoch": 0.6598173515981736, "grad_norm": 47947.84765625, "learning_rate": 5.6418255515358486e-05, "loss": 0.9341, "step": 289 }, { "epoch": 0.6621004566210046, "grad_norm": 69258.109375, "learning_rate": 5.574304693498346e-05, "loss": 1.1231, "step": 290 }, { "epoch": 0.6643835616438356, "grad_norm": 48883.62109375, "learning_rate": 5.507033827770225e-05, "loss": 1.0446, "step": 291 }, { "epoch": 0.6666666666666666, "grad_norm": 48850.08203125, "learning_rate": 5.4400167542513636e-05, "loss": 1.0385, "step": 292 }, { "epoch": 0.6689497716894978, "grad_norm": 58763.58984375, "learning_rate": 5.3732572585057974e-05, "loss": 1.3623, "step": 293 }, { "epoch": 0.6712328767123288, "grad_norm": 53041.3046875, "learning_rate": 5.306759111547881e-05, "loss": 1.1103, "step": 294 }, { "epoch": 0.6735159817351598, "grad_norm": 54536.75390625, "learning_rate": 5.240526069629265e-05, "loss": 1.359, "step": 295 }, { "epoch": 0.6757990867579908, "grad_norm": 55320.109375, "learning_rate": 5.174561874026741e-05, "loss": 1.1521, "step": 296 }, { "epoch": 0.678082191780822, "grad_norm": 56988.97265625, "learning_rate": 5.108870250830882e-05, "loss": 1.2104, "step": 297 }, { "epoch": 0.680365296803653, "grad_norm": 65441.95703125, "learning_rate": 5.0434549107355944e-05, "loss": 1.0381, "step": 298 }, { "epoch": 0.682648401826484, "grad_norm": 54248.94921875, "learning_rate": 4.978319548828504e-05, "loss": 1.0877, "step": 299 }, { "epoch": 0.684931506849315, "grad_norm": 62817.375, "learning_rate": 4.9134678443822166e-05, "loss": 1.1492, "step": 300 }, { "epoch": 0.6872146118721462, "grad_norm": 55446.65234375, "learning_rate": 4.8489034606465225e-05, "loss": 1.078, "step": 301 }, { "epoch": 0.6894977168949772, "grad_norm": 61516.6484375, "learning_rate": 4.784630044641435e-05, "loss": 1.0592, "step": 302 }, { "epoch": 0.6917808219178082, "grad_norm": 53431.09375, "learning_rate": 4.7206512269512124e-05, "loss": 1.2311, "step": 303 }, { "epoch": 0.6940639269406392, "grad_norm": 68350.4609375, "learning_rate": 4.65697062151927e-05, "loss": 0.9918, "step": 304 }, { "epoch": 0.6963470319634704, "grad_norm": 53722.33984375, "learning_rate": 4.593591825444028e-05, "loss": 1.0563, "step": 305 }, { "epoch": 0.6986301369863014, "grad_norm": 73647.765625, "learning_rate": 4.530518418775733e-05, "loss": 1.1067, "step": 306 }, { "epoch": 0.7009132420091324, "grad_norm": 60410.765625, "learning_rate": 4.4677539643142454e-05, "loss": 1.2064, "step": 307 }, { "epoch": 0.7031963470319634, "grad_norm": 74342.1328125, "learning_rate": 4.40530200740777e-05, "loss": 1.2019, "step": 308 }, { "epoch": 0.7054794520547946, "grad_norm": 58122.41796875, "learning_rate": 4.343166075752605e-05, "loss": 1.0697, "step": 309 }, { "epoch": 0.7077625570776256, "grad_norm": 68661.671875, "learning_rate": 4.281349679193861e-05, "loss": 1.444, "step": 310 }, { "epoch": 0.7100456621004566, "grad_norm": 60156.05859375, "learning_rate": 4.2198563095272116e-05, "loss": 1.0489, "step": 311 }, { "epoch": 0.7123287671232876, "grad_norm": 72793.8359375, "learning_rate": 4.158689440301657e-05, "loss": 1.225, "step": 312 }, { "epoch": 0.7146118721461188, "grad_norm": 61000.60546875, "learning_rate": 4.097852526623307e-05, "loss": 1.1954, "step": 313 }, { "epoch": 0.7168949771689498, "grad_norm": 60486.58203125, "learning_rate": 4.0373490049602204e-05, "loss": 1.0551, "step": 314 }, { "epoch": 0.7191780821917808, "grad_norm": 91112.53125, "learning_rate": 3.977182292948283e-05, "loss": 1.2928, "step": 315 }, { "epoch": 0.7214611872146118, "grad_norm": 60165.15234375, "learning_rate": 3.9173557891981573e-05, "loss": 0.9197, "step": 316 }, { "epoch": 0.723744292237443, "grad_norm": 63943.1875, "learning_rate": 3.857872873103322e-05, "loss": 1.0148, "step": 317 }, { "epoch": 0.726027397260274, "grad_norm": 66577.5234375, "learning_rate": 3.7987369046491684e-05, "loss": 1.1527, "step": 318 }, { "epoch": 0.728310502283105, "grad_norm": 46133.828125, "learning_rate": 3.7399512242231995e-05, "loss": 0.6874, "step": 319 }, { "epoch": 0.730593607305936, "grad_norm": 54615.4296875, "learning_rate": 3.6815191524263624e-05, "loss": 1.2463, "step": 320 }, { "epoch": 0.7328767123287672, "grad_norm": 60014.6015625, "learning_rate": 3.623443989885462e-05, "loss": 1.3196, "step": 321 }, { "epoch": 0.7351598173515982, "grad_norm": 53860.69140625, "learning_rate": 3.565729017066729e-05, "loss": 1.0468, "step": 322 }, { "epoch": 0.7374429223744292, "grad_norm": 55772.921875, "learning_rate": 3.508377494090521e-05, "loss": 1.0851, "step": 323 }, { "epoch": 0.7397260273972602, "grad_norm": 53785.3203125, "learning_rate": 3.45139266054715e-05, "loss": 1.2084, "step": 324 }, { "epoch": 0.7420091324200914, "grad_norm": 57143.98046875, "learning_rate": 3.394777735313919e-05, "loss": 1.031, "step": 325 }, { "epoch": 0.7442922374429224, "grad_norm": 60805.40625, "learning_rate": 3.338535916373266e-05, "loss": 1.2843, "step": 326 }, { "epoch": 0.7465753424657534, "grad_norm": 56165.6875, "learning_rate": 3.2826703806321525e-05, "loss": 0.9815, "step": 327 }, { "epoch": 0.7488584474885844, "grad_norm": 65290.7421875, "learning_rate": 3.227184283742591e-05, "loss": 1.2205, "step": 328 }, { "epoch": 0.7511415525114156, "grad_norm": 59236.3828125, "learning_rate": 3.17208075992339e-05, "loss": 1.0087, "step": 329 }, { "epoch": 0.7534246575342466, "grad_norm": 68966.4921875, "learning_rate": 3.117362921783134e-05, "loss": 1.2744, "step": 330 }, { "epoch": 0.7534246575342466, "eval_loss": 1.1098405122756958, "eval_runtime": 5.9855, "eval_samples_per_second": 16.707, "eval_steps_per_second": 8.354, "step": 330 }, { "epoch": 0.7557077625570776, "grad_norm": 57247.7265625, "learning_rate": 3.063033860144339e-05, "loss": 1.3947, "step": 331 }, { "epoch": 0.7579908675799086, "grad_norm": 60022.890625, "learning_rate": 3.0090966438688772e-05, "loss": 1.4142, "step": 332 }, { "epoch": 0.7602739726027398, "grad_norm": 48251.8359375, "learning_rate": 2.9555543196846292e-05, "loss": 0.9917, "step": 333 }, { "epoch": 0.7625570776255708, "grad_norm": 50926.33203125, "learning_rate": 2.9024099120133673e-05, "loss": 1.0497, "step": 334 }, { "epoch": 0.7648401826484018, "grad_norm": 51535.24609375, "learning_rate": 2.8496664227999415e-05, "loss": 1.0956, "step": 335 }, { "epoch": 0.7671232876712328, "grad_norm": 47902.59375, "learning_rate": 2.7973268313426837e-05, "loss": 1.182, "step": 336 }, { "epoch": 0.769406392694064, "grad_norm": 57668.80859375, "learning_rate": 2.745394094125141e-05, "loss": 1.1184, "step": 337 }, { "epoch": 0.771689497716895, "grad_norm": 65797.921875, "learning_rate": 2.6938711446490606e-05, "loss": 1.524, "step": 338 }, { "epoch": 0.773972602739726, "grad_norm": 44397.484375, "learning_rate": 2.6427608932686843e-05, "loss": 1.0515, "step": 339 }, { "epoch": 0.776255707762557, "grad_norm": 88163.90625, "learning_rate": 2.5920662270263653e-05, "loss": 1.291, "step": 340 }, { "epoch": 0.7785388127853882, "grad_norm": 91966.3046875, "learning_rate": 2.5417900094894744e-05, "loss": 1.5279, "step": 341 }, { "epoch": 0.7808219178082192, "grad_norm": 54689.46875, "learning_rate": 2.4919350805886577e-05, "loss": 1.0487, "step": 342 }, { "epoch": 0.7831050228310502, "grad_norm": 60095.0703125, "learning_rate": 2.4425042564574184e-05, "loss": 1.3099, "step": 343 }, { "epoch": 0.7853881278538812, "grad_norm": 54892.97265625, "learning_rate": 2.3935003292730296e-05, "loss": 1.1576, "step": 344 }, { "epoch": 0.7876712328767124, "grad_norm": 99210.53125, "learning_rate": 2.344926067098836e-05, "loss": 1.846, "step": 345 }, { "epoch": 0.7899543378995434, "grad_norm": 65337.87890625, "learning_rate": 2.2967842137278706e-05, "loss": 1.1789, "step": 346 }, { "epoch": 0.7922374429223744, "grad_norm": 71690.0625, "learning_rate": 2.2490774885278908e-05, "loss": 1.3072, "step": 347 }, { "epoch": 0.7945205479452054, "grad_norm": 58516.3828125, "learning_rate": 2.201808586287757e-05, "loss": 1.0443, "step": 348 }, { "epoch": 0.7968036529680366, "grad_norm": 72086.8828125, "learning_rate": 2.15498017706521e-05, "loss": 1.1584, "step": 349 }, { "epoch": 0.7990867579908676, "grad_norm": 58905.08984375, "learning_rate": 2.1085949060360654e-05, "loss": 1.2844, "step": 350 }, { "epoch": 0.8013698630136986, "grad_norm": 56423.5703125, "learning_rate": 2.0626553933447734e-05, "loss": 1.1755, "step": 351 }, { "epoch": 0.8036529680365296, "grad_norm": 55340.45703125, "learning_rate": 2.01716423395644e-05, "loss": 1.08, "step": 352 }, { "epoch": 0.8059360730593608, "grad_norm": 56130.5859375, "learning_rate": 1.9721239975102313e-05, "loss": 1.1129, "step": 353 }, { "epoch": 0.8082191780821918, "grad_norm": 58951.8046875, "learning_rate": 1.9275372281742242e-05, "loss": 0.9749, "step": 354 }, { "epoch": 0.8105022831050228, "grad_norm": 58070.171875, "learning_rate": 1.8834064445016953e-05, "loss": 1.3564, "step": 355 }, { "epoch": 0.8127853881278538, "grad_norm": 56061.4921875, "learning_rate": 1.839734139288868e-05, "loss": 1.2425, "step": 356 }, { "epoch": 0.815068493150685, "grad_norm": 56343.59375, "learning_rate": 1.7965227794340877e-05, "loss": 0.9662, "step": 357 }, { "epoch": 0.817351598173516, "grad_norm": 61289.234375, "learning_rate": 1.753774805798486e-05, "loss": 1.1741, "step": 358 }, { "epoch": 0.819634703196347, "grad_norm": 98183.703125, "learning_rate": 1.7114926330680957e-05, "loss": 0.9826, "step": 359 }, { "epoch": 0.821917808219178, "grad_norm": 55233.2890625, "learning_rate": 1.6696786496174578e-05, "loss": 1.0732, "step": 360 }, { "epoch": 0.8242009132420092, "grad_norm": 45176.41015625, "learning_rate": 1.6283352173747145e-05, "loss": 1.0048, "step": 361 }, { "epoch": 0.8264840182648402, "grad_norm": 66357.2109375, "learning_rate": 1.587464671688187e-05, "loss": 1.1619, "step": 362 }, { "epoch": 0.8287671232876712, "grad_norm": 57376.734375, "learning_rate": 1.5470693211944643e-05, "loss": 1.1175, "step": 363 }, { "epoch": 0.8310502283105022, "grad_norm": 54212.546875, "learning_rate": 1.5071514476879878e-05, "loss": 0.9726, "step": 364 }, { "epoch": 0.8333333333333334, "grad_norm": 65802.8515625, "learning_rate": 1.4677133059921632e-05, "loss": 1.3092, "step": 365 }, { "epoch": 0.8356164383561644, "grad_norm": 49888.99609375, "learning_rate": 1.4287571238320053e-05, "loss": 0.5516, "step": 366 }, { "epoch": 0.8378995433789954, "grad_norm": 66406.1875, "learning_rate": 1.3902851017082864e-05, "loss": 0.9952, "step": 367 }, { "epoch": 0.8401826484018264, "grad_norm": 57395.77734375, "learning_rate": 1.3522994127732414e-05, "loss": 1.1279, "step": 368 }, { "epoch": 0.8424657534246576, "grad_norm": 79530.40625, "learning_rate": 1.3148022027078222e-05, "loss": 1.8326, "step": 369 }, { "epoch": 0.8447488584474886, "grad_norm": 63298.96875, "learning_rate": 1.2777955896004812e-05, "loss": 1.0041, "step": 370 }, { "epoch": 0.8470319634703196, "grad_norm": 74275.5234375, "learning_rate": 1.2412816638275404e-05, "loss": 1.3752, "step": 371 }, { "epoch": 0.8493150684931506, "grad_norm": 57181.0390625, "learning_rate": 1.2052624879351104e-05, "loss": 1.3001, "step": 372 }, { "epoch": 0.8515981735159818, "grad_norm": 56421.484375, "learning_rate": 1.1697400965225747e-05, "loss": 1.1496, "step": 373 }, { "epoch": 0.8538812785388128, "grad_norm": 55708.265625, "learning_rate": 1.134716496127679e-05, "loss": 1.0953, "step": 374 }, { "epoch": 0.8561643835616438, "grad_norm": 64520.75, "learning_rate": 1.1001936651131717e-05, "loss": 0.9648, "step": 375 }, { "epoch": 0.8584474885844748, "grad_norm": 67208.2578125, "learning_rate": 1.0661735535550666e-05, "loss": 1.465, "step": 376 }, { "epoch": 0.860730593607306, "grad_norm": 69055.765625, "learning_rate": 1.0326580831324817e-05, "loss": 1.1437, "step": 377 }, { "epoch": 0.863013698630137, "grad_norm": 56862.21875, "learning_rate": 9.996491470190917e-06, "loss": 1.1829, "step": 378 }, { "epoch": 0.865296803652968, "grad_norm": 43474.109375, "learning_rate": 9.671486097761917e-06, "loss": 0.7937, "step": 379 }, { "epoch": 0.867579908675799, "grad_norm": 58035.9140625, "learning_rate": 9.351583072473713e-06, "loss": 1.2742, "step": 380 }, { "epoch": 0.8698630136986302, "grad_norm": 59717.421875, "learning_rate": 9.036800464548157e-06, "loss": 1.245, "step": 381 }, { "epoch": 0.8721461187214612, "grad_norm": 100361.84375, "learning_rate": 8.727156054972374e-06, "loss": 0.9628, "step": 382 }, { "epoch": 0.8744292237442922, "grad_norm": 46780.90234375, "learning_rate": 8.422667334494249e-06, "loss": 0.7594, "step": 383 }, { "epoch": 0.8767123287671232, "grad_norm": 58426.484375, "learning_rate": 8.123351502634625e-06, "loss": 1.3035, "step": 384 }, { "epoch": 0.8789954337899544, "grad_norm": 55850.3046875, "learning_rate": 7.82922546671555e-06, "loss": 1.3021, "step": 385 }, { "epoch": 0.8812785388127854, "grad_norm": 65162.09375, "learning_rate": 7.54030584090537e-06, "loss": 1.214, "step": 386 }, { "epoch": 0.8835616438356164, "grad_norm": 50473.97265625, "learning_rate": 7.256608945280319e-06, "loss": 1.1281, "step": 387 }, { "epoch": 0.8858447488584474, "grad_norm": 50409.5390625, "learning_rate": 6.97815080490245e-06, "loss": 1.2092, "step": 388 }, { "epoch": 0.8881278538812786, "grad_norm": 56921.63671875, "learning_rate": 6.704947148914609e-06, "loss": 1.0073, "step": 389 }, { "epoch": 0.8904109589041096, "grad_norm": 72200.1875, "learning_rate": 6.437013409651849e-06, "loss": 1.0158, "step": 390 }, { "epoch": 0.8926940639269406, "grad_norm": 48175.20703125, "learning_rate": 6.174364721769743e-06, "loss": 0.939, "step": 391 }, { "epoch": 0.8949771689497716, "grad_norm": 53166.9296875, "learning_rate": 5.917015921389568e-06, "loss": 1.0468, "step": 392 }, { "epoch": 0.8972602739726028, "grad_norm": 58858.93359375, "learning_rate": 5.664981545260073e-06, "loss": 1.0791, "step": 393 }, { "epoch": 0.8995433789954338, "grad_norm": 49101.5078125, "learning_rate": 5.418275829936537e-06, "loss": 1.1875, "step": 394 }, { "epoch": 0.9018264840182648, "grad_norm": 65572.046875, "learning_rate": 5.176912710976467e-06, "loss": 1.326, "step": 395 }, { "epoch": 0.9041095890410958, "grad_norm": 61459.63671875, "learning_rate": 4.940905822152453e-06, "loss": 1.0855, "step": 396 }, { "epoch": 0.906392694063927, "grad_norm": 64900.1796875, "learning_rate": 4.710268494682146e-06, "loss": 1.3393, "step": 397 }, { "epoch": 0.908675799086758, "grad_norm": 52331.64453125, "learning_rate": 4.485013756475076e-06, "loss": 1.0434, "step": 398 }, { "epoch": 0.910958904109589, "grad_norm": 71229.609375, "learning_rate": 4.2651543313968145e-06, "loss": 1.2118, "step": 399 }, { "epoch": 0.91324200913242, "grad_norm": 63443.84375, "learning_rate": 4.050702638550275e-06, "loss": 1.1458, "step": 400 }, { "epoch": 0.9155251141552512, "grad_norm": 64409.73828125, "learning_rate": 3.841670791574137e-06, "loss": 1.1705, "step": 401 }, { "epoch": 0.9178082191780822, "grad_norm": 81092.15625, "learning_rate": 3.638070597958665e-06, "loss": 1.2991, "step": 402 }, { "epoch": 0.9200913242009132, "grad_norm": 41609.2109375, "learning_rate": 3.4399135583787043e-06, "loss": 0.4622, "step": 403 }, { "epoch": 0.9223744292237442, "grad_norm": 58560.71875, "learning_rate": 3.2472108660439706e-06, "loss": 0.8759, "step": 404 }, { "epoch": 0.9246575342465754, "grad_norm": 65780.9609375, "learning_rate": 3.059973406066963e-06, "loss": 1.3583, "step": 405 }, { "epoch": 0.9269406392694064, "grad_norm": 53281.921875, "learning_rate": 2.878211754847926e-06, "loss": 0.971, "step": 406 }, { "epoch": 0.9292237442922374, "grad_norm": 62492.41015625, "learning_rate": 2.7019361794775156e-06, "loss": 1.1152, "step": 407 }, { "epoch": 0.9315068493150684, "grad_norm": 57761.875, "learning_rate": 2.5311566371568507e-06, "loss": 1.1735, "step": 408 }, { "epoch": 0.9337899543378996, "grad_norm": 53913.1640625, "learning_rate": 2.365882774634998e-06, "loss": 0.9759, "step": 409 }, { "epoch": 0.9360730593607306, "grad_norm": 57200.4375, "learning_rate": 2.206123927664161e-06, "loss": 1.262, "step": 410 }, { "epoch": 0.9383561643835616, "grad_norm": 64508.68359375, "learning_rate": 2.0518891204722168e-06, "loss": 1.1443, "step": 411 }, { "epoch": 0.9406392694063926, "grad_norm": 68100.71875, "learning_rate": 1.903187065253076e-06, "loss": 1.156, "step": 412 }, { "epoch": 0.9429223744292238, "grad_norm": 46483.26171875, "learning_rate": 1.7600261616745106e-06, "loss": 0.9084, "step": 413 }, { "epoch": 0.9452054794520548, "grad_norm": 75558.0078125, "learning_rate": 1.6224144964036681e-06, "loss": 1.3323, "step": 414 }, { "epoch": 0.9474885844748858, "grad_norm": 61868.96484375, "learning_rate": 1.4903598426503241e-06, "loss": 1.6048, "step": 415 }, { "epoch": 0.9497716894977168, "grad_norm": 48915.76953125, "learning_rate": 1.3638696597277679e-06, "loss": 0.9012, "step": 416 }, { "epoch": 0.952054794520548, "grad_norm": 75366.640625, "learning_rate": 1.2429510926314836e-06, "loss": 1.132, "step": 417 }, { "epoch": 0.954337899543379, "grad_norm": 60648.18359375, "learning_rate": 1.1276109716355287e-06, "loss": 1.0748, "step": 418 }, { "epoch": 0.95662100456621, "grad_norm": 55365.9765625, "learning_rate": 1.0178558119067315e-06, "loss": 0.827, "step": 419 }, { "epoch": 0.958904109589041, "grad_norm": 58780.57421875, "learning_rate": 9.136918131366412e-07, "loss": 1.2993, "step": 420 }, { "epoch": 0.9611872146118722, "grad_norm": 47850.671875, "learning_rate": 8.151248591913518e-07, "loss": 1.0673, "step": 421 }, { "epoch": 0.9634703196347032, "grad_norm": 50755.59765625, "learning_rate": 7.221605177791691e-07, "loss": 1.0819, "step": 422 }, { "epoch": 0.9657534246575342, "grad_norm": 68484.7265625, "learning_rate": 6.348040401360833e-07, "loss": 1.3769, "step": 423 }, { "epoch": 0.9680365296803652, "grad_norm": 59223.92578125, "learning_rate": 5.530603607290851e-07, "loss": 1.3159, "step": 424 }, { "epoch": 0.9703196347031964, "grad_norm": 63650.15234375, "learning_rate": 4.76934096977566e-07, "loss": 1.3134, "step": 425 }, { "epoch": 0.9726027397260274, "grad_norm": 62978.29296875, "learning_rate": 4.0642954899238197e-07, "loss": 0.9593, "step": 426 }, { "epoch": 0.9748858447488584, "grad_norm": 63336.55078125, "learning_rate": 3.415506993330153e-07, "loss": 1.1585, "step": 427 }, { "epoch": 0.9771689497716894, "grad_norm": 58021.15625, "learning_rate": 2.8230121278257637e-07, "loss": 1.2236, "step": 428 }, { "epoch": 0.9794520547945206, "grad_norm": 56498.8515625, "learning_rate": 2.2868443614082469e-07, "loss": 1.0606, "step": 429 }, { "epoch": 0.9817351598173516, "grad_norm": 48720.328125, "learning_rate": 1.8070339803509807e-07, "loss": 1.123, "step": 430 }, { "epoch": 0.9840182648401826, "grad_norm": 60693.24609375, "learning_rate": 1.3836080874926049e-07, "loss": 0.5269, "step": 431 }, { "epoch": 0.9863013698630136, "grad_norm": 123484.75, "learning_rate": 1.0165906007056914e-07, "loss": 1.2241, "step": 432 }, { "epoch": 0.9885844748858448, "grad_norm": 52265.00390625, "learning_rate": 7.060022515460451e-08, "loss": 0.8173, "step": 433 }, { "epoch": 0.9908675799086758, "grad_norm": 54763.09375, "learning_rate": 4.518605840815315e-08, "loss": 1.2235, "step": 434 }, { "epoch": 0.9931506849315068, "grad_norm": 67981.8515625, "learning_rate": 2.5417995390086824e-08, "loss": 1.3369, "step": 435 }, { "epoch": 0.9954337899543378, "grad_norm": 53286.59375, "learning_rate": 1.129715273033849e-08, "loss": 0.9584, "step": 436 }, { "epoch": 0.997716894977169, "grad_norm": 54882.32421875, "learning_rate": 2.824328066730608e-09, "loss": 1.0711, "step": 437 }, { "epoch": 1.0, "grad_norm": 54870.58984375, "learning_rate": 0.0, "loss": 1.1062, "step": 438 } ], "logging_steps": 1, "max_steps": 438, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 7972410614906880.0, "train_batch_size": 2, "trial_name": null, "trial_params": null }