diff --git "a/trainer_state.json" "b/trainer_state.json" --- "a/trainer_state.json" +++ "b/trainer_state.json" @@ -1,3964 +1,2485 @@ { "best_metric": null, "best_model_checkpoint": null, - "epoch": 9.98185117967332, + "epoch": 15.0, "eval_steps": 500, - "global_step": 2750, + "global_step": 1665, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { - "epoch": 0.003629764065335753, - "grad_norm": 516.0, - "learning_rate": 7.272727272727273e-07, - "loss": 51.5982, + "epoch": 0.009009009009009009, + "grad_norm": 608.0, + "learning_rate": 1.1976047904191619e-06, + "loss": 58.5641, "step": 1 }, { - "epoch": 0.018148820326678767, - "grad_norm": 454.0, - "learning_rate": 3.636363636363636e-06, - "loss": 52.3483, + "epoch": 0.04504504504504504, + "grad_norm": 536.0, + "learning_rate": 5.9880239520958085e-06, + "loss": 54.74, "step": 5 }, { - "epoch": 0.036297640653357534, - "grad_norm": 536.0, - "learning_rate": 7.272727272727272e-06, - "loss": 53.1669, + "epoch": 0.09009009009009009, + "grad_norm": 488.0, + "learning_rate": 1.1976047904191617e-05, + "loss": 52.9637, "step": 10 }, { - "epoch": 0.0544464609800363, - "grad_norm": 440.0, - "learning_rate": 1.0909090909090909e-05, - "loss": 46.8624, + "epoch": 0.13513513513513514, + "grad_norm": 392.0, + "learning_rate": 1.7964071856287426e-05, + "loss": 40.7219, "step": 15 }, { - "epoch": 0.07259528130671507, - "grad_norm": 306.0, - "learning_rate": 1.4545454545454545e-05, - "loss": 35.5942, + "epoch": 0.18018018018018017, + "grad_norm": 46.5, + "learning_rate": 2.3952095808383234e-05, + "loss": 28.1416, "step": 20 }, { - "epoch": 0.09074410163339383, - "grad_norm": 41.5, - "learning_rate": 1.8181818181818182e-05, - "loss": 25.8916, + "epoch": 0.22522522522522523, + "grad_norm": 40.25, + "learning_rate": 2.994011976047904e-05, + "loss": 24.2461, "step": 25 }, { - "epoch": 0.1088929219600726, - "grad_norm": 44.0, - "learning_rate": 2.1818181818181818e-05, - "loss": 24.4549, + "epoch": 0.2702702702702703, + "grad_norm": 24.125, + "learning_rate": 3.592814371257485e-05, + "loss": 22.2451, "step": 30 }, { - "epoch": 0.12704174228675136, - "grad_norm": 24.375, - "learning_rate": 2.5454545454545454e-05, - "loss": 22.3596, + "epoch": 0.3153153153153153, + "grad_norm": 13.5625, + "learning_rate": 4.191616766467066e-05, + "loss": 20.2199, "step": 35 }, { - "epoch": 0.14519056261343014, - "grad_norm": 20.0, - "learning_rate": 2.909090909090909e-05, - "loss": 21.008, + "epoch": 0.36036036036036034, + "grad_norm": 7.5625, + "learning_rate": 4.790419161676647e-05, + "loss": 19.4029, "step": 40 }, { - "epoch": 0.16333938294010888, - "grad_norm": 7.4375, - "learning_rate": 3.272727272727273e-05, - "loss": 20.0885, + "epoch": 0.40540540540540543, + "grad_norm": 9.1875, + "learning_rate": 5.389221556886228e-05, + "loss": 18.5345, "step": 45 }, { - "epoch": 0.18148820326678766, - "grad_norm": 7.28125, - "learning_rate": 3.6363636363636364e-05, - "loss": 19.1877, + "epoch": 0.45045045045045046, + "grad_norm": 15.75, + "learning_rate": 5.988023952095808e-05, + "loss": 17.2917, "step": 50 }, { - "epoch": 0.1996370235934664, - "grad_norm": 10.25, - "learning_rate": 4e-05, - "loss": 18.5964, + "epoch": 0.4954954954954955, + "grad_norm": 37.25, + "learning_rate": 6.58682634730539e-05, + "loss": 15.8479, "step": 55 }, { - "epoch": 0.2177858439201452, - "grad_norm": 14.5, - "learning_rate": 4.3636363636363636e-05, - "loss": 17.6581, + "epoch": 0.5405405405405406, + "grad_norm": 59.75, + "learning_rate": 7.18562874251497e-05, + "loss": 10.5192, "step": 60 }, { - "epoch": 0.23593466424682397, - "grad_norm": 26.5, - "learning_rate": 4.7272727272727275e-05, - "loss": 16.1181, + "epoch": 0.5855855855855856, + "grad_norm": 13.375, + "learning_rate": 7.784431137724552e-05, + "loss": 3.399, "step": 65 }, { - "epoch": 0.2540834845735027, - "grad_norm": 49.75, - "learning_rate": 5.090909090909091e-05, - "loss": 13.6493, + "epoch": 0.6306306306306306, + "grad_norm": 2.84375, + "learning_rate": 8.383233532934131e-05, + "loss": 2.1324, "step": 70 }, { - "epoch": 0.27223230490018147, - "grad_norm": 50.75, - "learning_rate": 5.4545454545454546e-05, - "loss": 7.9128, + "epoch": 0.6756756756756757, + "grad_norm": 2.125, + "learning_rate": 8.982035928143712e-05, + "loss": 1.8224, "step": 75 }, { - "epoch": 0.29038112522686027, - "grad_norm": 6.625, - "learning_rate": 5.818181818181818e-05, - "loss": 2.7543, + "epoch": 0.7207207207207207, + "grad_norm": 3.703125, + "learning_rate": 9.580838323353294e-05, + "loss": 1.6034, "step": 80 }, { - "epoch": 0.308529945553539, - "grad_norm": 12.375, - "learning_rate": 6.181818181818182e-05, - "loss": 2.0995, + "epoch": 0.7657657657657657, + "grad_norm": 2.75, + "learning_rate": 0.00010179640718562875, + "loss": 1.4811, "step": 85 }, { - "epoch": 0.32667876588021777, - "grad_norm": 2.53125, - "learning_rate": 6.545454545454546e-05, - "loss": 1.8007, + "epoch": 0.8108108108108109, + "grad_norm": 7.53125, + "learning_rate": 0.00010778443113772456, + "loss": 1.4187, "step": 90 }, { - "epoch": 0.3448275862068966, - "grad_norm": 1.984375, - "learning_rate": 6.90909090909091e-05, - "loss": 1.6358, + "epoch": 0.8558558558558559, + "grad_norm": 2.234375, + "learning_rate": 0.00011377245508982037, + "loss": 1.3517, "step": 95 }, { - "epoch": 0.3629764065335753, - "grad_norm": 2.828125, - "learning_rate": 7.272727272727273e-05, - "loss": 1.509, + "epoch": 0.9009009009009009, + "grad_norm": 8.375, + "learning_rate": 0.00011976047904191617, + "loss": 1.2567, "step": 100 }, { - "epoch": 0.3811252268602541, - "grad_norm": 4.0, - "learning_rate": 7.636363636363637e-05, - "loss": 1.4215, + "epoch": 0.9459459459459459, + "grad_norm": 5.90625, + "learning_rate": 0.00012574850299401196, + "loss": 1.2634, "step": 105 }, { - "epoch": 0.3992740471869328, - "grad_norm": 3.21875, - "learning_rate": 8e-05, - "loss": 1.3173, + "epoch": 0.990990990990991, + "grad_norm": 4.1875, + "learning_rate": 0.0001317365269461078, + "loss": 1.2272, "step": 110 }, { - "epoch": 0.41742286751361163, - "grad_norm": 2.625, - "learning_rate": 8.363636363636364e-05, - "loss": 1.2786, + "epoch": 1.0, + "eval_loss": 2.3900420665740967, + "eval_runtime": 1.0174, + "eval_samples_per_second": 4.914, + "eval_steps_per_second": 1.966, + "step": 111 + }, + { + "epoch": 1.0360360360360361, + "grad_norm": 3.234375, + "learning_rate": 0.00013772455089820359, + "loss": 1.1598, "step": 115 }, { - "epoch": 0.4355716878402904, - "grad_norm": 2.8125, - "learning_rate": 8.727272727272727e-05, - "loss": 1.2548, + "epoch": 1.0810810810810811, + "grad_norm": 2.3125, + "learning_rate": 0.0001437125748502994, + "loss": 1.1474, "step": 120 }, { - "epoch": 0.4537205081669691, - "grad_norm": 2.96875, - "learning_rate": 9.090909090909092e-05, - "loss": 1.1995, + "epoch": 1.1261261261261262, + "grad_norm": 6.375, + "learning_rate": 0.0001497005988023952, + "loss": 1.1051, "step": 125 }, { - "epoch": 0.47186932849364793, - "grad_norm": 1.390625, - "learning_rate": 9.454545454545455e-05, - "loss": 1.1818, + "epoch": 1.1711711711711712, + "grad_norm": 6.625, + "learning_rate": 0.00015568862275449103, + "loss": 1.0869, "step": 130 }, { - "epoch": 0.4900181488203267, - "grad_norm": 5.78125, - "learning_rate": 9.818181818181818e-05, - "loss": 1.1444, + "epoch": 1.2162162162162162, + "grad_norm": 1.0703125, + "learning_rate": 0.00016167664670658683, + "loss": 1.0864, "step": 135 }, { - "epoch": 0.5081669691470054, - "grad_norm": 7.34375, - "learning_rate": 0.00010181818181818181, - "loss": 1.0977, + "epoch": 1.2612612612612613, + "grad_norm": 6.1875, + "learning_rate": 0.00016766467065868263, + "loss": 1.0419, "step": 140 }, { - "epoch": 0.5263157894736842, - "grad_norm": 5.1875, - "learning_rate": 0.00010545454545454545, - "loss": 1.0818, + "epoch": 1.3063063063063063, + "grad_norm": 2.84375, + "learning_rate": 0.00017365269461077845, + "loss": 1.0647, "step": 145 }, { - "epoch": 0.5444646098003629, - "grad_norm": 1.984375, - "learning_rate": 0.00010909090909090909, - "loss": 1.0524, + "epoch": 1.3513513513513513, + "grad_norm": 24.0, + "learning_rate": 0.00017964071856287425, + "loss": 1.0682, "step": 150 }, { - "epoch": 0.5626134301270418, - "grad_norm": 20.5, - "learning_rate": 0.00011272727272727272, - "loss": 1.0397, + "epoch": 1.3963963963963963, + "grad_norm": 5.3125, + "learning_rate": 0.00018562874251497007, + "loss": 1.0375, "step": 155 }, { - "epoch": 0.5807622504537205, - "grad_norm": 7.71875, - "learning_rate": 0.00011636363636363636, - "loss": 1.0695, + "epoch": 1.4414414414414414, + "grad_norm": 2.5625, + "learning_rate": 0.00019161676646706587, + "loss": 1.0111, "step": 160 }, { - "epoch": 0.5989110707803993, - "grad_norm": 5.28125, - "learning_rate": 0.00012, - "loss": 1.0426, + "epoch": 1.4864864864864864, + "grad_norm": 4.25, + "learning_rate": 0.0001976047904191617, + "loss": 0.9605, "step": 165 }, { - "epoch": 0.617059891107078, - "grad_norm": 1.453125, - "learning_rate": 0.00012363636363636364, - "loss": 1.0295, + "epoch": 1.5315315315315314, + "grad_norm": 6.15625, + "learning_rate": 0.00019999802081131313, + "loss": 0.9843, "step": 170 }, { - "epoch": 0.6352087114337568, - "grad_norm": 17.75, - "learning_rate": 0.00012727272727272728, - "loss": 0.9987, + "epoch": 1.5765765765765765, + "grad_norm": 2.484375, + "learning_rate": 0.00019998592605305064, + "loss": 1.0225, "step": 175 }, { - "epoch": 0.6533575317604355, - "grad_norm": 1.3515625, - "learning_rate": 0.00013090909090909093, - "loss": 0.9895, + "epoch": 1.6216216216216215, + "grad_norm": 2.3125, + "learning_rate": 0.00019996283741405458, + "loss": 0.9908, "step": 180 }, { - "epoch": 0.6715063520871143, - "grad_norm": 1.2421875, - "learning_rate": 0.00013454545454545455, - "loss": 1.0001, + "epoch": 1.6666666666666665, + "grad_norm": 28.375, + "learning_rate": 0.00019992875743301967, + "loss": 1.0038, "step": 185 }, { - "epoch": 0.6896551724137931, - "grad_norm": 1.3828125, - "learning_rate": 0.0001381818181818182, - "loss": 0.9797, + "epoch": 1.7117117117117115, + "grad_norm": 2.390625, + "learning_rate": 0.00019988368985718576, + "loss": 0.9688, "step": 190 }, { - "epoch": 0.7078039927404719, - "grad_norm": 0.84375, - "learning_rate": 0.00014181818181818184, - "loss": 0.9846, + "epoch": 1.7567567567567568, + "grad_norm": 1.75, + "learning_rate": 0.00019982763964192585, + "loss": 0.9638, "step": 195 }, { - "epoch": 0.7259528130671506, - "grad_norm": 1.8671875, - "learning_rate": 0.00014545454545454546, - "loss": 0.9978, + "epoch": 1.8018018018018018, + "grad_norm": 3.125, + "learning_rate": 0.00019976061295020106, + "loss": 0.9878, "step": 200 }, { - "epoch": 0.7441016333938294, - "grad_norm": 1.1640625, - "learning_rate": 0.0001490909090909091, - "loss": 0.9639, + "epoch": 1.8468468468468469, + "grad_norm": 1.921875, + "learning_rate": 0.00019968261715188324, + "loss": 0.9394, "step": 205 }, { - "epoch": 0.7622504537205081, - "grad_norm": 0.78125, - "learning_rate": 0.00015272727272727275, - "loss": 0.9302, + "epoch": 1.8918918918918919, + "grad_norm": 1.546875, + "learning_rate": 0.00019959366082294438, + "loss": 0.9285, "step": 210 }, { - "epoch": 0.7803992740471869, - "grad_norm": 2.390625, - "learning_rate": 0.00015636363636363637, - "loss": 0.9798, + "epoch": 1.936936936936937, + "grad_norm": 3.53125, + "learning_rate": 0.00019949375374451377, + "loss": 0.929, "step": 215 }, { - "epoch": 0.7985480943738656, - "grad_norm": 2.390625, - "learning_rate": 0.00016, - "loss": 0.9477, + "epoch": 1.981981981981982, + "grad_norm": 1.625, + "learning_rate": 0.00019938290690180247, + "loss": 0.9374, "step": 220 }, { - "epoch": 0.8166969147005445, - "grad_norm": 3.265625, - "learning_rate": 0.00016363636363636366, - "loss": 0.9237, + "epoch": 2.0, + "eval_loss": 2.192767381668091, + "eval_runtime": 1.0169, + "eval_samples_per_second": 4.917, + "eval_steps_per_second": 1.967, + "step": 222 + }, + { + "epoch": 2.027027027027027, + "grad_norm": 2.25, + "learning_rate": 0.0001992611324828956, + "loss": 0.8849, "step": 225 }, { - "epoch": 0.8348457350272233, - "grad_norm": 4.34375, - "learning_rate": 0.00016727272727272728, - "loss": 0.9335, + "epoch": 2.0720720720720722, + "grad_norm": 1.8125, + "learning_rate": 0.00019912844387741195, + "loss": 0.864, "step": 230 }, { - "epoch": 0.852994555353902, - "grad_norm": 3.640625, - "learning_rate": 0.0001709090909090909, - "loss": 0.9452, + "epoch": 2.1171171171171173, + "grad_norm": 1.09375, + "learning_rate": 0.00019898485567503187, + "loss": 0.8497, "step": 235 }, { - "epoch": 0.8711433756805808, - "grad_norm": 3.265625, - "learning_rate": 0.00017454545454545454, - "loss": 0.9112, + "epoch": 2.1621621621621623, + "grad_norm": 1.9609375, + "learning_rate": 0.00019883038366389313, + "loss": 0.8844, "step": 240 }, { - "epoch": 0.8892921960072595, - "grad_norm": 3.796875, - "learning_rate": 0.0001781818181818182, - "loss": 0.9211, + "epoch": 2.2072072072072073, + "grad_norm": 1.0234375, + "learning_rate": 0.0001986650448288548, + "loss": 0.8509, "step": 245 }, { - "epoch": 0.9074410163339383, - "grad_norm": 6.125, - "learning_rate": 0.00018181818181818183, - "loss": 0.9188, + "epoch": 2.2522522522522523, + "grad_norm": 3.078125, + "learning_rate": 0.00019848885734962982, + "loss": 0.8718, "step": 250 }, { - "epoch": 0.925589836660617, - "grad_norm": 2.609375, - "learning_rate": 0.00018545454545454545, - "loss": 0.937, + "epoch": 2.2972972972972974, + "grad_norm": 3.390625, + "learning_rate": 0.00019830184059878606, + "loss": 0.881, "step": 255 }, { - "epoch": 0.9437386569872959, - "grad_norm": 3.03125, - "learning_rate": 0.0001890909090909091, - "loss": 0.9138, + "epoch": 2.3423423423423424, + "grad_norm": 2.75, + "learning_rate": 0.0001981040151396161, + "loss": 0.854, "step": 260 }, { - "epoch": 0.9618874773139746, - "grad_norm": 1.875, - "learning_rate": 0.00019272727272727274, - "loss": 0.9222, + "epoch": 2.3873873873873874, + "grad_norm": 1.25, + "learning_rate": 0.0001978954027238763, + "loss": 0.8856, "step": 265 }, { - "epoch": 0.9800362976406534, - "grad_norm": 3.84375, - "learning_rate": 0.00019636363636363636, - "loss": 0.9238, + "epoch": 2.4324324324324325, + "grad_norm": 1.640625, + "learning_rate": 0.00019767602628939507, + "loss": 0.8797, "step": 270 }, { - "epoch": 0.9981851179673321, - "grad_norm": 2.6875, - "learning_rate": 0.0002, - "loss": 0.9152, - "step": 275 - }, - { - "epoch": 0.9981851179673321, - "eval_loss": 2.195019245147705, - "eval_runtime": 1.0238, - "eval_samples_per_second": 4.884, - "eval_steps_per_second": 1.953, + "epoch": 2.4774774774774775, + "grad_norm": 2.28125, + "learning_rate": 0.00019744590995755088, + "loss": 0.8829, "step": 275 }, { - "epoch": 1.0163339382940109, - "grad_norm": 1.875, - "learning_rate": 0.00019999798600729064, - "loss": 0.8799, + "epoch": 2.5225225225225225, + "grad_norm": 1.265625, + "learning_rate": 0.00019720507903061974, + "loss": 0.8319, "step": 280 }, { - "epoch": 1.0344827586206897, - "grad_norm": 4.28125, - "learning_rate": 0.00019999194411028594, - "loss": 0.87, + "epoch": 2.5675675675675675, + "grad_norm": 1.6328125, + "learning_rate": 0.00019695355998899345, + "loss": 0.8489, "step": 285 }, { - "epoch": 1.0526315789473684, - "grad_norm": 3.28125, - "learning_rate": 0.0001999818745523526, - "loss": 0.8781, + "epoch": 2.6126126126126126, + "grad_norm": 1.1328125, + "learning_rate": 0.00019669138048826766, + "loss": 0.8409, "step": 290 }, { - "epoch": 1.0707803992740472, - "grad_norm": 0.87109375, - "learning_rate": 0.00019996777773909093, - "loss": 0.8886, + "epoch": 2.6576576576576576, + "grad_norm": 1.1640625, + "learning_rate": 0.00019641856935620117, + "loss": 0.8454, "step": 295 }, { - "epoch": 1.0889292196007259, - "grad_norm": 4.375, - "learning_rate": 0.00019994965423831854, - "loss": 0.9042, + "epoch": 2.7027027027027026, + "grad_norm": 0.625, + "learning_rate": 0.00019613515658954624, + "loss": 0.8501, "step": 300 }, { - "epoch": 1.1070780399274047, - "grad_norm": 1.2265625, - "learning_rate": 0.00019992750478004738, - "loss": 0.8569, + "epoch": 2.7477477477477477, + "grad_norm": 1.0859375, + "learning_rate": 0.0001958411733507502, + "loss": 0.8223, "step": 305 }, { - "epoch": 1.1252268602540836, - "grad_norm": 1.828125, - "learning_rate": 0.0001999013302564544, - "loss": 0.8641, + "epoch": 2.7927927927927927, + "grad_norm": 0.875, + "learning_rate": 0.00019553665196452903, + "loss": 0.8396, "step": 310 }, { - "epoch": 1.1433756805807622, - "grad_norm": 1.03125, - "learning_rate": 0.00019987113172184563, - "loss": 0.8521, + "epoch": 2.8378378378378377, + "grad_norm": 1.328125, + "learning_rate": 0.0001952216259143132, + "loss": 0.836, "step": 315 }, { - "epoch": 1.161524500907441, - "grad_norm": 1.15625, - "learning_rate": 0.00019983691039261357, - "loss": 0.8658, + "epoch": 2.8828828828828827, + "grad_norm": 2.09375, + "learning_rate": 0.0001948961298385659, + "loss": 0.8453, "step": 320 }, { - "epoch": 1.1796733212341197, - "grad_norm": 12.5, - "learning_rate": 0.00019979866764718843, - "loss": 0.8702, + "epoch": 2.9279279279279278, + "grad_norm": 0.62109375, + "learning_rate": 0.00019456019952697462, + "loss": 0.824, "step": 325 }, { - "epoch": 1.1978221415607986, - "grad_norm": 1.8203125, - "learning_rate": 0.00019975640502598244, - "loss": 0.874, + "epoch": 2.972972972972973, + "grad_norm": 1.7578125, + "learning_rate": 0.0001942138719165156, + "loss": 0.8471, "step": 330 }, { - "epoch": 1.2159709618874772, - "grad_norm": 2.109375, - "learning_rate": 0.00019971012423132775, - "loss": 0.8552, + "epoch": 3.0, + "eval_loss": 2.168168544769287, + "eval_runtime": 1.0188, + "eval_samples_per_second": 4.908, + "eval_steps_per_second": 1.963, + "step": 333 + }, + { + "epoch": 3.018018018018018, + "grad_norm": 1.5078125, + "learning_rate": 0.00019385718508739262, + "loss": 0.8242, "step": 335 }, { - "epoch": 1.234119782214156, - "grad_norm": 4.5625, - "learning_rate": 0.00019965982712740808, - "loss": 0.8459, + "epoch": 3.063063063063063, + "grad_norm": 1.4375, + "learning_rate": 0.00019349017825885, + "loss": 0.7439, "step": 340 }, { - "epoch": 1.252268602540835, - "grad_norm": 2.046875, - "learning_rate": 0.0001996055157401834, - "loss": 0.8528, + "epoch": 3.108108108108108, + "grad_norm": 0.67578125, + "learning_rate": 0.00019311289178486018, + "loss": 0.7545, "step": 345 }, { - "epoch": 1.2704174228675136, - "grad_norm": 1.3203125, - "learning_rate": 0.00019954719225730847, - "loss": 0.8687, + "epoch": 3.153153153153153, + "grad_norm": 1.71875, + "learning_rate": 0.00019272536714968658, + "loss": 0.7531, "step": 350 }, { - "epoch": 1.2885662431941924, - "grad_norm": 1.3828125, - "learning_rate": 0.0001994848590280447, - "loss": 0.8539, + "epoch": 3.1981981981981984, + "grad_norm": 2.15625, + "learning_rate": 0.0001923276469633223, + "loss": 0.7731, "step": 355 }, { - "epoch": 1.306715063520871, - "grad_norm": 1.515625, - "learning_rate": 0.00019941851856316548, - "loss": 0.8335, + "epoch": 3.2432432432432434, + "grad_norm": 1.453125, + "learning_rate": 0.000191919774956805, + "loss": 0.7558, "step": 360 }, { - "epoch": 1.32486388384755, - "grad_norm": 12.75, - "learning_rate": 0.00019934817353485501, - "loss": 0.8529, + "epoch": 3.2882882882882885, + "grad_norm": 1.71875, + "learning_rate": 0.0001915017959774084, + "loss": 0.782, "step": 365 }, { - "epoch": 1.3430127041742286, - "grad_norm": 0.94140625, - "learning_rate": 0.00019927382677660088, - "loss": 0.861, + "epoch": 3.3333333333333335, + "grad_norm": 0.7265625, + "learning_rate": 0.00019107375598371112, + "loss": 0.7795, "step": 370 }, { - "epoch": 1.3611615245009074, - "grad_norm": 2.4375, - "learning_rate": 0.00019919548128307954, - "loss": 0.8304, + "epoch": 3.3783783783783785, + "grad_norm": 1.71875, + "learning_rate": 0.0001906357020405435, + "loss": 0.7638, "step": 375 }, { - "epoch": 1.3793103448275863, - "grad_norm": 0.93359375, - "learning_rate": 0.00019911314021003613, - "loss": 0.8179, + "epoch": 3.4234234234234235, + "grad_norm": 1.7109375, + "learning_rate": 0.00019018768231381238, + "loss": 0.7511, "step": 380 }, { - "epoch": 1.397459165154265, - "grad_norm": 1.5234375, - "learning_rate": 0.00019902680687415705, - "loss": 0.8295, + "epoch": 3.4684684684684686, + "grad_norm": 0.6015625, + "learning_rate": 0.0001897297460652053, + "loss": 0.7591, "step": 385 }, { - "epoch": 1.4156079854809438, - "grad_norm": 1.703125, - "learning_rate": 0.00019893648475293648, - "loss": 0.8607, + "epoch": 3.5135135135135136, + "grad_norm": 7.625, + "learning_rate": 0.0001892619436467738, + "loss": 0.7433, "step": 390 }, { - "epoch": 1.4337568058076224, - "grad_norm": 1.1015625, - "learning_rate": 0.00019884217748453623, - "loss": 0.8223, + "epoch": 3.5585585585585586, + "grad_norm": 1.171875, + "learning_rate": 0.00018878432649539696, + "loss": 0.7673, "step": 395 }, { - "epoch": 1.4519056261343013, - "grad_norm": 0.796875, - "learning_rate": 0.00019874388886763944, - "loss": 0.8603, + "epoch": 3.6036036036036037, + "grad_norm": 0.734375, + "learning_rate": 0.0001882969471271259, + "loss": 0.7633, "step": 400 }, { - "epoch": 1.47005444646098, - "grad_norm": 0.640625, - "learning_rate": 0.0001986416228612972, - "loss": 0.8393, + "epoch": 3.6486486486486487, + "grad_norm": 0.93359375, + "learning_rate": 0.00018779985913140924, + "loss": 0.7751, "step": 405 }, { - "epoch": 1.4882032667876588, - "grad_norm": 1.5390625, - "learning_rate": 0.00019853538358476932, - "loss": 0.8363, + "epoch": 3.6936936936936937, + "grad_norm": 0.9140625, + "learning_rate": 0.00018729311716520074, + "loss": 0.7628, "step": 410 }, { - "epoch": 1.5063520871143377, + "epoch": 3.7387387387387387, "grad_norm": 1.09375, - "learning_rate": 0.00019842517531735838, - "loss": 0.8403, + "learning_rate": 0.00018677677694694958, + "loss": 0.7553, "step": 415 }, { - "epoch": 1.5245009074410163, - "grad_norm": 1.296875, - "learning_rate": 0.00019831100249823733, - "loss": 0.8198, + "epoch": 3.7837837837837838, + "grad_norm": 0.7578125, + "learning_rate": 0.00018625089525047385, + "loss": 0.7636, "step": 420 }, { - "epoch": 1.542649727767695, - "grad_norm": 17.5, - "learning_rate": 0.00019819286972627066, - "loss": 0.8203, + "epoch": 3.828828828828829, + "grad_norm": 0.76953125, + "learning_rate": 0.00018571552989871806, + "loss": 0.7764, "step": 425 }, { - "epoch": 1.560798548094374, - "grad_norm": 4.3125, - "learning_rate": 0.00019807078175982924, - "loss": 0.8595, + "epoch": 3.873873873873874, + "grad_norm": 0.84765625, + "learning_rate": 0.00018517073975739514, + "loss": 0.7715, "step": 430 }, { - "epoch": 1.5789473684210527, - "grad_norm": 1.890625, - "learning_rate": 0.00019794474351659852, - "loss": 0.8302, + "epoch": 3.918918918918919, + "grad_norm": 0.80859375, + "learning_rate": 0.0001846165847285141, + "loss": 0.77, "step": 435 }, { - "epoch": 1.5970961887477313, - "grad_norm": 1.125, - "learning_rate": 0.00019781476007338058, - "loss": 0.8248, + "epoch": 3.963963963963964, + "grad_norm": 3.0, + "learning_rate": 0.0001840531257437934, + "loss": 0.7873, "step": 440 }, { - "epoch": 1.6152450090744102, - "grad_norm": 1.0234375, - "learning_rate": 0.00019768083666588953, - "loss": 0.8099, + "epoch": 4.0, + "eval_loss": 2.2036337852478027, + "eval_runtime": 1.0177, + "eval_samples_per_second": 4.913, + "eval_steps_per_second": 1.965, + "step": 444 + }, + { + "epoch": 4.009009009009009, + "grad_norm": 2.0625, + "learning_rate": 0.00018348042475796122, + "loss": 0.7489, "step": 445 }, { - "epoch": 1.633393829401089, - "grad_norm": 1.4921875, - "learning_rate": 0.00019754297868854073, - "loss": 0.8127, + "epoch": 4.054054054054054, + "grad_norm": 0.6484375, + "learning_rate": 0.00018289854474194347, + "loss": 0.6615, "step": 450 }, { - "epoch": 1.6515426497277677, - "grad_norm": 1.5078125, - "learning_rate": 0.00019740119169423337, - "loss": 0.8199, + "epoch": 4.099099099099099, + "grad_norm": 1.2421875, + "learning_rate": 0.00018230754967593963, + "loss": 0.6595, "step": 455 }, { - "epoch": 1.6696914700544465, - "grad_norm": 0.6640625, - "learning_rate": 0.00019725548139412692, - "loss": 0.8185, + "epoch": 4.1441441441441444, + "grad_norm": 0.7578125, + "learning_rate": 0.00018170750454238793, + "loss": 0.6718, "step": 460 }, { - "epoch": 1.6878402903811254, - "grad_norm": 0.71875, - "learning_rate": 0.00019710585365741103, - "loss": 0.8347, + "epoch": 4.1891891891891895, + "grad_norm": 0.89453125, + "learning_rate": 0.00018109847531882047, + "loss": 0.6528, "step": 465 }, { - "epoch": 1.705989110707804, - "grad_norm": 1.046875, - "learning_rate": 0.00019695231451106912, - "loss": 0.8217, + "epoch": 4.2342342342342345, + "grad_norm": 1.15625, + "learning_rate": 0.0001804805289706083, + "loss": 0.6505, "step": 470 }, { - "epoch": 1.7241379310344827, - "grad_norm": 0.87109375, - "learning_rate": 0.00019679487013963564, - "loss": 0.8093, + "epoch": 4.2792792792792795, + "grad_norm": 1.25, + "learning_rate": 0.0001798537334435986, + "loss": 0.6708, "step": 475 }, { - "epoch": 1.7422867513611615, - "grad_norm": 0.6953125, - "learning_rate": 0.00019663352688494684, - "loss": 0.8139, + "epoch": 4.324324324324325, + "grad_norm": 0.89453125, + "learning_rate": 0.0001792181576566437, + "loss": 0.6702, "step": 480 }, { - "epoch": 1.7604355716878404, - "grad_norm": 0.6796875, - "learning_rate": 0.0001964682912458856, - "loss": 0.8232, + "epoch": 4.36936936936937, + "grad_norm": 0.734375, + "learning_rate": 0.0001785738714940231, + "loss": 0.6755, "step": 485 }, { - "epoch": 1.778584392014519, - "grad_norm": 0.75, - "learning_rate": 0.00019629916987811926, - "loss": 0.8143, + "epoch": 4.414414414414415, + "grad_norm": 1.46875, + "learning_rate": 0.00017792094579775926, + "loss": 0.6825, "step": 490 }, { - "epoch": 1.7967332123411979, - "grad_norm": 1.046875, - "learning_rate": 0.0001961261695938319, - "loss": 0.8022, + "epoch": 4.45945945945946, + "grad_norm": 1.0390625, + "learning_rate": 0.00017725945235982856, + "loss": 0.6864, "step": 495 }, { - "epoch": 1.8148820326678767, - "grad_norm": 0.88671875, - "learning_rate": 0.00019594929736144976, - "loss": 0.8233, + "epoch": 4.504504504504505, + "grad_norm": 0.62890625, + "learning_rate": 0.0001765894639142671, + "loss": 0.6758, "step": 500 }, { - "epoch": 1.8330308529945554, + "epoch": 4.54954954954955, "grad_norm": 0.71875, - "learning_rate": 0.00019576856030536054, - "loss": 0.8185, + "learning_rate": 0.0001759110541291736, + "loss": 0.6823, "step": 505 }, { - "epoch": 1.851179673321234, - "grad_norm": 1.0234375, - "learning_rate": 0.0001955839657056265, - "loss": 0.8208, + "epoch": 4.594594594594595, + "grad_norm": 0.6875, + "learning_rate": 0.000175224297598609, + "loss": 0.6882, "step": 510 }, { - "epoch": 1.8693284936479129, - "grad_norm": 1.546875, - "learning_rate": 0.00019539552099769126, - "loss": 0.8178, + "epoch": 4.63963963963964, + "grad_norm": 0.65234375, + "learning_rate": 0.00017452926983439475, + "loss": 0.6922, "step": 515 }, { - "epoch": 1.8874773139745917, - "grad_norm": 2.5625, - "learning_rate": 0.00019520323377208017, - "loss": 0.8086, + "epoch": 4.684684684684685, + "grad_norm": 0.90234375, + "learning_rate": 0.0001738260472578097, + "loss": 0.6787, "step": 520 }, { - "epoch": 1.9056261343012704, - "grad_norm": 0.9140625, - "learning_rate": 0.00019500711177409454, - "loss": 0.8057, + "epoch": 4.72972972972973, + "grad_norm": 0.84765625, + "learning_rate": 0.00017311470719118765, + "loss": 0.7028, "step": 525 }, { - "epoch": 1.9237749546279492, - "grad_norm": 2.15625, - "learning_rate": 0.00019480716290349995, - "loss": 0.8139, + "epoch": 4.774774774774775, + "grad_norm": 0.80859375, + "learning_rate": 0.00017239532784941495, + "loss": 0.6851, "step": 530 }, { - "epoch": 1.941923774954628, - "grad_norm": 0.86328125, - "learning_rate": 0.00019460339521420772, - "loss": 0.8028, + "epoch": 4.81981981981982, + "grad_norm": 1.15625, + "learning_rate": 0.0001716679883313308, + "loss": 0.6756, "step": 535 }, { - "epoch": 1.9600725952813067, - "grad_norm": 1.5546875, - "learning_rate": 0.00019439581691395067, - "loss": 0.8212, + "epoch": 4.864864864864865, + "grad_norm": 1.34375, + "learning_rate": 0.00017093276861102988, + "loss": 0.6697, "step": 540 }, { - "epoch": 1.9782214156079854, - "grad_norm": 0.984375, - "learning_rate": 0.00019418443636395248, - "loss": 0.8107, + "epoch": 4.90990990990991, + "grad_norm": 1.4296875, + "learning_rate": 0.00017018974952906884, + "loss": 0.689, "step": 545 }, { - "epoch": 1.9963702359346642, - "grad_norm": 0.546875, - "learning_rate": 0.00019396926207859084, - "loss": 0.8104, + "epoch": 4.954954954954955, + "grad_norm": 0.68359375, + "learning_rate": 0.0001694390127835775, + "loss": 0.6894, "step": 550 }, { - "epoch": 2.0, - "eval_loss": 2.1405208110809326, - "eval_runtime": 1.0236, - "eval_samples_per_second": 4.885, - "eval_steps_per_second": 1.954, - "step": 551 + "epoch": 5.0, + "grad_norm": 1.390625, + "learning_rate": 0.00016868064092127587, + "loss": 0.685, + "step": 555 }, { - "epoch": 2.014519056261343, - "grad_norm": 0.85546875, - "learning_rate": 0.00019375030272505463, - "loss": 0.7501, + "epoch": 5.0, + "eval_loss": 2.297661304473877, + "eval_runtime": 1.0166, + "eval_samples_per_second": 4.918, + "eval_steps_per_second": 1.967, "step": 555 }, { - "epoch": 2.0326678765880217, - "grad_norm": 0.69921875, - "learning_rate": 0.00019352756712299468, - "loss": 0.7576, + "epoch": 5.045045045045045, + "grad_norm": 1.328125, + "learning_rate": 0.00016791471732839776, + "loss": 0.5743, "step": 560 }, { - "epoch": 2.0508166969147004, - "grad_norm": 1.8828125, - "learning_rate": 0.00019330106424416852, - "loss": 0.7446, + "epoch": 5.09009009009009, + "grad_norm": 0.7734375, + "learning_rate": 0.00016714132622152206, + "loss": 0.5415, "step": 565 }, { - "epoch": 2.0689655172413794, - "grad_norm": 5.9375, - "learning_rate": 0.00019307080321207912, - "loss": 0.7617, + "epoch": 5.135135135135135, + "grad_norm": 0.78515625, + "learning_rate": 0.00016636055263831285, + "loss": 0.5752, "step": 570 }, { - "epoch": 2.087114337568058, - "grad_norm": 0.734375, - "learning_rate": 0.00019283679330160726, - "loss": 0.7408, + "epoch": 5.18018018018018, + "grad_norm": 0.91015625, + "learning_rate": 0.00016557248242816903, + "loss": 0.5778, "step": 575 }, { - "epoch": 2.1052631578947367, - "grad_norm": 1.2734375, - "learning_rate": 0.00019259904393863802, - "loss": 0.7426, + "epoch": 5.225225225225225, + "grad_norm": 1.015625, + "learning_rate": 0.00016477720224278492, + "loss": 0.5924, "step": 580 }, { - "epoch": 2.123411978221416, - "grad_norm": 1.25, - "learning_rate": 0.0001923575646996811, - "loss": 0.7396, + "epoch": 5.27027027027027, + "grad_norm": 0.8046875, + "learning_rate": 0.00016397479952662248, + "loss": 0.5683, "step": 585 }, { - "epoch": 2.1415607985480944, - "grad_norm": 2.03125, - "learning_rate": 0.000192112365311485, - "loss": 0.7581, + "epoch": 5.315315315315315, + "grad_norm": 0.734375, + "learning_rate": 0.0001631653625072965, + "loss": 0.595, "step": 590 }, { - "epoch": 2.159709618874773, - "grad_norm": 0.921875, - "learning_rate": 0.00019186345565064535, - "loss": 0.7494, + "epoch": 5.36036036036036, + "grad_norm": 0.92578125, + "learning_rate": 0.00016234898018587337, + "loss": 0.5862, "step": 595 }, { - "epoch": 2.1778584392014517, - "grad_norm": 0.73046875, - "learning_rate": 0.00019161084574320696, - "loss": 0.7356, + "epoch": 5.405405405405405, + "grad_norm": 1.3671875, + "learning_rate": 0.00016152574232708534, + "loss": 0.5806, "step": 600 }, { - "epoch": 2.196007259528131, - "grad_norm": 1.3125, - "learning_rate": 0.0001913545457642601, - "loss": 0.7502, + "epoch": 5.45045045045045, + "grad_norm": 1.578125, + "learning_rate": 0.00016069573944946025, + "loss": 0.6002, "step": 605 }, { - "epoch": 2.2141560798548094, - "grad_norm": 2.078125, - "learning_rate": 0.0001910945660375305, - "loss": 0.7566, + "epoch": 5.495495495495495, + "grad_norm": 1.1875, + "learning_rate": 0.00015985906281536875, + "loss": 0.6061, "step": 610 }, { - "epoch": 2.232304900181488, - "grad_norm": 1.109375, - "learning_rate": 0.0001908309170349637, - "loss": 0.7429, + "epoch": 5.54054054054054, + "grad_norm": 1.9765625, + "learning_rate": 0.00015901580442098968, + "loss": 0.6029, "step": 615 }, { - "epoch": 2.250453720508167, - "grad_norm": 0.87109375, - "learning_rate": 0.0001905636093763031, - "loss": 0.7449, + "epoch": 5.585585585585585, + "grad_norm": 0.78515625, + "learning_rate": 0.00015816605698619452, + "loss": 0.5985, "step": 620 }, { - "epoch": 2.268602540834846, - "grad_norm": 0.78515625, - "learning_rate": 0.00019029265382866214, - "loss": 0.735, + "epoch": 5.63063063063063, + "grad_norm": 0.73046875, + "learning_rate": 0.0001573099139443525, + "loss": 0.5863, "step": 625 }, { - "epoch": 2.2867513611615244, - "grad_norm": 1.5546875, - "learning_rate": 0.0001900180613060908, - "loss": 0.7585, + "epoch": 5.675675675675675, + "grad_norm": 0.84375, + "learning_rate": 0.00015644746943205734, + "loss": 0.5927, "step": 630 }, { - "epoch": 2.304900181488203, - "grad_norm": 2.015625, - "learning_rate": 0.00018973984286913584, - "loss": 0.7429, + "epoch": 5.7207207207207205, + "grad_norm": 0.796875, + "learning_rate": 0.00015557881827877633, + "loss": 0.6072, "step": 635 }, { - "epoch": 2.323049001814882, - "grad_norm": 0.7890625, - "learning_rate": 0.00018945800972439538, - "loss": 0.758, + "epoch": 5.7657657657657655, + "grad_norm": 0.76171875, + "learning_rate": 0.0001547040559964236, + "loss": 0.6032, "step": 640 }, { - "epoch": 2.341197822141561, - "grad_norm": 1.1015625, - "learning_rate": 0.00018917257322406734, - "loss": 0.757, + "epoch": 5.8108108108108105, + "grad_norm": 0.7109375, + "learning_rate": 0.00015382327876885797, + "loss": 0.6135, "step": 645 }, { - "epoch": 2.3593466424682394, - "grad_norm": 0.9296875, - "learning_rate": 0.00018888354486549237, - "loss": 0.7506, + "epoch": 5.8558558558558556, + "grad_norm": 0.7265625, + "learning_rate": 0.00015293658344130734, + "loss": 0.6058, "step": 650 }, { - "epoch": 2.3774954627949185, - "grad_norm": 1.0625, - "learning_rate": 0.00018859093629069058, - "loss": 0.756, + "epoch": 5.900900900900901, + "grad_norm": 0.69140625, + "learning_rate": 0.00015204406750972, + "loss": 0.5948, "step": 655 }, { - "epoch": 2.395644283121597, - "grad_norm": 0.609375, - "learning_rate": 0.00018829475928589271, - "loss": 0.7411, + "epoch": 5.945945945945946, + "grad_norm": 0.65234375, + "learning_rate": 0.00015114582911004466, + "loss": 0.6108, "step": 660 }, { - "epoch": 2.413793103448276, - "grad_norm": 0.58203125, - "learning_rate": 0.00018799502578106534, - "loss": 0.7426, + "epoch": 5.990990990990991, + "grad_norm": 0.79296875, + "learning_rate": 0.0001502419670074398, + "loss": 0.6223, "step": 665 }, { - "epoch": 2.4319419237749544, - "grad_norm": 0.8359375, - "learning_rate": 0.0001876917478494303, - "loss": 0.7359, + "epoch": 6.0, + "eval_loss": 2.4441299438476562, + "eval_runtime": 1.0183, + "eval_samples_per_second": 4.91, + "eval_steps_per_second": 1.964, + "step": 666 + }, + { + "epoch": 6.036036036036036, + "grad_norm": 1.109375, + "learning_rate": 0.00014933258058541415, + "loss": 0.5455, "step": 670 }, { - "epoch": 2.4500907441016335, - "grad_norm": 1.1640625, - "learning_rate": 0.00018738493770697852, - "loss": 0.7516, + "epoch": 6.081081081081081, + "grad_norm": 0.703125, + "learning_rate": 0.000148417769834899, + "loss": 0.5079, "step": 675 }, { - "epoch": 2.468239564428312, - "grad_norm": 0.82421875, - "learning_rate": 0.00018707460771197774, - "loss": 0.7524, + "epoch": 6.126126126126126, + "grad_norm": 0.9375, + "learning_rate": 0.0001474976353432539, + "loss": 0.4906, "step": 680 }, { - "epoch": 2.486388384754991, - "grad_norm": 1.0078125, - "learning_rate": 0.00018676077036447494, - "loss": 0.7655, + "epoch": 6.171171171171171, + "grad_norm": 0.96875, + "learning_rate": 0.00014657227828320635, + "loss": 0.5065, "step": 685 }, { - "epoch": 2.50453720508167, - "grad_norm": 0.546875, - "learning_rate": 0.0001864434383057927, - "loss": 0.7457, + "epoch": 6.216216216216216, + "grad_norm": 0.75, + "learning_rate": 0.0001456418004017278, + "loss": 0.4978, "step": 690 }, { - "epoch": 2.5226860254083485, - "grad_norm": 0.921875, - "learning_rate": 0.00018612262431802007, - "loss": 0.7608, + "epoch": 6.261261261261261, + "grad_norm": 0.75390625, + "learning_rate": 0.00014470630400884575, + "loss": 0.4966, "step": 695 }, { - "epoch": 2.540834845735027, - "grad_norm": 2.390625, - "learning_rate": 0.00018579834132349772, - "loss": 0.7437, + "epoch": 6.306306306306306, + "grad_norm": 0.7578125, + "learning_rate": 0.00014376589196639467, + "loss": 0.4981, "step": 700 }, { - "epoch": 2.558983666061706, - "grad_norm": 1.7890625, - "learning_rate": 0.00018547060238429736, - "loss": 0.7586, + "epoch": 6.351351351351352, + "grad_norm": 0.85546875, + "learning_rate": 0.00014282066767670572, + "loss": 0.5238, "step": 705 }, { - "epoch": 2.577132486388385, - "grad_norm": 3.3125, - "learning_rate": 0.0001851394207016957, - "loss": 0.752, + "epoch": 6.396396396396397, + "grad_norm": 0.66015625, + "learning_rate": 0.0001418707350712372, + "loss": 0.5052, "step": 710 }, { - "epoch": 2.5952813067150635, - "grad_norm": 0.58984375, - "learning_rate": 0.0001848048096156426, - "loss": 0.7484, + "epoch": 6.441441441441442, + "grad_norm": 0.87890625, + "learning_rate": 0.00014091619859914694, + "loss": 0.5045, "step": 715 }, { - "epoch": 2.613430127041742, - "grad_norm": 6.3125, - "learning_rate": 0.00018446678260422385, - "loss": 0.7554, + "epoch": 6.486486486486487, + "grad_norm": 0.8515625, + "learning_rate": 0.0001399571632158076, + "loss": 0.5169, "step": 720 }, { - "epoch": 2.6315789473684212, - "grad_norm": 0.62109375, - "learning_rate": 0.00018412535328311814, - "loss": 0.7442, + "epoch": 6.531531531531532, + "grad_norm": 0.703125, + "learning_rate": 0.00013899373437126637, + "loss": 0.5208, "step": 725 }, { - "epoch": 2.6497277676951, - "grad_norm": 0.8828125, - "learning_rate": 0.00018378053540504873, - "loss": 0.7581, + "epoch": 6.576576576576577, + "grad_norm": 0.66015625, + "learning_rate": 0.0001380260179986504, + "loss": 0.5192, "step": 730 }, { - "epoch": 2.6678765880217785, - "grad_norm": 1.7265625, - "learning_rate": 0.00018343234285922953, - "loss": 0.7727, + "epoch": 6.621621621621622, + "grad_norm": 0.77734375, + "learning_rate": 0.00013705412050251883, + "loss": 0.526, "step": 735 }, { - "epoch": 2.686025408348457, - "grad_norm": 0.58203125, - "learning_rate": 0.00018308078967080546, - "loss": 0.7525, + "epoch": 6.666666666666667, + "grad_norm": 0.84375, + "learning_rate": 0.00013607814874716325, + "loss": 0.541, "step": 740 }, { - "epoch": 2.7041742286751362, - "grad_norm": 1.0703125, - "learning_rate": 0.00018272589000028772, - "loss": 0.737, + "epoch": 6.711711711711712, + "grad_norm": 0.875, + "learning_rate": 0.0001350982100448575, + "loss": 0.5119, "step": 745 }, { - "epoch": 2.722323049001815, + "epoch": 6.756756756756757, "grad_norm": 0.7578125, - "learning_rate": 0.0001823676581429833, - "loss": 0.7724, + "learning_rate": 0.00013411441214405829, + "loss": 0.5192, "step": 750 }, { - "epoch": 2.7404718693284935, - "grad_norm": 1.640625, - "learning_rate": 0.00018200610852841913, - "loss": 0.746, + "epoch": 6.801801801801802, + "grad_norm": 1.0703125, + "learning_rate": 0.00013312686321755761, + "loss": 0.5353, "step": 755 }, { - "epoch": 2.7586206896551726, - "grad_norm": 0.7265625, - "learning_rate": 0.00018164125571976098, - "loss": 0.7609, + "epoch": 6.846846846846847, + "grad_norm": 1.0078125, + "learning_rate": 0.00013213567185058888, + "loss": 0.5286, "step": 760 }, { - "epoch": 2.7767695099818512, - "grad_norm": 1.1640625, - "learning_rate": 0.0001812731144132268, - "loss": 0.7623, + "epoch": 6.891891891891892, + "grad_norm": 0.88671875, + "learning_rate": 0.0001311409470288874, + "loss": 0.5286, "step": 765 }, { - "epoch": 2.79491833030853, - "grad_norm": 0.6328125, - "learning_rate": 0.00018090169943749476, - "loss": 0.7668, + "epoch": 6.936936936936937, + "grad_norm": 0.796875, + "learning_rate": 0.00013014279812670703, + "loss": 0.5353, "step": 770 }, { - "epoch": 2.8130671506352085, - "grad_norm": 0.546875, - "learning_rate": 0.00018052702575310588, - "loss": 0.747, + "epoch": 6.981981981981982, + "grad_norm": 0.6796875, + "learning_rate": 0.00012914133489479385, + "loss": 0.5378, "step": 775 }, { - "epoch": 2.8312159709618876, - "grad_norm": 1.09375, - "learning_rate": 0.00018014910845186153, - "loss": 0.7533, + "epoch": 7.0, + "eval_loss": 2.6714916229248047, + "eval_runtime": 1.0156, + "eval_samples_per_second": 4.923, + "eval_steps_per_second": 1.969, + "step": 777 + }, + { + "epoch": 7.027027027027027, + "grad_norm": 0.74609375, + "learning_rate": 0.0001281366674483187, + "loss": 0.4628, "step": 780 }, { - "epoch": 2.8493647912885662, - "grad_norm": 0.83984375, - "learning_rate": 0.00017976796275621555, - "loss": 0.7435, + "epoch": 7.072072072072072, + "grad_norm": 0.75, + "learning_rate": 0.00012712890625476953, + "loss": 0.4205, "step": 785 }, { - "epoch": 2.867513611615245, - "grad_norm": 0.73046875, - "learning_rate": 0.00017938360401866093, - "loss": 0.7424, + "epoch": 7.117117117117117, + "grad_norm": 0.84765625, + "learning_rate": 0.0001261181621218051, + "loss": 0.4286, "step": 790 }, { - "epoch": 2.885662431941924, - "grad_norm": 0.75, - "learning_rate": 0.00017899604772111163, - "loss": 0.7321, + "epoch": 7.162162162162162, + "grad_norm": 0.8515625, + "learning_rate": 0.00012510454618507106, + "loss": 0.4393, "step": 795 }, { - "epoch": 2.9038112522686026, - "grad_norm": 0.703125, - "learning_rate": 0.00017860530947427875, - "loss": 0.7474, + "epoch": 7.207207207207207, + "grad_norm": 0.765625, + "learning_rate": 0.0001240881698959802, + "loss": 0.4276, "step": 800 }, { - "epoch": 2.9219600725952812, - "grad_norm": 0.63671875, - "learning_rate": 0.00017821140501704194, - "loss": 0.7502, + "epoch": 7.252252252252252, + "grad_norm": 0.72265625, + "learning_rate": 0.00012306914500945788, + "loss": 0.4322, "step": 805 }, { - "epoch": 2.94010889292196, - "grad_norm": 1.1328125, - "learning_rate": 0.00017781435021581527, - "loss": 0.7432, + "epoch": 7.297297297297297, + "grad_norm": 0.87890625, + "learning_rate": 0.00012204758357165409, + "loss": 0.442, "step": 810 }, { - "epoch": 2.958257713248639, - "grad_norm": 1.046875, - "learning_rate": 0.00017741416106390826, - "loss": 0.7529, + "epoch": 7.342342342342342, + "grad_norm": 0.7578125, + "learning_rate": 0.00012102359790762347, + "loss": 0.4401, "step": 815 }, { - "epoch": 2.9764065335753176, - "grad_norm": 0.921875, - "learning_rate": 0.00017701085368088156, - "loss": 0.7633, + "epoch": 7.387387387387387, + "grad_norm": 0.73828125, + "learning_rate": 0.00011999730060897476, + "loss": 0.435, "step": 820 }, { - "epoch": 2.9945553539019962, - "grad_norm": 0.5859375, - "learning_rate": 0.0001766044443118978, - "loss": 0.7914, + "epoch": 7.4324324324324325, + "grad_norm": 0.8671875, + "learning_rate": 0.00011896880452149077, + "loss": 0.4448, "step": 825 }, { - "epoch": 2.9981851179673322, - "eval_loss": 2.1592297554016113, - "eval_runtime": 1.0245, - "eval_samples_per_second": 4.881, - "eval_steps_per_second": 1.952, - "step": 826 - }, - { - "epoch": 3.0127041742286753, - "grad_norm": 0.53515625, - "learning_rate": 0.0001761949493270671, - "loss": 0.6962, + "epoch": 7.4774774774774775, + "grad_norm": 0.8203125, + "learning_rate": 0.00011793822273272066, + "loss": 0.4448, "step": 830 }, { - "epoch": 3.030852994555354, - "grad_norm": 0.61328125, - "learning_rate": 0.0001757823852207877, - "loss": 0.6788, + "epoch": 7.5225225225225225, + "grad_norm": 0.78125, + "learning_rate": 0.00011690566855954523, + "loss": 0.4414, "step": 835 }, { - "epoch": 3.0490018148820326, - "grad_norm": 0.6953125, - "learning_rate": 0.00017536676861108164, - "loss": 0.6785, + "epoch": 7.5675675675675675, + "grad_norm": 0.8046875, + "learning_rate": 0.0001158712555357175, + "loss": 0.4461, "step": 840 }, { - "epoch": 3.0671506352087112, - "grad_norm": 0.796875, - "learning_rate": 0.0001749481162389254, - "loss": 0.6784, + "epoch": 7.612612612612613, + "grad_norm": 0.78515625, + "learning_rate": 0.00011483509739937903, + "loss": 0.4455, "step": 845 }, { - "epoch": 3.0852994555353903, - "grad_norm": 1.3515625, - "learning_rate": 0.0001745264449675755, - "loss": 0.6796, + "epoch": 7.657657657657658, + "grad_norm": 0.84765625, + "learning_rate": 0.000113797308080554, + "loss": 0.4521, "step": 850 }, { - "epoch": 3.103448275862069, - "grad_norm": 1.7109375, - "learning_rate": 0.00017410177178188918, - "loss": 0.6682, + "epoch": 7.702702702702703, + "grad_norm": 0.81640625, + "learning_rate": 0.00011275800168862214, + "loss": 0.4446, "step": 855 }, { - "epoch": 3.1215970961887476, - "grad_norm": 0.76953125, - "learning_rate": 0.0001736741137876405, - "loss": 0.6941, + "epoch": 7.747747747747748, + "grad_norm": 0.76171875, + "learning_rate": 0.00011171729249977169, + "loss": 0.4333, "step": 860 }, { - "epoch": 3.1397459165154267, - "grad_norm": 1.1015625, - "learning_rate": 0.0001732434882108311, - "loss": 0.6757, + "epoch": 7.792792792792793, + "grad_norm": 0.7734375, + "learning_rate": 0.00011067529494443463, + "loss": 0.4494, "step": 865 }, { - "epoch": 3.1578947368421053, - "grad_norm": 1.0859375, - "learning_rate": 0.00017280991239699642, - "loss": 0.6772, + "epoch": 7.837837837837838, + "grad_norm": 1.1796875, + "learning_rate": 0.00010963212359470421, + "loss": 0.4423, "step": 870 }, { - "epoch": 3.176043557168784, - "grad_norm": 0.79296875, - "learning_rate": 0.00017237340381050703, - "loss": 0.6951, + "epoch": 7.882882882882883, + "grad_norm": 1.15625, + "learning_rate": 0.00010858789315173745, + "loss": 0.4424, "step": 875 }, { - "epoch": 3.1941923774954626, - "grad_norm": 1.390625, - "learning_rate": 0.0001719339800338651, - "loss": 0.7005, + "epoch": 7.927927927927928, + "grad_norm": 0.76953125, + "learning_rate": 0.00010754271843314325, + "loss": 0.4491, "step": 880 }, { - "epoch": 3.2123411978221417, - "grad_norm": 1.46875, - "learning_rate": 0.00017149165876699635, - "loss": 0.6948, + "epoch": 7.972972972972973, + "grad_norm": 0.9296875, + "learning_rate": 0.00010649671436035753, + "loss": 0.458, "step": 885 }, { - "epoch": 3.2304900181488203, - "grad_norm": 0.80078125, - "learning_rate": 0.0001710464578265369, - "loss": 0.6809, + "epoch": 8.0, + "eval_loss": 2.955517292022705, + "eval_runtime": 1.015, + "eval_samples_per_second": 4.926, + "eval_steps_per_second": 1.97, + "step": 888 + }, + { + "epoch": 8.018018018018019, + "grad_norm": 0.90625, + "learning_rate": 0.0001054499959460074, + "loss": 0.4181, "step": 890 }, { - "epoch": 3.248638838475499, - "grad_norm": 0.79296875, - "learning_rate": 0.00017059839514511565, - "loss": 0.6848, + "epoch": 8.063063063063064, + "grad_norm": 0.87109375, + "learning_rate": 0.00010440267828126478, + "loss": 0.3664, "step": 895 }, { - "epoch": 3.266787658802178, - "grad_norm": 0.66015625, - "learning_rate": 0.00017014748877063214, - "loss": 0.6695, + "epoch": 8.108108108108109, + "grad_norm": 1.0546875, + "learning_rate": 0.00010335487652319182, + "loss": 0.3723, "step": 900 }, { - "epoch": 3.2849364791288567, - "grad_norm": 0.5859375, - "learning_rate": 0.00016969375686552937, - "loss": 0.6939, + "epoch": 8.153153153153154, + "grad_norm": 0.734375, + "learning_rate": 0.00010230670588207873, + "loss": 0.3588, "step": 905 }, { - "epoch": 3.3030852994555353, - "grad_norm": 0.83984375, - "learning_rate": 0.00016923721770606228, - "loss": 0.6792, + "epoch": 8.198198198198199, + "grad_norm": 0.859375, + "learning_rate": 0.00010125828160877604, + "loss": 0.3619, "step": 910 }, { - "epoch": 3.321234119782214, - "grad_norm": 0.81640625, - "learning_rate": 0.0001687778896815617, - "loss": 0.7015, + "epoch": 8.243243243243244, + "grad_norm": 0.765625, + "learning_rate": 0.00010020971898202218, + "loss": 0.3651, "step": 915 }, { - "epoch": 3.339382940108893, - "grad_norm": 0.64453125, - "learning_rate": 0.00016831579129369346, - "loss": 0.6828, + "epoch": 8.288288288288289, + "grad_norm": 0.90625, + "learning_rate": 9.916113329576818e-05, + "loss": 0.3742, "step": 920 }, { - "epoch": 3.3575317604355717, - "grad_norm": 0.6015625, - "learning_rate": 0.00016785094115571322, - "loss": 0.6704, + "epoch": 8.333333333333334, + "grad_norm": 0.8359375, + "learning_rate": 9.811263984650056e-05, + "loss": 0.3752, "step": 925 }, { - "epoch": 3.3756805807622503, - "grad_norm": 2.15625, - "learning_rate": 0.00016738335799171682, - "loss": 0.6796, + "epoch": 8.378378378378379, + "grad_norm": 1.015625, + "learning_rate": 9.706435392056394e-05, + "loss": 0.3746, "step": 930 }, { - "epoch": 3.3938294010889294, - "grad_norm": 0.671875, - "learning_rate": 0.00016691306063588583, - "loss": 0.6866, + "epoch": 8.423423423423424, + "grad_norm": 0.8046875, + "learning_rate": 9.601639078148489e-05, + "loss": 0.3744, "step": 935 }, { - "epoch": 3.411978221415608, - "grad_norm": 1.609375, - "learning_rate": 0.00016644006803172924, - "loss": 0.6788, + "epoch": 8.468468468468469, + "grad_norm": 0.82421875, + "learning_rate": 9.496886565729823e-05, + "loss": 0.3765, "step": 940 }, { - "epoch": 3.4301270417422867, - "grad_norm": 0.6484375, - "learning_rate": 0.00016596439923132017, - "loss": 0.7052, + "epoch": 8.513513513513514, + "grad_norm": 0.73828125, + "learning_rate": 9.392189372787706e-05, + "loss": 0.371, "step": 945 }, { - "epoch": 3.4482758620689653, - "grad_norm": 0.859375, - "learning_rate": 0.00016548607339452853, - "loss": 0.672, + "epoch": 8.558558558558559, + "grad_norm": 0.796875, + "learning_rate": 9.287559011226843e-05, + "loss": 0.3747, "step": 950 }, { - "epoch": 3.4664246823956444, - "grad_norm": 0.56640625, - "learning_rate": 0.00016500510978824926, - "loss": 0.6981, + "epoch": 8.603603603603604, + "grad_norm": 0.80859375, + "learning_rate": 9.183006985603545e-05, + "loss": 0.372, "step": 955 }, { - "epoch": 3.484573502722323, - "grad_norm": 0.609375, - "learning_rate": 0.0001645215277856263, - "loss": 0.6958, + "epoch": 8.64864864864865, + "grad_norm": 0.7578125, + "learning_rate": 9.078544791860739e-05, + "loss": 0.3764, "step": 960 }, { - "epoch": 3.5027223230490017, - "grad_norm": 0.7890625, - "learning_rate": 0.00016403534686527225, - "loss": 0.6779, + "epoch": 8.693693693693694, + "grad_norm": 0.81640625, + "learning_rate": 8.974183916063968e-05, + "loss": 0.375, "step": 965 }, { - "epoch": 3.5208711433756807, - "grad_norm": 0.5703125, - "learning_rate": 0.00016354658661048364, - "loss": 0.6773, + "epoch": 8.73873873873874, + "grad_norm": 0.75, + "learning_rate": 8.869935833138414e-05, + "loss": 0.3836, "step": 970 }, { - "epoch": 3.5390199637023594, - "grad_norm": 0.703125, - "learning_rate": 0.00016305526670845226, - "loss": 0.6883, + "epoch": 8.783783783783784, + "grad_norm": 0.75390625, + "learning_rate": 8.765812005607217e-05, + "loss": 0.3804, "step": 975 }, { - "epoch": 3.557168784029038, - "grad_norm": 0.765625, - "learning_rate": 0.00016256140694947217, - "loss": 0.6962, + "epoch": 8.82882882882883, + "grad_norm": 0.78515625, + "learning_rate": 8.6618238823311e-05, + "loss": 0.3857, "step": 980 }, { - "epoch": 3.5753176043557167, - "grad_norm": 0.703125, - "learning_rate": 0.00016206502722614238, - "loss": 0.6783, + "epoch": 8.873873873873874, + "grad_norm": 0.81640625, + "learning_rate": 8.557982897249513e-05, + "loss": 0.383, "step": 985 }, { - "epoch": 3.5934664246823957, - "grad_norm": 0.53515625, - "learning_rate": 0.0001615661475325658, - "loss": 0.6933, + "epoch": 8.91891891891892, + "grad_norm": 0.8671875, + "learning_rate": 8.45430046812344e-05, + "loss": 0.3831, "step": 990 }, { - "epoch": 3.6116152450090744, - "grad_norm": 0.67578125, - "learning_rate": 0.00016106478796354382, - "loss": 0.6791, + "epoch": 8.963963963963964, + "grad_norm": 0.83203125, + "learning_rate": 8.35078799527996e-05, + "loss": 0.3843, "step": 995 }, { - "epoch": 3.629764065335753, - "grad_norm": 0.859375, - "learning_rate": 0.00016056096871376667, - "loss": 0.7038, + "epoch": 9.0, + "eval_loss": 3.43645977973938, + "eval_runtime": 1.0159, + "eval_samples_per_second": 4.922, + "eval_steps_per_second": 1.969, + "step": 999 + }, + { + "epoch": 9.00900900900901, + "grad_norm": 0.734375, + "learning_rate": 8.247456860358725e-05, + "loss": 0.3691, "step": 1000 }, { - "epoch": 3.647912885662432, - "grad_norm": 0.6796875, - "learning_rate": 0.00016005471007700031, - "loss": 0.6966, + "epoch": 9.054054054054054, + "grad_norm": 1.0546875, + "learning_rate": 8.144318425060509e-05, + "loss": 0.32, "step": 1005 }, { - "epoch": 3.6660617059891107, - "grad_norm": 1.1875, - "learning_rate": 0.0001595460324452688, - "loss": 0.6856, + "epoch": 9.0990990990991, + "grad_norm": 0.7890625, + "learning_rate": 8.04138402989794e-05, + "loss": 0.3192, "step": 1010 }, { - "epoch": 3.6842105263157894, - "grad_norm": 0.9609375, - "learning_rate": 0.000159034956308033, - "loss": 0.6949, + "epoch": 9.144144144144144, + "grad_norm": 0.88671875, + "learning_rate": 7.938664992948549e-05, + "loss": 0.3166, "step": 1015 }, { - "epoch": 3.702359346642468, - "grad_norm": 0.6171875, - "learning_rate": 0.00015852150225136518, - "loss": 0.6962, + "epoch": 9.18918918918919, + "grad_norm": 0.72265625, + "learning_rate": 7.83617260861032e-05, + "loss": 0.3123, "step": 1020 }, { - "epoch": 3.720508166969147, - "grad_norm": 0.50390625, - "learning_rate": 0.00015800569095711982, - "loss": 0.6948, + "epoch": 9.234234234234235, + "grad_norm": 0.87890625, + "learning_rate": 7.733918146359815e-05, + "loss": 0.3192, "step": 1025 }, { - "epoch": 3.7386569872958257, - "grad_norm": 0.6328125, - "learning_rate": 0.00015748754320210072, - "loss": 0.695, + "epoch": 9.27927927927928, + "grad_norm": 0.8046875, + "learning_rate": 7.631912849513025e-05, + "loss": 0.3127, "step": 1030 }, { - "epoch": 3.7568058076225044, - "grad_norm": 0.71484375, - "learning_rate": 0.0001569670798572239, - "loss": 0.6778, + "epoch": 9.324324324324325, + "grad_norm": 0.765625, + "learning_rate": 7.530167933989161e-05, + "loss": 0.3149, "step": 1035 }, { - "epoch": 3.7749546279491835, - "grad_norm": 0.87109375, - "learning_rate": 0.00015644432188667695, - "loss": 0.6932, + "epoch": 9.36936936936937, + "grad_norm": 0.7890625, + "learning_rate": 7.428694587077371e-05, + "loss": 0.318, "step": 1040 }, { - "epoch": 3.793103448275862, - "grad_norm": 1.0703125, - "learning_rate": 0.0001559192903470747, - "loss": 0.7006, + "epoch": 9.414414414414415, + "grad_norm": 0.8125, + "learning_rate": 7.32750396620669e-05, + "loss": 0.3136, "step": 1045 }, { - "epoch": 3.8112522686025407, - "grad_norm": 0.796875, - "learning_rate": 0.00015539200638661104, - "loss": 0.6905, + "epoch": 9.45945945945946, + "grad_norm": 0.859375, + "learning_rate": 7.226607197719202e-05, + "loss": 0.3193, "step": 1050 }, { - "epoch": 3.8294010889292194, - "grad_norm": 0.6328125, - "learning_rate": 0.000154862491244207, - "loss": 0.684, + "epoch": 9.504504504504505, + "grad_norm": 0.75390625, + "learning_rate": 7.126015375646666e-05, + "loss": 0.319, "step": 1055 }, { - "epoch": 3.8475499092558985, - "grad_norm": 0.9765625, - "learning_rate": 0.00015433076624865531, - "loss": 0.6838, + "epoch": 9.54954954954955, + "grad_norm": 0.75, + "learning_rate": 7.025739560490675e-05, + "loss": 0.3246, "step": 1060 }, { - "epoch": 3.865698729582577, - "grad_norm": 0.8046875, - "learning_rate": 0.00015379685281776125, - "loss": 0.6807, + "epoch": 9.594594594594595, + "grad_norm": 0.79296875, + "learning_rate": 6.925790778006518e-05, + "loss": 0.3223, "step": 1065 }, { - "epoch": 3.8838475499092557, - "grad_norm": 0.59375, - "learning_rate": 0.00015326077245747999, - "loss": 0.6966, + "epoch": 9.63963963963964, + "grad_norm": 0.77734375, + "learning_rate": 6.826180017990828e-05, + "loss": 0.3238, "step": 1070 }, { - "epoch": 3.901996370235935, - "grad_norm": 0.6875, - "learning_rate": 0.00015272254676105025, - "loss": 0.6782, + "epoch": 9.684684684684685, + "grad_norm": 0.80078125, + "learning_rate": 6.726918233073231e-05, + "loss": 0.321, "step": 1075 }, { - "epoch": 3.9201451905626135, - "grad_norm": 0.8984375, - "learning_rate": 0.0001521821974081246, - "loss": 0.6908, + "epoch": 9.72972972972973, + "grad_norm": 0.8203125, + "learning_rate": 6.62801633751204e-05, + "loss": 0.3161, "step": 1080 }, { - "epoch": 3.938294010889292, - "grad_norm": 0.58203125, - "learning_rate": 0.0001516397461638962, - "loss": 0.7119, + "epoch": 9.774774774774775, + "grad_norm": 0.83203125, + "learning_rate": 6.52948520599419e-05, + "loss": 0.3221, "step": 1085 }, { - "epoch": 3.9564428312159707, - "grad_norm": 0.86328125, - "learning_rate": 0.00015109521487822206, - "loss": 0.6926, + "epoch": 9.81981981981982, + "grad_norm": 0.7734375, + "learning_rate": 6.431335672439531e-05, + "loss": 0.3251, "step": 1090 }, { - "epoch": 3.97459165154265, - "grad_norm": 0.62109375, - "learning_rate": 0.000150548625484743, - "loss": 0.6885, + "epoch": 9.864864864864865, + "grad_norm": 0.7890625, + "learning_rate": 6.333578528809574e-05, + "loss": 0.3246, "step": 1095 }, { - "epoch": 3.9927404718693285, - "grad_norm": 0.640625, - "learning_rate": 0.00015000000000000001, - "loss": 0.6978, + "epoch": 9.90990990990991, + "grad_norm": 0.79296875, + "learning_rate": 6.236224523920879e-05, + "loss": 0.3238, "step": 1100 }, { - "epoch": 4.0, - "eval_loss": 2.2176449298858643, - "eval_runtime": 1.0239, - "eval_samples_per_second": 4.883, - "eval_steps_per_second": 1.953, - "step": 1102 + "epoch": 9.954954954954955, + "grad_norm": 0.765625, + "learning_rate": 6.139284362263185e-05, + "loss": 0.3194, + "step": 1105 }, { - "epoch": 4.0108892921960075, - "grad_norm": 0.6015625, - "learning_rate": 0.0001494493605225477, - "loss": 0.6515, - "step": 1105 + "epoch": 10.0, + "grad_norm": 0.7109375, + "learning_rate": 6.042768702822381e-05, + "loss": 0.3241, + "step": 1110 }, { - "epoch": 4.029038112522686, - "grad_norm": 0.7421875, - "learning_rate": 0.0001488967292320639, - "loss": 0.6052, + "epoch": 10.0, + "eval_loss": 3.8823204040527344, + "eval_runtime": 1.0143, + "eval_samples_per_second": 4.929, + "eval_steps_per_second": 1.972, "step": 1110 }, { - "epoch": 4.047186932849365, - "grad_norm": 0.640625, - "learning_rate": 0.00014834212838845637, - "loss": 0.6075, + "epoch": 10.045045045045045, + "grad_norm": 0.7890625, + "learning_rate": 5.9466881579085275e-05, + "loss": 0.2776, "step": 1115 }, { - "epoch": 4.0653357531760435, - "grad_norm": 0.65234375, - "learning_rate": 0.00014778558033096633, - "loss": 0.6193, + "epoch": 10.09009009009009, + "grad_norm": 0.7890625, + "learning_rate": 5.8510532919889804e-05, + "loss": 0.2767, "step": 1120 }, { - "epoch": 4.083484573502722, - "grad_norm": 1.125, - "learning_rate": 0.0001472271074772683, - "loss": 0.6051, + "epoch": 10.135135135135135, + "grad_norm": 0.69140625, + "learning_rate": 5.7558746205267756e-05, + "loss": 0.2734, "step": 1125 }, { - "epoch": 4.101633393829401, - "grad_norm": 0.9765625, - "learning_rate": 0.00014666673232256738, - "loss": 0.6112, + "epoch": 10.18018018018018, + "grad_norm": 0.8203125, + "learning_rate": 5.6611626088244194e-05, + "loss": 0.2792, "step": 1130 }, { - "epoch": 4.11978221415608, - "grad_norm": 0.68359375, - "learning_rate": 0.00014610447743869314, - "loss": 0.6208, + "epoch": 10.225225225225225, + "grad_norm": 0.734375, + "learning_rate": 5.56692767087318e-05, + "loss": 0.2753, "step": 1135 }, { - "epoch": 4.137931034482759, - "grad_norm": 0.89453125, - "learning_rate": 0.00014554036547319033, - "loss": 0.6185, + "epoch": 10.27027027027027, + "grad_norm": 0.70703125, + "learning_rate": 5.4731801682080206e-05, + "loss": 0.2763, "step": 1140 }, { - "epoch": 4.1560798548094375, - "grad_norm": 1.09375, - "learning_rate": 0.0001449744191484066, - "loss": 0.6209, + "epoch": 10.315315315315315, + "grad_norm": 0.7265625, + "learning_rate": 5.3799304087683146e-05, + "loss": 0.2772, "step": 1145 }, { - "epoch": 4.174228675136116, - "grad_norm": 0.87109375, - "learning_rate": 0.00014440666126057744, - "loss": 0.6134, + "epoch": 10.36036036036036, + "grad_norm": 0.79296875, + "learning_rate": 5.28718864576442e-05, + "loss": 0.2763, "step": 1150 }, { - "epoch": 4.192377495462795, - "grad_norm": 0.76171875, - "learning_rate": 0.00014383711467890774, - "loss": 0.6234, + "epoch": 10.405405405405405, + "grad_norm": 0.75, + "learning_rate": 5.194965076550323e-05, + "loss": 0.2766, "step": 1155 }, { - "epoch": 4.2105263157894735, - "grad_norm": 0.640625, - "learning_rate": 0.00014326580234465085, - "loss": 0.6104, + "epoch": 10.45045045045045, + "grad_norm": 0.84375, + "learning_rate": 5.1032698415023963e-05, + "loss": 0.2802, "step": 1160 }, { - "epoch": 4.228675136116152, - "grad_norm": 0.9296875, - "learning_rate": 0.0001426927472701842, - "loss": 0.6183, + "epoch": 10.495495495495495, + "grad_norm": 0.8203125, + "learning_rate": 5.0121130229043786e-05, + "loss": 0.276, "step": 1165 }, { - "epoch": 4.246823956442832, - "grad_norm": 0.88671875, - "learning_rate": 0.00014211797253808268, - "loss": 0.6201, + "epoch": 10.54054054054054, + "grad_norm": 0.79296875, + "learning_rate": 4.921504643838839e-05, + "loss": 0.2801, "step": 1170 }, { - "epoch": 4.26497277676951, - "grad_norm": 0.89453125, - "learning_rate": 0.00014154150130018866, - "loss": 0.6291, + "epoch": 10.585585585585585, + "grad_norm": 0.7109375, + "learning_rate": 4.8314546670850594e-05, + "loss": 0.2775, "step": 1175 }, { - "epoch": 4.283121597096189, - "grad_norm": 1.046875, - "learning_rate": 0.00014096335677667954, - "loss": 0.6248, + "epoch": 10.63063063063063, + "grad_norm": 0.7421875, + "learning_rate": 4.741972994023601e-05, + "loss": 0.2806, "step": 1180 }, { - "epoch": 4.3012704174228675, - "grad_norm": 0.8984375, - "learning_rate": 0.00014038356225513248, - "loss": 0.6321, + "epoch": 10.675675675675675, + "grad_norm": 0.7421875, + "learning_rate": 4.6530694635476e-05, + "loss": 0.2775, "step": 1185 }, { - "epoch": 4.319419237749546, - "grad_norm": 0.61328125, - "learning_rate": 0.00013980214108958624, - "loss": 0.6284, + "epoch": 10.72072072072072, + "grad_norm": 0.7421875, + "learning_rate": 4.5647538509809416e-05, + "loss": 0.2802, "step": 1190 }, { - "epoch": 4.337568058076225, - "grad_norm": 0.79296875, - "learning_rate": 0.00013921911669960055, - "loss": 0.6294, + "epoch": 10.765765765765765, + "grad_norm": 0.73828125, + "learning_rate": 4.477035867003405e-05, + "loss": 0.2751, "step": 1195 }, { - "epoch": 4.3557168784029034, - "grad_norm": 0.80859375, - "learning_rate": 0.00013863451256931287, - "loss": 0.6166, + "epoch": 10.81081081081081, + "grad_norm": 0.7734375, + "learning_rate": 4.38992515658297e-05, + "loss": 0.2802, "step": 1200 }, { - "epoch": 4.373865698729583, - "grad_norm": 0.60546875, - "learning_rate": 0.0001380483522464923, - "loss": 0.6254, + "epoch": 10.855855855855856, + "grad_norm": 0.79296875, + "learning_rate": 4.303431297915277e-05, + "loss": 0.2748, "step": 1205 }, { - "epoch": 4.392014519056262, - "grad_norm": 0.875, - "learning_rate": 0.00013746065934159123, - "loss": 0.6194, + "epoch": 10.9009009009009, + "grad_norm": 0.77734375, + "learning_rate": 4.2175638013704655e-05, + "loss": 0.2808, "step": 1210 }, { - "epoch": 4.41016333938294, - "grad_norm": 0.69921875, - "learning_rate": 0.0001368714575267941, - "loss": 0.6287, + "epoch": 10.945945945945946, + "grad_norm": 0.7421875, + "learning_rate": 4.1323321084474896e-05, + "loss": 0.2795, "step": 1215 }, { - "epoch": 4.428312159709619, - "grad_norm": 0.55859375, - "learning_rate": 0.0001362807705350641, - "loss": 0.6306, + "epoch": 10.99099099099099, + "grad_norm": 0.76953125, + "learning_rate": 4.0477455907359715e-05, + "loss": 0.2825, "step": 1220 }, { - "epoch": 4.4464609800362975, - "grad_norm": 0.7421875, - "learning_rate": 0.00013568862215918717, - "loss": 0.6298, + "epoch": 11.0, + "eval_loss": 4.404350757598877, + "eval_runtime": 1.0163, + "eval_samples_per_second": 4.92, + "eval_steps_per_second": 1.968, + "step": 1221 + }, + { + "epoch": 11.036036036036036, + "grad_norm": 0.609375, + "learning_rate": 3.963813548885751e-05, + "loss": 0.2565, "step": 1225 }, { - "epoch": 4.464609800362976, - "grad_norm": 0.62109375, - "learning_rate": 0.00013509503625081358, - "loss": 0.6162, + "epoch": 11.08108108108108, + "grad_norm": 0.7109375, + "learning_rate": 3.880545211584251e-05, + "loss": 0.2458, "step": 1230 }, { - "epoch": 4.482758620689655, - "grad_norm": 0.6484375, - "learning_rate": 0.00013450003671949706, - "loss": 0.6235, + "epoch": 11.126126126126126, + "grad_norm": 0.6796875, + "learning_rate": 3.7979497345417234e-05, + "loss": 0.257, "step": 1235 }, { - "epoch": 4.500907441016334, - "grad_norm": 0.7578125, - "learning_rate": 0.00013390364753173206, - "loss": 0.6217, + "epoch": 11.17117117117117, + "grad_norm": 0.6640625, + "learning_rate": 3.716036199484574e-05, + "loss": 0.2449, "step": 1240 }, { - "epoch": 4.519056261343013, - "grad_norm": 0.69921875, - "learning_rate": 0.00013330589270998808, - "loss": 0.6119, + "epoch": 11.216216216216216, + "grad_norm": 0.83203125, + "learning_rate": 3.634813613156753e-05, + "loss": 0.2529, "step": 1245 }, { - "epoch": 4.537205081669692, - "grad_norm": 0.6484375, - "learning_rate": 0.00013270679633174218, - "loss": 0.6272, + "epoch": 11.26126126126126, + "grad_norm": 0.7421875, + "learning_rate": 3.554290906329438e-05, + "loss": 0.2519, "step": 1250 }, { - "epoch": 4.55535390199637, - "grad_norm": 0.60546875, - "learning_rate": 0.00013210638252850908, - "loss": 0.6244, + "epoch": 11.306306306306306, + "grad_norm": 0.734375, + "learning_rate": 3.474476932819062e-05, + "loss": 0.2551, "step": 1255 }, { - "epoch": 4.573502722323049, - "grad_norm": 0.609375, - "learning_rate": 0.0001315046754848693, - "loss": 0.6335, + "epoch": 11.35135135135135, + "grad_norm": 0.69140625, + "learning_rate": 3.395380468513805e-05, + "loss": 0.2492, "step": 1260 }, { - "epoch": 4.5916515426497275, - "grad_norm": 0.59375, - "learning_rate": 0.00013090169943749476, - "loss": 0.6282, + "epoch": 11.396396396396396, + "grad_norm": 0.703125, + "learning_rate": 3.317010210408617e-05, + "loss": 0.2501, "step": 1265 }, { - "epoch": 4.609800362976406, - "grad_norm": 0.59375, - "learning_rate": 0.00013029747867417276, - "loss": 0.628, + "epoch": 11.441441441441441, + "grad_norm": 0.734375, + "learning_rate": 3.2393747756489845e-05, + "loss": 0.2493, "step": 1270 }, { - "epoch": 4.627949183303086, - "grad_norm": 0.6171875, - "learning_rate": 0.0001296920375328275, - "loss": 0.6257, + "epoch": 11.486486486486486, + "grad_norm": 0.70703125, + "learning_rate": 3.162482700583402e-05, + "loss": 0.2475, "step": 1275 }, { - "epoch": 4.646098003629764, - "grad_norm": 0.57421875, - "learning_rate": 0.0001290854004005399, - "loss": 0.6229, + "epoch": 11.531531531531531, + "grad_norm": 0.7578125, + "learning_rate": 3.0863424398248064e-05, + "loss": 0.2524, "step": 1280 }, { - "epoch": 4.664246823956443, - "grad_norm": 0.640625, - "learning_rate": 0.00012847759171256523, - "loss": 0.6401, + "epoch": 11.576576576576576, + "grad_norm": 0.6953125, + "learning_rate": 3.0109623653209163e-05, + "loss": 0.2489, "step": 1285 }, { - "epoch": 4.682395644283122, - "grad_norm": 0.5859375, - "learning_rate": 0.0001278686359513488, - "loss": 0.627, + "epoch": 11.621621621621621, + "grad_norm": 0.72265625, + "learning_rate": 2.936350765433713e-05, + "loss": 0.2507, "step": 1290 }, { - "epoch": 4.7005444646098, - "grad_norm": 0.6171875, - "learning_rate": 0.0001272585576455398, - "loss": 0.6368, + "epoch": 11.666666666666666, + "grad_norm": 0.68359375, + "learning_rate": 2.862515844028111e-05, + "loss": 0.2478, "step": 1295 }, { - "epoch": 4.718693284936479, - "grad_norm": 0.65234375, - "learning_rate": 0.00012664738136900348, - "loss": 0.6378, + "epoch": 11.711711711711711, + "grad_norm": 0.7109375, + "learning_rate": 2.789465719569906e-05, + "loss": 0.2521, "step": 1300 }, { - "epoch": 4.7368421052631575, - "grad_norm": 0.609375, - "learning_rate": 0.0001260351317398312, - "loss": 0.6318, + "epoch": 11.756756756756756, + "grad_norm": 0.6875, + "learning_rate": 2.717208424233091e-05, + "loss": 0.2525, "step": 1305 }, { - "epoch": 4.754990925589837, - "grad_norm": 0.59375, - "learning_rate": 0.00012542183341934872, - "loss": 0.6277, + "epoch": 11.801801801801801, + "grad_norm": 0.7421875, + "learning_rate": 2.6457519030167143e-05, + "loss": 0.2566, "step": 1310 }, { - "epoch": 4.773139745916516, - "grad_norm": 0.59375, - "learning_rate": 0.0001248075111111229, - "loss": 0.6407, + "epoch": 11.846846846846846, + "grad_norm": 0.7109375, + "learning_rate": 2.57510401287128e-05, + "loss": 0.2495, "step": 1315 }, { - "epoch": 4.791288566243194, - "grad_norm": 0.73828125, - "learning_rate": 0.00012419218955996676, - "loss": 0.6421, + "epoch": 11.891891891891891, + "grad_norm": 0.6796875, + "learning_rate": 2.5052725218348394e-05, + "loss": 0.2526, "step": 1320 }, { - "epoch": 4.809437386569873, - "grad_norm": 0.80859375, - "learning_rate": 0.00012357589355094275, - "loss": 0.6323, + "epoch": 11.936936936936936, + "grad_norm": 0.71484375, + "learning_rate": 2.4362651081788667e-05, + "loss": 0.249, "step": 1325 }, { - "epoch": 4.827586206896552, - "grad_norm": 0.59765625, - "learning_rate": 0.0001229586479083641, - "loss": 0.6437, + "epoch": 11.981981981981981, + "grad_norm": 0.77734375, + "learning_rate": 2.3680893595640043e-05, + "loss": 0.2549, "step": 1330 }, { - "epoch": 4.84573502722323, - "grad_norm": 0.5625, - "learning_rate": 0.00012234047749479544, - "loss": 0.6415, + "epoch": 12.0, + "eval_loss": 4.838156700134277, + "eval_runtime": 1.0155, + "eval_samples_per_second": 4.923, + "eval_steps_per_second": 1.969, + "step": 1332 + }, + { + "epoch": 12.027027027027026, + "grad_norm": 0.625, + "learning_rate": 2.3007527722057488e-05, + "loss": 0.2452, "step": 1335 }, { - "epoch": 4.863883847549909, - "grad_norm": 0.59375, - "learning_rate": 0.00012172140721005079, - "loss": 0.6293, + "epoch": 12.072072072072071, + "grad_norm": 0.6484375, + "learning_rate": 2.234262750050241e-05, + "loss": 0.2436, "step": 1340 }, { - "epoch": 4.882032667876588, - "grad_norm": 0.609375, - "learning_rate": 0.000121101461990191, - "loss": 0.6245, + "epoch": 12.117117117117116, + "grad_norm": 0.6171875, + "learning_rate": 2.168626603960138e-05, + "loss": 0.2396, "step": 1345 }, { - "epoch": 4.900181488203267, - "grad_norm": 0.5703125, - "learning_rate": 0.00012048066680651908, - "loss": 0.6146, + "epoch": 12.162162162162161, + "grad_norm": 0.671875, + "learning_rate": 2.1038515509107736e-05, + "loss": 0.2344, "step": 1350 }, { - "epoch": 4.918330308529946, - "grad_norm": 0.65625, - "learning_rate": 0.00011985904666457455, - "loss": 0.6425, + "epoch": 12.207207207207206, + "grad_norm": 0.71875, + "learning_rate": 2.0399447131966132e-05, + "loss": 0.2359, "step": 1355 }, { - "epoch": 4.936479128856624, - "grad_norm": 0.80078125, - "learning_rate": 0.00011923662660312611, - "loss": 0.64, + "epoch": 12.252252252252251, + "grad_norm": 0.65234375, + "learning_rate": 1.976913117648128e-05, + "loss": 0.2374, "step": 1360 }, { - "epoch": 4.954627949183303, - "grad_norm": 0.54296875, - "learning_rate": 0.00011861343169316301, - "loss": 0.6316, + "epoch": 12.297297297297296, + "grad_norm": 0.6015625, + "learning_rate": 1.9147636948591653e-05, + "loss": 0.2406, "step": 1365 }, { - "epoch": 4.972776769509982, - "grad_norm": 0.76171875, - "learning_rate": 0.00011798948703688539, - "loss": 0.629, + "epoch": 12.342342342342342, + "grad_norm": 0.65234375, + "learning_rate": 1.8535032784249028e-05, + "loss": 0.2392, "step": 1370 }, { - "epoch": 4.99092558983666, - "grad_norm": 0.625, - "learning_rate": 0.00011736481776669306, - "loss": 0.6386, + "epoch": 12.387387387387387, + "grad_norm": 0.703125, + "learning_rate": 1.7931386041904506e-05, + "loss": 0.2347, "step": 1375 }, { - "epoch": 4.998185117967332, - "eval_loss": 2.3272275924682617, - "eval_runtime": 1.0255, - "eval_samples_per_second": 4.876, - "eval_steps_per_second": 1.95, - "step": 1377 - }, - { - "epoch": 5.00907441016334, - "grad_norm": 0.77734375, - "learning_rate": 0.00011673944904417308, - "loss": 0.5952, + "epoch": 12.432432432432432, + "grad_norm": 0.66796875, + "learning_rate": 1.733676309510245e-05, + "loss": 0.241, "step": 1380 }, { - "epoch": 5.027223230490018, - "grad_norm": 0.79296875, - "learning_rate": 0.00011611340605908642, - "loss": 0.5449, + "epoch": 12.477477477477478, + "grad_norm": 0.71875, + "learning_rate": 1.6751229325182195e-05, + "loss": 0.2392, "step": 1385 }, { - "epoch": 5.045372050816697, + "epoch": 12.522522522522522, "grad_norm": 0.65234375, - "learning_rate": 0.00011548671402835325, - "loss": 0.5476, + "learning_rate": 1.6174849114089064e-05, + "loss": 0.2389, "step": 1390 }, { - "epoch": 5.063520871143376, - "grad_norm": 0.72265625, - "learning_rate": 0.00011485939819503717, - "loss": 0.5626, + "epoch": 12.567567567567568, + "grad_norm": 0.703125, + "learning_rate": 1.5607685837295516e-05, + "loss": 0.2397, "step": 1395 }, { - "epoch": 5.081669691470054, - "grad_norm": 0.5703125, - "learning_rate": 0.00011423148382732853, - "loss": 0.5616, + "epoch": 12.612612612612612, + "grad_norm": 0.6640625, + "learning_rate": 1.5049801856832557e-05, + "loss": 0.2343, "step": 1400 }, { - "epoch": 5.099818511796733, - "grad_norm": 0.61328125, - "learning_rate": 0.00011360299621752644, - "loss": 0.569, + "epoch": 12.657657657657658, + "grad_norm": 0.67578125, + "learning_rate": 1.4501258514432836e-05, + "loss": 0.2364, "step": 1405 }, { - "epoch": 5.117967332123412, - "grad_norm": 0.6953125, - "learning_rate": 0.00011297396068102017, - "loss": 0.5644, + "epoch": 12.702702702702704, + "grad_norm": 0.69140625, + "learning_rate": 1.3962116124785863e-05, + "loss": 0.2416, "step": 1410 }, { - "epoch": 5.136116152450091, - "grad_norm": 0.64453125, - "learning_rate": 0.00011234440255526948, - "loss": 0.5577, + "epoch": 12.747747747747749, + "grad_norm": 0.6796875, + "learning_rate": 1.3432433968906044e-05, + "loss": 0.238, "step": 1415 }, { - "epoch": 5.15426497277677, - "grad_norm": 0.66015625, - "learning_rate": 0.00011171434719878384, - "loss": 0.5484, + "epoch": 12.792792792792794, + "grad_norm": 0.625, + "learning_rate": 1.2912270287614736e-05, + "loss": 0.2355, "step": 1420 }, { - "epoch": 5.172413793103448, - "grad_norm": 0.671875, - "learning_rate": 0.00011108381999010111, - "loss": 0.5599, + "epoch": 12.837837837837839, + "grad_norm": 0.65625, + "learning_rate": 1.240168227513614e-05, + "loss": 0.2388, "step": 1425 }, { - "epoch": 5.190562613430127, - "grad_norm": 0.62109375, - "learning_rate": 0.00011045284632676536, - "loss": 0.5588, + "epoch": 12.882882882882884, + "grad_norm": 0.6640625, + "learning_rate": 1.1900726072808666e-05, + "loss": 0.2351, "step": 1430 }, { - "epoch": 5.208711433756806, - "grad_norm": 0.625, - "learning_rate": 0.00010982145162430373, - "loss": 0.5562, + "epoch": 12.927927927927929, + "grad_norm": 0.65234375, + "learning_rate": 1.1409456762911985e-05, + "loss": 0.2393, "step": 1435 }, { - "epoch": 5.226860254083484, - "grad_norm": 0.59375, - "learning_rate": 0.00010918966131520277, - "loss": 0.549, + "epoch": 12.972972972972974, + "grad_norm": 0.65234375, + "learning_rate": 1.0927928362610462e-05, + "loss": 0.2408, "step": 1440 }, { - "epoch": 5.245009074410163, - "grad_norm": 0.640625, - "learning_rate": 0.00010855750084788398, - "loss": 0.5667, + "epoch": 13.0, + "eval_loss": 5.061110496520996, + "eval_runtime": 1.0149, + "eval_samples_per_second": 4.927, + "eval_steps_per_second": 1.971, + "step": 1443 + }, + { + "epoch": 13.018018018018019, + "grad_norm": 0.68359375, + "learning_rate": 1.0456193818013726e-05, + "loss": 0.2395, "step": 1445 }, { - "epoch": 5.2631578947368425, - "grad_norm": 0.64453125, - "learning_rate": 0.00010792499568567884, - "loss": 0.5611, + "epoch": 13.063063063063064, + "grad_norm": 0.66015625, + "learning_rate": 9.99430499835503e-06, + "loss": 0.2342, "step": 1450 }, { - "epoch": 5.281306715063521, - "grad_norm": 0.7421875, - "learning_rate": 0.0001072921713058031, - "loss": 0.5577, + "epoch": 13.108108108108109, + "grad_norm": 0.75390625, + "learning_rate": 9.542312690288036e-06, + "loss": 0.2342, "step": 1455 }, { - "epoch": 5.2994555353902, - "grad_norm": 0.65234375, - "learning_rate": 0.00010665905319833041, - "loss": 0.559, + "epoch": 13.153153153153154, + "grad_norm": 0.6015625, + "learning_rate": 9.100266592302542e-06, + "loss": 0.2332, "step": 1460 }, { - "epoch": 5.317604355716878, - "grad_norm": 0.6953125, - "learning_rate": 0.00010602566686516586, - "loss": 0.552, + "epoch": 13.198198198198199, + "grad_norm": 0.671875, + "learning_rate": 8.668215309259997e-06, + "loss": 0.2327, "step": 1465 }, { - "epoch": 5.335753176043557, - "grad_norm": 1.0546875, - "learning_rate": 0.00010539203781901861, - "loss": 0.5516, + "epoch": 13.243243243243244, + "grad_norm": 0.62890625, + "learning_rate": 8.246206347049079e-06, + "loss": 0.2353, "step": 1470 }, { - "epoch": 5.353901996370236, - "grad_norm": 0.88671875, - "learning_rate": 0.00010475819158237425, - "loss": 0.5646, + "epoch": 13.288288288288289, + "grad_norm": 0.66796875, + "learning_rate": 7.834286107362343e-06, + "loss": 0.2331, "step": 1475 }, { - "epoch": 5.372050816696914, - "grad_norm": 0.73046875, - "learning_rate": 0.00010412415368646673, - "loss": 0.5614, + "epoch": 13.333333333333334, + "grad_norm": 0.67578125, + "learning_rate": 7.432499882594091e-06, + "loss": 0.2323, "step": 1480 }, { - "epoch": 5.390199637023594, - "grad_norm": 0.75, - "learning_rate": 0.00010348994967025012, - "loss": 0.5532, + "epoch": 13.378378378378379, + "grad_norm": 0.68359375, + "learning_rate": 7.040891850860287e-06, + "loss": 0.2354, "step": 1485 }, { - "epoch": 5.4083484573502725, - "grad_norm": 0.91796875, - "learning_rate": 0.00010285560507936961, - "loss": 0.5631, + "epoch": 13.423423423423424, + "grad_norm": 0.6328125, + "learning_rate": 6.659505071140959e-06, + "loss": 0.2345, "step": 1490 }, { - "epoch": 5.426497277676951, - "grad_norm": 0.80078125, - "learning_rate": 0.00010222114546513295, - "loss": 0.581, + "epoch": 13.468468468468469, + "grad_norm": 0.62109375, + "learning_rate": 6.2883814785457575e-06, + "loss": 0.2376, "step": 1495 }, { - "epoch": 5.44464609800363, - "grad_norm": 0.61328125, - "learning_rate": 0.00010158659638348081, - "loss": 0.5503, + "epoch": 13.513513513513514, + "grad_norm": 0.64453125, + "learning_rate": 5.927561879702903e-06, + "loss": 0.2361, "step": 1500 }, { - "epoch": 5.462794918330308, - "grad_norm": 0.69921875, - "learning_rate": 0.00010095198339395769, - "loss": 0.5645, + "epoch": 13.558558558558559, + "grad_norm": 0.72265625, + "learning_rate": 5.577085948272431e-06, + "loss": 0.2331, "step": 1505 }, { - "epoch": 5.480943738656987, - "grad_norm": 0.65234375, - "learning_rate": 0.00010031733205868224, - "loss": 0.565, + "epoch": 13.603603603603604, + "grad_norm": 0.609375, + "learning_rate": 5.236992220583814e-06, + "loss": 0.2312, "step": 1510 }, { - "epoch": 5.499092558983666, - "grad_norm": 0.64453125, - "learning_rate": 9.968266794131777e-05, - "loss": 0.5616, + "epoch": 13.64864864864865, + "grad_norm": 0.6953125, + "learning_rate": 4.90731809139876e-06, + "loss": 0.2347, "step": 1515 }, { - "epoch": 5.517241379310345, - "grad_norm": 0.640625, - "learning_rate": 9.904801660604234e-05, - "loss": 0.5601, + "epoch": 13.693693693693694, + "grad_norm": 0.62890625, + "learning_rate": 4.588099809799551e-06, + "loss": 0.2359, "step": 1520 }, { - "epoch": 5.535390199637024, - "grad_norm": 0.7109375, - "learning_rate": 9.84134036165192e-05, - "loss": 0.5661, + "epoch": 13.73873873873874, + "grad_norm": 0.62890625, + "learning_rate": 4.279372475203181e-06, + "loss": 0.2321, "step": 1525 }, { - "epoch": 5.5535390199637025, - "grad_norm": 0.65234375, - "learning_rate": 9.777885453486706e-05, - "loss": 0.5756, + "epoch": 13.783783783783784, + "grad_norm": 0.65625, + "learning_rate": 3.981170033502158e-06, + "loss": 0.2326, "step": 1530 }, { - "epoch": 5.571687840290381, - "grad_norm": 0.625, - "learning_rate": 9.71443949206304e-05, - "loss": 0.5702, + "epoch": 13.82882882882883, + "grad_norm": 0.6328125, + "learning_rate": 3.6935252733319213e-06, + "loss": 0.2341, "step": 1535 }, { - "epoch": 5.58983666061706, - "grad_norm": 0.5859375, - "learning_rate": 9.651005032974994e-05, - "loss": 0.561, + "epoch": 13.873873873873874, + "grad_norm": 0.6171875, + "learning_rate": 3.4164698224656066e-06, + "loss": 0.2308, "step": 1540 }, { - "epoch": 5.607985480943738, - "grad_norm": 0.6953125, - "learning_rate": 9.587584631353329e-05, - "loss": 0.5747, + "epoch": 13.91891891891892, + "grad_norm": 0.66796875, + "learning_rate": 3.1500341443364556e-06, + "loss": 0.2338, "step": 1545 }, { - "epoch": 5.626134301270417, - "grad_norm": 0.6875, - "learning_rate": 9.524180841762577e-05, - "loss": 0.562, + "epoch": 13.963963963963964, + "grad_norm": 0.66796875, + "learning_rate": 2.8942475346882435e-06, + "loss": 0.2361, "step": 1550 }, { - "epoch": 5.6442831215970966, - "grad_norm": 0.76953125, - "learning_rate": 9.460796218098143e-05, - "loss": 0.5608, + "epoch": 14.0, + "eval_loss": 5.106085777282715, + "eval_runtime": 1.0149, + "eval_samples_per_second": 4.927, + "eval_steps_per_second": 1.971, + "step": 1554 + }, + { + "epoch": 14.00900900900901, + "grad_norm": 0.61328125, + "learning_rate": 2.649138118354011e-06, + "loss": 0.2352, "step": 1555 }, { - "epoch": 5.662431941923775, - "grad_norm": 1.03125, - "learning_rate": 9.397433313483416e-05, - "loss": 0.5677, + "epoch": 14.054054054054054, + "grad_norm": 0.6796875, + "learning_rate": 2.4147328461637144e-06, + "loss": 0.2335, "step": 1560 }, { - "epoch": 5.680580762250454, - "grad_norm": 0.734375, - "learning_rate": 9.334094680166962e-05, - "loss": 0.575, + "epoch": 14.0990990990991, + "grad_norm": 0.6484375, + "learning_rate": 2.1910574919808304e-06, + "loss": 0.2308, "step": 1565 }, { - "epoch": 5.6987295825771325, - "grad_norm": 0.9296875, - "learning_rate": 9.270782869419694e-05, - "loss": 0.5709, + "epoch": 14.144144144144144, + "grad_norm": 0.62890625, + "learning_rate": 1.978136649868345e-06, + "loss": 0.2345, "step": 1570 }, { - "epoch": 5.716878402903811, - "grad_norm": 0.86328125, - "learning_rate": 9.207500431432115e-05, - "loss": 0.5556, + "epoch": 14.18918918918919, + "grad_norm": 0.66015625, + "learning_rate": 1.775993731384662e-06, + "loss": 0.235, "step": 1575 }, { - "epoch": 5.73502722323049, - "grad_norm": 0.8828125, - "learning_rate": 9.144249915211605e-05, - "loss": 0.5694, + "epoch": 14.234234234234235, + "grad_norm": 0.6796875, + "learning_rate": 1.584650963009271e-06, + "loss": 0.2337, "step": 1580 }, { - "epoch": 5.753176043557168, - "grad_norm": 0.65625, - "learning_rate": 9.081033868479727e-05, - "loss": 0.5652, + "epoch": 14.27927927927928, + "grad_norm": 0.68359375, + "learning_rate": 1.4041293836989377e-06, + "loss": 0.2319, "step": 1585 }, { - "epoch": 5.771324863883848, - "grad_norm": 0.8125, - "learning_rate": 9.01785483756963e-05, - "loss": 0.568, + "epoch": 14.324324324324325, + "grad_norm": 0.64453125, + "learning_rate": 1.234448842574365e-06, + "loss": 0.2332, "step": 1590 }, { - "epoch": 5.7894736842105265, - "grad_norm": 0.74609375, - "learning_rate": 8.954715367323468e-05, - "loss": 0.5638, + "epoch": 14.36936936936937, + "grad_norm": 0.6328125, + "learning_rate": 1.075627996737627e-06, + "loss": 0.23, "step": 1595 }, { - "epoch": 5.807622504537205, - "grad_norm": 0.671875, - "learning_rate": 8.891618000989891e-05, - "loss": 0.5668, + "epoch": 14.414414414414415, + "grad_norm": 0.6875, + "learning_rate": 9.276843092208553e-07, + "loss": 0.2353, "step": 1600 }, { - "epoch": 5.825771324863884, + "epoch": 14.45945945945946, "grad_norm": 0.640625, - "learning_rate": 8.828565280121617e-05, - "loss": 0.5714, + "learning_rate": 7.90634047066019e-07, + "loss": 0.2339, "step": 1605 }, { - "epoch": 5.8439201451905625, - "grad_norm": 0.69921875, - "learning_rate": 8.765559744473053e-05, - "loss": 0.5709, + "epoch": 14.504504504504505, + "grad_norm": 0.6796875, + "learning_rate": 6.644922795363218e-07, + "loss": 0.2381, "step": 1610 }, { - "epoch": 5.862068965517241, - "grad_norm": 0.69140625, - "learning_rate": 8.702603931897982e-05, - "loss": 0.5834, + "epoch": 14.54954954954955, + "grad_norm": 0.66015625, + "learning_rate": 5.492728764592725e-07, + "loss": 0.2335, "step": 1615 }, { - "epoch": 5.88021778584392, - "grad_norm": 0.64453125, - "learning_rate": 8.639700378247361e-05, - "loss": 0.5673, + "epoch": 14.594594594594595, + "grad_norm": 0.65234375, + "learning_rate": 4.4498850670164906e-07, + "loss": 0.2343, "step": 1620 }, { - "epoch": 5.898366606170599, - "grad_norm": 0.84765625, - "learning_rate": 8.57685161726715e-05, - "loss": 0.5702, + "epoch": 14.63963963963964, + "grad_norm": 0.6015625, + "learning_rate": 3.5165063677645714e-07, + "loss": 0.2369, "step": 1625 }, { - "epoch": 5.916515426497278, - "grad_norm": 0.6328125, - "learning_rate": 8.514060180496285e-05, - "loss": 0.5671, + "epoch": 14.684684684684685, + "grad_norm": 0.65234375, + "learning_rate": 2.6926952958221674e-07, + "loss": 0.2367, "step": 1630 }, { - "epoch": 5.9346642468239565, - "grad_norm": 0.640625, - "learning_rate": 8.451328597164679e-05, - "loss": 0.5686, + "epoch": 14.72972972972973, + "grad_norm": 0.6328125, + "learning_rate": 1.9785424327440906e-07, + "loss": 0.2302, "step": 1635 }, { - "epoch": 5.952813067150635, - "grad_norm": 0.70703125, - "learning_rate": 8.38865939409136e-05, - "loss": 0.5686, + "epoch": 14.774774774774775, + "grad_norm": 0.6640625, + "learning_rate": 1.374126302696066e-07, + "loss": 0.237, "step": 1640 }, { - "epoch": 5.970961887477314, - "grad_norm": 0.6796875, - "learning_rate": 8.326055095582694e-05, - "loss": 0.5685, + "epoch": 14.81981981981982, + "grad_norm": 0.65625, + "learning_rate": 8.795133638197506e-08, + "loss": 0.2298, "step": 1645 }, { - "epoch": 5.9891107078039925, - "grad_norm": 0.625, - "learning_rate": 8.263518223330697e-05, - "loss": 0.5725, + "epoch": 14.864864864864865, + "grad_norm": 0.7109375, + "learning_rate": 4.94758000925799e-08, + "loss": 0.2372, "step": 1650 }, { - "epoch": 6.0, - "eval_loss": 2.4713406562805176, - "eval_runtime": 1.0255, - "eval_samples_per_second": 4.876, - "eval_steps_per_second": 1.95, - "step": 1653 - }, - { - "epoch": 6.007259528130671, - "grad_norm": 0.640625, - "learning_rate": 8.201051296311462e-05, - "loss": 0.5327, + "epoch": 14.90990990990991, + "grad_norm": 0.62890625, + "learning_rate": 2.1990251951398144e-08, + "loss": 0.2312, "step": 1655 }, { - "epoch": 6.025408348457351, - "grad_norm": 0.7265625, - "learning_rate": 8.1386568306837e-05, - "loss": 0.5063, + "epoch": 14.954954954954955, + "grad_norm": 0.62890625, + "learning_rate": 5.497714112157048e-09, + "loss": 0.2311, "step": 1660 }, { - "epoch": 6.043557168784029, - "grad_norm": 0.703125, - "learning_rate": 8.076337339687394e-05, - "loss": 0.4988, + "epoch": 15.0, + "grad_norm": 0.62109375, + "learning_rate": 0.0, + "loss": 0.2319, "step": 1665 }, { - "epoch": 6.061705989110708, - "grad_norm": 0.7421875, - "learning_rate": 8.014095333542548e-05, - "loss": 0.4931, - "step": 1670 - }, - { - "epoch": 6.0798548094373865, - "grad_norm": 0.62890625, - "learning_rate": 7.951933319348095e-05, - "loss": 0.4928, - "step": 1675 - }, - { - "epoch": 6.098003629764065, - "grad_norm": 0.67578125, - "learning_rate": 7.889853800980904e-05, - "loss": 0.4987, - "step": 1680 - }, - { - "epoch": 6.116152450090744, - "grad_norm": 0.734375, - "learning_rate": 7.827859278994925e-05, - "loss": 0.502, - "step": 1685 - }, - { - "epoch": 6.1343012704174225, - "grad_norm": 0.73828125, - "learning_rate": 7.765952250520459e-05, - "loss": 0.494, - "step": 1690 - }, - { - "epoch": 6.152450090744102, - "grad_norm": 0.71875, - "learning_rate": 7.704135209163589e-05, - "loss": 0.4876, - "step": 1695 - }, - { - "epoch": 6.170598911070781, - "grad_norm": 0.6875, - "learning_rate": 7.642410644905726e-05, - "loss": 0.51, - "step": 1700 - }, - { - "epoch": 6.188747731397459, - "grad_norm": 0.64453125, - "learning_rate": 7.580781044003324e-05, - "loss": 0.4999, - "step": 1705 - }, - { - "epoch": 6.206896551724138, - "grad_norm": 0.69140625, - "learning_rate": 7.519248888887716e-05, - "loss": 0.5019, - "step": 1710 - }, - { - "epoch": 6.2250453720508165, - "grad_norm": 0.65234375, - "learning_rate": 7.457816658065134e-05, - "loss": 0.4964, - "step": 1715 - }, - { - "epoch": 6.243194192377495, - "grad_norm": 0.6953125, - "learning_rate": 7.39648682601688e-05, - "loss": 0.4976, - "step": 1720 - }, - { - "epoch": 6.261343012704174, - "grad_norm": 0.62890625, - "learning_rate": 7.335261863099651e-05, - "loss": 0.5046, - "step": 1725 - }, - { - "epoch": 6.279491833030853, - "grad_norm": 0.69140625, - "learning_rate": 7.274144235446023e-05, - "loss": 0.4968, - "step": 1730 - }, - { - "epoch": 6.297640653357532, - "grad_norm": 0.703125, - "learning_rate": 7.213136404865124e-05, - "loss": 0.508, - "step": 1735 - }, - { - "epoch": 6.315789473684211, - "grad_norm": 0.68359375, - "learning_rate": 7.152240828743477e-05, - "loss": 0.4968, - "step": 1740 - }, - { - "epoch": 6.333938294010889, - "grad_norm": 0.83984375, - "learning_rate": 7.09145995994601e-05, - "loss": 0.5009, - "step": 1745 - }, - { - "epoch": 6.352087114337568, - "grad_norm": 0.671875, - "learning_rate": 7.030796246717255e-05, - "loss": 0.4925, - "step": 1750 - }, - { - "epoch": 6.3702359346642465, - "grad_norm": 0.6953125, - "learning_rate": 6.970252132582728e-05, - "loss": 0.5057, - "step": 1755 - }, - { - "epoch": 6.388384754990925, - "grad_norm": 0.6875, - "learning_rate": 6.909830056250527e-05, - "loss": 0.505, - "step": 1760 - }, - { - "epoch": 6.406533575317605, - "grad_norm": 0.69140625, - "learning_rate": 6.849532451513074e-05, - "loss": 0.4962, - "step": 1765 - }, - { - "epoch": 6.424682395644283, - "grad_norm": 0.66796875, - "learning_rate": 6.789361747149093e-05, - "loss": 0.5082, - "step": 1770 - }, - { - "epoch": 6.442831215970962, - "grad_norm": 0.640625, - "learning_rate": 6.729320366825784e-05, - "loss": 0.5006, - "step": 1775 - }, - { - "epoch": 6.460980036297641, - "grad_norm": 0.78515625, - "learning_rate": 6.669410729001193e-05, - "loss": 0.5089, - "step": 1780 - }, - { - "epoch": 6.479128856624319, - "grad_norm": 0.76171875, - "learning_rate": 6.609635246826794e-05, - "loss": 0.5005, - "step": 1785 - }, - { - "epoch": 6.497277676950998, - "grad_norm": 0.7890625, - "learning_rate": 6.549996328050296e-05, - "loss": 0.5045, - "step": 1790 - }, - { - "epoch": 6.5154264972776765, - "grad_norm": 0.78125, - "learning_rate": 6.490496374918647e-05, - "loss": 0.5117, - "step": 1795 - }, - { - "epoch": 6.533575317604356, - "grad_norm": 0.69921875, - "learning_rate": 6.431137784081282e-05, - "loss": 0.5074, - "step": 1800 - }, - { - "epoch": 6.551724137931035, - "grad_norm": 0.65625, - "learning_rate": 6.371922946493591e-05, - "loss": 0.5078, - "step": 1805 - }, - { - "epoch": 6.569872958257713, - "grad_norm": 0.69140625, - "learning_rate": 6.312854247320595e-05, - "loss": 0.5032, - "step": 1810 - }, - { - "epoch": 6.588021778584392, - "grad_norm": 0.7109375, - "learning_rate": 6.25393406584088e-05, - "loss": 0.507, - "step": 1815 - }, - { - "epoch": 6.606170598911071, - "grad_norm": 0.71875, - "learning_rate": 6.19516477535077e-05, - "loss": 0.5106, - "step": 1820 - }, - { - "epoch": 6.624319419237749, - "grad_norm": 0.66796875, - "learning_rate": 6.136548743068713e-05, - "loss": 0.4985, - "step": 1825 - }, - { - "epoch": 6.642468239564428, - "grad_norm": 0.66796875, - "learning_rate": 6.078088330039945e-05, - "loss": 0.5001, - "step": 1830 - }, - { - "epoch": 6.660617059891107, - "grad_norm": 0.68359375, - "learning_rate": 6.019785891041381e-05, - "loss": 0.5002, - "step": 1835 - }, - { - "epoch": 6.678765880217786, - "grad_norm": 0.66015625, - "learning_rate": 5.9616437744867535e-05, - "loss": 0.5107, - "step": 1840 - }, - { - "epoch": 6.696914700544465, - "grad_norm": 0.7890625, - "learning_rate": 5.9036643223320475e-05, - "loss": 0.4967, - "step": 1845 - }, - { - "epoch": 6.715063520871143, - "grad_norm": 0.765625, - "learning_rate": 5.845849869981137e-05, - "loss": 0.4988, - "step": 1850 - }, - { - "epoch": 6.733212341197822, - "grad_norm": 0.7265625, - "learning_rate": 5.788202746191734e-05, - "loss": 0.5003, - "step": 1855 - }, - { - "epoch": 6.751361161524501, - "grad_norm": 0.76171875, - "learning_rate": 5.7307252729815833e-05, - "loss": 0.5062, - "step": 1860 - }, - { - "epoch": 6.769509981851179, - "grad_norm": 0.80859375, - "learning_rate": 5.6734197655349156e-05, - "loss": 0.5056, - "step": 1865 - }, - { - "epoch": 6.787658802177859, - "grad_norm": 0.72265625, - "learning_rate": 5.616288532109225e-05, - "loss": 0.5081, - "step": 1870 - }, - { - "epoch": 6.805807622504537, - "grad_norm": 0.6640625, - "learning_rate": 5.559333873942259e-05, - "loss": 0.502, - "step": 1875 - }, - { - "epoch": 6.823956442831216, - "grad_norm": 0.76171875, - "learning_rate": 5.5025580851593436e-05, - "loss": 0.5091, - "step": 1880 - }, - { - "epoch": 6.842105263157895, - "grad_norm": 0.7265625, - "learning_rate": 5.445963452680973e-05, - "loss": 0.5017, - "step": 1885 - }, - { - "epoch": 6.860254083484573, - "grad_norm": 0.7890625, - "learning_rate": 5.38955225613069e-05, - "loss": 0.5148, - "step": 1890 - }, - { - "epoch": 6.878402903811252, - "grad_norm": 0.6484375, - "learning_rate": 5.333326767743263e-05, - "loss": 0.5044, - "step": 1895 - }, - { - "epoch": 6.896551724137931, - "grad_norm": 0.66015625, - "learning_rate": 5.277289252273174e-05, - "loss": 0.5094, - "step": 1900 - }, - { - "epoch": 6.91470054446461, - "grad_norm": 0.69140625, - "learning_rate": 5.221441966903371e-05, - "loss": 0.5056, - "step": 1905 - }, - { - "epoch": 6.932849364791289, - "grad_norm": 0.78515625, - "learning_rate": 5.1657871611543605e-05, - "loss": 0.5108, - "step": 1910 - }, - { - "epoch": 6.950998185117967, - "grad_norm": 0.796875, - "learning_rate": 5.110327076793613e-05, - "loss": 0.5068, - "step": 1915 - }, - { - "epoch": 6.969147005444646, - "grad_norm": 0.72265625, - "learning_rate": 5.055063947745233e-05, - "loss": 0.5038, - "step": 1920 - }, - { - "epoch": 6.987295825771325, - "grad_norm": 0.67578125, - "learning_rate": 5.000000000000002e-05, - "loss": 0.5089, - "step": 1925 - }, - { - "epoch": 6.998185117967332, - "eval_loss": 2.6490731239318848, - "eval_runtime": 1.0248, - "eval_samples_per_second": 4.879, - "eval_steps_per_second": 1.952, - "step": 1928 - }, - { - "epoch": 7.005444646098003, - "grad_norm": 0.65625, - "learning_rate": 4.945137451525707e-05, - "loss": 0.4914, - "step": 1930 - }, - { - "epoch": 7.023593466424682, - "grad_norm": 0.87109375, - "learning_rate": 4.890478512177795e-05, - "loss": 0.4567, - "step": 1935 - }, - { - "epoch": 7.0417422867513615, - "grad_norm": 0.7109375, - "learning_rate": 4.836025383610382e-05, - "loss": 0.4459, - "step": 1940 - }, - { - "epoch": 7.05989110707804, - "grad_norm": 0.6796875, - "learning_rate": 4.7817802591875426e-05, - "loss": 0.4523, - "step": 1945 - }, - { - "epoch": 7.078039927404719, - "grad_norm": 0.71484375, - "learning_rate": 4.727745323894976e-05, - "loss": 0.4551, - "step": 1950 - }, - { - "epoch": 7.096188747731397, - "grad_norm": 0.72265625, - "learning_rate": 4.673922754252002e-05, - "loss": 0.4435, - "step": 1955 - }, - { - "epoch": 7.114337568058076, - "grad_norm": 0.74609375, - "learning_rate": 4.620314718223876e-05, - "loss": 0.4572, - "step": 1960 - }, - { - "epoch": 7.132486388384755, - "grad_norm": 0.78515625, - "learning_rate": 4.566923375134472e-05, - "loss": 0.4564, - "step": 1965 - }, - { - "epoch": 7.150635208711433, - "grad_norm": 0.73828125, - "learning_rate": 4.513750875579303e-05, - "loss": 0.4482, - "step": 1970 - }, - { - "epoch": 7.168784029038113, - "grad_norm": 0.6875, - "learning_rate": 4.4607993613388976e-05, - "loss": 0.4536, - "step": 1975 - }, - { - "epoch": 7.1869328493647915, - "grad_norm": 0.7578125, - "learning_rate": 4.4080709652925336e-05, - "loss": 0.4571, - "step": 1980 - }, - { - "epoch": 7.20508166969147, - "grad_norm": 0.69921875, - "learning_rate": 4.355567811332311e-05, - "loss": 0.4548, - "step": 1985 - }, - { - "epoch": 7.223230490018149, - "grad_norm": 0.6796875, - "learning_rate": 4.3032920142776125e-05, - "loss": 0.4538, - "step": 1990 - }, - { - "epoch": 7.241379310344827, - "grad_norm": 0.71875, - "learning_rate": 4.251245679789928e-05, - "loss": 0.4497, - "step": 1995 - }, - { - "epoch": 7.259528130671506, - "grad_norm": 0.74609375, - "learning_rate": 4.19943090428802e-05, - "loss": 0.4546, - "step": 2000 - }, - { - "epoch": 7.277676950998185, - "grad_norm": 0.73828125, - "learning_rate": 4.147849774863488e-05, - "loss": 0.4612, - "step": 2005 - }, - { - "epoch": 7.295825771324864, - "grad_norm": 0.7265625, - "learning_rate": 4.096504369196704e-05, - "loss": 0.4437, - "step": 2010 - }, - { - "epoch": 7.313974591651543, - "grad_norm": 0.765625, - "learning_rate": 4.045396755473121e-05, - "loss": 0.4468, - "step": 2015 - }, - { - "epoch": 7.3321234119782215, - "grad_norm": 0.73828125, - "learning_rate": 3.994528992299971e-05, - "loss": 0.4456, - "step": 2020 - }, - { - "epoch": 7.3502722323049, - "grad_norm": 0.671875, - "learning_rate": 3.943903128623335e-05, - "loss": 0.4458, - "step": 2025 - }, - { - "epoch": 7.368421052631579, - "grad_norm": 0.7265625, - "learning_rate": 3.893521203645618e-05, - "loss": 0.4524, - "step": 2030 - }, - { - "epoch": 7.386569872958257, - "grad_norm": 0.73828125, - "learning_rate": 3.843385246743417e-05, - "loss": 0.4534, - "step": 2035 - }, - { - "epoch": 7.404718693284936, - "grad_norm": 0.76171875, - "learning_rate": 3.7934972773857634e-05, - "loss": 0.4577, - "step": 2040 - }, - { - "epoch": 7.422867513611616, - "grad_norm": 0.78515625, - "learning_rate": 3.7438593050527845e-05, - "loss": 0.455, - "step": 2045 - }, - { - "epoch": 7.441016333938294, - "grad_norm": 0.703125, - "learning_rate": 3.694473329154778e-05, - "loss": 0.462, - "step": 2050 - }, - { - "epoch": 7.459165154264973, - "grad_norm": 0.6796875, - "learning_rate": 3.645341338951639e-05, - "loss": 0.4538, - "step": 2055 - }, - { - "epoch": 7.4773139745916515, - "grad_norm": 0.72265625, - "learning_rate": 3.5964653134727776e-05, - "loss": 0.455, - "step": 2060 - }, - { - "epoch": 7.49546279491833, - "grad_norm": 0.7265625, - "learning_rate": 3.547847221437372e-05, - "loss": 0.459, - "step": 2065 - }, - { - "epoch": 7.513611615245009, - "grad_norm": 0.765625, - "learning_rate": 3.4994890211750754e-05, - "loss": 0.4491, - "step": 2070 - }, - { - "epoch": 7.531760435571687, - "grad_norm": 0.7421875, - "learning_rate": 3.45139266054715e-05, - "loss": 0.4535, - "step": 2075 - }, - { - "epoch": 7.549909255898367, - "grad_norm": 0.79296875, - "learning_rate": 3.4035600768679855e-05, - "loss": 0.4517, - "step": 2080 - }, - { - "epoch": 7.568058076225046, - "grad_norm": 0.67578125, - "learning_rate": 3.3559931968270753e-05, - "loss": 0.4535, - "step": 2085 - }, - { - "epoch": 7.586206896551724, - "grad_norm": 0.6875, - "learning_rate": 3.308693936411421e-05, - "loss": 0.4486, - "step": 2090 - }, - { - "epoch": 7.604355716878403, - "grad_norm": 0.7265625, - "learning_rate": 3.2616642008283213e-05, - "loss": 0.459, - "step": 2095 - }, - { - "epoch": 7.6225045372050815, - "grad_norm": 0.76953125, - "learning_rate": 3.21490588442868e-05, - "loss": 0.4579, - "step": 2100 - }, - { - "epoch": 7.64065335753176, - "grad_norm": 0.81640625, - "learning_rate": 3.1684208706306574e-05, - "loss": 0.4431, - "step": 2105 - }, - { - "epoch": 7.658802177858439, - "grad_norm": 0.734375, - "learning_rate": 3.1222110318438304e-05, - "loss": 0.4554, - "step": 2110 - }, - { - "epoch": 7.676950998185118, - "grad_norm": 0.734375, - "learning_rate": 3.076278229393773e-05, - "loss": 0.4574, - "step": 2115 - }, - { - "epoch": 7.695099818511797, - "grad_norm": 0.75, - "learning_rate": 3.030624313447067e-05, - "loss": 0.466, - "step": 2120 - }, - { - "epoch": 7.713248638838476, - "grad_norm": 0.72265625, - "learning_rate": 2.9852511229367865e-05, - "loss": 0.4592, - "step": 2125 - }, - { - "epoch": 7.731397459165154, - "grad_norm": 0.67578125, - "learning_rate": 2.9401604854884357e-05, - "loss": 0.4622, - "step": 2130 - }, - { - "epoch": 7.749546279491833, - "grad_norm": 0.7890625, - "learning_rate": 2.8953542173463133e-05, - "loss": 0.4527, - "step": 2135 - }, - { - "epoch": 7.7676950998185115, - "grad_norm": 0.73828125, - "learning_rate": 2.8508341233003654e-05, - "loss": 0.4565, - "step": 2140 - }, - { - "epoch": 7.78584392014519, - "grad_norm": 0.7734375, - "learning_rate": 2.8066019966134904e-05, - "loss": 0.4548, - "step": 2145 - }, - { - "epoch": 7.80399274047187, - "grad_norm": 0.73046875, - "learning_rate": 2.7626596189492983e-05, - "loss": 0.4584, - "step": 2150 - }, - { - "epoch": 7.822141560798548, - "grad_norm": 0.79296875, - "learning_rate": 2.719008760300359e-05, - "loss": 0.4523, - "step": 2155 - }, - { - "epoch": 7.840290381125227, - "grad_norm": 0.70703125, - "learning_rate": 2.6756511789168925e-05, - "loss": 0.4503, - "step": 2160 - }, - { - "epoch": 7.8584392014519056, - "grad_norm": 0.73046875, - "learning_rate": 2.6325886212359498e-05, - "loss": 0.451, - "step": 2165 - }, - { - "epoch": 7.876588021778584, - "grad_norm": 0.75390625, - "learning_rate": 2.589822821811083e-05, - "loss": 0.4574, - "step": 2170 - }, - { - "epoch": 7.894736842105263, - "grad_norm": 0.7109375, - "learning_rate": 2.5473555032424533e-05, - "loss": 0.4692, - "step": 2175 - }, - { - "epoch": 7.9128856624319415, - "grad_norm": 0.76953125, - "learning_rate": 2.5051883761074614e-05, - "loss": 0.4595, - "step": 2180 - }, - { - "epoch": 7.931034482758621, - "grad_norm": 0.66796875, - "learning_rate": 2.4633231388918378e-05, - "loss": 0.4543, - "step": 2185 - }, - { - "epoch": 7.9491833030853, - "grad_norm": 0.7265625, - "learning_rate": 2.4217614779212315e-05, - "loss": 0.4552, - "step": 2190 - }, - { - "epoch": 7.967332123411978, - "grad_norm": 0.8203125, - "learning_rate": 2.3805050672932928e-05, - "loss": 0.4542, - "step": 2195 - }, - { - "epoch": 7.985480943738657, - "grad_norm": 0.72265625, - "learning_rate": 2.339555568810221e-05, - "loss": 0.4678, - "step": 2200 - }, - { - "epoch": 8.0, - "eval_loss": 2.8434367179870605, - "eval_runtime": 1.027, - "eval_samples_per_second": 4.869, - "eval_steps_per_second": 1.947, - "step": 2204 - }, - { - "epoch": 8.003629764065336, - "grad_norm": 0.6875, - "learning_rate": 2.2989146319118425e-05, - "loss": 0.4538, - "step": 2205 - }, - { - "epoch": 8.021778584392015, - "grad_norm": 0.6484375, - "learning_rate": 2.2585838936091754e-05, - "loss": 0.4327, - "step": 2210 - }, - { - "epoch": 8.039927404718693, - "grad_norm": 0.71875, - "learning_rate": 2.2185649784184746e-05, - "loss": 0.4241, - "step": 2215 - }, - { - "epoch": 8.058076225045372, - "grad_norm": 0.76171875, - "learning_rate": 2.178859498295809e-05, - "loss": 0.4201, - "step": 2220 - }, - { - "epoch": 8.07622504537205, - "grad_norm": 0.77734375, - "learning_rate": 2.139469052572127e-05, - "loss": 0.4283, - "step": 2225 - }, - { - "epoch": 8.09437386569873, - "grad_norm": 0.703125, - "learning_rate": 2.1003952278888382e-05, - "loss": 0.4324, - "step": 2230 - }, - { - "epoch": 8.11252268602541, - "grad_norm": 0.6953125, - "learning_rate": 2.0616395981339075e-05, - "loss": 0.4288, - "step": 2235 - }, - { - "epoch": 8.130671506352087, - "grad_norm": 0.7421875, - "learning_rate": 2.0232037243784475e-05, - "loss": 0.4235, - "step": 2240 - }, - { - "epoch": 8.148820326678766, - "grad_norm": 0.7734375, - "learning_rate": 1.985089154813846e-05, - "loss": 0.4291, - "step": 2245 - }, - { - "epoch": 8.166969147005444, - "grad_norm": 0.71875, - "learning_rate": 1.947297424689414e-05, - "loss": 0.4272, - "step": 2250 - }, - { - "epoch": 8.185117967332124, - "grad_norm": 0.69921875, - "learning_rate": 1.9098300562505266e-05, - "loss": 0.4274, - "step": 2255 - }, - { - "epoch": 8.203266787658801, - "grad_norm": 0.74609375, - "learning_rate": 1.8726885586773212e-05, - "loss": 0.4336, - "step": 2260 - }, - { - "epoch": 8.221415607985481, - "grad_norm": 0.6953125, - "learning_rate": 1.835874428023905e-05, - "loss": 0.4359, - "step": 2265 - }, - { - "epoch": 8.23956442831216, - "grad_norm": 0.703125, - "learning_rate": 1.7993891471580893e-05, - "loss": 0.4275, - "step": 2270 - }, - { - "epoch": 8.257713248638838, - "grad_norm": 0.6796875, - "learning_rate": 1.763234185701673e-05, - "loss": 0.4315, - "step": 2275 - }, - { - "epoch": 8.275862068965518, - "grad_norm": 0.70703125, - "learning_rate": 1.7274109999712295e-05, - "loss": 0.4249, - "step": 2280 - }, - { - "epoch": 8.294010889292196, - "grad_norm": 0.734375, - "learning_rate": 1.6919210329194533e-05, - "loss": 0.4276, - "step": 2285 - }, - { - "epoch": 8.312159709618875, - "grad_norm": 0.765625, - "learning_rate": 1.6567657140770475e-05, - "loss": 0.4315, - "step": 2290 - }, - { - "epoch": 8.330308529945553, - "grad_norm": 0.71875, - "learning_rate": 1.621946459495127e-05, - "loss": 0.4247, - "step": 2295 - }, - { - "epoch": 8.348457350272232, - "grad_norm": 0.76171875, - "learning_rate": 1.587464671688187e-05, - "loss": 0.4441, - "step": 2300 - }, - { - "epoch": 8.366606170598912, - "grad_norm": 0.6875, - "learning_rate": 1.553321739577619e-05, - "loss": 0.4267, - "step": 2305 - }, - { - "epoch": 8.38475499092559, - "grad_norm": 0.66796875, - "learning_rate": 1.5195190384357404e-05, - "loss": 0.4224, - "step": 2310 - }, - { - "epoch": 8.40290381125227, - "grad_norm": 0.73828125, - "learning_rate": 1.4860579298304312e-05, - "loss": 0.4348, - "step": 2315 - }, - { - "epoch": 8.421052631578947, - "grad_norm": 0.703125, - "learning_rate": 1.4529397615702656e-05, - "loss": 0.4321, - "step": 2320 - }, - { - "epoch": 8.439201451905626, - "grad_norm": 0.73828125, - "learning_rate": 1.4201658676502294e-05, - "loss": 0.4386, - "step": 2325 - }, - { - "epoch": 8.457350272232304, - "grad_norm": 0.77734375, - "learning_rate": 1.3877375681979943e-05, - "loss": 0.4366, - "step": 2330 - }, - { - "epoch": 8.475499092558984, - "grad_norm": 0.72265625, - "learning_rate": 1.3556561694207338e-05, - "loss": 0.4336, - "step": 2335 - }, - { - "epoch": 8.493647912885663, - "grad_norm": 0.75, - "learning_rate": 1.3239229635525074e-05, - "loss": 0.4274, - "step": 2340 - }, - { - "epoch": 8.511796733212341, - "grad_norm": 0.7109375, - "learning_rate": 1.2925392288022298e-05, - "loss": 0.4259, - "step": 2345 - }, - { - "epoch": 8.52994555353902, - "grad_norm": 0.66796875, - "learning_rate": 1.2615062293021507e-05, - "loss": 0.427, - "step": 2350 - }, - { - "epoch": 8.548094373865698, - "grad_norm": 0.67578125, - "learning_rate": 1.230825215056971e-05, - "loss": 0.4226, - "step": 2355 - }, - { - "epoch": 8.566243194192378, - "grad_norm": 0.7421875, - "learning_rate": 1.2004974218934695e-05, - "loss": 0.4361, - "step": 2360 - }, - { - "epoch": 8.584392014519056, - "grad_norm": 0.65625, - "learning_rate": 1.1705240714107302e-05, - "loss": 0.4291, - "step": 2365 - }, - { - "epoch": 8.602540834845735, - "grad_norm": 1.171875, - "learning_rate": 1.1409063709309442e-05, - "loss": 0.4416, - "step": 2370 - }, - { - "epoch": 8.620689655172415, - "grad_norm": 0.69921875, - "learning_rate": 1.1116455134507664e-05, - "loss": 0.425, - "step": 2375 - }, - { - "epoch": 8.638838475499092, - "grad_norm": 0.74609375, - "learning_rate": 1.0827426775932658e-05, - "loss": 0.4316, - "step": 2380 - }, - { - "epoch": 8.656987295825772, - "grad_norm": 0.703125, - "learning_rate": 1.054199027560463e-05, - "loss": 0.425, - "step": 2385 - }, - { - "epoch": 8.67513611615245, - "grad_norm": 0.6953125, - "learning_rate": 1.026015713086418e-05, - "loss": 0.4344, - "step": 2390 - }, - { - "epoch": 8.69328493647913, - "grad_norm": 0.76171875, - "learning_rate": 9.98193869390922e-06, - "loss": 0.4261, - "step": 2395 - }, - { - "epoch": 8.711433756805807, - "grad_norm": 0.75, - "learning_rate": 9.707346171337894e-06, - "loss": 0.4323, - "step": 2400 - }, - { - "epoch": 8.729582577132486, - "grad_norm": 0.71484375, - "learning_rate": 9.436390623696911e-06, - "loss": 0.4309, - "step": 2405 - }, - { - "epoch": 8.747731397459166, - "grad_norm": 0.76171875, - "learning_rate": 9.16908296503628e-06, - "loss": 0.4278, - "step": 2410 - }, - { - "epoch": 8.765880217785844, - "grad_norm": 0.765625, - "learning_rate": 8.905433962469489e-06, - "loss": 0.4289, - "step": 2415 - }, - { - "epoch": 8.784029038112523, - "grad_norm": 0.76171875, - "learning_rate": 8.645454235739903e-06, - "loss": 0.4323, - "step": 2420 - }, - { - "epoch": 8.802177858439201, - "grad_norm": 0.66796875, - "learning_rate": 8.38915425679304e-06, - "loss": 0.4286, - "step": 2425 - }, - { - "epoch": 8.82032667876588, - "grad_norm": 0.69140625, - "learning_rate": 8.13654434935467e-06, - "loss": 0.4267, - "step": 2430 - }, - { - "epoch": 8.838475499092558, - "grad_norm": 0.75390625, - "learning_rate": 7.887634688515e-06, - "loss": 0.4351, - "step": 2435 - }, - { - "epoch": 8.856624319419238, - "grad_norm": 0.7265625, - "learning_rate": 7.642435300318907e-06, - "loss": 0.4294, - "step": 2440 - }, - { - "epoch": 8.874773139745917, - "grad_norm": 0.73828125, - "learning_rate": 7.400956061361974e-06, - "loss": 0.4301, - "step": 2445 - }, - { - "epoch": 8.892921960072595, - "grad_norm": 0.69921875, - "learning_rate": 7.163206698392744e-06, - "loss": 0.4306, - "step": 2450 - }, - { - "epoch": 8.911070780399275, - "grad_norm": 0.7265625, - "learning_rate": 6.929196787920899e-06, - "loss": 0.433, - "step": 2455 - }, - { - "epoch": 8.929219600725952, - "grad_norm": 0.7265625, - "learning_rate": 6.698935755831492e-06, - "loss": 0.4332, - "step": 2460 - }, - { - "epoch": 8.947368421052632, - "grad_norm": 0.70703125, - "learning_rate": 6.472432877005341e-06, - "loss": 0.4364, - "step": 2465 - }, - { - "epoch": 8.96551724137931, - "grad_norm": 0.72265625, - "learning_rate": 6.2496972749453766e-06, - "loss": 0.434, - "step": 2470 - }, - { - "epoch": 8.98366606170599, - "grad_norm": 0.65625, - "learning_rate": 6.030737921409169e-06, - "loss": 0.433, - "step": 2475 - }, - { - "epoch": 8.998185117967331, - "eval_loss": 2.9603512287139893, - "eval_runtime": 1.0237, - "eval_samples_per_second": 4.884, - "eval_steps_per_second": 1.954, - "step": 2479 - }, - { - "epoch": 9.001814882032669, - "grad_norm": 0.68359375, - "learning_rate": 5.8155636360475385e-06, - "loss": 0.4369, - "step": 2480 - }, - { - "epoch": 9.019963702359346, - "grad_norm": 0.68359375, - "learning_rate": 5.604183086049342e-06, - "loss": 0.4232, - "step": 2485 - }, - { - "epoch": 9.038112522686026, - "grad_norm": 0.7109375, - "learning_rate": 5.396604785792281e-06, - "loss": 0.4265, - "step": 2490 - }, - { - "epoch": 9.056261343012704, - "grad_norm": 0.73828125, - "learning_rate": 5.192837096500058e-06, - "loss": 0.4255, - "step": 2495 - }, - { - "epoch": 9.074410163339383, - "grad_norm": 0.7265625, - "learning_rate": 4.992888225905468e-06, - "loss": 0.4207, - "step": 2500 - }, - { - "epoch": 9.092558983666061, - "grad_norm": 0.6875, - "learning_rate": 4.796766227919857e-06, - "loss": 0.4223, - "step": 2505 - }, - { - "epoch": 9.11070780399274, - "grad_norm": 0.72265625, - "learning_rate": 4.604479002308737e-06, - "loss": 0.4237, - "step": 2510 - }, - { - "epoch": 9.12885662431942, - "grad_norm": 0.7421875, - "learning_rate": 4.416034294373472e-06, - "loss": 0.4417, - "step": 2515 - }, - { - "epoch": 9.147005444646098, - "grad_norm": 0.69140625, - "learning_rate": 4.231439694639483e-06, - "loss": 0.4247, - "step": 2520 - }, - { - "epoch": 9.165154264972777, - "grad_norm": 0.70703125, - "learning_rate": 4.050702638550275e-06, - "loss": 0.4244, - "step": 2525 - }, - { - "epoch": 9.183303085299455, - "grad_norm": 0.69921875, - "learning_rate": 3.873830406168111e-06, - "loss": 0.4309, - "step": 2530 - }, - { - "epoch": 9.201451905626135, - "grad_norm": 0.71875, - "learning_rate": 3.7008301218807716e-06, - "loss": 0.4264, - "step": 2535 - }, - { - "epoch": 9.219600725952812, - "grad_norm": 0.73046875, - "learning_rate": 3.5317087541144377e-06, - "loss": 0.4226, - "step": 2540 - }, - { - "epoch": 9.237749546279492, - "grad_norm": 0.6953125, - "learning_rate": 3.3664731150531482e-06, - "loss": 0.4263, - "step": 2545 - }, - { - "epoch": 9.255898366606171, - "grad_norm": 0.7578125, - "learning_rate": 3.2051298603643753e-06, - "loss": 0.4272, - "step": 2550 - }, - { - "epoch": 9.27404718693285, - "grad_norm": 0.84375, - "learning_rate": 3.047685488930874e-06, - "loss": 0.4207, - "step": 2555 - }, - { - "epoch": 9.292196007259529, - "grad_norm": 0.73046875, - "learning_rate": 2.894146342588977e-06, - "loss": 0.426, - "step": 2560 - }, - { - "epoch": 9.310344827586206, - "grad_norm": 0.75390625, - "learning_rate": 2.744518605873092e-06, - "loss": 0.4308, - "step": 2565 - }, - { - "epoch": 9.328493647912886, - "grad_norm": 0.76953125, - "learning_rate": 2.5988083057666533e-06, - "loss": 0.4262, - "step": 2570 - }, - { - "epoch": 9.346642468239564, - "grad_norm": 0.75, - "learning_rate": 2.4570213114592954e-06, - "loss": 0.4225, - "step": 2575 - }, - { - "epoch": 9.364791288566243, - "grad_norm": 0.73046875, - "learning_rate": 2.3191633341104856e-06, - "loss": 0.4198, - "step": 2580 - }, - { - "epoch": 9.382940108892923, - "grad_norm": 0.69921875, - "learning_rate": 2.1852399266194314e-06, - "loss": 0.4168, - "step": 2585 - }, - { - "epoch": 9.4010889292196, - "grad_norm": 0.69921875, - "learning_rate": 2.05525648340148e-06, - "loss": 0.4234, - "step": 2590 - }, - { - "epoch": 9.41923774954628, - "grad_norm": 0.75, - "learning_rate": 1.9292182401707603e-06, - "loss": 0.4383, - "step": 2595 - }, - { - "epoch": 9.437386569872958, - "grad_norm": 0.68359375, - "learning_rate": 1.8071302737293295e-06, - "loss": 0.4203, - "step": 2600 - }, - { - "epoch": 9.455535390199637, - "grad_norm": 0.7265625, - "learning_rate": 1.6889975017626903e-06, - "loss": 0.4152, - "step": 2605 - }, - { - "epoch": 9.473684210526315, - "grad_norm": 0.6875, - "learning_rate": 1.574824682641629e-06, - "loss": 0.4275, - "step": 2610 - }, - { - "epoch": 9.491833030852995, - "grad_norm": 0.6640625, - "learning_rate": 1.4646164152307018e-06, - "loss": 0.4209, - "step": 2615 - }, - { - "epoch": 9.509981851179674, - "grad_norm": 0.734375, - "learning_rate": 1.3583771387028265e-06, - "loss": 0.429, - "step": 2620 - }, - { - "epoch": 9.528130671506352, - "grad_norm": 0.69140625, - "learning_rate": 1.2561111323605712e-06, - "loss": 0.4182, - "step": 2625 - }, - { - "epoch": 9.546279491833031, - "grad_norm": 0.7109375, - "learning_rate": 1.157822515463758e-06, - "loss": 0.424, - "step": 2630 - }, - { - "epoch": 9.56442831215971, - "grad_norm": 0.6875, - "learning_rate": 1.0635152470635512e-06, - "loss": 0.4211, - "step": 2635 - }, - { - "epoch": 9.582577132486389, - "grad_norm": 0.69140625, - "learning_rate": 9.731931258429638e-07, - "loss": 0.4202, - "step": 2640 - }, - { - "epoch": 9.600725952813066, - "grad_norm": 0.6953125, - "learning_rate": 8.868597899638898e-07, - "loss": 0.4183, - "step": 2645 - }, - { - "epoch": 9.618874773139746, - "grad_norm": 0.6796875, - "learning_rate": 8.04518716920466e-07, - "loss": 0.4258, - "step": 2650 - }, - { - "epoch": 9.637023593466425, - "grad_norm": 0.6953125, - "learning_rate": 7.261732233991513e-07, - "loss": 0.426, - "step": 2655 - }, - { - "epoch": 9.655172413793103, - "grad_norm": 0.69921875, - "learning_rate": 6.518264651449779e-07, - "loss": 0.4341, - "step": 2660 - }, - { - "epoch": 9.673321234119783, - "grad_norm": 0.7265625, - "learning_rate": 5.814814368345412e-07, - "loss": 0.4311, - "step": 2665 - }, - { - "epoch": 9.69147005444646, - "grad_norm": 0.72265625, - "learning_rate": 5.151409719553079e-07, - "loss": 0.4219, - "step": 2670 - }, - { - "epoch": 9.70961887477314, - "grad_norm": 0.6796875, - "learning_rate": 4.5280774269154115e-07, - "loss": 0.4205, - "step": 2675 - }, - { - "epoch": 9.727767695099818, - "grad_norm": 0.74609375, - "learning_rate": 3.9448425981661876e-07, - "loss": 0.4268, - "step": 2680 - }, - { - "epoch": 9.745916515426497, - "grad_norm": 0.69921875, - "learning_rate": 3.401728725919373e-07, - "loss": 0.4258, - "step": 2685 - }, - { - "epoch": 9.764065335753177, - "grad_norm": 0.73046875, - "learning_rate": 2.898757686722542e-07, - "loss": 0.4191, - "step": 2690 - }, - { - "epoch": 9.782214156079855, - "grad_norm": 0.78125, - "learning_rate": 2.4359497401758024e-07, - "loss": 0.4277, - "step": 2695 - }, - { - "epoch": 9.800362976406534, - "grad_norm": 0.68359375, - "learning_rate": 2.0133235281156736e-07, - "loss": 0.434, - "step": 2700 - }, - { - "epoch": 9.818511796733212, - "grad_norm": 0.69921875, - "learning_rate": 1.630896073864352e-07, - "loss": 0.4307, - "step": 2705 - }, - { - "epoch": 9.836660617059891, - "grad_norm": 0.71484375, - "learning_rate": 1.2886827815440372e-07, - "loss": 0.4291, - "step": 2710 - }, - { - "epoch": 9.85480943738657, - "grad_norm": 0.73046875, - "learning_rate": 9.866974354560965e-08, - "loss": 0.4284, - "step": 2715 - }, - { - "epoch": 9.872958257713249, - "grad_norm": 0.796875, - "learning_rate": 7.249521995263964e-08, - "loss": 0.4319, - "step": 2720 - }, - { - "epoch": 9.891107078039928, - "grad_norm": 0.71484375, - "learning_rate": 5.0345761681491746e-08, - "loss": 0.4329, - "step": 2725 - }, - { - "epoch": 9.909255898366606, - "grad_norm": 0.7421875, - "learning_rate": 3.2222260909087196e-08, - "loss": 0.4218, - "step": 2730 - }, - { - "epoch": 9.927404718693285, - "grad_norm": 1.0, - "learning_rate": 1.81254476474213e-08, - "loss": 0.4236, - "step": 2735 - }, - { - "epoch": 9.945553539019963, - "grad_norm": 0.69921875, - "learning_rate": 8.055889714064791e-09, - "loss": 0.424, - "step": 2740 - }, - { - "epoch": 9.963702359346643, - "grad_norm": 0.71875, - "learning_rate": 2.0139927093487664e-09, - "loss": 0.4362, - "step": 2745 - }, - { - "epoch": 9.98185117967332, - "grad_norm": 0.69140625, - "learning_rate": 0.0, - "loss": 0.4229, - "step": 2750 - }, - { - "epoch": 9.98185117967332, - "eval_loss": 2.9801077842712402, - "eval_runtime": 1.0256, - "eval_samples_per_second": 4.875, - "eval_steps_per_second": 1.95, - "step": 2750 + "epoch": 15.0, + "eval_loss": 5.111132621765137, + "eval_runtime": 1.0137, + "eval_samples_per_second": 4.932, + "eval_steps_per_second": 1.973, + "step": 1665 }, { - "epoch": 9.98185117967332, - "step": 2750, - "total_flos": 4.2044012259841147e+18, - "train_loss": 1.3294648733139038, - "train_runtime": 22350.6599, - "train_samples_per_second": 1.971, - "train_steps_per_second": 0.123 + "epoch": 15.0, + "step": 1665, + "total_flos": 2.545573832974926e+18, + "train_loss": 1.4823458194016694, + "train_runtime": 13248.7618, + "train_samples_per_second": 2.007, + "train_steps_per_second": 0.126 } ], "logging_steps": 5, - "max_steps": 2750, + "max_steps": 1665, "num_input_tokens_seen": 0, - "num_train_epochs": 10, + "num_train_epochs": 15, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { @@ -3972,7 +2493,7 @@ "attributes": {} } }, - "total_flos": 4.2044012259841147e+18, + "total_flos": 2.545573832974926e+18, "train_batch_size": 4, "trial_name": null, "trial_params": null