{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.9998557067962099, "eval_steps": 500, "global_step": 5197, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.00019239093838680199, "grad_norm": 39.856727253059226, "learning_rate": 1.282051282051282e-07, "loss": 1.7886, "step": 1 }, { "epoch": 0.00038478187677360397, "grad_norm": 43.99718465803916, "learning_rate": 2.564102564102564e-07, "loss": 1.9234, "step": 2 }, { "epoch": 0.0005771728151604059, "grad_norm": 32.389076335497585, "learning_rate": 3.846153846153847e-07, "loss": 1.851, "step": 3 }, { "epoch": 0.0007695637535472079, "grad_norm": 36.69384278283701, "learning_rate": 5.128205128205128e-07, "loss": 1.8184, "step": 4 }, { "epoch": 0.0009619546919340099, "grad_norm": 36.36792193192248, "learning_rate": 6.41025641025641e-07, "loss": 1.9427, "step": 5 }, { "epoch": 0.0011543456303208118, "grad_norm": 36.11364989047238, "learning_rate": 7.692307692307694e-07, "loss": 1.8813, "step": 6 }, { "epoch": 0.0013467365687076138, "grad_norm": 31.392391061688958, "learning_rate": 8.974358974358975e-07, "loss": 1.8073, "step": 7 }, { "epoch": 0.0015391275070944159, "grad_norm": 29.647126114614526, "learning_rate": 1.0256410256410257e-06, "loss": 1.7741, "step": 8 }, { "epoch": 0.001731518445481218, "grad_norm": 31.45726929004091, "learning_rate": 1.153846153846154e-06, "loss": 1.8229, "step": 9 }, { "epoch": 0.0019239093838680197, "grad_norm": 25.833914314971167, "learning_rate": 1.282051282051282e-06, "loss": 1.6523, "step": 10 }, { "epoch": 0.002116300322254822, "grad_norm": 21.913983740282774, "learning_rate": 1.4102564102564104e-06, "loss": 1.6231, "step": 11 }, { "epoch": 0.0023086912606416236, "grad_norm": 20.956554346104312, "learning_rate": 1.5384615384615387e-06, "loss": 1.6949, "step": 12 }, { "epoch": 0.0025010821990284256, "grad_norm": 8.890485569053753, "learning_rate": 1.6666666666666667e-06, "loss": 1.6307, "step": 13 }, { "epoch": 0.0026934731374152277, "grad_norm": 9.103272141497593, "learning_rate": 1.794871794871795e-06, "loss": 1.507, "step": 14 }, { "epoch": 0.0028858640758020297, "grad_norm": 7.92450736032254, "learning_rate": 1.9230769230769234e-06, "loss": 1.5732, "step": 15 }, { "epoch": 0.0030782550141888318, "grad_norm": 7.6073379960965735, "learning_rate": 2.0512820512820513e-06, "loss": 1.5783, "step": 16 }, { "epoch": 0.003270645952575634, "grad_norm": 1.3008203691710705, "learning_rate": 2.1794871794871797e-06, "loss": 0.8074, "step": 17 }, { "epoch": 0.003463036890962436, "grad_norm": 4.791383699296221, "learning_rate": 2.307692307692308e-06, "loss": 1.4051, "step": 18 }, { "epoch": 0.0036554278293492374, "grad_norm": 5.365124215656911, "learning_rate": 2.435897435897436e-06, "loss": 1.4299, "step": 19 }, { "epoch": 0.0038478187677360395, "grad_norm": 5.715042732562032, "learning_rate": 2.564102564102564e-06, "loss": 1.4515, "step": 20 }, { "epoch": 0.004040209706122842, "grad_norm": 4.871153654770502, "learning_rate": 2.6923076923076923e-06, "loss": 1.3966, "step": 21 }, { "epoch": 0.004232600644509644, "grad_norm": 5.175730400245004, "learning_rate": 2.8205128205128207e-06, "loss": 1.4942, "step": 22 }, { "epoch": 0.004424991582896445, "grad_norm": 4.955843317998343, "learning_rate": 2.948717948717949e-06, "loss": 1.3065, "step": 23 }, { "epoch": 0.004617382521283247, "grad_norm": 4.381100652246126, "learning_rate": 3.0769230769230774e-06, "loss": 1.4412, "step": 24 }, { "epoch": 0.004809773459670049, "grad_norm": 3.5575427258372687, "learning_rate": 3.205128205128206e-06, "loss": 1.4477, "step": 25 }, { "epoch": 0.005002164398056851, "grad_norm": 3.7489971646457976, "learning_rate": 3.3333333333333333e-06, "loss": 1.3719, "step": 26 }, { "epoch": 0.005194555336443653, "grad_norm": 3.471365163148772, "learning_rate": 3.4615384615384617e-06, "loss": 1.3902, "step": 27 }, { "epoch": 0.005386946274830455, "grad_norm": 4.046531170440466, "learning_rate": 3.58974358974359e-06, "loss": 1.3699, "step": 28 }, { "epoch": 0.005579337213217257, "grad_norm": 2.9975705062839797, "learning_rate": 3.7179487179487184e-06, "loss": 1.2983, "step": 29 }, { "epoch": 0.0057717281516040594, "grad_norm": 3.2957922651947005, "learning_rate": 3.846153846153847e-06, "loss": 1.4126, "step": 30 }, { "epoch": 0.0059641190899908615, "grad_norm": 3.7184653587418572, "learning_rate": 3.974358974358974e-06, "loss": 1.3331, "step": 31 }, { "epoch": 0.0061565100283776635, "grad_norm": 3.351193359447979, "learning_rate": 4.102564102564103e-06, "loss": 1.3589, "step": 32 }, { "epoch": 0.0063489009667644656, "grad_norm": 2.839250676000173, "learning_rate": 4.230769230769231e-06, "loss": 1.3368, "step": 33 }, { "epoch": 0.006541291905151268, "grad_norm": 2.8342203871986458, "learning_rate": 4.358974358974359e-06, "loss": 1.2768, "step": 34 }, { "epoch": 0.00673368284353807, "grad_norm": 7.398450472002764, "learning_rate": 4.487179487179488e-06, "loss": 1.2251, "step": 35 }, { "epoch": 0.006926073781924872, "grad_norm": 4.600384449671116, "learning_rate": 4.615384615384616e-06, "loss": 1.2473, "step": 36 }, { "epoch": 0.007118464720311674, "grad_norm": 2.4228194571161192, "learning_rate": 4.743589743589744e-06, "loss": 1.3354, "step": 37 }, { "epoch": 0.007310855658698475, "grad_norm": 2.4000946660201077, "learning_rate": 4.871794871794872e-06, "loss": 1.292, "step": 38 }, { "epoch": 0.007503246597085277, "grad_norm": 2.3488003165971523, "learning_rate": 5e-06, "loss": 1.2718, "step": 39 }, { "epoch": 0.007695637535472079, "grad_norm": 2.274536073797015, "learning_rate": 5.128205128205128e-06, "loss": 1.2174, "step": 40 }, { "epoch": 0.007888028473858881, "grad_norm": 2.2462107178462216, "learning_rate": 5.256410256410257e-06, "loss": 1.252, "step": 41 }, { "epoch": 0.008080419412245684, "grad_norm": 1.899016322234258, "learning_rate": 5.384615384615385e-06, "loss": 1.1814, "step": 42 }, { "epoch": 0.008272810350632485, "grad_norm": 1.850672649056904, "learning_rate": 5.512820512820514e-06, "loss": 1.2675, "step": 43 }, { "epoch": 0.008465201289019288, "grad_norm": 2.0671292460046167, "learning_rate": 5.641025641025641e-06, "loss": 1.2131, "step": 44 }, { "epoch": 0.00865759222740609, "grad_norm": 2.1400070943287868, "learning_rate": 5.769230769230769e-06, "loss": 1.2174, "step": 45 }, { "epoch": 0.00884998316579289, "grad_norm": 2.1789334244158995, "learning_rate": 5.897435897435898e-06, "loss": 1.2416, "step": 46 }, { "epoch": 0.009042374104179693, "grad_norm": 2.9158967106185743, "learning_rate": 6.025641025641026e-06, "loss": 1.3689, "step": 47 }, { "epoch": 0.009234765042566494, "grad_norm": 2.059633778739421, "learning_rate": 6.153846153846155e-06, "loss": 1.1999, "step": 48 }, { "epoch": 0.009427155980953297, "grad_norm": 2.0919510733316806, "learning_rate": 6.282051282051282e-06, "loss": 1.3064, "step": 49 }, { "epoch": 0.009619546919340098, "grad_norm": 1.8944065817360585, "learning_rate": 6.410256410256412e-06, "loss": 1.2751, "step": 50 }, { "epoch": 0.009811937857726901, "grad_norm": 2.077146778804178, "learning_rate": 6.538461538461539e-06, "loss": 1.2519, "step": 51 }, { "epoch": 0.010004328796113703, "grad_norm": 2.055924387328907, "learning_rate": 6.666666666666667e-06, "loss": 1.2509, "step": 52 }, { "epoch": 0.010196719734500505, "grad_norm": 2.013398816823918, "learning_rate": 6.794871794871796e-06, "loss": 1.1987, "step": 53 }, { "epoch": 0.010389110672887307, "grad_norm": 1.9122663528057509, "learning_rate": 6.923076923076923e-06, "loss": 1.2263, "step": 54 }, { "epoch": 0.01058150161127411, "grad_norm": 2.2331999720599254, "learning_rate": 7.051282051282053e-06, "loss": 1.3284, "step": 55 }, { "epoch": 0.01077389254966091, "grad_norm": 1.8907043771192067, "learning_rate": 7.17948717948718e-06, "loss": 1.2885, "step": 56 }, { "epoch": 0.010966283488047714, "grad_norm": 1.92179040593395, "learning_rate": 7.307692307692308e-06, "loss": 1.2194, "step": 57 }, { "epoch": 0.011158674426434515, "grad_norm": 1.9288136589965414, "learning_rate": 7.435897435897437e-06, "loss": 1.2684, "step": 58 }, { "epoch": 0.011351065364821318, "grad_norm": 1.9758658355758019, "learning_rate": 7.564102564102564e-06, "loss": 1.2582, "step": 59 }, { "epoch": 0.011543456303208119, "grad_norm": 1.8187490862845184, "learning_rate": 7.692307692307694e-06, "loss": 1.2103, "step": 60 }, { "epoch": 0.01173584724159492, "grad_norm": 2.139067342197671, "learning_rate": 7.820512820512822e-06, "loss": 1.2314, "step": 61 }, { "epoch": 0.011928238179981723, "grad_norm": 1.6612094573020384, "learning_rate": 7.948717948717949e-06, "loss": 1.2866, "step": 62 }, { "epoch": 0.012120629118368524, "grad_norm": 2.1421365947898288, "learning_rate": 8.076923076923077e-06, "loss": 1.2747, "step": 63 }, { "epoch": 0.012313020056755327, "grad_norm": 1.8587006818697545, "learning_rate": 8.205128205128205e-06, "loss": 1.1847, "step": 64 }, { "epoch": 0.012505410995142128, "grad_norm": 1.7369243955992775, "learning_rate": 8.333333333333334e-06, "loss": 1.1494, "step": 65 }, { "epoch": 0.012697801933528931, "grad_norm": 2.2246880669238984, "learning_rate": 8.461538461538462e-06, "loss": 1.2235, "step": 66 }, { "epoch": 0.012890192871915732, "grad_norm": 1.8680117021581786, "learning_rate": 8.58974358974359e-06, "loss": 1.187, "step": 67 }, { "epoch": 0.013082583810302535, "grad_norm": 1.914526909017439, "learning_rate": 8.717948717948719e-06, "loss": 1.2273, "step": 68 }, { "epoch": 0.013274974748689336, "grad_norm": 1.8467174968068547, "learning_rate": 8.846153846153847e-06, "loss": 1.2412, "step": 69 }, { "epoch": 0.01346736568707614, "grad_norm": 1.7811175413005071, "learning_rate": 8.974358974358976e-06, "loss": 1.1859, "step": 70 }, { "epoch": 0.01365975662546294, "grad_norm": 1.8170091793598775, "learning_rate": 9.102564102564104e-06, "loss": 1.1528, "step": 71 }, { "epoch": 0.013852147563849743, "grad_norm": 2.007838107192024, "learning_rate": 9.230769230769232e-06, "loss": 1.2334, "step": 72 }, { "epoch": 0.014044538502236545, "grad_norm": 1.9119397633609467, "learning_rate": 9.358974358974359e-06, "loss": 1.1795, "step": 73 }, { "epoch": 0.014236929440623347, "grad_norm": 1.9197211313506093, "learning_rate": 9.487179487179487e-06, "loss": 1.1924, "step": 74 }, { "epoch": 0.014429320379010149, "grad_norm": 1.9716401904917051, "learning_rate": 9.615384615384616e-06, "loss": 1.1737, "step": 75 }, { "epoch": 0.01462171131739695, "grad_norm": 2.11779054468896, "learning_rate": 9.743589743589744e-06, "loss": 1.2214, "step": 76 }, { "epoch": 0.014814102255783753, "grad_norm": 1.7455813131697298, "learning_rate": 9.871794871794872e-06, "loss": 1.1749, "step": 77 }, { "epoch": 0.015006493194170554, "grad_norm": 1.6258802000847046, "learning_rate": 1e-05, "loss": 1.2528, "step": 78 }, { "epoch": 0.015198884132557357, "grad_norm": 1.8089159074501866, "learning_rate": 1.012820512820513e-05, "loss": 1.2199, "step": 79 }, { "epoch": 0.015391275070944158, "grad_norm": 1.983216157555372, "learning_rate": 1.0256410256410256e-05, "loss": 1.1582, "step": 80 }, { "epoch": 0.01558366600933096, "grad_norm": 1.7537139947168217, "learning_rate": 1.0384615384615386e-05, "loss": 1.2097, "step": 81 }, { "epoch": 0.015776056947717762, "grad_norm": 1.7995399386661501, "learning_rate": 1.0512820512820514e-05, "loss": 1.1454, "step": 82 }, { "epoch": 0.015968447886104565, "grad_norm": 1.8309937633078612, "learning_rate": 1.0641025641025643e-05, "loss": 1.161, "step": 83 }, { "epoch": 0.016160838824491368, "grad_norm": 2.0239031617208814, "learning_rate": 1.076923076923077e-05, "loss": 1.202, "step": 84 }, { "epoch": 0.016353229762878167, "grad_norm": 1.5810688781902729, "learning_rate": 1.0897435897435898e-05, "loss": 0.8564, "step": 85 }, { "epoch": 0.01654562070126497, "grad_norm": 2.186988724544659, "learning_rate": 1.1025641025641028e-05, "loss": 1.1952, "step": 86 }, { "epoch": 0.016738011639651773, "grad_norm": 1.772276699140469, "learning_rate": 1.1153846153846154e-05, "loss": 1.1902, "step": 87 }, { "epoch": 0.016930402578038576, "grad_norm": 1.7263318983697422, "learning_rate": 1.1282051282051283e-05, "loss": 1.2021, "step": 88 }, { "epoch": 0.017122793516425375, "grad_norm": 1.810582714856199, "learning_rate": 1.1410256410256411e-05, "loss": 1.223, "step": 89 }, { "epoch": 0.01731518445481218, "grad_norm": 1.9015989230496069, "learning_rate": 1.1538461538461538e-05, "loss": 1.1937, "step": 90 }, { "epoch": 0.01750757539319898, "grad_norm": 1.9848568571050291, "learning_rate": 1.1666666666666668e-05, "loss": 1.2445, "step": 91 }, { "epoch": 0.01769996633158578, "grad_norm": 1.8777361279218723, "learning_rate": 1.1794871794871796e-05, "loss": 1.2157, "step": 92 }, { "epoch": 0.017892357269972584, "grad_norm": 2.2378879743831845, "learning_rate": 1.1923076923076925e-05, "loss": 1.2185, "step": 93 }, { "epoch": 0.018084748208359386, "grad_norm": 1.9637153215100094, "learning_rate": 1.2051282051282051e-05, "loss": 1.143, "step": 94 }, { "epoch": 0.01827713914674619, "grad_norm": 1.907184379666683, "learning_rate": 1.217948717948718e-05, "loss": 1.195, "step": 95 }, { "epoch": 0.01846953008513299, "grad_norm": 1.9110430117645854, "learning_rate": 1.230769230769231e-05, "loss": 1.1938, "step": 96 }, { "epoch": 0.01866192102351979, "grad_norm": 1.6092919397060848, "learning_rate": 1.2435897435897436e-05, "loss": 1.1375, "step": 97 }, { "epoch": 0.018854311961906595, "grad_norm": 1.8194099385293594, "learning_rate": 1.2564102564102565e-05, "loss": 1.1754, "step": 98 }, { "epoch": 0.019046702900293398, "grad_norm": 1.6312583977543342, "learning_rate": 1.2692307692307693e-05, "loss": 1.1577, "step": 99 }, { "epoch": 0.019239093838680197, "grad_norm": 1.765312628653466, "learning_rate": 1.2820512820512823e-05, "loss": 1.1877, "step": 100 }, { "epoch": 0.019431484777067, "grad_norm": 2.066766998255024, "learning_rate": 1.294871794871795e-05, "loss": 1.2673, "step": 101 }, { "epoch": 0.019623875715453803, "grad_norm": 1.7911087077904286, "learning_rate": 1.3076923076923078e-05, "loss": 1.1754, "step": 102 }, { "epoch": 0.019816266653840606, "grad_norm": 1.847640055882121, "learning_rate": 1.3205128205128207e-05, "loss": 1.1878, "step": 103 }, { "epoch": 0.020008657592227405, "grad_norm": 1.722862361164871, "learning_rate": 1.3333333333333333e-05, "loss": 1.1189, "step": 104 }, { "epoch": 0.020201048530614208, "grad_norm": 1.887161151439792, "learning_rate": 1.3461538461538463e-05, "loss": 1.2401, "step": 105 }, { "epoch": 0.02039343946900101, "grad_norm": 1.6763075137281496, "learning_rate": 1.3589743589743592e-05, "loss": 1.1772, "step": 106 }, { "epoch": 0.02058583040738781, "grad_norm": 1.7790067680368, "learning_rate": 1.3717948717948718e-05, "loss": 1.2369, "step": 107 }, { "epoch": 0.020778221345774613, "grad_norm": 1.834615413114737, "learning_rate": 1.3846153846153847e-05, "loss": 1.102, "step": 108 }, { "epoch": 0.020970612284161416, "grad_norm": 1.9135026265139552, "learning_rate": 1.3974358974358975e-05, "loss": 1.1528, "step": 109 }, { "epoch": 0.02116300322254822, "grad_norm": 1.805937170756496, "learning_rate": 1.4102564102564105e-05, "loss": 1.1588, "step": 110 }, { "epoch": 0.02135539416093502, "grad_norm": 1.894847928238082, "learning_rate": 1.4230769230769232e-05, "loss": 1.1751, "step": 111 }, { "epoch": 0.02154778509932182, "grad_norm": 1.5649396876014035, "learning_rate": 1.435897435897436e-05, "loss": 1.1746, "step": 112 }, { "epoch": 0.021740176037708624, "grad_norm": 1.8448182964851807, "learning_rate": 1.4487179487179489e-05, "loss": 1.0655, "step": 113 }, { "epoch": 0.021932566976095427, "grad_norm": 1.9404500935141968, "learning_rate": 1.4615384615384615e-05, "loss": 1.2331, "step": 114 }, { "epoch": 0.022124957914482227, "grad_norm": 1.783339270368585, "learning_rate": 1.4743589743589745e-05, "loss": 1.142, "step": 115 }, { "epoch": 0.02231734885286903, "grad_norm": 1.7165180750284237, "learning_rate": 1.4871794871794874e-05, "loss": 1.1819, "step": 116 }, { "epoch": 0.022509739791255833, "grad_norm": 1.8266509202712566, "learning_rate": 1.5000000000000002e-05, "loss": 1.1377, "step": 117 }, { "epoch": 0.022702130729642635, "grad_norm": 1.8654812043075542, "learning_rate": 1.5128205128205129e-05, "loss": 1.1473, "step": 118 }, { "epoch": 0.022894521668029435, "grad_norm": 1.9533074456707924, "learning_rate": 1.5256410256410257e-05, "loss": 1.1846, "step": 119 }, { "epoch": 0.023086912606416238, "grad_norm": 1.799152373918723, "learning_rate": 1.5384615384615387e-05, "loss": 1.1813, "step": 120 }, { "epoch": 0.02327930354480304, "grad_norm": 1.7909588606589146, "learning_rate": 1.5512820512820516e-05, "loss": 1.191, "step": 121 }, { "epoch": 0.02347169448318984, "grad_norm": 1.7838605814221162, "learning_rate": 1.5641025641025644e-05, "loss": 1.11, "step": 122 }, { "epoch": 0.023664085421576643, "grad_norm": 2.1477057735326976, "learning_rate": 1.576923076923077e-05, "loss": 1.1833, "step": 123 }, { "epoch": 0.023856476359963446, "grad_norm": 1.793826972701178, "learning_rate": 1.5897435897435897e-05, "loss": 1.2041, "step": 124 }, { "epoch": 0.02404886729835025, "grad_norm": 1.9298981599231941, "learning_rate": 1.602564102564103e-05, "loss": 1.171, "step": 125 }, { "epoch": 0.02424125823673705, "grad_norm": 1.8817471684596963, "learning_rate": 1.6153846153846154e-05, "loss": 1.1942, "step": 126 }, { "epoch": 0.02443364917512385, "grad_norm": 1.8590182865421865, "learning_rate": 1.6282051282051282e-05, "loss": 1.1525, "step": 127 }, { "epoch": 0.024626040113510654, "grad_norm": 1.691636768527045, "learning_rate": 1.641025641025641e-05, "loss": 1.1206, "step": 128 }, { "epoch": 0.024818431051897457, "grad_norm": 1.8084276542758917, "learning_rate": 1.653846153846154e-05, "loss": 1.1553, "step": 129 }, { "epoch": 0.025010821990284256, "grad_norm": 1.9130668121313184, "learning_rate": 1.6666666666666667e-05, "loss": 1.2104, "step": 130 }, { "epoch": 0.02520321292867106, "grad_norm": 1.752616214298338, "learning_rate": 1.6794871794871796e-05, "loss": 1.0907, "step": 131 }, { "epoch": 0.025395603867057862, "grad_norm": 1.9018514487228027, "learning_rate": 1.6923076923076924e-05, "loss": 1.2178, "step": 132 }, { "epoch": 0.025587994805444665, "grad_norm": 1.800275643488801, "learning_rate": 1.7051282051282053e-05, "loss": 1.178, "step": 133 }, { "epoch": 0.025780385743831465, "grad_norm": 1.819028354525209, "learning_rate": 1.717948717948718e-05, "loss": 1.1689, "step": 134 }, { "epoch": 0.025972776682218267, "grad_norm": 2.01869737872508, "learning_rate": 1.730769230769231e-05, "loss": 1.2199, "step": 135 }, { "epoch": 0.02616516762060507, "grad_norm": 1.360698440790348, "learning_rate": 1.7435897435897438e-05, "loss": 0.9037, "step": 136 }, { "epoch": 0.02635755855899187, "grad_norm": 1.2609284916014871, "learning_rate": 1.7564102564102566e-05, "loss": 0.8683, "step": 137 }, { "epoch": 0.026549949497378673, "grad_norm": 1.6185789257039382, "learning_rate": 1.7692307692307694e-05, "loss": 1.2012, "step": 138 }, { "epoch": 0.026742340435765476, "grad_norm": 1.817660120275698, "learning_rate": 1.7820512820512823e-05, "loss": 1.2087, "step": 139 }, { "epoch": 0.02693473137415228, "grad_norm": 1.8020101821926362, "learning_rate": 1.794871794871795e-05, "loss": 1.1842, "step": 140 }, { "epoch": 0.027127122312539078, "grad_norm": 1.6277276712135862, "learning_rate": 1.807692307692308e-05, "loss": 1.1272, "step": 141 }, { "epoch": 0.02731951325092588, "grad_norm": 2.392001464100763, "learning_rate": 1.8205128205128208e-05, "loss": 1.1561, "step": 142 }, { "epoch": 0.027511904189312684, "grad_norm": 1.7507780509454565, "learning_rate": 1.8333333333333333e-05, "loss": 1.0651, "step": 143 }, { "epoch": 0.027704295127699487, "grad_norm": 1.872192114731274, "learning_rate": 1.8461538461538465e-05, "loss": 1.1501, "step": 144 }, { "epoch": 0.027896686066086286, "grad_norm": 1.9860735961669016, "learning_rate": 1.8589743589743593e-05, "loss": 1.0919, "step": 145 }, { "epoch": 0.02808907700447309, "grad_norm": 1.7057407946770782, "learning_rate": 1.8717948717948718e-05, "loss": 1.0796, "step": 146 }, { "epoch": 0.028281467942859892, "grad_norm": 1.707285053453735, "learning_rate": 1.8846153846153846e-05, "loss": 1.1756, "step": 147 }, { "epoch": 0.028473858881246695, "grad_norm": 1.875995549223614, "learning_rate": 1.8974358974358975e-05, "loss": 1.0843, "step": 148 }, { "epoch": 0.028666249819633494, "grad_norm": 1.8579332246339644, "learning_rate": 1.9102564102564106e-05, "loss": 1.2316, "step": 149 }, { "epoch": 0.028858640758020297, "grad_norm": 3.4403289494492193, "learning_rate": 1.923076923076923e-05, "loss": 1.157, "step": 150 }, { "epoch": 0.0290510316964071, "grad_norm": 1.8898716041746375, "learning_rate": 1.935897435897436e-05, "loss": 1.0952, "step": 151 }, { "epoch": 0.0292434226347939, "grad_norm": 1.7146030018079543, "learning_rate": 1.9487179487179488e-05, "loss": 1.136, "step": 152 }, { "epoch": 0.029435813573180702, "grad_norm": 1.6981082297404997, "learning_rate": 1.9615384615384617e-05, "loss": 1.1609, "step": 153 }, { "epoch": 0.029628204511567505, "grad_norm": 1.7677924829767588, "learning_rate": 1.9743589743589745e-05, "loss": 1.1557, "step": 154 }, { "epoch": 0.02982059544995431, "grad_norm": 1.70411514034604, "learning_rate": 1.9871794871794873e-05, "loss": 1.1471, "step": 155 }, { "epoch": 0.030012986388341108, "grad_norm": 1.9538384739294636, "learning_rate": 2e-05, "loss": 1.1884, "step": 156 }, { "epoch": 0.03020537732672791, "grad_norm": 1.755446576171191, "learning_rate": 1.9999998058057616e-05, "loss": 1.158, "step": 157 }, { "epoch": 0.030397768265114714, "grad_norm": 1.7570445652315725, "learning_rate": 1.9999992232231216e-05, "loss": 1.1366, "step": 158 }, { "epoch": 0.030590159203501516, "grad_norm": 1.782900823163102, "learning_rate": 1.999998252252306e-05, "loss": 1.06, "step": 159 }, { "epoch": 0.030782550141888316, "grad_norm": 1.7812856210466579, "learning_rate": 1.9999968928936924e-05, "loss": 1.1767, "step": 160 }, { "epoch": 0.03097494108027512, "grad_norm": 1.6259844393641125, "learning_rate": 1.999995145147809e-05, "loss": 1.088, "step": 161 }, { "epoch": 0.03116733201866192, "grad_norm": 1.6499499349906308, "learning_rate": 1.9999930090153335e-05, "loss": 1.0494, "step": 162 }, { "epoch": 0.031359722957048725, "grad_norm": 1.8776847528808032, "learning_rate": 1.9999904844970963e-05, "loss": 1.1198, "step": 163 }, { "epoch": 0.031552113895435524, "grad_norm": 1.924298065275549, "learning_rate": 1.999987571594078e-05, "loss": 1.1568, "step": 164 }, { "epoch": 0.03174450483382233, "grad_norm": 1.8099072284653077, "learning_rate": 1.99998427030741e-05, "loss": 1.1826, "step": 165 }, { "epoch": 0.03193689577220913, "grad_norm": 1.6733312420012516, "learning_rate": 1.999980580638374e-05, "loss": 1.0785, "step": 166 }, { "epoch": 0.03212928671059593, "grad_norm": 1.7930270825478092, "learning_rate": 1.999976502588403e-05, "loss": 1.169, "step": 167 }, { "epoch": 0.032321677648982736, "grad_norm": 1.809873532668896, "learning_rate": 1.9999720361590812e-05, "loss": 1.1803, "step": 168 }, { "epoch": 0.032514068587369535, "grad_norm": 0.864928293452628, "learning_rate": 1.9999671813521435e-05, "loss": 0.8951, "step": 169 }, { "epoch": 0.032706459525756335, "grad_norm": 1.9065555724520393, "learning_rate": 1.999961938169475e-05, "loss": 1.1167, "step": 170 }, { "epoch": 0.03289885046414314, "grad_norm": 1.9313131434408026, "learning_rate": 1.9999563066131124e-05, "loss": 1.1341, "step": 171 }, { "epoch": 0.03309124140252994, "grad_norm": 1.837193814735097, "learning_rate": 1.9999502866852427e-05, "loss": 1.1661, "step": 172 }, { "epoch": 0.03328363234091674, "grad_norm": 1.8525168738006605, "learning_rate": 1.999943878388204e-05, "loss": 1.1129, "step": 173 }, { "epoch": 0.033476023279303546, "grad_norm": 1.8679610702227327, "learning_rate": 1.9999370817244853e-05, "loss": 1.1853, "step": 174 }, { "epoch": 0.033668414217690346, "grad_norm": 1.82474137450198, "learning_rate": 1.9999298966967264e-05, "loss": 1.1032, "step": 175 }, { "epoch": 0.03386080515607715, "grad_norm": 1.6782117198234074, "learning_rate": 1.9999223233077178e-05, "loss": 1.1733, "step": 176 }, { "epoch": 0.03405319609446395, "grad_norm": 1.6303062008396523, "learning_rate": 1.999914361560401e-05, "loss": 1.0962, "step": 177 }, { "epoch": 0.03424558703285075, "grad_norm": 1.6695822925832988, "learning_rate": 1.9999060114578682e-05, "loss": 1.1628, "step": 178 }, { "epoch": 0.03443797797123756, "grad_norm": 1.6851060252569636, "learning_rate": 1.9998972730033624e-05, "loss": 1.147, "step": 179 }, { "epoch": 0.03463036890962436, "grad_norm": 1.7154477920367113, "learning_rate": 1.9998881462002778e-05, "loss": 1.1445, "step": 180 }, { "epoch": 0.034822759848011156, "grad_norm": 1.6701136984403484, "learning_rate": 1.9998786310521585e-05, "loss": 1.1297, "step": 181 }, { "epoch": 0.03501515078639796, "grad_norm": 1.7978636767232463, "learning_rate": 1.9998687275627008e-05, "loss": 1.1135, "step": 182 }, { "epoch": 0.03520754172478476, "grad_norm": 1.728573539212089, "learning_rate": 1.9998584357357503e-05, "loss": 1.1612, "step": 183 }, { "epoch": 0.03539993266317156, "grad_norm": 1.7868619372820678, "learning_rate": 1.9998477555753054e-05, "loss": 1.2448, "step": 184 }, { "epoch": 0.03559232360155837, "grad_norm": 1.667606777204198, "learning_rate": 1.9998366870855134e-05, "loss": 1.0958, "step": 185 }, { "epoch": 0.03578471453994517, "grad_norm": 1.8367279410872854, "learning_rate": 1.999825230270673e-05, "loss": 1.0857, "step": 186 }, { "epoch": 0.035977105478331974, "grad_norm": 1.6343370981796232, "learning_rate": 1.9998133851352342e-05, "loss": 1.1776, "step": 187 }, { "epoch": 0.03616949641671877, "grad_norm": 1.7887109158739116, "learning_rate": 1.9998011516837974e-05, "loss": 1.1115, "step": 188 }, { "epoch": 0.03636188735510557, "grad_norm": 1.8981321516193486, "learning_rate": 1.999788529921114e-05, "loss": 1.1542, "step": 189 }, { "epoch": 0.03655427829349238, "grad_norm": 1.818293025219953, "learning_rate": 1.999775519852086e-05, "loss": 1.1093, "step": 190 }, { "epoch": 0.03674666923187918, "grad_norm": 1.6359255945611482, "learning_rate": 1.999762121481767e-05, "loss": 1.1039, "step": 191 }, { "epoch": 0.03693906017026598, "grad_norm": 1.7024667678238687, "learning_rate": 1.99974833481536e-05, "loss": 1.1497, "step": 192 }, { "epoch": 0.037131451108652784, "grad_norm": 1.814675158209907, "learning_rate": 1.9997341598582197e-05, "loss": 1.082, "step": 193 }, { "epoch": 0.03732384204703958, "grad_norm": 1.9435593187267395, "learning_rate": 1.9997195966158518e-05, "loss": 1.0956, "step": 194 }, { "epoch": 0.03751623298542639, "grad_norm": 1.595627633642553, "learning_rate": 1.9997046450939122e-05, "loss": 1.1453, "step": 195 }, { "epoch": 0.03770862392381319, "grad_norm": 1.9317920074916064, "learning_rate": 1.9996893052982083e-05, "loss": 1.1386, "step": 196 }, { "epoch": 0.03790101486219999, "grad_norm": 1.7850196600255248, "learning_rate": 1.9996735772346973e-05, "loss": 1.0433, "step": 197 }, { "epoch": 0.038093405800586795, "grad_norm": 1.8343097101448291, "learning_rate": 1.9996574609094887e-05, "loss": 1.1318, "step": 198 }, { "epoch": 0.038285796738973595, "grad_norm": 1.8996772913266364, "learning_rate": 1.9996409563288404e-05, "loss": 1.1788, "step": 199 }, { "epoch": 0.038478187677360394, "grad_norm": 1.947549902533559, "learning_rate": 1.9996240634991645e-05, "loss": 1.0491, "step": 200 }, { "epoch": 0.0386705786157472, "grad_norm": 1.8581331157702963, "learning_rate": 1.9996067824270204e-05, "loss": 1.1473, "step": 201 }, { "epoch": 0.038862969554134, "grad_norm": 1.6882485665169518, "learning_rate": 1.999589113119121e-05, "loss": 1.1497, "step": 202 }, { "epoch": 0.0390553604925208, "grad_norm": 1.8000227047432582, "learning_rate": 1.9995710555823277e-05, "loss": 1.2016, "step": 203 }, { "epoch": 0.039247751430907606, "grad_norm": 1.7846935282840164, "learning_rate": 1.999552609823655e-05, "loss": 1.1059, "step": 204 }, { "epoch": 0.039440142369294405, "grad_norm": 1.792954396769008, "learning_rate": 1.999533775850266e-05, "loss": 1.1061, "step": 205 }, { "epoch": 0.03963253330768121, "grad_norm": 1.761046599622354, "learning_rate": 1.9995145536694764e-05, "loss": 1.1569, "step": 206 }, { "epoch": 0.03982492424606801, "grad_norm": 0.9772077728036879, "learning_rate": 1.9994949432887512e-05, "loss": 0.8703, "step": 207 }, { "epoch": 0.04001731518445481, "grad_norm": 1.840054889010289, "learning_rate": 1.999474944715708e-05, "loss": 1.1039, "step": 208 }, { "epoch": 0.04020970612284162, "grad_norm": 1.838876838515895, "learning_rate": 1.9994545579581125e-05, "loss": 1.1381, "step": 209 }, { "epoch": 0.040402097061228416, "grad_norm": 1.68259167871146, "learning_rate": 1.9994337830238836e-05, "loss": 1.1769, "step": 210 }, { "epoch": 0.040594487999615216, "grad_norm": 1.7775367638106732, "learning_rate": 1.9994126199210897e-05, "loss": 1.1809, "step": 211 }, { "epoch": 0.04078687893800202, "grad_norm": 1.7050383884421774, "learning_rate": 1.999391068657951e-05, "loss": 1.0948, "step": 212 }, { "epoch": 0.04097926987638882, "grad_norm": 1.8313812163300809, "learning_rate": 1.9993691292428364e-05, "loss": 1.0523, "step": 213 }, { "epoch": 0.04117166081477562, "grad_norm": 1.7748285215469513, "learning_rate": 1.9993468016842684e-05, "loss": 1.1651, "step": 214 }, { "epoch": 0.04136405175316243, "grad_norm": 1.728804391921365, "learning_rate": 1.999324085990918e-05, "loss": 1.1584, "step": 215 }, { "epoch": 0.04155644269154923, "grad_norm": 1.7879220868832149, "learning_rate": 1.9993009821716076e-05, "loss": 1.0863, "step": 216 }, { "epoch": 0.04174883362993603, "grad_norm": 1.7830275646199585, "learning_rate": 1.9992774902353104e-05, "loss": 1.0892, "step": 217 }, { "epoch": 0.04194122456832283, "grad_norm": 1.705295500158426, "learning_rate": 1.999253610191151e-05, "loss": 1.1917, "step": 218 }, { "epoch": 0.04213361550670963, "grad_norm": 1.9705869758034644, "learning_rate": 1.999229342048404e-05, "loss": 1.1538, "step": 219 }, { "epoch": 0.04232600644509644, "grad_norm": 1.812002719964561, "learning_rate": 1.9992046858164942e-05, "loss": 1.1019, "step": 220 }, { "epoch": 0.04251839738348324, "grad_norm": 0.8879505406730074, "learning_rate": 1.999179641504999e-05, "loss": 0.8794, "step": 221 }, { "epoch": 0.04271078832187004, "grad_norm": 1.7055380092748138, "learning_rate": 1.9991542091236438e-05, "loss": 1.1109, "step": 222 }, { "epoch": 0.042903179260256843, "grad_norm": 1.862730988978415, "learning_rate": 1.9991283886823075e-05, "loss": 1.2603, "step": 223 }, { "epoch": 0.04309557019864364, "grad_norm": 1.9617073472734794, "learning_rate": 1.9991021801910177e-05, "loss": 1.1208, "step": 224 }, { "epoch": 0.04328796113703045, "grad_norm": 1.8621541252942821, "learning_rate": 1.999075583659954e-05, "loss": 1.1431, "step": 225 }, { "epoch": 0.04348035207541725, "grad_norm": 1.6982211965432938, "learning_rate": 1.999048599099446e-05, "loss": 1.1049, "step": 226 }, { "epoch": 0.04367274301380405, "grad_norm": 1.734144377696869, "learning_rate": 1.9990212265199738e-05, "loss": 1.1177, "step": 227 }, { "epoch": 0.043865133952190855, "grad_norm": 1.6806209271770016, "learning_rate": 1.998993465932169e-05, "loss": 1.1911, "step": 228 }, { "epoch": 0.044057524890577654, "grad_norm": 1.7578791864271754, "learning_rate": 1.9989653173468137e-05, "loss": 1.0976, "step": 229 }, { "epoch": 0.04424991582896445, "grad_norm": 1.9357782828307153, "learning_rate": 1.99893678077484e-05, "loss": 1.1418, "step": 230 }, { "epoch": 0.04444230676735126, "grad_norm": 1.7918432543677907, "learning_rate": 1.9989078562273313e-05, "loss": 1.153, "step": 231 }, { "epoch": 0.04463469770573806, "grad_norm": 1.6702046935677408, "learning_rate": 1.9988785437155222e-05, "loss": 1.0408, "step": 232 }, { "epoch": 0.04482708864412486, "grad_norm": 1.6431894796580215, "learning_rate": 1.9988488432507963e-05, "loss": 1.0836, "step": 233 }, { "epoch": 0.045019479582511665, "grad_norm": 1.8259771123744863, "learning_rate": 1.9988187548446895e-05, "loss": 1.0374, "step": 234 }, { "epoch": 0.045211870520898464, "grad_norm": 1.875136703343282, "learning_rate": 1.998788278508888e-05, "loss": 1.168, "step": 235 }, { "epoch": 0.04540426145928527, "grad_norm": 1.9989871484094068, "learning_rate": 1.9987574142552274e-05, "loss": 1.0963, "step": 236 }, { "epoch": 0.04559665239767207, "grad_norm": 1.735665359664887, "learning_rate": 1.9987261620956964e-05, "loss": 1.1146, "step": 237 }, { "epoch": 0.04578904333605887, "grad_norm": 1.8524929841282003, "learning_rate": 1.9986945220424326e-05, "loss": 1.1526, "step": 238 }, { "epoch": 0.045981434274445676, "grad_norm": 1.8572948218834517, "learning_rate": 1.998662494107724e-05, "loss": 1.1164, "step": 239 }, { "epoch": 0.046173825212832476, "grad_norm": 1.7928926944635277, "learning_rate": 1.99863007830401e-05, "loss": 1.108, "step": 240 }, { "epoch": 0.046366216151219275, "grad_norm": 1.701086082774569, "learning_rate": 1.9985972746438815e-05, "loss": 1.0649, "step": 241 }, { "epoch": 0.04655860708960608, "grad_norm": 1.6858330002405497, "learning_rate": 1.9985640831400778e-05, "loss": 1.1861, "step": 242 }, { "epoch": 0.04675099802799288, "grad_norm": 1.8307969812731244, "learning_rate": 1.998530503805491e-05, "loss": 1.1046, "step": 243 }, { "epoch": 0.04694338896637968, "grad_norm": 1.7039260289347749, "learning_rate": 1.9984965366531624e-05, "loss": 1.1698, "step": 244 }, { "epoch": 0.04713577990476649, "grad_norm": 1.6606476133952373, "learning_rate": 1.9984621816962843e-05, "loss": 1.1327, "step": 245 }, { "epoch": 0.047328170843153286, "grad_norm": 1.7502710658753098, "learning_rate": 1.9984274389482005e-05, "loss": 1.1098, "step": 246 }, { "epoch": 0.04752056178154009, "grad_norm": 1.7013513216620337, "learning_rate": 1.9983923084224047e-05, "loss": 1.0518, "step": 247 }, { "epoch": 0.04771295271992689, "grad_norm": 0.8650014160443987, "learning_rate": 1.9983567901325404e-05, "loss": 0.8683, "step": 248 }, { "epoch": 0.04790534365831369, "grad_norm": 1.6833514652880774, "learning_rate": 1.9983208840924028e-05, "loss": 1.1263, "step": 249 }, { "epoch": 0.0480977345967005, "grad_norm": 1.7045282842135472, "learning_rate": 1.998284590315937e-05, "loss": 1.173, "step": 250 }, { "epoch": 0.0482901255350873, "grad_norm": 1.6980013869050161, "learning_rate": 1.9982479088172403e-05, "loss": 1.1253, "step": 251 }, { "epoch": 0.0484825164734741, "grad_norm": 1.577981358397524, "learning_rate": 1.9982108396105584e-05, "loss": 1.0468, "step": 252 }, { "epoch": 0.0486749074118609, "grad_norm": 0.8387380343203026, "learning_rate": 1.9981733827102884e-05, "loss": 0.8749, "step": 253 }, { "epoch": 0.0488672983502477, "grad_norm": 1.7848092818812484, "learning_rate": 1.998135538130979e-05, "loss": 1.161, "step": 254 }, { "epoch": 0.04905968928863451, "grad_norm": 1.7261864427908635, "learning_rate": 1.998097305887328e-05, "loss": 1.1723, "step": 255 }, { "epoch": 0.04925208022702131, "grad_norm": 0.8487515992465713, "learning_rate": 1.9980586859941846e-05, "loss": 0.8769, "step": 256 }, { "epoch": 0.04944447116540811, "grad_norm": 0.7958926147173413, "learning_rate": 1.998019678466548e-05, "loss": 0.8776, "step": 257 }, { "epoch": 0.049636862103794914, "grad_norm": 0.7524053139523467, "learning_rate": 1.997980283319568e-05, "loss": 0.7637, "step": 258 }, { "epoch": 0.04982925304218171, "grad_norm": 1.933242403227823, "learning_rate": 1.9979405005685466e-05, "loss": 1.1036, "step": 259 }, { "epoch": 0.05002164398056851, "grad_norm": 1.7535154163762932, "learning_rate": 1.9979003302289336e-05, "loss": 1.156, "step": 260 }, { "epoch": 0.05021403491895532, "grad_norm": 1.5241609657325839, "learning_rate": 1.997859772316331e-05, "loss": 1.1242, "step": 261 }, { "epoch": 0.05040642585734212, "grad_norm": 1.7411171418678992, "learning_rate": 1.9978188268464912e-05, "loss": 1.0796, "step": 262 }, { "epoch": 0.05059881679572892, "grad_norm": 1.8755025516192252, "learning_rate": 1.997777493835317e-05, "loss": 1.0994, "step": 263 }, { "epoch": 0.050791207734115724, "grad_norm": 1.608972178963089, "learning_rate": 1.9977357732988616e-05, "loss": 1.047, "step": 264 }, { "epoch": 0.050983598672502524, "grad_norm": 1.6340109955203368, "learning_rate": 1.9976936652533288e-05, "loss": 1.1419, "step": 265 }, { "epoch": 0.05117598961088933, "grad_norm": 1.7460469039289643, "learning_rate": 1.997651169715073e-05, "loss": 1.0426, "step": 266 }, { "epoch": 0.05136838054927613, "grad_norm": 1.6667698464596121, "learning_rate": 1.9976082867005985e-05, "loss": 1.1441, "step": 267 }, { "epoch": 0.05156077148766293, "grad_norm": 1.6640744408604946, "learning_rate": 1.997565016226561e-05, "loss": 1.008, "step": 268 }, { "epoch": 0.051753162426049736, "grad_norm": 1.9577462439917643, "learning_rate": 1.997521358309766e-05, "loss": 1.1167, "step": 269 }, { "epoch": 0.051945553364436535, "grad_norm": 1.6313500304691848, "learning_rate": 1.99747731296717e-05, "loss": 1.1141, "step": 270 }, { "epoch": 0.052137944302823334, "grad_norm": 1.5709215621941617, "learning_rate": 1.9974328802158798e-05, "loss": 1.0391, "step": 271 }, { "epoch": 0.05233033524121014, "grad_norm": 1.7868699429588981, "learning_rate": 1.997388060073152e-05, "loss": 1.0674, "step": 272 }, { "epoch": 0.05252272617959694, "grad_norm": 1.5781387122138604, "learning_rate": 1.9973428525563948e-05, "loss": 1.1525, "step": 273 }, { "epoch": 0.05271511711798374, "grad_norm": 1.5998284148741961, "learning_rate": 1.9972972576831656e-05, "loss": 1.1097, "step": 274 }, { "epoch": 0.052907508056370546, "grad_norm": 1.773229331593496, "learning_rate": 1.9972512754711738e-05, "loss": 1.0862, "step": 275 }, { "epoch": 0.053099898994757345, "grad_norm": 1.7064664778706438, "learning_rate": 1.997204905938278e-05, "loss": 1.1535, "step": 276 }, { "epoch": 0.05329228993314415, "grad_norm": 1.510860284584126, "learning_rate": 1.9971581491024873e-05, "loss": 1.0795, "step": 277 }, { "epoch": 0.05348468087153095, "grad_norm": 1.7148115245451485, "learning_rate": 1.997111004981962e-05, "loss": 1.0934, "step": 278 }, { "epoch": 0.05367707180991775, "grad_norm": 1.7120819409198682, "learning_rate": 1.9970634735950117e-05, "loss": 1.0806, "step": 279 }, { "epoch": 0.05386946274830456, "grad_norm": 1.6842597955308214, "learning_rate": 1.9970155549600978e-05, "loss": 1.1253, "step": 280 }, { "epoch": 0.05406185368669136, "grad_norm": 1.559548764215912, "learning_rate": 1.9969672490958304e-05, "loss": 1.0323, "step": 281 }, { "epoch": 0.054254244625078156, "grad_norm": 1.7599560039773927, "learning_rate": 1.996918556020972e-05, "loss": 1.1156, "step": 282 }, { "epoch": 0.05444663556346496, "grad_norm": 1.687067205922634, "learning_rate": 1.996869475754434e-05, "loss": 1.1264, "step": 283 }, { "epoch": 0.05463902650185176, "grad_norm": 1.67213254019715, "learning_rate": 1.9968200083152784e-05, "loss": 1.0548, "step": 284 }, { "epoch": 0.05483141744023857, "grad_norm": 1.5546216162207025, "learning_rate": 1.9967701537227175e-05, "loss": 1.0643, "step": 285 }, { "epoch": 0.05502380837862537, "grad_norm": 1.566308822179121, "learning_rate": 1.996719911996115e-05, "loss": 1.1135, "step": 286 }, { "epoch": 0.05521619931701217, "grad_norm": 0.762928226656591, "learning_rate": 1.996669283154984e-05, "loss": 0.7676, "step": 287 }, { "epoch": 0.05540859025539897, "grad_norm": 1.5926164805065226, "learning_rate": 1.996618267218988e-05, "loss": 1.1371, "step": 288 }, { "epoch": 0.05560098119378577, "grad_norm": 0.8052732407784513, "learning_rate": 1.996566864207941e-05, "loss": 0.8463, "step": 289 }, { "epoch": 0.05579337213217257, "grad_norm": 1.5452797820635085, "learning_rate": 1.9965150741418072e-05, "loss": 1.0627, "step": 290 }, { "epoch": 0.05598576307055938, "grad_norm": 1.8061757218331242, "learning_rate": 1.9964628970407018e-05, "loss": 1.168, "step": 291 }, { "epoch": 0.05617815400894618, "grad_norm": 1.7025889337667304, "learning_rate": 1.9964103329248892e-05, "loss": 1.1132, "step": 292 }, { "epoch": 0.05637054494733298, "grad_norm": 1.636130125486776, "learning_rate": 1.996357381814785e-05, "loss": 1.182, "step": 293 }, { "epoch": 0.056562935885719784, "grad_norm": 1.7418066070167273, "learning_rate": 1.996304043730955e-05, "loss": 1.1658, "step": 294 }, { "epoch": 0.05675532682410658, "grad_norm": 1.5407714986591736, "learning_rate": 1.9962503186941143e-05, "loss": 1.1011, "step": 295 }, { "epoch": 0.05694771776249339, "grad_norm": 1.7815415395216665, "learning_rate": 1.9961962067251298e-05, "loss": 1.0434, "step": 296 }, { "epoch": 0.05714010870088019, "grad_norm": 1.8786325298032998, "learning_rate": 1.9961417078450177e-05, "loss": 1.1271, "step": 297 }, { "epoch": 0.05733249963926699, "grad_norm": 1.5793900701796861, "learning_rate": 1.996086822074945e-05, "loss": 1.0537, "step": 298 }, { "epoch": 0.057524890577653795, "grad_norm": 1.537696001987639, "learning_rate": 1.9960315494362286e-05, "loss": 1.1178, "step": 299 }, { "epoch": 0.057717281516040594, "grad_norm": 1.4004156977980802, "learning_rate": 1.9959758899503355e-05, "loss": 1.0204, "step": 300 }, { "epoch": 0.057909672454427394, "grad_norm": 1.5868613933706661, "learning_rate": 1.995919843638883e-05, "loss": 1.0573, "step": 301 }, { "epoch": 0.0581020633928142, "grad_norm": 1.6745674805221693, "learning_rate": 1.9958634105236395e-05, "loss": 1.1207, "step": 302 }, { "epoch": 0.058294454331201, "grad_norm": 1.8482128542415242, "learning_rate": 1.9958065906265228e-05, "loss": 1.181, "step": 303 }, { "epoch": 0.0584868452695878, "grad_norm": 1.6826377798309042, "learning_rate": 1.9957493839696013e-05, "loss": 1.0565, "step": 304 }, { "epoch": 0.058679236207974605, "grad_norm": 1.7334151984341004, "learning_rate": 1.9956917905750926e-05, "loss": 1.106, "step": 305 }, { "epoch": 0.058871627146361405, "grad_norm": 1.6691018138950187, "learning_rate": 1.995633810465366e-05, "loss": 1.1149, "step": 306 }, { "epoch": 0.05906401808474821, "grad_norm": 1.6539722529406384, "learning_rate": 1.99557544366294e-05, "loss": 1.127, "step": 307 }, { "epoch": 0.05925640902313501, "grad_norm": 1.5122892311806257, "learning_rate": 1.9955166901904838e-05, "loss": 1.0766, "step": 308 }, { "epoch": 0.05944879996152181, "grad_norm": 1.6239986284102597, "learning_rate": 1.9954575500708164e-05, "loss": 1.1113, "step": 309 }, { "epoch": 0.05964119089990862, "grad_norm": 1.6679384180655221, "learning_rate": 1.995398023326907e-05, "loss": 1.2119, "step": 310 }, { "epoch": 0.059833581838295416, "grad_norm": 1.7433233453619017, "learning_rate": 1.9953381099818756e-05, "loss": 1.1871, "step": 311 }, { "epoch": 0.060025972776682215, "grad_norm": 0.7978928000809967, "learning_rate": 1.9952778100589912e-05, "loss": 0.8423, "step": 312 }, { "epoch": 0.06021836371506902, "grad_norm": 1.5175405910509532, "learning_rate": 1.9952171235816747e-05, "loss": 1.0143, "step": 313 }, { "epoch": 0.06041075465345582, "grad_norm": 1.7195416584441428, "learning_rate": 1.9951560505734948e-05, "loss": 1.1095, "step": 314 }, { "epoch": 0.06060314559184263, "grad_norm": 1.619072957767398, "learning_rate": 1.9950945910581718e-05, "loss": 1.1185, "step": 315 }, { "epoch": 0.06079553653022943, "grad_norm": 1.6783799365244796, "learning_rate": 1.9950327450595766e-05, "loss": 1.1267, "step": 316 }, { "epoch": 0.060987927468616226, "grad_norm": 1.5727983508571761, "learning_rate": 1.9949705126017286e-05, "loss": 0.9497, "step": 317 }, { "epoch": 0.06118031840700303, "grad_norm": 1.5523444370924058, "learning_rate": 1.9949078937087988e-05, "loss": 1.0544, "step": 318 }, { "epoch": 0.06137270934538983, "grad_norm": 1.6759742782989033, "learning_rate": 1.994844888405107e-05, "loss": 1.1787, "step": 319 }, { "epoch": 0.06156510028377663, "grad_norm": 1.6901722033455862, "learning_rate": 1.9947814967151246e-05, "loss": 1.1452, "step": 320 }, { "epoch": 0.06175749122216344, "grad_norm": 1.7487789325968108, "learning_rate": 1.9947177186634716e-05, "loss": 1.1384, "step": 321 }, { "epoch": 0.06194988216055024, "grad_norm": 1.583278159738727, "learning_rate": 1.9946535542749187e-05, "loss": 1.0335, "step": 322 }, { "epoch": 0.06214227309893704, "grad_norm": 1.6921183866053526, "learning_rate": 1.9945890035743866e-05, "loss": 1.1275, "step": 323 }, { "epoch": 0.06233466403732384, "grad_norm": 0.8713728905579831, "learning_rate": 1.9945240665869465e-05, "loss": 0.9136, "step": 324 }, { "epoch": 0.06252705497571065, "grad_norm": 0.8191798382657365, "learning_rate": 1.9944587433378187e-05, "loss": 0.8659, "step": 325 }, { "epoch": 0.06271944591409745, "grad_norm": 1.571423984274342, "learning_rate": 1.994393033852374e-05, "loss": 1.0381, "step": 326 }, { "epoch": 0.06291183685248425, "grad_norm": 1.5851840462886717, "learning_rate": 1.9943269381561334e-05, "loss": 0.9654, "step": 327 }, { "epoch": 0.06310422779087105, "grad_norm": 1.6321781617926094, "learning_rate": 1.994260456274768e-05, "loss": 1.1587, "step": 328 }, { "epoch": 0.06329661872925785, "grad_norm": 1.773577560170391, "learning_rate": 1.9941935882340976e-05, "loss": 1.1139, "step": 329 }, { "epoch": 0.06348900966764466, "grad_norm": 1.6380430836506674, "learning_rate": 1.994126334060094e-05, "loss": 1.1136, "step": 330 }, { "epoch": 0.06368140060603146, "grad_norm": 1.6100599372811317, "learning_rate": 1.994058693778878e-05, "loss": 1.0061, "step": 331 }, { "epoch": 0.06387379154441826, "grad_norm": 1.7606250161324577, "learning_rate": 1.9939906674167192e-05, "loss": 1.1191, "step": 332 }, { "epoch": 0.06406618248280506, "grad_norm": 1.6694206556959363, "learning_rate": 1.993922255000039e-05, "loss": 1.1234, "step": 333 }, { "epoch": 0.06425857342119186, "grad_norm": 1.5220961956944252, "learning_rate": 1.993853456555408e-05, "loss": 1.0644, "step": 334 }, { "epoch": 0.06445096435957866, "grad_norm": 1.4497059360197637, "learning_rate": 1.9937842721095468e-05, "loss": 1.0235, "step": 335 }, { "epoch": 0.06464335529796547, "grad_norm": 1.5992632559911901, "learning_rate": 1.9937147016893257e-05, "loss": 1.1389, "step": 336 }, { "epoch": 0.06483574623635227, "grad_norm": 0.7966917149770965, "learning_rate": 1.9936447453217646e-05, "loss": 0.8426, "step": 337 }, { "epoch": 0.06502813717473907, "grad_norm": 1.68872370144609, "learning_rate": 1.9935744030340347e-05, "loss": 1.0877, "step": 338 }, { "epoch": 0.06522052811312587, "grad_norm": 1.8002003257805452, "learning_rate": 1.9935036748534555e-05, "loss": 1.0819, "step": 339 }, { "epoch": 0.06541291905151267, "grad_norm": 1.754454993321424, "learning_rate": 1.993432560807497e-05, "loss": 1.1774, "step": 340 }, { "epoch": 0.06560530998989948, "grad_norm": 1.6949156956828577, "learning_rate": 1.993361060923779e-05, "loss": 1.146, "step": 341 }, { "epoch": 0.06579770092828628, "grad_norm": 1.5747600507876345, "learning_rate": 1.9932891752300717e-05, "loss": 1.0816, "step": 342 }, { "epoch": 0.06599009186667308, "grad_norm": 1.6804095667389307, "learning_rate": 1.9932169037542947e-05, "loss": 1.1965, "step": 343 }, { "epoch": 0.06618248280505988, "grad_norm": 1.662294334822539, "learning_rate": 1.9931442465245164e-05, "loss": 1.1397, "step": 344 }, { "epoch": 0.06637487374344668, "grad_norm": 1.5974012143567102, "learning_rate": 1.9930712035689576e-05, "loss": 1.126, "step": 345 }, { "epoch": 0.06656726468183348, "grad_norm": 1.6541977558335939, "learning_rate": 1.992997774915986e-05, "loss": 1.1174, "step": 346 }, { "epoch": 0.06675965562022029, "grad_norm": 1.6142416582274428, "learning_rate": 1.992923960594121e-05, "loss": 1.0913, "step": 347 }, { "epoch": 0.06695204655860709, "grad_norm": 0.8165280076402156, "learning_rate": 1.9928497606320308e-05, "loss": 0.9132, "step": 348 }, { "epoch": 0.06714443749699389, "grad_norm": 0.7729442114240119, "learning_rate": 1.992775175058535e-05, "loss": 0.8586, "step": 349 }, { "epoch": 0.06733682843538069, "grad_norm": 1.7143130573962522, "learning_rate": 1.9927002039026002e-05, "loss": 1.0691, "step": 350 }, { "epoch": 0.06752921937376749, "grad_norm": 1.8428838885558987, "learning_rate": 1.9926248471933453e-05, "loss": 1.0541, "step": 351 }, { "epoch": 0.0677216103121543, "grad_norm": 1.7005966659552052, "learning_rate": 1.9925491049600382e-05, "loss": 1.1355, "step": 352 }, { "epoch": 0.0679140012505411, "grad_norm": 1.7696554070160964, "learning_rate": 1.9924729772320953e-05, "loss": 1.1542, "step": 353 }, { "epoch": 0.0681063921889279, "grad_norm": 1.469442438707683, "learning_rate": 1.9923964640390846e-05, "loss": 1.0704, "step": 354 }, { "epoch": 0.0682987831273147, "grad_norm": 1.6307349779430962, "learning_rate": 1.9923195654107227e-05, "loss": 1.1271, "step": 355 }, { "epoch": 0.0684911740657015, "grad_norm": 1.5101622216978898, "learning_rate": 1.992242281376876e-05, "loss": 1.1292, "step": 356 }, { "epoch": 0.0686835650040883, "grad_norm": 1.6759407275817033, "learning_rate": 1.9921646119675606e-05, "loss": 1.1446, "step": 357 }, { "epoch": 0.06887595594247511, "grad_norm": 1.670678163376096, "learning_rate": 1.9920865572129426e-05, "loss": 1.1266, "step": 358 }, { "epoch": 0.06906834688086191, "grad_norm": 1.6375624848333663, "learning_rate": 1.9920081171433377e-05, "loss": 1.1074, "step": 359 }, { "epoch": 0.06926073781924871, "grad_norm": 1.74340802580963, "learning_rate": 1.991929291789211e-05, "loss": 1.1825, "step": 360 }, { "epoch": 0.06945312875763551, "grad_norm": 1.491082128244184, "learning_rate": 1.9918500811811778e-05, "loss": 1.0609, "step": 361 }, { "epoch": 0.06964551969602231, "grad_norm": 1.6090058552168711, "learning_rate": 1.991770485350002e-05, "loss": 1.0215, "step": 362 }, { "epoch": 0.06983791063440913, "grad_norm": 1.7367570176198255, "learning_rate": 1.991690504326597e-05, "loss": 1.1564, "step": 363 }, { "epoch": 0.07003030157279592, "grad_norm": 0.8270942010887061, "learning_rate": 1.9916101381420285e-05, "loss": 0.8425, "step": 364 }, { "epoch": 0.07022269251118272, "grad_norm": 1.6752506019853053, "learning_rate": 1.9915293868275083e-05, "loss": 1.0584, "step": 365 }, { "epoch": 0.07041508344956952, "grad_norm": 1.938509136914827, "learning_rate": 1.9914482504143996e-05, "loss": 1.0353, "step": 366 }, { "epoch": 0.07060747438795632, "grad_norm": 0.8106840433434672, "learning_rate": 1.9913667289342147e-05, "loss": 0.8724, "step": 367 }, { "epoch": 0.07079986532634312, "grad_norm": 1.5355092758175588, "learning_rate": 1.991284822418616e-05, "loss": 0.9982, "step": 368 }, { "epoch": 0.07099225626472994, "grad_norm": 1.6218540079284545, "learning_rate": 1.9912025308994146e-05, "loss": 1.0824, "step": 369 }, { "epoch": 0.07118464720311674, "grad_norm": 1.612943051416389, "learning_rate": 1.9911198544085723e-05, "loss": 1.1372, "step": 370 }, { "epoch": 0.07137703814150353, "grad_norm": 1.4586150270184854, "learning_rate": 1.991036792978199e-05, "loss": 1.0088, "step": 371 }, { "epoch": 0.07156942907989033, "grad_norm": 0.7938525815010766, "learning_rate": 1.990953346640555e-05, "loss": 0.8887, "step": 372 }, { "epoch": 0.07176182001827713, "grad_norm": 1.6652050793875313, "learning_rate": 1.9908695154280496e-05, "loss": 1.2014, "step": 373 }, { "epoch": 0.07195421095666395, "grad_norm": 1.4973407360926279, "learning_rate": 1.9907852993732425e-05, "loss": 1.0991, "step": 374 }, { "epoch": 0.07214660189505075, "grad_norm": 1.6934650747516096, "learning_rate": 1.990700698508842e-05, "loss": 1.1485, "step": 375 }, { "epoch": 0.07233899283343755, "grad_norm": 1.6680943550024225, "learning_rate": 1.990615712867706e-05, "loss": 1.1573, "step": 376 }, { "epoch": 0.07253138377182435, "grad_norm": 1.6607842487743674, "learning_rate": 1.9905303424828418e-05, "loss": 1.0831, "step": 377 }, { "epoch": 0.07272377471021114, "grad_norm": 1.559895225529676, "learning_rate": 1.9904445873874068e-05, "loss": 1.0546, "step": 378 }, { "epoch": 0.07291616564859794, "grad_norm": 1.6214357787355018, "learning_rate": 1.9903584476147066e-05, "loss": 1.1601, "step": 379 }, { "epoch": 0.07310855658698476, "grad_norm": 1.6113844278452782, "learning_rate": 1.9902719231981975e-05, "loss": 1.1222, "step": 380 }, { "epoch": 0.07330094752537156, "grad_norm": 1.7393536946555008, "learning_rate": 1.9901850141714843e-05, "loss": 1.1518, "step": 381 }, { "epoch": 0.07349333846375836, "grad_norm": 1.7410198697145907, "learning_rate": 1.9900977205683213e-05, "loss": 1.1037, "step": 382 }, { "epoch": 0.07368572940214516, "grad_norm": 1.675477421176322, "learning_rate": 1.9900100424226124e-05, "loss": 1.1789, "step": 383 }, { "epoch": 0.07387812034053196, "grad_norm": 0.8329130394444415, "learning_rate": 1.9899219797684113e-05, "loss": 0.8012, "step": 384 }, { "epoch": 0.07407051127891877, "grad_norm": 1.5929892935935157, "learning_rate": 1.98983353263992e-05, "loss": 1.0382, "step": 385 }, { "epoch": 0.07426290221730557, "grad_norm": 1.5512378107229137, "learning_rate": 1.9897447010714905e-05, "loss": 1.0287, "step": 386 }, { "epoch": 0.07445529315569237, "grad_norm": 1.479404962094613, "learning_rate": 1.989655485097624e-05, "loss": 1.1251, "step": 387 }, { "epoch": 0.07464768409407917, "grad_norm": 1.5535708845529694, "learning_rate": 1.989565884752971e-05, "loss": 1.0887, "step": 388 }, { "epoch": 0.07484007503246597, "grad_norm": 1.5766511675800903, "learning_rate": 1.9894759000723308e-05, "loss": 1.1304, "step": 389 }, { "epoch": 0.07503246597085278, "grad_norm": 1.5499251636703317, "learning_rate": 1.9893855310906526e-05, "loss": 1.0255, "step": 390 }, { "epoch": 0.07522485690923958, "grad_norm": 1.4510394024555313, "learning_rate": 1.9892947778430352e-05, "loss": 1.0364, "step": 391 }, { "epoch": 0.07541724784762638, "grad_norm": 1.838970646740351, "learning_rate": 1.9892036403647256e-05, "loss": 1.0982, "step": 392 }, { "epoch": 0.07560963878601318, "grad_norm": 1.616275422824502, "learning_rate": 1.9891121186911207e-05, "loss": 1.091, "step": 393 }, { "epoch": 0.07580202972439998, "grad_norm": 1.5350870980235092, "learning_rate": 1.9890202128577664e-05, "loss": 1.067, "step": 394 }, { "epoch": 0.07599442066278678, "grad_norm": 1.4410373908679792, "learning_rate": 1.988927922900358e-05, "loss": 1.111, "step": 395 }, { "epoch": 0.07618681160117359, "grad_norm": 1.745975593807465, "learning_rate": 1.9888352488547397e-05, "loss": 1.1465, "step": 396 }, { "epoch": 0.07637920253956039, "grad_norm": 1.833197230708042, "learning_rate": 1.988742190756905e-05, "loss": 1.0607, "step": 397 }, { "epoch": 0.07657159347794719, "grad_norm": 1.5292190691496872, "learning_rate": 1.9886487486429966e-05, "loss": 1.1718, "step": 398 }, { "epoch": 0.07676398441633399, "grad_norm": 1.5082909233264947, "learning_rate": 1.9885549225493064e-05, "loss": 1.1314, "step": 399 }, { "epoch": 0.07695637535472079, "grad_norm": 1.7126865836556926, "learning_rate": 1.9884607125122753e-05, "loss": 1.0899, "step": 400 }, { "epoch": 0.0771487662931076, "grad_norm": 1.5375897527242126, "learning_rate": 1.988366118568494e-05, "loss": 1.094, "step": 401 }, { "epoch": 0.0773411572314944, "grad_norm": 1.7859592124340835, "learning_rate": 1.988271140754701e-05, "loss": 1.0923, "step": 402 }, { "epoch": 0.0775335481698812, "grad_norm": 1.515594999940806, "learning_rate": 1.9881757791077848e-05, "loss": 1.0361, "step": 403 }, { "epoch": 0.077725939108268, "grad_norm": 1.8126268115824689, "learning_rate": 1.9880800336647825e-05, "loss": 1.1616, "step": 404 }, { "epoch": 0.0779183300466548, "grad_norm": 1.6353509637014088, "learning_rate": 1.987983904462881e-05, "loss": 1.0254, "step": 405 }, { "epoch": 0.0781107209850416, "grad_norm": 1.68198684742007, "learning_rate": 1.9878873915394154e-05, "loss": 1.1622, "step": 406 }, { "epoch": 0.07830311192342841, "grad_norm": 1.6833857587355576, "learning_rate": 1.9877904949318704e-05, "loss": 1.1294, "step": 407 }, { "epoch": 0.07849550286181521, "grad_norm": 1.7136942978639933, "learning_rate": 1.9876932146778796e-05, "loss": 1.0118, "step": 408 }, { "epoch": 0.07868789380020201, "grad_norm": 1.4611474885621862, "learning_rate": 1.9875955508152254e-05, "loss": 1.0876, "step": 409 }, { "epoch": 0.07888028473858881, "grad_norm": 1.9486731403246937, "learning_rate": 1.987497503381839e-05, "loss": 1.1149, "step": 410 }, { "epoch": 0.07907267567697561, "grad_norm": 1.65185606686191, "learning_rate": 1.9873990724158014e-05, "loss": 1.1437, "step": 411 }, { "epoch": 0.07926506661536242, "grad_norm": 1.672348765782603, "learning_rate": 1.987300257955342e-05, "loss": 1.0344, "step": 412 }, { "epoch": 0.07945745755374922, "grad_norm": 1.6311551292995257, "learning_rate": 1.987201060038839e-05, "loss": 1.1405, "step": 413 }, { "epoch": 0.07964984849213602, "grad_norm": 1.6516725975423914, "learning_rate": 1.9871014787048197e-05, "loss": 1.1228, "step": 414 }, { "epoch": 0.07984223943052282, "grad_norm": 1.6671965437858707, "learning_rate": 1.9870015139919606e-05, "loss": 1.0957, "step": 415 }, { "epoch": 0.08003463036890962, "grad_norm": 1.7436407625308992, "learning_rate": 1.9869011659390866e-05, "loss": 1.1084, "step": 416 }, { "epoch": 0.08022702130729642, "grad_norm": 1.588593571746427, "learning_rate": 1.9868004345851716e-05, "loss": 1.0941, "step": 417 }, { "epoch": 0.08041941224568323, "grad_norm": 1.5097061429805656, "learning_rate": 1.9866993199693393e-05, "loss": 1.0626, "step": 418 }, { "epoch": 0.08061180318407003, "grad_norm": 0.8287565114734501, "learning_rate": 1.98659782213086e-05, "loss": 0.8356, "step": 419 }, { "epoch": 0.08080419412245683, "grad_norm": 1.874989329488625, "learning_rate": 1.986495941109156e-05, "loss": 1.192, "step": 420 }, { "epoch": 0.08099658506084363, "grad_norm": 1.7073800067215992, "learning_rate": 1.9863936769437956e-05, "loss": 1.0899, "step": 421 }, { "epoch": 0.08118897599923043, "grad_norm": 1.5845792486411263, "learning_rate": 1.986291029674497e-05, "loss": 1.1287, "step": 422 }, { "epoch": 0.08138136693761724, "grad_norm": 1.5602978173506343, "learning_rate": 1.986187999341128e-05, "loss": 1.0187, "step": 423 }, { "epoch": 0.08157375787600404, "grad_norm": 1.5692289439267835, "learning_rate": 1.9860845859837034e-05, "loss": 1.023, "step": 424 }, { "epoch": 0.08176614881439084, "grad_norm": 0.8111823371290139, "learning_rate": 1.985980789642388e-05, "loss": 0.8726, "step": 425 }, { "epoch": 0.08195853975277764, "grad_norm": 1.6634122671506413, "learning_rate": 1.985876610357496e-05, "loss": 1.0369, "step": 426 }, { "epoch": 0.08215093069116444, "grad_norm": 1.6298950199839592, "learning_rate": 1.9857720481694887e-05, "loss": 1.136, "step": 427 }, { "epoch": 0.08234332162955124, "grad_norm": 1.6925355288786392, "learning_rate": 1.9856671031189765e-05, "loss": 1.1462, "step": 428 }, { "epoch": 0.08253571256793805, "grad_norm": 1.419923539845589, "learning_rate": 1.98556177524672e-05, "loss": 1.0893, "step": 429 }, { "epoch": 0.08272810350632485, "grad_norm": 1.5895272878404043, "learning_rate": 1.9854560645936262e-05, "loss": 1.1012, "step": 430 }, { "epoch": 0.08292049444471165, "grad_norm": 1.7961885292015582, "learning_rate": 1.9853499712007523e-05, "loss": 1.1296, "step": 431 }, { "epoch": 0.08311288538309845, "grad_norm": 1.5439158348060376, "learning_rate": 1.9852434951093035e-05, "loss": 1.1113, "step": 432 }, { "epoch": 0.08330527632148525, "grad_norm": 1.5728965445661418, "learning_rate": 1.985136636360635e-05, "loss": 1.1304, "step": 433 }, { "epoch": 0.08349766725987207, "grad_norm": 1.5756250178570221, "learning_rate": 1.985029394996248e-05, "loss": 1.0449, "step": 434 }, { "epoch": 0.08369005819825887, "grad_norm": 1.555957039604662, "learning_rate": 1.9849217710577945e-05, "loss": 1.1135, "step": 435 }, { "epoch": 0.08388244913664566, "grad_norm": 0.7819355303355624, "learning_rate": 1.9848137645870745e-05, "loss": 0.8908, "step": 436 }, { "epoch": 0.08407484007503246, "grad_norm": 1.6163797048509785, "learning_rate": 1.9847053756260363e-05, "loss": 1.0536, "step": 437 }, { "epoch": 0.08426723101341926, "grad_norm": 1.5855570719275534, "learning_rate": 1.984596604216777e-05, "loss": 1.1087, "step": 438 }, { "epoch": 0.08445962195180606, "grad_norm": 1.6095717453385394, "learning_rate": 1.984487450401542e-05, "loss": 1.0749, "step": 439 }, { "epoch": 0.08465201289019288, "grad_norm": 1.5242346408910767, "learning_rate": 1.9843779142227258e-05, "loss": 1.1213, "step": 440 }, { "epoch": 0.08484440382857968, "grad_norm": 1.4751560091804796, "learning_rate": 1.9842679957228706e-05, "loss": 1.0805, "step": 441 }, { "epoch": 0.08503679476696648, "grad_norm": 1.5607457550130048, "learning_rate": 1.9841576949446675e-05, "loss": 1.1608, "step": 442 }, { "epoch": 0.08522918570535327, "grad_norm": 1.5378799396502496, "learning_rate": 1.984047011930956e-05, "loss": 1.083, "step": 443 }, { "epoch": 0.08542157664374007, "grad_norm": 1.6270068379359262, "learning_rate": 1.9839359467247243e-05, "loss": 1.0869, "step": 444 }, { "epoch": 0.08561396758212689, "grad_norm": 1.7003755500757285, "learning_rate": 1.983824499369109e-05, "loss": 1.0178, "step": 445 }, { "epoch": 0.08580635852051369, "grad_norm": 1.5668506411147258, "learning_rate": 1.9837126699073948e-05, "loss": 1.1574, "step": 446 }, { "epoch": 0.08599874945890049, "grad_norm": 1.7057497115883145, "learning_rate": 1.9836004583830146e-05, "loss": 1.1358, "step": 447 }, { "epoch": 0.08619114039728729, "grad_norm": 1.5339243699884866, "learning_rate": 1.9834878648395507e-05, "loss": 1.0977, "step": 448 }, { "epoch": 0.08638353133567409, "grad_norm": 1.5682130955259999, "learning_rate": 1.9833748893207326e-05, "loss": 1.1461, "step": 449 }, { "epoch": 0.0865759222740609, "grad_norm": 1.583114298748472, "learning_rate": 1.9832615318704388e-05, "loss": 1.0897, "step": 450 }, { "epoch": 0.0867683132124477, "grad_norm": 1.528997698337866, "learning_rate": 1.9831477925326962e-05, "loss": 1.025, "step": 451 }, { "epoch": 0.0869607041508345, "grad_norm": 1.4919148933472652, "learning_rate": 1.98303367135168e-05, "loss": 1.097, "step": 452 }, { "epoch": 0.0871530950892213, "grad_norm": 1.745176487179496, "learning_rate": 1.9829191683717133e-05, "loss": 1.1374, "step": 453 }, { "epoch": 0.0873454860276081, "grad_norm": 0.8027574480122367, "learning_rate": 1.9828042836372677e-05, "loss": 0.8394, "step": 454 }, { "epoch": 0.0875378769659949, "grad_norm": 0.7657243608022711, "learning_rate": 1.9826890171929634e-05, "loss": 0.9127, "step": 455 }, { "epoch": 0.08773026790438171, "grad_norm": 1.665395215674917, "learning_rate": 1.982573369083568e-05, "loss": 1.0818, "step": 456 }, { "epoch": 0.08792265884276851, "grad_norm": 1.6163279181130235, "learning_rate": 1.9824573393539984e-05, "loss": 1.0139, "step": 457 }, { "epoch": 0.08811504978115531, "grad_norm": 1.8161674980824332, "learning_rate": 1.982340928049319e-05, "loss": 1.1117, "step": 458 }, { "epoch": 0.08830744071954211, "grad_norm": 1.6514827302913233, "learning_rate": 1.9822241352147426e-05, "loss": 1.1943, "step": 459 }, { "epoch": 0.0884998316579289, "grad_norm": 1.5322294268008614, "learning_rate": 1.9821069608956307e-05, "loss": 1.1367, "step": 460 }, { "epoch": 0.08869222259631572, "grad_norm": 1.7657150617567232, "learning_rate": 1.9819894051374917e-05, "loss": 1.0915, "step": 461 }, { "epoch": 0.08888461353470252, "grad_norm": 1.7378084785959476, "learning_rate": 1.981871467985983e-05, "loss": 1.1021, "step": 462 }, { "epoch": 0.08907700447308932, "grad_norm": 1.7204447990608556, "learning_rate": 1.9817531494869105e-05, "loss": 1.0914, "step": 463 }, { "epoch": 0.08926939541147612, "grad_norm": 1.5514063153470394, "learning_rate": 1.9816344496862272e-05, "loss": 1.0548, "step": 464 }, { "epoch": 0.08946178634986292, "grad_norm": 1.6012777020963498, "learning_rate": 1.9815153686300352e-05, "loss": 1.0909, "step": 465 }, { "epoch": 0.08965417728824972, "grad_norm": 0.73191776818113, "learning_rate": 1.981395906364584e-05, "loss": 0.8175, "step": 466 }, { "epoch": 0.08984656822663653, "grad_norm": 1.663048571855859, "learning_rate": 1.9812760629362714e-05, "loss": 1.1283, "step": 467 }, { "epoch": 0.09003895916502333, "grad_norm": 1.7845010247894173, "learning_rate": 1.981155838391643e-05, "loss": 1.0957, "step": 468 }, { "epoch": 0.09023135010341013, "grad_norm": 1.5299064596197585, "learning_rate": 1.9810352327773935e-05, "loss": 1.0574, "step": 469 }, { "epoch": 0.09042374104179693, "grad_norm": 1.5295305553856127, "learning_rate": 1.9809142461403635e-05, "loss": 1.1212, "step": 470 }, { "epoch": 0.09061613198018373, "grad_norm": 1.6684853907752544, "learning_rate": 1.9807928785275433e-05, "loss": 1.0901, "step": 471 }, { "epoch": 0.09080852291857054, "grad_norm": 1.5875810018021976, "learning_rate": 1.980671129986071e-05, "loss": 1.0965, "step": 472 }, { "epoch": 0.09100091385695734, "grad_norm": 1.6648352955572006, "learning_rate": 1.9805490005632323e-05, "loss": 1.0716, "step": 473 }, { "epoch": 0.09119330479534414, "grad_norm": 1.550427747929866, "learning_rate": 1.98042649030646e-05, "loss": 1.137, "step": 474 }, { "epoch": 0.09138569573373094, "grad_norm": 1.4455033920785822, "learning_rate": 1.9803035992633366e-05, "loss": 1.0252, "step": 475 }, { "epoch": 0.09157808667211774, "grad_norm": 1.5134227017410835, "learning_rate": 1.9801803274815915e-05, "loss": 1.104, "step": 476 }, { "epoch": 0.09177047761050454, "grad_norm": 1.6355273360899945, "learning_rate": 1.9800566750091018e-05, "loss": 1.0055, "step": 477 }, { "epoch": 0.09196286854889135, "grad_norm": 1.7041987656360573, "learning_rate": 1.9799326418938924e-05, "loss": 1.0982, "step": 478 }, { "epoch": 0.09215525948727815, "grad_norm": 1.6071590184905755, "learning_rate": 1.979808228184137e-05, "loss": 1.0745, "step": 479 }, { "epoch": 0.09234765042566495, "grad_norm": 1.7605922834891587, "learning_rate": 1.9796834339281557e-05, "loss": 1.1335, "step": 480 }, { "epoch": 0.09254004136405175, "grad_norm": 1.7392892510979492, "learning_rate": 1.979558259174418e-05, "loss": 1.1104, "step": 481 }, { "epoch": 0.09273243230243855, "grad_norm": 0.799367712359431, "learning_rate": 1.9794327039715395e-05, "loss": 0.8637, "step": 482 }, { "epoch": 0.09292482324082536, "grad_norm": 1.7275909192422816, "learning_rate": 1.979306768368285e-05, "loss": 1.1134, "step": 483 }, { "epoch": 0.09311721417921216, "grad_norm": 1.5843826150129907, "learning_rate": 1.9791804524135663e-05, "loss": 1.0945, "step": 484 }, { "epoch": 0.09330960511759896, "grad_norm": 1.6774243685906585, "learning_rate": 1.979053756156443e-05, "loss": 1.0389, "step": 485 }, { "epoch": 0.09350199605598576, "grad_norm": 0.7648146054422538, "learning_rate": 1.9789266796461222e-05, "loss": 0.8431, "step": 486 }, { "epoch": 0.09369438699437256, "grad_norm": 1.6067533046103935, "learning_rate": 1.9787992229319594e-05, "loss": 1.1251, "step": 487 }, { "epoch": 0.09388677793275936, "grad_norm": 1.6959235173077816, "learning_rate": 1.978671386063457e-05, "loss": 1.0922, "step": 488 }, { "epoch": 0.09407916887114617, "grad_norm": 1.3800902599135796, "learning_rate": 1.978543169090265e-05, "loss": 1.0313, "step": 489 }, { "epoch": 0.09427155980953297, "grad_norm": 1.748896549147447, "learning_rate": 1.9784145720621827e-05, "loss": 1.1182, "step": 490 }, { "epoch": 0.09446395074791977, "grad_norm": 1.4846254654576254, "learning_rate": 1.9782855950291542e-05, "loss": 1.1394, "step": 491 }, { "epoch": 0.09465634168630657, "grad_norm": 1.5685640788892097, "learning_rate": 1.978156238041274e-05, "loss": 1.0881, "step": 492 }, { "epoch": 0.09484873262469337, "grad_norm": 1.471050721778769, "learning_rate": 1.9780265011487822e-05, "loss": 1.1678, "step": 493 }, { "epoch": 0.09504112356308018, "grad_norm": 1.6565366776887211, "learning_rate": 1.9778963844020668e-05, "loss": 1.1094, "step": 494 }, { "epoch": 0.09523351450146698, "grad_norm": 1.4735105729870293, "learning_rate": 1.977765887851664e-05, "loss": 1.0643, "step": 495 }, { "epoch": 0.09542590543985378, "grad_norm": 1.5792347293820075, "learning_rate": 1.977635011548257e-05, "loss": 1.1118, "step": 496 }, { "epoch": 0.09561829637824058, "grad_norm": 1.6471994623014026, "learning_rate": 1.9775037555426772e-05, "loss": 1.0713, "step": 497 }, { "epoch": 0.09581068731662738, "grad_norm": 0.8038268715199373, "learning_rate": 1.9773721198859024e-05, "loss": 0.8501, "step": 498 }, { "epoch": 0.09600307825501418, "grad_norm": 1.5155774329321483, "learning_rate": 1.9772401046290584e-05, "loss": 1.095, "step": 499 }, { "epoch": 0.096195469193401, "grad_norm": 1.5981052886401215, "learning_rate": 1.9771077098234187e-05, "loss": 1.0623, "step": 500 }, { "epoch": 0.0963878601317878, "grad_norm": 1.5734191057749545, "learning_rate": 1.9769749355204034e-05, "loss": 1.068, "step": 501 }, { "epoch": 0.0965802510701746, "grad_norm": 1.6253252621195515, "learning_rate": 1.976841781771581e-05, "loss": 1.0862, "step": 502 }, { "epoch": 0.0967726420085614, "grad_norm": 1.5952002800945937, "learning_rate": 1.9767082486286667e-05, "loss": 1.0841, "step": 503 }, { "epoch": 0.0969650329469482, "grad_norm": 1.6141289405713546, "learning_rate": 1.9765743361435234e-05, "loss": 1.1563, "step": 504 }, { "epoch": 0.097157423885335, "grad_norm": 1.5325673898120205, "learning_rate": 1.9764400443681607e-05, "loss": 1.0571, "step": 505 }, { "epoch": 0.0973498148237218, "grad_norm": 1.685945829672869, "learning_rate": 1.9763053733547367e-05, "loss": 1.1138, "step": 506 }, { "epoch": 0.0975422057621086, "grad_norm": 1.5055641405812477, "learning_rate": 1.976170323155555e-05, "loss": 1.0704, "step": 507 }, { "epoch": 0.0977345967004954, "grad_norm": 1.5219049504337634, "learning_rate": 1.976034893823069e-05, "loss": 1.0285, "step": 508 }, { "epoch": 0.0979269876388822, "grad_norm": 0.8801762878099432, "learning_rate": 1.975899085409876e-05, "loss": 0.9086, "step": 509 }, { "epoch": 0.09811937857726902, "grad_norm": 1.7831852670916235, "learning_rate": 1.9757628979687247e-05, "loss": 1.1491, "step": 510 }, { "epoch": 0.09831176951565582, "grad_norm": 1.6795909865667955, "learning_rate": 1.975626331552507e-05, "loss": 1.0816, "step": 511 }, { "epoch": 0.09850416045404262, "grad_norm": 1.452701895968579, "learning_rate": 1.9754893862142643e-05, "loss": 1.115, "step": 512 }, { "epoch": 0.09869655139242942, "grad_norm": 1.5679055952994008, "learning_rate": 1.9753520620071846e-05, "loss": 0.9992, "step": 513 }, { "epoch": 0.09888894233081622, "grad_norm": 1.8637214561472175, "learning_rate": 1.9752143589846027e-05, "loss": 1.0285, "step": 514 }, { "epoch": 0.09908133326920301, "grad_norm": 1.686382969120432, "learning_rate": 1.9750762772000014e-05, "loss": 1.1521, "step": 515 }, { "epoch": 0.09927372420758983, "grad_norm": 1.5748844237544477, "learning_rate": 1.9749378167070097e-05, "loss": 1.0893, "step": 516 }, { "epoch": 0.09946611514597663, "grad_norm": 1.7505916419915812, "learning_rate": 1.9747989775594044e-05, "loss": 1.1042, "step": 517 }, { "epoch": 0.09965850608436343, "grad_norm": 1.6228528155958506, "learning_rate": 1.974659759811109e-05, "loss": 1.0633, "step": 518 }, { "epoch": 0.09985089702275023, "grad_norm": 1.708788120001719, "learning_rate": 1.9745201635161938e-05, "loss": 1.1843, "step": 519 }, { "epoch": 0.10004328796113703, "grad_norm": 1.7454388493313868, "learning_rate": 1.9743801887288762e-05, "loss": 1.0968, "step": 520 }, { "epoch": 0.10023567889952384, "grad_norm": 1.5686953721628265, "learning_rate": 1.9742398355035212e-05, "loss": 1.0924, "step": 521 }, { "epoch": 0.10042806983791064, "grad_norm": 1.5160981593646703, "learning_rate": 1.9740991038946404e-05, "loss": 1.0037, "step": 522 }, { "epoch": 0.10062046077629744, "grad_norm": 1.606927189300518, "learning_rate": 1.973957993956892e-05, "loss": 1.0913, "step": 523 }, { "epoch": 0.10081285171468424, "grad_norm": 1.6333364050531254, "learning_rate": 1.9738165057450817e-05, "loss": 1.0579, "step": 524 }, { "epoch": 0.10100524265307104, "grad_norm": 1.5055594524426983, "learning_rate": 1.9736746393141617e-05, "loss": 1.0488, "step": 525 }, { "epoch": 0.10119763359145784, "grad_norm": 1.5850442119446149, "learning_rate": 1.9735323947192317e-05, "loss": 1.1785, "step": 526 }, { "epoch": 0.10139002452984465, "grad_norm": 1.732193746716268, "learning_rate": 1.9733897720155377e-05, "loss": 1.0978, "step": 527 }, { "epoch": 0.10158241546823145, "grad_norm": 1.378627523006674, "learning_rate": 1.9732467712584723e-05, "loss": 1.0526, "step": 528 }, { "epoch": 0.10177480640661825, "grad_norm": 1.56310950691073, "learning_rate": 1.973103392503576e-05, "loss": 1.0238, "step": 529 }, { "epoch": 0.10196719734500505, "grad_norm": 1.5637841856200838, "learning_rate": 1.9729596358065347e-05, "loss": 1.1349, "step": 530 }, { "epoch": 0.10215958828339185, "grad_norm": 1.6349236739385073, "learning_rate": 1.9728155012231825e-05, "loss": 1.0978, "step": 531 }, { "epoch": 0.10235197922177866, "grad_norm": 1.4177988258499936, "learning_rate": 1.9726709888094994e-05, "loss": 1.098, "step": 532 }, { "epoch": 0.10254437016016546, "grad_norm": 1.6469031579719793, "learning_rate": 1.972526098621612e-05, "loss": 1.046, "step": 533 }, { "epoch": 0.10273676109855226, "grad_norm": 1.7474587430676558, "learning_rate": 1.972380830715795e-05, "loss": 1.0493, "step": 534 }, { "epoch": 0.10292915203693906, "grad_norm": 1.6802161653802248, "learning_rate": 1.9722351851484677e-05, "loss": 1.0573, "step": 535 }, { "epoch": 0.10312154297532586, "grad_norm": 1.6086682969361632, "learning_rate": 1.9720891619761974e-05, "loss": 1.0687, "step": 536 }, { "epoch": 0.10331393391371266, "grad_norm": 1.5370258309652889, "learning_rate": 1.9719427612556982e-05, "loss": 1.1295, "step": 537 }, { "epoch": 0.10350632485209947, "grad_norm": 1.5659429613470524, "learning_rate": 1.9717959830438302e-05, "loss": 1.0313, "step": 538 }, { "epoch": 0.10369871579048627, "grad_norm": 1.6766288359846353, "learning_rate": 1.9716488273976006e-05, "loss": 1.0616, "step": 539 }, { "epoch": 0.10389110672887307, "grad_norm": 1.5440806632603619, "learning_rate": 1.971501294374162e-05, "loss": 1.0525, "step": 540 }, { "epoch": 0.10408349766725987, "grad_norm": 1.529673713311215, "learning_rate": 1.971353384030816e-05, "loss": 1.0349, "step": 541 }, { "epoch": 0.10427588860564667, "grad_norm": 1.5264727666981908, "learning_rate": 1.9712050964250083e-05, "loss": 1.1283, "step": 542 }, { "epoch": 0.10446827954403348, "grad_norm": 1.4231664505212454, "learning_rate": 1.9710564316143323e-05, "loss": 1.144, "step": 543 }, { "epoch": 0.10466067048242028, "grad_norm": 1.6036783516795112, "learning_rate": 1.9709073896565276e-05, "loss": 1.1135, "step": 544 }, { "epoch": 0.10485306142080708, "grad_norm": 1.5937290667256547, "learning_rate": 1.9707579706094807e-05, "loss": 1.1125, "step": 545 }, { "epoch": 0.10504545235919388, "grad_norm": 1.5700145027184211, "learning_rate": 1.970608174531224e-05, "loss": 1.1059, "step": 546 }, { "epoch": 0.10523784329758068, "grad_norm": 1.5152664219277434, "learning_rate": 1.970458001479937e-05, "loss": 1.1019, "step": 547 }, { "epoch": 0.10543023423596748, "grad_norm": 1.4035219140330684, "learning_rate": 1.9703074515139445e-05, "loss": 0.9753, "step": 548 }, { "epoch": 0.10562262517435429, "grad_norm": 1.6620292071252563, "learning_rate": 1.9701565246917184e-05, "loss": 1.0767, "step": 549 }, { "epoch": 0.10581501611274109, "grad_norm": 1.4961098327007254, "learning_rate": 1.9700052210718775e-05, "loss": 1.0971, "step": 550 }, { "epoch": 0.10600740705112789, "grad_norm": 1.4510201165760381, "learning_rate": 1.969853540713186e-05, "loss": 1.072, "step": 551 }, { "epoch": 0.10619979798951469, "grad_norm": 1.6022152408256327, "learning_rate": 1.9697014836745552e-05, "loss": 1.1457, "step": 552 }, { "epoch": 0.10639218892790149, "grad_norm": 1.7161355200794195, "learning_rate": 1.9695490500150418e-05, "loss": 1.1175, "step": 553 }, { "epoch": 0.1065845798662883, "grad_norm": 1.6797972226597333, "learning_rate": 1.9693962397938495e-05, "loss": 1.1095, "step": 554 }, { "epoch": 0.1067769708046751, "grad_norm": 1.4657183493982409, "learning_rate": 1.9692430530703282e-05, "loss": 1.1202, "step": 555 }, { "epoch": 0.1069693617430619, "grad_norm": 1.5134809530737936, "learning_rate": 1.9690894899039735e-05, "loss": 1.0922, "step": 556 }, { "epoch": 0.1071617526814487, "grad_norm": 1.6074880869953214, "learning_rate": 1.9689355503544277e-05, "loss": 1.0838, "step": 557 }, { "epoch": 0.1073541436198355, "grad_norm": 1.63396134901111, "learning_rate": 1.968781234481479e-05, "loss": 1.1436, "step": 558 }, { "epoch": 0.1075465345582223, "grad_norm": 0.8693154674204089, "learning_rate": 1.9686265423450624e-05, "loss": 0.8137, "step": 559 }, { "epoch": 0.10773892549660911, "grad_norm": 1.4819639131379114, "learning_rate": 1.9684714740052584e-05, "loss": 1.1665, "step": 560 }, { "epoch": 0.10793131643499591, "grad_norm": 1.6179340624579839, "learning_rate": 1.9683160295222934e-05, "loss": 1.0818, "step": 561 }, { "epoch": 0.10812370737338271, "grad_norm": 1.7979527189484714, "learning_rate": 1.9681602089565403e-05, "loss": 1.0707, "step": 562 }, { "epoch": 0.10831609831176951, "grad_norm": 1.6267976065783536, "learning_rate": 1.968004012368518e-05, "loss": 1.1169, "step": 563 }, { "epoch": 0.10850848925015631, "grad_norm": 1.6483302487974651, "learning_rate": 1.967847439818892e-05, "loss": 1.1162, "step": 564 }, { "epoch": 0.10870088018854313, "grad_norm": 1.4940425536137527, "learning_rate": 1.9676904913684725e-05, "loss": 1.125, "step": 565 }, { "epoch": 0.10889327112692992, "grad_norm": 1.4808153295970115, "learning_rate": 1.967533167078217e-05, "loss": 1.0388, "step": 566 }, { "epoch": 0.10908566206531672, "grad_norm": 1.5181728161017536, "learning_rate": 1.9673754670092283e-05, "loss": 1.1114, "step": 567 }, { "epoch": 0.10927805300370352, "grad_norm": 1.5076373299272972, "learning_rate": 1.9672173912227556e-05, "loss": 0.9634, "step": 568 }, { "epoch": 0.10947044394209032, "grad_norm": 1.5855682654814183, "learning_rate": 1.967058939780193e-05, "loss": 1.1416, "step": 569 }, { "epoch": 0.10966283488047714, "grad_norm": 0.7896120098330134, "learning_rate": 1.966900112743082e-05, "loss": 0.8629, "step": 570 }, { "epoch": 0.10985522581886394, "grad_norm": 1.7750831098504174, "learning_rate": 1.966740910173108e-05, "loss": 1.1191, "step": 571 }, { "epoch": 0.11004761675725074, "grad_norm": 1.5257394024525308, "learning_rate": 1.9665813321321054e-05, "loss": 1.0003, "step": 572 }, { "epoch": 0.11024000769563753, "grad_norm": 1.5262705214308443, "learning_rate": 1.9664213786820502e-05, "loss": 1.0571, "step": 573 }, { "epoch": 0.11043239863402433, "grad_norm": 1.6686577660550823, "learning_rate": 1.9662610498850684e-05, "loss": 1.1133, "step": 574 }, { "epoch": 0.11062478957241113, "grad_norm": 1.4371526690019962, "learning_rate": 1.9661003458034288e-05, "loss": 1.0626, "step": 575 }, { "epoch": 0.11081718051079795, "grad_norm": 1.6117202842853495, "learning_rate": 1.965939266499547e-05, "loss": 1.0787, "step": 576 }, { "epoch": 0.11100957144918475, "grad_norm": 1.4370562836097123, "learning_rate": 1.9657778120359848e-05, "loss": 1.1189, "step": 577 }, { "epoch": 0.11120196238757155, "grad_norm": 1.6537740572386446, "learning_rate": 1.965615982475449e-05, "loss": 1.0925, "step": 578 }, { "epoch": 0.11139435332595835, "grad_norm": 1.4993735008927884, "learning_rate": 1.9654537778807924e-05, "loss": 1.0115, "step": 579 }, { "epoch": 0.11158674426434514, "grad_norm": 1.492040543464859, "learning_rate": 1.9652911983150135e-05, "loss": 1.1095, "step": 580 }, { "epoch": 0.11177913520273196, "grad_norm": 1.5482259838832895, "learning_rate": 1.965128243841256e-05, "loss": 1.0695, "step": 581 }, { "epoch": 0.11197152614111876, "grad_norm": 1.5799945252017078, "learning_rate": 1.96496491452281e-05, "loss": 1.1581, "step": 582 }, { "epoch": 0.11216391707950556, "grad_norm": 1.5956967538731086, "learning_rate": 1.9648012104231106e-05, "loss": 1.0297, "step": 583 }, { "epoch": 0.11235630801789236, "grad_norm": 1.5927658164412823, "learning_rate": 1.964637131605738e-05, "loss": 1.1235, "step": 584 }, { "epoch": 0.11254869895627916, "grad_norm": 1.5502126736517772, "learning_rate": 1.9644726781344197e-05, "loss": 1.1054, "step": 585 }, { "epoch": 0.11274108989466596, "grad_norm": 1.6545255032468225, "learning_rate": 1.964307850073026e-05, "loss": 1.0152, "step": 586 }, { "epoch": 0.11293348083305277, "grad_norm": 1.4002017484195404, "learning_rate": 1.964142647485576e-05, "loss": 1.0693, "step": 587 }, { "epoch": 0.11312587177143957, "grad_norm": 1.5120480302141144, "learning_rate": 1.9639770704362305e-05, "loss": 0.983, "step": 588 }, { "epoch": 0.11331826270982637, "grad_norm": 1.5983007240103448, "learning_rate": 1.9638111189892994e-05, "loss": 1.0019, "step": 589 }, { "epoch": 0.11351065364821317, "grad_norm": 1.5344813617269681, "learning_rate": 1.9636447932092354e-05, "loss": 1.0463, "step": 590 }, { "epoch": 0.11370304458659997, "grad_norm": 1.7639948847390157, "learning_rate": 1.963478093160638e-05, "loss": 1.193, "step": 591 }, { "epoch": 0.11389543552498678, "grad_norm": 1.5769030108147404, "learning_rate": 1.9633110189082515e-05, "loss": 1.0385, "step": 592 }, { "epoch": 0.11408782646337358, "grad_norm": 1.5283284224522693, "learning_rate": 1.963143570516965e-05, "loss": 1.0041, "step": 593 }, { "epoch": 0.11428021740176038, "grad_norm": 1.6193650946375413, "learning_rate": 1.9629757480518144e-05, "loss": 1.0597, "step": 594 }, { "epoch": 0.11447260834014718, "grad_norm": 0.772630743659607, "learning_rate": 1.9628075515779796e-05, "loss": 0.852, "step": 595 }, { "epoch": 0.11466499927853398, "grad_norm": 1.540537749020827, "learning_rate": 1.962638981160786e-05, "loss": 1.051, "step": 596 }, { "epoch": 0.11485739021692078, "grad_norm": 1.6400144051687358, "learning_rate": 1.9624700368657045e-05, "loss": 1.009, "step": 597 }, { "epoch": 0.11504978115530759, "grad_norm": 1.470964025074987, "learning_rate": 1.9623007187583518e-05, "loss": 1.0756, "step": 598 }, { "epoch": 0.11524217209369439, "grad_norm": 1.55213259079993, "learning_rate": 1.962131026904488e-05, "loss": 1.0795, "step": 599 }, { "epoch": 0.11543456303208119, "grad_norm": 1.4862980220155029, "learning_rate": 1.96196096137002e-05, "loss": 1.0898, "step": 600 }, { "epoch": 0.11562695397046799, "grad_norm": 1.7009368181144953, "learning_rate": 1.9617905222209998e-05, "loss": 1.0221, "step": 601 }, { "epoch": 0.11581934490885479, "grad_norm": 1.5252241184528634, "learning_rate": 1.961619709523623e-05, "loss": 1.1196, "step": 602 }, { "epoch": 0.1160117358472416, "grad_norm": 0.7930232766420391, "learning_rate": 1.9614485233442316e-05, "loss": 0.9247, "step": 603 }, { "epoch": 0.1162041267856284, "grad_norm": 1.574365099005017, "learning_rate": 1.961276963749313e-05, "loss": 1.1644, "step": 604 }, { "epoch": 0.1163965177240152, "grad_norm": 1.7271164740805833, "learning_rate": 1.9611050308054982e-05, "loss": 1.0351, "step": 605 }, { "epoch": 0.116588908662402, "grad_norm": 1.59690012346693, "learning_rate": 1.9609327245795642e-05, "loss": 1.1792, "step": 606 }, { "epoch": 0.1167812996007888, "grad_norm": 1.62610073121876, "learning_rate": 1.9607600451384327e-05, "loss": 1.0517, "step": 607 }, { "epoch": 0.1169736905391756, "grad_norm": 1.4885090568450594, "learning_rate": 1.960586992549171e-05, "loss": 1.1057, "step": 608 }, { "epoch": 0.11716608147756241, "grad_norm": 1.4254606310484985, "learning_rate": 1.9604135668789897e-05, "loss": 1.1027, "step": 609 }, { "epoch": 0.11735847241594921, "grad_norm": 1.572222953635354, "learning_rate": 1.9602397681952462e-05, "loss": 1.1145, "step": 610 }, { "epoch": 0.11755086335433601, "grad_norm": 1.6536698053398815, "learning_rate": 1.9600655965654413e-05, "loss": 1.0498, "step": 611 }, { "epoch": 0.11774325429272281, "grad_norm": 0.7960597117779793, "learning_rate": 1.959891052057222e-05, "loss": 0.8528, "step": 612 }, { "epoch": 0.11793564523110961, "grad_norm": 1.5096062265195969, "learning_rate": 1.9597161347383783e-05, "loss": 1.0409, "step": 613 }, { "epoch": 0.11812803616949642, "grad_norm": 1.696268014848814, "learning_rate": 1.959540844676847e-05, "loss": 1.0661, "step": 614 }, { "epoch": 0.11832042710788322, "grad_norm": 1.5822510447910942, "learning_rate": 1.9593651819407084e-05, "loss": 1.007, "step": 615 }, { "epoch": 0.11851281804627002, "grad_norm": 1.5408638252885944, "learning_rate": 1.959189146598188e-05, "loss": 1.0915, "step": 616 }, { "epoch": 0.11870520898465682, "grad_norm": 1.5382595896468747, "learning_rate": 1.9590127387176556e-05, "loss": 1.0921, "step": 617 }, { "epoch": 0.11889759992304362, "grad_norm": 1.4047288057465954, "learning_rate": 1.9588359583676263e-05, "loss": 1.0956, "step": 618 }, { "epoch": 0.11908999086143042, "grad_norm": 1.6150343114545793, "learning_rate": 1.9586588056167595e-05, "loss": 1.0575, "step": 619 }, { "epoch": 0.11928238179981723, "grad_norm": 1.5157146996375834, "learning_rate": 1.958481280533859e-05, "loss": 1.0175, "step": 620 }, { "epoch": 0.11947477273820403, "grad_norm": 1.7099614451483824, "learning_rate": 1.958303383187874e-05, "loss": 1.058, "step": 621 }, { "epoch": 0.11966716367659083, "grad_norm": 1.5616816678324081, "learning_rate": 1.9581251136478974e-05, "loss": 1.0512, "step": 622 }, { "epoch": 0.11985955461497763, "grad_norm": 0.8012181552007294, "learning_rate": 1.9579464719831668e-05, "loss": 0.8672, "step": 623 }, { "epoch": 0.12005194555336443, "grad_norm": 1.6376519118507944, "learning_rate": 1.9577674582630653e-05, "loss": 1.1304, "step": 624 }, { "epoch": 0.12024433649175124, "grad_norm": 1.6380617763584955, "learning_rate": 1.957588072557119e-05, "loss": 1.0853, "step": 625 }, { "epoch": 0.12043672743013804, "grad_norm": 0.7369367936524942, "learning_rate": 1.957408314935e-05, "loss": 0.8146, "step": 626 }, { "epoch": 0.12062911836852484, "grad_norm": 1.681252568828316, "learning_rate": 1.9572281854665233e-05, "loss": 1.1463, "step": 627 }, { "epoch": 0.12082150930691164, "grad_norm": 1.657792669317273, "learning_rate": 1.95704768422165e-05, "loss": 1.1088, "step": 628 }, { "epoch": 0.12101390024529844, "grad_norm": 1.7481632990677995, "learning_rate": 1.956866811270484e-05, "loss": 1.1586, "step": 629 }, { "epoch": 0.12120629118368526, "grad_norm": 1.5706819288063434, "learning_rate": 1.9566855666832743e-05, "loss": 1.1406, "step": 630 }, { "epoch": 0.12139868212207205, "grad_norm": 0.7349205697007025, "learning_rate": 1.9565039505304145e-05, "loss": 0.8594, "step": 631 }, { "epoch": 0.12159107306045885, "grad_norm": 1.5141372421031714, "learning_rate": 1.956321962882442e-05, "loss": 1.0557, "step": 632 }, { "epoch": 0.12178346399884565, "grad_norm": 1.5395010586637843, "learning_rate": 1.956139603810039e-05, "loss": 1.0529, "step": 633 }, { "epoch": 0.12197585493723245, "grad_norm": 0.7875660139129536, "learning_rate": 1.9559568733840317e-05, "loss": 0.8229, "step": 634 }, { "epoch": 0.12216824587561925, "grad_norm": 1.480672338233897, "learning_rate": 1.9557737716753898e-05, "loss": 1.0778, "step": 635 }, { "epoch": 0.12236063681400607, "grad_norm": 1.5110425003766386, "learning_rate": 1.9555902987552283e-05, "loss": 1.0474, "step": 636 }, { "epoch": 0.12255302775239287, "grad_norm": 1.5039232260599453, "learning_rate": 1.9554064546948064e-05, "loss": 1.1476, "step": 637 }, { "epoch": 0.12274541869077966, "grad_norm": 1.5893689480042532, "learning_rate": 1.9552222395655262e-05, "loss": 1.1062, "step": 638 }, { "epoch": 0.12293780962916646, "grad_norm": 1.6598623088713007, "learning_rate": 1.9550376534389355e-05, "loss": 1.1523, "step": 639 }, { "epoch": 0.12313020056755326, "grad_norm": 1.5756058346789263, "learning_rate": 1.9548526963867253e-05, "loss": 1.0325, "step": 640 }, { "epoch": 0.12332259150594008, "grad_norm": 1.4829332002527897, "learning_rate": 1.9546673684807303e-05, "loss": 1.142, "step": 641 }, { "epoch": 0.12351498244432688, "grad_norm": 1.5307077371050215, "learning_rate": 1.95448166979293e-05, "loss": 1.0195, "step": 642 }, { "epoch": 0.12370737338271368, "grad_norm": 1.6376162253679134, "learning_rate": 1.9542956003954477e-05, "loss": 1.0352, "step": 643 }, { "epoch": 0.12389976432110048, "grad_norm": 1.5502236176828252, "learning_rate": 1.9541091603605508e-05, "loss": 1.0966, "step": 644 }, { "epoch": 0.12409215525948727, "grad_norm": 1.4998782558127122, "learning_rate": 1.95392234976065e-05, "loss": 1.0476, "step": 645 }, { "epoch": 0.12428454619787407, "grad_norm": 1.4590796009752662, "learning_rate": 1.9537351686683003e-05, "loss": 1.0977, "step": 646 }, { "epoch": 0.12447693713626089, "grad_norm": 1.6129364630851144, "learning_rate": 1.9535476171562014e-05, "loss": 1.002, "step": 647 }, { "epoch": 0.12466932807464769, "grad_norm": 1.5142947887833962, "learning_rate": 1.9533596952971955e-05, "loss": 1.0702, "step": 648 }, { "epoch": 0.12486171901303449, "grad_norm": 1.6847874168387993, "learning_rate": 1.9531714031642698e-05, "loss": 1.0513, "step": 649 }, { "epoch": 0.1250541099514213, "grad_norm": 0.8402929667087204, "learning_rate": 1.9529827408305542e-05, "loss": 0.8628, "step": 650 }, { "epoch": 0.1252465008898081, "grad_norm": 1.5387059638407856, "learning_rate": 1.9527937083693233e-05, "loss": 0.9911, "step": 651 }, { "epoch": 0.1254388918281949, "grad_norm": 1.4945129170611784, "learning_rate": 1.952604305853995e-05, "loss": 1.0887, "step": 652 }, { "epoch": 0.1256312827665817, "grad_norm": 1.4209101696606887, "learning_rate": 1.9524145333581315e-05, "loss": 1.054, "step": 653 }, { "epoch": 0.1258236737049685, "grad_norm": 1.6293225203468324, "learning_rate": 1.9522243909554375e-05, "loss": 1.062, "step": 654 }, { "epoch": 0.1260160646433553, "grad_norm": 1.6308432674032605, "learning_rate": 1.952033878719763e-05, "loss": 1.0675, "step": 655 }, { "epoch": 0.1262084555817421, "grad_norm": 1.5685221498261532, "learning_rate": 1.9518429967251e-05, "loss": 1.0734, "step": 656 }, { "epoch": 0.1264008465201289, "grad_norm": 1.5783609985290348, "learning_rate": 1.9516517450455853e-05, "loss": 1.1325, "step": 657 }, { "epoch": 0.1265932374585157, "grad_norm": 1.5421063825392898, "learning_rate": 1.951460123755499e-05, "loss": 1.0719, "step": 658 }, { "epoch": 0.1267856283969025, "grad_norm": 1.3913919693765386, "learning_rate": 1.9512681329292635e-05, "loss": 1.0077, "step": 659 }, { "epoch": 0.12697801933528932, "grad_norm": 1.3905438586060295, "learning_rate": 1.951075772641447e-05, "loss": 1.0813, "step": 660 }, { "epoch": 0.12717041027367612, "grad_norm": 1.5322362364983162, "learning_rate": 1.95088304296676e-05, "loss": 1.0641, "step": 661 }, { "epoch": 0.12736280121206292, "grad_norm": 1.7891329594899699, "learning_rate": 1.950689943980056e-05, "loss": 1.0791, "step": 662 }, { "epoch": 0.12755519215044972, "grad_norm": 1.4250279030967914, "learning_rate": 1.9504964757563322e-05, "loss": 1.09, "step": 663 }, { "epoch": 0.12774758308883652, "grad_norm": 0.7827180494527548, "learning_rate": 1.95030263837073e-05, "loss": 0.7979, "step": 664 }, { "epoch": 0.12793997402722332, "grad_norm": 1.5557846825421613, "learning_rate": 1.9501084318985335e-05, "loss": 1.0674, "step": 665 }, { "epoch": 0.12813236496561012, "grad_norm": 1.6075100744200541, "learning_rate": 1.94991385641517e-05, "loss": 1.1643, "step": 666 }, { "epoch": 0.12832475590399692, "grad_norm": 0.7569881746941038, "learning_rate": 1.9497189119962105e-05, "loss": 0.8426, "step": 667 }, { "epoch": 0.12851714684238372, "grad_norm": 1.8203543116997092, "learning_rate": 1.9495235987173693e-05, "loss": 1.1577, "step": 668 }, { "epoch": 0.12870953778077052, "grad_norm": 1.3780936795174683, "learning_rate": 1.949327916654504e-05, "loss": 0.9821, "step": 669 }, { "epoch": 0.12890192871915732, "grad_norm": 1.5094653997962613, "learning_rate": 1.949131865883614e-05, "loss": 1.1026, "step": 670 }, { "epoch": 0.12909431965754414, "grad_norm": 1.5938605550059206, "learning_rate": 1.948935446480845e-05, "loss": 1.1062, "step": 671 }, { "epoch": 0.12928671059593094, "grad_norm": 1.6135861796190019, "learning_rate": 1.948738658522483e-05, "loss": 1.1183, "step": 672 }, { "epoch": 0.12947910153431774, "grad_norm": 1.3874964982895048, "learning_rate": 1.9485415020849583e-05, "loss": 0.9835, "step": 673 }, { "epoch": 0.12967149247270454, "grad_norm": 1.5614684798918277, "learning_rate": 1.9483439772448444e-05, "loss": 1.0705, "step": 674 }, { "epoch": 0.12986388341109134, "grad_norm": 1.6449712622564643, "learning_rate": 1.9481460840788573e-05, "loss": 1.1175, "step": 675 }, { "epoch": 0.13005627434947814, "grad_norm": 1.6089028271890946, "learning_rate": 1.9479478226638565e-05, "loss": 1.06, "step": 676 }, { "epoch": 0.13024866528786494, "grad_norm": 1.4716584310704317, "learning_rate": 1.947749193076845e-05, "loss": 1.0945, "step": 677 }, { "epoch": 0.13044105622625174, "grad_norm": 1.6417636056527611, "learning_rate": 1.9475501953949674e-05, "loss": 1.1146, "step": 678 }, { "epoch": 0.13063344716463854, "grad_norm": 1.5404014934508237, "learning_rate": 1.9473508296955126e-05, "loss": 1.0741, "step": 679 }, { "epoch": 0.13082583810302534, "grad_norm": 1.4299719083344433, "learning_rate": 1.9471510960559122e-05, "loss": 1.1087, "step": 680 }, { "epoch": 0.13101822904141214, "grad_norm": 1.7057863592461517, "learning_rate": 1.9469509945537395e-05, "loss": 1.1097, "step": 681 }, { "epoch": 0.13121061997979896, "grad_norm": 1.572464158294302, "learning_rate": 1.9467505252667126e-05, "loss": 1.1184, "step": 682 }, { "epoch": 0.13140301091818576, "grad_norm": 1.4413072116743453, "learning_rate": 1.9465496882726913e-05, "loss": 0.9706, "step": 683 }, { "epoch": 0.13159540185657256, "grad_norm": 0.8650564931785216, "learning_rate": 1.946348483649678e-05, "loss": 0.8487, "step": 684 }, { "epoch": 0.13178779279495936, "grad_norm": 1.5086132330775563, "learning_rate": 1.9461469114758184e-05, "loss": 1.0143, "step": 685 }, { "epoch": 0.13198018373334616, "grad_norm": 1.62217296278039, "learning_rate": 1.9459449718294008e-05, "loss": 1.1277, "step": 686 }, { "epoch": 0.13217257467173296, "grad_norm": 1.3056235251525445, "learning_rate": 1.945742664788856e-05, "loss": 1.03, "step": 687 }, { "epoch": 0.13236496561011976, "grad_norm": 1.5695192431356968, "learning_rate": 1.9455399904327585e-05, "loss": 1.0183, "step": 688 }, { "epoch": 0.13255735654850656, "grad_norm": 1.4813435664459456, "learning_rate": 1.945336948839824e-05, "loss": 1.0829, "step": 689 }, { "epoch": 0.13274974748689336, "grad_norm": 1.7785547232159655, "learning_rate": 1.9451335400889114e-05, "loss": 1.2065, "step": 690 }, { "epoch": 0.13294213842528016, "grad_norm": 1.4560024459723093, "learning_rate": 1.944929764259023e-05, "loss": 1.0461, "step": 691 }, { "epoch": 0.13313452936366696, "grad_norm": 1.5365279541648753, "learning_rate": 1.9447256214293026e-05, "loss": 1.0207, "step": 692 }, { "epoch": 0.1333269203020538, "grad_norm": 1.3568830574550264, "learning_rate": 1.9445211116790365e-05, "loss": 0.9719, "step": 693 }, { "epoch": 0.13351931124044059, "grad_norm": 1.6305041776427924, "learning_rate": 1.9443162350876544e-05, "loss": 1.1186, "step": 694 }, { "epoch": 0.13371170217882739, "grad_norm": 1.526765790859847, "learning_rate": 1.944110991734728e-05, "loss": 1.0469, "step": 695 }, { "epoch": 0.13390409311721418, "grad_norm": 1.6523795393626437, "learning_rate": 1.9439053816999717e-05, "loss": 1.1229, "step": 696 }, { "epoch": 0.13409648405560098, "grad_norm": 1.4541003556069252, "learning_rate": 1.9436994050632415e-05, "loss": 0.9751, "step": 697 }, { "epoch": 0.13428887499398778, "grad_norm": 1.5080875551455624, "learning_rate": 1.9434930619045367e-05, "loss": 1.0624, "step": 698 }, { "epoch": 0.13448126593237458, "grad_norm": 1.5924767486750668, "learning_rate": 1.9432863523039986e-05, "loss": 1.0213, "step": 699 }, { "epoch": 0.13467365687076138, "grad_norm": 1.4132583858434784, "learning_rate": 1.9430792763419105e-05, "loss": 0.9644, "step": 700 }, { "epoch": 0.13486604780914818, "grad_norm": 1.5610910637101645, "learning_rate": 1.942871834098699e-05, "loss": 1.0542, "step": 701 }, { "epoch": 0.13505843874753498, "grad_norm": 1.637554648261022, "learning_rate": 1.9426640256549313e-05, "loss": 1.0897, "step": 702 }, { "epoch": 0.13525082968592178, "grad_norm": 1.6031289927451928, "learning_rate": 1.9424558510913186e-05, "loss": 1.0615, "step": 703 }, { "epoch": 0.1354432206243086, "grad_norm": 1.4262853009665797, "learning_rate": 1.9422473104887133e-05, "loss": 1.0075, "step": 704 }, { "epoch": 0.1356356115626954, "grad_norm": 1.5765811416029898, "learning_rate": 1.9420384039281103e-05, "loss": 0.9619, "step": 705 }, { "epoch": 0.1358280025010822, "grad_norm": 1.5743840156098274, "learning_rate": 1.941829131490646e-05, "loss": 1.0801, "step": 706 }, { "epoch": 0.136020393439469, "grad_norm": 1.5956556588431408, "learning_rate": 1.9416194932576e-05, "loss": 1.0496, "step": 707 }, { "epoch": 0.1362127843778558, "grad_norm": 1.3953618119310658, "learning_rate": 1.941409489310393e-05, "loss": 1.0204, "step": 708 }, { "epoch": 0.1364051753162426, "grad_norm": 1.7923874049269375, "learning_rate": 1.9411991197305878e-05, "loss": 1.0932, "step": 709 }, { "epoch": 0.1365975662546294, "grad_norm": 1.460806817411089, "learning_rate": 1.9409883845998905e-05, "loss": 1.0274, "step": 710 }, { "epoch": 0.1367899571930162, "grad_norm": 1.5094575717970724, "learning_rate": 1.9407772840001473e-05, "loss": 1.0933, "step": 711 }, { "epoch": 0.136982348131403, "grad_norm": 1.486863781931974, "learning_rate": 1.9405658180133477e-05, "loss": 1.0347, "step": 712 }, { "epoch": 0.1371747390697898, "grad_norm": 1.7025382654071202, "learning_rate": 1.9403539867216226e-05, "loss": 1.0573, "step": 713 }, { "epoch": 0.1373671300081766, "grad_norm": 1.799036907543484, "learning_rate": 1.9401417902072447e-05, "loss": 1.1126, "step": 714 }, { "epoch": 0.13755952094656343, "grad_norm": 1.6387378912272426, "learning_rate": 1.9399292285526286e-05, "loss": 1.1387, "step": 715 }, { "epoch": 0.13775191188495023, "grad_norm": 1.4428747292611883, "learning_rate": 1.939716301840331e-05, "loss": 1.0055, "step": 716 }, { "epoch": 0.13794430282333703, "grad_norm": 1.359242651280295, "learning_rate": 1.9395030101530504e-05, "loss": 1.0514, "step": 717 }, { "epoch": 0.13813669376172383, "grad_norm": 1.5865231903968953, "learning_rate": 1.939289353573626e-05, "loss": 1.1227, "step": 718 }, { "epoch": 0.13832908470011063, "grad_norm": 1.5184687156211027, "learning_rate": 1.9390753321850404e-05, "loss": 1.1054, "step": 719 }, { "epoch": 0.13852147563849743, "grad_norm": 1.5877364000273295, "learning_rate": 1.938860946070417e-05, "loss": 1.037, "step": 720 }, { "epoch": 0.13871386657688423, "grad_norm": 1.51135949018525, "learning_rate": 1.93864619531302e-05, "loss": 1.0613, "step": 721 }, { "epoch": 0.13890625751527103, "grad_norm": 1.4833423075157561, "learning_rate": 1.9384310799962575e-05, "loss": 1.1039, "step": 722 }, { "epoch": 0.13909864845365782, "grad_norm": 1.4259450186483018, "learning_rate": 1.9382156002036764e-05, "loss": 1.1005, "step": 723 }, { "epoch": 0.13929103939204462, "grad_norm": 1.3967699114936596, "learning_rate": 1.9379997560189677e-05, "loss": 1.0022, "step": 724 }, { "epoch": 0.13948343033043142, "grad_norm": 1.6566081130458767, "learning_rate": 1.937783547525962e-05, "loss": 1.1218, "step": 725 }, { "epoch": 0.13967582126881825, "grad_norm": 0.8609807132446312, "learning_rate": 1.9375669748086326e-05, "loss": 0.8408, "step": 726 }, { "epoch": 0.13986821220720505, "grad_norm": 1.6938869270635657, "learning_rate": 1.937350037951094e-05, "loss": 1.0465, "step": 727 }, { "epoch": 0.14006060314559185, "grad_norm": 1.3829439527209744, "learning_rate": 1.9371327370376018e-05, "loss": 1.0233, "step": 728 }, { "epoch": 0.14025299408397865, "grad_norm": 1.65198359477732, "learning_rate": 1.936915072152553e-05, "loss": 1.0403, "step": 729 }, { "epoch": 0.14044538502236545, "grad_norm": 1.4968564050025623, "learning_rate": 1.936697043380486e-05, "loss": 1.1277, "step": 730 }, { "epoch": 0.14063777596075225, "grad_norm": 1.4825575639052833, "learning_rate": 1.936478650806081e-05, "loss": 1.0555, "step": 731 }, { "epoch": 0.14083016689913905, "grad_norm": 1.4994750067176423, "learning_rate": 1.936259894514159e-05, "loss": 1.0681, "step": 732 }, { "epoch": 0.14102255783752585, "grad_norm": 1.5179890129234983, "learning_rate": 1.9360407745896828e-05, "loss": 1.0894, "step": 733 }, { "epoch": 0.14121494877591265, "grad_norm": 1.548434905385124, "learning_rate": 1.9358212911177556e-05, "loss": 0.9852, "step": 734 }, { "epoch": 0.14140733971429945, "grad_norm": 1.5589217289672017, "learning_rate": 1.935601444183622e-05, "loss": 0.9616, "step": 735 }, { "epoch": 0.14159973065268625, "grad_norm": 1.4957686433610307, "learning_rate": 1.935381233872669e-05, "loss": 1.1736, "step": 736 }, { "epoch": 0.14179212159107307, "grad_norm": 1.6570037542226186, "learning_rate": 1.935160660270423e-05, "loss": 1.0755, "step": 737 }, { "epoch": 0.14198451252945987, "grad_norm": 1.5152782676164092, "learning_rate": 1.934939723462552e-05, "loss": 1.0715, "step": 738 }, { "epoch": 0.14217690346784667, "grad_norm": 1.4469665038417279, "learning_rate": 1.9347184235348663e-05, "loss": 1.0533, "step": 739 }, { "epoch": 0.14236929440623347, "grad_norm": 1.6719570358364082, "learning_rate": 1.9344967605733154e-05, "loss": 1.0388, "step": 740 }, { "epoch": 0.14256168534462027, "grad_norm": 1.8122511265292558, "learning_rate": 1.934274734663991e-05, "loss": 1.0083, "step": 741 }, { "epoch": 0.14275407628300707, "grad_norm": 1.527077434668499, "learning_rate": 1.934052345893125e-05, "loss": 1.1309, "step": 742 }, { "epoch": 0.14294646722139387, "grad_norm": 1.4564137506157924, "learning_rate": 1.9338295943470915e-05, "loss": 1.0644, "step": 743 }, { "epoch": 0.14313885815978067, "grad_norm": 1.610957636806743, "learning_rate": 1.9336064801124034e-05, "loss": 1.0364, "step": 744 }, { "epoch": 0.14333124909816747, "grad_norm": 1.6352138029600805, "learning_rate": 1.933383003275717e-05, "loss": 1.0647, "step": 745 }, { "epoch": 0.14352364003655427, "grad_norm": 1.7221198447501387, "learning_rate": 1.933159163923827e-05, "loss": 1.1185, "step": 746 }, { "epoch": 0.14371603097494107, "grad_norm": 1.5296706765558383, "learning_rate": 1.9329349621436708e-05, "loss": 1.1165, "step": 747 }, { "epoch": 0.1439084219133279, "grad_norm": 1.699437853351202, "learning_rate": 1.9327103980223255e-05, "loss": 1.0702, "step": 748 }, { "epoch": 0.1441008128517147, "grad_norm": 1.5163045000138418, "learning_rate": 1.932485471647009e-05, "loss": 1.0394, "step": 749 }, { "epoch": 0.1442932037901015, "grad_norm": 0.7386449668425761, "learning_rate": 1.9322601831050804e-05, "loss": 0.8403, "step": 750 }, { "epoch": 0.1444855947284883, "grad_norm": 1.524610559662008, "learning_rate": 1.9320345324840396e-05, "loss": 1.0861, "step": 751 }, { "epoch": 0.1446779856668751, "grad_norm": 1.5545806111843412, "learning_rate": 1.9318085198715257e-05, "loss": 1.0079, "step": 752 }, { "epoch": 0.1448703766052619, "grad_norm": 1.5159631930433766, "learning_rate": 1.93158214535532e-05, "loss": 1.0587, "step": 753 }, { "epoch": 0.1450627675436487, "grad_norm": 1.5541752460546128, "learning_rate": 1.9313554090233436e-05, "loss": 0.9934, "step": 754 }, { "epoch": 0.1452551584820355, "grad_norm": 1.487577927275408, "learning_rate": 1.9311283109636586e-05, "loss": 1.1128, "step": 755 }, { "epoch": 0.1454475494204223, "grad_norm": 1.5218683013312806, "learning_rate": 1.9309008512644668e-05, "loss": 1.0648, "step": 756 }, { "epoch": 0.1456399403588091, "grad_norm": 1.5370731950738898, "learning_rate": 1.930673030014111e-05, "loss": 1.0737, "step": 757 }, { "epoch": 0.1458323312971959, "grad_norm": 1.5386416510927954, "learning_rate": 1.930444847301075e-05, "loss": 0.9811, "step": 758 }, { "epoch": 0.14602472223558272, "grad_norm": 0.7739367420764496, "learning_rate": 1.9302163032139813e-05, "loss": 0.7803, "step": 759 }, { "epoch": 0.14621711317396952, "grad_norm": 1.515827558408712, "learning_rate": 1.9299873978415947e-05, "loss": 1.019, "step": 760 }, { "epoch": 0.14640950411235631, "grad_norm": 1.5530492469217634, "learning_rate": 1.9297581312728187e-05, "loss": 1.0509, "step": 761 }, { "epoch": 0.14660189505074311, "grad_norm": 1.527726320876282, "learning_rate": 1.929528503596698e-05, "loss": 1.0652, "step": 762 }, { "epoch": 0.1467942859891299, "grad_norm": 1.6172456713799417, "learning_rate": 1.929298514902418e-05, "loss": 1.0682, "step": 763 }, { "epoch": 0.1469866769275167, "grad_norm": 1.4781734943068925, "learning_rate": 1.929068165279303e-05, "loss": 1.0943, "step": 764 }, { "epoch": 0.1471790678659035, "grad_norm": 1.5194772902302374, "learning_rate": 1.928837454816818e-05, "loss": 1.1232, "step": 765 }, { "epoch": 0.1473714588042903, "grad_norm": 1.5991523515441277, "learning_rate": 1.9286063836045687e-05, "loss": 1.0257, "step": 766 }, { "epoch": 0.1475638497426771, "grad_norm": 1.647858197534753, "learning_rate": 1.9283749517323e-05, "loss": 1.0664, "step": 767 }, { "epoch": 0.1477562406810639, "grad_norm": 1.60454798644335, "learning_rate": 1.928143159289898e-05, "loss": 1.0495, "step": 768 }, { "epoch": 0.1479486316194507, "grad_norm": 1.5299097970634694, "learning_rate": 1.927911006367388e-05, "loss": 1.0809, "step": 769 }, { "epoch": 0.14814102255783754, "grad_norm": 1.4624477568422074, "learning_rate": 1.927678493054935e-05, "loss": 1.0033, "step": 770 }, { "epoch": 0.14833341349622434, "grad_norm": 1.499294543935229, "learning_rate": 1.9274456194428454e-05, "loss": 1.0804, "step": 771 }, { "epoch": 0.14852580443461114, "grad_norm": 1.4662522942435685, "learning_rate": 1.9272123856215643e-05, "loss": 1.0356, "step": 772 }, { "epoch": 0.14871819537299794, "grad_norm": 1.4125423162316666, "learning_rate": 1.9269787916816764e-05, "loss": 1.0699, "step": 773 }, { "epoch": 0.14891058631138474, "grad_norm": 1.4659005790182438, "learning_rate": 1.9267448377139074e-05, "loss": 1.0903, "step": 774 }, { "epoch": 0.14910297724977153, "grad_norm": 1.4880602403953913, "learning_rate": 1.9265105238091227e-05, "loss": 1.0257, "step": 775 }, { "epoch": 0.14929536818815833, "grad_norm": 1.500050553865249, "learning_rate": 1.9262758500583265e-05, "loss": 1.0202, "step": 776 }, { "epoch": 0.14948775912654513, "grad_norm": 1.5409036419097897, "learning_rate": 1.9260408165526638e-05, "loss": 1.0916, "step": 777 }, { "epoch": 0.14968015006493193, "grad_norm": 1.7164467451702485, "learning_rate": 1.9258054233834184e-05, "loss": 0.9931, "step": 778 }, { "epoch": 0.14987254100331873, "grad_norm": 1.5022981618766598, "learning_rate": 1.9255696706420147e-05, "loss": 1.0647, "step": 779 }, { "epoch": 0.15006493194170556, "grad_norm": 1.596288404101551, "learning_rate": 1.9253335584200165e-05, "loss": 1.0923, "step": 780 }, { "epoch": 0.15025732288009236, "grad_norm": 1.4509410503660665, "learning_rate": 1.9250970868091268e-05, "loss": 1.0889, "step": 781 }, { "epoch": 0.15044971381847916, "grad_norm": 1.5933412391549777, "learning_rate": 1.924860255901188e-05, "loss": 1.0694, "step": 782 }, { "epoch": 0.15064210475686596, "grad_norm": 1.4832660722057442, "learning_rate": 1.9246230657881834e-05, "loss": 1.0783, "step": 783 }, { "epoch": 0.15083449569525276, "grad_norm": 1.4664246546863766, "learning_rate": 1.9243855165622345e-05, "loss": 1.1159, "step": 784 }, { "epoch": 0.15102688663363956, "grad_norm": 1.5525456561483382, "learning_rate": 1.9241476083156026e-05, "loss": 1.0804, "step": 785 }, { "epoch": 0.15121927757202636, "grad_norm": 1.5559792282158613, "learning_rate": 1.9239093411406885e-05, "loss": 1.0047, "step": 786 }, { "epoch": 0.15141166851041316, "grad_norm": 1.6379355380971083, "learning_rate": 1.9236707151300326e-05, "loss": 1.1244, "step": 787 }, { "epoch": 0.15160405944879995, "grad_norm": 1.5141992809128297, "learning_rate": 1.9234317303763145e-05, "loss": 1.0689, "step": 788 }, { "epoch": 0.15179645038718675, "grad_norm": 1.4057964682201596, "learning_rate": 1.9231923869723528e-05, "loss": 1.0772, "step": 789 }, { "epoch": 0.15198884132557355, "grad_norm": 1.3040700348473229, "learning_rate": 1.922952685011106e-05, "loss": 1.0191, "step": 790 }, { "epoch": 0.15218123226396038, "grad_norm": 1.4384326083405943, "learning_rate": 1.9227126245856716e-05, "loss": 1.089, "step": 791 }, { "epoch": 0.15237362320234718, "grad_norm": 1.4693784891017518, "learning_rate": 1.922472205789286e-05, "loss": 1.0398, "step": 792 }, { "epoch": 0.15256601414073398, "grad_norm": 1.6113454685917898, "learning_rate": 1.9222314287153255e-05, "loss": 1.0156, "step": 793 }, { "epoch": 0.15275840507912078, "grad_norm": 1.4761853498482034, "learning_rate": 1.9219902934573048e-05, "loss": 1.0624, "step": 794 }, { "epoch": 0.15295079601750758, "grad_norm": 1.6890239872760242, "learning_rate": 1.9217488001088784e-05, "loss": 1.0992, "step": 795 }, { "epoch": 0.15314318695589438, "grad_norm": 1.4928552078922768, "learning_rate": 1.9215069487638396e-05, "loss": 1.0901, "step": 796 }, { "epoch": 0.15333557789428118, "grad_norm": 1.4101416502380797, "learning_rate": 1.92126473951612e-05, "loss": 1.1151, "step": 797 }, { "epoch": 0.15352796883266798, "grad_norm": 1.5575195050894186, "learning_rate": 1.921022172459791e-05, "loss": 1.0766, "step": 798 }, { "epoch": 0.15372035977105478, "grad_norm": 1.4081883521340637, "learning_rate": 1.920779247689064e-05, "loss": 1.0835, "step": 799 }, { "epoch": 0.15391275070944158, "grad_norm": 1.6057795020956116, "learning_rate": 1.9205359652982868e-05, "loss": 1.0379, "step": 800 }, { "epoch": 0.15410514164782838, "grad_norm": 1.6253121989066237, "learning_rate": 1.9202923253819482e-05, "loss": 1.1126, "step": 801 }, { "epoch": 0.1542975325862152, "grad_norm": 1.4208153165356558, "learning_rate": 1.920048328034675e-05, "loss": 1.0447, "step": 802 }, { "epoch": 0.154489923524602, "grad_norm": 1.6036010005783645, "learning_rate": 1.9198039733512326e-05, "loss": 1.1045, "step": 803 }, { "epoch": 0.1546823144629888, "grad_norm": 1.5639924789476194, "learning_rate": 1.9195592614265262e-05, "loss": 0.9616, "step": 804 }, { "epoch": 0.1548747054013756, "grad_norm": 1.679644045187824, "learning_rate": 1.9193141923555984e-05, "loss": 1.0798, "step": 805 }, { "epoch": 0.1550670963397624, "grad_norm": 1.6159629537841023, "learning_rate": 1.919068766233632e-05, "loss": 1.0796, "step": 806 }, { "epoch": 0.1552594872781492, "grad_norm": 1.6037432972740124, "learning_rate": 1.9188229831559468e-05, "loss": 1.0338, "step": 807 }, { "epoch": 0.155451878216536, "grad_norm": 1.5955429428984638, "learning_rate": 1.9185768432180026e-05, "loss": 1.0252, "step": 808 }, { "epoch": 0.1556442691549228, "grad_norm": 1.5971985015293153, "learning_rate": 1.9183303465153972e-05, "loss": 1.0371, "step": 809 }, { "epoch": 0.1558366600933096, "grad_norm": 1.6615159492307017, "learning_rate": 1.9180834931438673e-05, "loss": 1.099, "step": 810 }, { "epoch": 0.1560290510316964, "grad_norm": 1.4845125535731458, "learning_rate": 1.917836283199288e-05, "loss": 1.0516, "step": 811 }, { "epoch": 0.1562214419700832, "grad_norm": 1.441484551589866, "learning_rate": 1.917588716777672e-05, "loss": 1.0457, "step": 812 }, { "epoch": 0.15641383290847002, "grad_norm": 1.589752460819522, "learning_rate": 1.917340793975172e-05, "loss": 1.0214, "step": 813 }, { "epoch": 0.15660622384685682, "grad_norm": 1.5035737821427109, "learning_rate": 1.917092514888078e-05, "loss": 1.0379, "step": 814 }, { "epoch": 0.15679861478524362, "grad_norm": 1.6278849614579216, "learning_rate": 1.9168438796128193e-05, "loss": 1.0539, "step": 815 }, { "epoch": 0.15699100572363042, "grad_norm": 1.4394579347568328, "learning_rate": 1.9165948882459623e-05, "loss": 1.1071, "step": 816 }, { "epoch": 0.15718339666201722, "grad_norm": 1.4628947099764906, "learning_rate": 1.9163455408842123e-05, "loss": 0.9631, "step": 817 }, { "epoch": 0.15737578760040402, "grad_norm": 1.5845054749525174, "learning_rate": 1.9160958376244138e-05, "loss": 1.0975, "step": 818 }, { "epoch": 0.15756817853879082, "grad_norm": 1.5546428437204995, "learning_rate": 1.9158457785635478e-05, "loss": 0.9832, "step": 819 }, { "epoch": 0.15776056947717762, "grad_norm": 1.5642827144323346, "learning_rate": 1.915595363798735e-05, "loss": 1.1062, "step": 820 }, { "epoch": 0.15795296041556442, "grad_norm": 1.6017185091246073, "learning_rate": 1.915344593427233e-05, "loss": 1.0837, "step": 821 }, { "epoch": 0.15814535135395122, "grad_norm": 1.6985013279949495, "learning_rate": 1.9150934675464384e-05, "loss": 1.1464, "step": 822 }, { "epoch": 0.15833774229233802, "grad_norm": 1.5417874291954128, "learning_rate": 1.9148419862538858e-05, "loss": 1.012, "step": 823 }, { "epoch": 0.15853013323072485, "grad_norm": 1.5335596994355865, "learning_rate": 1.9145901496472474e-05, "loss": 0.9317, "step": 824 }, { "epoch": 0.15872252416911165, "grad_norm": 0.7810821065764861, "learning_rate": 1.9143379578243335e-05, "loss": 0.8814, "step": 825 }, { "epoch": 0.15891491510749844, "grad_norm": 1.705973136741719, "learning_rate": 1.914085410883093e-05, "loss": 1.0896, "step": 826 }, { "epoch": 0.15910730604588524, "grad_norm": 1.5845100002228947, "learning_rate": 1.9138325089216118e-05, "loss": 0.9671, "step": 827 }, { "epoch": 0.15929969698427204, "grad_norm": 1.3234725093291357, "learning_rate": 1.913579252038114e-05, "loss": 1.019, "step": 828 }, { "epoch": 0.15949208792265884, "grad_norm": 1.622874525088459, "learning_rate": 1.9133256403309627e-05, "loss": 1.028, "step": 829 }, { "epoch": 0.15968447886104564, "grad_norm": 1.359555077119314, "learning_rate": 1.913071673898656e-05, "loss": 1.0551, "step": 830 }, { "epoch": 0.15987686979943244, "grad_norm": 1.4158042398400492, "learning_rate": 1.912817352839833e-05, "loss": 1.0078, "step": 831 }, { "epoch": 0.16006926073781924, "grad_norm": 1.579908965081598, "learning_rate": 1.9125626772532683e-05, "loss": 1.0822, "step": 832 }, { "epoch": 0.16026165167620604, "grad_norm": 1.5062015441585428, "learning_rate": 1.9123076472378753e-05, "loss": 1.0081, "step": 833 }, { "epoch": 0.16045404261459284, "grad_norm": 1.4216579628484576, "learning_rate": 1.9120522628927047e-05, "loss": 1.1373, "step": 834 }, { "epoch": 0.16064643355297967, "grad_norm": 1.5780469991400041, "learning_rate": 1.9117965243169444e-05, "loss": 1.1546, "step": 835 }, { "epoch": 0.16083882449136647, "grad_norm": 1.4811087442230917, "learning_rate": 1.9115404316099212e-05, "loss": 1.0035, "step": 836 }, { "epoch": 0.16103121542975327, "grad_norm": 1.4902491410400762, "learning_rate": 1.9112839848710978e-05, "loss": 0.9984, "step": 837 }, { "epoch": 0.16122360636814007, "grad_norm": 1.564179572225476, "learning_rate": 1.9110271842000755e-05, "loss": 1.0665, "step": 838 }, { "epoch": 0.16141599730652686, "grad_norm": 1.5805214209568568, "learning_rate": 1.9107700296965926e-05, "loss": 1.066, "step": 839 }, { "epoch": 0.16160838824491366, "grad_norm": 1.4957479387891657, "learning_rate": 1.910512521460525e-05, "loss": 1.0781, "step": 840 }, { "epoch": 0.16180077918330046, "grad_norm": 1.4793872901688228, "learning_rate": 1.9102546595918857e-05, "loss": 1.085, "step": 841 }, { "epoch": 0.16199317012168726, "grad_norm": 1.6241219970742236, "learning_rate": 1.9099964441908258e-05, "loss": 1.0549, "step": 842 }, { "epoch": 0.16218556106007406, "grad_norm": 1.5436056246796772, "learning_rate": 1.9097378753576327e-05, "loss": 1.0647, "step": 843 }, { "epoch": 0.16237795199846086, "grad_norm": 1.6029525001784573, "learning_rate": 1.9094789531927315e-05, "loss": 1.0589, "step": 844 }, { "epoch": 0.16257034293684766, "grad_norm": 1.5984474527742862, "learning_rate": 1.909219677796685e-05, "loss": 1.0897, "step": 845 }, { "epoch": 0.1627627338752345, "grad_norm": 1.626250683023961, "learning_rate": 1.9089600492701926e-05, "loss": 1.0674, "step": 846 }, { "epoch": 0.1629551248136213, "grad_norm": 1.4153769403106005, "learning_rate": 1.908700067714091e-05, "loss": 1.0959, "step": 847 }, { "epoch": 0.1631475157520081, "grad_norm": 1.3569252829673961, "learning_rate": 1.9084397332293537e-05, "loss": 1.0988, "step": 848 }, { "epoch": 0.1633399066903949, "grad_norm": 1.4679199967204215, "learning_rate": 1.9081790459170926e-05, "loss": 1.0443, "step": 849 }, { "epoch": 0.1635322976287817, "grad_norm": 1.5215756457214642, "learning_rate": 1.9079180058785547e-05, "loss": 1.1159, "step": 850 }, { "epoch": 0.16372468856716849, "grad_norm": 1.422085455602293, "learning_rate": 1.9076566132151255e-05, "loss": 0.9979, "step": 851 }, { "epoch": 0.16391707950555529, "grad_norm": 1.4781075560929375, "learning_rate": 1.907394868028326e-05, "loss": 1.0414, "step": 852 }, { "epoch": 0.16410947044394208, "grad_norm": 1.421938295934273, "learning_rate": 1.9071327704198163e-05, "loss": 0.9831, "step": 853 }, { "epoch": 0.16430186138232888, "grad_norm": 1.3958449470659444, "learning_rate": 1.906870320491391e-05, "loss": 1.1526, "step": 854 }, { "epoch": 0.16449425232071568, "grad_norm": 1.7390630792883184, "learning_rate": 1.9066075183449835e-05, "loss": 1.0765, "step": 855 }, { "epoch": 0.16468664325910248, "grad_norm": 1.5561666198880326, "learning_rate": 1.9063443640826624e-05, "loss": 1.0511, "step": 856 }, { "epoch": 0.1648790341974893, "grad_norm": 1.4936397155601018, "learning_rate": 1.906080857806634e-05, "loss": 1.015, "step": 857 }, { "epoch": 0.1650714251358761, "grad_norm": 0.7487230299374287, "learning_rate": 1.905816999619242e-05, "loss": 0.8227, "step": 858 }, { "epoch": 0.1652638160742629, "grad_norm": 1.6118002986654445, "learning_rate": 1.9055527896229642e-05, "loss": 1.0825, "step": 859 }, { "epoch": 0.1654562070126497, "grad_norm": 1.4482047091492118, "learning_rate": 1.905288227920418e-05, "loss": 1.0129, "step": 860 }, { "epoch": 0.1656485979510365, "grad_norm": 1.458951538708241, "learning_rate": 1.9050233146143554e-05, "loss": 1.0658, "step": 861 }, { "epoch": 0.1658409888894233, "grad_norm": 1.5375097907107684, "learning_rate": 1.9047580498076663e-05, "loss": 1.0924, "step": 862 }, { "epoch": 0.1660333798278101, "grad_norm": 1.613825043639672, "learning_rate": 1.904492433603376e-05, "loss": 1.1022, "step": 863 }, { "epoch": 0.1662257707661969, "grad_norm": 1.4601214774502882, "learning_rate": 1.904226466104647e-05, "loss": 1.1053, "step": 864 }, { "epoch": 0.1664181617045837, "grad_norm": 1.5009363690188942, "learning_rate": 1.903960147414778e-05, "loss": 1.1526, "step": 865 }, { "epoch": 0.1666105526429705, "grad_norm": 1.573210158051203, "learning_rate": 1.903693477637204e-05, "loss": 1.0692, "step": 866 }, { "epoch": 0.1668029435813573, "grad_norm": 1.5072407293547951, "learning_rate": 1.9034264568754967e-05, "loss": 1.1248, "step": 867 }, { "epoch": 0.16699533451974413, "grad_norm": 1.2762237224317925, "learning_rate": 1.9031590852333637e-05, "loss": 0.9714, "step": 868 }, { "epoch": 0.16718772545813093, "grad_norm": 1.5508631100407024, "learning_rate": 1.9028913628146487e-05, "loss": 1.0452, "step": 869 }, { "epoch": 0.16738011639651773, "grad_norm": 1.4203831628387282, "learning_rate": 1.902623289723333e-05, "loss": 0.9784, "step": 870 }, { "epoch": 0.16757250733490453, "grad_norm": 1.5248875883444213, "learning_rate": 1.902354866063532e-05, "loss": 1.0621, "step": 871 }, { "epoch": 0.16776489827329133, "grad_norm": 1.5598486587795675, "learning_rate": 1.9020860919394992e-05, "loss": 1.1008, "step": 872 }, { "epoch": 0.16795728921167813, "grad_norm": 1.5887070292902576, "learning_rate": 1.9018169674556228e-05, "loss": 1.0214, "step": 873 }, { "epoch": 0.16814968015006493, "grad_norm": 1.5624531410136762, "learning_rate": 1.901547492716428e-05, "loss": 1.005, "step": 874 }, { "epoch": 0.16834207108845173, "grad_norm": 1.5087477239767728, "learning_rate": 1.9012776678265756e-05, "loss": 1.0424, "step": 875 }, { "epoch": 0.16853446202683853, "grad_norm": 1.496668715745799, "learning_rate": 1.9010074928908624e-05, "loss": 1.1141, "step": 876 }, { "epoch": 0.16872685296522533, "grad_norm": 1.6827460640903986, "learning_rate": 1.900736968014221e-05, "loss": 1.031, "step": 877 }, { "epoch": 0.16891924390361213, "grad_norm": 1.5207631959100218, "learning_rate": 1.9004660933017208e-05, "loss": 1.0513, "step": 878 }, { "epoch": 0.16911163484199895, "grad_norm": 1.5549818947630913, "learning_rate": 1.900194868858566e-05, "loss": 1.0667, "step": 879 }, { "epoch": 0.16930402578038575, "grad_norm": 0.7999213171544891, "learning_rate": 1.8999232947900968e-05, "loss": 0.8676, "step": 880 }, { "epoch": 0.16949641671877255, "grad_norm": 1.38725828104696, "learning_rate": 1.89965137120179e-05, "loss": 0.9846, "step": 881 }, { "epoch": 0.16968880765715935, "grad_norm": 1.5273587576345953, "learning_rate": 1.899379098199257e-05, "loss": 1.1355, "step": 882 }, { "epoch": 0.16988119859554615, "grad_norm": 1.5257086033924527, "learning_rate": 1.899106475888246e-05, "loss": 1.1414, "step": 883 }, { "epoch": 0.17007358953393295, "grad_norm": 1.5583166310519796, "learning_rate": 1.89883350437464e-05, "loss": 1.0358, "step": 884 }, { "epoch": 0.17026598047231975, "grad_norm": 1.494683073452839, "learning_rate": 1.8985601837644586e-05, "loss": 1.0882, "step": 885 }, { "epoch": 0.17045837141070655, "grad_norm": 1.6325981279357977, "learning_rate": 1.8982865141638557e-05, "loss": 1.0321, "step": 886 }, { "epoch": 0.17065076234909335, "grad_norm": 1.6010527861397577, "learning_rate": 1.8980124956791216e-05, "loss": 1.0424, "step": 887 }, { "epoch": 0.17084315328748015, "grad_norm": 1.496982714296233, "learning_rate": 1.8977381284166818e-05, "loss": 1.1272, "step": 888 }, { "epoch": 0.17103554422586695, "grad_norm": 1.3583086200835208, "learning_rate": 1.897463412483098e-05, "loss": 1.0582, "step": 889 }, { "epoch": 0.17122793516425378, "grad_norm": 1.5134695731869765, "learning_rate": 1.897188347985066e-05, "loss": 1.0201, "step": 890 }, { "epoch": 0.17142032610264057, "grad_norm": 0.7380445539102235, "learning_rate": 1.896912935029418e-05, "loss": 0.839, "step": 891 }, { "epoch": 0.17161271704102737, "grad_norm": 1.4818580243329114, "learning_rate": 1.896637173723121e-05, "loss": 1.0636, "step": 892 }, { "epoch": 0.17180510797941417, "grad_norm": 0.7594490568483634, "learning_rate": 1.8963610641732777e-05, "loss": 0.8802, "step": 893 }, { "epoch": 0.17199749891780097, "grad_norm": 1.5664110474657231, "learning_rate": 1.8960846064871257e-05, "loss": 1.0991, "step": 894 }, { "epoch": 0.17218988985618777, "grad_norm": 0.7690336324180602, "learning_rate": 1.8958078007720387e-05, "loss": 0.8585, "step": 895 }, { "epoch": 0.17238228079457457, "grad_norm": 1.3615642320807837, "learning_rate": 1.895530647135524e-05, "loss": 1.0573, "step": 896 }, { "epoch": 0.17257467173296137, "grad_norm": 1.5683038202634831, "learning_rate": 1.8952531456852248e-05, "loss": 1.1154, "step": 897 }, { "epoch": 0.17276706267134817, "grad_norm": 1.3629443672995403, "learning_rate": 1.8949752965289197e-05, "loss": 1.0315, "step": 898 }, { "epoch": 0.17295945360973497, "grad_norm": 1.5757096548511562, "learning_rate": 1.894697099774523e-05, "loss": 1.0276, "step": 899 }, { "epoch": 0.1731518445481218, "grad_norm": 1.5727496652201685, "learning_rate": 1.894418555530082e-05, "loss": 1.0832, "step": 900 }, { "epoch": 0.1733442354865086, "grad_norm": 1.4153849661222835, "learning_rate": 1.89413966390378e-05, "loss": 1.028, "step": 901 }, { "epoch": 0.1735366264248954, "grad_norm": 1.5437258853496687, "learning_rate": 1.8938604250039362e-05, "loss": 0.9863, "step": 902 }, { "epoch": 0.1737290173632822, "grad_norm": 0.763958231660208, "learning_rate": 1.8935808389390032e-05, "loss": 0.7741, "step": 903 }, { "epoch": 0.173921408301669, "grad_norm": 1.4696317751390715, "learning_rate": 1.893300905817569e-05, "loss": 0.99, "step": 904 }, { "epoch": 0.1741137992400558, "grad_norm": 1.588824832414014, "learning_rate": 1.8930206257483566e-05, "loss": 1.0869, "step": 905 }, { "epoch": 0.1743061901784426, "grad_norm": 1.59738626717053, "learning_rate": 1.8927399988402233e-05, "loss": 1.0911, "step": 906 }, { "epoch": 0.1744985811168294, "grad_norm": 1.5156795773250729, "learning_rate": 1.8924590252021614e-05, "loss": 1.0415, "step": 907 }, { "epoch": 0.1746909720552162, "grad_norm": 1.5627318151630976, "learning_rate": 1.8921777049432985e-05, "loss": 1.0698, "step": 908 }, { "epoch": 0.174883362993603, "grad_norm": 1.4405773131461637, "learning_rate": 1.8918960381728947e-05, "loss": 1.033, "step": 909 }, { "epoch": 0.1750757539319898, "grad_norm": 1.4544636924361571, "learning_rate": 1.8916140250003475e-05, "loss": 1.0797, "step": 910 }, { "epoch": 0.17526814487037662, "grad_norm": 1.4878271473296392, "learning_rate": 1.891331665535187e-05, "loss": 0.9909, "step": 911 }, { "epoch": 0.17546053580876342, "grad_norm": 1.7279543759959448, "learning_rate": 1.8910489598870784e-05, "loss": 0.9356, "step": 912 }, { "epoch": 0.17565292674715022, "grad_norm": 1.4481439860102605, "learning_rate": 1.8907659081658214e-05, "loss": 1.0126, "step": 913 }, { "epoch": 0.17584531768553702, "grad_norm": 1.5137237865565238, "learning_rate": 1.8904825104813497e-05, "loss": 1.0183, "step": 914 }, { "epoch": 0.17603770862392382, "grad_norm": 0.8279610641718373, "learning_rate": 1.8901987669437322e-05, "loss": 0.9192, "step": 915 }, { "epoch": 0.17623009956231062, "grad_norm": 1.6016255434219355, "learning_rate": 1.889914677663171e-05, "loss": 1.1289, "step": 916 }, { "epoch": 0.17642249050069742, "grad_norm": 1.6221305740979939, "learning_rate": 1.8896302427500042e-05, "loss": 1.1623, "step": 917 }, { "epoch": 0.17661488143908421, "grad_norm": 1.5776084402189714, "learning_rate": 1.8893454623147017e-05, "loss": 1.0822, "step": 918 }, { "epoch": 0.17680727237747101, "grad_norm": 1.489260117237681, "learning_rate": 1.88906033646787e-05, "loss": 1.0234, "step": 919 }, { "epoch": 0.1769996633158578, "grad_norm": 1.421838494749104, "learning_rate": 1.8887748653202478e-05, "loss": 1.0419, "step": 920 }, { "epoch": 0.1771920542542446, "grad_norm": 1.584209837678964, "learning_rate": 1.8884890489827097e-05, "loss": 1.0354, "step": 921 }, { "epoch": 0.17738444519263144, "grad_norm": 1.413185419671685, "learning_rate": 1.8882028875662627e-05, "loss": 1.0591, "step": 922 }, { "epoch": 0.17757683613101824, "grad_norm": 1.3757618745502262, "learning_rate": 1.8879163811820493e-05, "loss": 1.0116, "step": 923 }, { "epoch": 0.17776922706940504, "grad_norm": 1.5745943575377999, "learning_rate": 1.8876295299413445e-05, "loss": 1.0393, "step": 924 }, { "epoch": 0.17796161800779184, "grad_norm": 1.4744912135955752, "learning_rate": 1.8873423339555584e-05, "loss": 1.0629, "step": 925 }, { "epoch": 0.17815400894617864, "grad_norm": 1.492358600479478, "learning_rate": 1.8870547933362352e-05, "loss": 1.0357, "step": 926 }, { "epoch": 0.17834639988456544, "grad_norm": 1.4957398872350742, "learning_rate": 1.886766908195051e-05, "loss": 0.9951, "step": 927 }, { "epoch": 0.17853879082295224, "grad_norm": 1.3475847067989541, "learning_rate": 1.8864786786438187e-05, "loss": 1.0454, "step": 928 }, { "epoch": 0.17873118176133904, "grad_norm": 1.4807129323057662, "learning_rate": 1.886190104794482e-05, "loss": 1.0185, "step": 929 }, { "epoch": 0.17892357269972584, "grad_norm": 0.7924076050770938, "learning_rate": 1.8859011867591203e-05, "loss": 0.8928, "step": 930 }, { "epoch": 0.17911596363811264, "grad_norm": 1.365357935727793, "learning_rate": 1.885611924649946e-05, "loss": 1.0814, "step": 931 }, { "epoch": 0.17930835457649943, "grad_norm": 1.405504099051713, "learning_rate": 1.885322318579305e-05, "loss": 1.1006, "step": 932 }, { "epoch": 0.17950074551488626, "grad_norm": 1.4412377869537083, "learning_rate": 1.8850323686596766e-05, "loss": 1.0219, "step": 933 }, { "epoch": 0.17969313645327306, "grad_norm": 1.4256646536980258, "learning_rate": 1.8847420750036748e-05, "loss": 1.0502, "step": 934 }, { "epoch": 0.17988552739165986, "grad_norm": 1.572970207025588, "learning_rate": 1.884451437724046e-05, "loss": 1.1334, "step": 935 }, { "epoch": 0.18007791833004666, "grad_norm": 1.5510613886385367, "learning_rate": 1.8841604569336702e-05, "loss": 1.0478, "step": 936 }, { "epoch": 0.18027030926843346, "grad_norm": 1.3875026653468912, "learning_rate": 1.883869132745561e-05, "loss": 0.9741, "step": 937 }, { "epoch": 0.18046270020682026, "grad_norm": 1.4937429374279116, "learning_rate": 1.883577465272866e-05, "loss": 1.0854, "step": 938 }, { "epoch": 0.18065509114520706, "grad_norm": 1.4503539647798445, "learning_rate": 1.8832854546288642e-05, "loss": 1.0159, "step": 939 }, { "epoch": 0.18084748208359386, "grad_norm": 0.7853359873592057, "learning_rate": 1.8829931009269707e-05, "loss": 0.8492, "step": 940 }, { "epoch": 0.18103987302198066, "grad_norm": 1.4024667176438383, "learning_rate": 1.882700404280731e-05, "loss": 0.9999, "step": 941 }, { "epoch": 0.18123226396036746, "grad_norm": 1.549385215560397, "learning_rate": 1.8824073648038258e-05, "loss": 0.9606, "step": 942 }, { "epoch": 0.18142465489875426, "grad_norm": 1.4877679503465469, "learning_rate": 1.882113982610068e-05, "loss": 1.0319, "step": 943 }, { "epoch": 0.18161704583714108, "grad_norm": 1.428352609738142, "learning_rate": 1.881820257813404e-05, "loss": 1.0728, "step": 944 }, { "epoch": 0.18180943677552788, "grad_norm": 1.411321151316002, "learning_rate": 1.8815261905279133e-05, "loss": 1.0169, "step": 945 }, { "epoch": 0.18200182771391468, "grad_norm": 1.4334001916054553, "learning_rate": 1.8812317808678075e-05, "loss": 1.0629, "step": 946 }, { "epoch": 0.18219421865230148, "grad_norm": 1.4465991251894705, "learning_rate": 1.8809370289474327e-05, "loss": 1.0233, "step": 947 }, { "epoch": 0.18238660959068828, "grad_norm": 1.4585508718268827, "learning_rate": 1.8806419348812673e-05, "loss": 1.135, "step": 948 }, { "epoch": 0.18257900052907508, "grad_norm": 1.6326986942509087, "learning_rate": 1.8803464987839217e-05, "loss": 1.0627, "step": 949 }, { "epoch": 0.18277139146746188, "grad_norm": 1.3361420569415279, "learning_rate": 1.88005072077014e-05, "loss": 0.9883, "step": 950 }, { "epoch": 0.18296378240584868, "grad_norm": 1.447675219327409, "learning_rate": 1.8797546009547996e-05, "loss": 1.0752, "step": 951 }, { "epoch": 0.18315617334423548, "grad_norm": 1.3672985380530156, "learning_rate": 1.879458139452909e-05, "loss": 1.0692, "step": 952 }, { "epoch": 0.18334856428262228, "grad_norm": 1.4898933695534797, "learning_rate": 1.8791613363796118e-05, "loss": 1.1373, "step": 953 }, { "epoch": 0.18354095522100908, "grad_norm": 1.4864500431129493, "learning_rate": 1.8788641918501817e-05, "loss": 1.0266, "step": 954 }, { "epoch": 0.1837333461593959, "grad_norm": 1.5302478720534543, "learning_rate": 1.8785667059800264e-05, "loss": 0.9886, "step": 955 }, { "epoch": 0.1839257370977827, "grad_norm": 1.5408700477082928, "learning_rate": 1.8782688788846865e-05, "loss": 0.9787, "step": 956 }, { "epoch": 0.1841181280361695, "grad_norm": 0.7636008919308716, "learning_rate": 1.877970710679834e-05, "loss": 0.8794, "step": 957 }, { "epoch": 0.1843105189745563, "grad_norm": 1.536723911838152, "learning_rate": 1.877672201481275e-05, "loss": 1.044, "step": 958 }, { "epoch": 0.1845029099129431, "grad_norm": 1.5438338497300623, "learning_rate": 1.877373351404946e-05, "loss": 1.0222, "step": 959 }, { "epoch": 0.1846953008513299, "grad_norm": 1.5878653080727445, "learning_rate": 1.8770741605669173e-05, "loss": 1.0665, "step": 960 }, { "epoch": 0.1848876917897167, "grad_norm": 1.6817843441079854, "learning_rate": 1.876774629083391e-05, "loss": 1.0376, "step": 961 }, { "epoch": 0.1850800827281035, "grad_norm": 1.5291600236164198, "learning_rate": 1.8764747570707017e-05, "loss": 1.1222, "step": 962 }, { "epoch": 0.1852724736664903, "grad_norm": 1.558068323594428, "learning_rate": 1.8761745446453167e-05, "loss": 1.1165, "step": 963 }, { "epoch": 0.1854648646048771, "grad_norm": 1.4964842149311817, "learning_rate": 1.875873991923835e-05, "loss": 1.0713, "step": 964 }, { "epoch": 0.1856572555432639, "grad_norm": 1.493565851143512, "learning_rate": 1.875573099022987e-05, "loss": 0.982, "step": 965 }, { "epoch": 0.18584964648165073, "grad_norm": 1.4320190762311922, "learning_rate": 1.8752718660596367e-05, "loss": 1.0954, "step": 966 }, { "epoch": 0.18604203742003753, "grad_norm": 1.5416273275261139, "learning_rate": 1.8749702931507797e-05, "loss": 1.0375, "step": 967 }, { "epoch": 0.18623442835842433, "grad_norm": 1.5131194197837936, "learning_rate": 1.874668380413543e-05, "loss": 1.0934, "step": 968 }, { "epoch": 0.18642681929681112, "grad_norm": 1.5437573097533848, "learning_rate": 1.8743661279651856e-05, "loss": 1.0897, "step": 969 }, { "epoch": 0.18661921023519792, "grad_norm": 1.527285117477586, "learning_rate": 1.8740635359231e-05, "loss": 1.0739, "step": 970 }, { "epoch": 0.18681160117358472, "grad_norm": 1.4549242545520737, "learning_rate": 1.8737606044048086e-05, "loss": 1.1017, "step": 971 }, { "epoch": 0.18700399211197152, "grad_norm": 1.5193772672802055, "learning_rate": 1.873457333527967e-05, "loss": 1.0431, "step": 972 }, { "epoch": 0.18719638305035832, "grad_norm": 1.53428450123911, "learning_rate": 1.873153723410362e-05, "loss": 1.0711, "step": 973 }, { "epoch": 0.18738877398874512, "grad_norm": 1.377990622360118, "learning_rate": 1.8728497741699115e-05, "loss": 1.0151, "step": 974 }, { "epoch": 0.18758116492713192, "grad_norm": 1.4481427234110449, "learning_rate": 1.872545485924667e-05, "loss": 1.0814, "step": 975 }, { "epoch": 0.18777355586551872, "grad_norm": 1.5122188732979565, "learning_rate": 1.8722408587928104e-05, "loss": 1.0938, "step": 976 }, { "epoch": 0.18796594680390555, "grad_norm": 1.4941778880349021, "learning_rate": 1.8719358928926546e-05, "loss": 0.9748, "step": 977 }, { "epoch": 0.18815833774229235, "grad_norm": 1.555354460353274, "learning_rate": 1.8716305883426456e-05, "loss": 1.0414, "step": 978 }, { "epoch": 0.18835072868067915, "grad_norm": 1.5996639611282173, "learning_rate": 1.87132494526136e-05, "loss": 1.0933, "step": 979 }, { "epoch": 0.18854311961906595, "grad_norm": 1.4920152112519318, "learning_rate": 1.8710189637675055e-05, "loss": 1.0597, "step": 980 }, { "epoch": 0.18873551055745275, "grad_norm": 1.4351042883924143, "learning_rate": 1.8707126439799225e-05, "loss": 1.0618, "step": 981 }, { "epoch": 0.18892790149583955, "grad_norm": 1.44762096921506, "learning_rate": 1.870405986017582e-05, "loss": 0.9881, "step": 982 }, { "epoch": 0.18912029243422634, "grad_norm": 1.3272676896890687, "learning_rate": 1.8700989899995857e-05, "loss": 0.9925, "step": 983 }, { "epoch": 0.18931268337261314, "grad_norm": 0.8424295396361104, "learning_rate": 1.8697916560451682e-05, "loss": 0.8937, "step": 984 }, { "epoch": 0.18950507431099994, "grad_norm": 1.6016105094346256, "learning_rate": 1.869483984273694e-05, "loss": 1.0141, "step": 985 }, { "epoch": 0.18969746524938674, "grad_norm": 1.3733338542662057, "learning_rate": 1.8691759748046594e-05, "loss": 1.0238, "step": 986 }, { "epoch": 0.18988985618777354, "grad_norm": 1.4479573995824289, "learning_rate": 1.8688676277576916e-05, "loss": 0.9989, "step": 987 }, { "epoch": 0.19008224712616037, "grad_norm": 1.3720379577662056, "learning_rate": 1.868558943252549e-05, "loss": 0.9062, "step": 988 }, { "epoch": 0.19027463806454717, "grad_norm": 1.4252012793187065, "learning_rate": 1.868249921409122e-05, "loss": 1.0244, "step": 989 }, { "epoch": 0.19046702900293397, "grad_norm": 0.7256270413544507, "learning_rate": 1.8679405623474294e-05, "loss": 0.8135, "step": 990 }, { "epoch": 0.19065941994132077, "grad_norm": 1.456769684782095, "learning_rate": 1.8676308661876242e-05, "loss": 1.0099, "step": 991 }, { "epoch": 0.19085181087970757, "grad_norm": 1.5516260765470042, "learning_rate": 1.8673208330499884e-05, "loss": 1.0944, "step": 992 }, { "epoch": 0.19104420181809437, "grad_norm": 1.5697275725133024, "learning_rate": 1.8670104630549348e-05, "loss": 1.0707, "step": 993 }, { "epoch": 0.19123659275648117, "grad_norm": 1.561574626304262, "learning_rate": 1.866699756323008e-05, "loss": 1.0649, "step": 994 }, { "epoch": 0.19142898369486797, "grad_norm": 1.4733719719159772, "learning_rate": 1.866388712974883e-05, "loss": 1.1399, "step": 995 }, { "epoch": 0.19162137463325477, "grad_norm": 1.6438152479733625, "learning_rate": 1.866077333131365e-05, "loss": 1.0088, "step": 996 }, { "epoch": 0.19181376557164156, "grad_norm": 1.3800926632255617, "learning_rate": 1.8657656169133908e-05, "loss": 1.024, "step": 997 }, { "epoch": 0.19200615651002836, "grad_norm": 1.4684242229859639, "learning_rate": 1.8654535644420276e-05, "loss": 1.0393, "step": 998 }, { "epoch": 0.1921985474484152, "grad_norm": 1.499825220002056, "learning_rate": 1.8651411758384718e-05, "loss": 1.0725, "step": 999 }, { "epoch": 0.192390938386802, "grad_norm": 1.5098064756416685, "learning_rate": 1.8648284512240527e-05, "loss": 1.0483, "step": 1000 }, { "epoch": 0.1925833293251888, "grad_norm": 1.3211020845688197, "learning_rate": 1.8645153907202285e-05, "loss": 1.0336, "step": 1001 }, { "epoch": 0.1927757202635756, "grad_norm": 1.5661823376437634, "learning_rate": 1.8642019944485884e-05, "loss": 1.052, "step": 1002 }, { "epoch": 0.1929681112019624, "grad_norm": 1.2953533605489833, "learning_rate": 1.863888262530852e-05, "loss": 1.0163, "step": 1003 }, { "epoch": 0.1931605021403492, "grad_norm": 1.3891298508647705, "learning_rate": 1.863574195088869e-05, "loss": 1.0661, "step": 1004 }, { "epoch": 0.193352893078736, "grad_norm": 1.3374249536887635, "learning_rate": 1.8632597922446195e-05, "loss": 1.0406, "step": 1005 }, { "epoch": 0.1935452840171228, "grad_norm": 1.440663227084882, "learning_rate": 1.8629450541202142e-05, "loss": 0.9961, "step": 1006 }, { "epoch": 0.1937376749555096, "grad_norm": 1.5458544851432583, "learning_rate": 1.8626299808378933e-05, "loss": 1.0167, "step": 1007 }, { "epoch": 0.1939300658938964, "grad_norm": 1.5904691063733976, "learning_rate": 1.862314572520028e-05, "loss": 1.0691, "step": 1008 }, { "epoch": 0.19412245683228319, "grad_norm": 1.5818308772068097, "learning_rate": 1.861998829289119e-05, "loss": 1.0486, "step": 1009 }, { "epoch": 0.19431484777067, "grad_norm": 1.3740464027534744, "learning_rate": 1.861682751267798e-05, "loss": 1.0458, "step": 1010 }, { "epoch": 0.1945072387090568, "grad_norm": 1.6281799444114138, "learning_rate": 1.861366338578825e-05, "loss": 1.0504, "step": 1011 }, { "epoch": 0.1946996296474436, "grad_norm": 1.4081289064530769, "learning_rate": 1.8610495913450922e-05, "loss": 0.9999, "step": 1012 }, { "epoch": 0.1948920205858304, "grad_norm": 1.3791356298269917, "learning_rate": 1.8607325096896197e-05, "loss": 1.0628, "step": 1013 }, { "epoch": 0.1950844115242172, "grad_norm": 1.4771926054974283, "learning_rate": 1.8604150937355588e-05, "loss": 0.9903, "step": 1014 }, { "epoch": 0.195276802462604, "grad_norm": 1.4220598911008115, "learning_rate": 1.86009734360619e-05, "loss": 1.0523, "step": 1015 }, { "epoch": 0.1954691934009908, "grad_norm": 1.4619540574986885, "learning_rate": 1.8597792594249237e-05, "loss": 1.0764, "step": 1016 }, { "epoch": 0.1956615843393776, "grad_norm": 1.552323858209436, "learning_rate": 1.8594608413153e-05, "loss": 1.0181, "step": 1017 }, { "epoch": 0.1958539752777644, "grad_norm": 0.7263171224546403, "learning_rate": 1.8591420894009897e-05, "loss": 0.8058, "step": 1018 }, { "epoch": 0.1960463662161512, "grad_norm": 0.7896134077064728, "learning_rate": 1.8588230038057913e-05, "loss": 0.8518, "step": 1019 }, { "epoch": 0.19623875715453803, "grad_norm": 1.3793296314097618, "learning_rate": 1.8585035846536347e-05, "loss": 1.0446, "step": 1020 }, { "epoch": 0.19643114809292483, "grad_norm": 1.4393527732298448, "learning_rate": 1.8581838320685782e-05, "loss": 1.0626, "step": 1021 }, { "epoch": 0.19662353903131163, "grad_norm": 1.559292676247122, "learning_rate": 1.8578637461748105e-05, "loss": 1.0859, "step": 1022 }, { "epoch": 0.19681592996969843, "grad_norm": 1.5573906949693181, "learning_rate": 1.857543327096649e-05, "loss": 1.1631, "step": 1023 }, { "epoch": 0.19700832090808523, "grad_norm": 1.5694969835257884, "learning_rate": 1.85722257495854e-05, "loss": 1.0773, "step": 1024 }, { "epoch": 0.19720071184647203, "grad_norm": 1.389610714728074, "learning_rate": 1.856901489885061e-05, "loss": 1.0801, "step": 1025 }, { "epoch": 0.19739310278485883, "grad_norm": 1.5063328258196238, "learning_rate": 1.856580072000918e-05, "loss": 1.1599, "step": 1026 }, { "epoch": 0.19758549372324563, "grad_norm": 1.5218621907169616, "learning_rate": 1.8562583214309447e-05, "loss": 1.0517, "step": 1027 }, { "epoch": 0.19777788466163243, "grad_norm": 1.4683370692212216, "learning_rate": 1.855936238300106e-05, "loss": 1.0728, "step": 1028 }, { "epoch": 0.19797027560001923, "grad_norm": 1.5585773743278912, "learning_rate": 1.8556138227334957e-05, "loss": 1.0821, "step": 1029 }, { "epoch": 0.19816266653840603, "grad_norm": 1.456151415996446, "learning_rate": 1.855291074856336e-05, "loss": 1.082, "step": 1030 }, { "epoch": 0.19835505747679286, "grad_norm": 1.443008966599867, "learning_rate": 1.8549679947939778e-05, "loss": 1.0465, "step": 1031 }, { "epoch": 0.19854744841517966, "grad_norm": 1.4170307866422505, "learning_rate": 1.8546445826719023e-05, "loss": 1.0575, "step": 1032 }, { "epoch": 0.19873983935356646, "grad_norm": 1.617649460350673, "learning_rate": 1.8543208386157195e-05, "loss": 1.046, "step": 1033 }, { "epoch": 0.19893223029195325, "grad_norm": 1.6969965846908064, "learning_rate": 1.853996762751167e-05, "loss": 1.0725, "step": 1034 }, { "epoch": 0.19912462123034005, "grad_norm": 1.4157309136723457, "learning_rate": 1.8536723552041124e-05, "loss": 1.0377, "step": 1035 }, { "epoch": 0.19931701216872685, "grad_norm": 1.3253660951839896, "learning_rate": 1.853347616100552e-05, "loss": 0.9657, "step": 1036 }, { "epoch": 0.19950940310711365, "grad_norm": 0.7641298539206717, "learning_rate": 1.8530225455666103e-05, "loss": 0.8574, "step": 1037 }, { "epoch": 0.19970179404550045, "grad_norm": 1.718369114299437, "learning_rate": 1.8526971437285416e-05, "loss": 1.1653, "step": 1038 }, { "epoch": 0.19989418498388725, "grad_norm": 1.557032338480856, "learning_rate": 1.8523714107127278e-05, "loss": 1.1104, "step": 1039 }, { "epoch": 0.20008657592227405, "grad_norm": 1.5608805756344484, "learning_rate": 1.8520453466456797e-05, "loss": 1.0395, "step": 1040 }, { "epoch": 0.20027896686066085, "grad_norm": 1.4026226407035047, "learning_rate": 1.8517189516540376e-05, "loss": 1.0074, "step": 1041 }, { "epoch": 0.20047135779904768, "grad_norm": 0.7880722703503641, "learning_rate": 1.8513922258645687e-05, "loss": 0.8905, "step": 1042 }, { "epoch": 0.20066374873743448, "grad_norm": 1.5003521111531286, "learning_rate": 1.8510651694041702e-05, "loss": 1.0152, "step": 1043 }, { "epoch": 0.20085613967582128, "grad_norm": 1.4391685240685, "learning_rate": 1.8507377823998664e-05, "loss": 0.9783, "step": 1044 }, { "epoch": 0.20104853061420808, "grad_norm": 1.5840219560974373, "learning_rate": 1.850410064978811e-05, "loss": 1.0381, "step": 1045 }, { "epoch": 0.20124092155259488, "grad_norm": 1.4698151335578409, "learning_rate": 1.8500820172682858e-05, "loss": 1.0228, "step": 1046 }, { "epoch": 0.20143331249098168, "grad_norm": 1.4622869095270308, "learning_rate": 1.8497536393957005e-05, "loss": 1.083, "step": 1047 }, { "epoch": 0.20162570342936847, "grad_norm": 1.5332500522810644, "learning_rate": 1.8494249314885932e-05, "loss": 1.0445, "step": 1048 }, { "epoch": 0.20181809436775527, "grad_norm": 1.358209771548359, "learning_rate": 1.8490958936746304e-05, "loss": 1.0649, "step": 1049 }, { "epoch": 0.20201048530614207, "grad_norm": 1.466036683975362, "learning_rate": 1.848766526081607e-05, "loss": 1.0534, "step": 1050 }, { "epoch": 0.20220287624452887, "grad_norm": 1.3965982653466449, "learning_rate": 1.8484368288374452e-05, "loss": 1.0563, "step": 1051 }, { "epoch": 0.20239526718291567, "grad_norm": 1.4051142978509625, "learning_rate": 1.8481068020701954e-05, "loss": 0.9444, "step": 1052 }, { "epoch": 0.2025876581213025, "grad_norm": 1.5611128383008295, "learning_rate": 1.8477764459080364e-05, "loss": 1.032, "step": 1053 }, { "epoch": 0.2027800490596893, "grad_norm": 1.5435074087453176, "learning_rate": 1.8474457604792746e-05, "loss": 1.1449, "step": 1054 }, { "epoch": 0.2029724399980761, "grad_norm": 1.4369623143552108, "learning_rate": 1.8471147459123447e-05, "loss": 1.0155, "step": 1055 }, { "epoch": 0.2031648309364629, "grad_norm": 1.3885675879325186, "learning_rate": 1.8467834023358088e-05, "loss": 1.0119, "step": 1056 }, { "epoch": 0.2033572218748497, "grad_norm": 1.4349782137496572, "learning_rate": 1.846451729878357e-05, "loss": 1.1018, "step": 1057 }, { "epoch": 0.2035496128132365, "grad_norm": 1.412706801814163, "learning_rate": 1.846119728668807e-05, "loss": 1.0723, "step": 1058 }, { "epoch": 0.2037420037516233, "grad_norm": 1.458678114856342, "learning_rate": 1.845787398836104e-05, "loss": 1.0285, "step": 1059 }, { "epoch": 0.2039343946900101, "grad_norm": 0.8751770290617329, "learning_rate": 1.8454547405093212e-05, "loss": 0.9037, "step": 1060 }, { "epoch": 0.2041267856283969, "grad_norm": 1.349224638204598, "learning_rate": 1.8451217538176597e-05, "loss": 1.0274, "step": 1061 }, { "epoch": 0.2043191765667837, "grad_norm": 1.5605248568315582, "learning_rate": 1.844788438890447e-05, "loss": 1.0487, "step": 1062 }, { "epoch": 0.2045115675051705, "grad_norm": 1.345730368707232, "learning_rate": 1.8444547958571396e-05, "loss": 1.0239, "step": 1063 }, { "epoch": 0.20470395844355732, "grad_norm": 1.4051985017798796, "learning_rate": 1.84412082484732e-05, "loss": 1.059, "step": 1064 }, { "epoch": 0.20489634938194412, "grad_norm": 1.2729986287117638, "learning_rate": 1.8437865259906987e-05, "loss": 1.0774, "step": 1065 }, { "epoch": 0.20508874032033092, "grad_norm": 1.4462599399895988, "learning_rate": 1.8434518994171136e-05, "loss": 1.0084, "step": 1066 }, { "epoch": 0.20528113125871772, "grad_norm": 1.5773370571066345, "learning_rate": 1.84311694525653e-05, "loss": 0.974, "step": 1067 }, { "epoch": 0.20547352219710452, "grad_norm": 1.3366521825275495, "learning_rate": 1.84278166363904e-05, "loss": 1.034, "step": 1068 }, { "epoch": 0.20566591313549132, "grad_norm": 1.620557876771007, "learning_rate": 1.8424460546948632e-05, "loss": 1.156, "step": 1069 }, { "epoch": 0.20585830407387812, "grad_norm": 1.458714830701919, "learning_rate": 1.8421101185543463e-05, "loss": 0.9994, "step": 1070 }, { "epoch": 0.20605069501226492, "grad_norm": 1.5151067726703762, "learning_rate": 1.841773855347963e-05, "loss": 1.0322, "step": 1071 }, { "epoch": 0.20624308595065172, "grad_norm": 1.4306480431349498, "learning_rate": 1.841437265206314e-05, "loss": 1.1058, "step": 1072 }, { "epoch": 0.20643547688903852, "grad_norm": 1.3800137324863677, "learning_rate": 1.841100348260127e-05, "loss": 1.0329, "step": 1073 }, { "epoch": 0.20662786782742532, "grad_norm": 1.4634616235680138, "learning_rate": 1.840763104640257e-05, "loss": 1.018, "step": 1074 }, { "epoch": 0.20682025876581214, "grad_norm": 1.4709280107012312, "learning_rate": 1.8404255344776853e-05, "loss": 0.9949, "step": 1075 }, { "epoch": 0.20701264970419894, "grad_norm": 1.488898372211007, "learning_rate": 1.84008763790352e-05, "loss": 1.1096, "step": 1076 }, { "epoch": 0.20720504064258574, "grad_norm": 1.280834807199373, "learning_rate": 1.8397494150489963e-05, "loss": 0.9836, "step": 1077 }, { "epoch": 0.20739743158097254, "grad_norm": 1.33719284798762, "learning_rate": 1.8394108660454766e-05, "loss": 1.028, "step": 1078 }, { "epoch": 0.20758982251935934, "grad_norm": 1.300961559232701, "learning_rate": 1.8390719910244487e-05, "loss": 1.0345, "step": 1079 }, { "epoch": 0.20778221345774614, "grad_norm": 1.3632104543225425, "learning_rate": 1.8387327901175286e-05, "loss": 0.9655, "step": 1080 }, { "epoch": 0.20797460439613294, "grad_norm": 1.4871457591986856, "learning_rate": 1.838393263456457e-05, "loss": 1.0686, "step": 1081 }, { "epoch": 0.20816699533451974, "grad_norm": 1.6110243419358108, "learning_rate": 1.838053411173103e-05, "loss": 1.0299, "step": 1082 }, { "epoch": 0.20835938627290654, "grad_norm": 1.444248012630939, "learning_rate": 1.8377132333994606e-05, "loss": 1.0569, "step": 1083 }, { "epoch": 0.20855177721129334, "grad_norm": 1.472732401100989, "learning_rate": 1.837372730267652e-05, "loss": 1.078, "step": 1084 }, { "epoch": 0.20874416814968014, "grad_norm": 0.8063836497558428, "learning_rate": 1.8370319019099236e-05, "loss": 0.8461, "step": 1085 }, { "epoch": 0.20893655908806696, "grad_norm": 1.5962462088797753, "learning_rate": 1.8366907484586497e-05, "loss": 1.0342, "step": 1086 }, { "epoch": 0.20912895002645376, "grad_norm": 1.624549604225716, "learning_rate": 1.83634927004633e-05, "loss": 1.0962, "step": 1087 }, { "epoch": 0.20932134096484056, "grad_norm": 1.3879997897269656, "learning_rate": 1.8360074668055915e-05, "loss": 1.0398, "step": 1088 }, { "epoch": 0.20951373190322736, "grad_norm": 1.5449465246855758, "learning_rate": 1.8356653388691857e-05, "loss": 0.9372, "step": 1089 }, { "epoch": 0.20970612284161416, "grad_norm": 1.6237063536696854, "learning_rate": 1.8353228863699922e-05, "loss": 1.0955, "step": 1090 }, { "epoch": 0.20989851378000096, "grad_norm": 1.3876131301758015, "learning_rate": 1.8349801094410148e-05, "loss": 0.9975, "step": 1091 }, { "epoch": 0.21009090471838776, "grad_norm": 1.735068213418668, "learning_rate": 1.8346370082153843e-05, "loss": 1.0444, "step": 1092 }, { "epoch": 0.21028329565677456, "grad_norm": 1.9119148473057686, "learning_rate": 1.8342935828263574e-05, "loss": 0.9749, "step": 1093 }, { "epoch": 0.21047568659516136, "grad_norm": 1.5663890460691487, "learning_rate": 1.8339498334073166e-05, "loss": 0.9878, "step": 1094 }, { "epoch": 0.21066807753354816, "grad_norm": 1.335878103933352, "learning_rate": 1.83360576009177e-05, "loss": 1.0264, "step": 1095 }, { "epoch": 0.21086046847193496, "grad_norm": 1.40304980158215, "learning_rate": 1.833261363013352e-05, "loss": 1.0028, "step": 1096 }, { "epoch": 0.21105285941032179, "grad_norm": 1.509177878331129, "learning_rate": 1.832916642305822e-05, "loss": 1.0908, "step": 1097 }, { "epoch": 0.21124525034870859, "grad_norm": 1.6287223609981698, "learning_rate": 1.832571598103066e-05, "loss": 1.0778, "step": 1098 }, { "epoch": 0.21143764128709538, "grad_norm": 1.4911577712748232, "learning_rate": 1.8322262305390948e-05, "loss": 1.0326, "step": 1099 }, { "epoch": 0.21163003222548218, "grad_norm": 1.5897177252250891, "learning_rate": 1.8318805397480455e-05, "loss": 1.0891, "step": 1100 }, { "epoch": 0.21182242316386898, "grad_norm": 1.5090689067774465, "learning_rate": 1.8315345258641802e-05, "loss": 1.1011, "step": 1101 }, { "epoch": 0.21201481410225578, "grad_norm": 1.432180251254999, "learning_rate": 1.8311881890218873e-05, "loss": 1.0671, "step": 1102 }, { "epoch": 0.21220720504064258, "grad_norm": 1.3804234802963913, "learning_rate": 1.830841529355679e-05, "loss": 1.078, "step": 1103 }, { "epoch": 0.21239959597902938, "grad_norm": 1.4385609769093197, "learning_rate": 1.8304945470001948e-05, "loss": 1.0639, "step": 1104 }, { "epoch": 0.21259198691741618, "grad_norm": 1.4450954830522835, "learning_rate": 1.8301472420901985e-05, "loss": 1.0966, "step": 1105 }, { "epoch": 0.21278437785580298, "grad_norm": 1.5172074998079368, "learning_rate": 1.8297996147605787e-05, "loss": 1.0365, "step": 1106 }, { "epoch": 0.21297676879418978, "grad_norm": 0.7537553060191139, "learning_rate": 1.829451665146351e-05, "loss": 0.8464, "step": 1107 }, { "epoch": 0.2131691597325766, "grad_norm": 1.5279477790606972, "learning_rate": 1.8291033933826535e-05, "loss": 1.0194, "step": 1108 }, { "epoch": 0.2133615506709634, "grad_norm": 1.6483583486216091, "learning_rate": 1.8287547996047523e-05, "loss": 1.0338, "step": 1109 }, { "epoch": 0.2135539416093502, "grad_norm": 1.3553498236885633, "learning_rate": 1.8284058839480363e-05, "loss": 1.0148, "step": 1110 }, { "epoch": 0.213746332547737, "grad_norm": 1.502815124515043, "learning_rate": 1.8280566465480206e-05, "loss": 1.0443, "step": 1111 }, { "epoch": 0.2139387234861238, "grad_norm": 1.4213156990193632, "learning_rate": 1.8277070875403455e-05, "loss": 1.0075, "step": 1112 }, { "epoch": 0.2141311144245106, "grad_norm": 1.4383777776060906, "learning_rate": 1.8273572070607756e-05, "loss": 0.9758, "step": 1113 }, { "epoch": 0.2143235053628974, "grad_norm": 1.477429986740852, "learning_rate": 1.8270070052451995e-05, "loss": 1.0756, "step": 1114 }, { "epoch": 0.2145158963012842, "grad_norm": 1.7164589001277644, "learning_rate": 1.8266564822296323e-05, "loss": 1.0565, "step": 1115 }, { "epoch": 0.214708287239671, "grad_norm": 1.5552877561499043, "learning_rate": 1.826305638150213e-05, "loss": 1.0486, "step": 1116 }, { "epoch": 0.2149006781780578, "grad_norm": 1.3859235483748127, "learning_rate": 1.825954473143205e-05, "loss": 0.998, "step": 1117 }, { "epoch": 0.2150930691164446, "grad_norm": 1.6281637465251093, "learning_rate": 1.8256029873449976e-05, "loss": 1.0506, "step": 1118 }, { "epoch": 0.21528546005483143, "grad_norm": 0.8171724106360425, "learning_rate": 1.825251180892103e-05, "loss": 0.894, "step": 1119 }, { "epoch": 0.21547785099321823, "grad_norm": 1.5178985552748194, "learning_rate": 1.8248990539211596e-05, "loss": 1.0643, "step": 1120 }, { "epoch": 0.21567024193160503, "grad_norm": 1.5016005212670094, "learning_rate": 1.8245466065689282e-05, "loss": 1.099, "step": 1121 }, { "epoch": 0.21586263286999183, "grad_norm": 0.7141949418653588, "learning_rate": 1.824193838972297e-05, "loss": 0.8723, "step": 1122 }, { "epoch": 0.21605502380837863, "grad_norm": 1.4309569499308432, "learning_rate": 1.823840751268275e-05, "loss": 1.0548, "step": 1123 }, { "epoch": 0.21624741474676543, "grad_norm": 1.4539576012247828, "learning_rate": 1.8234873435939987e-05, "loss": 0.9858, "step": 1124 }, { "epoch": 0.21643980568515223, "grad_norm": 1.5696339760380367, "learning_rate": 1.8231336160867275e-05, "loss": 1.0622, "step": 1125 }, { "epoch": 0.21663219662353903, "grad_norm": 0.7694306551612524, "learning_rate": 1.8227795688838446e-05, "loss": 0.8655, "step": 1126 }, { "epoch": 0.21682458756192582, "grad_norm": 1.4630947683777988, "learning_rate": 1.822425202122858e-05, "loss": 1.0535, "step": 1127 }, { "epoch": 0.21701697850031262, "grad_norm": 1.5515995517092422, "learning_rate": 1.8220705159413996e-05, "loss": 1.0825, "step": 1128 }, { "epoch": 0.21720936943869945, "grad_norm": 1.5122834083250793, "learning_rate": 1.8217155104772256e-05, "loss": 0.9496, "step": 1129 }, { "epoch": 0.21740176037708625, "grad_norm": 1.4903777628549906, "learning_rate": 1.8213601858682158e-05, "loss": 1.0776, "step": 1130 }, { "epoch": 0.21759415131547305, "grad_norm": 1.6053111286488595, "learning_rate": 1.8210045422523744e-05, "loss": 0.997, "step": 1131 }, { "epoch": 0.21778654225385985, "grad_norm": 1.239810580897348, "learning_rate": 1.8206485797678294e-05, "loss": 0.9772, "step": 1132 }, { "epoch": 0.21797893319224665, "grad_norm": 1.3556230175760804, "learning_rate": 1.820292298552832e-05, "loss": 1.0398, "step": 1133 }, { "epoch": 0.21817132413063345, "grad_norm": 1.500717868639847, "learning_rate": 1.819935698745759e-05, "loss": 1.0882, "step": 1134 }, { "epoch": 0.21836371506902025, "grad_norm": 1.4486034179164367, "learning_rate": 1.8195787804851076e-05, "loss": 0.974, "step": 1135 }, { "epoch": 0.21855610600740705, "grad_norm": 1.387934205353269, "learning_rate": 1.8192215439095025e-05, "loss": 1.0669, "step": 1136 }, { "epoch": 0.21874849694579385, "grad_norm": 1.5153615880072413, "learning_rate": 1.8188639891576893e-05, "loss": 1.0689, "step": 1137 }, { "epoch": 0.21894088788418065, "grad_norm": 1.264765917330258, "learning_rate": 1.8185061163685386e-05, "loss": 0.9698, "step": 1138 }, { "epoch": 0.21913327882256745, "grad_norm": 1.494635978250613, "learning_rate": 1.818147925681044e-05, "loss": 1.1016, "step": 1139 }, { "epoch": 0.21932566976095427, "grad_norm": 0.727547796748458, "learning_rate": 1.8177894172343227e-05, "loss": 0.8266, "step": 1140 }, { "epoch": 0.21951806069934107, "grad_norm": 1.5627488596804806, "learning_rate": 1.817430591167615e-05, "loss": 1.004, "step": 1141 }, { "epoch": 0.21971045163772787, "grad_norm": 1.4284905171684972, "learning_rate": 1.8170714476202848e-05, "loss": 0.9483, "step": 1142 }, { "epoch": 0.21990284257611467, "grad_norm": 1.60090738553087, "learning_rate": 1.8167119867318197e-05, "loss": 1.0395, "step": 1143 }, { "epoch": 0.22009523351450147, "grad_norm": 1.4738766650813255, "learning_rate": 1.81635220864183e-05, "loss": 1.1381, "step": 1144 }, { "epoch": 0.22028762445288827, "grad_norm": 1.5672724019066044, "learning_rate": 1.8159921134900486e-05, "loss": 1.0697, "step": 1145 }, { "epoch": 0.22048001539127507, "grad_norm": 1.473458147290815, "learning_rate": 1.8156317014163337e-05, "loss": 1.1163, "step": 1146 }, { "epoch": 0.22067240632966187, "grad_norm": 1.437273818147347, "learning_rate": 1.8152709725606642e-05, "loss": 1.0451, "step": 1147 }, { "epoch": 0.22086479726804867, "grad_norm": 1.5647260482221546, "learning_rate": 1.8149099270631434e-05, "loss": 1.0694, "step": 1148 }, { "epoch": 0.22105718820643547, "grad_norm": 1.4835054188188757, "learning_rate": 1.8145485650639973e-05, "loss": 1.0265, "step": 1149 }, { "epoch": 0.22124957914482227, "grad_norm": 1.651451694505899, "learning_rate": 1.8141868867035745e-05, "loss": 1.0734, "step": 1150 }, { "epoch": 0.2214419700832091, "grad_norm": 1.2633343277822378, "learning_rate": 1.8138248921223465e-05, "loss": 1.0811, "step": 1151 }, { "epoch": 0.2216343610215959, "grad_norm": 1.3650972995555135, "learning_rate": 1.8134625814609084e-05, "loss": 0.9953, "step": 1152 }, { "epoch": 0.2218267519599827, "grad_norm": 1.4343751581186952, "learning_rate": 1.8130999548599767e-05, "loss": 1.0428, "step": 1153 }, { "epoch": 0.2220191428983695, "grad_norm": 1.4418628578971313, "learning_rate": 1.8127370124603927e-05, "loss": 0.9932, "step": 1154 }, { "epoch": 0.2222115338367563, "grad_norm": 1.6117876549439671, "learning_rate": 1.8123737544031178e-05, "loss": 1.0617, "step": 1155 }, { "epoch": 0.2224039247751431, "grad_norm": 1.5060096508602219, "learning_rate": 1.8120101808292373e-05, "loss": 0.9574, "step": 1156 }, { "epoch": 0.2225963157135299, "grad_norm": 1.5423668070747891, "learning_rate": 1.81164629187996e-05, "loss": 1.0591, "step": 1157 }, { "epoch": 0.2227887066519167, "grad_norm": 0.7913029984442924, "learning_rate": 1.811282087696615e-05, "loss": 0.8788, "step": 1158 }, { "epoch": 0.2229810975903035, "grad_norm": 1.4351014408980591, "learning_rate": 1.8109175684206558e-05, "loss": 1.0679, "step": 1159 }, { "epoch": 0.2231734885286903, "grad_norm": 1.6169477603025917, "learning_rate": 1.8105527341936574e-05, "loss": 1.0302, "step": 1160 }, { "epoch": 0.2233658794670771, "grad_norm": 1.5453030617959151, "learning_rate": 1.810187585157317e-05, "loss": 1.0554, "step": 1161 }, { "epoch": 0.22355827040546392, "grad_norm": 1.4607267522307412, "learning_rate": 1.8098221214534543e-05, "loss": 1.0728, "step": 1162 }, { "epoch": 0.22375066134385072, "grad_norm": 1.4171333784879798, "learning_rate": 1.8094563432240107e-05, "loss": 1.0346, "step": 1163 }, { "epoch": 0.22394305228223751, "grad_norm": 1.5318659231064777, "learning_rate": 1.8090902506110513e-05, "loss": 1.065, "step": 1164 }, { "epoch": 0.22413544322062431, "grad_norm": 1.4507182866934505, "learning_rate": 1.8087238437567614e-05, "loss": 0.9934, "step": 1165 }, { "epoch": 0.2243278341590111, "grad_norm": 1.4062934814861385, "learning_rate": 1.8083571228034498e-05, "loss": 1.0098, "step": 1166 }, { "epoch": 0.2245202250973979, "grad_norm": 1.476736341564346, "learning_rate": 1.807990087893546e-05, "loss": 1.0204, "step": 1167 }, { "epoch": 0.2247126160357847, "grad_norm": 1.3432456952752612, "learning_rate": 1.807622739169603e-05, "loss": 1.0375, "step": 1168 }, { "epoch": 0.2249050069741715, "grad_norm": 1.2629110501395215, "learning_rate": 1.807255076774294e-05, "loss": 1.0517, "step": 1169 }, { "epoch": 0.2250973979125583, "grad_norm": 1.461618441510121, "learning_rate": 1.8068871008504153e-05, "loss": 1.0188, "step": 1170 }, { "epoch": 0.2252897888509451, "grad_norm": 1.4048288438030587, "learning_rate": 1.8065188115408844e-05, "loss": 1.0162, "step": 1171 }, { "epoch": 0.2254821797893319, "grad_norm": 0.7693811283767767, "learning_rate": 1.8061502089887406e-05, "loss": 0.7766, "step": 1172 }, { "epoch": 0.22567457072771874, "grad_norm": 1.4810620941804122, "learning_rate": 1.805781293337145e-05, "loss": 1.0469, "step": 1173 }, { "epoch": 0.22586696166610554, "grad_norm": 1.5435789610317963, "learning_rate": 1.8054120647293798e-05, "loss": 1.0093, "step": 1174 }, { "epoch": 0.22605935260449234, "grad_norm": 1.5228183427701976, "learning_rate": 1.8050425233088496e-05, "loss": 1.0558, "step": 1175 }, { "epoch": 0.22625174354287914, "grad_norm": 1.6399015736538989, "learning_rate": 1.80467266921908e-05, "loss": 1.1018, "step": 1176 }, { "epoch": 0.22644413448126594, "grad_norm": 1.3164372306585983, "learning_rate": 1.8043025026037178e-05, "loss": 0.9752, "step": 1177 }, { "epoch": 0.22663652541965273, "grad_norm": 1.3563248190105006, "learning_rate": 1.8039320236065314e-05, "loss": 1.032, "step": 1178 }, { "epoch": 0.22682891635803953, "grad_norm": 1.5647756194144178, "learning_rate": 1.803561232371411e-05, "loss": 0.9416, "step": 1179 }, { "epoch": 0.22702130729642633, "grad_norm": 1.5156138411578042, "learning_rate": 1.803190129042367e-05, "loss": 0.9661, "step": 1180 }, { "epoch": 0.22721369823481313, "grad_norm": 1.5482636429413674, "learning_rate": 1.8028187137635325e-05, "loss": 1.0701, "step": 1181 }, { "epoch": 0.22740608917319993, "grad_norm": 1.5195889270121168, "learning_rate": 1.8024469866791602e-05, "loss": 0.9858, "step": 1182 }, { "epoch": 0.22759848011158673, "grad_norm": 1.5190305402866175, "learning_rate": 1.802074947933625e-05, "loss": 1.0401, "step": 1183 }, { "epoch": 0.22779087104997356, "grad_norm": 1.423826236720167, "learning_rate": 1.801702597671422e-05, "loss": 0.9539, "step": 1184 }, { "epoch": 0.22798326198836036, "grad_norm": 1.658838451091318, "learning_rate": 1.8013299360371685e-05, "loss": 1.0595, "step": 1185 }, { "epoch": 0.22817565292674716, "grad_norm": 1.5283792118505235, "learning_rate": 1.8009569631756013e-05, "loss": 1.0218, "step": 1186 }, { "epoch": 0.22836804386513396, "grad_norm": 1.380667036360989, "learning_rate": 1.8005836792315793e-05, "loss": 1.0678, "step": 1187 }, { "epoch": 0.22856043480352076, "grad_norm": 1.4623659995624196, "learning_rate": 1.800210084350081e-05, "loss": 1.0054, "step": 1188 }, { "epoch": 0.22875282574190756, "grad_norm": 1.5180266184818914, "learning_rate": 1.799836178676207e-05, "loss": 1.0347, "step": 1189 }, { "epoch": 0.22894521668029436, "grad_norm": 1.5099953146180676, "learning_rate": 1.799461962355178e-05, "loss": 1.05, "step": 1190 }, { "epoch": 0.22913760761868116, "grad_norm": 1.4548522080125614, "learning_rate": 1.7990874355323345e-05, "loss": 1.0455, "step": 1191 }, { "epoch": 0.22932999855706795, "grad_norm": 1.3531405527939924, "learning_rate": 1.7987125983531393e-05, "loss": 1.0143, "step": 1192 }, { "epoch": 0.22952238949545475, "grad_norm": 0.7352075698840201, "learning_rate": 1.7983374509631742e-05, "loss": 0.8225, "step": 1193 }, { "epoch": 0.22971478043384155, "grad_norm": 1.6231772999345095, "learning_rate": 1.7979619935081424e-05, "loss": 0.9624, "step": 1194 }, { "epoch": 0.22990717137222838, "grad_norm": 1.3706132435975826, "learning_rate": 1.797586226133867e-05, "loss": 1.0964, "step": 1195 }, { "epoch": 0.23009956231061518, "grad_norm": 1.745336374611114, "learning_rate": 1.7972101489862924e-05, "loss": 1.0835, "step": 1196 }, { "epoch": 0.23029195324900198, "grad_norm": 1.4538493130358998, "learning_rate": 1.7968337622114824e-05, "loss": 1.0333, "step": 1197 }, { "epoch": 0.23048434418738878, "grad_norm": 1.4681760408608426, "learning_rate": 1.7964570659556206e-05, "loss": 1.0905, "step": 1198 }, { "epoch": 0.23067673512577558, "grad_norm": 1.3110094138510802, "learning_rate": 1.796080060365012e-05, "loss": 0.9168, "step": 1199 }, { "epoch": 0.23086912606416238, "grad_norm": 1.3722863768617009, "learning_rate": 1.7957027455860815e-05, "loss": 1.0918, "step": 1200 }, { "epoch": 0.23106151700254918, "grad_norm": 1.4167677819777604, "learning_rate": 1.795325121765373e-05, "loss": 1.093, "step": 1201 }, { "epoch": 0.23125390794093598, "grad_norm": 1.462531327517025, "learning_rate": 1.794947189049552e-05, "loss": 0.9806, "step": 1202 }, { "epoch": 0.23144629887932278, "grad_norm": 1.6021866138638379, "learning_rate": 1.7945689475854033e-05, "loss": 1.0449, "step": 1203 }, { "epoch": 0.23163868981770958, "grad_norm": 1.49411370073779, "learning_rate": 1.7941903975198305e-05, "loss": 0.991, "step": 1204 }, { "epoch": 0.23183108075609637, "grad_norm": 1.5923366446492513, "learning_rate": 1.7938115389998595e-05, "loss": 1.0396, "step": 1205 }, { "epoch": 0.2320234716944832, "grad_norm": 1.6532608661446135, "learning_rate": 1.7934323721726334e-05, "loss": 0.9632, "step": 1206 }, { "epoch": 0.23221586263287, "grad_norm": 1.4411759192012579, "learning_rate": 1.7930528971854166e-05, "loss": 1.0291, "step": 1207 }, { "epoch": 0.2324082535712568, "grad_norm": 1.3426640524434894, "learning_rate": 1.792673114185593e-05, "loss": 1.0902, "step": 1208 }, { "epoch": 0.2326006445096436, "grad_norm": 1.5615263878659953, "learning_rate": 1.7922930233206656e-05, "loss": 0.9974, "step": 1209 }, { "epoch": 0.2327930354480304, "grad_norm": 1.5098102737651575, "learning_rate": 1.7919126247382576e-05, "loss": 1.0251, "step": 1210 }, { "epoch": 0.2329854263864172, "grad_norm": 1.4469833128094984, "learning_rate": 1.791531918586112e-05, "loss": 1.0095, "step": 1211 }, { "epoch": 0.233177817324804, "grad_norm": 1.4749956102581063, "learning_rate": 1.7911509050120892e-05, "loss": 1.0605, "step": 1212 }, { "epoch": 0.2333702082631908, "grad_norm": 1.442165901499154, "learning_rate": 1.7907695841641716e-05, "loss": 1.0328, "step": 1213 }, { "epoch": 0.2335625992015776, "grad_norm": 1.483044953510192, "learning_rate": 1.7903879561904597e-05, "loss": 1.1383, "step": 1214 }, { "epoch": 0.2337549901399644, "grad_norm": 1.3651031343301963, "learning_rate": 1.790006021239173e-05, "loss": 0.9872, "step": 1215 }, { "epoch": 0.2339473810783512, "grad_norm": 1.4725662166329792, "learning_rate": 1.789623779458651e-05, "loss": 1.0033, "step": 1216 }, { "epoch": 0.23413977201673802, "grad_norm": 1.4971699649731887, "learning_rate": 1.789241230997352e-05, "loss": 1.0108, "step": 1217 }, { "epoch": 0.23433216295512482, "grad_norm": 1.3799070264569053, "learning_rate": 1.7888583760038534e-05, "loss": 1.0177, "step": 1218 }, { "epoch": 0.23452455389351162, "grad_norm": 1.4711086095248433, "learning_rate": 1.7884752146268513e-05, "loss": 1.0252, "step": 1219 }, { "epoch": 0.23471694483189842, "grad_norm": 1.4809053069972204, "learning_rate": 1.7880917470151614e-05, "loss": 1.0532, "step": 1220 }, { "epoch": 0.23490933577028522, "grad_norm": 1.5721823947408944, "learning_rate": 1.7877079733177185e-05, "loss": 1.0303, "step": 1221 }, { "epoch": 0.23510172670867202, "grad_norm": 1.352211656845598, "learning_rate": 1.7873238936835754e-05, "loss": 1.0448, "step": 1222 }, { "epoch": 0.23529411764705882, "grad_norm": 1.5500802015369788, "learning_rate": 1.786939508261904e-05, "loss": 1.046, "step": 1223 }, { "epoch": 0.23548650858544562, "grad_norm": 1.3375113508571501, "learning_rate": 1.786554817201996e-05, "loss": 1.0171, "step": 1224 }, { "epoch": 0.23567889952383242, "grad_norm": 1.4638072302960075, "learning_rate": 1.78616982065326e-05, "loss": 1.0359, "step": 1225 }, { "epoch": 0.23587129046221922, "grad_norm": 1.454627817316177, "learning_rate": 1.785784518765225e-05, "loss": 1.0811, "step": 1226 }, { "epoch": 0.23606368140060602, "grad_norm": 1.3994216085067837, "learning_rate": 1.7853989116875373e-05, "loss": 1.0707, "step": 1227 }, { "epoch": 0.23625607233899285, "grad_norm": 1.2900228665446662, "learning_rate": 1.7850129995699626e-05, "loss": 0.9778, "step": 1228 }, { "epoch": 0.23644846327737964, "grad_norm": 1.3903992917218064, "learning_rate": 1.7846267825623843e-05, "loss": 1.0076, "step": 1229 }, { "epoch": 0.23664085421576644, "grad_norm": 1.5314129623118709, "learning_rate": 1.7842402608148053e-05, "loss": 1.1026, "step": 1230 }, { "epoch": 0.23683324515415324, "grad_norm": 1.2865681871758765, "learning_rate": 1.7838534344773453e-05, "loss": 0.9834, "step": 1231 }, { "epoch": 0.23702563609254004, "grad_norm": 1.4408709500606829, "learning_rate": 1.7834663037002444e-05, "loss": 1.0745, "step": 1232 }, { "epoch": 0.23721802703092684, "grad_norm": 1.5689863280697538, "learning_rate": 1.7830788686338586e-05, "loss": 1.0232, "step": 1233 }, { "epoch": 0.23741041796931364, "grad_norm": 1.4002780462196118, "learning_rate": 1.7826911294286636e-05, "loss": 1.1255, "step": 1234 }, { "epoch": 0.23760280890770044, "grad_norm": 1.4385969932949005, "learning_rate": 1.782303086235253e-05, "loss": 1.0826, "step": 1235 }, { "epoch": 0.23779519984608724, "grad_norm": 1.4191651634491074, "learning_rate": 1.781914739204338e-05, "loss": 1.1376, "step": 1236 }, { "epoch": 0.23798759078447404, "grad_norm": 1.461240208116727, "learning_rate": 1.7815260884867486e-05, "loss": 0.9685, "step": 1237 }, { "epoch": 0.23817998172286084, "grad_norm": 1.571606734114084, "learning_rate": 1.781137134233432e-05, "loss": 1.0585, "step": 1238 }, { "epoch": 0.23837237266124767, "grad_norm": 1.4562746065878358, "learning_rate": 1.7807478765954532e-05, "loss": 1.0425, "step": 1239 }, { "epoch": 0.23856476359963447, "grad_norm": 1.546835245933223, "learning_rate": 1.7803583157239958e-05, "loss": 0.9873, "step": 1240 }, { "epoch": 0.23875715453802127, "grad_norm": 1.4199528812492046, "learning_rate": 1.7799684517703605e-05, "loss": 1.1, "step": 1241 }, { "epoch": 0.23894954547640807, "grad_norm": 1.5342081777243528, "learning_rate": 1.779578284885966e-05, "loss": 1.0406, "step": 1242 }, { "epoch": 0.23914193641479486, "grad_norm": 1.382995508981523, "learning_rate": 1.779187815222349e-05, "loss": 1.0376, "step": 1243 }, { "epoch": 0.23933432735318166, "grad_norm": 1.3391974578402686, "learning_rate": 1.778797042931163e-05, "loss": 1.0314, "step": 1244 }, { "epoch": 0.23952671829156846, "grad_norm": 1.339305262223316, "learning_rate": 1.7784059681641798e-05, "loss": 0.9874, "step": 1245 }, { "epoch": 0.23971910922995526, "grad_norm": 1.452325720752906, "learning_rate": 1.778014591073288e-05, "loss": 1.0131, "step": 1246 }, { "epoch": 0.23991150016834206, "grad_norm": 1.4683542869288926, "learning_rate": 1.777622911810494e-05, "loss": 1.0254, "step": 1247 }, { "epoch": 0.24010389110672886, "grad_norm": 1.4226730605110078, "learning_rate": 1.777230930527922e-05, "loss": 0.9919, "step": 1248 }, { "epoch": 0.2402962820451157, "grad_norm": 1.3477746654098457, "learning_rate": 1.7768386473778124e-05, "loss": 1.0363, "step": 1249 }, { "epoch": 0.2404886729835025, "grad_norm": 1.37912675664878, "learning_rate": 1.7764460625125236e-05, "loss": 1.0474, "step": 1250 }, { "epoch": 0.2406810639218893, "grad_norm": 1.566597795103946, "learning_rate": 1.776053176084531e-05, "loss": 1.037, "step": 1251 }, { "epoch": 0.2408734548602761, "grad_norm": 1.577147894459967, "learning_rate": 1.7756599882464274e-05, "loss": 1.0833, "step": 1252 }, { "epoch": 0.2410658457986629, "grad_norm": 0.811166996581583, "learning_rate": 1.7752664991509224e-05, "loss": 0.893, "step": 1253 }, { "epoch": 0.24125823673704969, "grad_norm": 1.432378755885441, "learning_rate": 1.7748727089508423e-05, "loss": 0.9291, "step": 1254 }, { "epoch": 0.24145062767543649, "grad_norm": 1.5747202334079584, "learning_rate": 1.7744786177991307e-05, "loss": 1.0599, "step": 1255 }, { "epoch": 0.24164301861382328, "grad_norm": 1.6579278140790383, "learning_rate": 1.774084225848849e-05, "loss": 1.0663, "step": 1256 }, { "epoch": 0.24183540955221008, "grad_norm": 1.4731256195264997, "learning_rate": 1.773689533253173e-05, "loss": 1.0782, "step": 1257 }, { "epoch": 0.24202780049059688, "grad_norm": 1.4410968508089546, "learning_rate": 1.7732945401653978e-05, "loss": 1.0884, "step": 1258 }, { "epoch": 0.24222019142898368, "grad_norm": 1.4532027939399814, "learning_rate": 1.7728992467389342e-05, "loss": 1.0293, "step": 1259 }, { "epoch": 0.2424125823673705, "grad_norm": 1.4953252629173928, "learning_rate": 1.7725036531273087e-05, "loss": 1.0013, "step": 1260 }, { "epoch": 0.2426049733057573, "grad_norm": 1.2563076656884689, "learning_rate": 1.7721077594841663e-05, "loss": 1.0068, "step": 1261 }, { "epoch": 0.2427973642441441, "grad_norm": 1.5093401902702792, "learning_rate": 1.771711565963267e-05, "loss": 1.0001, "step": 1262 }, { "epoch": 0.2429897551825309, "grad_norm": 0.7535535919891584, "learning_rate": 1.7713150727184878e-05, "loss": 0.8412, "step": 1263 }, { "epoch": 0.2431821461209177, "grad_norm": 1.3408877041795604, "learning_rate": 1.770918279903822e-05, "loss": 0.9918, "step": 1264 }, { "epoch": 0.2433745370593045, "grad_norm": 1.4665890318310413, "learning_rate": 1.77052118767338e-05, "loss": 1.0311, "step": 1265 }, { "epoch": 0.2435669279976913, "grad_norm": 1.316744882496198, "learning_rate": 1.7701237961813874e-05, "loss": 1.0737, "step": 1266 }, { "epoch": 0.2437593189360781, "grad_norm": 1.3886020785293887, "learning_rate": 1.7697261055821864e-05, "loss": 0.9684, "step": 1267 }, { "epoch": 0.2439517098744649, "grad_norm": 1.3840834126414299, "learning_rate": 1.7693281160302354e-05, "loss": 0.9447, "step": 1268 }, { "epoch": 0.2441441008128517, "grad_norm": 1.500113462129856, "learning_rate": 1.7689298276801095e-05, "loss": 1.0362, "step": 1269 }, { "epoch": 0.2443364917512385, "grad_norm": 1.489131446884849, "learning_rate": 1.7685312406864986e-05, "loss": 1.0869, "step": 1270 }, { "epoch": 0.24452888268962533, "grad_norm": 1.298899513696176, "learning_rate": 1.7681323552042094e-05, "loss": 1.0307, "step": 1271 }, { "epoch": 0.24472127362801213, "grad_norm": 1.3963441547643658, "learning_rate": 1.767733171388165e-05, "loss": 1.0398, "step": 1272 }, { "epoch": 0.24491366456639893, "grad_norm": 1.4756814915966159, "learning_rate": 1.7673336893934033e-05, "loss": 1.0691, "step": 1273 }, { "epoch": 0.24510605550478573, "grad_norm": 1.3765330811499443, "learning_rate": 1.7669339093750786e-05, "loss": 1.0811, "step": 1274 }, { "epoch": 0.24529844644317253, "grad_norm": 1.40006673727285, "learning_rate": 1.766533831488461e-05, "loss": 1.1058, "step": 1275 }, { "epoch": 0.24549083738155933, "grad_norm": 1.7282653585425836, "learning_rate": 1.7661334558889357e-05, "loss": 0.9823, "step": 1276 }, { "epoch": 0.24568322831994613, "grad_norm": 1.4568149687739653, "learning_rate": 1.7657327827320046e-05, "loss": 1.0947, "step": 1277 }, { "epoch": 0.24587561925833293, "grad_norm": 1.3259935527877253, "learning_rate": 1.765331812173284e-05, "loss": 1.015, "step": 1278 }, { "epoch": 0.24606801019671973, "grad_norm": 1.3801309445753056, "learning_rate": 1.7649305443685068e-05, "loss": 1.1103, "step": 1279 }, { "epoch": 0.24626040113510653, "grad_norm": 0.7424625694570207, "learning_rate": 1.76452897947352e-05, "loss": 0.854, "step": 1280 }, { "epoch": 0.24645279207349333, "grad_norm": 0.7693359937556596, "learning_rate": 1.7641271176442876e-05, "loss": 0.85, "step": 1281 }, { "epoch": 0.24664518301188015, "grad_norm": 1.6007427924918467, "learning_rate": 1.7637249590368878e-05, "loss": 1.1028, "step": 1282 }, { "epoch": 0.24683757395026695, "grad_norm": 1.489172918598833, "learning_rate": 1.763322503807514e-05, "loss": 1.0102, "step": 1283 }, { "epoch": 0.24702996488865375, "grad_norm": 1.575597987241676, "learning_rate": 1.7629197521124758e-05, "loss": 1.1052, "step": 1284 }, { "epoch": 0.24722235582704055, "grad_norm": 1.538573302258429, "learning_rate": 1.7625167041081967e-05, "loss": 1.0485, "step": 1285 }, { "epoch": 0.24741474676542735, "grad_norm": 1.4983842251469686, "learning_rate": 1.7621133599512163e-05, "loss": 1.048, "step": 1286 }, { "epoch": 0.24760713770381415, "grad_norm": 1.5873449679857088, "learning_rate": 1.761709719798189e-05, "loss": 0.9648, "step": 1287 }, { "epoch": 0.24779952864220095, "grad_norm": 1.5451107590199638, "learning_rate": 1.761305783805883e-05, "loss": 1.0318, "step": 1288 }, { "epoch": 0.24799191958058775, "grad_norm": 1.4104865155782187, "learning_rate": 1.7609015521311836e-05, "loss": 1.118, "step": 1289 }, { "epoch": 0.24818431051897455, "grad_norm": 1.5207821011683391, "learning_rate": 1.7604970249310893e-05, "loss": 0.976, "step": 1290 }, { "epoch": 0.24837670145736135, "grad_norm": 1.4116224538216846, "learning_rate": 1.7600922023627137e-05, "loss": 1.078, "step": 1291 }, { "epoch": 0.24856909239574815, "grad_norm": 1.4727367843095893, "learning_rate": 1.759687084583285e-05, "loss": 1.0955, "step": 1292 }, { "epoch": 0.24876148333413498, "grad_norm": 1.4889226455734998, "learning_rate": 1.759281671750147e-05, "loss": 0.9901, "step": 1293 }, { "epoch": 0.24895387427252177, "grad_norm": 1.3863227275328827, "learning_rate": 1.7588759640207564e-05, "loss": 1.0792, "step": 1294 }, { "epoch": 0.24914626521090857, "grad_norm": 1.357711001144357, "learning_rate": 1.7584699615526857e-05, "loss": 1.1102, "step": 1295 }, { "epoch": 0.24933865614929537, "grad_norm": 1.3384759274735385, "learning_rate": 1.7580636645036224e-05, "loss": 1.0578, "step": 1296 }, { "epoch": 0.24953104708768217, "grad_norm": 1.35750756025264, "learning_rate": 1.757657073031367e-05, "loss": 1.0074, "step": 1297 }, { "epoch": 0.24972343802606897, "grad_norm": 1.4522513433227695, "learning_rate": 1.7572501872938343e-05, "loss": 0.9924, "step": 1298 }, { "epoch": 0.24991582896445577, "grad_norm": 1.513979756519628, "learning_rate": 1.756843007449055e-05, "loss": 1.0121, "step": 1299 }, { "epoch": 0.2501082199028426, "grad_norm": 1.5287657973901576, "learning_rate": 1.7564355336551727e-05, "loss": 1.0465, "step": 1300 }, { "epoch": 0.2503006108412294, "grad_norm": 0.8224832805082704, "learning_rate": 1.7560277660704455e-05, "loss": 0.8311, "step": 1301 }, { "epoch": 0.2504930017796162, "grad_norm": 1.6589487165241814, "learning_rate": 1.755619704853246e-05, "loss": 1.0597, "step": 1302 }, { "epoch": 0.250685392718003, "grad_norm": 1.4814964498168535, "learning_rate": 1.7552113501620595e-05, "loss": 1.0112, "step": 1303 }, { "epoch": 0.2508777836563898, "grad_norm": 0.7525312229130054, "learning_rate": 1.7548027021554874e-05, "loss": 0.8731, "step": 1304 }, { "epoch": 0.2510701745947766, "grad_norm": 1.5195009888642992, "learning_rate": 1.754393760992243e-05, "loss": 1.0945, "step": 1305 }, { "epoch": 0.2512625655331634, "grad_norm": 1.5076653439427619, "learning_rate": 1.7539845268311548e-05, "loss": 1.0394, "step": 1306 }, { "epoch": 0.2514549564715502, "grad_norm": 1.5833940924127787, "learning_rate": 1.7535749998311645e-05, "loss": 1.0754, "step": 1307 }, { "epoch": 0.251647347409937, "grad_norm": 1.4554047681359819, "learning_rate": 1.753165180151328e-05, "loss": 1.0897, "step": 1308 }, { "epoch": 0.2518397383483238, "grad_norm": 1.5286953014213485, "learning_rate": 1.752755067950814e-05, "loss": 1.1206, "step": 1309 }, { "epoch": 0.2520321292867106, "grad_norm": 1.5644126338409432, "learning_rate": 1.752344663388906e-05, "loss": 0.986, "step": 1310 }, { "epoch": 0.2522245202250974, "grad_norm": 1.399273675139637, "learning_rate": 1.7519339666249997e-05, "loss": 0.9773, "step": 1311 }, { "epoch": 0.2524169111634842, "grad_norm": 1.528144178267268, "learning_rate": 1.7515229778186052e-05, "loss": 1.0575, "step": 1312 }, { "epoch": 0.252609302101871, "grad_norm": 1.4666185240235012, "learning_rate": 1.7511116971293463e-05, "loss": 1.072, "step": 1313 }, { "epoch": 0.2528016930402578, "grad_norm": 1.4874718842979089, "learning_rate": 1.7507001247169587e-05, "loss": 1.1143, "step": 1314 }, { "epoch": 0.2529940839786446, "grad_norm": 1.531427427127143, "learning_rate": 1.7502882607412933e-05, "loss": 1.0221, "step": 1315 }, { "epoch": 0.2531864749170314, "grad_norm": 1.375049493719332, "learning_rate": 1.749876105362313e-05, "loss": 0.9933, "step": 1316 }, { "epoch": 0.2533788658554182, "grad_norm": 1.493255967685993, "learning_rate": 1.7494636587400942e-05, "loss": 1.0234, "step": 1317 }, { "epoch": 0.253571256793805, "grad_norm": 1.545550577869909, "learning_rate": 1.749050921034826e-05, "loss": 1.0516, "step": 1318 }, { "epoch": 0.2537636477321918, "grad_norm": 1.5233366584818613, "learning_rate": 1.7486378924068123e-05, "loss": 1.0299, "step": 1319 }, { "epoch": 0.25395603867057864, "grad_norm": 1.648423009004788, "learning_rate": 1.748224573016467e-05, "loss": 1.1394, "step": 1320 }, { "epoch": 0.25414842960896544, "grad_norm": 1.4761511153615716, "learning_rate": 1.7478109630243195e-05, "loss": 1.0235, "step": 1321 }, { "epoch": 0.25434082054735224, "grad_norm": 0.8040582990711634, "learning_rate": 1.747397062591011e-05, "loss": 0.8341, "step": 1322 }, { "epoch": 0.25453321148573904, "grad_norm": 1.4576657169247718, "learning_rate": 1.746982871877296e-05, "loss": 1.0097, "step": 1323 }, { "epoch": 0.25472560242412584, "grad_norm": 1.532010662410696, "learning_rate": 1.7465683910440405e-05, "loss": 1.0491, "step": 1324 }, { "epoch": 0.25491799336251264, "grad_norm": 1.3629717934955854, "learning_rate": 1.7461536202522248e-05, "loss": 0.9974, "step": 1325 }, { "epoch": 0.25511038430089944, "grad_norm": 1.4528481025839295, "learning_rate": 1.745738559662941e-05, "loss": 1.1075, "step": 1326 }, { "epoch": 0.25530277523928624, "grad_norm": 1.44955153929119, "learning_rate": 1.7453232094373936e-05, "loss": 0.9523, "step": 1327 }, { "epoch": 0.25549516617767304, "grad_norm": 1.4441556568183946, "learning_rate": 1.7449075697369005e-05, "loss": 1.0936, "step": 1328 }, { "epoch": 0.25568755711605984, "grad_norm": 1.486758231603601, "learning_rate": 1.7444916407228904e-05, "loss": 1.026, "step": 1329 }, { "epoch": 0.25587994805444664, "grad_norm": 1.4111192779673063, "learning_rate": 1.744075422556906e-05, "loss": 0.9851, "step": 1330 }, { "epoch": 0.25607233899283344, "grad_norm": 1.3322593011438488, "learning_rate": 1.7436589154006014e-05, "loss": 0.9852, "step": 1331 }, { "epoch": 0.25626472993122024, "grad_norm": 1.4113849112142935, "learning_rate": 1.743242119415743e-05, "loss": 0.967, "step": 1332 }, { "epoch": 0.25645712086960704, "grad_norm": 1.4498481173829527, "learning_rate": 1.7428250347642102e-05, "loss": 1.05, "step": 1333 }, { "epoch": 0.25664951180799384, "grad_norm": 1.519749407773537, "learning_rate": 1.7424076616079933e-05, "loss": 1.0512, "step": 1334 }, { "epoch": 0.25684190274638063, "grad_norm": 1.4435940116366208, "learning_rate": 1.7419900001091953e-05, "loss": 1.09, "step": 1335 }, { "epoch": 0.25703429368476743, "grad_norm": 1.6515648444244102, "learning_rate": 1.7415720504300314e-05, "loss": 1.0639, "step": 1336 }, { "epoch": 0.25722668462315423, "grad_norm": 1.40922454266578, "learning_rate": 1.741153812732828e-05, "loss": 1.0876, "step": 1337 }, { "epoch": 0.25741907556154103, "grad_norm": 1.5830091973319411, "learning_rate": 1.7407352871800246e-05, "loss": 0.9996, "step": 1338 }, { "epoch": 0.25761146649992783, "grad_norm": 1.281772691194008, "learning_rate": 1.7403164739341708e-05, "loss": 1.019, "step": 1339 }, { "epoch": 0.25780385743831463, "grad_norm": 1.4470628724049217, "learning_rate": 1.739897373157929e-05, "loss": 1.0666, "step": 1340 }, { "epoch": 0.25799624837670143, "grad_norm": 1.4796892976233238, "learning_rate": 1.7394779850140736e-05, "loss": 1.0783, "step": 1341 }, { "epoch": 0.2581886393150883, "grad_norm": 1.4366775622154593, "learning_rate": 1.7390583096654895e-05, "loss": 1.0499, "step": 1342 }, { "epoch": 0.2583810302534751, "grad_norm": 1.5411305934059698, "learning_rate": 1.7386383472751745e-05, "loss": 1.0411, "step": 1343 }, { "epoch": 0.2585734211918619, "grad_norm": 1.4272761878303448, "learning_rate": 1.7382180980062365e-05, "loss": 1.0526, "step": 1344 }, { "epoch": 0.2587658121302487, "grad_norm": 1.4741469035533414, "learning_rate": 1.7377975620218954e-05, "loss": 1.0433, "step": 1345 }, { "epoch": 0.2589582030686355, "grad_norm": 1.4129175019367946, "learning_rate": 1.7373767394854836e-05, "loss": 1.1119, "step": 1346 }, { "epoch": 0.2591505940070223, "grad_norm": 1.5557673859687484, "learning_rate": 1.7369556305604422e-05, "loss": 1.0242, "step": 1347 }, { "epoch": 0.2593429849454091, "grad_norm": 1.5506428733049455, "learning_rate": 1.736534235410326e-05, "loss": 0.9888, "step": 1348 }, { "epoch": 0.2595353758837959, "grad_norm": 1.4590070129364017, "learning_rate": 1.7361125541988e-05, "loss": 0.9764, "step": 1349 }, { "epoch": 0.2597277668221827, "grad_norm": 1.6535721593806332, "learning_rate": 1.7356905870896407e-05, "loss": 1.1552, "step": 1350 }, { "epoch": 0.2599201577605695, "grad_norm": 1.4015363795183016, "learning_rate": 1.735268334246734e-05, "loss": 1.0682, "step": 1351 }, { "epoch": 0.2601125486989563, "grad_norm": 0.8395271070481146, "learning_rate": 1.7348457958340792e-05, "loss": 0.8939, "step": 1352 }, { "epoch": 0.2603049396373431, "grad_norm": 1.3714895057985923, "learning_rate": 1.7344229720157846e-05, "loss": 1.0595, "step": 1353 }, { "epoch": 0.2604973305757299, "grad_norm": 1.5241705545916044, "learning_rate": 1.7339998629560705e-05, "loss": 1.1002, "step": 1354 }, { "epoch": 0.2606897215141167, "grad_norm": 1.3702578119487046, "learning_rate": 1.7335764688192676e-05, "loss": 0.988, "step": 1355 }, { "epoch": 0.2608821124525035, "grad_norm": 1.3664293627417166, "learning_rate": 1.733152789769817e-05, "loss": 1.0865, "step": 1356 }, { "epoch": 0.2610745033908903, "grad_norm": 1.4780072396412494, "learning_rate": 1.7327288259722714e-05, "loss": 1.0259, "step": 1357 }, { "epoch": 0.2612668943292771, "grad_norm": 1.5675724926714258, "learning_rate": 1.7323045775912927e-05, "loss": 1.0838, "step": 1358 }, { "epoch": 0.2614592852676639, "grad_norm": 1.484186741286413, "learning_rate": 1.7318800447916543e-05, "loss": 1.0624, "step": 1359 }, { "epoch": 0.2616516762060507, "grad_norm": 1.472229023381236, "learning_rate": 1.7314552277382403e-05, "loss": 1.1228, "step": 1360 }, { "epoch": 0.2618440671444375, "grad_norm": 1.5240953397224064, "learning_rate": 1.7310301265960446e-05, "loss": 1.1208, "step": 1361 }, { "epoch": 0.2620364580828243, "grad_norm": 1.2236279143364845, "learning_rate": 1.7306047415301706e-05, "loss": 1.0636, "step": 1362 }, { "epoch": 0.2622288490212111, "grad_norm": 1.4829016468417042, "learning_rate": 1.7301790727058344e-05, "loss": 1.0514, "step": 1363 }, { "epoch": 0.26242123995959793, "grad_norm": 1.3699763793897557, "learning_rate": 1.7297531202883598e-05, "loss": 1.0472, "step": 1364 }, { "epoch": 0.26261363089798473, "grad_norm": 1.406983714366048, "learning_rate": 1.7293268844431826e-05, "loss": 1.0145, "step": 1365 }, { "epoch": 0.26280602183637153, "grad_norm": 1.294805444663406, "learning_rate": 1.7289003653358472e-05, "loss": 1.036, "step": 1366 }, { "epoch": 0.26299841277475833, "grad_norm": 1.364007288321854, "learning_rate": 1.7284735631320093e-05, "loss": 1.0371, "step": 1367 }, { "epoch": 0.2631908037131451, "grad_norm": 1.3980234183408744, "learning_rate": 1.7280464779974335e-05, "loss": 1.0248, "step": 1368 }, { "epoch": 0.2633831946515319, "grad_norm": 1.4307965973115049, "learning_rate": 1.7276191100979952e-05, "loss": 1.0378, "step": 1369 }, { "epoch": 0.2635755855899187, "grad_norm": 1.508545543380773, "learning_rate": 1.7271914595996784e-05, "loss": 1.0702, "step": 1370 }, { "epoch": 0.2637679765283055, "grad_norm": 1.479744385348807, "learning_rate": 1.7267635266685782e-05, "loss": 1.0811, "step": 1371 }, { "epoch": 0.2639603674666923, "grad_norm": 1.41190205304537, "learning_rate": 1.7263353114708993e-05, "loss": 1.0725, "step": 1372 }, { "epoch": 0.2641527584050791, "grad_norm": 1.5081202973643513, "learning_rate": 1.7259068141729542e-05, "loss": 0.9614, "step": 1373 }, { "epoch": 0.2643451493434659, "grad_norm": 1.4445659339249515, "learning_rate": 1.7254780349411677e-05, "loss": 0.9684, "step": 1374 }, { "epoch": 0.2645375402818527, "grad_norm": 0.728241900569844, "learning_rate": 1.7250489739420718e-05, "loss": 0.7926, "step": 1375 }, { "epoch": 0.2647299312202395, "grad_norm": 1.520007068100136, "learning_rate": 1.7246196313423095e-05, "loss": 1.1247, "step": 1376 }, { "epoch": 0.2649223221586263, "grad_norm": 1.5013752492699783, "learning_rate": 1.7241900073086318e-05, "loss": 1.0246, "step": 1377 }, { "epoch": 0.2651147130970131, "grad_norm": 1.4269051119453027, "learning_rate": 1.7237601020079003e-05, "loss": 1.0466, "step": 1378 }, { "epoch": 0.2653071040353999, "grad_norm": 1.4843565034750927, "learning_rate": 1.7233299156070852e-05, "loss": 0.9857, "step": 1379 }, { "epoch": 0.2654994949737867, "grad_norm": 1.4414317942444896, "learning_rate": 1.7228994482732653e-05, "loss": 1.0683, "step": 1380 }, { "epoch": 0.2656918859121735, "grad_norm": 1.5356764991901275, "learning_rate": 1.72246870017363e-05, "loss": 1.0257, "step": 1381 }, { "epoch": 0.2658842768505603, "grad_norm": 1.5035870799897157, "learning_rate": 1.7220376714754766e-05, "loss": 1.0888, "step": 1382 }, { "epoch": 0.2660766677889471, "grad_norm": 1.5050511387223855, "learning_rate": 1.7216063623462112e-05, "loss": 1.0089, "step": 1383 }, { "epoch": 0.2662690587273339, "grad_norm": 1.3395040968404441, "learning_rate": 1.7211747729533504e-05, "loss": 1.0472, "step": 1384 }, { "epoch": 0.2664614496657207, "grad_norm": 1.3850783882250453, "learning_rate": 1.7207429034645176e-05, "loss": 1.0448, "step": 1385 }, { "epoch": 0.2666538406041076, "grad_norm": 1.3752864788713826, "learning_rate": 1.720310754047446e-05, "loss": 1.0744, "step": 1386 }, { "epoch": 0.26684623154249437, "grad_norm": 1.3295652416773391, "learning_rate": 1.719878324869978e-05, "loss": 1.0143, "step": 1387 }, { "epoch": 0.26703862248088117, "grad_norm": 1.5593381965742104, "learning_rate": 1.7194456161000634e-05, "loss": 0.9809, "step": 1388 }, { "epoch": 0.26723101341926797, "grad_norm": 1.4642286140815226, "learning_rate": 1.719012627905762e-05, "loss": 1.0145, "step": 1389 }, { "epoch": 0.26742340435765477, "grad_norm": 1.450626154011079, "learning_rate": 1.718579360455241e-05, "loss": 1.1278, "step": 1390 }, { "epoch": 0.26761579529604157, "grad_norm": 1.4806373946239924, "learning_rate": 1.7181458139167767e-05, "loss": 1.0032, "step": 1391 }, { "epoch": 0.26780818623442837, "grad_norm": 1.3211206171220964, "learning_rate": 1.7177119884587536e-05, "loss": 1.012, "step": 1392 }, { "epoch": 0.26800057717281517, "grad_norm": 1.4903068397253014, "learning_rate": 1.717277884249664e-05, "loss": 1.0363, "step": 1393 }, { "epoch": 0.26819296811120197, "grad_norm": 1.3895472606200676, "learning_rate": 1.716843501458109e-05, "loss": 1.0352, "step": 1394 }, { "epoch": 0.26838535904958877, "grad_norm": 1.3188775633731211, "learning_rate": 1.716408840252799e-05, "loss": 0.9865, "step": 1395 }, { "epoch": 0.26857774998797557, "grad_norm": 1.2787730490485945, "learning_rate": 1.7159739008025503e-05, "loss": 1.1048, "step": 1396 }, { "epoch": 0.26877014092636237, "grad_norm": 1.5485803675035419, "learning_rate": 1.7155386832762892e-05, "loss": 1.1002, "step": 1397 }, { "epoch": 0.26896253186474917, "grad_norm": 1.4876926328985252, "learning_rate": 1.715103187843048e-05, "loss": 0.9288, "step": 1398 }, { "epoch": 0.26915492280313597, "grad_norm": 1.3737366271655933, "learning_rate": 1.7146674146719688e-05, "loss": 1.0191, "step": 1399 }, { "epoch": 0.26934731374152276, "grad_norm": 1.3117680525307718, "learning_rate": 1.7142313639323012e-05, "loss": 1.1083, "step": 1400 }, { "epoch": 0.26953970467990956, "grad_norm": 1.512901620469821, "learning_rate": 1.7137950357934017e-05, "loss": 1.0824, "step": 1401 }, { "epoch": 0.26973209561829636, "grad_norm": 1.6040423725320756, "learning_rate": 1.7133584304247354e-05, "loss": 1.0683, "step": 1402 }, { "epoch": 0.26992448655668316, "grad_norm": 1.417296718730569, "learning_rate": 1.7129215479958747e-05, "loss": 1.0379, "step": 1403 }, { "epoch": 0.27011687749506996, "grad_norm": 1.276078950449191, "learning_rate": 1.7124843886765e-05, "loss": 1.0428, "step": 1404 }, { "epoch": 0.27030926843345676, "grad_norm": 1.5941812945636864, "learning_rate": 1.712046952636398e-05, "loss": 1.0129, "step": 1405 }, { "epoch": 0.27050165937184356, "grad_norm": 1.4237080802928124, "learning_rate": 1.7116092400454655e-05, "loss": 1.0586, "step": 1406 }, { "epoch": 0.27069405031023036, "grad_norm": 0.7638757614448921, "learning_rate": 1.7111712510737035e-05, "loss": 0.8352, "step": 1407 }, { "epoch": 0.2708864412486172, "grad_norm": 1.3488282969645427, "learning_rate": 1.7107329858912226e-05, "loss": 1.0034, "step": 1408 }, { "epoch": 0.271078832187004, "grad_norm": 1.4158659247181955, "learning_rate": 1.7102944446682393e-05, "loss": 1.0732, "step": 1409 }, { "epoch": 0.2712712231253908, "grad_norm": 1.3995914983018163, "learning_rate": 1.709855627575079e-05, "loss": 1.0082, "step": 1410 }, { "epoch": 0.2714636140637776, "grad_norm": 1.3352108809230918, "learning_rate": 1.7094165347821724e-05, "loss": 1.0428, "step": 1411 }, { "epoch": 0.2716560050021644, "grad_norm": 1.5583755893261577, "learning_rate": 1.7089771664600584e-05, "loss": 0.9974, "step": 1412 }, { "epoch": 0.2718483959405512, "grad_norm": 1.3883366407369337, "learning_rate": 1.708537522779382e-05, "loss": 1.045, "step": 1413 }, { "epoch": 0.272040786878938, "grad_norm": 1.3251358030554878, "learning_rate": 1.7080976039108964e-05, "loss": 1.0136, "step": 1414 }, { "epoch": 0.2722331778173248, "grad_norm": 1.439108937620389, "learning_rate": 1.7076574100254614e-05, "loss": 0.9624, "step": 1415 }, { "epoch": 0.2724255687557116, "grad_norm": 1.6052676514061424, "learning_rate": 1.707216941294042e-05, "loss": 1.0198, "step": 1416 }, { "epoch": 0.2726179596940984, "grad_norm": 1.4692913116384196, "learning_rate": 1.706776197887712e-05, "loss": 0.9781, "step": 1417 }, { "epoch": 0.2728103506324852, "grad_norm": 1.547225801343285, "learning_rate": 1.7063351799776514e-05, "loss": 1.0655, "step": 1418 }, { "epoch": 0.273002741570872, "grad_norm": 1.397832883125895, "learning_rate": 1.7058938877351456e-05, "loss": 1.0216, "step": 1419 }, { "epoch": 0.2731951325092588, "grad_norm": 1.441218612510157, "learning_rate": 1.705452321331588e-05, "loss": 1.0558, "step": 1420 }, { "epoch": 0.2733875234476456, "grad_norm": 1.5730467996440347, "learning_rate": 1.7050104809384774e-05, "loss": 1.0354, "step": 1421 }, { "epoch": 0.2735799143860324, "grad_norm": 1.3634646370822558, "learning_rate": 1.70456836672742e-05, "loss": 0.976, "step": 1422 }, { "epoch": 0.2737723053244192, "grad_norm": 1.3068860882575508, "learning_rate": 1.704125978870128e-05, "loss": 1.0529, "step": 1423 }, { "epoch": 0.273964696262806, "grad_norm": 1.3655193115449256, "learning_rate": 1.7036833175384192e-05, "loss": 1.0867, "step": 1424 }, { "epoch": 0.2741570872011928, "grad_norm": 1.5304338424850477, "learning_rate": 1.7032403829042182e-05, "loss": 1.0132, "step": 1425 }, { "epoch": 0.2743494781395796, "grad_norm": 1.3744981520897648, "learning_rate": 1.7027971751395563e-05, "loss": 1.0289, "step": 1426 }, { "epoch": 0.2745418690779664, "grad_norm": 1.4534306729350788, "learning_rate": 1.7023536944165697e-05, "loss": 1.0423, "step": 1427 }, { "epoch": 0.2747342600163532, "grad_norm": 1.4430673132109346, "learning_rate": 1.7019099409075014e-05, "loss": 1.0062, "step": 1428 }, { "epoch": 0.27492665095474006, "grad_norm": 0.7209587567825678, "learning_rate": 1.7014659147847005e-05, "loss": 0.8699, "step": 1429 }, { "epoch": 0.27511904189312686, "grad_norm": 1.5814260930543185, "learning_rate": 1.701021616220621e-05, "loss": 0.9954, "step": 1430 }, { "epoch": 0.27531143283151366, "grad_norm": 1.4608348643598408, "learning_rate": 1.7005770453878234e-05, "loss": 0.9961, "step": 1431 }, { "epoch": 0.27550382376990046, "grad_norm": 1.5127683647220538, "learning_rate": 1.7001322024589742e-05, "loss": 1.0776, "step": 1432 }, { "epoch": 0.27569621470828726, "grad_norm": 1.4579722837255735, "learning_rate": 1.6996870876068455e-05, "loss": 0.9947, "step": 1433 }, { "epoch": 0.27588860564667406, "grad_norm": 1.4162169072934752, "learning_rate": 1.6992417010043144e-05, "loss": 1.055, "step": 1434 }, { "epoch": 0.27608099658506086, "grad_norm": 1.6176770198569514, "learning_rate": 1.6987960428243637e-05, "loss": 1.0131, "step": 1435 }, { "epoch": 0.27627338752344766, "grad_norm": 1.3895776932509725, "learning_rate": 1.6983501132400825e-05, "loss": 1.083, "step": 1436 }, { "epoch": 0.27646577846183445, "grad_norm": 1.3491888373263254, "learning_rate": 1.6979039124246643e-05, "loss": 0.934, "step": 1437 }, { "epoch": 0.27665816940022125, "grad_norm": 1.5304601416686408, "learning_rate": 1.6974574405514083e-05, "loss": 1.0744, "step": 1438 }, { "epoch": 0.27685056033860805, "grad_norm": 1.6221720047182466, "learning_rate": 1.6970106977937192e-05, "loss": 1.0997, "step": 1439 }, { "epoch": 0.27704295127699485, "grad_norm": 1.5691991378722592, "learning_rate": 1.696563684325107e-05, "loss": 0.9666, "step": 1440 }, { "epoch": 0.27723534221538165, "grad_norm": 1.5709252913579215, "learning_rate": 1.6961164003191862e-05, "loss": 0.9895, "step": 1441 }, { "epoch": 0.27742773315376845, "grad_norm": 1.6913966377944438, "learning_rate": 1.6956688459496767e-05, "loss": 0.9806, "step": 1442 }, { "epoch": 0.27762012409215525, "grad_norm": 1.326203219920161, "learning_rate": 1.695221021390404e-05, "loss": 1.0394, "step": 1443 }, { "epoch": 0.27781251503054205, "grad_norm": 1.4365893565978887, "learning_rate": 1.6947729268152972e-05, "loss": 0.9788, "step": 1444 }, { "epoch": 0.27800490596892885, "grad_norm": 1.4413406686885732, "learning_rate": 1.6943245623983918e-05, "loss": 1.0505, "step": 1445 }, { "epoch": 0.27819729690731565, "grad_norm": 1.490939026930701, "learning_rate": 1.6938759283138268e-05, "loss": 0.9799, "step": 1446 }, { "epoch": 0.27838968784570245, "grad_norm": 1.3788884406813449, "learning_rate": 1.693427024735847e-05, "loss": 0.9776, "step": 1447 }, { "epoch": 0.27858207878408925, "grad_norm": 1.5194253243560132, "learning_rate": 1.692977851838801e-05, "loss": 1.0204, "step": 1448 }, { "epoch": 0.27877446972247605, "grad_norm": 1.459304572910906, "learning_rate": 1.6925284097971427e-05, "loss": 1.0461, "step": 1449 }, { "epoch": 0.27896686066086285, "grad_norm": 1.5094646256212183, "learning_rate": 1.6920786987854296e-05, "loss": 1.0321, "step": 1450 }, { "epoch": 0.2791592515992497, "grad_norm": 1.3807352543455822, "learning_rate": 1.691628718978325e-05, "loss": 1.0409, "step": 1451 }, { "epoch": 0.2793516425376365, "grad_norm": 0.8061223337447446, "learning_rate": 1.691178470550596e-05, "loss": 0.882, "step": 1452 }, { "epoch": 0.2795440334760233, "grad_norm": 1.3840168485370732, "learning_rate": 1.6907279536771127e-05, "loss": 1.006, "step": 1453 }, { "epoch": 0.2797364244144101, "grad_norm": 1.4365692979011615, "learning_rate": 1.6902771685328524e-05, "loss": 0.9955, "step": 1454 }, { "epoch": 0.2799288153527969, "grad_norm": 1.3990283449593965, "learning_rate": 1.6898261152928933e-05, "loss": 0.9332, "step": 1455 }, { "epoch": 0.2801212062911837, "grad_norm": 1.52961808307168, "learning_rate": 1.6893747941324197e-05, "loss": 0.9995, "step": 1456 }, { "epoch": 0.2803135972295705, "grad_norm": 1.4434819138868151, "learning_rate": 1.6889232052267203e-05, "loss": 1.0842, "step": 1457 }, { "epoch": 0.2805059881679573, "grad_norm": 1.4611471852260838, "learning_rate": 1.688471348751186e-05, "loss": 0.9606, "step": 1458 }, { "epoch": 0.2806983791063441, "grad_norm": 1.47189345861329, "learning_rate": 1.688019224881313e-05, "loss": 1.0705, "step": 1459 }, { "epoch": 0.2808907700447309, "grad_norm": 1.4033452887720401, "learning_rate": 1.6875668337927014e-05, "loss": 1.056, "step": 1460 }, { "epoch": 0.2810831609831177, "grad_norm": 1.3954192885763694, "learning_rate": 1.6871141756610544e-05, "loss": 1.0351, "step": 1461 }, { "epoch": 0.2812755519215045, "grad_norm": 1.4407208498100283, "learning_rate": 1.6866612506621788e-05, "loss": 1.0396, "step": 1462 }, { "epoch": 0.2814679428598913, "grad_norm": 1.542225210979105, "learning_rate": 1.6862080589719863e-05, "loss": 1.0456, "step": 1463 }, { "epoch": 0.2816603337982781, "grad_norm": 1.2579984834114455, "learning_rate": 1.6857546007664908e-05, "loss": 0.946, "step": 1464 }, { "epoch": 0.2818527247366649, "grad_norm": 1.3779486409985542, "learning_rate": 1.6853008762218103e-05, "loss": 1.0644, "step": 1465 }, { "epoch": 0.2820451156750517, "grad_norm": 1.2912569170760764, "learning_rate": 1.684846885514166e-05, "loss": 1.0582, "step": 1466 }, { "epoch": 0.2822375066134385, "grad_norm": 1.4011561718966121, "learning_rate": 1.6843926288198828e-05, "loss": 1.0399, "step": 1467 }, { "epoch": 0.2824298975518253, "grad_norm": 1.3983413749570257, "learning_rate": 1.683938106315389e-05, "loss": 0.9976, "step": 1468 }, { "epoch": 0.2826222884902121, "grad_norm": 1.4589952789289686, "learning_rate": 1.683483318177216e-05, "loss": 1.0286, "step": 1469 }, { "epoch": 0.2828146794285989, "grad_norm": 1.4371672998100542, "learning_rate": 1.6830282645819974e-05, "loss": 1.0267, "step": 1470 }, { "epoch": 0.2830070703669857, "grad_norm": 1.368683848436029, "learning_rate": 1.6825729457064718e-05, "loss": 1.0983, "step": 1471 }, { "epoch": 0.2831994613053725, "grad_norm": 1.4254737268933872, "learning_rate": 1.6821173617274793e-05, "loss": 0.9904, "step": 1472 }, { "epoch": 0.28339185224375935, "grad_norm": 1.4055569393598546, "learning_rate": 1.6816615128219635e-05, "loss": 1.0073, "step": 1473 }, { "epoch": 0.28358424318214615, "grad_norm": 1.4365342800648186, "learning_rate": 1.681205399166971e-05, "loss": 0.9547, "step": 1474 }, { "epoch": 0.28377663412053294, "grad_norm": 1.3798036043555824, "learning_rate": 1.6807490209396506e-05, "loss": 0.9716, "step": 1475 }, { "epoch": 0.28396902505891974, "grad_norm": 1.5084469602227735, "learning_rate": 1.6802923783172553e-05, "loss": 0.995, "step": 1476 }, { "epoch": 0.28416141599730654, "grad_norm": 1.4685784729555538, "learning_rate": 1.679835471477139e-05, "loss": 1.0081, "step": 1477 }, { "epoch": 0.28435380693569334, "grad_norm": 1.5531396246346691, "learning_rate": 1.6793783005967593e-05, "loss": 0.9624, "step": 1478 }, { "epoch": 0.28454619787408014, "grad_norm": 0.7965007615936762, "learning_rate": 1.678920865853676e-05, "loss": 0.8531, "step": 1479 }, { "epoch": 0.28473858881246694, "grad_norm": 1.2561142070068514, "learning_rate": 1.678463167425552e-05, "loss": 1.0049, "step": 1480 }, { "epoch": 0.28493097975085374, "grad_norm": 1.4741180544264805, "learning_rate": 1.6780052054901512e-05, "loss": 1.0318, "step": 1481 }, { "epoch": 0.28512337068924054, "grad_norm": 1.5159454503443452, "learning_rate": 1.6775469802253416e-05, "loss": 1.0538, "step": 1482 }, { "epoch": 0.28531576162762734, "grad_norm": 1.4699318267915875, "learning_rate": 1.6770884918090923e-05, "loss": 0.9703, "step": 1483 }, { "epoch": 0.28550815256601414, "grad_norm": 1.4620492809496288, "learning_rate": 1.6766297404194745e-05, "loss": 1.0356, "step": 1484 }, { "epoch": 0.28570054350440094, "grad_norm": 1.4709366745034502, "learning_rate": 1.6761707262346624e-05, "loss": 0.9779, "step": 1485 }, { "epoch": 0.28589293444278774, "grad_norm": 1.5381245548141633, "learning_rate": 1.675711449432932e-05, "loss": 1.0391, "step": 1486 }, { "epoch": 0.28608532538117454, "grad_norm": 1.4838834881644238, "learning_rate": 1.6752519101926606e-05, "loss": 0.9014, "step": 1487 }, { "epoch": 0.28627771631956134, "grad_norm": 1.4828990197072076, "learning_rate": 1.6747921086923284e-05, "loss": 1.0178, "step": 1488 }, { "epoch": 0.28647010725794814, "grad_norm": 1.4890957486216145, "learning_rate": 1.674332045110517e-05, "loss": 1.0537, "step": 1489 }, { "epoch": 0.28666249819633494, "grad_norm": 1.499331378175137, "learning_rate": 1.6738717196259092e-05, "loss": 1.0327, "step": 1490 }, { "epoch": 0.28685488913472174, "grad_norm": 1.344927195971191, "learning_rate": 1.673411132417291e-05, "loss": 1.0048, "step": 1491 }, { "epoch": 0.28704728007310853, "grad_norm": 1.564968564910984, "learning_rate": 1.672950283663548e-05, "loss": 1.0033, "step": 1492 }, { "epoch": 0.28723967101149533, "grad_norm": 1.5206943444840895, "learning_rate": 1.6724891735436697e-05, "loss": 1.0009, "step": 1493 }, { "epoch": 0.28743206194988213, "grad_norm": 1.5374211355727014, "learning_rate": 1.6720278022367453e-05, "loss": 1.038, "step": 1494 }, { "epoch": 0.287624452888269, "grad_norm": 1.446644465313128, "learning_rate": 1.6715661699219664e-05, "loss": 1.0553, "step": 1495 }, { "epoch": 0.2878168438266558, "grad_norm": 1.4898206193646917, "learning_rate": 1.6711042767786257e-05, "loss": 1.0474, "step": 1496 }, { "epoch": 0.2880092347650426, "grad_norm": 1.4818855937889186, "learning_rate": 1.6706421229861168e-05, "loss": 1.0628, "step": 1497 }, { "epoch": 0.2882016257034294, "grad_norm": 1.4617702597627467, "learning_rate": 1.6701797087239354e-05, "loss": 0.9827, "step": 1498 }, { "epoch": 0.2883940166418162, "grad_norm": 1.405349892207289, "learning_rate": 1.6697170341716772e-05, "loss": 1.0227, "step": 1499 }, { "epoch": 0.288586407580203, "grad_norm": 1.462210224330947, "learning_rate": 1.6692540995090403e-05, "loss": 1.0645, "step": 1500 }, { "epoch": 0.2887787985185898, "grad_norm": 1.319613154914654, "learning_rate": 1.668790904915823e-05, "loss": 1.03, "step": 1501 }, { "epoch": 0.2889711894569766, "grad_norm": 1.3977553663374307, "learning_rate": 1.6683274505719248e-05, "loss": 0.9602, "step": 1502 }, { "epoch": 0.2891635803953634, "grad_norm": 1.4504910499893322, "learning_rate": 1.6678637366573455e-05, "loss": 1.0172, "step": 1503 }, { "epoch": 0.2893559713337502, "grad_norm": 0.7560098050925573, "learning_rate": 1.667399763352187e-05, "loss": 0.8092, "step": 1504 }, { "epoch": 0.289548362272137, "grad_norm": 1.4693427966405477, "learning_rate": 1.666935530836651e-05, "loss": 1.016, "step": 1505 }, { "epoch": 0.2897407532105238, "grad_norm": 1.4542779970932598, "learning_rate": 1.6664710392910396e-05, "loss": 1.0633, "step": 1506 }, { "epoch": 0.2899331441489106, "grad_norm": 1.3058564545642644, "learning_rate": 1.6660062888957564e-05, "loss": 0.9874, "step": 1507 }, { "epoch": 0.2901255350872974, "grad_norm": 1.3295567552552077, "learning_rate": 1.665541279831305e-05, "loss": 0.9988, "step": 1508 }, { "epoch": 0.2903179260256842, "grad_norm": 1.3993838470133302, "learning_rate": 1.6650760122782898e-05, "loss": 1.0605, "step": 1509 }, { "epoch": 0.290510316964071, "grad_norm": 1.3867559988985299, "learning_rate": 1.6646104864174147e-05, "loss": 0.9608, "step": 1510 }, { "epoch": 0.2907027079024578, "grad_norm": 1.4449616360336721, "learning_rate": 1.664144702429485e-05, "loss": 1.0285, "step": 1511 }, { "epoch": 0.2908950988408446, "grad_norm": 1.538471261146775, "learning_rate": 1.663678660495406e-05, "loss": 1.1294, "step": 1512 }, { "epoch": 0.2910874897792314, "grad_norm": 1.4413518490893122, "learning_rate": 1.663212360796183e-05, "loss": 1.0759, "step": 1513 }, { "epoch": 0.2912798807176182, "grad_norm": 1.3387348190670052, "learning_rate": 1.662745803512921e-05, "loss": 1.0735, "step": 1514 }, { "epoch": 0.291472271656005, "grad_norm": 1.4722026833484718, "learning_rate": 1.662278988826826e-05, "loss": 1.0546, "step": 1515 }, { "epoch": 0.2916646625943918, "grad_norm": 1.4761840635574246, "learning_rate": 1.6618119169192027e-05, "loss": 1.0751, "step": 1516 }, { "epoch": 0.29185705353277863, "grad_norm": 1.3813538676028874, "learning_rate": 1.661344587971457e-05, "loss": 1.0132, "step": 1517 }, { "epoch": 0.29204944447116543, "grad_norm": 1.564327293895, "learning_rate": 1.6608770021650945e-05, "loss": 1.0874, "step": 1518 }, { "epoch": 0.29224183540955223, "grad_norm": 1.3982307995882168, "learning_rate": 1.6604091596817193e-05, "loss": 1.0636, "step": 1519 }, { "epoch": 0.29243422634793903, "grad_norm": 1.4030691698698992, "learning_rate": 1.6599410607030363e-05, "loss": 1.0165, "step": 1520 }, { "epoch": 0.29262661728632583, "grad_norm": 1.5164456341109849, "learning_rate": 1.6594727054108498e-05, "loss": 1.1281, "step": 1521 }, { "epoch": 0.29281900822471263, "grad_norm": 1.4407218267048385, "learning_rate": 1.659004093987064e-05, "loss": 1.031, "step": 1522 }, { "epoch": 0.29301139916309943, "grad_norm": 1.4246568123411545, "learning_rate": 1.6585352266136814e-05, "loss": 0.9893, "step": 1523 }, { "epoch": 0.29320379010148623, "grad_norm": 1.4210666360008992, "learning_rate": 1.6580661034728055e-05, "loss": 1.0526, "step": 1524 }, { "epoch": 0.293396181039873, "grad_norm": 1.4936003969830856, "learning_rate": 1.6575967247466376e-05, "loss": 1.0355, "step": 1525 }, { "epoch": 0.2935885719782598, "grad_norm": 1.3863442860002357, "learning_rate": 1.657127090617479e-05, "loss": 0.981, "step": 1526 }, { "epoch": 0.2937809629166466, "grad_norm": 1.3706403286658044, "learning_rate": 1.656657201267731e-05, "loss": 1.001, "step": 1527 }, { "epoch": 0.2939733538550334, "grad_norm": 1.4282887352945268, "learning_rate": 1.6561870568798927e-05, "loss": 1.0163, "step": 1528 }, { "epoch": 0.2941657447934202, "grad_norm": 0.7848103504125624, "learning_rate": 1.655716657636562e-05, "loss": 0.8267, "step": 1529 }, { "epoch": 0.294358135731807, "grad_norm": 1.554047510259495, "learning_rate": 1.6552460037204382e-05, "loss": 1.0922, "step": 1530 }, { "epoch": 0.2945505266701938, "grad_norm": 1.4928854117139758, "learning_rate": 1.6547750953143168e-05, "loss": 1.0716, "step": 1531 }, { "epoch": 0.2947429176085806, "grad_norm": 1.388785549563299, "learning_rate": 1.654303932601093e-05, "loss": 1.0005, "step": 1532 }, { "epoch": 0.2949353085469674, "grad_norm": 1.3537312404037007, "learning_rate": 1.6538325157637614e-05, "loss": 1.0159, "step": 1533 }, { "epoch": 0.2951276994853542, "grad_norm": 1.3430247418297352, "learning_rate": 1.653360844985415e-05, "loss": 0.9863, "step": 1534 }, { "epoch": 0.295320090423741, "grad_norm": 1.391353341313584, "learning_rate": 1.652888920449245e-05, "loss": 0.9394, "step": 1535 }, { "epoch": 0.2955124813621278, "grad_norm": 1.2907120616643173, "learning_rate": 1.6524167423385414e-05, "loss": 0.9876, "step": 1536 }, { "epoch": 0.2957048723005146, "grad_norm": 1.388239724450648, "learning_rate": 1.651944310836693e-05, "loss": 0.9916, "step": 1537 }, { "epoch": 0.2958972632389014, "grad_norm": 1.3109623229848333, "learning_rate": 1.6514716261271866e-05, "loss": 0.9838, "step": 1538 }, { "epoch": 0.2960896541772883, "grad_norm": 1.4252796897419002, "learning_rate": 1.6509986883936073e-05, "loss": 0.9692, "step": 1539 }, { "epoch": 0.2962820451156751, "grad_norm": 1.5231154676076881, "learning_rate": 1.650525497819639e-05, "loss": 1.084, "step": 1540 }, { "epoch": 0.2964744360540619, "grad_norm": 1.6274211008574133, "learning_rate": 1.6500520545890634e-05, "loss": 1.0421, "step": 1541 }, { "epoch": 0.2966668269924487, "grad_norm": 1.4584810284466527, "learning_rate": 1.6495783588857605e-05, "loss": 1.0187, "step": 1542 }, { "epoch": 0.2968592179308355, "grad_norm": 1.498269848288568, "learning_rate": 1.649104410893708e-05, "loss": 1.0238, "step": 1543 }, { "epoch": 0.2970516088692223, "grad_norm": 1.4880611020142316, "learning_rate": 1.648630210796982e-05, "loss": 1.0058, "step": 1544 }, { "epoch": 0.29724399980760907, "grad_norm": 1.4036755861337764, "learning_rate": 1.6481557587797562e-05, "loss": 0.9494, "step": 1545 }, { "epoch": 0.29743639074599587, "grad_norm": 1.3161592171944274, "learning_rate": 1.6476810550263023e-05, "loss": 1.0095, "step": 1546 }, { "epoch": 0.29762878168438267, "grad_norm": 1.4410426248854211, "learning_rate": 1.6472060997209898e-05, "loss": 1.0048, "step": 1547 }, { "epoch": 0.29782117262276947, "grad_norm": 0.7488633306560424, "learning_rate": 1.6467308930482863e-05, "loss": 0.8443, "step": 1548 }, { "epoch": 0.29801356356115627, "grad_norm": 1.3756679464061332, "learning_rate": 1.6462554351927558e-05, "loss": 0.9812, "step": 1549 }, { "epoch": 0.29820595449954307, "grad_norm": 1.428423870643744, "learning_rate": 1.6457797263390613e-05, "loss": 1.0238, "step": 1550 }, { "epoch": 0.29839834543792987, "grad_norm": 1.546188124158174, "learning_rate": 1.6453037666719624e-05, "loss": 0.9468, "step": 1551 }, { "epoch": 0.29859073637631667, "grad_norm": 1.5761407813331947, "learning_rate": 1.6448275563763162e-05, "loss": 1.08, "step": 1552 }, { "epoch": 0.29878312731470347, "grad_norm": 1.3395943028680113, "learning_rate": 1.644351095637078e-05, "loss": 0.9991, "step": 1553 }, { "epoch": 0.29897551825309027, "grad_norm": 1.4071833537409364, "learning_rate": 1.6438743846392987e-05, "loss": 0.9899, "step": 1554 }, { "epoch": 0.29916790919147707, "grad_norm": 1.3590975843973305, "learning_rate": 1.6433974235681274e-05, "loss": 1.0799, "step": 1555 }, { "epoch": 0.29936030012986387, "grad_norm": 1.406805375135603, "learning_rate": 1.6429202126088112e-05, "loss": 0.9235, "step": 1556 }, { "epoch": 0.29955269106825066, "grad_norm": 1.5015449882379555, "learning_rate": 1.6424427519466925e-05, "loss": 1.0332, "step": 1557 }, { "epoch": 0.29974508200663746, "grad_norm": 0.7737334275720307, "learning_rate": 1.641965041767212e-05, "loss": 0.8751, "step": 1558 }, { "epoch": 0.29993747294502426, "grad_norm": 1.5616551827193164, "learning_rate": 1.6414870822559064e-05, "loss": 0.9881, "step": 1559 }, { "epoch": 0.3001298638834111, "grad_norm": 1.4593694911315889, "learning_rate": 1.6410088735984103e-05, "loss": 0.9251, "step": 1560 }, { "epoch": 0.3003222548217979, "grad_norm": 1.3617678975645495, "learning_rate": 1.6405304159804534e-05, "loss": 1.1142, "step": 1561 }, { "epoch": 0.3005146457601847, "grad_norm": 1.6068789369065641, "learning_rate": 1.6400517095878644e-05, "loss": 1.0257, "step": 1562 }, { "epoch": 0.3007070366985715, "grad_norm": 1.406977434737403, "learning_rate": 1.6395727546065665e-05, "loss": 0.9996, "step": 1563 }, { "epoch": 0.3008994276369583, "grad_norm": 0.7804846768899183, "learning_rate": 1.6390935512225806e-05, "loss": 0.8786, "step": 1564 }, { "epoch": 0.3010918185753451, "grad_norm": 1.6949620910192766, "learning_rate": 1.6386140996220232e-05, "loss": 1.0724, "step": 1565 }, { "epoch": 0.3012842095137319, "grad_norm": 1.5006865507475888, "learning_rate": 1.6381343999911088e-05, "loss": 1.0408, "step": 1566 }, { "epoch": 0.3014766004521187, "grad_norm": 1.5136741435457577, "learning_rate": 1.6376544525161463e-05, "loss": 1.0286, "step": 1567 }, { "epoch": 0.3016689913905055, "grad_norm": 1.3517652039411914, "learning_rate": 1.6371742573835426e-05, "loss": 0.9801, "step": 1568 }, { "epoch": 0.3018613823288923, "grad_norm": 1.4626513715401752, "learning_rate": 1.636693814779799e-05, "loss": 1.0683, "step": 1569 }, { "epoch": 0.3020537732672791, "grad_norm": 1.4205389205553771, "learning_rate": 1.6362131248915145e-05, "loss": 1.0503, "step": 1570 }, { "epoch": 0.3022461642056659, "grad_norm": 1.5321469890748738, "learning_rate": 1.6357321879053833e-05, "loss": 1.0125, "step": 1571 }, { "epoch": 0.3024385551440527, "grad_norm": 1.3217638042427384, "learning_rate": 1.6352510040081962e-05, "loss": 0.9793, "step": 1572 }, { "epoch": 0.3026309460824395, "grad_norm": 1.436952076013133, "learning_rate": 1.634769573386839e-05, "loss": 1.083, "step": 1573 }, { "epoch": 0.3028233370208263, "grad_norm": 1.433205449757716, "learning_rate": 1.634287896228294e-05, "loss": 1.0254, "step": 1574 }, { "epoch": 0.3030157279592131, "grad_norm": 1.321431622791033, "learning_rate": 1.6338059727196386e-05, "loss": 1.029, "step": 1575 }, { "epoch": 0.3032081188975999, "grad_norm": 1.432442041930054, "learning_rate": 1.6333238030480473e-05, "loss": 1.0879, "step": 1576 }, { "epoch": 0.3034005098359867, "grad_norm": 1.5094086550077543, "learning_rate": 1.6328413874007884e-05, "loss": 1.0414, "step": 1577 }, { "epoch": 0.3035929007743735, "grad_norm": 1.3520089549475427, "learning_rate": 1.6323587259652267e-05, "loss": 1.0085, "step": 1578 }, { "epoch": 0.3037852917127603, "grad_norm": 1.4142789517686236, "learning_rate": 1.6318758189288227e-05, "loss": 1.0162, "step": 1579 }, { "epoch": 0.3039776826511471, "grad_norm": 1.4566440821002653, "learning_rate": 1.6313926664791316e-05, "loss": 0.9781, "step": 1580 }, { "epoch": 0.3041700735895339, "grad_norm": 0.719462554093406, "learning_rate": 1.6309092688038047e-05, "loss": 0.8547, "step": 1581 }, { "epoch": 0.30436246452792076, "grad_norm": 0.6649996749710015, "learning_rate": 1.6304256260905872e-05, "loss": 0.8128, "step": 1582 }, { "epoch": 0.30455485546630756, "grad_norm": 1.370293641998078, "learning_rate": 1.6299417385273216e-05, "loss": 1.0285, "step": 1583 }, { "epoch": 0.30474724640469436, "grad_norm": 1.3925876517613116, "learning_rate": 1.629457606301943e-05, "loss": 1.0254, "step": 1584 }, { "epoch": 0.30493963734308116, "grad_norm": 1.5252024465731668, "learning_rate": 1.6289732296024837e-05, "loss": 1.0417, "step": 1585 }, { "epoch": 0.30513202828146796, "grad_norm": 1.5169950240781416, "learning_rate": 1.6284886086170697e-05, "loss": 1.0522, "step": 1586 }, { "epoch": 0.30532441921985476, "grad_norm": 1.4336062440040038, "learning_rate": 1.628003743533922e-05, "loss": 0.9841, "step": 1587 }, { "epoch": 0.30551681015824156, "grad_norm": 1.6796820653557303, "learning_rate": 1.6275186345413566e-05, "loss": 1.0848, "step": 1588 }, { "epoch": 0.30570920109662836, "grad_norm": 1.305505362383458, "learning_rate": 1.627033281827785e-05, "loss": 0.9756, "step": 1589 }, { "epoch": 0.30590159203501516, "grad_norm": 1.4309760051630545, "learning_rate": 1.6265476855817116e-05, "loss": 1.0646, "step": 1590 }, { "epoch": 0.30609398297340196, "grad_norm": 0.7738351121220538, "learning_rate": 1.6260618459917366e-05, "loss": 0.8237, "step": 1591 }, { "epoch": 0.30628637391178876, "grad_norm": 1.3958036499256439, "learning_rate": 1.6255757632465553e-05, "loss": 1.0505, "step": 1592 }, { "epoch": 0.30647876485017556, "grad_norm": 1.6161448029083076, "learning_rate": 1.625089437534956e-05, "loss": 1.0724, "step": 1593 }, { "epoch": 0.30667115578856236, "grad_norm": 1.3711015501286334, "learning_rate": 1.624602869045822e-05, "loss": 1.014, "step": 1594 }, { "epoch": 0.30686354672694915, "grad_norm": 1.3056600423021636, "learning_rate": 1.624116057968131e-05, "loss": 0.9446, "step": 1595 }, { "epoch": 0.30705593766533595, "grad_norm": 1.3430538715333231, "learning_rate": 1.6236290044909543e-05, "loss": 1.078, "step": 1596 }, { "epoch": 0.30724832860372275, "grad_norm": 1.462063209620216, "learning_rate": 1.6231417088034585e-05, "loss": 1.0689, "step": 1597 }, { "epoch": 0.30744071954210955, "grad_norm": 1.2739387016855872, "learning_rate": 1.622654171094904e-05, "loss": 0.9477, "step": 1598 }, { "epoch": 0.30763311048049635, "grad_norm": 1.3280060725230802, "learning_rate": 1.6221663915546437e-05, "loss": 0.9707, "step": 1599 }, { "epoch": 0.30782550141888315, "grad_norm": 1.5363483949549532, "learning_rate": 1.6216783703721265e-05, "loss": 1.0539, "step": 1600 }, { "epoch": 0.30801789235726995, "grad_norm": 1.514838738533436, "learning_rate": 1.6211901077368937e-05, "loss": 1.0947, "step": 1601 }, { "epoch": 0.30821028329565675, "grad_norm": 1.3753710084276756, "learning_rate": 1.620701603838581e-05, "loss": 0.9503, "step": 1602 }, { "epoch": 0.30840267423404355, "grad_norm": 1.4319681837890432, "learning_rate": 1.6202128588669177e-05, "loss": 0.9757, "step": 1603 }, { "epoch": 0.3085950651724304, "grad_norm": 1.5367387177868608, "learning_rate": 1.619723873011727e-05, "loss": 1.0256, "step": 1604 }, { "epoch": 0.3087874561108172, "grad_norm": 1.340701566095807, "learning_rate": 1.6192346464629247e-05, "loss": 1.0188, "step": 1605 }, { "epoch": 0.308979847049204, "grad_norm": 1.4368167847941749, "learning_rate": 1.6187451794105212e-05, "loss": 0.9671, "step": 1606 }, { "epoch": 0.3091722379875908, "grad_norm": 1.322960181369264, "learning_rate": 1.61825547204462e-05, "loss": 1.0149, "step": 1607 }, { "epoch": 0.3093646289259776, "grad_norm": 1.484180146677743, "learning_rate": 1.6177655245554177e-05, "loss": 0.9946, "step": 1608 }, { "epoch": 0.3095570198643644, "grad_norm": 1.506733805601682, "learning_rate": 1.617275337133204e-05, "loss": 1.0186, "step": 1609 }, { "epoch": 0.3097494108027512, "grad_norm": 1.4829252093747456, "learning_rate": 1.6167849099683623e-05, "loss": 1.0639, "step": 1610 }, { "epoch": 0.309941801741138, "grad_norm": 1.4781555151620014, "learning_rate": 1.6162942432513687e-05, "loss": 1.0041, "step": 1611 }, { "epoch": 0.3101341926795248, "grad_norm": 0.729462219093717, "learning_rate": 1.6158033371727924e-05, "loss": 0.8686, "step": 1612 }, { "epoch": 0.3103265836179116, "grad_norm": 0.7300361986617847, "learning_rate": 1.6153121919232962e-05, "loss": 0.8334, "step": 1613 }, { "epoch": 0.3105189745562984, "grad_norm": 1.4628795068175782, "learning_rate": 1.614820807693635e-05, "loss": 0.9862, "step": 1614 }, { "epoch": 0.3107113654946852, "grad_norm": 1.6684005740977672, "learning_rate": 1.6143291846746563e-05, "loss": 0.9448, "step": 1615 }, { "epoch": 0.310903756433072, "grad_norm": 1.555919231489896, "learning_rate": 1.613837323057301e-05, "loss": 1.0747, "step": 1616 }, { "epoch": 0.3110961473714588, "grad_norm": 1.3603835184432114, "learning_rate": 1.6133452230326035e-05, "loss": 1.0211, "step": 1617 }, { "epoch": 0.3112885383098456, "grad_norm": 1.4748521456987351, "learning_rate": 1.6128528847916883e-05, "loss": 0.9946, "step": 1618 }, { "epoch": 0.3114809292482324, "grad_norm": 1.4071979453027368, "learning_rate": 1.6123603085257746e-05, "loss": 1.0784, "step": 1619 }, { "epoch": 0.3116733201866192, "grad_norm": 1.4261672752588943, "learning_rate": 1.6118674944261732e-05, "loss": 1.1041, "step": 1620 }, { "epoch": 0.311865711125006, "grad_norm": 1.4073590388211845, "learning_rate": 1.6113744426842882e-05, "loss": 1.0317, "step": 1621 }, { "epoch": 0.3120581020633928, "grad_norm": 1.4590779981610187, "learning_rate": 1.6108811534916137e-05, "loss": 0.9589, "step": 1622 }, { "epoch": 0.3122504930017796, "grad_norm": 1.5960441976633952, "learning_rate": 1.6103876270397387e-05, "loss": 1.0409, "step": 1623 }, { "epoch": 0.3124428839401664, "grad_norm": 1.4005677138843908, "learning_rate": 1.609893863520343e-05, "loss": 1.0828, "step": 1624 }, { "epoch": 0.3126352748785532, "grad_norm": 1.2772842880173096, "learning_rate": 1.609399863125198e-05, "loss": 0.976, "step": 1625 }, { "epoch": 0.31282766581694005, "grad_norm": 0.800440409028499, "learning_rate": 1.6089056260461687e-05, "loss": 0.7822, "step": 1626 }, { "epoch": 0.31302005675532685, "grad_norm": 1.3912069773168012, "learning_rate": 1.6084111524752107e-05, "loss": 1.041, "step": 1627 }, { "epoch": 0.31321244769371365, "grad_norm": 1.3156171514995856, "learning_rate": 1.607916442604372e-05, "loss": 0.9898, "step": 1628 }, { "epoch": 0.31340483863210045, "grad_norm": 1.431760273681409, "learning_rate": 1.6074214966257914e-05, "loss": 1.0577, "step": 1629 }, { "epoch": 0.31359722957048725, "grad_norm": 1.3933144157344517, "learning_rate": 1.6069263147317015e-05, "loss": 0.9487, "step": 1630 }, { "epoch": 0.31378962050887405, "grad_norm": 1.4435412730438208, "learning_rate": 1.6064308971144236e-05, "loss": 1.1439, "step": 1631 }, { "epoch": 0.31398201144726084, "grad_norm": 0.7740173920643733, "learning_rate": 1.605935243966374e-05, "loss": 0.8868, "step": 1632 }, { "epoch": 0.31417440238564764, "grad_norm": 1.3517594836855034, "learning_rate": 1.6054393554800574e-05, "loss": 1.0476, "step": 1633 }, { "epoch": 0.31436679332403444, "grad_norm": 1.4181301753568794, "learning_rate": 1.604943231848072e-05, "loss": 1.0531, "step": 1634 }, { "epoch": 0.31455918426242124, "grad_norm": 1.2481593261408845, "learning_rate": 1.604446873263106e-05, "loss": 1.0483, "step": 1635 }, { "epoch": 0.31475157520080804, "grad_norm": 1.4818205241954392, "learning_rate": 1.6039502799179394e-05, "loss": 0.9545, "step": 1636 }, { "epoch": 0.31494396613919484, "grad_norm": 1.292045505827004, "learning_rate": 1.6034534520054435e-05, "loss": 1.0541, "step": 1637 }, { "epoch": 0.31513635707758164, "grad_norm": 1.4733808702373201, "learning_rate": 1.60295638971858e-05, "loss": 1.0272, "step": 1638 }, { "epoch": 0.31532874801596844, "grad_norm": 1.4686150674676437, "learning_rate": 1.602459093250403e-05, "loss": 0.9466, "step": 1639 }, { "epoch": 0.31552113895435524, "grad_norm": 1.3386328548442004, "learning_rate": 1.601961562794056e-05, "loss": 0.9848, "step": 1640 }, { "epoch": 0.31571352989274204, "grad_norm": 1.648510823162265, "learning_rate": 1.601463798542775e-05, "loss": 1.0282, "step": 1641 }, { "epoch": 0.31590592083112884, "grad_norm": 1.3021663302277657, "learning_rate": 1.6009658006898848e-05, "loss": 1.0691, "step": 1642 }, { "epoch": 0.31609831176951564, "grad_norm": 1.2364263979718249, "learning_rate": 1.600467569428803e-05, "loss": 1.034, "step": 1643 }, { "epoch": 0.31629070270790244, "grad_norm": 1.472547262199496, "learning_rate": 1.599969104953036e-05, "loss": 0.9816, "step": 1644 }, { "epoch": 0.31648309364628924, "grad_norm": 1.3469561094199716, "learning_rate": 1.599470407456182e-05, "loss": 1.0168, "step": 1645 }, { "epoch": 0.31667548458467604, "grad_norm": 1.4766366463883047, "learning_rate": 1.5989714771319297e-05, "loss": 0.9848, "step": 1646 }, { "epoch": 0.31686787552306284, "grad_norm": 1.3894384425629982, "learning_rate": 1.5984723141740578e-05, "loss": 1.0786, "step": 1647 }, { "epoch": 0.3170602664614497, "grad_norm": 1.3499531438489254, "learning_rate": 1.597972918776435e-05, "loss": 1.0247, "step": 1648 }, { "epoch": 0.3172526573998365, "grad_norm": 1.448161843632479, "learning_rate": 1.5974732911330208e-05, "loss": 1.051, "step": 1649 }, { "epoch": 0.3174450483382233, "grad_norm": 1.467118253843496, "learning_rate": 1.5969734314378654e-05, "loss": 1.0369, "step": 1650 }, { "epoch": 0.3176374392766101, "grad_norm": 1.394533636106525, "learning_rate": 1.5964733398851078e-05, "loss": 0.9996, "step": 1651 }, { "epoch": 0.3178298302149969, "grad_norm": 1.390265928306952, "learning_rate": 1.5959730166689783e-05, "loss": 1.0139, "step": 1652 }, { "epoch": 0.3180222211533837, "grad_norm": 1.4522701684516883, "learning_rate": 1.5954724619837966e-05, "loss": 1.0878, "step": 1653 }, { "epoch": 0.3182146120917705, "grad_norm": 1.5078137515671608, "learning_rate": 1.5949716760239722e-05, "loss": 1.0481, "step": 1654 }, { "epoch": 0.3184070030301573, "grad_norm": 1.3066122571955094, "learning_rate": 1.5944706589840046e-05, "loss": 0.9613, "step": 1655 }, { "epoch": 0.3185993939685441, "grad_norm": 1.4199795082808866, "learning_rate": 1.5939694110584833e-05, "loss": 1.0614, "step": 1656 }, { "epoch": 0.3187917849069309, "grad_norm": 1.4557375461618787, "learning_rate": 1.593467932442087e-05, "loss": 0.9949, "step": 1657 }, { "epoch": 0.3189841758453177, "grad_norm": 1.4880854833122872, "learning_rate": 1.5929662233295846e-05, "loss": 0.9779, "step": 1658 }, { "epoch": 0.3191765667837045, "grad_norm": 1.4047495457887462, "learning_rate": 1.5924642839158334e-05, "loss": 1.0664, "step": 1659 }, { "epoch": 0.3193689577220913, "grad_norm": 1.3761423002817101, "learning_rate": 1.591962114395781e-05, "loss": 1.0673, "step": 1660 }, { "epoch": 0.3195613486604781, "grad_norm": 1.43176306968897, "learning_rate": 1.5914597149644654e-05, "loss": 1.0469, "step": 1661 }, { "epoch": 0.3197537395988649, "grad_norm": 1.5022712283787598, "learning_rate": 1.5909570858170115e-05, "loss": 1.0526, "step": 1662 }, { "epoch": 0.3199461305372517, "grad_norm": 1.4071702275937534, "learning_rate": 1.5904542271486346e-05, "loss": 1.0026, "step": 1663 }, { "epoch": 0.3201385214756385, "grad_norm": 1.3963073448256407, "learning_rate": 1.5899511391546403e-05, "loss": 0.9595, "step": 1664 }, { "epoch": 0.3203309124140253, "grad_norm": 1.3339628663977072, "learning_rate": 1.5894478220304215e-05, "loss": 1.1252, "step": 1665 }, { "epoch": 0.3205233033524121, "grad_norm": 1.429435113163273, "learning_rate": 1.5889442759714603e-05, "loss": 1.0207, "step": 1666 }, { "epoch": 0.3207156942907989, "grad_norm": 1.4151532053806457, "learning_rate": 1.5884405011733294e-05, "loss": 1.0007, "step": 1667 }, { "epoch": 0.3209080852291857, "grad_norm": 1.4427140737124449, "learning_rate": 1.587936497831688e-05, "loss": 0.984, "step": 1668 }, { "epoch": 0.32110047616757253, "grad_norm": 1.2901994733002522, "learning_rate": 1.5874322661422856e-05, "loss": 0.9691, "step": 1669 }, { "epoch": 0.32129286710595933, "grad_norm": 1.372830371345152, "learning_rate": 1.5869278063009602e-05, "loss": 0.9391, "step": 1670 }, { "epoch": 0.32148525804434613, "grad_norm": 1.49999540041299, "learning_rate": 1.586423118503638e-05, "loss": 1.0288, "step": 1671 }, { "epoch": 0.32167764898273293, "grad_norm": 0.7500230291106335, "learning_rate": 1.585918202946334e-05, "loss": 0.8457, "step": 1672 }, { "epoch": 0.32187003992111973, "grad_norm": 1.4026440209928785, "learning_rate": 1.5854130598251514e-05, "loss": 1.0472, "step": 1673 }, { "epoch": 0.32206243085950653, "grad_norm": 1.4941787520473677, "learning_rate": 1.5849076893362822e-05, "loss": 1.0001, "step": 1674 }, { "epoch": 0.32225482179789333, "grad_norm": 1.5157775828664672, "learning_rate": 1.584402091676006e-05, "loss": 0.9864, "step": 1675 }, { "epoch": 0.32244721273628013, "grad_norm": 1.5921365646945007, "learning_rate": 1.5838962670406918e-05, "loss": 1.052, "step": 1676 }, { "epoch": 0.32263960367466693, "grad_norm": 1.566677072002966, "learning_rate": 1.5833902156267956e-05, "loss": 1.015, "step": 1677 }, { "epoch": 0.32283199461305373, "grad_norm": 1.3067359303825214, "learning_rate": 1.582883937630862e-05, "loss": 1.0478, "step": 1678 }, { "epoch": 0.32302438555144053, "grad_norm": 1.4662902534315192, "learning_rate": 1.5823774332495236e-05, "loss": 1.0117, "step": 1679 }, { "epoch": 0.32321677648982733, "grad_norm": 1.4028590460607604, "learning_rate": 1.581870702679501e-05, "loss": 0.9821, "step": 1680 }, { "epoch": 0.32340916742821413, "grad_norm": 1.4773183731374318, "learning_rate": 1.581363746117602e-05, "loss": 0.9063, "step": 1681 }, { "epoch": 0.3236015583666009, "grad_norm": 1.476944991706024, "learning_rate": 1.580856563760724e-05, "loss": 1.0668, "step": 1682 }, { "epoch": 0.3237939493049877, "grad_norm": 1.4918186508989708, "learning_rate": 1.5803491558058486e-05, "loss": 1.0589, "step": 1683 }, { "epoch": 0.3239863402433745, "grad_norm": 1.3718034857754733, "learning_rate": 1.579841522450049e-05, "loss": 1.0301, "step": 1684 }, { "epoch": 0.3241787311817613, "grad_norm": 1.4901051459666472, "learning_rate": 1.5793336638904838e-05, "loss": 1.0378, "step": 1685 }, { "epoch": 0.3243711221201481, "grad_norm": 1.478949767704853, "learning_rate": 1.578825580324399e-05, "loss": 0.964, "step": 1686 }, { "epoch": 0.3245635130585349, "grad_norm": 1.4467477132378679, "learning_rate": 1.5783172719491288e-05, "loss": 1.0492, "step": 1687 }, { "epoch": 0.3247559039969217, "grad_norm": 1.4149417616184246, "learning_rate": 1.577808738962094e-05, "loss": 1.0912, "step": 1688 }, { "epoch": 0.3249482949353085, "grad_norm": 1.4868467099221812, "learning_rate": 1.577299981560803e-05, "loss": 1.0408, "step": 1689 }, { "epoch": 0.3251406858736953, "grad_norm": 1.3124418179465398, "learning_rate": 1.5767909999428513e-05, "loss": 0.9865, "step": 1690 }, { "epoch": 0.3253330768120822, "grad_norm": 1.620702392529636, "learning_rate": 1.576281794305922e-05, "loss": 1.0356, "step": 1691 }, { "epoch": 0.325525467750469, "grad_norm": 1.3742168818750917, "learning_rate": 1.575772364847784e-05, "loss": 1.0485, "step": 1692 }, { "epoch": 0.3257178586888558, "grad_norm": 1.4109387500511006, "learning_rate": 1.575262711766294e-05, "loss": 1.0632, "step": 1693 }, { "epoch": 0.3259102496272426, "grad_norm": 1.4230189067125354, "learning_rate": 1.5747528352593956e-05, "loss": 0.9978, "step": 1694 }, { "epoch": 0.3261026405656294, "grad_norm": 1.3401815829589412, "learning_rate": 1.574242735525119e-05, "loss": 1.0308, "step": 1695 }, { "epoch": 0.3262950315040162, "grad_norm": 1.3485186173441202, "learning_rate": 1.5737324127615808e-05, "loss": 1.0042, "step": 1696 }, { "epoch": 0.326487422442403, "grad_norm": 1.3920699598906232, "learning_rate": 1.5732218671669847e-05, "loss": 0.9779, "step": 1697 }, { "epoch": 0.3266798133807898, "grad_norm": 1.4646059780100036, "learning_rate": 1.5727110989396205e-05, "loss": 0.9876, "step": 1698 }, { "epoch": 0.3268722043191766, "grad_norm": 1.4742656067761435, "learning_rate": 1.5722001082778645e-05, "loss": 1.04, "step": 1699 }, { "epoch": 0.3270645952575634, "grad_norm": 1.4792093975850589, "learning_rate": 1.5716888953801805e-05, "loss": 1.0484, "step": 1700 }, { "epoch": 0.3272569861959502, "grad_norm": 1.5603120176080498, "learning_rate": 1.5711774604451168e-05, "loss": 0.9663, "step": 1701 }, { "epoch": 0.32744937713433697, "grad_norm": 1.4480240119784213, "learning_rate": 1.5706658036713093e-05, "loss": 0.9872, "step": 1702 }, { "epoch": 0.32764176807272377, "grad_norm": 1.3748217587358114, "learning_rate": 1.5701539252574795e-05, "loss": 0.9367, "step": 1703 }, { "epoch": 0.32783415901111057, "grad_norm": 1.4373384573731065, "learning_rate": 1.5696418254024344e-05, "loss": 0.9785, "step": 1704 }, { "epoch": 0.32802654994949737, "grad_norm": 1.4136029818939144, "learning_rate": 1.569129504305069e-05, "loss": 0.9822, "step": 1705 }, { "epoch": 0.32821894088788417, "grad_norm": 1.4975329362860696, "learning_rate": 1.568616962164362e-05, "loss": 0.9182, "step": 1706 }, { "epoch": 0.32841133182627097, "grad_norm": 1.5514502375618964, "learning_rate": 1.5681041991793788e-05, "loss": 1.0518, "step": 1707 }, { "epoch": 0.32860372276465777, "grad_norm": 1.4743246139616928, "learning_rate": 1.567591215549271e-05, "loss": 1.0841, "step": 1708 }, { "epoch": 0.32879611370304457, "grad_norm": 1.3788091879799032, "learning_rate": 1.567078011473276e-05, "loss": 0.9941, "step": 1709 }, { "epoch": 0.32898850464143137, "grad_norm": 1.5316578111819372, "learning_rate": 1.5665645871507152e-05, "loss": 1.0413, "step": 1710 }, { "epoch": 0.32918089557981817, "grad_norm": 0.7245212429477952, "learning_rate": 1.5660509427809973e-05, "loss": 0.8455, "step": 1711 }, { "epoch": 0.32937328651820497, "grad_norm": 1.4370766196161797, "learning_rate": 1.565537078563616e-05, "loss": 0.9413, "step": 1712 }, { "epoch": 0.3295656774565918, "grad_norm": 1.4418048288224055, "learning_rate": 1.56502299469815e-05, "loss": 1.0972, "step": 1713 }, { "epoch": 0.3297580683949786, "grad_norm": 1.52286002450575, "learning_rate": 1.564508691384264e-05, "loss": 1.0396, "step": 1714 }, { "epoch": 0.3299504593333654, "grad_norm": 1.4324221478129322, "learning_rate": 1.5639941688217063e-05, "loss": 0.9676, "step": 1715 }, { "epoch": 0.3301428502717522, "grad_norm": 0.7889335654053979, "learning_rate": 1.5634794272103126e-05, "loss": 0.8771, "step": 1716 }, { "epoch": 0.330335241210139, "grad_norm": 1.4130672169620442, "learning_rate": 1.562964466750003e-05, "loss": 0.9021, "step": 1717 }, { "epoch": 0.3305276321485258, "grad_norm": 1.581733035444508, "learning_rate": 1.562449287640781e-05, "loss": 0.9581, "step": 1718 }, { "epoch": 0.3307200230869126, "grad_norm": 1.4907309908082393, "learning_rate": 1.5619338900827368e-05, "loss": 1.0712, "step": 1719 }, { "epoch": 0.3309124140252994, "grad_norm": 1.2615922564839235, "learning_rate": 1.5614182742760448e-05, "loss": 0.9734, "step": 1720 }, { "epoch": 0.3311048049636862, "grad_norm": 1.3679092807905369, "learning_rate": 1.5609024404209643e-05, "loss": 0.9635, "step": 1721 }, { "epoch": 0.331297195902073, "grad_norm": 1.4949246216819314, "learning_rate": 1.5603863887178393e-05, "loss": 1.0263, "step": 1722 }, { "epoch": 0.3314895868404598, "grad_norm": 1.3569628125549338, "learning_rate": 1.5598701193670983e-05, "loss": 1.003, "step": 1723 }, { "epoch": 0.3316819777788466, "grad_norm": 1.3515078204554531, "learning_rate": 1.559353632569254e-05, "loss": 1.109, "step": 1724 }, { "epoch": 0.3318743687172334, "grad_norm": 1.5656706229011985, "learning_rate": 1.5588369285249048e-05, "loss": 1.1135, "step": 1725 }, { "epoch": 0.3320667596556202, "grad_norm": 1.4122907733163341, "learning_rate": 1.5583200074347318e-05, "loss": 1.0106, "step": 1726 }, { "epoch": 0.332259150594007, "grad_norm": 1.363145533202102, "learning_rate": 1.557802869499501e-05, "loss": 0.9389, "step": 1727 }, { "epoch": 0.3324515415323938, "grad_norm": 0.7583280207245634, "learning_rate": 1.5572855149200637e-05, "loss": 0.8814, "step": 1728 }, { "epoch": 0.3326439324707806, "grad_norm": 1.4489298638432935, "learning_rate": 1.5567679438973543e-05, "loss": 1.0131, "step": 1729 }, { "epoch": 0.3328363234091674, "grad_norm": 1.3949926967902706, "learning_rate": 1.5562501566323906e-05, "loss": 0.9363, "step": 1730 }, { "epoch": 0.3330287143475542, "grad_norm": 1.5597441910032044, "learning_rate": 1.555732153326276e-05, "loss": 1.038, "step": 1731 }, { "epoch": 0.333221105285941, "grad_norm": 1.4540597921188787, "learning_rate": 1.5552139341801965e-05, "loss": 0.9151, "step": 1732 }, { "epoch": 0.3334134962243278, "grad_norm": 1.4629071741395876, "learning_rate": 1.554695499395423e-05, "loss": 1.0955, "step": 1733 }, { "epoch": 0.3336058871627146, "grad_norm": 1.464242371524627, "learning_rate": 1.5541768491733092e-05, "loss": 1.0125, "step": 1734 }, { "epoch": 0.33379827810110146, "grad_norm": 1.4899158649201698, "learning_rate": 1.5536579837152927e-05, "loss": 0.902, "step": 1735 }, { "epoch": 0.33399066903948826, "grad_norm": 1.3319287240535607, "learning_rate": 1.5531389032228955e-05, "loss": 0.928, "step": 1736 }, { "epoch": 0.33418305997787506, "grad_norm": 1.4356792157794829, "learning_rate": 1.552619607897722e-05, "loss": 1.0324, "step": 1737 }, { "epoch": 0.33437545091626186, "grad_norm": 1.4592370494164577, "learning_rate": 1.55210009794146e-05, "loss": 0.9446, "step": 1738 }, { "epoch": 0.33456784185464866, "grad_norm": 1.4452107074304832, "learning_rate": 1.5515803735558827e-05, "loss": 1.0443, "step": 1739 }, { "epoch": 0.33476023279303546, "grad_norm": 1.3467574714609445, "learning_rate": 1.5510604349428438e-05, "loss": 1.0772, "step": 1740 }, { "epoch": 0.33495262373142226, "grad_norm": 1.4745277803679444, "learning_rate": 1.550540282304282e-05, "loss": 1.0252, "step": 1741 }, { "epoch": 0.33514501466980906, "grad_norm": 1.412921262151277, "learning_rate": 1.550019915842218e-05, "loss": 1.0184, "step": 1742 }, { "epoch": 0.33533740560819586, "grad_norm": 1.6105631291636944, "learning_rate": 1.549499335758757e-05, "loss": 1.053, "step": 1743 }, { "epoch": 0.33552979654658266, "grad_norm": 1.3168084644730853, "learning_rate": 1.548978542256086e-05, "loss": 0.946, "step": 1744 }, { "epoch": 0.33572218748496946, "grad_norm": 1.2879803542671495, "learning_rate": 1.5484575355364744e-05, "loss": 1.0537, "step": 1745 }, { "epoch": 0.33591457842335626, "grad_norm": 1.4022201750966405, "learning_rate": 1.5479363158022763e-05, "loss": 1.0387, "step": 1746 }, { "epoch": 0.33610696936174306, "grad_norm": 1.4940708199373411, "learning_rate": 1.547414883255927e-05, "loss": 1.1016, "step": 1747 }, { "epoch": 0.33629936030012986, "grad_norm": 1.3657835431421543, "learning_rate": 1.546893238099945e-05, "loss": 0.9745, "step": 1748 }, { "epoch": 0.33649175123851666, "grad_norm": 1.4177336177866497, "learning_rate": 1.5463713805369312e-05, "loss": 1.019, "step": 1749 }, { "epoch": 0.33668414217690346, "grad_norm": 1.4272542822225736, "learning_rate": 1.5458493107695688e-05, "loss": 1.0622, "step": 1750 }, { "epoch": 0.33687653311529026, "grad_norm": 1.5826406380805496, "learning_rate": 1.5453270290006237e-05, "loss": 1.0449, "step": 1751 }, { "epoch": 0.33706892405367705, "grad_norm": 1.3947180740907188, "learning_rate": 1.544804535432945e-05, "loss": 1.0438, "step": 1752 }, { "epoch": 0.33726131499206385, "grad_norm": 1.2831097086751122, "learning_rate": 1.544281830269462e-05, "loss": 0.9388, "step": 1753 }, { "epoch": 0.33745370593045065, "grad_norm": 1.4250386535140689, "learning_rate": 1.5437589137131882e-05, "loss": 1.0046, "step": 1754 }, { "epoch": 0.33764609686883745, "grad_norm": 1.54885859556202, "learning_rate": 1.5432357859672177e-05, "loss": 0.9877, "step": 1755 }, { "epoch": 0.33783848780722425, "grad_norm": 1.4731371000268485, "learning_rate": 1.542712447234728e-05, "loss": 1.0649, "step": 1756 }, { "epoch": 0.3380308787456111, "grad_norm": 1.3586470594091495, "learning_rate": 1.542188897718977e-05, "loss": 1.0215, "step": 1757 }, { "epoch": 0.3382232696839979, "grad_norm": 1.4408003303868955, "learning_rate": 1.5416651376233062e-05, "loss": 0.9749, "step": 1758 }, { "epoch": 0.3384156606223847, "grad_norm": 1.44960607128323, "learning_rate": 1.5411411671511376e-05, "loss": 1.0525, "step": 1759 }, { "epoch": 0.3386080515607715, "grad_norm": 1.3997892924056634, "learning_rate": 1.5406169865059747e-05, "loss": 1.023, "step": 1760 }, { "epoch": 0.3388004424991583, "grad_norm": 1.4822974403412397, "learning_rate": 1.5400925958914045e-05, "loss": 1.0696, "step": 1761 }, { "epoch": 0.3389928334375451, "grad_norm": 1.5742683717104404, "learning_rate": 1.5395679955110927e-05, "loss": 1.0712, "step": 1762 }, { "epoch": 0.3391852243759319, "grad_norm": 1.482044572262077, "learning_rate": 1.53904318556879e-05, "loss": 0.9464, "step": 1763 }, { "epoch": 0.3393776153143187, "grad_norm": 1.405775752343729, "learning_rate": 1.5385181662683244e-05, "loss": 0.9303, "step": 1764 }, { "epoch": 0.3395700062527055, "grad_norm": 1.4208148028603549, "learning_rate": 1.5379929378136088e-05, "loss": 1.0378, "step": 1765 }, { "epoch": 0.3397623971910923, "grad_norm": 0.7589019082684197, "learning_rate": 1.5374675004086353e-05, "loss": 0.8401, "step": 1766 }, { "epoch": 0.3399547881294791, "grad_norm": 1.5522488363821219, "learning_rate": 1.5369418542574782e-05, "loss": 1.0692, "step": 1767 }, { "epoch": 0.3401471790678659, "grad_norm": 1.4229861876405205, "learning_rate": 1.536415999564292e-05, "loss": 1.0562, "step": 1768 }, { "epoch": 0.3403395700062527, "grad_norm": 1.4526589035790713, "learning_rate": 1.5358899365333123e-05, "loss": 0.9844, "step": 1769 }, { "epoch": 0.3405319609446395, "grad_norm": 1.6616082097922862, "learning_rate": 1.5353636653688563e-05, "loss": 0.9895, "step": 1770 }, { "epoch": 0.3407243518830263, "grad_norm": 0.6613703375279393, "learning_rate": 1.534837186275322e-05, "loss": 0.7976, "step": 1771 }, { "epoch": 0.3409167428214131, "grad_norm": 1.582024889675277, "learning_rate": 1.5343104994571877e-05, "loss": 0.9368, "step": 1772 }, { "epoch": 0.3411091337597999, "grad_norm": 1.432638666964044, "learning_rate": 1.533783605119012e-05, "loss": 0.9318, "step": 1773 }, { "epoch": 0.3413015246981867, "grad_norm": 1.349813275749781, "learning_rate": 1.5332565034654344e-05, "loss": 0.9684, "step": 1774 }, { "epoch": 0.3414939156365735, "grad_norm": 1.4529513387323907, "learning_rate": 1.5327291947011763e-05, "loss": 1.0163, "step": 1775 }, { "epoch": 0.3416863065749603, "grad_norm": 1.3951018823224395, "learning_rate": 1.5322016790310373e-05, "loss": 0.9862, "step": 1776 }, { "epoch": 0.3418786975133471, "grad_norm": 1.4097089880479272, "learning_rate": 1.5316739566598985e-05, "loss": 0.9048, "step": 1777 }, { "epoch": 0.3420710884517339, "grad_norm": 1.296209596250269, "learning_rate": 1.531146027792722e-05, "loss": 1.0297, "step": 1778 }, { "epoch": 0.34226347939012075, "grad_norm": 1.3057872333712153, "learning_rate": 1.530617892634548e-05, "loss": 0.9104, "step": 1779 }, { "epoch": 0.34245587032850755, "grad_norm": 0.7853451127586505, "learning_rate": 1.5300895513904993e-05, "loss": 0.9322, "step": 1780 }, { "epoch": 0.34264826126689435, "grad_norm": 1.3996212722161212, "learning_rate": 1.529561004265777e-05, "loss": 0.9453, "step": 1781 }, { "epoch": 0.34284065220528115, "grad_norm": 1.3966400688821974, "learning_rate": 1.5290322514656624e-05, "loss": 1.0458, "step": 1782 }, { "epoch": 0.34303304314366795, "grad_norm": 1.4361521305684353, "learning_rate": 1.5285032931955177e-05, "loss": 0.9512, "step": 1783 }, { "epoch": 0.34322543408205475, "grad_norm": 1.31319767123844, "learning_rate": 1.527974129660784e-05, "loss": 0.973, "step": 1784 }, { "epoch": 0.34341782502044155, "grad_norm": 1.3663792466479951, "learning_rate": 1.527444761066982e-05, "loss": 1.0208, "step": 1785 }, { "epoch": 0.34361021595882835, "grad_norm": 0.7072048310673656, "learning_rate": 1.5269151876197127e-05, "loss": 0.8228, "step": 1786 }, { "epoch": 0.34380260689721515, "grad_norm": 1.3530519505852872, "learning_rate": 1.5263854095246557e-05, "loss": 0.9891, "step": 1787 }, { "epoch": 0.34399499783560195, "grad_norm": 1.4227412654430291, "learning_rate": 1.5258554269875716e-05, "loss": 1.0259, "step": 1788 }, { "epoch": 0.34418738877398874, "grad_norm": 1.40352622160582, "learning_rate": 1.5253252402142989e-05, "loss": 1.0356, "step": 1789 }, { "epoch": 0.34437977971237554, "grad_norm": 1.3200038402178036, "learning_rate": 1.5247948494107566e-05, "loss": 0.9258, "step": 1790 }, { "epoch": 0.34457217065076234, "grad_norm": 1.475153345240784, "learning_rate": 1.5242642547829416e-05, "loss": 0.995, "step": 1791 }, { "epoch": 0.34476456158914914, "grad_norm": 1.3182404123121192, "learning_rate": 1.523733456536931e-05, "loss": 1.0398, "step": 1792 }, { "epoch": 0.34495695252753594, "grad_norm": 1.2991829831348825, "learning_rate": 1.5232024548788813e-05, "loss": 0.9882, "step": 1793 }, { "epoch": 0.34514934346592274, "grad_norm": 1.4538626184537253, "learning_rate": 1.5226712500150267e-05, "loss": 1.0217, "step": 1794 }, { "epoch": 0.34534173440430954, "grad_norm": 1.319414152541442, "learning_rate": 1.5221398421516816e-05, "loss": 0.9783, "step": 1795 }, { "epoch": 0.34553412534269634, "grad_norm": 1.5053172847022942, "learning_rate": 1.5216082314952383e-05, "loss": 1.0681, "step": 1796 }, { "epoch": 0.34572651628108314, "grad_norm": 1.3346177495757119, "learning_rate": 1.521076418252168e-05, "loss": 1.0505, "step": 1797 }, { "epoch": 0.34591890721946994, "grad_norm": 0.7392235048296377, "learning_rate": 1.5205444026290218e-05, "loss": 0.8853, "step": 1798 }, { "epoch": 0.34611129815785674, "grad_norm": 0.7271597917839728, "learning_rate": 1.5200121848324276e-05, "loss": 0.8506, "step": 1799 }, { "epoch": 0.3463036890962436, "grad_norm": 1.380519164445062, "learning_rate": 1.5194797650690926e-05, "loss": 1.025, "step": 1800 }, { "epoch": 0.3464960800346304, "grad_norm": 1.4853255319978744, "learning_rate": 1.5189471435458032e-05, "loss": 0.9823, "step": 1801 }, { "epoch": 0.3466884709730172, "grad_norm": 1.3640357715432347, "learning_rate": 1.5184143204694231e-05, "loss": 1.0004, "step": 1802 }, { "epoch": 0.346880861911404, "grad_norm": 0.6842685307489399, "learning_rate": 1.5178812960468945e-05, "loss": 0.82, "step": 1803 }, { "epoch": 0.3470732528497908, "grad_norm": 1.429257399841582, "learning_rate": 1.5173480704852379e-05, "loss": 0.9555, "step": 1804 }, { "epoch": 0.3472656437881776, "grad_norm": 1.3258486900077135, "learning_rate": 1.5168146439915525e-05, "loss": 0.9998, "step": 1805 }, { "epoch": 0.3474580347265644, "grad_norm": 1.3699589496129514, "learning_rate": 1.5162810167730144e-05, "loss": 0.972, "step": 1806 }, { "epoch": 0.3476504256649512, "grad_norm": 1.365955053094885, "learning_rate": 1.5157471890368785e-05, "loss": 0.9414, "step": 1807 }, { "epoch": 0.347842816603338, "grad_norm": 1.4808655433966669, "learning_rate": 1.5152131609904773e-05, "loss": 1.0413, "step": 1808 }, { "epoch": 0.3480352075417248, "grad_norm": 1.5032056444527175, "learning_rate": 1.5146789328412213e-05, "loss": 0.983, "step": 1809 }, { "epoch": 0.3482275984801116, "grad_norm": 1.3577932412237865, "learning_rate": 1.5141445047965984e-05, "loss": 1.0461, "step": 1810 }, { "epoch": 0.3484199894184984, "grad_norm": 0.6909105867527786, "learning_rate": 1.5136098770641741e-05, "loss": 0.8091, "step": 1811 }, { "epoch": 0.3486123803568852, "grad_norm": 1.4110036267782136, "learning_rate": 1.513075049851592e-05, "loss": 1.0513, "step": 1812 }, { "epoch": 0.348804771295272, "grad_norm": 1.417210294461217, "learning_rate": 1.5125400233665728e-05, "loss": 1.0885, "step": 1813 }, { "epoch": 0.3489971622336588, "grad_norm": 1.3353108952970567, "learning_rate": 1.5120047978169146e-05, "loss": 0.9839, "step": 1814 }, { "epoch": 0.3491895531720456, "grad_norm": 1.4125328347610604, "learning_rate": 1.5114693734104926e-05, "loss": 1.0098, "step": 1815 }, { "epoch": 0.3493819441104324, "grad_norm": 1.4139342111872315, "learning_rate": 1.5109337503552594e-05, "loss": 0.974, "step": 1816 }, { "epoch": 0.3495743350488192, "grad_norm": 1.3617505773119793, "learning_rate": 1.5103979288592454e-05, "loss": 0.9972, "step": 1817 }, { "epoch": 0.349766725987206, "grad_norm": 1.3282439555430674, "learning_rate": 1.5098619091305571e-05, "loss": 0.8975, "step": 1818 }, { "epoch": 0.3499591169255928, "grad_norm": 1.6585238785045908, "learning_rate": 1.5093256913773786e-05, "loss": 1.0496, "step": 1819 }, { "epoch": 0.3501515078639796, "grad_norm": 1.3543185908502522, "learning_rate": 1.50878927580797e-05, "loss": 0.9959, "step": 1820 }, { "epoch": 0.3503438988023664, "grad_norm": 1.4809857997972191, "learning_rate": 1.5082526626306698e-05, "loss": 0.9666, "step": 1821 }, { "epoch": 0.35053628974075324, "grad_norm": 1.4850550832444478, "learning_rate": 1.5077158520538921e-05, "loss": 1.0204, "step": 1822 }, { "epoch": 0.35072868067914004, "grad_norm": 1.4209297867929045, "learning_rate": 1.5071788442861277e-05, "loss": 1.045, "step": 1823 }, { "epoch": 0.35092107161752684, "grad_norm": 1.484903805814293, "learning_rate": 1.5066416395359444e-05, "loss": 0.8946, "step": 1824 }, { "epoch": 0.35111346255591364, "grad_norm": 1.2879709777382484, "learning_rate": 1.5061042380119864e-05, "loss": 0.9797, "step": 1825 }, { "epoch": 0.35130585349430044, "grad_norm": 1.3855134938170168, "learning_rate": 1.5055666399229743e-05, "loss": 1.0564, "step": 1826 }, { "epoch": 0.35149824443268723, "grad_norm": 1.3399269380888357, "learning_rate": 1.5050288454777047e-05, "loss": 0.9418, "step": 1827 }, { "epoch": 0.35169063537107403, "grad_norm": 1.369718400772686, "learning_rate": 1.504490854885051e-05, "loss": 1.0175, "step": 1828 }, { "epoch": 0.35188302630946083, "grad_norm": 1.343448561392123, "learning_rate": 1.5039526683539627e-05, "loss": 1.0041, "step": 1829 }, { "epoch": 0.35207541724784763, "grad_norm": 1.3929572526146399, "learning_rate": 1.5034142860934649e-05, "loss": 1.0547, "step": 1830 }, { "epoch": 0.35226780818623443, "grad_norm": 1.405884501719027, "learning_rate": 1.5028757083126594e-05, "loss": 1.0739, "step": 1831 }, { "epoch": 0.35246019912462123, "grad_norm": 1.366670857728935, "learning_rate": 1.5023369352207229e-05, "loss": 0.9706, "step": 1832 }, { "epoch": 0.35265259006300803, "grad_norm": 1.4260635592667876, "learning_rate": 1.5017979670269096e-05, "loss": 0.9956, "step": 1833 }, { "epoch": 0.35284498100139483, "grad_norm": 1.4054608810318365, "learning_rate": 1.501258803940548e-05, "loss": 1.0517, "step": 1834 }, { "epoch": 0.35303737193978163, "grad_norm": 1.458656929024581, "learning_rate": 1.500719446171043e-05, "loss": 0.8483, "step": 1835 }, { "epoch": 0.35322976287816843, "grad_norm": 1.466611845087547, "learning_rate": 1.500179893927875e-05, "loss": 0.9678, "step": 1836 }, { "epoch": 0.35342215381655523, "grad_norm": 1.4286946019324251, "learning_rate": 1.4996401474205997e-05, "loss": 1.0222, "step": 1837 }, { "epoch": 0.35361454475494203, "grad_norm": 1.4158641460363726, "learning_rate": 1.4991002068588484e-05, "loss": 0.9646, "step": 1838 }, { "epoch": 0.35380693569332883, "grad_norm": 1.4588894277241253, "learning_rate": 1.4985600724523282e-05, "loss": 0.989, "step": 1839 }, { "epoch": 0.3539993266317156, "grad_norm": 1.4044481317866861, "learning_rate": 1.4980197444108205e-05, "loss": 0.9832, "step": 1840 }, { "epoch": 0.3541917175701024, "grad_norm": 1.6081936542936035, "learning_rate": 1.4974792229441826e-05, "loss": 1.0373, "step": 1841 }, { "epoch": 0.3543841085084892, "grad_norm": 1.4062373483246684, "learning_rate": 1.4969385082623473e-05, "loss": 0.928, "step": 1842 }, { "epoch": 0.354576499446876, "grad_norm": 1.4638332674914085, "learning_rate": 1.4963976005753216e-05, "loss": 1.0213, "step": 1843 }, { "epoch": 0.3547688903852629, "grad_norm": 1.3725344810093554, "learning_rate": 1.4958565000931877e-05, "loss": 1.0031, "step": 1844 }, { "epoch": 0.3549612813236497, "grad_norm": 0.7800257655274754, "learning_rate": 1.4953152070261027e-05, "loss": 0.8259, "step": 1845 }, { "epoch": 0.3551536722620365, "grad_norm": 1.5587417120117881, "learning_rate": 1.494773721584299e-05, "loss": 0.9366, "step": 1846 }, { "epoch": 0.3553460632004233, "grad_norm": 1.6180874877239744, "learning_rate": 1.4942320439780833e-05, "loss": 0.992, "step": 1847 }, { "epoch": 0.3555384541388101, "grad_norm": 1.3814227470036675, "learning_rate": 1.4936901744178367e-05, "loss": 0.9565, "step": 1848 }, { "epoch": 0.3557308450771969, "grad_norm": 1.4334923388099883, "learning_rate": 1.4931481131140149e-05, "loss": 1.037, "step": 1849 }, { "epoch": 0.3559232360155837, "grad_norm": 1.4410528020729212, "learning_rate": 1.4926058602771484e-05, "loss": 1.0926, "step": 1850 }, { "epoch": 0.3561156269539705, "grad_norm": 1.3626762346681862, "learning_rate": 1.4920634161178424e-05, "loss": 0.9994, "step": 1851 }, { "epoch": 0.3563080178923573, "grad_norm": 1.4322898895377356, "learning_rate": 1.4915207808467756e-05, "loss": 1.0456, "step": 1852 }, { "epoch": 0.3565004088307441, "grad_norm": 1.599088112486679, "learning_rate": 1.4909779546747011e-05, "loss": 1.079, "step": 1853 }, { "epoch": 0.3566927997691309, "grad_norm": 1.3223100060908428, "learning_rate": 1.4904349378124467e-05, "loss": 1.0202, "step": 1854 }, { "epoch": 0.3568851907075177, "grad_norm": 1.2369587162106153, "learning_rate": 1.489891730470914e-05, "loss": 0.9901, "step": 1855 }, { "epoch": 0.3570775816459045, "grad_norm": 1.4198624737497918, "learning_rate": 1.4893483328610778e-05, "loss": 1.0285, "step": 1856 }, { "epoch": 0.3572699725842913, "grad_norm": 1.4044080173676363, "learning_rate": 1.488804745193988e-05, "loss": 0.9543, "step": 1857 }, { "epoch": 0.3574623635226781, "grad_norm": 1.3686641682003702, "learning_rate": 1.4882609676807675e-05, "loss": 0.9756, "step": 1858 }, { "epoch": 0.35765475446106487, "grad_norm": 1.3695396585063369, "learning_rate": 1.4877170005326136e-05, "loss": 1.0407, "step": 1859 }, { "epoch": 0.35784714539945167, "grad_norm": 0.7631407343022654, "learning_rate": 1.4871728439607967e-05, "loss": 0.8477, "step": 1860 }, { "epoch": 0.35803953633783847, "grad_norm": 1.3007914060990113, "learning_rate": 1.4866284981766607e-05, "loss": 1.037, "step": 1861 }, { "epoch": 0.35823192727622527, "grad_norm": 1.3397240083086446, "learning_rate": 1.4860839633916236e-05, "loss": 0.9742, "step": 1862 }, { "epoch": 0.35842431821461207, "grad_norm": 1.3309030274141942, "learning_rate": 1.4855392398171762e-05, "loss": 1.0409, "step": 1863 }, { "epoch": 0.35861670915299887, "grad_norm": 1.3426032419443255, "learning_rate": 1.484994327664883e-05, "loss": 0.9672, "step": 1864 }, { "epoch": 0.35880910009138567, "grad_norm": 1.4282720278434706, "learning_rate": 1.4844492271463814e-05, "loss": 0.966, "step": 1865 }, { "epoch": 0.3590014910297725, "grad_norm": 1.4263532746105674, "learning_rate": 1.4839039384733821e-05, "loss": 1.0192, "step": 1866 }, { "epoch": 0.3591938819681593, "grad_norm": 1.5097856208258111, "learning_rate": 1.4833584618576695e-05, "loss": 1.1151, "step": 1867 }, { "epoch": 0.3593862729065461, "grad_norm": 0.7561751909924007, "learning_rate": 1.4828127975111e-05, "loss": 0.8608, "step": 1868 }, { "epoch": 0.3595786638449329, "grad_norm": 1.3315605776004789, "learning_rate": 1.4822669456456031e-05, "loss": 0.9968, "step": 1869 }, { "epoch": 0.3597710547833197, "grad_norm": 1.44955645495706, "learning_rate": 1.4817209064731819e-05, "loss": 0.9248, "step": 1870 }, { "epoch": 0.3599634457217065, "grad_norm": 1.4042481421993336, "learning_rate": 1.4811746802059115e-05, "loss": 1.0935, "step": 1871 }, { "epoch": 0.3601558366600933, "grad_norm": 1.4914934475484274, "learning_rate": 1.48062826705594e-05, "loss": 1.0711, "step": 1872 }, { "epoch": 0.3603482275984801, "grad_norm": 1.4821204529951832, "learning_rate": 1.4800816672354876e-05, "loss": 0.9913, "step": 1873 }, { "epoch": 0.3605406185368669, "grad_norm": 1.53452125153387, "learning_rate": 1.4795348809568477e-05, "loss": 1.0071, "step": 1874 }, { "epoch": 0.3607330094752537, "grad_norm": 1.4031788709355106, "learning_rate": 1.4789879084323858e-05, "loss": 0.9316, "step": 1875 }, { "epoch": 0.3609254004136405, "grad_norm": 1.4799421011664744, "learning_rate": 1.4784407498745394e-05, "loss": 1.0351, "step": 1876 }, { "epoch": 0.3611177913520273, "grad_norm": 1.453293757351648, "learning_rate": 1.477893405495819e-05, "loss": 1.0085, "step": 1877 }, { "epoch": 0.3613101822904141, "grad_norm": 1.521153632045585, "learning_rate": 1.4773458755088068e-05, "loss": 0.9985, "step": 1878 }, { "epoch": 0.3615025732288009, "grad_norm": 1.4425380727191215, "learning_rate": 1.4767981601261567e-05, "loss": 1.1113, "step": 1879 }, { "epoch": 0.3616949641671877, "grad_norm": 1.4037129244685458, "learning_rate": 1.4762502595605957e-05, "loss": 0.9216, "step": 1880 }, { "epoch": 0.3618873551055745, "grad_norm": 1.3627063939714539, "learning_rate": 1.4757021740249213e-05, "loss": 0.9531, "step": 1881 }, { "epoch": 0.3620797460439613, "grad_norm": 1.3718237562623994, "learning_rate": 1.4751539037320044e-05, "loss": 1.0533, "step": 1882 }, { "epoch": 0.3622721369823481, "grad_norm": 1.4097400614496616, "learning_rate": 1.4746054488947863e-05, "loss": 0.9901, "step": 1883 }, { "epoch": 0.3624645279207349, "grad_norm": 1.597133363922736, "learning_rate": 1.4740568097262811e-05, "loss": 1.0698, "step": 1884 }, { "epoch": 0.3626569188591217, "grad_norm": 0.7843829803998575, "learning_rate": 1.473507986439573e-05, "loss": 0.9075, "step": 1885 }, { "epoch": 0.3628493097975085, "grad_norm": 1.4356826032604755, "learning_rate": 1.4729589792478193e-05, "loss": 0.9892, "step": 1886 }, { "epoch": 0.3630417007358953, "grad_norm": 1.468067743619024, "learning_rate": 1.4724097883642482e-05, "loss": 0.9668, "step": 1887 }, { "epoch": 0.36323409167428217, "grad_norm": 1.4131594110215049, "learning_rate": 1.4718604140021588e-05, "loss": 0.9766, "step": 1888 }, { "epoch": 0.36342648261266897, "grad_norm": 1.4218171451183255, "learning_rate": 1.471310856374922e-05, "loss": 1.0658, "step": 1889 }, { "epoch": 0.36361887355105577, "grad_norm": 1.4684896127576568, "learning_rate": 1.470761115695979e-05, "loss": 1.047, "step": 1890 }, { "epoch": 0.36381126448944257, "grad_norm": 1.5173516448451152, "learning_rate": 1.4702111921788437e-05, "loss": 0.9586, "step": 1891 }, { "epoch": 0.36400365542782936, "grad_norm": 1.56984433467405, "learning_rate": 1.4696610860370997e-05, "loss": 0.9662, "step": 1892 }, { "epoch": 0.36419604636621616, "grad_norm": 1.3606378184264531, "learning_rate": 1.4691107974844015e-05, "loss": 1.0002, "step": 1893 }, { "epoch": 0.36438843730460296, "grad_norm": 0.7960513984119993, "learning_rate": 1.468560326734475e-05, "loss": 0.8769, "step": 1894 }, { "epoch": 0.36458082824298976, "grad_norm": 1.4106268804240334, "learning_rate": 1.4680096740011172e-05, "loss": 1.0458, "step": 1895 }, { "epoch": 0.36477321918137656, "grad_norm": 1.415537969280195, "learning_rate": 1.4674588394981948e-05, "loss": 1.0596, "step": 1896 }, { "epoch": 0.36496561011976336, "grad_norm": 1.4710907167923089, "learning_rate": 1.4669078234396454e-05, "loss": 1.0746, "step": 1897 }, { "epoch": 0.36515800105815016, "grad_norm": 1.3642302701132085, "learning_rate": 1.4663566260394775e-05, "loss": 0.9854, "step": 1898 }, { "epoch": 0.36535039199653696, "grad_norm": 1.4368053086068273, "learning_rate": 1.4658052475117704e-05, "loss": 1.0192, "step": 1899 }, { "epoch": 0.36554278293492376, "grad_norm": 1.2667060479657353, "learning_rate": 1.4652536880706723e-05, "loss": 1.0321, "step": 1900 }, { "epoch": 0.36573517387331056, "grad_norm": 1.3063916016919388, "learning_rate": 1.4647019479304028e-05, "loss": 1.0586, "step": 1901 }, { "epoch": 0.36592756481169736, "grad_norm": 1.4491677415076878, "learning_rate": 1.4641500273052516e-05, "loss": 1.0504, "step": 1902 }, { "epoch": 0.36611995575008416, "grad_norm": 1.3232857305611345, "learning_rate": 1.463597926409578e-05, "loss": 1.0183, "step": 1903 }, { "epoch": 0.36631234668847096, "grad_norm": 1.5984984752065683, "learning_rate": 1.4630456454578122e-05, "loss": 1.0816, "step": 1904 }, { "epoch": 0.36650473762685776, "grad_norm": 1.4925310351634564, "learning_rate": 1.462493184664453e-05, "loss": 1.0645, "step": 1905 }, { "epoch": 0.36669712856524456, "grad_norm": 1.3617981725349106, "learning_rate": 1.4619405442440702e-05, "loss": 0.9965, "step": 1906 }, { "epoch": 0.36688951950363136, "grad_norm": 1.448914924870538, "learning_rate": 1.4613877244113033e-05, "loss": 1.0104, "step": 1907 }, { "epoch": 0.36708191044201816, "grad_norm": 1.4352301315726033, "learning_rate": 1.4608347253808605e-05, "loss": 1.0844, "step": 1908 }, { "epoch": 0.367274301380405, "grad_norm": 1.3819715779158677, "learning_rate": 1.460281547367521e-05, "loss": 0.9664, "step": 1909 }, { "epoch": 0.3674666923187918, "grad_norm": 0.6815584780096472, "learning_rate": 1.4597281905861318e-05, "loss": 0.8669, "step": 1910 }, { "epoch": 0.3676590832571786, "grad_norm": 1.324259967139786, "learning_rate": 1.4591746552516109e-05, "loss": 1.0243, "step": 1911 }, { "epoch": 0.3678514741955654, "grad_norm": 1.2720367425679273, "learning_rate": 1.4586209415789452e-05, "loss": 0.973, "step": 1912 }, { "epoch": 0.3680438651339522, "grad_norm": 1.4011628097800435, "learning_rate": 1.4580670497831904e-05, "loss": 1.0788, "step": 1913 }, { "epoch": 0.368236256072339, "grad_norm": 0.6467861991013064, "learning_rate": 1.4575129800794718e-05, "loss": 0.7739, "step": 1914 }, { "epoch": 0.3684286470107258, "grad_norm": 1.3616443098271318, "learning_rate": 1.4569587326829834e-05, "loss": 0.9728, "step": 1915 }, { "epoch": 0.3686210379491126, "grad_norm": 1.389491818700812, "learning_rate": 1.4564043078089891e-05, "loss": 1.0219, "step": 1916 }, { "epoch": 0.3688134288874994, "grad_norm": 1.4160188901948694, "learning_rate": 1.4558497056728205e-05, "loss": 0.9988, "step": 1917 }, { "epoch": 0.3690058198258862, "grad_norm": 1.333101128282083, "learning_rate": 1.4552949264898795e-05, "loss": 0.9399, "step": 1918 }, { "epoch": 0.369198210764273, "grad_norm": 1.311965840796405, "learning_rate": 1.4547399704756348e-05, "loss": 0.9581, "step": 1919 }, { "epoch": 0.3693906017026598, "grad_norm": 1.255570888531546, "learning_rate": 1.4541848378456255e-05, "loss": 0.9796, "step": 1920 }, { "epoch": 0.3695829926410466, "grad_norm": 1.4513995437585814, "learning_rate": 1.4536295288154594e-05, "loss": 1.0033, "step": 1921 }, { "epoch": 0.3697753835794334, "grad_norm": 1.4072946415499314, "learning_rate": 1.4530740436008111e-05, "loss": 1.0318, "step": 1922 }, { "epoch": 0.3699677745178202, "grad_norm": 0.7816117698872312, "learning_rate": 1.452518382417425e-05, "loss": 0.8608, "step": 1923 }, { "epoch": 0.370160165456207, "grad_norm": 1.4568841621240547, "learning_rate": 1.4519625454811135e-05, "loss": 0.9989, "step": 1924 }, { "epoch": 0.3703525563945938, "grad_norm": 0.686355957028045, "learning_rate": 1.4514065330077575e-05, "loss": 0.8002, "step": 1925 }, { "epoch": 0.3705449473329806, "grad_norm": 1.4180533361411911, "learning_rate": 1.4508503452133053e-05, "loss": 1.0686, "step": 1926 }, { "epoch": 0.3707373382713674, "grad_norm": 1.3751804586985403, "learning_rate": 1.4502939823137744e-05, "loss": 0.9938, "step": 1927 }, { "epoch": 0.3709297292097542, "grad_norm": 1.4453044579534173, "learning_rate": 1.4497374445252496e-05, "loss": 1.0319, "step": 1928 }, { "epoch": 0.371122120148141, "grad_norm": 1.3936080501366204, "learning_rate": 1.4491807320638835e-05, "loss": 1.0019, "step": 1929 }, { "epoch": 0.3713145110865278, "grad_norm": 1.3436501228519067, "learning_rate": 1.4486238451458972e-05, "loss": 1.0364, "step": 1930 }, { "epoch": 0.37150690202491465, "grad_norm": 1.3403836826220867, "learning_rate": 1.4480667839875786e-05, "loss": 1.0482, "step": 1931 }, { "epoch": 0.37169929296330145, "grad_norm": 1.4010403119542245, "learning_rate": 1.4475095488052843e-05, "loss": 1.019, "step": 1932 }, { "epoch": 0.37189168390168825, "grad_norm": 0.8233818897284773, "learning_rate": 1.4469521398154381e-05, "loss": 0.883, "step": 1933 }, { "epoch": 0.37208407484007505, "grad_norm": 1.3950715795951036, "learning_rate": 1.4463945572345308e-05, "loss": 0.9661, "step": 1934 }, { "epoch": 0.37227646577846185, "grad_norm": 1.3354664040386615, "learning_rate": 1.4458368012791213e-05, "loss": 1.0439, "step": 1935 }, { "epoch": 0.37246885671684865, "grad_norm": 1.4742402451985919, "learning_rate": 1.4452788721658355e-05, "loss": 1.0114, "step": 1936 }, { "epoch": 0.37266124765523545, "grad_norm": 1.3849636599532693, "learning_rate": 1.4447207701113669e-05, "loss": 1.0228, "step": 1937 }, { "epoch": 0.37285363859362225, "grad_norm": 1.3680645143589973, "learning_rate": 1.4441624953324755e-05, "loss": 0.9513, "step": 1938 }, { "epoch": 0.37304602953200905, "grad_norm": 1.4215326991692847, "learning_rate": 1.4436040480459891e-05, "loss": 0.9297, "step": 1939 }, { "epoch": 0.37323842047039585, "grad_norm": 1.4614649396692212, "learning_rate": 1.443045428468802e-05, "loss": 1.0813, "step": 1940 }, { "epoch": 0.37343081140878265, "grad_norm": 1.4267045684991513, "learning_rate": 1.4424866368178761e-05, "loss": 0.991, "step": 1941 }, { "epoch": 0.37362320234716945, "grad_norm": 1.309712312794925, "learning_rate": 1.441927673310239e-05, "loss": 0.8894, "step": 1942 }, { "epoch": 0.37381559328555625, "grad_norm": 1.3686337206580022, "learning_rate": 1.4413685381629855e-05, "loss": 0.9281, "step": 1943 }, { "epoch": 0.37400798422394305, "grad_norm": 1.3402402880066766, "learning_rate": 1.440809231593278e-05, "loss": 1.0358, "step": 1944 }, { "epoch": 0.37420037516232985, "grad_norm": 1.3586720083286536, "learning_rate": 1.4402497538183444e-05, "loss": 1.0057, "step": 1945 }, { "epoch": 0.37439276610071665, "grad_norm": 1.3798401290478928, "learning_rate": 1.4396901050554794e-05, "loss": 1.0778, "step": 1946 }, { "epoch": 0.37458515703910344, "grad_norm": 1.407076062611882, "learning_rate": 1.4391302855220442e-05, "loss": 0.9782, "step": 1947 }, { "epoch": 0.37477754797749024, "grad_norm": 1.4491675612805983, "learning_rate": 1.4385702954354662e-05, "loss": 1.0117, "step": 1948 }, { "epoch": 0.37496993891587704, "grad_norm": 1.390637245868594, "learning_rate": 1.438010135013239e-05, "loss": 1.0326, "step": 1949 }, { "epoch": 0.37516232985426384, "grad_norm": 1.4340102624162476, "learning_rate": 1.4374498044729225e-05, "loss": 0.9359, "step": 1950 }, { "epoch": 0.37535472079265064, "grad_norm": 1.4623624214280264, "learning_rate": 1.4368893040321428e-05, "loss": 1.0634, "step": 1951 }, { "epoch": 0.37554711173103744, "grad_norm": 1.3915613710539396, "learning_rate": 1.4363286339085915e-05, "loss": 1.0143, "step": 1952 }, { "epoch": 0.3757395026694243, "grad_norm": 1.3671090643104957, "learning_rate": 1.435767794320027e-05, "loss": 0.9506, "step": 1953 }, { "epoch": 0.3759318936078111, "grad_norm": 1.360581150696953, "learning_rate": 1.4352067854842724e-05, "loss": 1.0237, "step": 1954 }, { "epoch": 0.3761242845461979, "grad_norm": 1.3822787085820931, "learning_rate": 1.434645607619217e-05, "loss": 1.0844, "step": 1955 }, { "epoch": 0.3763166754845847, "grad_norm": 1.4511896560569295, "learning_rate": 1.434084260942816e-05, "loss": 1.0516, "step": 1956 }, { "epoch": 0.3765090664229715, "grad_norm": 1.4305611834621117, "learning_rate": 1.4335227456730902e-05, "loss": 1.0384, "step": 1957 }, { "epoch": 0.3767014573613583, "grad_norm": 1.4106280217359122, "learning_rate": 1.4329610620281253e-05, "loss": 0.9994, "step": 1958 }, { "epoch": 0.3768938482997451, "grad_norm": 1.3434196229984297, "learning_rate": 1.4323992102260733e-05, "loss": 0.9822, "step": 1959 }, { "epoch": 0.3770862392381319, "grad_norm": 0.7257094706218611, "learning_rate": 1.4318371904851502e-05, "loss": 0.8414, "step": 1960 }, { "epoch": 0.3772786301765187, "grad_norm": 1.3919710094103661, "learning_rate": 1.4312750030236382e-05, "loss": 1.0158, "step": 1961 }, { "epoch": 0.3774710211149055, "grad_norm": 1.440594780449693, "learning_rate": 1.4307126480598852e-05, "loss": 1.0831, "step": 1962 }, { "epoch": 0.3776634120532923, "grad_norm": 1.381760816608156, "learning_rate": 1.4301501258123024e-05, "loss": 1.0394, "step": 1963 }, { "epoch": 0.3778558029916791, "grad_norm": 1.3609925541006163, "learning_rate": 1.4295874364993672e-05, "loss": 0.9397, "step": 1964 }, { "epoch": 0.3780481939300659, "grad_norm": 0.7282663352112597, "learning_rate": 1.4290245803396221e-05, "loss": 0.799, "step": 1965 }, { "epoch": 0.3782405848684527, "grad_norm": 0.712578940160904, "learning_rate": 1.4284615575516737e-05, "loss": 0.7639, "step": 1966 }, { "epoch": 0.3784329758068395, "grad_norm": 1.344150819254214, "learning_rate": 1.4278983683541934e-05, "loss": 0.9994, "step": 1967 }, { "epoch": 0.3786253667452263, "grad_norm": 1.334318950369608, "learning_rate": 1.4273350129659173e-05, "loss": 1.0231, "step": 1968 }, { "epoch": 0.3788177576836131, "grad_norm": 1.306421463037682, "learning_rate": 1.4267714916056465e-05, "loss": 0.9068, "step": 1969 }, { "epoch": 0.3790101486219999, "grad_norm": 1.4432841486803476, "learning_rate": 1.426207804492246e-05, "loss": 1.0221, "step": 1970 }, { "epoch": 0.3792025395603867, "grad_norm": 1.4557997604148725, "learning_rate": 1.4256439518446456e-05, "loss": 1.0062, "step": 1971 }, { "epoch": 0.3793949304987735, "grad_norm": 1.3347622111176207, "learning_rate": 1.4250799338818388e-05, "loss": 1.0838, "step": 1972 }, { "epoch": 0.3795873214371603, "grad_norm": 1.3886461828660361, "learning_rate": 1.424515750822884e-05, "loss": 0.9821, "step": 1973 }, { "epoch": 0.3797797123755471, "grad_norm": 1.4482133397709676, "learning_rate": 1.4239514028869032e-05, "loss": 1.0189, "step": 1974 }, { "epoch": 0.37997210331393394, "grad_norm": 1.3259025887673128, "learning_rate": 1.4233868902930827e-05, "loss": 0.974, "step": 1975 }, { "epoch": 0.38016449425232074, "grad_norm": 1.4299569713407476, "learning_rate": 1.4228222132606729e-05, "loss": 0.9913, "step": 1976 }, { "epoch": 0.38035688519070754, "grad_norm": 1.4248559016004863, "learning_rate": 1.4222573720089874e-05, "loss": 0.946, "step": 1977 }, { "epoch": 0.38054927612909434, "grad_norm": 1.344847010953489, "learning_rate": 1.4216923667574042e-05, "loss": 0.9712, "step": 1978 }, { "epoch": 0.38074166706748114, "grad_norm": 1.4223338527190945, "learning_rate": 1.4211271977253653e-05, "loss": 0.9956, "step": 1979 }, { "epoch": 0.38093405800586794, "grad_norm": 1.4131386849654466, "learning_rate": 1.4205618651323753e-05, "loss": 0.9531, "step": 1980 }, { "epoch": 0.38112644894425474, "grad_norm": 1.356800967964039, "learning_rate": 1.4199963691980027e-05, "loss": 1.0438, "step": 1981 }, { "epoch": 0.38131883988264154, "grad_norm": 1.4682234238735048, "learning_rate": 1.4194307101418805e-05, "loss": 1.0501, "step": 1982 }, { "epoch": 0.38151123082102834, "grad_norm": 1.4508955531218748, "learning_rate": 1.4188648881837033e-05, "loss": 1.0314, "step": 1983 }, { "epoch": 0.38170362175941513, "grad_norm": 1.3282515726859514, "learning_rate": 1.4182989035432299e-05, "loss": 0.9277, "step": 1984 }, { "epoch": 0.38189601269780193, "grad_norm": 1.3255823930367763, "learning_rate": 1.4177327564402825e-05, "loss": 0.9599, "step": 1985 }, { "epoch": 0.38208840363618873, "grad_norm": 1.5611793738329527, "learning_rate": 1.4171664470947464e-05, "loss": 0.9634, "step": 1986 }, { "epoch": 0.38228079457457553, "grad_norm": 1.4080306937681553, "learning_rate": 1.416599975726569e-05, "loss": 1.0961, "step": 1987 }, { "epoch": 0.38247318551296233, "grad_norm": 1.3394714507599093, "learning_rate": 1.4160333425557616e-05, "loss": 1.0266, "step": 1988 }, { "epoch": 0.38266557645134913, "grad_norm": 1.3871591475773342, "learning_rate": 1.4154665478023977e-05, "loss": 0.9555, "step": 1989 }, { "epoch": 0.38285796738973593, "grad_norm": 1.4657713801115002, "learning_rate": 1.4148995916866139e-05, "loss": 1.0102, "step": 1990 }, { "epoch": 0.38305035832812273, "grad_norm": 1.2982889446455355, "learning_rate": 1.41433247442861e-05, "loss": 0.9921, "step": 1991 }, { "epoch": 0.38324274926650953, "grad_norm": 1.3629789197486382, "learning_rate": 1.4137651962486472e-05, "loss": 1.0098, "step": 1992 }, { "epoch": 0.38343514020489633, "grad_norm": 1.3585875652233632, "learning_rate": 1.4131977573670499e-05, "loss": 1.0436, "step": 1993 }, { "epoch": 0.38362753114328313, "grad_norm": 1.373575039603788, "learning_rate": 1.412630158004205e-05, "loss": 0.9796, "step": 1994 }, { "epoch": 0.38381992208166993, "grad_norm": 1.3471754771601945, "learning_rate": 1.4120623983805617e-05, "loss": 1.0089, "step": 1995 }, { "epoch": 0.38401231302005673, "grad_norm": 1.4144171341829135, "learning_rate": 1.4114944787166307e-05, "loss": 1.0192, "step": 1996 }, { "epoch": 0.3842047039584436, "grad_norm": 1.5292715605986498, "learning_rate": 1.4109263992329858e-05, "loss": 1.0404, "step": 1997 }, { "epoch": 0.3843970948968304, "grad_norm": 1.5116641761646408, "learning_rate": 1.4103581601502629e-05, "loss": 0.9004, "step": 1998 }, { "epoch": 0.3845894858352172, "grad_norm": 1.3342240465319086, "learning_rate": 1.409789761689159e-05, "loss": 1.0554, "step": 1999 }, { "epoch": 0.384781876773604, "grad_norm": 1.2800451257449443, "learning_rate": 1.4092212040704336e-05, "loss": 1.0057, "step": 2000 }, { "epoch": 0.3849742677119908, "grad_norm": 1.386627729381219, "learning_rate": 1.408652487514908e-05, "loss": 1.0091, "step": 2001 }, { "epoch": 0.3851666586503776, "grad_norm": 1.3600845059193845, "learning_rate": 1.408083612243465e-05, "loss": 0.9811, "step": 2002 }, { "epoch": 0.3853590495887644, "grad_norm": 1.3781798060964794, "learning_rate": 1.4075145784770496e-05, "loss": 0.9284, "step": 2003 }, { "epoch": 0.3855514405271512, "grad_norm": 1.418077798246999, "learning_rate": 1.4069453864366678e-05, "loss": 0.9859, "step": 2004 }, { "epoch": 0.385743831465538, "grad_norm": 1.4277190283850743, "learning_rate": 1.4063760363433867e-05, "loss": 1.0491, "step": 2005 }, { "epoch": 0.3859362224039248, "grad_norm": 1.3023003551106773, "learning_rate": 1.405806528418336e-05, "loss": 0.9375, "step": 2006 }, { "epoch": 0.3861286133423116, "grad_norm": 0.784714085141282, "learning_rate": 1.4052368628827057e-05, "loss": 0.8826, "step": 2007 }, { "epoch": 0.3863210042806984, "grad_norm": 1.38934248793402, "learning_rate": 1.4046670399577478e-05, "loss": 0.9199, "step": 2008 }, { "epoch": 0.3865133952190852, "grad_norm": 1.601034548035233, "learning_rate": 1.4040970598647742e-05, "loss": 0.9509, "step": 2009 }, { "epoch": 0.386705786157472, "grad_norm": 1.469177097080399, "learning_rate": 1.4035269228251589e-05, "loss": 0.9617, "step": 2010 }, { "epoch": 0.3868981770958588, "grad_norm": 1.223062906393392, "learning_rate": 1.4029566290603368e-05, "loss": 0.9608, "step": 2011 }, { "epoch": 0.3870905680342456, "grad_norm": 0.7097338492273629, "learning_rate": 1.4023861787918031e-05, "loss": 0.8343, "step": 2012 }, { "epoch": 0.3872829589726324, "grad_norm": 1.3971256923242243, "learning_rate": 1.4018155722411144e-05, "loss": 1.0311, "step": 2013 }, { "epoch": 0.3874753499110192, "grad_norm": 1.4786420975462562, "learning_rate": 1.4012448096298874e-05, "loss": 0.9988, "step": 2014 }, { "epoch": 0.387667740849406, "grad_norm": 1.3145561158313954, "learning_rate": 1.4006738911798001e-05, "loss": 0.9929, "step": 2015 }, { "epoch": 0.3878601317877928, "grad_norm": 1.4148328535283319, "learning_rate": 1.40010281711259e-05, "loss": 1.0313, "step": 2016 }, { "epoch": 0.38805252272617957, "grad_norm": 1.2525833787695704, "learning_rate": 1.3995315876500565e-05, "loss": 1.0202, "step": 2017 }, { "epoch": 0.38824491366456637, "grad_norm": 1.3848087632219723, "learning_rate": 1.3989602030140581e-05, "loss": 1.0326, "step": 2018 }, { "epoch": 0.3884373046029532, "grad_norm": 1.4698683220153919, "learning_rate": 1.398388663426514e-05, "loss": 1.0213, "step": 2019 }, { "epoch": 0.38862969554134, "grad_norm": 1.420703494940033, "learning_rate": 1.3978169691094037e-05, "loss": 1.0303, "step": 2020 }, { "epoch": 0.3888220864797268, "grad_norm": 0.8017221117372137, "learning_rate": 1.3972451202847665e-05, "loss": 0.8613, "step": 2021 }, { "epoch": 0.3890144774181136, "grad_norm": 1.5580890625903716, "learning_rate": 1.3966731171747024e-05, "loss": 0.9807, "step": 2022 }, { "epoch": 0.3892068683565004, "grad_norm": 1.435003222120676, "learning_rate": 1.3961009600013702e-05, "loss": 0.955, "step": 2023 }, { "epoch": 0.3893992592948872, "grad_norm": 1.5374181788520123, "learning_rate": 1.3955286489869894e-05, "loss": 1.1002, "step": 2024 }, { "epoch": 0.389591650233274, "grad_norm": 1.3881423585890171, "learning_rate": 1.394956184353839e-05, "loss": 1.0624, "step": 2025 }, { "epoch": 0.3897840411716608, "grad_norm": 1.470877876224803, "learning_rate": 1.3943835663242577e-05, "loss": 1.0277, "step": 2026 }, { "epoch": 0.3899764321100476, "grad_norm": 1.3752295749511065, "learning_rate": 1.3938107951206438e-05, "loss": 0.9859, "step": 2027 }, { "epoch": 0.3901688230484344, "grad_norm": 1.3432255808375524, "learning_rate": 1.3932378709654548e-05, "loss": 0.9691, "step": 2028 }, { "epoch": 0.3903612139868212, "grad_norm": 1.403635754220005, "learning_rate": 1.3926647940812081e-05, "loss": 1.0315, "step": 2029 }, { "epoch": 0.390553604925208, "grad_norm": 1.3337140786979311, "learning_rate": 1.39209156469048e-05, "loss": 1.016, "step": 2030 }, { "epoch": 0.3907459958635948, "grad_norm": 1.334613879122619, "learning_rate": 1.3915181830159061e-05, "loss": 0.9136, "step": 2031 }, { "epoch": 0.3909383868019816, "grad_norm": 0.7024266766085506, "learning_rate": 1.3909446492801819e-05, "loss": 0.8424, "step": 2032 }, { "epoch": 0.3911307777403684, "grad_norm": 1.4387454606485943, "learning_rate": 1.3903709637060605e-05, "loss": 0.9223, "step": 2033 }, { "epoch": 0.3913231686787552, "grad_norm": 1.471401216100159, "learning_rate": 1.3897971265163546e-05, "loss": 1.0442, "step": 2034 }, { "epoch": 0.391515559617142, "grad_norm": 0.7091647453942185, "learning_rate": 1.3892231379339369e-05, "loss": 0.8332, "step": 2035 }, { "epoch": 0.3917079505555288, "grad_norm": 1.3455841135589968, "learning_rate": 1.3886489981817375e-05, "loss": 0.9836, "step": 2036 }, { "epoch": 0.3919003414939156, "grad_norm": 1.5077665662142836, "learning_rate": 1.3880747074827454e-05, "loss": 1.0935, "step": 2037 }, { "epoch": 0.3920927324323024, "grad_norm": 0.8197434425579366, "learning_rate": 1.3875002660600085e-05, "loss": 0.8521, "step": 2038 }, { "epoch": 0.3922851233706892, "grad_norm": 1.4097256766915822, "learning_rate": 1.386925674136634e-05, "loss": 1.0641, "step": 2039 }, { "epoch": 0.39247751430907607, "grad_norm": 0.7211520723564182, "learning_rate": 1.3863509319357857e-05, "loss": 0.8344, "step": 2040 }, { "epoch": 0.39266990524746287, "grad_norm": 1.4851202798751233, "learning_rate": 1.3857760396806876e-05, "loss": 1.0178, "step": 2041 }, { "epoch": 0.39286229618584967, "grad_norm": 1.3611015992080993, "learning_rate": 1.3852009975946209e-05, "loss": 1.0681, "step": 2042 }, { "epoch": 0.39305468712423647, "grad_norm": 1.5811894757103406, "learning_rate": 1.3846258059009252e-05, "loss": 0.9559, "step": 2043 }, { "epoch": 0.39324707806262327, "grad_norm": 1.3917932900611247, "learning_rate": 1.384050464822999e-05, "loss": 0.9864, "step": 2044 }, { "epoch": 0.39343946900101007, "grad_norm": 1.3427141718503341, "learning_rate": 1.383474974584297e-05, "loss": 1.0544, "step": 2045 }, { "epoch": 0.39363185993939687, "grad_norm": 1.3066471065357215, "learning_rate": 1.3828993354083342e-05, "loss": 1.0302, "step": 2046 }, { "epoch": 0.39382425087778367, "grad_norm": 1.4011552241231187, "learning_rate": 1.3823235475186816e-05, "loss": 0.9389, "step": 2047 }, { "epoch": 0.39401664181617047, "grad_norm": 1.6137216109461727, "learning_rate": 1.3817476111389685e-05, "loss": 1.032, "step": 2048 }, { "epoch": 0.39420903275455726, "grad_norm": 1.3567351474947387, "learning_rate": 1.3811715264928824e-05, "loss": 0.946, "step": 2049 }, { "epoch": 0.39440142369294406, "grad_norm": 1.294158607144472, "learning_rate": 1.3805952938041674e-05, "loss": 0.9834, "step": 2050 }, { "epoch": 0.39459381463133086, "grad_norm": 1.3684256944261293, "learning_rate": 1.3800189132966257e-05, "loss": 1.0121, "step": 2051 }, { "epoch": 0.39478620556971766, "grad_norm": 1.466594247884106, "learning_rate": 1.3794423851941174e-05, "loss": 0.9694, "step": 2052 }, { "epoch": 0.39497859650810446, "grad_norm": 0.7368677514276096, "learning_rate": 1.378865709720559e-05, "loss": 0.8435, "step": 2053 }, { "epoch": 0.39517098744649126, "grad_norm": 1.4028305292409444, "learning_rate": 1.3782888870999245e-05, "loss": 0.9507, "step": 2054 }, { "epoch": 0.39536337838487806, "grad_norm": 1.383029381473231, "learning_rate": 1.377711917556245e-05, "loss": 0.9502, "step": 2055 }, { "epoch": 0.39555576932326486, "grad_norm": 1.3634966533364767, "learning_rate": 1.3771348013136096e-05, "loss": 0.9276, "step": 2056 }, { "epoch": 0.39574816026165166, "grad_norm": 1.400924038153296, "learning_rate": 1.3765575385961627e-05, "loss": 0.9937, "step": 2057 }, { "epoch": 0.39594055120003846, "grad_norm": 1.5434204481497362, "learning_rate": 1.3759801296281072e-05, "loss": 1.0681, "step": 2058 }, { "epoch": 0.39613294213842526, "grad_norm": 1.3464756940941742, "learning_rate": 1.3754025746337014e-05, "loss": 0.9835, "step": 2059 }, { "epoch": 0.39632533307681206, "grad_norm": 1.5036268901970136, "learning_rate": 1.3748248738372616e-05, "loss": 0.9521, "step": 2060 }, { "epoch": 0.39651772401519886, "grad_norm": 1.4939063764087734, "learning_rate": 1.3742470274631599e-05, "loss": 0.9911, "step": 2061 }, { "epoch": 0.3967101149535857, "grad_norm": 1.4324449840021627, "learning_rate": 1.3736690357358253e-05, "loss": 0.9455, "step": 2062 }, { "epoch": 0.3969025058919725, "grad_norm": 1.4453043441029922, "learning_rate": 1.3730908988797427e-05, "loss": 1.1206, "step": 2063 }, { "epoch": 0.3970948968303593, "grad_norm": 1.4785530967525702, "learning_rate": 1.3725126171194543e-05, "loss": 0.9631, "step": 2064 }, { "epoch": 0.3972872877687461, "grad_norm": 1.4272471122815786, "learning_rate": 1.371934190679558e-05, "loss": 1.0212, "step": 2065 }, { "epoch": 0.3974796787071329, "grad_norm": 1.343044969915809, "learning_rate": 1.3713556197847076e-05, "loss": 0.986, "step": 2066 }, { "epoch": 0.3976720696455197, "grad_norm": 1.5479881783618552, "learning_rate": 1.3707769046596136e-05, "loss": 1.0565, "step": 2067 }, { "epoch": 0.3978644605839065, "grad_norm": 1.4988831965286287, "learning_rate": 1.3701980455290425e-05, "loss": 1.0284, "step": 2068 }, { "epoch": 0.3980568515222933, "grad_norm": 1.4458517849963957, "learning_rate": 1.3696190426178162e-05, "loss": 0.953, "step": 2069 }, { "epoch": 0.3982492424606801, "grad_norm": 1.292239716522804, "learning_rate": 1.3690398961508128e-05, "loss": 0.9237, "step": 2070 }, { "epoch": 0.3984416333990669, "grad_norm": 1.44175057798763, "learning_rate": 1.3684606063529662e-05, "loss": 0.9414, "step": 2071 }, { "epoch": 0.3986340243374537, "grad_norm": 1.3676237536062659, "learning_rate": 1.3678811734492659e-05, "loss": 0.9965, "step": 2072 }, { "epoch": 0.3988264152758405, "grad_norm": 1.5001176013965818, "learning_rate": 1.367301597664757e-05, "loss": 0.9707, "step": 2073 }, { "epoch": 0.3990188062142273, "grad_norm": 1.4089170219874096, "learning_rate": 1.36672187922454e-05, "loss": 0.9793, "step": 2074 }, { "epoch": 0.3992111971526141, "grad_norm": 1.3731373486973466, "learning_rate": 1.3661420183537705e-05, "loss": 0.9841, "step": 2075 }, { "epoch": 0.3994035880910009, "grad_norm": 1.3789915093560048, "learning_rate": 1.3655620152776605e-05, "loss": 1.0269, "step": 2076 }, { "epoch": 0.3995959790293877, "grad_norm": 1.452335375174253, "learning_rate": 1.364981870221476e-05, "loss": 1.0553, "step": 2077 }, { "epoch": 0.3997883699677745, "grad_norm": 0.8124206483050699, "learning_rate": 1.364401583410539e-05, "loss": 0.9227, "step": 2078 }, { "epoch": 0.3999807609061613, "grad_norm": 1.341209729986724, "learning_rate": 1.3638211550702256e-05, "loss": 0.9428, "step": 2079 }, { "epoch": 0.4001731518445481, "grad_norm": 1.4155660508750256, "learning_rate": 1.363240585425968e-05, "loss": 1.0013, "step": 2080 }, { "epoch": 0.4003655427829349, "grad_norm": 1.3980864531220885, "learning_rate": 1.362659874703253e-05, "loss": 0.9893, "step": 2081 }, { "epoch": 0.4005579337213217, "grad_norm": 1.4278471124448529, "learning_rate": 1.3620790231276213e-05, "loss": 0.9962, "step": 2082 }, { "epoch": 0.4007503246597085, "grad_norm": 1.357881406527882, "learning_rate": 1.3614980309246692e-05, "loss": 1.0595, "step": 2083 }, { "epoch": 0.40094271559809536, "grad_norm": 1.2575737699952552, "learning_rate": 1.3609168983200474e-05, "loss": 1.0511, "step": 2084 }, { "epoch": 0.40113510653648216, "grad_norm": 1.4504850048409452, "learning_rate": 1.3603356255394613e-05, "loss": 1.0101, "step": 2085 }, { "epoch": 0.40132749747486895, "grad_norm": 1.4350224859794687, "learning_rate": 1.3597542128086702e-05, "loss": 1.0429, "step": 2086 }, { "epoch": 0.40151988841325575, "grad_norm": 1.373757720456018, "learning_rate": 1.3591726603534885e-05, "loss": 1.0095, "step": 2087 }, { "epoch": 0.40171227935164255, "grad_norm": 1.3116876962071768, "learning_rate": 1.3585909683997842e-05, "loss": 1.0089, "step": 2088 }, { "epoch": 0.40190467029002935, "grad_norm": 1.5945042222056847, "learning_rate": 1.3580091371734798e-05, "loss": 0.9715, "step": 2089 }, { "epoch": 0.40209706122841615, "grad_norm": 1.414653564142914, "learning_rate": 1.357427166900552e-05, "loss": 1.0263, "step": 2090 }, { "epoch": 0.40228945216680295, "grad_norm": 1.3940756720433143, "learning_rate": 1.3568450578070309e-05, "loss": 1.0816, "step": 2091 }, { "epoch": 0.40248184310518975, "grad_norm": 1.3465342351931795, "learning_rate": 1.3562628101190015e-05, "loss": 0.9286, "step": 2092 }, { "epoch": 0.40267423404357655, "grad_norm": 1.50455172318564, "learning_rate": 1.3556804240626019e-05, "loss": 1.1041, "step": 2093 }, { "epoch": 0.40286662498196335, "grad_norm": 1.3894044778820998, "learning_rate": 1.3550978998640241e-05, "loss": 0.984, "step": 2094 }, { "epoch": 0.40305901592035015, "grad_norm": 1.4090912915578322, "learning_rate": 1.3545152377495136e-05, "loss": 1.0324, "step": 2095 }, { "epoch": 0.40325140685873695, "grad_norm": 1.3796517515838613, "learning_rate": 1.3539324379453698e-05, "loss": 0.8934, "step": 2096 }, { "epoch": 0.40344379779712375, "grad_norm": 1.4372099038090216, "learning_rate": 1.3533495006779455e-05, "loss": 1.0543, "step": 2097 }, { "epoch": 0.40363618873551055, "grad_norm": 1.3249282764318457, "learning_rate": 1.3527664261736471e-05, "loss": 0.9352, "step": 2098 }, { "epoch": 0.40382857967389735, "grad_norm": 1.2739624941164576, "learning_rate": 1.3521832146589335e-05, "loss": 0.9138, "step": 2099 }, { "epoch": 0.40402097061228415, "grad_norm": 1.3955709998197101, "learning_rate": 1.3515998663603174e-05, "loss": 0.9616, "step": 2100 }, { "epoch": 0.40421336155067095, "grad_norm": 1.3979897744658987, "learning_rate": 1.3510163815043647e-05, "loss": 1.0381, "step": 2101 }, { "epoch": 0.40440575248905775, "grad_norm": 1.5689581437261642, "learning_rate": 1.3504327603176943e-05, "loss": 0.9947, "step": 2102 }, { "epoch": 0.40459814342744455, "grad_norm": 1.3074413151015345, "learning_rate": 1.3498490030269782e-05, "loss": 0.917, "step": 2103 }, { "epoch": 0.40479053436583134, "grad_norm": 1.4325312446951086, "learning_rate": 1.3492651098589398e-05, "loss": 1.0428, "step": 2104 }, { "epoch": 0.40498292530421814, "grad_norm": 1.5745884793934461, "learning_rate": 1.3486810810403578e-05, "loss": 0.9676, "step": 2105 }, { "epoch": 0.405175316242605, "grad_norm": 1.377648709919026, "learning_rate": 1.348096916798062e-05, "loss": 0.9804, "step": 2106 }, { "epoch": 0.4053677071809918, "grad_norm": 1.3039382333385796, "learning_rate": 1.3475126173589343e-05, "loss": 0.8743, "step": 2107 }, { "epoch": 0.4055600981193786, "grad_norm": 1.3833280388762712, "learning_rate": 1.3469281829499107e-05, "loss": 0.906, "step": 2108 }, { "epoch": 0.4057524890577654, "grad_norm": 1.369509341823081, "learning_rate": 1.3463436137979786e-05, "loss": 0.9888, "step": 2109 }, { "epoch": 0.4059448799961522, "grad_norm": 1.401327841657773, "learning_rate": 1.3457589101301776e-05, "loss": 0.9999, "step": 2110 }, { "epoch": 0.406137270934539, "grad_norm": 0.6964921092973995, "learning_rate": 1.3451740721736005e-05, "loss": 0.7835, "step": 2111 }, { "epoch": 0.4063296618729258, "grad_norm": 1.4382123374650742, "learning_rate": 1.3445891001553905e-05, "loss": 0.9676, "step": 2112 }, { "epoch": 0.4065220528113126, "grad_norm": 1.4376650399718882, "learning_rate": 1.3440039943027452e-05, "loss": 1.0682, "step": 2113 }, { "epoch": 0.4067144437496994, "grad_norm": 0.7421666377807957, "learning_rate": 1.3434187548429126e-05, "loss": 0.8428, "step": 2114 }, { "epoch": 0.4069068346880862, "grad_norm": 1.5578714048311686, "learning_rate": 1.3428333820031922e-05, "loss": 1.0238, "step": 2115 }, { "epoch": 0.407099225626473, "grad_norm": 1.3840685360117053, "learning_rate": 1.3422478760109371e-05, "loss": 1.0509, "step": 2116 }, { "epoch": 0.4072916165648598, "grad_norm": 1.2941498300383014, "learning_rate": 1.3416622370935507e-05, "loss": 0.9813, "step": 2117 }, { "epoch": 0.4074840075032466, "grad_norm": 1.3694936865754666, "learning_rate": 1.3410764654784885e-05, "loss": 1.0438, "step": 2118 }, { "epoch": 0.4076763984416334, "grad_norm": 1.5413850062207581, "learning_rate": 1.3404905613932573e-05, "loss": 0.9997, "step": 2119 }, { "epoch": 0.4078687893800202, "grad_norm": 1.2762379527097782, "learning_rate": 1.3399045250654152e-05, "loss": 0.9369, "step": 2120 }, { "epoch": 0.408061180318407, "grad_norm": 1.324599749903284, "learning_rate": 1.3393183567225724e-05, "loss": 0.9979, "step": 2121 }, { "epoch": 0.4082535712567938, "grad_norm": 1.4663408417421973, "learning_rate": 1.3387320565923901e-05, "loss": 1.0163, "step": 2122 }, { "epoch": 0.4084459621951806, "grad_norm": 1.3214333811429577, "learning_rate": 1.33814562490258e-05, "loss": 0.9735, "step": 2123 }, { "epoch": 0.4086383531335674, "grad_norm": 1.3456148180465206, "learning_rate": 1.3375590618809056e-05, "loss": 1.0145, "step": 2124 }, { "epoch": 0.4088307440719542, "grad_norm": 1.3768505625930307, "learning_rate": 1.3369723677551813e-05, "loss": 1.0527, "step": 2125 }, { "epoch": 0.409023135010341, "grad_norm": 1.4834576999723397, "learning_rate": 1.3363855427532724e-05, "loss": 0.9275, "step": 2126 }, { "epoch": 0.4092155259487278, "grad_norm": 1.2985223859771764, "learning_rate": 1.3357985871030948e-05, "loss": 0.9596, "step": 2127 }, { "epoch": 0.40940791688711464, "grad_norm": 1.4671133037212671, "learning_rate": 1.3352115010326155e-05, "loss": 1.1656, "step": 2128 }, { "epoch": 0.40960030782550144, "grad_norm": 1.334743551745226, "learning_rate": 1.3346242847698516e-05, "loss": 1.0702, "step": 2129 }, { "epoch": 0.40979269876388824, "grad_norm": 1.2764430159499478, "learning_rate": 1.3340369385428713e-05, "loss": 0.9633, "step": 2130 }, { "epoch": 0.40998508970227504, "grad_norm": 1.4069978079452659, "learning_rate": 1.3334494625797936e-05, "loss": 1.0162, "step": 2131 }, { "epoch": 0.41017748064066184, "grad_norm": 1.2849478799755207, "learning_rate": 1.3328618571087867e-05, "loss": 1.0086, "step": 2132 }, { "epoch": 0.41036987157904864, "grad_norm": 1.3527195384748214, "learning_rate": 1.33227412235807e-05, "loss": 0.9728, "step": 2133 }, { "epoch": 0.41056226251743544, "grad_norm": 1.247641500184482, "learning_rate": 1.3316862585559132e-05, "loss": 1.0381, "step": 2134 }, { "epoch": 0.41075465345582224, "grad_norm": 1.5183899086985482, "learning_rate": 1.3310982659306352e-05, "loss": 0.9738, "step": 2135 }, { "epoch": 0.41094704439420904, "grad_norm": 1.403815733106646, "learning_rate": 1.3305101447106064e-05, "loss": 0.9969, "step": 2136 }, { "epoch": 0.41113943533259584, "grad_norm": 1.3741128807227696, "learning_rate": 1.3299218951242456e-05, "loss": 1.0218, "step": 2137 }, { "epoch": 0.41133182627098264, "grad_norm": 1.3658180539513312, "learning_rate": 1.3293335174000226e-05, "loss": 1.047, "step": 2138 }, { "epoch": 0.41152421720936944, "grad_norm": 1.3723743609918257, "learning_rate": 1.328745011766456e-05, "loss": 0.995, "step": 2139 }, { "epoch": 0.41171660814775624, "grad_norm": 1.3194271257903918, "learning_rate": 1.3281563784521154e-05, "loss": 1.0387, "step": 2140 }, { "epoch": 0.41190899908614304, "grad_norm": 1.3828273008089857, "learning_rate": 1.3275676176856185e-05, "loss": 0.9873, "step": 2141 }, { "epoch": 0.41210139002452983, "grad_norm": 0.723246079227102, "learning_rate": 1.3269787296956333e-05, "loss": 0.8169, "step": 2142 }, { "epoch": 0.41229378096291663, "grad_norm": 1.4263041564013788, "learning_rate": 1.3263897147108778e-05, "loss": 1.0042, "step": 2143 }, { "epoch": 0.41248617190130343, "grad_norm": 1.349041350554299, "learning_rate": 1.3258005729601178e-05, "loss": 1.0371, "step": 2144 }, { "epoch": 0.41267856283969023, "grad_norm": 1.3611189504723227, "learning_rate": 1.3252113046721692e-05, "loss": 0.9387, "step": 2145 }, { "epoch": 0.41287095377807703, "grad_norm": 1.4018767910538386, "learning_rate": 1.3246219100758974e-05, "loss": 0.9706, "step": 2146 }, { "epoch": 0.41306334471646383, "grad_norm": 1.384698566515634, "learning_rate": 1.3240323894002166e-05, "loss": 1.0089, "step": 2147 }, { "epoch": 0.41325573565485063, "grad_norm": 1.5236018807870113, "learning_rate": 1.3234427428740895e-05, "loss": 0.957, "step": 2148 }, { "epoch": 0.4134481265932375, "grad_norm": 1.4079763408058374, "learning_rate": 1.3228529707265279e-05, "loss": 1.0558, "step": 2149 }, { "epoch": 0.4136405175316243, "grad_norm": 1.4738071067531118, "learning_rate": 1.322263073186593e-05, "loss": 1.0575, "step": 2150 }, { "epoch": 0.4138329084700111, "grad_norm": 1.440384001406343, "learning_rate": 1.3216730504833938e-05, "loss": 1.0, "step": 2151 }, { "epoch": 0.4140252994083979, "grad_norm": 1.308498901514196, "learning_rate": 1.3210829028460883e-05, "loss": 0.947, "step": 2152 }, { "epoch": 0.4142176903467847, "grad_norm": 1.343117893777874, "learning_rate": 1.3204926305038832e-05, "loss": 0.9928, "step": 2153 }, { "epoch": 0.4144100812851715, "grad_norm": 0.7609756072070647, "learning_rate": 1.3199022336860335e-05, "loss": 0.886, "step": 2154 }, { "epoch": 0.4146024722235583, "grad_norm": 1.4881611594907915, "learning_rate": 1.3193117126218425e-05, "loss": 0.9652, "step": 2155 }, { "epoch": 0.4147948631619451, "grad_norm": 1.3335725890780263, "learning_rate": 1.3187210675406617e-05, "loss": 0.975, "step": 2156 }, { "epoch": 0.4149872541003319, "grad_norm": 1.4339523729777819, "learning_rate": 1.318130298671891e-05, "loss": 1.0723, "step": 2157 }, { "epoch": 0.4151796450387187, "grad_norm": 1.2573747230553656, "learning_rate": 1.3175394062449777e-05, "loss": 0.9781, "step": 2158 }, { "epoch": 0.4153720359771055, "grad_norm": 1.4105580622434062, "learning_rate": 1.3169483904894185e-05, "loss": 1.0081, "step": 2159 }, { "epoch": 0.4155644269154923, "grad_norm": 1.5784582544627455, "learning_rate": 1.3163572516347565e-05, "loss": 1.086, "step": 2160 }, { "epoch": 0.4157568178538791, "grad_norm": 1.27205885945868, "learning_rate": 1.3157659899105835e-05, "loss": 0.9383, "step": 2161 }, { "epoch": 0.4159492087922659, "grad_norm": 1.4783811756194272, "learning_rate": 1.315174605546538e-05, "loss": 1.0188, "step": 2162 }, { "epoch": 0.4161415997306527, "grad_norm": 1.5146439146205017, "learning_rate": 1.3145830987723081e-05, "loss": 1.0084, "step": 2163 }, { "epoch": 0.4163339906690395, "grad_norm": 1.4076715626461342, "learning_rate": 1.3139914698176273e-05, "loss": 0.9738, "step": 2164 }, { "epoch": 0.4165263816074263, "grad_norm": 0.7887389680003992, "learning_rate": 1.3133997189122777e-05, "loss": 0.859, "step": 2165 }, { "epoch": 0.4167187725458131, "grad_norm": 1.4301037566680705, "learning_rate": 1.3128078462860887e-05, "loss": 0.9818, "step": 2166 }, { "epoch": 0.4169111634841999, "grad_norm": 1.44627339505306, "learning_rate": 1.3122158521689367e-05, "loss": 1.0029, "step": 2167 }, { "epoch": 0.4171035544225867, "grad_norm": 1.4927584108206593, "learning_rate": 1.3116237367907454e-05, "loss": 0.9541, "step": 2168 }, { "epoch": 0.4172959453609735, "grad_norm": 1.441925289151065, "learning_rate": 1.3110315003814855e-05, "loss": 1.0469, "step": 2169 }, { "epoch": 0.4174883362993603, "grad_norm": 1.3731147420870498, "learning_rate": 1.3104391431711748e-05, "loss": 1.0702, "step": 2170 }, { "epoch": 0.41768072723774713, "grad_norm": 1.514947919935472, "learning_rate": 1.309846665389878e-05, "loss": 0.9822, "step": 2171 }, { "epoch": 0.41787311817613393, "grad_norm": 1.299088552167726, "learning_rate": 1.309254067267707e-05, "loss": 0.9647, "step": 2172 }, { "epoch": 0.41806550911452073, "grad_norm": 1.3515943041417693, "learning_rate": 1.3086613490348198e-05, "loss": 0.973, "step": 2173 }, { "epoch": 0.4182579000529075, "grad_norm": 1.3603322618696791, "learning_rate": 1.3080685109214208e-05, "loss": 1.003, "step": 2174 }, { "epoch": 0.4184502909912943, "grad_norm": 1.3941118840586137, "learning_rate": 1.3074755531577628e-05, "loss": 1.0348, "step": 2175 }, { "epoch": 0.4186426819296811, "grad_norm": 1.4661659626468793, "learning_rate": 1.3068824759741428e-05, "loss": 0.9397, "step": 2176 }, { "epoch": 0.4188350728680679, "grad_norm": 1.4157095171524527, "learning_rate": 1.306289279600905e-05, "loss": 1.0641, "step": 2177 }, { "epoch": 0.4190274638064547, "grad_norm": 1.4842639228097347, "learning_rate": 1.3056959642684404e-05, "loss": 1.0063, "step": 2178 }, { "epoch": 0.4192198547448415, "grad_norm": 1.4715611632285774, "learning_rate": 1.305102530207186e-05, "loss": 1.0232, "step": 2179 }, { "epoch": 0.4194122456832283, "grad_norm": 1.3368743494531314, "learning_rate": 1.3045089776476246e-05, "loss": 1.0403, "step": 2180 }, { "epoch": 0.4196046366216151, "grad_norm": 1.3420181507116393, "learning_rate": 1.3039153068202853e-05, "loss": 0.9214, "step": 2181 }, { "epoch": 0.4197970275600019, "grad_norm": 1.2719028201246119, "learning_rate": 1.3033215179557424e-05, "loss": 1.034, "step": 2182 }, { "epoch": 0.4199894184983887, "grad_norm": 1.3634199452484577, "learning_rate": 1.3027276112846172e-05, "loss": 1.0208, "step": 2183 }, { "epoch": 0.4201818094367755, "grad_norm": 1.3948979961906123, "learning_rate": 1.3021335870375763e-05, "loss": 0.9342, "step": 2184 }, { "epoch": 0.4203742003751623, "grad_norm": 1.3474001703127778, "learning_rate": 1.3015394454453316e-05, "loss": 0.9844, "step": 2185 }, { "epoch": 0.4205665913135491, "grad_norm": 1.5023582279483854, "learning_rate": 1.3009451867386411e-05, "loss": 1.0398, "step": 2186 }, { "epoch": 0.4207589822519359, "grad_norm": 1.3841858596400847, "learning_rate": 1.3003508111483077e-05, "loss": 1.0314, "step": 2187 }, { "epoch": 0.4209513731903227, "grad_norm": 1.3135797729460852, "learning_rate": 1.29975631890518e-05, "loss": 1.0501, "step": 2188 }, { "epoch": 0.4211437641287095, "grad_norm": 1.4037812085451693, "learning_rate": 1.2991617102401524e-05, "loss": 1.042, "step": 2189 }, { "epoch": 0.4213361550670963, "grad_norm": 0.7165235234302098, "learning_rate": 1.2985669853841635e-05, "loss": 0.7955, "step": 2190 }, { "epoch": 0.4215285460054831, "grad_norm": 1.297666537349549, "learning_rate": 1.297972144568198e-05, "loss": 0.9998, "step": 2191 }, { "epoch": 0.4217209369438699, "grad_norm": 1.366709274652552, "learning_rate": 1.2973771880232853e-05, "loss": 1.0148, "step": 2192 }, { "epoch": 0.4219133278822568, "grad_norm": 1.3437015780905668, "learning_rate": 1.2967821159804994e-05, "loss": 1.0086, "step": 2193 }, { "epoch": 0.42210571882064357, "grad_norm": 1.2585855313649763, "learning_rate": 1.2961869286709594e-05, "loss": 1.0023, "step": 2194 }, { "epoch": 0.42229810975903037, "grad_norm": 1.24852729745958, "learning_rate": 1.295591626325829e-05, "loss": 0.9948, "step": 2195 }, { "epoch": 0.42249050069741717, "grad_norm": 1.3785717527011758, "learning_rate": 1.2949962091763174e-05, "loss": 1.0112, "step": 2196 }, { "epoch": 0.42268289163580397, "grad_norm": 1.4598108519932844, "learning_rate": 1.2944006774536773e-05, "loss": 0.9635, "step": 2197 }, { "epoch": 0.42287528257419077, "grad_norm": 1.390394187873638, "learning_rate": 1.2938050313892062e-05, "loss": 1.0004, "step": 2198 }, { "epoch": 0.42306767351257757, "grad_norm": 1.3538139812157148, "learning_rate": 1.2932092712142468e-05, "loss": 1.0151, "step": 2199 }, { "epoch": 0.42326006445096437, "grad_norm": 1.3432520559930676, "learning_rate": 1.292613397160185e-05, "loss": 0.914, "step": 2200 }, { "epoch": 0.42345245538935117, "grad_norm": 1.3287764502316006, "learning_rate": 1.2920174094584514e-05, "loss": 1.0055, "step": 2201 }, { "epoch": 0.42364484632773797, "grad_norm": 1.3775434711288281, "learning_rate": 1.2914213083405211e-05, "loss": 0.9629, "step": 2202 }, { "epoch": 0.42383723726612477, "grad_norm": 1.3804874130755376, "learning_rate": 1.2908250940379124e-05, "loss": 1.0638, "step": 2203 }, { "epoch": 0.42402962820451157, "grad_norm": 1.3581662376160641, "learning_rate": 1.2902287667821885e-05, "loss": 0.9074, "step": 2204 }, { "epoch": 0.42422201914289837, "grad_norm": 1.5196665640819949, "learning_rate": 1.289632326804956e-05, "loss": 1.0317, "step": 2205 }, { "epoch": 0.42441441008128516, "grad_norm": 1.479895595373699, "learning_rate": 1.2890357743378649e-05, "loss": 0.9503, "step": 2206 }, { "epoch": 0.42460680101967196, "grad_norm": 1.3720764963183212, "learning_rate": 1.2884391096126098e-05, "loss": 0.9826, "step": 2207 }, { "epoch": 0.42479919195805876, "grad_norm": 1.417390187365525, "learning_rate": 1.2878423328609281e-05, "loss": 0.9917, "step": 2208 }, { "epoch": 0.42499158289644556, "grad_norm": 1.4061486247254062, "learning_rate": 1.2872454443146015e-05, "loss": 1.055, "step": 2209 }, { "epoch": 0.42518397383483236, "grad_norm": 1.4707056694496023, "learning_rate": 1.286648444205454e-05, "loss": 1.0145, "step": 2210 }, { "epoch": 0.42537636477321916, "grad_norm": 1.413592309008556, "learning_rate": 1.2860513327653537e-05, "loss": 0.9715, "step": 2211 }, { "epoch": 0.42556875571160596, "grad_norm": 1.4156047917358188, "learning_rate": 1.2854541102262119e-05, "loss": 1.0013, "step": 2212 }, { "epoch": 0.42576114664999276, "grad_norm": 1.4059270896282012, "learning_rate": 1.284856776819983e-05, "loss": 1.0284, "step": 2213 }, { "epoch": 0.42595353758837956, "grad_norm": 1.3679099135313706, "learning_rate": 1.2842593327786649e-05, "loss": 1.0322, "step": 2214 }, { "epoch": 0.4261459285267664, "grad_norm": 1.348818884019458, "learning_rate": 1.2836617783342968e-05, "loss": 0.9948, "step": 2215 }, { "epoch": 0.4263383194651532, "grad_norm": 1.3083577469597105, "learning_rate": 1.2830641137189628e-05, "loss": 0.9938, "step": 2216 }, { "epoch": 0.42653071040354, "grad_norm": 1.249723330993427, "learning_rate": 1.282466339164789e-05, "loss": 0.9927, "step": 2217 }, { "epoch": 0.4267231013419268, "grad_norm": 1.4503693160181206, "learning_rate": 1.2818684549039437e-05, "loss": 1.0116, "step": 2218 }, { "epoch": 0.4269154922803136, "grad_norm": 1.5361661965850137, "learning_rate": 1.2812704611686386e-05, "loss": 1.0109, "step": 2219 }, { "epoch": 0.4271078832187004, "grad_norm": 1.4543007318358319, "learning_rate": 1.2806723581911274e-05, "loss": 1.0501, "step": 2220 }, { "epoch": 0.4273002741570872, "grad_norm": 1.4040450039762002, "learning_rate": 1.2800741462037065e-05, "loss": 1.0029, "step": 2221 }, { "epoch": 0.427492665095474, "grad_norm": 1.4047319688732316, "learning_rate": 1.2794758254387147e-05, "loss": 0.9499, "step": 2222 }, { "epoch": 0.4276850560338608, "grad_norm": 1.3620124657796135, "learning_rate": 1.2788773961285323e-05, "loss": 1.0079, "step": 2223 }, { "epoch": 0.4278774469722476, "grad_norm": 1.4370214063543927, "learning_rate": 1.2782788585055829e-05, "loss": 0.949, "step": 2224 }, { "epoch": 0.4280698379106344, "grad_norm": 1.521778129823936, "learning_rate": 1.2776802128023317e-05, "loss": 1.0012, "step": 2225 }, { "epoch": 0.4282622288490212, "grad_norm": 1.4621707652822207, "learning_rate": 1.2770814592512853e-05, "loss": 1.0187, "step": 2226 }, { "epoch": 0.428454619787408, "grad_norm": 1.4800448317806891, "learning_rate": 1.2764825980849931e-05, "loss": 1.0711, "step": 2227 }, { "epoch": 0.4286470107257948, "grad_norm": 1.427121500468448, "learning_rate": 1.2758836295360455e-05, "loss": 1.0008, "step": 2228 }, { "epoch": 0.4288394016641816, "grad_norm": 1.4059561093050221, "learning_rate": 1.2752845538370752e-05, "loss": 1.0553, "step": 2229 }, { "epoch": 0.4290317926025684, "grad_norm": 1.3282429055226717, "learning_rate": 1.2746853712207567e-05, "loss": 0.9964, "step": 2230 }, { "epoch": 0.4292241835409552, "grad_norm": 1.482903203833618, "learning_rate": 1.274086081919805e-05, "loss": 0.9429, "step": 2231 }, { "epoch": 0.429416574479342, "grad_norm": 1.4567975964417137, "learning_rate": 1.273486686166977e-05, "loss": 0.988, "step": 2232 }, { "epoch": 0.4296089654177288, "grad_norm": 0.759145146498389, "learning_rate": 1.2728871841950719e-05, "loss": 0.8583, "step": 2233 }, { "epoch": 0.4298013563561156, "grad_norm": 0.7541094475591062, "learning_rate": 1.2722875762369288e-05, "loss": 0.8714, "step": 2234 }, { "epoch": 0.4299937472945024, "grad_norm": 1.4323925883249442, "learning_rate": 1.2716878625254287e-05, "loss": 1.0428, "step": 2235 }, { "epoch": 0.4301861382328892, "grad_norm": 1.3501238047545396, "learning_rate": 1.2710880432934934e-05, "loss": 0.8631, "step": 2236 }, { "epoch": 0.43037852917127606, "grad_norm": 1.4936580265996149, "learning_rate": 1.270488118774086e-05, "loss": 1.0183, "step": 2237 }, { "epoch": 0.43057092010966286, "grad_norm": 1.2991587716530237, "learning_rate": 1.26988808920021e-05, "loss": 0.9798, "step": 2238 }, { "epoch": 0.43076331104804966, "grad_norm": 1.4066495232438307, "learning_rate": 1.26928795480491e-05, "loss": 1.0025, "step": 2239 }, { "epoch": 0.43095570198643646, "grad_norm": 1.4655446632694282, "learning_rate": 1.2686877158212715e-05, "loss": 0.884, "step": 2240 }, { "epoch": 0.43114809292482326, "grad_norm": 1.3876793620917263, "learning_rate": 1.26808737248242e-05, "loss": 0.926, "step": 2241 }, { "epoch": 0.43134048386321006, "grad_norm": 1.3026675628387328, "learning_rate": 1.2674869250215225e-05, "loss": 0.9413, "step": 2242 }, { "epoch": 0.43153287480159686, "grad_norm": 1.3673727871157544, "learning_rate": 1.2668863736717855e-05, "loss": 0.9925, "step": 2243 }, { "epoch": 0.43172526573998365, "grad_norm": 0.7601880718574062, "learning_rate": 1.2662857186664558e-05, "loss": 0.8941, "step": 2244 }, { "epoch": 0.43191765667837045, "grad_norm": 0.7749037249806784, "learning_rate": 1.2656849602388222e-05, "loss": 0.9063, "step": 2245 }, { "epoch": 0.43211004761675725, "grad_norm": 1.5486321909231258, "learning_rate": 1.2650840986222111e-05, "loss": 0.9771, "step": 2246 }, { "epoch": 0.43230243855514405, "grad_norm": 1.4846296908999184, "learning_rate": 1.2644831340499906e-05, "loss": 1.0147, "step": 2247 }, { "epoch": 0.43249482949353085, "grad_norm": 1.5431119996833196, "learning_rate": 1.2638820667555685e-05, "loss": 0.955, "step": 2248 }, { "epoch": 0.43268722043191765, "grad_norm": 1.517451104790015, "learning_rate": 1.2632808969723927e-05, "loss": 1.0376, "step": 2249 }, { "epoch": 0.43287961137030445, "grad_norm": 1.478201314517622, "learning_rate": 1.26267962493395e-05, "loss": 1.105, "step": 2250 }, { "epoch": 0.43307200230869125, "grad_norm": 0.6853982849309757, "learning_rate": 1.2620782508737678e-05, "loss": 0.7938, "step": 2251 }, { "epoch": 0.43326439324707805, "grad_norm": 1.3061580191405182, "learning_rate": 1.2614767750254129e-05, "loss": 0.9404, "step": 2252 }, { "epoch": 0.43345678418546485, "grad_norm": 1.5225358565231404, "learning_rate": 1.2608751976224916e-05, "loss": 1.0399, "step": 2253 }, { "epoch": 0.43364917512385165, "grad_norm": 1.4251961085721188, "learning_rate": 1.2602735188986498e-05, "loss": 1.0478, "step": 2254 }, { "epoch": 0.43384156606223845, "grad_norm": 1.3710189537906239, "learning_rate": 1.2596717390875721e-05, "loss": 1.0579, "step": 2255 }, { "epoch": 0.43403395700062525, "grad_norm": 1.363954160414454, "learning_rate": 1.2590698584229834e-05, "loss": 0.9348, "step": 2256 }, { "epoch": 0.43422634793901205, "grad_norm": 1.398528987629981, "learning_rate": 1.2584678771386467e-05, "loss": 1.0632, "step": 2257 }, { "epoch": 0.4344187388773989, "grad_norm": 1.4821995100824872, "learning_rate": 1.2578657954683651e-05, "loss": 0.9363, "step": 2258 }, { "epoch": 0.4346111298157857, "grad_norm": 1.303068458605695, "learning_rate": 1.2572636136459799e-05, "loss": 0.975, "step": 2259 }, { "epoch": 0.4348035207541725, "grad_norm": 1.4059515893469465, "learning_rate": 1.2566613319053713e-05, "loss": 0.9758, "step": 2260 }, { "epoch": 0.4349959116925593, "grad_norm": 1.5023804197390584, "learning_rate": 1.2560589504804592e-05, "loss": 1.0185, "step": 2261 }, { "epoch": 0.4351883026309461, "grad_norm": 0.7735895190874491, "learning_rate": 1.2554564696052011e-05, "loss": 0.917, "step": 2262 }, { "epoch": 0.4353806935693329, "grad_norm": 1.628310558961417, "learning_rate": 1.2548538895135942e-05, "loss": 0.9458, "step": 2263 }, { "epoch": 0.4355730845077197, "grad_norm": 1.516514706930992, "learning_rate": 1.254251210439673e-05, "loss": 0.9613, "step": 2264 }, { "epoch": 0.4357654754461065, "grad_norm": 1.3820139654039554, "learning_rate": 1.2536484326175114e-05, "loss": 1.0059, "step": 2265 }, { "epoch": 0.4359578663844933, "grad_norm": 1.3745051695118284, "learning_rate": 1.2530455562812214e-05, "loss": 1.0613, "step": 2266 }, { "epoch": 0.4361502573228801, "grad_norm": 1.3812574266406714, "learning_rate": 1.252442581664953e-05, "loss": 1.0398, "step": 2267 }, { "epoch": 0.4363426482612669, "grad_norm": 1.5395856963603651, "learning_rate": 1.2518395090028952e-05, "loss": 0.988, "step": 2268 }, { "epoch": 0.4365350391996537, "grad_norm": 1.346481244260342, "learning_rate": 1.2512363385292739e-05, "loss": 0.9583, "step": 2269 }, { "epoch": 0.4367274301380405, "grad_norm": 1.3766716404028279, "learning_rate": 1.2506330704783533e-05, "loss": 0.9508, "step": 2270 }, { "epoch": 0.4369198210764273, "grad_norm": 0.7360801502281307, "learning_rate": 1.2500297050844367e-05, "loss": 0.8685, "step": 2271 }, { "epoch": 0.4371122120148141, "grad_norm": 1.2793074854971802, "learning_rate": 1.2494262425818637e-05, "loss": 0.989, "step": 2272 }, { "epoch": 0.4373046029532009, "grad_norm": 1.5001791694864492, "learning_rate": 1.2488226832050116e-05, "loss": 1.0005, "step": 2273 }, { "epoch": 0.4374969938915877, "grad_norm": 1.3391370605127872, "learning_rate": 1.2482190271882973e-05, "loss": 0.883, "step": 2274 }, { "epoch": 0.4376893848299745, "grad_norm": 1.4205270295614414, "learning_rate": 1.2476152747661727e-05, "loss": 1.0365, "step": 2275 }, { "epoch": 0.4378817757683613, "grad_norm": 1.4228808560346147, "learning_rate": 1.2470114261731288e-05, "loss": 0.9953, "step": 2276 }, { "epoch": 0.4380741667067481, "grad_norm": 1.4770859064642525, "learning_rate": 1.246407481643693e-05, "loss": 0.8982, "step": 2277 }, { "epoch": 0.4382665576451349, "grad_norm": 1.3958642307170261, "learning_rate": 1.245803441412431e-05, "loss": 0.9805, "step": 2278 }, { "epoch": 0.4384589485835217, "grad_norm": 1.4406943605396905, "learning_rate": 1.2451993057139445e-05, "loss": 1.0008, "step": 2279 }, { "epoch": 0.43865133952190855, "grad_norm": 1.4897841134133432, "learning_rate": 1.2445950747828732e-05, "loss": 1.1044, "step": 2280 }, { "epoch": 0.43884373046029534, "grad_norm": 1.3975576249231274, "learning_rate": 1.2439907488538934e-05, "loss": 0.9145, "step": 2281 }, { "epoch": 0.43903612139868214, "grad_norm": 1.382209996787316, "learning_rate": 1.243386328161718e-05, "loss": 1.063, "step": 2282 }, { "epoch": 0.43922851233706894, "grad_norm": 1.2744291921357072, "learning_rate": 1.2427818129410975e-05, "loss": 0.9527, "step": 2283 }, { "epoch": 0.43942090327545574, "grad_norm": 1.5548763202596705, "learning_rate": 1.2421772034268187e-05, "loss": 1.058, "step": 2284 }, { "epoch": 0.43961329421384254, "grad_norm": 1.4711888084637261, "learning_rate": 1.2415724998537042e-05, "loss": 0.962, "step": 2285 }, { "epoch": 0.43980568515222934, "grad_norm": 1.4842219339483904, "learning_rate": 1.2409677024566145e-05, "loss": 0.9614, "step": 2286 }, { "epoch": 0.43999807609061614, "grad_norm": 1.3713664380615207, "learning_rate": 1.240362811470446e-05, "loss": 1.0071, "step": 2287 }, { "epoch": 0.44019046702900294, "grad_norm": 1.2249311281262734, "learning_rate": 1.2397578271301312e-05, "loss": 0.9897, "step": 2288 }, { "epoch": 0.44038285796738974, "grad_norm": 1.2424987533424643, "learning_rate": 1.2391527496706389e-05, "loss": 1.0231, "step": 2289 }, { "epoch": 0.44057524890577654, "grad_norm": 1.4054518501690811, "learning_rate": 1.2385475793269744e-05, "loss": 0.9441, "step": 2290 }, { "epoch": 0.44076763984416334, "grad_norm": 1.3669473336484408, "learning_rate": 1.2379423163341791e-05, "loss": 0.9629, "step": 2291 }, { "epoch": 0.44096003078255014, "grad_norm": 1.3902460005338455, "learning_rate": 1.2373369609273299e-05, "loss": 0.9951, "step": 2292 }, { "epoch": 0.44115242172093694, "grad_norm": 1.2751419300562163, "learning_rate": 1.2367315133415396e-05, "loss": 1.0283, "step": 2293 }, { "epoch": 0.44134481265932374, "grad_norm": 1.4162464933222094, "learning_rate": 1.2361259738119575e-05, "loss": 1.0876, "step": 2294 }, { "epoch": 0.44153720359771054, "grad_norm": 0.7591249546417853, "learning_rate": 1.2355203425737683e-05, "loss": 0.8943, "step": 2295 }, { "epoch": 0.44172959453609734, "grad_norm": 1.3275829893343256, "learning_rate": 1.2349146198621917e-05, "loss": 0.9688, "step": 2296 }, { "epoch": 0.44192198547448414, "grad_norm": 1.4157422831434103, "learning_rate": 1.2343088059124839e-05, "loss": 0.9539, "step": 2297 }, { "epoch": 0.44211437641287094, "grad_norm": 1.4621663692704245, "learning_rate": 1.2337029009599357e-05, "loss": 1.0483, "step": 2298 }, { "epoch": 0.44230676735125773, "grad_norm": 1.4624860307190704, "learning_rate": 1.2330969052398735e-05, "loss": 1.027, "step": 2299 }, { "epoch": 0.44249915828964453, "grad_norm": 1.378676717196178, "learning_rate": 1.2324908189876597e-05, "loss": 0.9545, "step": 2300 }, { "epoch": 0.44269154922803133, "grad_norm": 1.3407839116120306, "learning_rate": 1.2318846424386907e-05, "loss": 1.0207, "step": 2301 }, { "epoch": 0.4428839401664182, "grad_norm": 1.3504564808234545, "learning_rate": 1.2312783758283981e-05, "loss": 0.9684, "step": 2302 }, { "epoch": 0.443076331104805, "grad_norm": 1.3395870633127014, "learning_rate": 1.23067201939225e-05, "loss": 0.9678, "step": 2303 }, { "epoch": 0.4432687220431918, "grad_norm": 1.3588613423274867, "learning_rate": 1.2300655733657475e-05, "loss": 1.0116, "step": 2304 }, { "epoch": 0.4434611129815786, "grad_norm": 1.4395193453018664, "learning_rate": 1.2294590379844268e-05, "loss": 1.0275, "step": 2305 }, { "epoch": 0.4436535039199654, "grad_norm": 0.7312811478597813, "learning_rate": 1.2288524134838602e-05, "loss": 0.8754, "step": 2306 }, { "epoch": 0.4438458948583522, "grad_norm": 1.4789087486040824, "learning_rate": 1.2282457000996533e-05, "loss": 1.0225, "step": 2307 }, { "epoch": 0.444038285796739, "grad_norm": 0.7091799688018018, "learning_rate": 1.2276388980674465e-05, "loss": 0.8385, "step": 2308 }, { "epoch": 0.4442306767351258, "grad_norm": 1.36289172700742, "learning_rate": 1.227032007622915e-05, "loss": 0.888, "step": 2309 }, { "epoch": 0.4444230676735126, "grad_norm": 1.3578404614108328, "learning_rate": 1.2264250290017675e-05, "loss": 0.9892, "step": 2310 }, { "epoch": 0.4446154586118994, "grad_norm": 1.3802254062287913, "learning_rate": 1.2258179624397477e-05, "loss": 0.9348, "step": 2311 }, { "epoch": 0.4448078495502862, "grad_norm": 1.3193665831147328, "learning_rate": 1.2252108081726337e-05, "loss": 1.0223, "step": 2312 }, { "epoch": 0.445000240488673, "grad_norm": 1.4470738374870333, "learning_rate": 1.224603566436237e-05, "loss": 1.0221, "step": 2313 }, { "epoch": 0.4451926314270598, "grad_norm": 1.2616929841372757, "learning_rate": 1.2239962374664029e-05, "loss": 0.958, "step": 2314 }, { "epoch": 0.4453850223654466, "grad_norm": 1.5234047674687128, "learning_rate": 1.2233888214990113e-05, "loss": 1.0514, "step": 2315 }, { "epoch": 0.4455774133038334, "grad_norm": 0.8042343090118831, "learning_rate": 1.2227813187699757e-05, "loss": 0.8623, "step": 2316 }, { "epoch": 0.4457698042422202, "grad_norm": 1.554541140174422, "learning_rate": 1.222173729515243e-05, "loss": 1.0555, "step": 2317 }, { "epoch": 0.445962195180607, "grad_norm": 1.2578392568692036, "learning_rate": 1.2215660539707936e-05, "loss": 0.9864, "step": 2318 }, { "epoch": 0.4461545861189938, "grad_norm": 1.5632527454898808, "learning_rate": 1.2209582923726424e-05, "loss": 1.0347, "step": 2319 }, { "epoch": 0.4463469770573806, "grad_norm": 1.3683301781165225, "learning_rate": 1.2203504449568361e-05, "loss": 1.0141, "step": 2320 }, { "epoch": 0.4465393679957674, "grad_norm": 1.2010861359215237, "learning_rate": 1.2197425119594563e-05, "loss": 1.0019, "step": 2321 }, { "epoch": 0.4467317589341542, "grad_norm": 1.2813361106919208, "learning_rate": 1.219134493616617e-05, "loss": 0.9232, "step": 2322 }, { "epoch": 0.446924149872541, "grad_norm": 1.2691497415442254, "learning_rate": 1.2185263901644653e-05, "loss": 0.9664, "step": 2323 }, { "epoch": 0.44711654081092783, "grad_norm": 1.5151860660767114, "learning_rate": 1.217918201839182e-05, "loss": 1.0626, "step": 2324 }, { "epoch": 0.44730893174931463, "grad_norm": 1.3553346755202504, "learning_rate": 1.2173099288769799e-05, "loss": 1.0497, "step": 2325 }, { "epoch": 0.44750132268770143, "grad_norm": 1.3720504242149771, "learning_rate": 1.2167015715141057e-05, "loss": 0.9885, "step": 2326 }, { "epoch": 0.44769371362608823, "grad_norm": 1.485800163056368, "learning_rate": 1.216093129986838e-05, "loss": 1.0641, "step": 2327 }, { "epoch": 0.44788610456447503, "grad_norm": 1.4153989336925947, "learning_rate": 1.2154846045314885e-05, "loss": 0.9511, "step": 2328 }, { "epoch": 0.44807849550286183, "grad_norm": 1.444776151137511, "learning_rate": 1.214875995384402e-05, "loss": 0.9743, "step": 2329 }, { "epoch": 0.44827088644124863, "grad_norm": 1.4267463174612918, "learning_rate": 1.214267302781955e-05, "loss": 0.9559, "step": 2330 }, { "epoch": 0.4484632773796354, "grad_norm": 1.3599538988376956, "learning_rate": 1.2136585269605558e-05, "loss": 0.9871, "step": 2331 }, { "epoch": 0.4486556683180222, "grad_norm": 1.3123652705250402, "learning_rate": 1.2130496681566475e-05, "loss": 0.9878, "step": 2332 }, { "epoch": 0.448848059256409, "grad_norm": 1.3579479800407614, "learning_rate": 1.212440726606703e-05, "loss": 1.0276, "step": 2333 }, { "epoch": 0.4490404501947958, "grad_norm": 1.382497979871949, "learning_rate": 1.211831702547228e-05, "loss": 0.9611, "step": 2334 }, { "epoch": 0.4492328411331826, "grad_norm": 0.7274173427029371, "learning_rate": 1.2112225962147605e-05, "loss": 0.8199, "step": 2335 }, { "epoch": 0.4494252320715694, "grad_norm": 0.7358841636223049, "learning_rate": 1.210613407845871e-05, "loss": 0.876, "step": 2336 }, { "epoch": 0.4496176230099562, "grad_norm": 1.5669428789325484, "learning_rate": 1.2100041376771605e-05, "loss": 1.0282, "step": 2337 }, { "epoch": 0.449810013948343, "grad_norm": 1.4502012623171572, "learning_rate": 1.209394785945263e-05, "loss": 1.0008, "step": 2338 }, { "epoch": 0.4500024048867298, "grad_norm": 1.3357978805609414, "learning_rate": 1.2087853528868432e-05, "loss": 0.91, "step": 2339 }, { "epoch": 0.4501947958251166, "grad_norm": 1.3797938295689651, "learning_rate": 1.2081758387385982e-05, "loss": 1.04, "step": 2340 }, { "epoch": 0.4503871867635034, "grad_norm": 1.2949687369455458, "learning_rate": 1.2075662437372567e-05, "loss": 0.9833, "step": 2341 }, { "epoch": 0.4505795777018902, "grad_norm": 1.399390449028263, "learning_rate": 1.2069565681195776e-05, "loss": 0.9843, "step": 2342 }, { "epoch": 0.450771968640277, "grad_norm": 1.4237184780370566, "learning_rate": 1.206346812122352e-05, "loss": 1.0056, "step": 2343 }, { "epoch": 0.4509643595786638, "grad_norm": 1.4425804029478204, "learning_rate": 1.2057369759824025e-05, "loss": 1.0323, "step": 2344 }, { "epoch": 0.4511567505170506, "grad_norm": 1.284126229272528, "learning_rate": 1.2051270599365825e-05, "loss": 0.9428, "step": 2345 }, { "epoch": 0.4513491414554375, "grad_norm": 1.3594011445860634, "learning_rate": 1.2045170642217756e-05, "loss": 0.9431, "step": 2346 }, { "epoch": 0.4515415323938243, "grad_norm": 1.3770235448804862, "learning_rate": 1.2039069890748978e-05, "loss": 0.9797, "step": 2347 }, { "epoch": 0.4517339233322111, "grad_norm": 1.4344705598680594, "learning_rate": 1.2032968347328952e-05, "loss": 0.9614, "step": 2348 }, { "epoch": 0.4519263142705979, "grad_norm": 1.3172141742828571, "learning_rate": 1.2026866014327446e-05, "loss": 0.9643, "step": 2349 }, { "epoch": 0.4521187052089847, "grad_norm": 1.4699420750286165, "learning_rate": 1.2020762894114535e-05, "loss": 1.0656, "step": 2350 }, { "epoch": 0.45231109614737147, "grad_norm": 1.428785789197767, "learning_rate": 1.20146589890606e-05, "loss": 0.9494, "step": 2351 }, { "epoch": 0.45250348708575827, "grad_norm": 1.6101224270866428, "learning_rate": 1.2008554301536328e-05, "loss": 0.9968, "step": 2352 }, { "epoch": 0.45269587802414507, "grad_norm": 1.3976608492459717, "learning_rate": 1.2002448833912712e-05, "loss": 0.9449, "step": 2353 }, { "epoch": 0.45288826896253187, "grad_norm": 1.534351434558593, "learning_rate": 1.1996342588561042e-05, "loss": 1.0998, "step": 2354 }, { "epoch": 0.45308065990091867, "grad_norm": 1.3590023365727402, "learning_rate": 1.1990235567852917e-05, "loss": 1.0062, "step": 2355 }, { "epoch": 0.45327305083930547, "grad_norm": 1.343304892733799, "learning_rate": 1.1984127774160226e-05, "loss": 0.942, "step": 2356 }, { "epoch": 0.45346544177769227, "grad_norm": 1.3587548319322005, "learning_rate": 1.1978019209855174e-05, "loss": 0.9399, "step": 2357 }, { "epoch": 0.45365783271607907, "grad_norm": 1.5181526764532138, "learning_rate": 1.1971909877310253e-05, "loss": 0.9425, "step": 2358 }, { "epoch": 0.45385022365446587, "grad_norm": 0.7149217972360561, "learning_rate": 1.1965799778898258e-05, "loss": 0.8473, "step": 2359 }, { "epoch": 0.45404261459285267, "grad_norm": 1.468198914080035, "learning_rate": 1.1959688916992279e-05, "loss": 1.0873, "step": 2360 }, { "epoch": 0.45423500553123947, "grad_norm": 1.373645691048379, "learning_rate": 1.1953577293965707e-05, "loss": 1.0549, "step": 2361 }, { "epoch": 0.45442739646962627, "grad_norm": 1.4652796538920312, "learning_rate": 1.1947464912192228e-05, "loss": 0.9041, "step": 2362 }, { "epoch": 0.45461978740801307, "grad_norm": 1.5076353374569413, "learning_rate": 1.1941351774045815e-05, "loss": 0.9985, "step": 2363 }, { "epoch": 0.45481217834639986, "grad_norm": 1.381274151296404, "learning_rate": 1.1935237881900743e-05, "loss": 0.9933, "step": 2364 }, { "epoch": 0.45500456928478666, "grad_norm": 1.5418852411297483, "learning_rate": 1.1929123238131579e-05, "loss": 1.0646, "step": 2365 }, { "epoch": 0.45519696022317346, "grad_norm": 1.3514972779872392, "learning_rate": 1.1923007845113178e-05, "loss": 1.0324, "step": 2366 }, { "epoch": 0.45538935116156026, "grad_norm": 1.4386811447768753, "learning_rate": 1.1916891705220689e-05, "loss": 1.0366, "step": 2367 }, { "epoch": 0.4555817420999471, "grad_norm": 1.3201117233612591, "learning_rate": 1.191077482082955e-05, "loss": 1.0117, "step": 2368 }, { "epoch": 0.4557741330383339, "grad_norm": 1.324936682340966, "learning_rate": 1.1904657194315486e-05, "loss": 1.0557, "step": 2369 }, { "epoch": 0.4559665239767207, "grad_norm": 1.3973832873366763, "learning_rate": 1.1898538828054517e-05, "loss": 0.9792, "step": 2370 }, { "epoch": 0.4561589149151075, "grad_norm": 1.4021492853125817, "learning_rate": 1.1892419724422946e-05, "loss": 0.9114, "step": 2371 }, { "epoch": 0.4563513058534943, "grad_norm": 1.2482740682201539, "learning_rate": 1.1886299885797357e-05, "loss": 0.9357, "step": 2372 }, { "epoch": 0.4565436967918811, "grad_norm": 1.411078806635782, "learning_rate": 1.1880179314554629e-05, "loss": 0.9758, "step": 2373 }, { "epoch": 0.4567360877302679, "grad_norm": 1.3765073425624152, "learning_rate": 1.1874058013071923e-05, "loss": 0.9774, "step": 2374 }, { "epoch": 0.4569284786686547, "grad_norm": 1.4369818763376327, "learning_rate": 1.1867935983726676e-05, "loss": 1.0181, "step": 2375 }, { "epoch": 0.4571208696070415, "grad_norm": 1.5135251305864164, "learning_rate": 1.186181322889662e-05, "loss": 1.0398, "step": 2376 }, { "epoch": 0.4573132605454283, "grad_norm": 1.4710746895984261, "learning_rate": 1.1855689750959759e-05, "loss": 1.0307, "step": 2377 }, { "epoch": 0.4575056514838151, "grad_norm": 1.3130441856307797, "learning_rate": 1.1849565552294379e-05, "loss": 1.0441, "step": 2378 }, { "epoch": 0.4576980424222019, "grad_norm": 1.3570631341052402, "learning_rate": 1.1843440635279056e-05, "loss": 0.9911, "step": 2379 }, { "epoch": 0.4578904333605887, "grad_norm": 1.463806242720431, "learning_rate": 1.1837315002292629e-05, "loss": 0.9987, "step": 2380 }, { "epoch": 0.4580828242989755, "grad_norm": 1.3252700509830937, "learning_rate": 1.1831188655714225e-05, "loss": 1.0239, "step": 2381 }, { "epoch": 0.4582752152373623, "grad_norm": 1.4064252138215398, "learning_rate": 1.182506159792325e-05, "loss": 0.9696, "step": 2382 }, { "epoch": 0.4584676061757491, "grad_norm": 1.5457378256488286, "learning_rate": 1.1818933831299381e-05, "loss": 1.0438, "step": 2383 }, { "epoch": 0.4586599971141359, "grad_norm": 1.320859956232017, "learning_rate": 1.1812805358222571e-05, "loss": 0.9309, "step": 2384 }, { "epoch": 0.4588523880525227, "grad_norm": 1.3666868283281959, "learning_rate": 1.180667618107305e-05, "loss": 0.93, "step": 2385 }, { "epoch": 0.4590447789909095, "grad_norm": 1.504135340920485, "learning_rate": 1.1800546302231317e-05, "loss": 1.0547, "step": 2386 }, { "epoch": 0.4592371699292963, "grad_norm": 1.5222182504287312, "learning_rate": 1.1794415724078147e-05, "loss": 1.0542, "step": 2387 }, { "epoch": 0.4594295608676831, "grad_norm": 1.3714503620384113, "learning_rate": 1.1788284448994588e-05, "loss": 0.9796, "step": 2388 }, { "epoch": 0.45962195180606996, "grad_norm": 1.3213058455376487, "learning_rate": 1.1782152479361956e-05, "loss": 0.9644, "step": 2389 }, { "epoch": 0.45981434274445676, "grad_norm": 1.486558364657264, "learning_rate": 1.1776019817561834e-05, "loss": 0.977, "step": 2390 }, { "epoch": 0.46000673368284356, "grad_norm": 1.3304855599217513, "learning_rate": 1.1769886465976086e-05, "loss": 0.94, "step": 2391 }, { "epoch": 0.46019912462123036, "grad_norm": 1.387392595920755, "learning_rate": 1.1763752426986823e-05, "loss": 0.9722, "step": 2392 }, { "epoch": 0.46039151555961716, "grad_norm": 1.3521386277914071, "learning_rate": 1.1757617702976443e-05, "loss": 0.977, "step": 2393 }, { "epoch": 0.46058390649800396, "grad_norm": 1.4571204547234173, "learning_rate": 1.17514822963276e-05, "loss": 0.9987, "step": 2394 }, { "epoch": 0.46077629743639076, "grad_norm": 1.4534761830717835, "learning_rate": 1.1745346209423216e-05, "loss": 1.0015, "step": 2395 }, { "epoch": 0.46096868837477756, "grad_norm": 1.6046225352374042, "learning_rate": 1.1739209444646479e-05, "loss": 0.945, "step": 2396 }, { "epoch": 0.46116107931316436, "grad_norm": 0.7143193119940707, "learning_rate": 1.1733072004380827e-05, "loss": 0.8105, "step": 2397 }, { "epoch": 0.46135347025155116, "grad_norm": 1.4341494163000854, "learning_rate": 1.1726933891009985e-05, "loss": 1.0322, "step": 2398 }, { "epoch": 0.46154586118993796, "grad_norm": 1.3473616459116782, "learning_rate": 1.1720795106917917e-05, "loss": 0.905, "step": 2399 }, { "epoch": 0.46173825212832476, "grad_norm": 1.4136565940851207, "learning_rate": 1.171465565448886e-05, "loss": 1.0241, "step": 2400 }, { "epoch": 0.46193064306671155, "grad_norm": 1.424050454922619, "learning_rate": 1.1708515536107299e-05, "loss": 0.994, "step": 2401 }, { "epoch": 0.46212303400509835, "grad_norm": 1.3927225819594666, "learning_rate": 1.1702374754157998e-05, "loss": 0.9808, "step": 2402 }, { "epoch": 0.46231542494348515, "grad_norm": 1.3936139263655138, "learning_rate": 1.1696233311025957e-05, "loss": 0.9572, "step": 2403 }, { "epoch": 0.46250781588187195, "grad_norm": 1.3524995005050753, "learning_rate": 1.1690091209096441e-05, "loss": 1.0313, "step": 2404 }, { "epoch": 0.46270020682025875, "grad_norm": 1.4173286605967663, "learning_rate": 1.1683948450754976e-05, "loss": 0.9746, "step": 2405 }, { "epoch": 0.46289259775864555, "grad_norm": 1.4213193652600746, "learning_rate": 1.1677805038387337e-05, "loss": 1.0034, "step": 2406 }, { "epoch": 0.46308498869703235, "grad_norm": 1.4596710780317488, "learning_rate": 1.1671660974379554e-05, "loss": 1.0627, "step": 2407 }, { "epoch": 0.46327737963541915, "grad_norm": 1.3330582917447968, "learning_rate": 1.1665516261117914e-05, "loss": 0.9632, "step": 2408 }, { "epoch": 0.46346977057380595, "grad_norm": 1.3257880475878514, "learning_rate": 1.1659370900988946e-05, "loss": 0.9142, "step": 2409 }, { "epoch": 0.46366216151219275, "grad_norm": 1.4078904282257936, "learning_rate": 1.165322489637944e-05, "loss": 1.0164, "step": 2410 }, { "epoch": 0.4638545524505796, "grad_norm": 1.47718169751052, "learning_rate": 1.164707824967644e-05, "loss": 1.0377, "step": 2411 }, { "epoch": 0.4640469433889664, "grad_norm": 1.441653348685328, "learning_rate": 1.1640930963267226e-05, "loss": 0.9805, "step": 2412 }, { "epoch": 0.4642393343273532, "grad_norm": 1.3034661095347126, "learning_rate": 1.1634783039539328e-05, "loss": 0.9702, "step": 2413 }, { "epoch": 0.46443172526574, "grad_norm": 1.3822845886565378, "learning_rate": 1.162863448088054e-05, "loss": 1.0357, "step": 2414 }, { "epoch": 0.4646241162041268, "grad_norm": 0.7632915257158606, "learning_rate": 1.1622485289678886e-05, "loss": 0.9261, "step": 2415 }, { "epoch": 0.4648165071425136, "grad_norm": 0.7047900409673872, "learning_rate": 1.1616335468322641e-05, "loss": 0.8615, "step": 2416 }, { "epoch": 0.4650088980809004, "grad_norm": 1.4371162588612434, "learning_rate": 1.1610185019200324e-05, "loss": 1.0615, "step": 2417 }, { "epoch": 0.4652012890192872, "grad_norm": 1.5048995809891987, "learning_rate": 1.1604033944700701e-05, "loss": 0.9406, "step": 2418 }, { "epoch": 0.465393679957674, "grad_norm": 1.5628856582006103, "learning_rate": 1.1597882247212776e-05, "loss": 1.0134, "step": 2419 }, { "epoch": 0.4655860708960608, "grad_norm": 1.5316742552620422, "learning_rate": 1.15917299291258e-05, "loss": 1.0469, "step": 2420 }, { "epoch": 0.4657784618344476, "grad_norm": 0.6931214474639352, "learning_rate": 1.1585576992829261e-05, "loss": 0.8565, "step": 2421 }, { "epoch": 0.4659708527728344, "grad_norm": 1.4047367620297389, "learning_rate": 1.1579423440712887e-05, "loss": 1.0332, "step": 2422 }, { "epoch": 0.4661632437112212, "grad_norm": 1.3639562523866235, "learning_rate": 1.1573269275166652e-05, "loss": 0.9775, "step": 2423 }, { "epoch": 0.466355634649608, "grad_norm": 1.335862040922082, "learning_rate": 1.1567114498580758e-05, "loss": 0.9709, "step": 2424 }, { "epoch": 0.4665480255879948, "grad_norm": 1.4119058259695785, "learning_rate": 1.1560959113345649e-05, "loss": 1.058, "step": 2425 }, { "epoch": 0.4667404165263816, "grad_norm": 1.3921521802661059, "learning_rate": 1.1554803121852005e-05, "loss": 0.9982, "step": 2426 }, { "epoch": 0.4669328074647684, "grad_norm": 1.3310536849553092, "learning_rate": 1.1548646526490749e-05, "loss": 0.9705, "step": 2427 }, { "epoch": 0.4671251984031552, "grad_norm": 1.4508422004054713, "learning_rate": 1.1542489329653024e-05, "loss": 1.0051, "step": 2428 }, { "epoch": 0.467317589341542, "grad_norm": 1.3504910302674837, "learning_rate": 1.153633153373022e-05, "loss": 1.0069, "step": 2429 }, { "epoch": 0.4675099802799288, "grad_norm": 1.449459811365647, "learning_rate": 1.1530173141113947e-05, "loss": 1.0174, "step": 2430 }, { "epoch": 0.4677023712183156, "grad_norm": 1.3903148842897701, "learning_rate": 1.1524014154196063e-05, "loss": 0.9971, "step": 2431 }, { "epoch": 0.4678947621567024, "grad_norm": 1.506542507972318, "learning_rate": 1.1517854575368644e-05, "loss": 0.941, "step": 2432 }, { "epoch": 0.46808715309508925, "grad_norm": 1.4798229342029037, "learning_rate": 1.1511694407023994e-05, "loss": 1.0215, "step": 2433 }, { "epoch": 0.46827954403347605, "grad_norm": 1.3827468010363224, "learning_rate": 1.1505533651554654e-05, "loss": 0.9383, "step": 2434 }, { "epoch": 0.46847193497186285, "grad_norm": 1.4396763686218244, "learning_rate": 1.1499372311353398e-05, "loss": 0.941, "step": 2435 }, { "epoch": 0.46866432591024965, "grad_norm": 1.412368649702236, "learning_rate": 1.149321038881321e-05, "loss": 1.0261, "step": 2436 }, { "epoch": 0.46885671684863645, "grad_norm": 1.3498872828995518, "learning_rate": 1.1487047886327314e-05, "loss": 1.0673, "step": 2437 }, { "epoch": 0.46904910778702325, "grad_norm": 1.2903152993997618, "learning_rate": 1.1480884806289151e-05, "loss": 0.993, "step": 2438 }, { "epoch": 0.46924149872541004, "grad_norm": 1.407967253341879, "learning_rate": 1.1474721151092397e-05, "loss": 1.0188, "step": 2439 }, { "epoch": 0.46943388966379684, "grad_norm": 1.4458460230505725, "learning_rate": 1.1468556923130943e-05, "loss": 1.0003, "step": 2440 }, { "epoch": 0.46962628060218364, "grad_norm": 1.5638105115672185, "learning_rate": 1.14623921247989e-05, "loss": 0.9415, "step": 2441 }, { "epoch": 0.46981867154057044, "grad_norm": 1.3846928261400937, "learning_rate": 1.1456226758490603e-05, "loss": 1.0172, "step": 2442 }, { "epoch": 0.47001106247895724, "grad_norm": 1.6824627145634776, "learning_rate": 1.1450060826600618e-05, "loss": 1.0114, "step": 2443 }, { "epoch": 0.47020345341734404, "grad_norm": 1.4101443780370124, "learning_rate": 1.1443894331523718e-05, "loss": 0.9626, "step": 2444 }, { "epoch": 0.47039584435573084, "grad_norm": 1.4579676083428275, "learning_rate": 1.1437727275654893e-05, "loss": 1.0075, "step": 2445 }, { "epoch": 0.47058823529411764, "grad_norm": 1.3737167209651056, "learning_rate": 1.1431559661389362e-05, "loss": 0.9239, "step": 2446 }, { "epoch": 0.47078062623250444, "grad_norm": 1.4767010292529132, "learning_rate": 1.1425391491122557e-05, "loss": 1.0179, "step": 2447 }, { "epoch": 0.47097301717089124, "grad_norm": 1.4225390481417624, "learning_rate": 1.141922276725012e-05, "loss": 1.0459, "step": 2448 }, { "epoch": 0.47116540810927804, "grad_norm": 1.334625331355327, "learning_rate": 1.1413053492167915e-05, "loss": 0.9965, "step": 2449 }, { "epoch": 0.47135779904766484, "grad_norm": 1.338848443802459, "learning_rate": 1.1406883668272015e-05, "loss": 0.9786, "step": 2450 }, { "epoch": 0.47155018998605164, "grad_norm": 1.3491107380831453, "learning_rate": 1.140071329795871e-05, "loss": 0.9611, "step": 2451 }, { "epoch": 0.47174258092443844, "grad_norm": 1.404857236301926, "learning_rate": 1.13945423836245e-05, "loss": 1.0436, "step": 2452 }, { "epoch": 0.47193497186282524, "grad_norm": 1.4528779443868802, "learning_rate": 1.1388370927666102e-05, "loss": 1.028, "step": 2453 }, { "epoch": 0.47212736280121204, "grad_norm": 1.2739200894443543, "learning_rate": 1.1382198932480429e-05, "loss": 0.9558, "step": 2454 }, { "epoch": 0.4723197537395989, "grad_norm": 0.7668509647052022, "learning_rate": 1.1376026400464616e-05, "loss": 0.8643, "step": 2455 }, { "epoch": 0.4725121446779857, "grad_norm": 1.3314431663992574, "learning_rate": 1.136985333401601e-05, "loss": 1.0402, "step": 2456 }, { "epoch": 0.4727045356163725, "grad_norm": 1.4573647902613307, "learning_rate": 1.1363679735532151e-05, "loss": 0.9667, "step": 2457 }, { "epoch": 0.4728969265547593, "grad_norm": 1.5806634156328285, "learning_rate": 1.1357505607410797e-05, "loss": 0.9845, "step": 2458 }, { "epoch": 0.4730893174931461, "grad_norm": 1.5436187990023507, "learning_rate": 1.1351330952049908e-05, "loss": 1.0527, "step": 2459 }, { "epoch": 0.4732817084315329, "grad_norm": 1.29178239449431, "learning_rate": 1.1345155771847646e-05, "loss": 1.0016, "step": 2460 }, { "epoch": 0.4734740993699197, "grad_norm": 1.3391866137582962, "learning_rate": 1.1338980069202388e-05, "loss": 0.9879, "step": 2461 }, { "epoch": 0.4736664903083065, "grad_norm": 1.472346886029608, "learning_rate": 1.1332803846512697e-05, "loss": 0.9875, "step": 2462 }, { "epoch": 0.4738588812466933, "grad_norm": 1.6503972389638741, "learning_rate": 1.1326627106177348e-05, "loss": 0.9427, "step": 2463 }, { "epoch": 0.4740512721850801, "grad_norm": 1.438746572886431, "learning_rate": 1.132044985059532e-05, "loss": 0.9643, "step": 2464 }, { "epoch": 0.4742436631234669, "grad_norm": 1.3723497058948537, "learning_rate": 1.1314272082165785e-05, "loss": 1.0252, "step": 2465 }, { "epoch": 0.4744360540618537, "grad_norm": 1.32658905941618, "learning_rate": 1.1308093803288119e-05, "loss": 0.9611, "step": 2466 }, { "epoch": 0.4746284450002405, "grad_norm": 1.325857437306465, "learning_rate": 1.130191501636189e-05, "loss": 0.9727, "step": 2467 }, { "epoch": 0.4748208359386273, "grad_norm": 1.4984385601887515, "learning_rate": 1.1295735723786872e-05, "loss": 1.0676, "step": 2468 }, { "epoch": 0.4750132268770141, "grad_norm": 1.347699486538445, "learning_rate": 1.1289555927963032e-05, "loss": 0.9111, "step": 2469 }, { "epoch": 0.4752056178154009, "grad_norm": 1.552717814794428, "learning_rate": 1.1283375631290528e-05, "loss": 0.9809, "step": 2470 }, { "epoch": 0.4753980087537877, "grad_norm": 1.2591696419472898, "learning_rate": 1.1277194836169714e-05, "loss": 0.923, "step": 2471 }, { "epoch": 0.4755903996921745, "grad_norm": 1.363983664656227, "learning_rate": 1.1271013545001144e-05, "loss": 0.9444, "step": 2472 }, { "epoch": 0.4757827906305613, "grad_norm": 1.4824741117144418, "learning_rate": 1.1264831760185562e-05, "loss": 1.0344, "step": 2473 }, { "epoch": 0.4759751815689481, "grad_norm": 1.2929556236350255, "learning_rate": 1.1258649484123895e-05, "loss": 0.8061, "step": 2474 }, { "epoch": 0.4761675725073349, "grad_norm": 1.3005554953527074, "learning_rate": 1.1252466719217274e-05, "loss": 0.9757, "step": 2475 }, { "epoch": 0.4763599634457217, "grad_norm": 1.4431567578728735, "learning_rate": 1.1246283467867012e-05, "loss": 1.0989, "step": 2476 }, { "epoch": 0.47655235438410853, "grad_norm": 1.3134387581816205, "learning_rate": 1.1240099732474613e-05, "loss": 1.0054, "step": 2477 }, { "epoch": 0.47674474532249533, "grad_norm": 1.4573280320268518, "learning_rate": 1.1233915515441765e-05, "loss": 1.0135, "step": 2478 }, { "epoch": 0.47693713626088213, "grad_norm": 1.5008466341812876, "learning_rate": 1.1227730819170349e-05, "loss": 1.0014, "step": 2479 }, { "epoch": 0.47712952719926893, "grad_norm": 1.2789742219707658, "learning_rate": 1.1221545646062431e-05, "loss": 1.0026, "step": 2480 }, { "epoch": 0.47732191813765573, "grad_norm": 1.3984666340714997, "learning_rate": 1.121535999852026e-05, "loss": 0.9758, "step": 2481 }, { "epoch": 0.47751430907604253, "grad_norm": 1.4613382509652537, "learning_rate": 1.1209173878946271e-05, "loss": 1.0866, "step": 2482 }, { "epoch": 0.47770670001442933, "grad_norm": 1.4328744875959947, "learning_rate": 1.1202987289743078e-05, "loss": 0.9573, "step": 2483 }, { "epoch": 0.47789909095281613, "grad_norm": 1.2919895688214307, "learning_rate": 1.1196800233313488e-05, "loss": 0.9404, "step": 2484 }, { "epoch": 0.47809148189120293, "grad_norm": 1.440308581604877, "learning_rate": 1.1190612712060475e-05, "loss": 1.0425, "step": 2485 }, { "epoch": 0.47828387282958973, "grad_norm": 1.5838422637912102, "learning_rate": 1.1184424728387204e-05, "loss": 1.0203, "step": 2486 }, { "epoch": 0.47847626376797653, "grad_norm": 1.4764580476814653, "learning_rate": 1.1178236284697017e-05, "loss": 0.9762, "step": 2487 }, { "epoch": 0.47866865470636333, "grad_norm": 1.351575661571868, "learning_rate": 1.1172047383393434e-05, "loss": 0.8619, "step": 2488 }, { "epoch": 0.4788610456447501, "grad_norm": 1.6554127511904209, "learning_rate": 1.1165858026880151e-05, "loss": 1.0464, "step": 2489 }, { "epoch": 0.4790534365831369, "grad_norm": 1.2834749547050137, "learning_rate": 1.1159668217561048e-05, "loss": 1.0138, "step": 2490 }, { "epoch": 0.4792458275215237, "grad_norm": 1.3511988467541372, "learning_rate": 1.115347795784017e-05, "loss": 1.0279, "step": 2491 }, { "epoch": 0.4794382184599105, "grad_norm": 1.4388533167482234, "learning_rate": 1.1147287250121745e-05, "loss": 0.97, "step": 2492 }, { "epoch": 0.4796306093982973, "grad_norm": 1.4077988139578772, "learning_rate": 1.1141096096810174e-05, "loss": 1.0385, "step": 2493 }, { "epoch": 0.4798230003366841, "grad_norm": 1.4077889149352645, "learning_rate": 1.1134904500310029e-05, "loss": 1.0553, "step": 2494 }, { "epoch": 0.4800153912750709, "grad_norm": 1.4916961889977787, "learning_rate": 1.1128712463026048e-05, "loss": 1.0476, "step": 2495 }, { "epoch": 0.4802077822134577, "grad_norm": 1.4370536961888924, "learning_rate": 1.1122519987363156e-05, "loss": 0.9429, "step": 2496 }, { "epoch": 0.4804001731518445, "grad_norm": 1.3555397107073572, "learning_rate": 1.1116327075726436e-05, "loss": 0.9877, "step": 2497 }, { "epoch": 0.4805925640902314, "grad_norm": 1.3927556761934374, "learning_rate": 1.1110133730521142e-05, "loss": 1.0129, "step": 2498 }, { "epoch": 0.4807849550286182, "grad_norm": 1.3863712852507915, "learning_rate": 1.11039399541527e-05, "loss": 1.0099, "step": 2499 }, { "epoch": 0.480977345967005, "grad_norm": 1.368032666002305, "learning_rate": 1.10977457490267e-05, "loss": 0.989, "step": 2500 }, { "epoch": 0.4811697369053918, "grad_norm": 1.588774765497214, "learning_rate": 1.10915511175489e-05, "loss": 0.9683, "step": 2501 }, { "epoch": 0.4813621278437786, "grad_norm": 1.4313855076576099, "learning_rate": 1.1085356062125225e-05, "loss": 0.9934, "step": 2502 }, { "epoch": 0.4815545187821654, "grad_norm": 1.4609708498833722, "learning_rate": 1.1079160585161759e-05, "loss": 1.0697, "step": 2503 }, { "epoch": 0.4817469097205522, "grad_norm": 1.433773068234681, "learning_rate": 1.107296468906476e-05, "loss": 1.0372, "step": 2504 }, { "epoch": 0.481939300658939, "grad_norm": 0.7031701222523635, "learning_rate": 1.106676837624064e-05, "loss": 0.846, "step": 2505 }, { "epoch": 0.4821316915973258, "grad_norm": 1.3565248237112815, "learning_rate": 1.1060571649095972e-05, "loss": 1.0574, "step": 2506 }, { "epoch": 0.4823240825357126, "grad_norm": 1.4534935178120292, "learning_rate": 1.10543745100375e-05, "loss": 0.9622, "step": 2507 }, { "epoch": 0.48251647347409937, "grad_norm": 1.319687698139571, "learning_rate": 1.1048176961472114e-05, "loss": 0.9635, "step": 2508 }, { "epoch": 0.48270886441248617, "grad_norm": 1.2758579699181276, "learning_rate": 1.1041979005806876e-05, "loss": 0.9813, "step": 2509 }, { "epoch": 0.48290125535087297, "grad_norm": 1.4860429357293692, "learning_rate": 1.1035780645449001e-05, "loss": 0.9378, "step": 2510 }, { "epoch": 0.48309364628925977, "grad_norm": 1.5631368800180059, "learning_rate": 1.1029581882805857e-05, "loss": 1.0055, "step": 2511 }, { "epoch": 0.48328603722764657, "grad_norm": 0.7300709175886224, "learning_rate": 1.1023382720284973e-05, "loss": 0.8616, "step": 2512 }, { "epoch": 0.48347842816603337, "grad_norm": 1.3058169331111675, "learning_rate": 1.1017183160294033e-05, "loss": 0.9953, "step": 2513 }, { "epoch": 0.48367081910442017, "grad_norm": 0.7077453631401038, "learning_rate": 1.1010983205240878e-05, "loss": 0.8529, "step": 2514 }, { "epoch": 0.48386321004280697, "grad_norm": 1.3999816257361113, "learning_rate": 1.1004782857533488e-05, "loss": 0.9464, "step": 2515 }, { "epoch": 0.48405560098119377, "grad_norm": 1.3553650133993642, "learning_rate": 1.099858211958002e-05, "loss": 0.9468, "step": 2516 }, { "epoch": 0.48424799191958057, "grad_norm": 1.3722449767969824, "learning_rate": 1.0992380993788763e-05, "loss": 1.0067, "step": 2517 }, { "epoch": 0.48444038285796737, "grad_norm": 1.3680569041089135, "learning_rate": 1.0986179482568162e-05, "loss": 1.0092, "step": 2518 }, { "epoch": 0.48463277379635417, "grad_norm": 1.3937272258455724, "learning_rate": 1.0979977588326815e-05, "loss": 1.044, "step": 2519 }, { "epoch": 0.484825164734741, "grad_norm": 1.4295009470942528, "learning_rate": 1.0973775313473465e-05, "loss": 1.0249, "step": 2520 }, { "epoch": 0.4850175556731278, "grad_norm": 1.4038783747234784, "learning_rate": 1.0967572660417001e-05, "loss": 1.0967, "step": 2521 }, { "epoch": 0.4852099466115146, "grad_norm": 1.4023640550635388, "learning_rate": 1.0961369631566468e-05, "loss": 1.0262, "step": 2522 }, { "epoch": 0.4854023375499014, "grad_norm": 1.4393431521943532, "learning_rate": 1.0955166229331048e-05, "loss": 1.0489, "step": 2523 }, { "epoch": 0.4855947284882882, "grad_norm": 1.3558587874617603, "learning_rate": 1.0948962456120068e-05, "loss": 0.9753, "step": 2524 }, { "epoch": 0.485787119426675, "grad_norm": 1.4952103934329404, "learning_rate": 1.0942758314343007e-05, "loss": 0.9938, "step": 2525 }, { "epoch": 0.4859795103650618, "grad_norm": 1.5384646861668962, "learning_rate": 1.0936553806409482e-05, "loss": 1.0131, "step": 2526 }, { "epoch": 0.4861719013034486, "grad_norm": 1.4084235896831847, "learning_rate": 1.0930348934729249e-05, "loss": 0.9457, "step": 2527 }, { "epoch": 0.4863642922418354, "grad_norm": 0.7256333787040232, "learning_rate": 1.0924143701712211e-05, "loss": 0.8489, "step": 2528 }, { "epoch": 0.4865566831802222, "grad_norm": 1.2660664498605365, "learning_rate": 1.0917938109768404e-05, "loss": 1.0328, "step": 2529 }, { "epoch": 0.486749074118609, "grad_norm": 1.3318286724835546, "learning_rate": 1.0911732161308014e-05, "loss": 1.0192, "step": 2530 }, { "epoch": 0.4869414650569958, "grad_norm": 0.7185111658590905, "learning_rate": 1.0905525858741364e-05, "loss": 0.8222, "step": 2531 }, { "epoch": 0.4871338559953826, "grad_norm": 1.418068191975451, "learning_rate": 1.08993192044789e-05, "loss": 0.9168, "step": 2532 }, { "epoch": 0.4873262469337694, "grad_norm": 1.4135498826766482, "learning_rate": 1.089311220093122e-05, "loss": 0.926, "step": 2533 }, { "epoch": 0.4875186378721562, "grad_norm": 1.3315677013339033, "learning_rate": 1.0886904850509052e-05, "loss": 0.9828, "step": 2534 }, { "epoch": 0.487711028810543, "grad_norm": 1.4096585635639638, "learning_rate": 1.0880697155623264e-05, "loss": 1.0283, "step": 2535 }, { "epoch": 0.4879034197489298, "grad_norm": 1.418247602819718, "learning_rate": 1.0874489118684846e-05, "loss": 1.0224, "step": 2536 }, { "epoch": 0.4880958106873166, "grad_norm": 1.4884206013613601, "learning_rate": 1.086828074210493e-05, "loss": 1.0607, "step": 2537 }, { "epoch": 0.4882882016257034, "grad_norm": 1.4843893964244876, "learning_rate": 1.0862072028294777e-05, "loss": 0.9754, "step": 2538 }, { "epoch": 0.4884805925640902, "grad_norm": 1.242933020043668, "learning_rate": 1.0855862979665788e-05, "loss": 0.9827, "step": 2539 }, { "epoch": 0.488672983502477, "grad_norm": 1.4502100017587607, "learning_rate": 1.0849653598629477e-05, "loss": 0.9475, "step": 2540 }, { "epoch": 0.4888653744408638, "grad_norm": 1.3239676512501761, "learning_rate": 1.0843443887597495e-05, "loss": 0.9579, "step": 2541 }, { "epoch": 0.48905776537925066, "grad_norm": 1.3827222685948524, "learning_rate": 1.0837233848981632e-05, "loss": 0.987, "step": 2542 }, { "epoch": 0.48925015631763746, "grad_norm": 1.4396752970610216, "learning_rate": 1.0831023485193787e-05, "loss": 0.9642, "step": 2543 }, { "epoch": 0.48944254725602426, "grad_norm": 0.7607245077061977, "learning_rate": 1.0824812798645997e-05, "loss": 0.8218, "step": 2544 }, { "epoch": 0.48963493819441106, "grad_norm": 1.3908985500072997, "learning_rate": 1.0818601791750418e-05, "loss": 0.9685, "step": 2545 }, { "epoch": 0.48982732913279786, "grad_norm": 1.223222759444141, "learning_rate": 1.0812390466919337e-05, "loss": 0.9385, "step": 2546 }, { "epoch": 0.49001972007118466, "grad_norm": 1.332282444942835, "learning_rate": 1.0806178826565162e-05, "loss": 0.9753, "step": 2547 }, { "epoch": 0.49021211100957146, "grad_norm": 1.3622095045089915, "learning_rate": 1.0799966873100419e-05, "loss": 1.0022, "step": 2548 }, { "epoch": 0.49040450194795826, "grad_norm": 1.4686025807847816, "learning_rate": 1.0793754608937758e-05, "loss": 0.9608, "step": 2549 }, { "epoch": 0.49059689288634506, "grad_norm": 1.3203785708654834, "learning_rate": 1.0787542036489955e-05, "loss": 1.0026, "step": 2550 }, { "epoch": 0.49078928382473186, "grad_norm": 1.483488866893362, "learning_rate": 1.0781329158169902e-05, "loss": 0.984, "step": 2551 }, { "epoch": 0.49098167476311866, "grad_norm": 1.3801758981026886, "learning_rate": 1.0775115976390607e-05, "loss": 1.059, "step": 2552 }, { "epoch": 0.49117406570150546, "grad_norm": 1.4879414321535256, "learning_rate": 1.0768902493565197e-05, "loss": 0.9622, "step": 2553 }, { "epoch": 0.49136645663989226, "grad_norm": 1.5143538921648947, "learning_rate": 1.0762688712106918e-05, "loss": 0.9987, "step": 2554 }, { "epoch": 0.49155884757827906, "grad_norm": 1.5270329105281624, "learning_rate": 1.0756474634429133e-05, "loss": 0.9603, "step": 2555 }, { "epoch": 0.49175123851666586, "grad_norm": 1.514071144164441, "learning_rate": 1.0750260262945314e-05, "loss": 1.0598, "step": 2556 }, { "epoch": 0.49194362945505266, "grad_norm": 1.335130540084194, "learning_rate": 1.0744045600069055e-05, "loss": 0.9642, "step": 2557 }, { "epoch": 0.49213602039343946, "grad_norm": 1.3027558893424818, "learning_rate": 1.0737830648214063e-05, "loss": 0.9513, "step": 2558 }, { "epoch": 0.49232841133182625, "grad_norm": 1.394023470374505, "learning_rate": 1.0731615409794144e-05, "loss": 0.9815, "step": 2559 }, { "epoch": 0.49252080227021305, "grad_norm": 1.4211157795639113, "learning_rate": 1.0725399887223234e-05, "loss": 0.977, "step": 2560 }, { "epoch": 0.49271319320859985, "grad_norm": 1.3932187251831794, "learning_rate": 1.0719184082915364e-05, "loss": 0.9616, "step": 2561 }, { "epoch": 0.49290558414698665, "grad_norm": 1.4415573006246216, "learning_rate": 1.0712967999284682e-05, "loss": 0.9618, "step": 2562 }, { "epoch": 0.49309797508537345, "grad_norm": 1.378606308731851, "learning_rate": 1.0706751638745448e-05, "loss": 0.989, "step": 2563 }, { "epoch": 0.4932903660237603, "grad_norm": 1.589484763187014, "learning_rate": 1.0700535003712023e-05, "loss": 1.043, "step": 2564 }, { "epoch": 0.4934827569621471, "grad_norm": 1.4189116494779654, "learning_rate": 1.069431809659887e-05, "loss": 0.9906, "step": 2565 }, { "epoch": 0.4936751479005339, "grad_norm": 0.6983191770226435, "learning_rate": 1.068810091982057e-05, "loss": 0.8587, "step": 2566 }, { "epoch": 0.4938675388389207, "grad_norm": 1.385696672150538, "learning_rate": 1.0681883475791803e-05, "loss": 0.9744, "step": 2567 }, { "epoch": 0.4940599297773075, "grad_norm": 1.4568939848661928, "learning_rate": 1.067566576692735e-05, "loss": 0.9804, "step": 2568 }, { "epoch": 0.4942523207156943, "grad_norm": 1.3310637438119086, "learning_rate": 1.0669447795642103e-05, "loss": 1.0342, "step": 2569 }, { "epoch": 0.4944447116540811, "grad_norm": 1.4072713024533228, "learning_rate": 1.066322956435104e-05, "loss": 0.9673, "step": 2570 }, { "epoch": 0.4946371025924679, "grad_norm": 1.2914836827317577, "learning_rate": 1.065701107546926e-05, "loss": 1.0023, "step": 2571 }, { "epoch": 0.4948294935308547, "grad_norm": 1.3459609234118457, "learning_rate": 1.065079233141195e-05, "loss": 1.0353, "step": 2572 }, { "epoch": 0.4950218844692415, "grad_norm": 1.386146370649985, "learning_rate": 1.0644573334594395e-05, "loss": 1.0536, "step": 2573 }, { "epoch": 0.4952142754076283, "grad_norm": 1.3611206643210703, "learning_rate": 1.0638354087431986e-05, "loss": 0.9763, "step": 2574 }, { "epoch": 0.4954066663460151, "grad_norm": 1.3211336716436493, "learning_rate": 1.0632134592340204e-05, "loss": 1.0105, "step": 2575 }, { "epoch": 0.4955990572844019, "grad_norm": 1.4314658532644786, "learning_rate": 1.0625914851734632e-05, "loss": 0.9387, "step": 2576 }, { "epoch": 0.4957914482227887, "grad_norm": 1.3009948591747902, "learning_rate": 1.0619694868030943e-05, "loss": 0.9269, "step": 2577 }, { "epoch": 0.4959838391611755, "grad_norm": 1.3859920433094346, "learning_rate": 1.0613474643644907e-05, "loss": 0.9715, "step": 2578 }, { "epoch": 0.4961762300995623, "grad_norm": 1.1928321330810407, "learning_rate": 1.0607254180992391e-05, "loss": 0.8786, "step": 2579 }, { "epoch": 0.4963686210379491, "grad_norm": 0.699108999789214, "learning_rate": 1.0601033482489346e-05, "loss": 0.8233, "step": 2580 }, { "epoch": 0.4965610119763359, "grad_norm": 1.4200612045815255, "learning_rate": 1.0594812550551826e-05, "loss": 1.08, "step": 2581 }, { "epoch": 0.4967534029147227, "grad_norm": 1.3386121567948692, "learning_rate": 1.058859138759596e-05, "loss": 0.9812, "step": 2582 }, { "epoch": 0.4969457938531095, "grad_norm": 1.3963239697822964, "learning_rate": 1.0582369996037985e-05, "loss": 0.9821, "step": 2583 }, { "epoch": 0.4971381847914963, "grad_norm": 1.32158581575495, "learning_rate": 1.0576148378294213e-05, "loss": 0.9676, "step": 2584 }, { "epoch": 0.4973305757298831, "grad_norm": 1.298067356520405, "learning_rate": 1.056992653678105e-05, "loss": 0.9756, "step": 2585 }, { "epoch": 0.49752296666826995, "grad_norm": 1.2534285134329042, "learning_rate": 1.0563704473914986e-05, "loss": 0.9726, "step": 2586 }, { "epoch": 0.49771535760665675, "grad_norm": 0.7096818777682267, "learning_rate": 1.0557482192112603e-05, "loss": 0.8459, "step": 2587 }, { "epoch": 0.49790774854504355, "grad_norm": 1.3789927180391928, "learning_rate": 1.0551259693790556e-05, "loss": 1.034, "step": 2588 }, { "epoch": 0.49810013948343035, "grad_norm": 1.3479274934110945, "learning_rate": 1.0545036981365601e-05, "loss": 1.0177, "step": 2589 }, { "epoch": 0.49829253042181715, "grad_norm": 1.4954278932807314, "learning_rate": 1.053881405725456e-05, "loss": 1.032, "step": 2590 }, { "epoch": 0.49848492136020395, "grad_norm": 1.4399896507377916, "learning_rate": 1.0532590923874349e-05, "loss": 1.0451, "step": 2591 }, { "epoch": 0.49867731229859075, "grad_norm": 1.450442716752215, "learning_rate": 1.0526367583641958e-05, "loss": 1.0507, "step": 2592 }, { "epoch": 0.49886970323697755, "grad_norm": 1.4355280900476117, "learning_rate": 1.0520144038974468e-05, "loss": 0.8805, "step": 2593 }, { "epoch": 0.49906209417536435, "grad_norm": 1.440770268473864, "learning_rate": 1.0513920292289021e-05, "loss": 0.9979, "step": 2594 }, { "epoch": 0.49925448511375115, "grad_norm": 1.3084252024238836, "learning_rate": 1.0507696346002857e-05, "loss": 1.0154, "step": 2595 }, { "epoch": 0.49944687605213794, "grad_norm": 0.7485969204145442, "learning_rate": 1.0501472202533285e-05, "loss": 0.8374, "step": 2596 }, { "epoch": 0.49963926699052474, "grad_norm": 1.3489662073798787, "learning_rate": 1.0495247864297684e-05, "loss": 0.999, "step": 2597 }, { "epoch": 0.49983165792891154, "grad_norm": 1.5561157680542117, "learning_rate": 1.0489023333713522e-05, "loss": 1.0267, "step": 2598 }, { "epoch": 0.5000240488672983, "grad_norm": 1.4012798685938066, "learning_rate": 1.0482798613198328e-05, "loss": 0.9538, "step": 2599 }, { "epoch": 0.5002164398056852, "grad_norm": 1.3374187146160486, "learning_rate": 1.047657370516972e-05, "loss": 0.8949, "step": 2600 }, { "epoch": 0.5004088307440719, "grad_norm": 1.3617649711749356, "learning_rate": 1.0470348612045376e-05, "loss": 1.0492, "step": 2601 }, { "epoch": 0.5006012216824588, "grad_norm": 1.4861887683017332, "learning_rate": 1.0464123336243049e-05, "loss": 0.9627, "step": 2602 }, { "epoch": 0.5007936126208455, "grad_norm": 1.3910637213667745, "learning_rate": 1.0457897880180566e-05, "loss": 0.9213, "step": 2603 }, { "epoch": 0.5009860035592324, "grad_norm": 1.5416301913921513, "learning_rate": 1.0451672246275826e-05, "loss": 0.9687, "step": 2604 }, { "epoch": 0.5011783944976191, "grad_norm": 1.3403953867073242, "learning_rate": 1.0445446436946788e-05, "loss": 1.0263, "step": 2605 }, { "epoch": 0.501370785436006, "grad_norm": 1.3617095389136358, "learning_rate": 1.0439220454611486e-05, "loss": 1.0085, "step": 2606 }, { "epoch": 0.5015631763743927, "grad_norm": 1.3758073821725025, "learning_rate": 1.0432994301688021e-05, "loss": 0.9151, "step": 2607 }, { "epoch": 0.5017555673127796, "grad_norm": 1.3653462053976553, "learning_rate": 1.0426767980594559e-05, "loss": 0.9652, "step": 2608 }, { "epoch": 0.5019479582511663, "grad_norm": 1.3099860653643607, "learning_rate": 1.0420541493749332e-05, "loss": 0.9914, "step": 2609 }, { "epoch": 0.5021403491895532, "grad_norm": 1.3579240267861856, "learning_rate": 1.0414314843570634e-05, "loss": 0.9232, "step": 2610 }, { "epoch": 0.5023327401279399, "grad_norm": 1.3299752780917597, "learning_rate": 1.0408088032476822e-05, "loss": 0.9435, "step": 2611 }, { "epoch": 0.5025251310663268, "grad_norm": 1.3630643659629722, "learning_rate": 1.0401861062886324e-05, "loss": 0.9648, "step": 2612 }, { "epoch": 0.5027175220047135, "grad_norm": 1.3965732718048802, "learning_rate": 1.0395633937217622e-05, "loss": 0.9621, "step": 2613 }, { "epoch": 0.5029099129431004, "grad_norm": 1.3496883928006191, "learning_rate": 1.0389406657889254e-05, "loss": 1.0705, "step": 2614 }, { "epoch": 0.5031023038814871, "grad_norm": 1.3875373675677622, "learning_rate": 1.0383179227319826e-05, "loss": 0.9609, "step": 2615 }, { "epoch": 0.503294694819874, "grad_norm": 1.4412361866054857, "learning_rate": 1.0376951647928007e-05, "loss": 1.0649, "step": 2616 }, { "epoch": 0.5034870857582607, "grad_norm": 1.3532005355890515, "learning_rate": 1.0370723922132506e-05, "loss": 1.0248, "step": 2617 }, { "epoch": 0.5036794766966476, "grad_norm": 1.4172648209349539, "learning_rate": 1.036449605235211e-05, "loss": 1.0122, "step": 2618 }, { "epoch": 0.5038718676350344, "grad_norm": 1.3788958398149092, "learning_rate": 1.0358268041005644e-05, "loss": 0.9481, "step": 2619 }, { "epoch": 0.5040642585734212, "grad_norm": 1.4955651309759361, "learning_rate": 1.0352039890511997e-05, "loss": 0.979, "step": 2620 }, { "epoch": 0.504256649511808, "grad_norm": 1.4089019380993553, "learning_rate": 1.034581160329012e-05, "loss": 0.9629, "step": 2621 }, { "epoch": 0.5044490404501948, "grad_norm": 1.3969163357566976, "learning_rate": 1.0339583181758997e-05, "loss": 1.0448, "step": 2622 }, { "epoch": 0.5046414313885816, "grad_norm": 1.4340512992423484, "learning_rate": 1.033335462833768e-05, "loss": 0.9065, "step": 2623 }, { "epoch": 0.5048338223269684, "grad_norm": 1.4272492869720788, "learning_rate": 1.0327125945445265e-05, "loss": 0.9729, "step": 2624 }, { "epoch": 0.5050262132653552, "grad_norm": 1.3601587052578477, "learning_rate": 1.0320897135500904e-05, "loss": 0.9844, "step": 2625 }, { "epoch": 0.505218604203742, "grad_norm": 1.2776352443228984, "learning_rate": 1.0314668200923791e-05, "loss": 0.9638, "step": 2626 }, { "epoch": 0.5054109951421288, "grad_norm": 1.338979069416863, "learning_rate": 1.0308439144133177e-05, "loss": 0.994, "step": 2627 }, { "epoch": 0.5056033860805156, "grad_norm": 1.3978626559547496, "learning_rate": 1.0302209967548354e-05, "loss": 1.0625, "step": 2628 }, { "epoch": 0.5057957770189024, "grad_norm": 1.2747383697531163, "learning_rate": 1.029598067358866e-05, "loss": 0.9863, "step": 2629 }, { "epoch": 0.5059881679572892, "grad_norm": 1.4524225560829467, "learning_rate": 1.0289751264673485e-05, "loss": 0.9577, "step": 2630 }, { "epoch": 0.506180558895676, "grad_norm": 1.4609245114901726, "learning_rate": 1.0283521743222256e-05, "loss": 0.979, "step": 2631 }, { "epoch": 0.5063729498340628, "grad_norm": 1.4492231810641538, "learning_rate": 1.0277292111654447e-05, "loss": 0.9764, "step": 2632 }, { "epoch": 0.5065653407724496, "grad_norm": 1.317911343083288, "learning_rate": 1.0271062372389582e-05, "loss": 0.9926, "step": 2633 }, { "epoch": 0.5067577317108364, "grad_norm": 1.4479605682787198, "learning_rate": 1.0264832527847212e-05, "loss": 1.0403, "step": 2634 }, { "epoch": 0.5069501226492232, "grad_norm": 0.7024025276657847, "learning_rate": 1.0258602580446941e-05, "loss": 0.8727, "step": 2635 }, { "epoch": 0.50714251358761, "grad_norm": 1.341695737260812, "learning_rate": 1.0252372532608405e-05, "loss": 0.9572, "step": 2636 }, { "epoch": 0.5073349045259968, "grad_norm": 1.4373778665433439, "learning_rate": 1.024614238675129e-05, "loss": 1.02, "step": 2637 }, { "epoch": 0.5075272954643836, "grad_norm": 0.7079999043884816, "learning_rate": 1.0239912145295303e-05, "loss": 0.8274, "step": 2638 }, { "epoch": 0.5077196864027704, "grad_norm": 1.2874161835716278, "learning_rate": 1.0233681810660207e-05, "loss": 0.9714, "step": 2639 }, { "epoch": 0.5079120773411573, "grad_norm": 1.5411312241674104, "learning_rate": 1.0227451385265788e-05, "loss": 0.9301, "step": 2640 }, { "epoch": 0.508104468279544, "grad_norm": 0.7155837327554119, "learning_rate": 1.022122087153187e-05, "loss": 0.8553, "step": 2641 }, { "epoch": 0.5082968592179309, "grad_norm": 1.5851609999617648, "learning_rate": 1.0214990271878319e-05, "loss": 1.044, "step": 2642 }, { "epoch": 0.5084892501563176, "grad_norm": 1.4257320301143745, "learning_rate": 1.0208759588725016e-05, "loss": 0.9716, "step": 2643 }, { "epoch": 0.5086816410947045, "grad_norm": 1.2964659042176923, "learning_rate": 1.0202528824491899e-05, "loss": 0.9849, "step": 2644 }, { "epoch": 0.5088740320330912, "grad_norm": 1.508659258567574, "learning_rate": 1.0196297981598921e-05, "loss": 1.0429, "step": 2645 }, { "epoch": 0.5090664229714781, "grad_norm": 1.453270137014865, "learning_rate": 1.019006706246607e-05, "loss": 0.9816, "step": 2646 }, { "epoch": 0.5092588139098648, "grad_norm": 1.3706358086507422, "learning_rate": 1.018383606951336e-05, "loss": 1.0152, "step": 2647 }, { "epoch": 0.5094512048482517, "grad_norm": 1.4656218919053665, "learning_rate": 1.0177605005160837e-05, "loss": 1.0121, "step": 2648 }, { "epoch": 0.5096435957866384, "grad_norm": 1.4120020910827842, "learning_rate": 1.0171373871828578e-05, "loss": 0.9688, "step": 2649 }, { "epoch": 0.5098359867250253, "grad_norm": 1.4266473841421548, "learning_rate": 1.0165142671936685e-05, "loss": 1.0189, "step": 2650 }, { "epoch": 0.510028377663412, "grad_norm": 1.4705193422502834, "learning_rate": 1.0158911407905279e-05, "loss": 1.0047, "step": 2651 }, { "epoch": 0.5102207686017989, "grad_norm": 1.44205031727962, "learning_rate": 1.0152680082154514e-05, "loss": 0.8946, "step": 2652 }, { "epoch": 0.5104131595401856, "grad_norm": 1.5042937661036655, "learning_rate": 1.0146448697104561e-05, "loss": 1.0025, "step": 2653 }, { "epoch": 0.5106055504785725, "grad_norm": 1.3705968232056402, "learning_rate": 1.0140217255175626e-05, "loss": 0.9673, "step": 2654 }, { "epoch": 0.5107979414169592, "grad_norm": 1.5390328890450276, "learning_rate": 1.013398575878792e-05, "loss": 0.9891, "step": 2655 }, { "epoch": 0.5109903323553461, "grad_norm": 1.5138771822954014, "learning_rate": 1.0127754210361694e-05, "loss": 1.0059, "step": 2656 }, { "epoch": 0.5111827232937328, "grad_norm": 1.4583937542514065, "learning_rate": 1.0121522612317204e-05, "loss": 0.9264, "step": 2657 }, { "epoch": 0.5113751142321197, "grad_norm": 1.4869109850278674, "learning_rate": 1.011529096707473e-05, "loss": 0.9892, "step": 2658 }, { "epoch": 0.5115675051705064, "grad_norm": 1.307924479140096, "learning_rate": 1.0109059277054574e-05, "loss": 1.0579, "step": 2659 }, { "epoch": 0.5117598961088933, "grad_norm": 0.7017739518736756, "learning_rate": 1.010282754467705e-05, "loss": 0.826, "step": 2660 }, { "epoch": 0.51195228704728, "grad_norm": 1.306076300299477, "learning_rate": 1.0096595772362492e-05, "loss": 1.0484, "step": 2661 }, { "epoch": 0.5121446779856669, "grad_norm": 1.3098287054622215, "learning_rate": 1.0090363962531251e-05, "loss": 1.0209, "step": 2662 }, { "epoch": 0.5123370689240537, "grad_norm": 1.4121373477221513, "learning_rate": 1.0084132117603689e-05, "loss": 1.0732, "step": 2663 }, { "epoch": 0.5125294598624405, "grad_norm": 0.7353115629887503, "learning_rate": 1.0077900240000181e-05, "loss": 0.8548, "step": 2664 }, { "epoch": 0.5127218508008273, "grad_norm": 1.302512863469885, "learning_rate": 1.0071668332141115e-05, "loss": 1.0415, "step": 2665 }, { "epoch": 0.5129142417392141, "grad_norm": 1.4468077593871391, "learning_rate": 1.0065436396446899e-05, "loss": 0.9771, "step": 2666 }, { "epoch": 0.5131066326776009, "grad_norm": 1.5308987321256993, "learning_rate": 1.0059204435337938e-05, "loss": 1.0028, "step": 2667 }, { "epoch": 0.5132990236159877, "grad_norm": 1.4976417993164077, "learning_rate": 1.0052972451234656e-05, "loss": 1.0289, "step": 2668 }, { "epoch": 0.5134914145543745, "grad_norm": 1.4472643275279324, "learning_rate": 1.0046740446557485e-05, "loss": 1.0274, "step": 2669 }, { "epoch": 0.5136838054927613, "grad_norm": 1.3394633923883066, "learning_rate": 1.0040508423726865e-05, "loss": 1.0013, "step": 2670 }, { "epoch": 0.5138761964311481, "grad_norm": 1.3420528839338388, "learning_rate": 1.0034276385163238e-05, "loss": 1.0126, "step": 2671 }, { "epoch": 0.5140685873695349, "grad_norm": 1.4169355105786652, "learning_rate": 1.0028044333287056e-05, "loss": 0.9623, "step": 2672 }, { "epoch": 0.5142609783079217, "grad_norm": 1.3871876757254336, "learning_rate": 1.002181227051878e-05, "loss": 1.0003, "step": 2673 }, { "epoch": 0.5144533692463085, "grad_norm": 1.4034985477788762, "learning_rate": 1.0015580199278873e-05, "loss": 1.0269, "step": 2674 }, { "epoch": 0.5146457601846953, "grad_norm": 1.3446394313266168, "learning_rate": 1.0009348121987795e-05, "loss": 0.982, "step": 2675 }, { "epoch": 0.5148381511230821, "grad_norm": 1.3122991068502305, "learning_rate": 1.000311604106601e-05, "loss": 0.939, "step": 2676 }, { "epoch": 0.5150305420614689, "grad_norm": 1.2931062424109787, "learning_rate": 9.996883958933993e-06, "loss": 1.0286, "step": 2677 }, { "epoch": 0.5152229329998557, "grad_norm": 1.3403081373594334, "learning_rate": 9.99065187801221e-06, "loss": 0.8835, "step": 2678 }, { "epoch": 0.5154153239382425, "grad_norm": 1.4258266552869774, "learning_rate": 9.984419800721132e-06, "loss": 1.0255, "step": 2679 }, { "epoch": 0.5156077148766293, "grad_norm": 1.3779052007799963, "learning_rate": 9.978187729481218e-06, "loss": 0.953, "step": 2680 }, { "epoch": 0.5158001058150161, "grad_norm": 1.410007110345334, "learning_rate": 9.971955666712945e-06, "loss": 1.0157, "step": 2681 }, { "epoch": 0.5159924967534029, "grad_norm": 1.3939624902735597, "learning_rate": 9.965723614836764e-06, "loss": 0.9682, "step": 2682 }, { "epoch": 0.5161848876917897, "grad_norm": 1.4614866455443152, "learning_rate": 9.959491576273139e-06, "loss": 0.994, "step": 2683 }, { "epoch": 0.5163772786301766, "grad_norm": 1.3885373723376504, "learning_rate": 9.95325955344252e-06, "loss": 1.0476, "step": 2684 }, { "epoch": 0.5165696695685633, "grad_norm": 1.4192997538655538, "learning_rate": 9.947027548765347e-06, "loss": 1.0044, "step": 2685 }, { "epoch": 0.5167620605069502, "grad_norm": 1.5322782402393935, "learning_rate": 9.940795564662064e-06, "loss": 1.0821, "step": 2686 }, { "epoch": 0.5169544514453369, "grad_norm": 1.4036119320818057, "learning_rate": 9.934563603553103e-06, "loss": 1.0087, "step": 2687 }, { "epoch": 0.5171468423837238, "grad_norm": 1.4796285141024836, "learning_rate": 9.928331667858886e-06, "loss": 0.9961, "step": 2688 }, { "epoch": 0.5173392333221105, "grad_norm": 1.4262739348535896, "learning_rate": 9.922099759999822e-06, "loss": 1.0198, "step": 2689 }, { "epoch": 0.5175316242604974, "grad_norm": 1.6113513213839001, "learning_rate": 9.915867882396314e-06, "loss": 0.9607, "step": 2690 }, { "epoch": 0.5177240151988841, "grad_norm": 0.7272243232505152, "learning_rate": 9.909636037468754e-06, "loss": 0.8575, "step": 2691 }, { "epoch": 0.517916406137271, "grad_norm": 1.385352371315459, "learning_rate": 9.90340422763751e-06, "loss": 1.0049, "step": 2692 }, { "epoch": 0.5181087970756577, "grad_norm": 1.4060821070556624, "learning_rate": 9.897172455322953e-06, "loss": 0.9811, "step": 2693 }, { "epoch": 0.5183011880140446, "grad_norm": 1.576456516860468, "learning_rate": 9.890940722945429e-06, "loss": 0.9929, "step": 2694 }, { "epoch": 0.5184935789524313, "grad_norm": 1.4603464399924402, "learning_rate": 9.884709032925274e-06, "loss": 0.9552, "step": 2695 }, { "epoch": 0.5186859698908182, "grad_norm": 1.4242460248751008, "learning_rate": 9.878477387682801e-06, "loss": 1.085, "step": 2696 }, { "epoch": 0.5188783608292049, "grad_norm": 1.418631872937384, "learning_rate": 9.872245789638308e-06, "loss": 0.9478, "step": 2697 }, { "epoch": 0.5190707517675918, "grad_norm": 1.3249899184646052, "learning_rate": 9.866014241212078e-06, "loss": 0.9055, "step": 2698 }, { "epoch": 0.5192631427059785, "grad_norm": 1.347514542684691, "learning_rate": 9.859782744824376e-06, "loss": 0.9778, "step": 2699 }, { "epoch": 0.5194555336443654, "grad_norm": 1.481140213056934, "learning_rate": 9.85355130289544e-06, "loss": 1.0112, "step": 2700 }, { "epoch": 0.5196479245827521, "grad_norm": 1.4791501576864687, "learning_rate": 9.84731991784549e-06, "loss": 0.9905, "step": 2701 }, { "epoch": 0.519840315521139, "grad_norm": 1.3336027467769815, "learning_rate": 9.841088592094726e-06, "loss": 0.9476, "step": 2702 }, { "epoch": 0.5200327064595257, "grad_norm": 1.3138881017742254, "learning_rate": 9.834857328063316e-06, "loss": 0.9457, "step": 2703 }, { "epoch": 0.5202250973979126, "grad_norm": 1.368221908178884, "learning_rate": 9.828626128171422e-06, "loss": 0.9415, "step": 2704 }, { "epoch": 0.5204174883362993, "grad_norm": 1.3726749366910773, "learning_rate": 9.822394994839164e-06, "loss": 0.9364, "step": 2705 }, { "epoch": 0.5206098792746862, "grad_norm": 1.294093703526912, "learning_rate": 9.816163930486643e-06, "loss": 0.9543, "step": 2706 }, { "epoch": 0.520802270213073, "grad_norm": 1.3696779634778586, "learning_rate": 9.809932937533935e-06, "loss": 0.9574, "step": 2707 }, { "epoch": 0.5209946611514598, "grad_norm": 1.3586037125799755, "learning_rate": 9.803702018401084e-06, "loss": 1.036, "step": 2708 }, { "epoch": 0.5211870520898466, "grad_norm": 1.401246373112545, "learning_rate": 9.797471175508101e-06, "loss": 1.0093, "step": 2709 }, { "epoch": 0.5213794430282334, "grad_norm": 1.2998418708348103, "learning_rate": 9.791240411274982e-06, "loss": 0.9321, "step": 2710 }, { "epoch": 0.5215718339666202, "grad_norm": 1.3789622273023774, "learning_rate": 9.785009728121686e-06, "loss": 0.8878, "step": 2711 }, { "epoch": 0.521764224905007, "grad_norm": 1.2818721527145218, "learning_rate": 9.778779128468133e-06, "loss": 0.9492, "step": 2712 }, { "epoch": 0.5219566158433938, "grad_norm": 1.3947717399807689, "learning_rate": 9.772548614734217e-06, "loss": 0.9644, "step": 2713 }, { "epoch": 0.5221490067817806, "grad_norm": 1.2921766827264523, "learning_rate": 9.766318189339798e-06, "loss": 0.9928, "step": 2714 }, { "epoch": 0.5223413977201674, "grad_norm": 0.7239269599597173, "learning_rate": 9.760087854704697e-06, "loss": 0.8336, "step": 2715 }, { "epoch": 0.5225337886585542, "grad_norm": 1.2959239576962835, "learning_rate": 9.753857613248714e-06, "loss": 1.0387, "step": 2716 }, { "epoch": 0.522726179596941, "grad_norm": 1.411507592474316, "learning_rate": 9.747627467391596e-06, "loss": 1.0231, "step": 2717 }, { "epoch": 0.5229185705353278, "grad_norm": 1.3920385463663605, "learning_rate": 9.741397419553062e-06, "loss": 0.9605, "step": 2718 }, { "epoch": 0.5231109614737146, "grad_norm": 1.3130401145931008, "learning_rate": 9.735167472152793e-06, "loss": 1.0058, "step": 2719 }, { "epoch": 0.5233033524121014, "grad_norm": 1.2682079249482086, "learning_rate": 9.728937627610425e-06, "loss": 1.0419, "step": 2720 }, { "epoch": 0.5234957433504882, "grad_norm": 1.3779291868685841, "learning_rate": 9.722707888345553e-06, "loss": 1.0643, "step": 2721 }, { "epoch": 0.523688134288875, "grad_norm": 1.4229659465232556, "learning_rate": 9.716478256777749e-06, "loss": 1.0168, "step": 2722 }, { "epoch": 0.5238805252272618, "grad_norm": 1.4340114378496684, "learning_rate": 9.710248735326519e-06, "loss": 0.9955, "step": 2723 }, { "epoch": 0.5240729161656486, "grad_norm": 1.4416824739933998, "learning_rate": 9.704019326411344e-06, "loss": 1.0483, "step": 2724 }, { "epoch": 0.5242653071040354, "grad_norm": 1.2899640874227418, "learning_rate": 9.697790032451651e-06, "loss": 0.9586, "step": 2725 }, { "epoch": 0.5244576980424221, "grad_norm": 1.4229665032488785, "learning_rate": 9.691560855866826e-06, "loss": 1.0157, "step": 2726 }, { "epoch": 0.524650088980809, "grad_norm": 1.4061402464792507, "learning_rate": 9.685331799076208e-06, "loss": 0.9242, "step": 2727 }, { "epoch": 0.5248424799191959, "grad_norm": 1.4692579293996946, "learning_rate": 9.6791028644991e-06, "loss": 0.9059, "step": 2728 }, { "epoch": 0.5250348708575826, "grad_norm": 1.2920401438623499, "learning_rate": 9.672874054554738e-06, "loss": 0.9861, "step": 2729 }, { "epoch": 0.5252272617959695, "grad_norm": 1.3800044409763275, "learning_rate": 9.666645371662324e-06, "loss": 1.011, "step": 2730 }, { "epoch": 0.5254196527343562, "grad_norm": 1.2380566440302867, "learning_rate": 9.660416818241007e-06, "loss": 0.9494, "step": 2731 }, { "epoch": 0.5256120436727431, "grad_norm": 1.3596244550365941, "learning_rate": 9.654188396709882e-06, "loss": 0.9671, "step": 2732 }, { "epoch": 0.5258044346111298, "grad_norm": 1.3722406218764247, "learning_rate": 9.647960109488003e-06, "loss": 0.9932, "step": 2733 }, { "epoch": 0.5259968255495167, "grad_norm": 1.4191055751469506, "learning_rate": 9.64173195899436e-06, "loss": 1.0557, "step": 2734 }, { "epoch": 0.5261892164879034, "grad_norm": 1.3075838670390976, "learning_rate": 9.635503947647894e-06, "loss": 1.0364, "step": 2735 }, { "epoch": 0.5263816074262903, "grad_norm": 1.3951667334626074, "learning_rate": 9.629276077867497e-06, "loss": 1.0584, "step": 2736 }, { "epoch": 0.526573998364677, "grad_norm": 1.5024333002028947, "learning_rate": 9.623048352071998e-06, "loss": 0.9988, "step": 2737 }, { "epoch": 0.5267663893030639, "grad_norm": 0.7329374241533348, "learning_rate": 9.616820772680174e-06, "loss": 0.9054, "step": 2738 }, { "epoch": 0.5269587802414506, "grad_norm": 1.3098931341260822, "learning_rate": 9.610593342110746e-06, "loss": 0.9524, "step": 2739 }, { "epoch": 0.5271511711798375, "grad_norm": 1.3333275740989297, "learning_rate": 9.604366062782381e-06, "loss": 0.9574, "step": 2740 }, { "epoch": 0.5273435621182242, "grad_norm": 1.3847669584006819, "learning_rate": 9.598138937113677e-06, "loss": 0.9531, "step": 2741 }, { "epoch": 0.527535953056611, "grad_norm": 1.3142700431254988, "learning_rate": 9.59191196752318e-06, "loss": 0.8944, "step": 2742 }, { "epoch": 0.5277283439949978, "grad_norm": 1.365130127509295, "learning_rate": 9.58568515642937e-06, "loss": 1.0237, "step": 2743 }, { "epoch": 0.5279207349333846, "grad_norm": 1.4114933858110426, "learning_rate": 9.579458506250668e-06, "loss": 0.9361, "step": 2744 }, { "epoch": 0.5281131258717714, "grad_norm": 1.3691152705920475, "learning_rate": 9.573232019405441e-06, "loss": 0.9571, "step": 2745 }, { "epoch": 0.5283055168101582, "grad_norm": 0.6673033698167111, "learning_rate": 9.567005698311982e-06, "loss": 0.8103, "step": 2746 }, { "epoch": 0.528497907748545, "grad_norm": 1.465764199435743, "learning_rate": 9.560779545388517e-06, "loss": 0.94, "step": 2747 }, { "epoch": 0.5286902986869318, "grad_norm": 0.7118727311300314, "learning_rate": 9.554553563053217e-06, "loss": 0.8156, "step": 2748 }, { "epoch": 0.5288826896253187, "grad_norm": 1.4030647991175615, "learning_rate": 9.548327753724181e-06, "loss": 1.0155, "step": 2749 }, { "epoch": 0.5290750805637054, "grad_norm": 1.6318883156655015, "learning_rate": 9.542102119819436e-06, "loss": 1.0187, "step": 2750 }, { "epoch": 0.5292674715020923, "grad_norm": 1.4707285754514818, "learning_rate": 9.535876663756955e-06, "loss": 1.0598, "step": 2751 }, { "epoch": 0.529459862440479, "grad_norm": 1.4403658822747858, "learning_rate": 9.529651387954628e-06, "loss": 1.0266, "step": 2752 }, { "epoch": 0.5296522533788659, "grad_norm": 1.3923746452278185, "learning_rate": 9.523426294830284e-06, "loss": 0.963, "step": 2753 }, { "epoch": 0.5298446443172526, "grad_norm": 1.3759984876209634, "learning_rate": 9.517201386801675e-06, "loss": 1.0115, "step": 2754 }, { "epoch": 0.5300370352556395, "grad_norm": 1.5078568216104031, "learning_rate": 9.510976666286484e-06, "loss": 1.0237, "step": 2755 }, { "epoch": 0.5302294261940262, "grad_norm": 1.2936366065005434, "learning_rate": 9.504752135702318e-06, "loss": 0.9488, "step": 2756 }, { "epoch": 0.5304218171324131, "grad_norm": 1.3400078128348785, "learning_rate": 9.498527797466718e-06, "loss": 0.9906, "step": 2757 }, { "epoch": 0.5306142080707998, "grad_norm": 1.3380466526449346, "learning_rate": 9.492303653997146e-06, "loss": 0.9575, "step": 2758 }, { "epoch": 0.5308065990091867, "grad_norm": 1.3848821538412683, "learning_rate": 9.48607970771098e-06, "loss": 0.9312, "step": 2759 }, { "epoch": 0.5309989899475734, "grad_norm": 1.3510586696751339, "learning_rate": 9.479855961025538e-06, "loss": 1.1027, "step": 2760 }, { "epoch": 0.5311913808859603, "grad_norm": 1.534313644883511, "learning_rate": 9.473632416358045e-06, "loss": 1.0276, "step": 2761 }, { "epoch": 0.531383771824347, "grad_norm": 1.3380306576704935, "learning_rate": 9.467409076125653e-06, "loss": 1.0468, "step": 2762 }, { "epoch": 0.5315761627627339, "grad_norm": 1.3061207217231756, "learning_rate": 9.461185942745443e-06, "loss": 0.9551, "step": 2763 }, { "epoch": 0.5317685537011206, "grad_norm": 1.445314692270578, "learning_rate": 9.454963018634402e-06, "loss": 0.9745, "step": 2764 }, { "epoch": 0.5319609446395075, "grad_norm": 1.4809717533808162, "learning_rate": 9.448740306209447e-06, "loss": 1.0159, "step": 2765 }, { "epoch": 0.5321533355778942, "grad_norm": 1.5164036387824265, "learning_rate": 9.442517807887402e-06, "loss": 1.0381, "step": 2766 }, { "epoch": 0.5323457265162811, "grad_norm": 1.3633972587643521, "learning_rate": 9.436295526085016e-06, "loss": 0.9899, "step": 2767 }, { "epoch": 0.5325381174546678, "grad_norm": 1.5795827679228907, "learning_rate": 9.430073463218952e-06, "loss": 0.9596, "step": 2768 }, { "epoch": 0.5327305083930547, "grad_norm": 0.7491522306659684, "learning_rate": 9.423851621705789e-06, "loss": 0.8431, "step": 2769 }, { "epoch": 0.5329228993314414, "grad_norm": 1.4853321837055775, "learning_rate": 9.41763000396202e-06, "loss": 1.0014, "step": 2770 }, { "epoch": 0.5331152902698283, "grad_norm": 1.4537569922741391, "learning_rate": 9.411408612404043e-06, "loss": 0.9493, "step": 2771 }, { "epoch": 0.5333076812082151, "grad_norm": 1.310308826135544, "learning_rate": 9.40518744944818e-06, "loss": 0.911, "step": 2772 }, { "epoch": 0.5335000721466019, "grad_norm": 1.365480174989118, "learning_rate": 9.398966517510654e-06, "loss": 0.8551, "step": 2773 }, { "epoch": 0.5336924630849887, "grad_norm": 1.372322186300014, "learning_rate": 9.39274581900761e-06, "loss": 0.9608, "step": 2774 }, { "epoch": 0.5338848540233755, "grad_norm": 1.3819999871198685, "learning_rate": 9.386525356355095e-06, "loss": 0.9735, "step": 2775 }, { "epoch": 0.5340772449617623, "grad_norm": 1.3315780658436658, "learning_rate": 9.380305131969059e-06, "loss": 1.0133, "step": 2776 }, { "epoch": 0.5342696359001491, "grad_norm": 0.7815793900049088, "learning_rate": 9.374085148265372e-06, "loss": 0.9702, "step": 2777 }, { "epoch": 0.5344620268385359, "grad_norm": 0.7200904324203388, "learning_rate": 9.3678654076598e-06, "loss": 0.8542, "step": 2778 }, { "epoch": 0.5346544177769227, "grad_norm": 1.3555884039192811, "learning_rate": 9.361645912568015e-06, "loss": 0.9432, "step": 2779 }, { "epoch": 0.5348468087153095, "grad_norm": 1.2913145169058882, "learning_rate": 9.355426665405607e-06, "loss": 0.9611, "step": 2780 }, { "epoch": 0.5350391996536963, "grad_norm": 1.2989295935096927, "learning_rate": 9.349207668588053e-06, "loss": 0.9817, "step": 2781 }, { "epoch": 0.5352315905920831, "grad_norm": 1.365099137276945, "learning_rate": 9.342988924530742e-06, "loss": 1.0091, "step": 2782 }, { "epoch": 0.5354239815304699, "grad_norm": 1.34116634791925, "learning_rate": 9.336770435648963e-06, "loss": 0.9306, "step": 2783 }, { "epoch": 0.5356163724688567, "grad_norm": 1.344845833339101, "learning_rate": 9.330552204357904e-06, "loss": 0.9871, "step": 2784 }, { "epoch": 0.5358087634072435, "grad_norm": 1.450318937427637, "learning_rate": 9.32433423307265e-06, "loss": 1.0014, "step": 2785 }, { "epoch": 0.5360011543456303, "grad_norm": 1.3705147622388552, "learning_rate": 9.318116524208198e-06, "loss": 1.0082, "step": 2786 }, { "epoch": 0.5361935452840171, "grad_norm": 0.7158368630618961, "learning_rate": 9.311899080179433e-06, "loss": 0.8542, "step": 2787 }, { "epoch": 0.5363859362224039, "grad_norm": 1.332856535720958, "learning_rate": 9.305681903401133e-06, "loss": 0.8933, "step": 2788 }, { "epoch": 0.5365783271607907, "grad_norm": 1.367363785352865, "learning_rate": 9.299464996287984e-06, "loss": 0.987, "step": 2789 }, { "epoch": 0.5367707180991775, "grad_norm": 1.42326879439809, "learning_rate": 9.293248361254557e-06, "loss": 1.0283, "step": 2790 }, { "epoch": 0.5369631090375643, "grad_norm": 1.483852821058687, "learning_rate": 9.287032000715318e-06, "loss": 0.9204, "step": 2791 }, { "epoch": 0.5371554999759511, "grad_norm": 1.2672731936415957, "learning_rate": 9.28081591708464e-06, "loss": 0.98, "step": 2792 }, { "epoch": 0.537347890914338, "grad_norm": 1.3679223208570024, "learning_rate": 9.27460011277677e-06, "loss": 1.0207, "step": 2793 }, { "epoch": 0.5375402818527247, "grad_norm": 1.4100900896464568, "learning_rate": 9.268384590205858e-06, "loss": 0.923, "step": 2794 }, { "epoch": 0.5377326727911116, "grad_norm": 1.4136447623533943, "learning_rate": 9.262169351785944e-06, "loss": 0.926, "step": 2795 }, { "epoch": 0.5379250637294983, "grad_norm": 1.5128567710770284, "learning_rate": 9.255954399930948e-06, "loss": 1.0232, "step": 2796 }, { "epoch": 0.5381174546678852, "grad_norm": 1.2928582935754782, "learning_rate": 9.249739737054686e-06, "loss": 0.9184, "step": 2797 }, { "epoch": 0.5383098456062719, "grad_norm": 1.3851178781818674, "learning_rate": 9.24352536557087e-06, "loss": 0.9646, "step": 2798 }, { "epoch": 0.5385022365446588, "grad_norm": 1.3198479390342077, "learning_rate": 9.237311287893086e-06, "loss": 0.9157, "step": 2799 }, { "epoch": 0.5386946274830455, "grad_norm": 1.426470762585573, "learning_rate": 9.231097506434808e-06, "loss": 0.9775, "step": 2800 }, { "epoch": 0.5388870184214324, "grad_norm": 1.6528418862234067, "learning_rate": 9.224884023609398e-06, "loss": 1.0155, "step": 2801 }, { "epoch": 0.5390794093598191, "grad_norm": 1.3767849014895606, "learning_rate": 9.218670841830098e-06, "loss": 0.9937, "step": 2802 }, { "epoch": 0.539271800298206, "grad_norm": 1.4727361789925693, "learning_rate": 9.212457963510045e-06, "loss": 0.9758, "step": 2803 }, { "epoch": 0.5394641912365927, "grad_norm": 1.549485168948777, "learning_rate": 9.206245391062243e-06, "loss": 1.0437, "step": 2804 }, { "epoch": 0.5396565821749796, "grad_norm": 1.3081286937676941, "learning_rate": 9.200033126899585e-06, "loss": 0.9315, "step": 2805 }, { "epoch": 0.5398489731133663, "grad_norm": 1.3548552539584102, "learning_rate": 9.193821173434843e-06, "loss": 0.9224, "step": 2806 }, { "epoch": 0.5400413640517532, "grad_norm": 1.4620417348236272, "learning_rate": 9.187609533080668e-06, "loss": 0.9475, "step": 2807 }, { "epoch": 0.5402337549901399, "grad_norm": 1.3998841639540391, "learning_rate": 9.181398208249583e-06, "loss": 0.9791, "step": 2808 }, { "epoch": 0.5404261459285268, "grad_norm": 1.3330367376733245, "learning_rate": 9.175187201354005e-06, "loss": 0.9798, "step": 2809 }, { "epoch": 0.5406185368669135, "grad_norm": 0.6734967704225993, "learning_rate": 9.168976514806216e-06, "loss": 0.8307, "step": 2810 }, { "epoch": 0.5408109278053004, "grad_norm": 1.3418004639952097, "learning_rate": 9.162766151018372e-06, "loss": 0.9692, "step": 2811 }, { "epoch": 0.5410033187436871, "grad_norm": 0.735965217187769, "learning_rate": 9.156556112402508e-06, "loss": 0.913, "step": 2812 }, { "epoch": 0.541195709682074, "grad_norm": 1.4544166397653187, "learning_rate": 9.150346401370528e-06, "loss": 1.0491, "step": 2813 }, { "epoch": 0.5413881006204607, "grad_norm": 1.3381598694993917, "learning_rate": 9.144137020334214e-06, "loss": 0.9747, "step": 2814 }, { "epoch": 0.5415804915588476, "grad_norm": 1.3864617045615084, "learning_rate": 9.137927971705223e-06, "loss": 0.9942, "step": 2815 }, { "epoch": 0.5417728824972344, "grad_norm": 1.5836825416120415, "learning_rate": 9.131719257895074e-06, "loss": 0.975, "step": 2816 }, { "epoch": 0.5419652734356212, "grad_norm": 1.3467908879048818, "learning_rate": 9.125510881315159e-06, "loss": 1.03, "step": 2817 }, { "epoch": 0.542157664374008, "grad_norm": 1.2951405571444277, "learning_rate": 9.119302844376741e-06, "loss": 0.9685, "step": 2818 }, { "epoch": 0.5423500553123948, "grad_norm": 1.3663634534955815, "learning_rate": 9.113095149490951e-06, "loss": 0.9309, "step": 2819 }, { "epoch": 0.5425424462507816, "grad_norm": 1.3840132308559763, "learning_rate": 9.106887799068782e-06, "loss": 0.9237, "step": 2820 }, { "epoch": 0.5427348371891684, "grad_norm": 1.297779648687783, "learning_rate": 9.100680795521104e-06, "loss": 1.0096, "step": 2821 }, { "epoch": 0.5429272281275552, "grad_norm": 1.315255422838585, "learning_rate": 9.09447414125864e-06, "loss": 1.0011, "step": 2822 }, { "epoch": 0.543119619065942, "grad_norm": 1.4431583039198648, "learning_rate": 9.088267838691987e-06, "loss": 1.0072, "step": 2823 }, { "epoch": 0.5433120100043288, "grad_norm": 1.3154104387094947, "learning_rate": 9.0820618902316e-06, "loss": 0.9912, "step": 2824 }, { "epoch": 0.5435044009427156, "grad_norm": 1.3409444519602625, "learning_rate": 9.075856298287796e-06, "loss": 0.9646, "step": 2825 }, { "epoch": 0.5436967918811024, "grad_norm": 1.4763689096165615, "learning_rate": 9.069651065270753e-06, "loss": 0.9525, "step": 2826 }, { "epoch": 0.5438891828194892, "grad_norm": 1.4773209413404824, "learning_rate": 9.06344619359052e-06, "loss": 1.072, "step": 2827 }, { "epoch": 0.544081573757876, "grad_norm": 1.3917138984430504, "learning_rate": 9.057241685656995e-06, "loss": 0.8537, "step": 2828 }, { "epoch": 0.5442739646962628, "grad_norm": 1.3435707202245186, "learning_rate": 9.051037543879933e-06, "loss": 1.018, "step": 2829 }, { "epoch": 0.5444663556346496, "grad_norm": 1.414124180035908, "learning_rate": 9.044833770668957e-06, "loss": 1.0461, "step": 2830 }, { "epoch": 0.5446587465730364, "grad_norm": 1.5686984813071847, "learning_rate": 9.038630368433537e-06, "loss": 1.0114, "step": 2831 }, { "epoch": 0.5448511375114232, "grad_norm": 1.4215276614940036, "learning_rate": 9.032427339583e-06, "loss": 0.9647, "step": 2832 }, { "epoch": 0.54504352844981, "grad_norm": 1.4694713149902916, "learning_rate": 9.026224686526539e-06, "loss": 0.9343, "step": 2833 }, { "epoch": 0.5452359193881968, "grad_norm": 0.7282769446401178, "learning_rate": 9.020022411673186e-06, "loss": 0.8694, "step": 2834 }, { "epoch": 0.5454283103265836, "grad_norm": 1.4131761333455923, "learning_rate": 9.013820517431841e-06, "loss": 0.9887, "step": 2835 }, { "epoch": 0.5456207012649704, "grad_norm": 0.6762717042109169, "learning_rate": 9.00761900621124e-06, "loss": 0.8117, "step": 2836 }, { "epoch": 0.5458130922033573, "grad_norm": 1.2833635482665684, "learning_rate": 9.00141788041998e-06, "loss": 0.9636, "step": 2837 }, { "epoch": 0.546005483141744, "grad_norm": 0.7607429090325221, "learning_rate": 8.99521714246651e-06, "loss": 0.8444, "step": 2838 }, { "epoch": 0.5461978740801309, "grad_norm": 1.3202783039655046, "learning_rate": 8.989016794759127e-06, "loss": 0.9194, "step": 2839 }, { "epoch": 0.5463902650185176, "grad_norm": 1.3963609859935346, "learning_rate": 8.98281683970597e-06, "loss": 0.9352, "step": 2840 }, { "epoch": 0.5465826559569045, "grad_norm": 1.2847213486014348, "learning_rate": 8.97661727971503e-06, "loss": 1.0619, "step": 2841 }, { "epoch": 0.5467750468952912, "grad_norm": 1.3217423262012695, "learning_rate": 8.970418117194146e-06, "loss": 0.9097, "step": 2842 }, { "epoch": 0.5469674378336781, "grad_norm": 1.4235811958604436, "learning_rate": 8.964219354550999e-06, "loss": 1.0136, "step": 2843 }, { "epoch": 0.5471598287720648, "grad_norm": 1.385251616116583, "learning_rate": 8.958020994193124e-06, "loss": 0.9468, "step": 2844 }, { "epoch": 0.5473522197104517, "grad_norm": 1.5014382053299051, "learning_rate": 8.951823038527887e-06, "loss": 1.0418, "step": 2845 }, { "epoch": 0.5475446106488384, "grad_norm": 1.3841194806596773, "learning_rate": 8.945625489962503e-06, "loss": 1.0101, "step": 2846 }, { "epoch": 0.5477370015872253, "grad_norm": 1.4544637133150058, "learning_rate": 8.93942835090403e-06, "loss": 1.044, "step": 2847 }, { "epoch": 0.547929392525612, "grad_norm": 1.4261792148629233, "learning_rate": 8.933231623759365e-06, "loss": 0.9586, "step": 2848 }, { "epoch": 0.5481217834639989, "grad_norm": 1.3883545092123613, "learning_rate": 8.927035310935241e-06, "loss": 0.9307, "step": 2849 }, { "epoch": 0.5483141744023856, "grad_norm": 1.2743064644979811, "learning_rate": 8.920839414838243e-06, "loss": 0.9315, "step": 2850 }, { "epoch": 0.5485065653407725, "grad_norm": 1.5898378779164164, "learning_rate": 8.914643937874778e-06, "loss": 0.9928, "step": 2851 }, { "epoch": 0.5486989562791592, "grad_norm": 1.3490862649877295, "learning_rate": 8.908448882451104e-06, "loss": 0.9514, "step": 2852 }, { "epoch": 0.5488913472175461, "grad_norm": 1.24293354377101, "learning_rate": 8.902254250973306e-06, "loss": 0.9432, "step": 2853 }, { "epoch": 0.5490837381559328, "grad_norm": 1.4935016154318257, "learning_rate": 8.896060045847305e-06, "loss": 1.0009, "step": 2854 }, { "epoch": 0.5492761290943197, "grad_norm": 0.7400320289385627, "learning_rate": 8.88986626947886e-06, "loss": 0.8566, "step": 2855 }, { "epoch": 0.5494685200327064, "grad_norm": 1.4411122072165208, "learning_rate": 8.883672924273566e-06, "loss": 0.9564, "step": 2856 }, { "epoch": 0.5496609109710933, "grad_norm": 1.3859076205650098, "learning_rate": 8.877480012636847e-06, "loss": 0.9695, "step": 2857 }, { "epoch": 0.5498533019094801, "grad_norm": 1.4463831361905752, "learning_rate": 8.871287536973954e-06, "loss": 0.9529, "step": 2858 }, { "epoch": 0.5500456928478669, "grad_norm": 1.3842132524700002, "learning_rate": 8.865095499689978e-06, "loss": 0.9221, "step": 2859 }, { "epoch": 0.5502380837862537, "grad_norm": 1.3323623094383479, "learning_rate": 8.85890390318983e-06, "loss": 0.8983, "step": 2860 }, { "epoch": 0.5504304747246405, "grad_norm": 1.244992554909379, "learning_rate": 8.852712749878255e-06, "loss": 0.9153, "step": 2861 }, { "epoch": 0.5506228656630273, "grad_norm": 1.4126744248977612, "learning_rate": 8.846522042159833e-06, "loss": 1.058, "step": 2862 }, { "epoch": 0.5508152566014141, "grad_norm": 1.4229528062674275, "learning_rate": 8.840331782438954e-06, "loss": 0.9193, "step": 2863 }, { "epoch": 0.5510076475398009, "grad_norm": 1.259806700809623, "learning_rate": 8.83414197311985e-06, "loss": 0.8643, "step": 2864 }, { "epoch": 0.5512000384781877, "grad_norm": 1.341021570888399, "learning_rate": 8.82795261660657e-06, "loss": 0.9398, "step": 2865 }, { "epoch": 0.5513924294165745, "grad_norm": 1.4205031675523325, "learning_rate": 8.821763715302986e-06, "loss": 0.9485, "step": 2866 }, { "epoch": 0.5515848203549613, "grad_norm": 1.447025338675024, "learning_rate": 8.815575271612798e-06, "loss": 0.958, "step": 2867 }, { "epoch": 0.5517772112933481, "grad_norm": 1.3353490314335834, "learning_rate": 8.809387287939528e-06, "loss": 1.0012, "step": 2868 }, { "epoch": 0.5519696022317349, "grad_norm": 1.343750433970724, "learning_rate": 8.803199766686517e-06, "loss": 0.9495, "step": 2869 }, { "epoch": 0.5521619931701217, "grad_norm": 1.2928587476016409, "learning_rate": 8.797012710256923e-06, "loss": 1.0106, "step": 2870 }, { "epoch": 0.5523543841085085, "grad_norm": 1.4047544861499175, "learning_rate": 8.790826121053732e-06, "loss": 0.9811, "step": 2871 }, { "epoch": 0.5525467750468953, "grad_norm": 0.7918935790006756, "learning_rate": 8.784640001479741e-06, "loss": 0.8915, "step": 2872 }, { "epoch": 0.552739165985282, "grad_norm": 1.361054239352121, "learning_rate": 8.77845435393757e-06, "loss": 0.9141, "step": 2873 }, { "epoch": 0.5529315569236689, "grad_norm": 1.4531460499722229, "learning_rate": 8.772269180829653e-06, "loss": 0.9996, "step": 2874 }, { "epoch": 0.5531239478620557, "grad_norm": 1.319397434998939, "learning_rate": 8.766084484558237e-06, "loss": 0.9638, "step": 2875 }, { "epoch": 0.5533163388004425, "grad_norm": 1.382603099424192, "learning_rate": 8.759900267525393e-06, "loss": 0.9598, "step": 2876 }, { "epoch": 0.5535087297388293, "grad_norm": 1.3435578351419784, "learning_rate": 8.753716532132992e-06, "loss": 0.974, "step": 2877 }, { "epoch": 0.5537011206772161, "grad_norm": 1.3915677551862482, "learning_rate": 8.747533280782725e-06, "loss": 0.9414, "step": 2878 }, { "epoch": 0.5538935116156029, "grad_norm": 1.4557831089985858, "learning_rate": 8.741350515876103e-06, "loss": 0.9926, "step": 2879 }, { "epoch": 0.5540859025539897, "grad_norm": 1.4736450775584178, "learning_rate": 8.73516823981444e-06, "loss": 1.0045, "step": 2880 }, { "epoch": 0.5542782934923766, "grad_norm": 1.3084624215084977, "learning_rate": 8.728986454998858e-06, "loss": 0.9381, "step": 2881 }, { "epoch": 0.5544706844307633, "grad_norm": 1.419917248382354, "learning_rate": 8.72280516383029e-06, "loss": 0.9747, "step": 2882 }, { "epoch": 0.5546630753691502, "grad_norm": 1.2912076512543527, "learning_rate": 8.716624368709477e-06, "loss": 0.9837, "step": 2883 }, { "epoch": 0.5548554663075369, "grad_norm": 1.2889631775858572, "learning_rate": 8.71044407203697e-06, "loss": 0.8908, "step": 2884 }, { "epoch": 0.5550478572459238, "grad_norm": 1.4749509545084372, "learning_rate": 8.70426427621313e-06, "loss": 0.9772, "step": 2885 }, { "epoch": 0.5552402481843105, "grad_norm": 1.3808960234518568, "learning_rate": 8.698084983638111e-06, "loss": 0.9437, "step": 2886 }, { "epoch": 0.5554326391226974, "grad_norm": 1.4858510150739987, "learning_rate": 8.691906196711884e-06, "loss": 0.9107, "step": 2887 }, { "epoch": 0.5556250300610841, "grad_norm": 1.4172125690517658, "learning_rate": 8.685727917834218e-06, "loss": 0.9566, "step": 2888 }, { "epoch": 0.555817420999471, "grad_norm": 0.7104196455982649, "learning_rate": 8.679550149404685e-06, "loss": 0.8258, "step": 2889 }, { "epoch": 0.5560098119378577, "grad_norm": 1.3078173724467723, "learning_rate": 8.673372893822653e-06, "loss": 0.9562, "step": 2890 }, { "epoch": 0.5562022028762446, "grad_norm": 1.3320946962044546, "learning_rate": 8.667196153487308e-06, "loss": 0.9778, "step": 2891 }, { "epoch": 0.5563945938146313, "grad_norm": 1.4300741827855583, "learning_rate": 8.661019930797615e-06, "loss": 1.0192, "step": 2892 }, { "epoch": 0.5565869847530182, "grad_norm": 1.2868262658421714, "learning_rate": 8.654844228152355e-06, "loss": 0.9344, "step": 2893 }, { "epoch": 0.5567793756914049, "grad_norm": 1.3816768184116197, "learning_rate": 8.648669047950097e-06, "loss": 0.974, "step": 2894 }, { "epoch": 0.5569717666297918, "grad_norm": 1.3862070727521916, "learning_rate": 8.642494392589206e-06, "loss": 0.9603, "step": 2895 }, { "epoch": 0.5571641575681785, "grad_norm": 1.3654105696947632, "learning_rate": 8.63632026446785e-06, "loss": 0.9674, "step": 2896 }, { "epoch": 0.5573565485065654, "grad_norm": 1.4394829438292305, "learning_rate": 8.630146665983993e-06, "loss": 1.0143, "step": 2897 }, { "epoch": 0.5575489394449521, "grad_norm": 1.5334675620570337, "learning_rate": 8.623973599535385e-06, "loss": 0.9813, "step": 2898 }, { "epoch": 0.557741330383339, "grad_norm": 1.4271058477289997, "learning_rate": 8.617801067519575e-06, "loss": 0.9421, "step": 2899 }, { "epoch": 0.5579337213217257, "grad_norm": 1.4097659967199425, "learning_rate": 8.611629072333905e-06, "loss": 0.9406, "step": 2900 }, { "epoch": 0.5581261122601126, "grad_norm": 1.4713200845879988, "learning_rate": 8.605457616375503e-06, "loss": 0.9539, "step": 2901 }, { "epoch": 0.5583185031984994, "grad_norm": 1.5070556632587406, "learning_rate": 8.599286702041292e-06, "loss": 1.0992, "step": 2902 }, { "epoch": 0.5585108941368861, "grad_norm": 1.4299590998543803, "learning_rate": 8.593116331727987e-06, "loss": 1.0242, "step": 2903 }, { "epoch": 0.558703285075273, "grad_norm": 1.4118317820953654, "learning_rate": 8.586946507832088e-06, "loss": 1.0363, "step": 2904 }, { "epoch": 0.5588956760136597, "grad_norm": 1.2910893226211064, "learning_rate": 8.580777232749883e-06, "loss": 1.007, "step": 2905 }, { "epoch": 0.5590880669520466, "grad_norm": 1.3297934038765247, "learning_rate": 8.574608508877448e-06, "loss": 0.9634, "step": 2906 }, { "epoch": 0.5592804578904333, "grad_norm": 1.3632729038555742, "learning_rate": 8.568440338610638e-06, "loss": 0.9662, "step": 2907 }, { "epoch": 0.5594728488288202, "grad_norm": 1.3191551675736624, "learning_rate": 8.562272724345108e-06, "loss": 1.0139, "step": 2908 }, { "epoch": 0.559665239767207, "grad_norm": 1.3718851224604474, "learning_rate": 8.556105668476287e-06, "loss": 0.9374, "step": 2909 }, { "epoch": 0.5598576307055938, "grad_norm": 1.3518721025474567, "learning_rate": 8.549939173399385e-06, "loss": 0.9175, "step": 2910 }, { "epoch": 0.5600500216439805, "grad_norm": 1.3499458950522458, "learning_rate": 8.5437732415094e-06, "loss": 0.9213, "step": 2911 }, { "epoch": 0.5602424125823674, "grad_norm": 1.3304798341418747, "learning_rate": 8.537607875201106e-06, "loss": 0.9779, "step": 2912 }, { "epoch": 0.5604348035207541, "grad_norm": 1.4037078866101438, "learning_rate": 8.531443076869058e-06, "loss": 0.9745, "step": 2913 }, { "epoch": 0.560627194459141, "grad_norm": 1.3055893080592804, "learning_rate": 8.525278848907603e-06, "loss": 0.9602, "step": 2914 }, { "epoch": 0.5608195853975277, "grad_norm": 1.4352236901387991, "learning_rate": 8.51911519371085e-06, "loss": 1.0229, "step": 2915 }, { "epoch": 0.5610119763359146, "grad_norm": 1.5267748173849633, "learning_rate": 8.512952113672689e-06, "loss": 1.0828, "step": 2916 }, { "epoch": 0.5612043672743013, "grad_norm": 1.3890743668826673, "learning_rate": 8.506789611186794e-06, "loss": 0.9571, "step": 2917 }, { "epoch": 0.5613967582126882, "grad_norm": 1.4718772908169944, "learning_rate": 8.500627688646607e-06, "loss": 1.0287, "step": 2918 }, { "epoch": 0.5615891491510749, "grad_norm": 1.3094581009743946, "learning_rate": 8.494466348445345e-06, "loss": 1.0277, "step": 2919 }, { "epoch": 0.5617815400894618, "grad_norm": 1.5291342047358614, "learning_rate": 8.48830559297601e-06, "loss": 0.9494, "step": 2920 }, { "epoch": 0.5619739310278485, "grad_norm": 1.3974583173557313, "learning_rate": 8.48214542463136e-06, "loss": 1.0231, "step": 2921 }, { "epoch": 0.5621663219662354, "grad_norm": 1.3903588500684134, "learning_rate": 8.475985845803938e-06, "loss": 0.9987, "step": 2922 }, { "epoch": 0.5623587129046221, "grad_norm": 1.453802297467652, "learning_rate": 8.469826858886054e-06, "loss": 0.996, "step": 2923 }, { "epoch": 0.562551103843009, "grad_norm": 1.355428502912445, "learning_rate": 8.463668466269785e-06, "loss": 0.949, "step": 2924 }, { "epoch": 0.5627434947813958, "grad_norm": 1.423792730348024, "learning_rate": 8.457510670346976e-06, "loss": 1.0544, "step": 2925 }, { "epoch": 0.5629358857197826, "grad_norm": 1.2999498801328366, "learning_rate": 8.451353473509254e-06, "loss": 0.9194, "step": 2926 }, { "epoch": 0.5631282766581694, "grad_norm": 1.3536173442675672, "learning_rate": 8.445196878147997e-06, "loss": 0.9811, "step": 2927 }, { "epoch": 0.5633206675965562, "grad_norm": 1.36793095968629, "learning_rate": 8.439040886654354e-06, "loss": 0.9922, "step": 2928 }, { "epoch": 0.563513058534943, "grad_norm": 1.4210010680692085, "learning_rate": 8.432885501419248e-06, "loss": 0.9421, "step": 2929 }, { "epoch": 0.5637054494733298, "grad_norm": 1.4193493480913277, "learning_rate": 8.426730724833354e-06, "loss": 0.9912, "step": 2930 }, { "epoch": 0.5638978404117166, "grad_norm": 1.3413134023516264, "learning_rate": 8.420576559287112e-06, "loss": 0.9736, "step": 2931 }, { "epoch": 0.5640902313501034, "grad_norm": 1.4207161727311464, "learning_rate": 8.414423007170742e-06, "loss": 1.0304, "step": 2932 }, { "epoch": 0.5642826222884902, "grad_norm": 1.3959687427210212, "learning_rate": 8.408270070874201e-06, "loss": 1.0026, "step": 2933 }, { "epoch": 0.564475013226877, "grad_norm": 1.3249544595258291, "learning_rate": 8.402117752787225e-06, "loss": 0.9179, "step": 2934 }, { "epoch": 0.5646674041652638, "grad_norm": 1.387256713496102, "learning_rate": 8.395966055299302e-06, "loss": 0.9504, "step": 2935 }, { "epoch": 0.5648597951036506, "grad_norm": 1.284450932999201, "learning_rate": 8.389814980799679e-06, "loss": 0.9461, "step": 2936 }, { "epoch": 0.5650521860420374, "grad_norm": 1.328898603987695, "learning_rate": 8.38366453167736e-06, "loss": 0.977, "step": 2937 }, { "epoch": 0.5652445769804242, "grad_norm": 1.4701641652798008, "learning_rate": 8.377514710321117e-06, "loss": 0.8959, "step": 2938 }, { "epoch": 0.565436967918811, "grad_norm": 1.4023792161354967, "learning_rate": 8.371365519119463e-06, "loss": 0.9067, "step": 2939 }, { "epoch": 0.5656293588571978, "grad_norm": 1.4051310826675207, "learning_rate": 8.365216960460675e-06, "loss": 0.9612, "step": 2940 }, { "epoch": 0.5658217497955846, "grad_norm": 1.3529682114870665, "learning_rate": 8.359069036732781e-06, "loss": 0.9607, "step": 2941 }, { "epoch": 0.5660141407339714, "grad_norm": 1.4602227398333905, "learning_rate": 8.352921750323562e-06, "loss": 0.9928, "step": 2942 }, { "epoch": 0.5662065316723582, "grad_norm": 1.3268715106233853, "learning_rate": 8.346775103620559e-06, "loss": 0.9656, "step": 2943 }, { "epoch": 0.566398922610745, "grad_norm": 1.4366181248938898, "learning_rate": 8.340629099011057e-06, "loss": 1.0177, "step": 2944 }, { "epoch": 0.5665913135491318, "grad_norm": 1.3823573086153231, "learning_rate": 8.33448373888209e-06, "loss": 0.9134, "step": 2945 }, { "epoch": 0.5667837044875187, "grad_norm": 1.4631757433432182, "learning_rate": 8.328339025620449e-06, "loss": 0.9269, "step": 2946 }, { "epoch": 0.5669760954259054, "grad_norm": 1.5197324071946683, "learning_rate": 8.322194961612668e-06, "loss": 0.9483, "step": 2947 }, { "epoch": 0.5671684863642923, "grad_norm": 1.287541443505935, "learning_rate": 8.316051549245026e-06, "loss": 0.9581, "step": 2948 }, { "epoch": 0.567360877302679, "grad_norm": 1.374232303700878, "learning_rate": 8.309908790903562e-06, "loss": 0.924, "step": 2949 }, { "epoch": 0.5675532682410659, "grad_norm": 1.3312476112974918, "learning_rate": 8.303766688974047e-06, "loss": 0.9191, "step": 2950 }, { "epoch": 0.5677456591794526, "grad_norm": 1.4764057424978452, "learning_rate": 8.297625245842006e-06, "loss": 0.9705, "step": 2951 }, { "epoch": 0.5679380501178395, "grad_norm": 1.3183407289137028, "learning_rate": 8.291484463892703e-06, "loss": 0.9951, "step": 2952 }, { "epoch": 0.5681304410562262, "grad_norm": 1.3659051689515105, "learning_rate": 8.285344345511147e-06, "loss": 0.9933, "step": 2953 }, { "epoch": 0.5683228319946131, "grad_norm": 1.3790299532557675, "learning_rate": 8.279204893082083e-06, "loss": 1.0105, "step": 2954 }, { "epoch": 0.5685152229329998, "grad_norm": 1.34914374698744, "learning_rate": 8.273066108990017e-06, "loss": 0.972, "step": 2955 }, { "epoch": 0.5687076138713867, "grad_norm": 1.6054731555770432, "learning_rate": 8.266927995619175e-06, "loss": 0.9895, "step": 2956 }, { "epoch": 0.5689000048097734, "grad_norm": 1.3864696671587955, "learning_rate": 8.260790555353526e-06, "loss": 1.0152, "step": 2957 }, { "epoch": 0.5690923957481603, "grad_norm": 1.3887324390308162, "learning_rate": 8.254653790576787e-06, "loss": 0.9717, "step": 2958 }, { "epoch": 0.569284786686547, "grad_norm": 1.3954013639366822, "learning_rate": 8.248517703672405e-06, "loss": 1.0183, "step": 2959 }, { "epoch": 0.5694771776249339, "grad_norm": 1.3962013980965016, "learning_rate": 8.242382297023558e-06, "loss": 0.9314, "step": 2960 }, { "epoch": 0.5696695685633206, "grad_norm": 1.3205398340021248, "learning_rate": 8.23624757301318e-06, "loss": 0.9507, "step": 2961 }, { "epoch": 0.5698619595017075, "grad_norm": 1.4615532770204247, "learning_rate": 8.230113534023917e-06, "loss": 1.0237, "step": 2962 }, { "epoch": 0.5700543504400942, "grad_norm": 1.326881245584423, "learning_rate": 8.223980182438167e-06, "loss": 0.9591, "step": 2963 }, { "epoch": 0.5702467413784811, "grad_norm": 1.354323251660473, "learning_rate": 8.217847520638049e-06, "loss": 0.9528, "step": 2964 }, { "epoch": 0.5704391323168678, "grad_norm": 1.5215035778162498, "learning_rate": 8.211715551005414e-06, "loss": 1.0482, "step": 2965 }, { "epoch": 0.5706315232552547, "grad_norm": 1.3215733156244178, "learning_rate": 8.205584275921854e-06, "loss": 0.9273, "step": 2966 }, { "epoch": 0.5708239141936414, "grad_norm": 1.3882075368439453, "learning_rate": 8.199453697768686e-06, "loss": 0.9625, "step": 2967 }, { "epoch": 0.5710163051320283, "grad_norm": 1.3997566475167893, "learning_rate": 8.193323818926955e-06, "loss": 1.0486, "step": 2968 }, { "epoch": 0.5712086960704151, "grad_norm": 0.7776095487080199, "learning_rate": 8.187194641777432e-06, "loss": 0.8601, "step": 2969 }, { "epoch": 0.5714010870088019, "grad_norm": 1.4309301061108937, "learning_rate": 8.181066168700622e-06, "loss": 0.963, "step": 2970 }, { "epoch": 0.5715934779471887, "grad_norm": 1.3927544980333946, "learning_rate": 8.174938402076754e-06, "loss": 1.0046, "step": 2971 }, { "epoch": 0.5717858688855755, "grad_norm": 1.2343335316990074, "learning_rate": 8.168811344285776e-06, "loss": 0.9098, "step": 2972 }, { "epoch": 0.5719782598239623, "grad_norm": 1.2849155855737362, "learning_rate": 8.162684997707374e-06, "loss": 0.9392, "step": 2973 }, { "epoch": 0.5721706507623491, "grad_norm": 1.386831745591964, "learning_rate": 8.156559364720947e-06, "loss": 0.9586, "step": 2974 }, { "epoch": 0.5723630417007359, "grad_norm": 1.4676756340477672, "learning_rate": 8.150434447705623e-06, "loss": 0.9776, "step": 2975 }, { "epoch": 0.5725554326391227, "grad_norm": 1.3916914919926717, "learning_rate": 8.144310249040246e-06, "loss": 0.9385, "step": 2976 }, { "epoch": 0.5727478235775095, "grad_norm": 1.3437119208450636, "learning_rate": 8.138186771103382e-06, "loss": 0.9667, "step": 2977 }, { "epoch": 0.5729402145158963, "grad_norm": 1.336091987157466, "learning_rate": 8.132064016273325e-06, "loss": 0.9347, "step": 2978 }, { "epoch": 0.5731326054542831, "grad_norm": 1.4813534129424888, "learning_rate": 8.12594198692808e-06, "loss": 0.9804, "step": 2979 }, { "epoch": 0.5733249963926699, "grad_norm": 1.2584598919479972, "learning_rate": 8.119820685445373e-06, "loss": 0.9155, "step": 2980 }, { "epoch": 0.5735173873310567, "grad_norm": 1.3068293746434396, "learning_rate": 8.113700114202647e-06, "loss": 0.9544, "step": 2981 }, { "epoch": 0.5737097782694435, "grad_norm": 1.4082646502402727, "learning_rate": 8.107580275577059e-06, "loss": 0.9138, "step": 2982 }, { "epoch": 0.5739021692078303, "grad_norm": 1.460502011277777, "learning_rate": 8.101461171945483e-06, "loss": 0.8955, "step": 2983 }, { "epoch": 0.5740945601462171, "grad_norm": 1.2737071651007952, "learning_rate": 8.095342805684516e-06, "loss": 0.9426, "step": 2984 }, { "epoch": 0.5742869510846039, "grad_norm": 1.4309817120599362, "learning_rate": 8.089225179170454e-06, "loss": 0.9917, "step": 2985 }, { "epoch": 0.5744793420229907, "grad_norm": 0.7480468706421889, "learning_rate": 8.083108294779313e-06, "loss": 0.8265, "step": 2986 }, { "epoch": 0.5746717329613775, "grad_norm": 1.4051879450795057, "learning_rate": 8.076992154886826e-06, "loss": 0.9776, "step": 2987 }, { "epoch": 0.5748641238997643, "grad_norm": 1.4069046641044538, "learning_rate": 8.070876761868426e-06, "loss": 0.9606, "step": 2988 }, { "epoch": 0.5750565148381511, "grad_norm": 1.324891943374013, "learning_rate": 8.064762118099258e-06, "loss": 0.923, "step": 2989 }, { "epoch": 0.575248905776538, "grad_norm": 0.7607152630155444, "learning_rate": 8.058648225954188e-06, "loss": 0.8942, "step": 2990 }, { "epoch": 0.5754412967149247, "grad_norm": 0.7513757668896182, "learning_rate": 8.052535087807774e-06, "loss": 0.8622, "step": 2991 }, { "epoch": 0.5756336876533116, "grad_norm": 1.3860035111439435, "learning_rate": 8.046422706034294e-06, "loss": 0.8983, "step": 2992 }, { "epoch": 0.5758260785916983, "grad_norm": 1.4342084031950526, "learning_rate": 8.040311083007725e-06, "loss": 0.9366, "step": 2993 }, { "epoch": 0.5760184695300852, "grad_norm": 1.3288714233741958, "learning_rate": 8.034200221101746e-06, "loss": 0.9587, "step": 2994 }, { "epoch": 0.5762108604684719, "grad_norm": 1.4089208114068998, "learning_rate": 8.028090122689747e-06, "loss": 1.0207, "step": 2995 }, { "epoch": 0.5764032514068588, "grad_norm": 1.2972843431685745, "learning_rate": 8.021980790144828e-06, "loss": 0.9687, "step": 2996 }, { "epoch": 0.5765956423452455, "grad_norm": 1.372256506441452, "learning_rate": 8.015872225839776e-06, "loss": 0.9537, "step": 2997 }, { "epoch": 0.5767880332836324, "grad_norm": 1.3780654881545698, "learning_rate": 8.009764432147086e-06, "loss": 0.9747, "step": 2998 }, { "epoch": 0.5769804242220191, "grad_norm": 0.7280058411299929, "learning_rate": 8.003657411438961e-06, "loss": 0.874, "step": 2999 }, { "epoch": 0.577172815160406, "grad_norm": 1.6020770239000086, "learning_rate": 7.997551166087293e-06, "loss": 0.9056, "step": 3000 }, { "epoch": 0.5773652060987927, "grad_norm": 1.3312794939995007, "learning_rate": 7.991445698463672e-06, "loss": 0.9648, "step": 3001 }, { "epoch": 0.5775575970371796, "grad_norm": 1.2900124305039966, "learning_rate": 7.985341010939402e-06, "loss": 0.8839, "step": 3002 }, { "epoch": 0.5777499879755663, "grad_norm": 0.73765243995714, "learning_rate": 7.979237105885467e-06, "loss": 0.8586, "step": 3003 }, { "epoch": 0.5779423789139532, "grad_norm": 1.5573176993957643, "learning_rate": 7.973133985672558e-06, "loss": 1.0642, "step": 3004 }, { "epoch": 0.5781347698523399, "grad_norm": 1.3719039853746675, "learning_rate": 7.967031652671051e-06, "loss": 0.9517, "step": 3005 }, { "epoch": 0.5783271607907268, "grad_norm": 1.59667197012869, "learning_rate": 7.960930109251023e-06, "loss": 0.9673, "step": 3006 }, { "epoch": 0.5785195517291135, "grad_norm": 1.5548899322948173, "learning_rate": 7.954829357782243e-06, "loss": 0.9615, "step": 3007 }, { "epoch": 0.5787119426675004, "grad_norm": 1.357273896613859, "learning_rate": 7.948729400634178e-06, "loss": 0.9984, "step": 3008 }, { "epoch": 0.5789043336058871, "grad_norm": 1.3144583698470222, "learning_rate": 7.942630240175977e-06, "loss": 0.962, "step": 3009 }, { "epoch": 0.579096724544274, "grad_norm": 1.3508762185229373, "learning_rate": 7.936531878776484e-06, "loss": 0.9731, "step": 3010 }, { "epoch": 0.5792891154826608, "grad_norm": 1.41512663530986, "learning_rate": 7.930434318804229e-06, "loss": 0.9933, "step": 3011 }, { "epoch": 0.5794815064210476, "grad_norm": 1.250104182058119, "learning_rate": 7.924337562627435e-06, "loss": 0.983, "step": 3012 }, { "epoch": 0.5796738973594344, "grad_norm": 1.454098926575833, "learning_rate": 7.918241612614016e-06, "loss": 0.9637, "step": 3013 }, { "epoch": 0.5798662882978212, "grad_norm": 1.3184115645464967, "learning_rate": 7.91214647113157e-06, "loss": 0.9934, "step": 3014 }, { "epoch": 0.580058679236208, "grad_norm": 1.3405354506411729, "learning_rate": 7.906052140547373e-06, "loss": 0.9433, "step": 3015 }, { "epoch": 0.5802510701745948, "grad_norm": 1.323731992672675, "learning_rate": 7.899958623228398e-06, "loss": 0.9644, "step": 3016 }, { "epoch": 0.5804434611129816, "grad_norm": 1.3364404851614453, "learning_rate": 7.893865921541294e-06, "loss": 1.0116, "step": 3017 }, { "epoch": 0.5806358520513684, "grad_norm": 1.4681268078074523, "learning_rate": 7.887774037852395e-06, "loss": 1.0235, "step": 3018 }, { "epoch": 0.5808282429897552, "grad_norm": 1.4212125026448599, "learning_rate": 7.881682974527723e-06, "loss": 1.057, "step": 3019 }, { "epoch": 0.581020633928142, "grad_norm": 1.3842695545029284, "learning_rate": 7.875592733932972e-06, "loss": 0.9791, "step": 3020 }, { "epoch": 0.5812130248665288, "grad_norm": 1.3280857531065982, "learning_rate": 7.869503318433529e-06, "loss": 0.9934, "step": 3021 }, { "epoch": 0.5814054158049156, "grad_norm": 1.5221762708137518, "learning_rate": 7.863414730394444e-06, "loss": 1.0836, "step": 3022 }, { "epoch": 0.5815978067433024, "grad_norm": 1.383646674368079, "learning_rate": 7.857326972180455e-06, "loss": 0.9121, "step": 3023 }, { "epoch": 0.5817901976816892, "grad_norm": 1.233145532152077, "learning_rate": 7.85124004615598e-06, "loss": 0.8939, "step": 3024 }, { "epoch": 0.581982588620076, "grad_norm": 1.2745719438741636, "learning_rate": 7.845153954685114e-06, "loss": 0.8984, "step": 3025 }, { "epoch": 0.5821749795584628, "grad_norm": 1.3991392731603576, "learning_rate": 7.839068700131623e-06, "loss": 0.9693, "step": 3026 }, { "epoch": 0.5823673704968496, "grad_norm": 1.3600362528955092, "learning_rate": 7.832984284858946e-06, "loss": 0.948, "step": 3027 }, { "epoch": 0.5825597614352364, "grad_norm": 1.3519347069262482, "learning_rate": 7.826900711230204e-06, "loss": 0.9254, "step": 3028 }, { "epoch": 0.5827521523736232, "grad_norm": 1.3762946917463605, "learning_rate": 7.820817981608185e-06, "loss": 1.0178, "step": 3029 }, { "epoch": 0.58294454331201, "grad_norm": 1.4626053583665042, "learning_rate": 7.814736098355348e-06, "loss": 0.9668, "step": 3030 }, { "epoch": 0.5831369342503968, "grad_norm": 0.7400968203670436, "learning_rate": 7.808655063833833e-06, "loss": 0.7957, "step": 3031 }, { "epoch": 0.5833293251887836, "grad_norm": 1.5118769063397945, "learning_rate": 7.802574880405438e-06, "loss": 0.964, "step": 3032 }, { "epoch": 0.5835217161271704, "grad_norm": 1.346398997912494, "learning_rate": 7.79649555043164e-06, "loss": 0.9717, "step": 3033 }, { "epoch": 0.5837141070655573, "grad_norm": 1.4706078670318372, "learning_rate": 7.790417076273581e-06, "loss": 1.0289, "step": 3034 }, { "epoch": 0.583906498003944, "grad_norm": 0.7114337671478648, "learning_rate": 7.784339460292065e-06, "loss": 0.8817, "step": 3035 }, { "epoch": 0.5840988889423309, "grad_norm": 1.3559861339295036, "learning_rate": 7.77826270484757e-06, "loss": 0.9648, "step": 3036 }, { "epoch": 0.5842912798807176, "grad_norm": 1.2891042576886218, "learning_rate": 7.772186812300244e-06, "loss": 1.0002, "step": 3037 }, { "epoch": 0.5844836708191045, "grad_norm": 1.4597286335494764, "learning_rate": 7.766111785009888e-06, "loss": 1.0111, "step": 3038 }, { "epoch": 0.5846760617574912, "grad_norm": 1.357212518126331, "learning_rate": 7.760037625335973e-06, "loss": 0.867, "step": 3039 }, { "epoch": 0.5848684526958781, "grad_norm": 1.3387706788705684, "learning_rate": 7.753964335637634e-06, "loss": 1.0397, "step": 3040 }, { "epoch": 0.5850608436342648, "grad_norm": 1.5231948355844196, "learning_rate": 7.747891918273668e-06, "loss": 0.9601, "step": 3041 }, { "epoch": 0.5852532345726517, "grad_norm": 1.3860888518905607, "learning_rate": 7.741820375602524e-06, "loss": 1.0015, "step": 3042 }, { "epoch": 0.5854456255110384, "grad_norm": 1.3292904091869362, "learning_rate": 7.735749709982329e-06, "loss": 0.8993, "step": 3043 }, { "epoch": 0.5856380164494253, "grad_norm": 1.3618835905513869, "learning_rate": 7.729679923770855e-06, "loss": 0.9675, "step": 3044 }, { "epoch": 0.585830407387812, "grad_norm": 1.540493660772044, "learning_rate": 7.723611019325538e-06, "loss": 0.9889, "step": 3045 }, { "epoch": 0.5860227983261989, "grad_norm": 1.3086009729807078, "learning_rate": 7.71754299900347e-06, "loss": 0.99, "step": 3046 }, { "epoch": 0.5862151892645856, "grad_norm": 1.449543363902747, "learning_rate": 7.7114758651614e-06, "loss": 0.8863, "step": 3047 }, { "epoch": 0.5864075802029725, "grad_norm": 1.2375928906758165, "learning_rate": 7.705409620155733e-06, "loss": 0.9897, "step": 3048 }, { "epoch": 0.5865999711413592, "grad_norm": 1.5543181979922875, "learning_rate": 7.699344266342529e-06, "loss": 0.9784, "step": 3049 }, { "epoch": 0.586792362079746, "grad_norm": 1.420413551839509, "learning_rate": 7.693279806077504e-06, "loss": 0.9537, "step": 3050 }, { "epoch": 0.5869847530181328, "grad_norm": 1.4368872628441203, "learning_rate": 7.68721624171602e-06, "loss": 0.923, "step": 3051 }, { "epoch": 0.5871771439565197, "grad_norm": 1.4929306397437672, "learning_rate": 7.681153575613098e-06, "loss": 1.0439, "step": 3052 }, { "epoch": 0.5873695348949064, "grad_norm": 1.4705183177133339, "learning_rate": 7.675091810123404e-06, "loss": 1.0142, "step": 3053 }, { "epoch": 0.5875619258332933, "grad_norm": 1.5165822419247497, "learning_rate": 7.669030947601265e-06, "loss": 0.9465, "step": 3054 }, { "epoch": 0.5877543167716801, "grad_norm": 1.3919811617464695, "learning_rate": 7.662970990400647e-06, "loss": 0.9956, "step": 3055 }, { "epoch": 0.5879467077100669, "grad_norm": 1.4026415118701028, "learning_rate": 7.656911940875163e-06, "loss": 0.9921, "step": 3056 }, { "epoch": 0.5881390986484537, "grad_norm": 0.7256391014718786, "learning_rate": 7.650853801378084e-06, "loss": 0.8706, "step": 3057 }, { "epoch": 0.5883314895868405, "grad_norm": 1.5757986701616893, "learning_rate": 7.644796574262322e-06, "loss": 0.9739, "step": 3058 }, { "epoch": 0.5885238805252273, "grad_norm": 1.3638110980972198, "learning_rate": 7.638740261880423e-06, "loss": 0.9658, "step": 3059 }, { "epoch": 0.588716271463614, "grad_norm": 1.4095692963324211, "learning_rate": 7.632684866584606e-06, "loss": 0.9555, "step": 3060 }, { "epoch": 0.5889086624020009, "grad_norm": 1.376990405305991, "learning_rate": 7.626630390726704e-06, "loss": 0.9551, "step": 3061 }, { "epoch": 0.5891010533403876, "grad_norm": 1.2747157937758866, "learning_rate": 7.620576836658212e-06, "loss": 0.9329, "step": 3062 }, { "epoch": 0.5892934442787745, "grad_norm": 1.290819959524509, "learning_rate": 7.61452420673026e-06, "loss": 0.9804, "step": 3063 }, { "epoch": 0.5894858352171612, "grad_norm": 1.4053059931459424, "learning_rate": 7.608472503293615e-06, "loss": 1.0091, "step": 3064 }, { "epoch": 0.5896782261555481, "grad_norm": 1.3950833179556357, "learning_rate": 7.60242172869869e-06, "loss": 0.9973, "step": 3065 }, { "epoch": 0.5898706170939348, "grad_norm": 1.4989718539384205, "learning_rate": 7.596371885295542e-06, "loss": 0.9348, "step": 3066 }, { "epoch": 0.5900630080323217, "grad_norm": 1.4498759963663457, "learning_rate": 7.590322975433857e-06, "loss": 1.0168, "step": 3067 }, { "epoch": 0.5902553989707084, "grad_norm": 1.5676531492837682, "learning_rate": 7.584275001462961e-06, "loss": 1.0483, "step": 3068 }, { "epoch": 0.5904477899090953, "grad_norm": 1.474089139685289, "learning_rate": 7.578227965731819e-06, "loss": 1.0102, "step": 3069 }, { "epoch": 0.590640180847482, "grad_norm": 1.4367734911717982, "learning_rate": 7.572181870589028e-06, "loss": 1.0035, "step": 3070 }, { "epoch": 0.5908325717858689, "grad_norm": 1.3300323525132296, "learning_rate": 7.566136718382821e-06, "loss": 0.9516, "step": 3071 }, { "epoch": 0.5910249627242556, "grad_norm": 1.3520628913791237, "learning_rate": 7.560092511461069e-06, "loss": 0.9575, "step": 3072 }, { "epoch": 0.5912173536626425, "grad_norm": 0.7071973736985792, "learning_rate": 7.55404925217127e-06, "loss": 0.8056, "step": 3073 }, { "epoch": 0.5914097446010292, "grad_norm": 1.2958457352871944, "learning_rate": 7.548006942860557e-06, "loss": 0.9603, "step": 3074 }, { "epoch": 0.5916021355394161, "grad_norm": 1.2671736381617433, "learning_rate": 7.541965585875695e-06, "loss": 1.0292, "step": 3075 }, { "epoch": 0.5917945264778028, "grad_norm": 1.4941768454121214, "learning_rate": 7.535925183563073e-06, "loss": 0.9914, "step": 3076 }, { "epoch": 0.5919869174161897, "grad_norm": 1.6504471816491222, "learning_rate": 7.529885738268714e-06, "loss": 0.9178, "step": 3077 }, { "epoch": 0.5921793083545766, "grad_norm": 1.4028027891155492, "learning_rate": 7.523847252338274e-06, "loss": 0.9334, "step": 3078 }, { "epoch": 0.5923716992929633, "grad_norm": 1.4088675685091705, "learning_rate": 7.51780972811703e-06, "loss": 0.9168, "step": 3079 }, { "epoch": 0.5925640902313501, "grad_norm": 1.3949202404416274, "learning_rate": 7.511773167949885e-06, "loss": 0.9795, "step": 3080 }, { "epoch": 0.5927564811697369, "grad_norm": 1.3652895173755344, "learning_rate": 7.5057375741813685e-06, "loss": 0.9671, "step": 3081 }, { "epoch": 0.5929488721081237, "grad_norm": 1.5253617798380708, "learning_rate": 7.499702949155634e-06, "loss": 0.9222, "step": 3082 }, { "epoch": 0.5931412630465105, "grad_norm": 1.468456486091616, "learning_rate": 7.493669295216467e-06, "loss": 0.9644, "step": 3083 }, { "epoch": 0.5933336539848973, "grad_norm": 1.3213809925966342, "learning_rate": 7.487636614707265e-06, "loss": 0.9596, "step": 3084 }, { "epoch": 0.5935260449232841, "grad_norm": 1.3490828102337378, "learning_rate": 7.48160490997105e-06, "loss": 0.9751, "step": 3085 }, { "epoch": 0.593718435861671, "grad_norm": 1.4017000401618402, "learning_rate": 7.475574183350471e-06, "loss": 0.9413, "step": 3086 }, { "epoch": 0.5939108268000577, "grad_norm": 1.4638996229670547, "learning_rate": 7.46954443718779e-06, "loss": 1.0791, "step": 3087 }, { "epoch": 0.5941032177384445, "grad_norm": 1.5312693069616228, "learning_rate": 7.463515673824888e-06, "loss": 0.9688, "step": 3088 }, { "epoch": 0.5942956086768313, "grad_norm": 1.4831159517771761, "learning_rate": 7.457487895603273e-06, "loss": 0.9971, "step": 3089 }, { "epoch": 0.5944879996152181, "grad_norm": 1.358679144115142, "learning_rate": 7.451461104864061e-06, "loss": 0.9469, "step": 3090 }, { "epoch": 0.5946803905536049, "grad_norm": 1.2390434033621132, "learning_rate": 7.44543530394799e-06, "loss": 0.882, "step": 3091 }, { "epoch": 0.5948727814919917, "grad_norm": 1.3460610037271072, "learning_rate": 7.439410495195411e-06, "loss": 0.9863, "step": 3092 }, { "epoch": 0.5950651724303785, "grad_norm": 1.4537344462670512, "learning_rate": 7.433386680946288e-06, "loss": 0.9562, "step": 3093 }, { "epoch": 0.5952575633687653, "grad_norm": 1.3339860359304414, "learning_rate": 7.427363863540202e-06, "loss": 0.9288, "step": 3094 }, { "epoch": 0.5954499543071521, "grad_norm": 1.4064205990474692, "learning_rate": 7.421342045316351e-06, "loss": 0.9761, "step": 3095 }, { "epoch": 0.5956423452455389, "grad_norm": 1.453220084987934, "learning_rate": 7.415321228613534e-06, "loss": 1.0843, "step": 3096 }, { "epoch": 0.5958347361839257, "grad_norm": 1.4754258271715526, "learning_rate": 7.409301415770168e-06, "loss": 0.9408, "step": 3097 }, { "epoch": 0.5960271271223125, "grad_norm": 1.2738582955443165, "learning_rate": 7.403282609124281e-06, "loss": 0.9189, "step": 3098 }, { "epoch": 0.5962195180606994, "grad_norm": 0.7192958654143885, "learning_rate": 7.397264811013507e-06, "loss": 0.8487, "step": 3099 }, { "epoch": 0.5964119089990861, "grad_norm": 1.4531869492460343, "learning_rate": 7.391248023775084e-06, "loss": 0.9691, "step": 3100 }, { "epoch": 0.596604299937473, "grad_norm": 1.3431043214892058, "learning_rate": 7.385232249745873e-06, "loss": 0.966, "step": 3101 }, { "epoch": 0.5967966908758597, "grad_norm": 1.3725777132834538, "learning_rate": 7.379217491262325e-06, "loss": 0.9023, "step": 3102 }, { "epoch": 0.5969890818142466, "grad_norm": 1.3765501650341108, "learning_rate": 7.373203750660505e-06, "loss": 1.0047, "step": 3103 }, { "epoch": 0.5971814727526333, "grad_norm": 1.2764170191492106, "learning_rate": 7.36719103027608e-06, "loss": 0.8919, "step": 3104 }, { "epoch": 0.5973738636910202, "grad_norm": 1.4022305045998145, "learning_rate": 7.361179332444318e-06, "loss": 0.9454, "step": 3105 }, { "epoch": 0.5975662546294069, "grad_norm": 1.3725995304584169, "learning_rate": 7.355168659500094e-06, "loss": 0.9768, "step": 3106 }, { "epoch": 0.5977586455677938, "grad_norm": 1.3502608562274232, "learning_rate": 7.3491590137778915e-06, "loss": 0.9728, "step": 3107 }, { "epoch": 0.5979510365061805, "grad_norm": 1.4003691779943497, "learning_rate": 7.343150397611782e-06, "loss": 0.9079, "step": 3108 }, { "epoch": 0.5981434274445674, "grad_norm": 1.392030165200933, "learning_rate": 7.3371428133354435e-06, "loss": 0.9527, "step": 3109 }, { "epoch": 0.5983358183829541, "grad_norm": 1.3539865334678334, "learning_rate": 7.33113626328215e-06, "loss": 0.9004, "step": 3110 }, { "epoch": 0.598528209321341, "grad_norm": 1.3866005371728634, "learning_rate": 7.325130749784781e-06, "loss": 0.9213, "step": 3111 }, { "epoch": 0.5987206002597277, "grad_norm": 1.2989663056857184, "learning_rate": 7.3191262751758005e-06, "loss": 0.9433, "step": 3112 }, { "epoch": 0.5989129911981146, "grad_norm": 1.6009118583845776, "learning_rate": 7.3131228417872905e-06, "loss": 0.9539, "step": 3113 }, { "epoch": 0.5991053821365013, "grad_norm": 1.3672331640194912, "learning_rate": 7.307120451950902e-06, "loss": 0.8559, "step": 3114 }, { "epoch": 0.5992977730748882, "grad_norm": 1.5818660248380345, "learning_rate": 7.301119107997905e-06, "loss": 1.0499, "step": 3115 }, { "epoch": 0.5994901640132749, "grad_norm": 1.3819832644206103, "learning_rate": 7.295118812259145e-06, "loss": 0.92, "step": 3116 }, { "epoch": 0.5996825549516618, "grad_norm": 1.2842120895196985, "learning_rate": 7.289119567065068e-06, "loss": 0.9894, "step": 3117 }, { "epoch": 0.5998749458900485, "grad_norm": 1.3406273957132822, "learning_rate": 7.2831213747457155e-06, "loss": 0.9357, "step": 3118 }, { "epoch": 0.6000673368284354, "grad_norm": 1.4218622061380701, "learning_rate": 7.2771242376307125e-06, "loss": 0.9123, "step": 3119 }, { "epoch": 0.6002597277668222, "grad_norm": 1.5058255170628747, "learning_rate": 7.271128158049283e-06, "loss": 1.0819, "step": 3120 }, { "epoch": 0.600452118705209, "grad_norm": 1.421457855579803, "learning_rate": 7.2651331383302326e-06, "loss": 1.0284, "step": 3121 }, { "epoch": 0.6006445096435958, "grad_norm": 1.4491255744972014, "learning_rate": 7.2591391808019555e-06, "loss": 0.9291, "step": 3122 }, { "epoch": 0.6008369005819826, "grad_norm": 1.4385074230980635, "learning_rate": 7.253146287792434e-06, "loss": 1.0432, "step": 3123 }, { "epoch": 0.6010292915203694, "grad_norm": 1.2549131348485043, "learning_rate": 7.247154461629248e-06, "loss": 0.8697, "step": 3124 }, { "epoch": 0.6012216824587562, "grad_norm": 1.3451140658127914, "learning_rate": 7.241163704639547e-06, "loss": 0.959, "step": 3125 }, { "epoch": 0.601414073397143, "grad_norm": 1.4263092899863257, "learning_rate": 7.235174019150071e-06, "loss": 0.9759, "step": 3126 }, { "epoch": 0.6016064643355298, "grad_norm": 1.3238753460258363, "learning_rate": 7.229185407487149e-06, "loss": 0.9612, "step": 3127 }, { "epoch": 0.6017988552739166, "grad_norm": 1.28995157967355, "learning_rate": 7.2231978719766884e-06, "loss": 0.9278, "step": 3128 }, { "epoch": 0.6019912462123034, "grad_norm": 1.4387728460233784, "learning_rate": 7.217211414944171e-06, "loss": 1.0295, "step": 3129 }, { "epoch": 0.6021836371506902, "grad_norm": 1.544731120904611, "learning_rate": 7.2112260387146784e-06, "loss": 0.9185, "step": 3130 }, { "epoch": 0.602376028089077, "grad_norm": 1.348069336794233, "learning_rate": 7.2052417456128565e-06, "loss": 0.9368, "step": 3131 }, { "epoch": 0.6025684190274638, "grad_norm": 1.4474089140782356, "learning_rate": 7.199258537962936e-06, "loss": 0.8869, "step": 3132 }, { "epoch": 0.6027608099658506, "grad_norm": 1.5148075816079112, "learning_rate": 7.193276418088729e-06, "loss": 0.9977, "step": 3133 }, { "epoch": 0.6029532009042374, "grad_norm": 0.676970144939663, "learning_rate": 7.187295388313618e-06, "loss": 0.8305, "step": 3134 }, { "epoch": 0.6031455918426242, "grad_norm": 1.3512778481864958, "learning_rate": 7.181315450960562e-06, "loss": 0.9392, "step": 3135 }, { "epoch": 0.603337982781011, "grad_norm": 1.43057432729271, "learning_rate": 7.175336608352113e-06, "loss": 0.9028, "step": 3136 }, { "epoch": 0.6035303737193978, "grad_norm": 1.4588302167473455, "learning_rate": 7.169358862810374e-06, "loss": 0.9753, "step": 3137 }, { "epoch": 0.6037227646577846, "grad_norm": 1.3744013182176211, "learning_rate": 7.163382216657033e-06, "loss": 1.0239, "step": 3138 }, { "epoch": 0.6039151555961714, "grad_norm": 1.6156207325737233, "learning_rate": 7.1574066722133565e-06, "loss": 0.973, "step": 3139 }, { "epoch": 0.6041075465345582, "grad_norm": 1.2929906202919146, "learning_rate": 7.151432231800173e-06, "loss": 1.0629, "step": 3140 }, { "epoch": 0.604299937472945, "grad_norm": 1.4079765771120558, "learning_rate": 7.145458897737882e-06, "loss": 1.0401, "step": 3141 }, { "epoch": 0.6044923284113318, "grad_norm": 1.4278119255934802, "learning_rate": 7.139486672346466e-06, "loss": 0.9677, "step": 3142 }, { "epoch": 0.6046847193497187, "grad_norm": 1.4133523555642034, "learning_rate": 7.133515557945463e-06, "loss": 0.955, "step": 3143 }, { "epoch": 0.6048771102881054, "grad_norm": 1.414511498064203, "learning_rate": 7.12754555685399e-06, "loss": 0.9613, "step": 3144 }, { "epoch": 0.6050695012264923, "grad_norm": 1.5148676884556325, "learning_rate": 7.121576671390722e-06, "loss": 0.9999, "step": 3145 }, { "epoch": 0.605261892164879, "grad_norm": 1.272113158319942, "learning_rate": 7.115608903873905e-06, "loss": 0.9175, "step": 3146 }, { "epoch": 0.6054542831032659, "grad_norm": 1.3110133401328234, "learning_rate": 7.109642256621353e-06, "loss": 1.0038, "step": 3147 }, { "epoch": 0.6056466740416526, "grad_norm": 1.1376532888420914, "learning_rate": 7.103676731950443e-06, "loss": 0.8726, "step": 3148 }, { "epoch": 0.6058390649800395, "grad_norm": 0.6825398244374631, "learning_rate": 7.0977123321781176e-06, "loss": 0.8078, "step": 3149 }, { "epoch": 0.6060314559184262, "grad_norm": 1.4222550697095508, "learning_rate": 7.091749059620881e-06, "loss": 1.0143, "step": 3150 }, { "epoch": 0.6062238468568131, "grad_norm": 1.308833178561716, "learning_rate": 7.0857869165947945e-06, "loss": 0.9176, "step": 3151 }, { "epoch": 0.6064162377951998, "grad_norm": 1.3418378331204444, "learning_rate": 7.079825905415491e-06, "loss": 1.0303, "step": 3152 }, { "epoch": 0.6066086287335867, "grad_norm": 1.398545092678295, "learning_rate": 7.073866028398153e-06, "loss": 1.0128, "step": 3153 }, { "epoch": 0.6068010196719734, "grad_norm": 1.2618907368479628, "learning_rate": 7.067907287857535e-06, "loss": 0.8612, "step": 3154 }, { "epoch": 0.6069934106103603, "grad_norm": 1.3798189961880851, "learning_rate": 7.061949686107938e-06, "loss": 1.0354, "step": 3155 }, { "epoch": 0.607185801548747, "grad_norm": 1.5551116799662792, "learning_rate": 7.0559932254632315e-06, "loss": 1.0171, "step": 3156 }, { "epoch": 0.6073781924871339, "grad_norm": 1.5306031676473448, "learning_rate": 7.0500379082368305e-06, "loss": 0.9948, "step": 3157 }, { "epoch": 0.6075705834255206, "grad_norm": 1.3735651544899559, "learning_rate": 7.044083736741711e-06, "loss": 0.9508, "step": 3158 }, { "epoch": 0.6077629743639075, "grad_norm": 1.5835097189203962, "learning_rate": 7.03813071329041e-06, "loss": 1.0869, "step": 3159 }, { "epoch": 0.6079553653022942, "grad_norm": 0.7096522047583788, "learning_rate": 7.032178840195009e-06, "loss": 0.8788, "step": 3160 }, { "epoch": 0.6081477562406811, "grad_norm": 0.699402138392386, "learning_rate": 7.026228119767149e-06, "loss": 0.8427, "step": 3161 }, { "epoch": 0.6083401471790678, "grad_norm": 1.322370939127462, "learning_rate": 7.020278554318023e-06, "loss": 0.9728, "step": 3162 }, { "epoch": 0.6085325381174547, "grad_norm": 1.3811434260498778, "learning_rate": 7.014330146158367e-06, "loss": 0.979, "step": 3163 }, { "epoch": 0.6087249290558415, "grad_norm": 1.5347183664421662, "learning_rate": 7.008382897598477e-06, "loss": 0.9616, "step": 3164 }, { "epoch": 0.6089173199942283, "grad_norm": 1.5335543372163518, "learning_rate": 7.002436810948201e-06, "loss": 1.0162, "step": 3165 }, { "epoch": 0.6091097109326151, "grad_norm": 1.5166446673276177, "learning_rate": 6.996491888516927e-06, "loss": 1.0066, "step": 3166 }, { "epoch": 0.6093021018710019, "grad_norm": 1.4462722100923036, "learning_rate": 6.990548132613592e-06, "loss": 0.9979, "step": 3167 }, { "epoch": 0.6094944928093887, "grad_norm": 1.3136966665262881, "learning_rate": 6.984605545546686e-06, "loss": 0.9086, "step": 3168 }, { "epoch": 0.6096868837477755, "grad_norm": 1.4354878090397682, "learning_rate": 6.978664129624241e-06, "loss": 1.0275, "step": 3169 }, { "epoch": 0.6098792746861623, "grad_norm": 1.4491077488313222, "learning_rate": 6.972723887153828e-06, "loss": 0.975, "step": 3170 }, { "epoch": 0.6100716656245491, "grad_norm": 1.5486110465040546, "learning_rate": 6.9667848204425785e-06, "loss": 0.9198, "step": 3171 }, { "epoch": 0.6102640565629359, "grad_norm": 1.4942633437223167, "learning_rate": 6.960846931797152e-06, "loss": 0.9463, "step": 3172 }, { "epoch": 0.6104564475013227, "grad_norm": 1.5288721612069083, "learning_rate": 6.9549102235237565e-06, "loss": 0.9599, "step": 3173 }, { "epoch": 0.6106488384397095, "grad_norm": 1.4597819528461697, "learning_rate": 6.948974697928144e-06, "loss": 0.9635, "step": 3174 }, { "epoch": 0.6108412293780963, "grad_norm": 1.4709589888950094, "learning_rate": 6.943040357315598e-06, "loss": 0.9761, "step": 3175 }, { "epoch": 0.6110336203164831, "grad_norm": 0.7770767505876528, "learning_rate": 6.9371072039909515e-06, "loss": 0.8476, "step": 3176 }, { "epoch": 0.6112260112548699, "grad_norm": 1.385055164246724, "learning_rate": 6.931175240258576e-06, "loss": 0.9291, "step": 3177 }, { "epoch": 0.6114184021932567, "grad_norm": 0.6745112372980059, "learning_rate": 6.9252444684223765e-06, "loss": 0.82, "step": 3178 }, { "epoch": 0.6116107931316435, "grad_norm": 1.3096401150127128, "learning_rate": 6.919314890785793e-06, "loss": 0.9417, "step": 3179 }, { "epoch": 0.6118031840700303, "grad_norm": 1.2490373270822845, "learning_rate": 6.913386509651807e-06, "loss": 0.9623, "step": 3180 }, { "epoch": 0.6119955750084171, "grad_norm": 1.3563958388106019, "learning_rate": 6.907459327322934e-06, "loss": 0.9819, "step": 3181 }, { "epoch": 0.6121879659468039, "grad_norm": 1.4300067006016817, "learning_rate": 6.90153334610122e-06, "loss": 1.0088, "step": 3182 }, { "epoch": 0.6123803568851907, "grad_norm": 1.4716404534248277, "learning_rate": 6.895608568288255e-06, "loss": 0.9489, "step": 3183 }, { "epoch": 0.6125727478235775, "grad_norm": 1.3828120548022462, "learning_rate": 6.889684996185148e-06, "loss": 0.9548, "step": 3184 }, { "epoch": 0.6127651387619643, "grad_norm": 1.5474809869321475, "learning_rate": 6.88376263209255e-06, "loss": 0.9744, "step": 3185 }, { "epoch": 0.6129575297003511, "grad_norm": 1.3965857845279852, "learning_rate": 6.877841478310639e-06, "loss": 1.0332, "step": 3186 }, { "epoch": 0.613149920638738, "grad_norm": 1.3922900248585734, "learning_rate": 6.871921537139117e-06, "loss": 0.9886, "step": 3187 }, { "epoch": 0.6133423115771247, "grad_norm": 1.4278807769318753, "learning_rate": 6.866002810877224e-06, "loss": 0.9444, "step": 3188 }, { "epoch": 0.6135347025155116, "grad_norm": 1.3090462533470377, "learning_rate": 6.860085301823729e-06, "loss": 0.9316, "step": 3189 }, { "epoch": 0.6137270934538983, "grad_norm": 1.6296867239168122, "learning_rate": 6.854169012276923e-06, "loss": 0.9974, "step": 3190 }, { "epoch": 0.6139194843922852, "grad_norm": 1.5158971757143607, "learning_rate": 6.848253944534622e-06, "loss": 0.9032, "step": 3191 }, { "epoch": 0.6141118753306719, "grad_norm": 1.362204388991791, "learning_rate": 6.84234010089417e-06, "loss": 0.9687, "step": 3192 }, { "epoch": 0.6143042662690588, "grad_norm": 1.3723926172803296, "learning_rate": 6.836427483652436e-06, "loss": 0.9349, "step": 3193 }, { "epoch": 0.6144966572074455, "grad_norm": 1.3033006741580047, "learning_rate": 6.830516095105817e-06, "loss": 0.9608, "step": 3194 }, { "epoch": 0.6146890481458324, "grad_norm": 1.4250379624143077, "learning_rate": 6.824605937550224e-06, "loss": 0.9234, "step": 3195 }, { "epoch": 0.6148814390842191, "grad_norm": 1.3645071066432335, "learning_rate": 6.818697013281093e-06, "loss": 0.9777, "step": 3196 }, { "epoch": 0.615073830022606, "grad_norm": 1.4092467043783343, "learning_rate": 6.8127893245933864e-06, "loss": 0.9939, "step": 3197 }, { "epoch": 0.6152662209609927, "grad_norm": 1.433068645434575, "learning_rate": 6.806882873781579e-06, "loss": 0.9125, "step": 3198 }, { "epoch": 0.6154586118993796, "grad_norm": 1.4422189184202348, "learning_rate": 6.800977663139666e-06, "loss": 1.0032, "step": 3199 }, { "epoch": 0.6156510028377663, "grad_norm": 1.2937422270139343, "learning_rate": 6.795073694961171e-06, "loss": 0.9621, "step": 3200 }, { "epoch": 0.6158433937761532, "grad_norm": 1.4227581776432123, "learning_rate": 6.789170971539119e-06, "loss": 0.9897, "step": 3201 }, { "epoch": 0.6160357847145399, "grad_norm": 1.2467033543313863, "learning_rate": 6.783269495166066e-06, "loss": 0.9635, "step": 3202 }, { "epoch": 0.6162281756529268, "grad_norm": 1.3745619559711835, "learning_rate": 6.777369268134076e-06, "loss": 0.8607, "step": 3203 }, { "epoch": 0.6164205665913135, "grad_norm": 1.336496278670604, "learning_rate": 6.771470292734723e-06, "loss": 0.949, "step": 3204 }, { "epoch": 0.6166129575297004, "grad_norm": 1.3502394827415785, "learning_rate": 6.7655725712591055e-06, "loss": 0.9451, "step": 3205 }, { "epoch": 0.6168053484680871, "grad_norm": 1.3129716433625231, "learning_rate": 6.759676105997834e-06, "loss": 0.9546, "step": 3206 }, { "epoch": 0.616997739406474, "grad_norm": 1.3056456214073722, "learning_rate": 6.753780899241027e-06, "loss": 0.98, "step": 3207 }, { "epoch": 0.6171901303448608, "grad_norm": 1.39917563831135, "learning_rate": 6.747886953278311e-06, "loss": 0.9229, "step": 3208 }, { "epoch": 0.6173825212832476, "grad_norm": 1.293407142384187, "learning_rate": 6.741994270398826e-06, "loss": 1.0289, "step": 3209 }, { "epoch": 0.6175749122216344, "grad_norm": 1.293043087416766, "learning_rate": 6.736102852891227e-06, "loss": 0.8806, "step": 3210 }, { "epoch": 0.6177673031600212, "grad_norm": 1.363744633430354, "learning_rate": 6.730212703043666e-06, "loss": 0.9149, "step": 3211 }, { "epoch": 0.617959694098408, "grad_norm": 1.3411362015326005, "learning_rate": 6.7243238231438176e-06, "loss": 0.8919, "step": 3212 }, { "epoch": 0.6181520850367948, "grad_norm": 0.7650748513625317, "learning_rate": 6.718436215478849e-06, "loss": 0.8671, "step": 3213 }, { "epoch": 0.6183444759751816, "grad_norm": 1.3963619448683426, "learning_rate": 6.712549882335442e-06, "loss": 0.9172, "step": 3214 }, { "epoch": 0.6185368669135684, "grad_norm": 1.4240102469936202, "learning_rate": 6.70666482599978e-06, "loss": 0.9905, "step": 3215 }, { "epoch": 0.6187292578519552, "grad_norm": 1.3105949730967086, "learning_rate": 6.7007810487575475e-06, "loss": 0.9706, "step": 3216 }, { "epoch": 0.618921648790342, "grad_norm": 1.497652249642097, "learning_rate": 6.694898552893941e-06, "loss": 1.0458, "step": 3217 }, { "epoch": 0.6191140397287288, "grad_norm": 1.4713362357652298, "learning_rate": 6.6890173406936485e-06, "loss": 0.9392, "step": 3218 }, { "epoch": 0.6193064306671155, "grad_norm": 1.5091430232376146, "learning_rate": 6.683137414440872e-06, "loss": 1.0162, "step": 3219 }, { "epoch": 0.6194988216055024, "grad_norm": 1.4988717050895237, "learning_rate": 6.677258776419304e-06, "loss": 0.9868, "step": 3220 }, { "epoch": 0.6196912125438891, "grad_norm": 1.4089075525068895, "learning_rate": 6.671381428912138e-06, "loss": 1.075, "step": 3221 }, { "epoch": 0.619883603482276, "grad_norm": 1.2889057013374927, "learning_rate": 6.66550537420207e-06, "loss": 0.9868, "step": 3222 }, { "epoch": 0.6200759944206627, "grad_norm": 1.4214034717768889, "learning_rate": 6.659630614571287e-06, "loss": 0.9518, "step": 3223 }, { "epoch": 0.6202683853590496, "grad_norm": 1.3008713889273393, "learning_rate": 6.653757152301488e-06, "loss": 0.9569, "step": 3224 }, { "epoch": 0.6204607762974363, "grad_norm": 1.5821619374740536, "learning_rate": 6.647884989673849e-06, "loss": 1.0512, "step": 3225 }, { "epoch": 0.6206531672358232, "grad_norm": 1.323087044767629, "learning_rate": 6.642014128969055e-06, "loss": 0.9833, "step": 3226 }, { "epoch": 0.6208455581742099, "grad_norm": 1.3492600024527832, "learning_rate": 6.63614457246728e-06, "loss": 1.0126, "step": 3227 }, { "epoch": 0.6210379491125968, "grad_norm": 1.3574581392371416, "learning_rate": 6.630276322448188e-06, "loss": 0.9534, "step": 3228 }, { "epoch": 0.6212303400509837, "grad_norm": 1.3442165714577603, "learning_rate": 6.624409381190946e-06, "loss": 0.9463, "step": 3229 }, { "epoch": 0.6214227309893704, "grad_norm": 1.3296490099581768, "learning_rate": 6.618543750974202e-06, "loss": 0.9441, "step": 3230 }, { "epoch": 0.6216151219277573, "grad_norm": 1.2515691982643213, "learning_rate": 6.6126794340761025e-06, "loss": 0.9255, "step": 3231 }, { "epoch": 0.621807512866144, "grad_norm": 1.4544882899410412, "learning_rate": 6.606816432774279e-06, "loss": 1.0139, "step": 3232 }, { "epoch": 0.6219999038045309, "grad_norm": 1.3609800460674482, "learning_rate": 6.600954749345851e-06, "loss": 1.0891, "step": 3233 }, { "epoch": 0.6221922947429176, "grad_norm": 1.388510140750358, "learning_rate": 6.595094386067428e-06, "loss": 0.979, "step": 3234 }, { "epoch": 0.6223846856813045, "grad_norm": 1.5373889195963055, "learning_rate": 6.589235345215117e-06, "loss": 0.9291, "step": 3235 }, { "epoch": 0.6225770766196912, "grad_norm": 1.3490147012183553, "learning_rate": 6.583377629064494e-06, "loss": 1.0001, "step": 3236 }, { "epoch": 0.622769467558078, "grad_norm": 1.4691508738013617, "learning_rate": 6.5775212398906295e-06, "loss": 0.9053, "step": 3237 }, { "epoch": 0.6229618584964648, "grad_norm": 1.3676276508121847, "learning_rate": 6.571666179968079e-06, "loss": 1.0169, "step": 3238 }, { "epoch": 0.6231542494348516, "grad_norm": 1.4329732129631423, "learning_rate": 6.565812451570881e-06, "loss": 0.9906, "step": 3239 }, { "epoch": 0.6233466403732384, "grad_norm": 1.4778346848278745, "learning_rate": 6.5599600569725495e-06, "loss": 0.9744, "step": 3240 }, { "epoch": 0.6235390313116252, "grad_norm": 1.6627839674714158, "learning_rate": 6.554108998446096e-06, "loss": 1.0446, "step": 3241 }, { "epoch": 0.623731422250012, "grad_norm": 1.5429645747712737, "learning_rate": 6.548259278263999e-06, "loss": 0.9409, "step": 3242 }, { "epoch": 0.6239238131883988, "grad_norm": 1.4293341697764743, "learning_rate": 6.542410898698226e-06, "loss": 0.9488, "step": 3243 }, { "epoch": 0.6241162041267856, "grad_norm": 1.3459608632288484, "learning_rate": 6.536563862020218e-06, "loss": 0.9274, "step": 3244 }, { "epoch": 0.6243085950651724, "grad_norm": 1.3675965037587425, "learning_rate": 6.530718170500896e-06, "loss": 0.907, "step": 3245 }, { "epoch": 0.6245009860035592, "grad_norm": 1.359879525125751, "learning_rate": 6.524873826410658e-06, "loss": 0.9308, "step": 3246 }, { "epoch": 0.624693376941946, "grad_norm": 1.4834756055888103, "learning_rate": 6.519030832019383e-06, "loss": 1.0465, "step": 3247 }, { "epoch": 0.6248857678803328, "grad_norm": 1.4341663641298497, "learning_rate": 6.513189189596422e-06, "loss": 0.9835, "step": 3248 }, { "epoch": 0.6250781588187196, "grad_norm": 1.2931741145805422, "learning_rate": 6.507348901410604e-06, "loss": 0.9571, "step": 3249 }, { "epoch": 0.6252705497571064, "grad_norm": 1.4251856199180197, "learning_rate": 6.501509969730224e-06, "loss": 0.9714, "step": 3250 }, { "epoch": 0.6254629406954932, "grad_norm": 1.3828783165439409, "learning_rate": 6.495672396823061e-06, "loss": 0.9935, "step": 3251 }, { "epoch": 0.6256553316338801, "grad_norm": 1.3976345489099928, "learning_rate": 6.489836184956353e-06, "loss": 1.0124, "step": 3252 }, { "epoch": 0.6258477225722668, "grad_norm": 1.4127165504455672, "learning_rate": 6.484001336396828e-06, "loss": 1.0032, "step": 3253 }, { "epoch": 0.6260401135106537, "grad_norm": 1.5688457833722957, "learning_rate": 6.478167853410668e-06, "loss": 0.9988, "step": 3254 }, { "epoch": 0.6262325044490404, "grad_norm": 1.4058790425879701, "learning_rate": 6.472335738263534e-06, "loss": 0.9255, "step": 3255 }, { "epoch": 0.6264248953874273, "grad_norm": 1.3559337891148988, "learning_rate": 6.466504993220548e-06, "loss": 1.0134, "step": 3256 }, { "epoch": 0.626617286325814, "grad_norm": 1.312923904543076, "learning_rate": 6.460675620546305e-06, "loss": 0.9731, "step": 3257 }, { "epoch": 0.6268096772642009, "grad_norm": 1.4011612220626475, "learning_rate": 6.454847622504867e-06, "loss": 0.9766, "step": 3258 }, { "epoch": 0.6270020682025876, "grad_norm": 1.4373531596510463, "learning_rate": 6.4490210013597635e-06, "loss": 0.9398, "step": 3259 }, { "epoch": 0.6271944591409745, "grad_norm": 1.3182103483317604, "learning_rate": 6.4431957593739845e-06, "loss": 1.0598, "step": 3260 }, { "epoch": 0.6273868500793612, "grad_norm": 1.3959031110697413, "learning_rate": 6.4373718988099896e-06, "loss": 0.9542, "step": 3261 }, { "epoch": 0.6275792410177481, "grad_norm": 1.3249998340696894, "learning_rate": 6.431549421929694e-06, "loss": 0.9008, "step": 3262 }, { "epoch": 0.6277716319561348, "grad_norm": 1.2751543917263304, "learning_rate": 6.4257283309944804e-06, "loss": 0.9324, "step": 3263 }, { "epoch": 0.6279640228945217, "grad_norm": 1.3130206427797289, "learning_rate": 6.419908628265203e-06, "loss": 0.9499, "step": 3264 }, { "epoch": 0.6281564138329084, "grad_norm": 1.297331334749863, "learning_rate": 6.414090316002161e-06, "loss": 0.9692, "step": 3265 }, { "epoch": 0.6283488047712953, "grad_norm": 1.3751744795770093, "learning_rate": 6.4082733964651166e-06, "loss": 0.8855, "step": 3266 }, { "epoch": 0.628541195709682, "grad_norm": 1.3470209602765884, "learning_rate": 6.4024578719133e-06, "loss": 0.9611, "step": 3267 }, { "epoch": 0.6287335866480689, "grad_norm": 1.3981801426193006, "learning_rate": 6.396643744605391e-06, "loss": 0.9516, "step": 3268 }, { "epoch": 0.6289259775864556, "grad_norm": 1.3617662288062544, "learning_rate": 6.390831016799527e-06, "loss": 0.9313, "step": 3269 }, { "epoch": 0.6291183685248425, "grad_norm": 1.4351048869898977, "learning_rate": 6.385019690753311e-06, "loss": 0.9036, "step": 3270 }, { "epoch": 0.6293107594632292, "grad_norm": 1.6132091523338692, "learning_rate": 6.379209768723791e-06, "loss": 1.0378, "step": 3271 }, { "epoch": 0.6295031504016161, "grad_norm": 1.372717434860958, "learning_rate": 6.373401252967475e-06, "loss": 0.949, "step": 3272 }, { "epoch": 0.6296955413400029, "grad_norm": 1.5578068671756553, "learning_rate": 6.367594145740324e-06, "loss": 0.9931, "step": 3273 }, { "epoch": 0.6298879322783897, "grad_norm": 1.3457033186608627, "learning_rate": 6.361788449297748e-06, "loss": 1.0032, "step": 3274 }, { "epoch": 0.6300803232167765, "grad_norm": 1.384040487897845, "learning_rate": 6.355984165894613e-06, "loss": 0.8945, "step": 3275 }, { "epoch": 0.6302727141551633, "grad_norm": 1.332012142316043, "learning_rate": 6.350181297785242e-06, "loss": 0.9865, "step": 3276 }, { "epoch": 0.6304651050935501, "grad_norm": 1.371294605917315, "learning_rate": 6.344379847223398e-06, "loss": 0.9709, "step": 3277 }, { "epoch": 0.6306574960319369, "grad_norm": 0.7581303154989347, "learning_rate": 6.338579816462298e-06, "loss": 0.8423, "step": 3278 }, { "epoch": 0.6308498869703237, "grad_norm": 1.456604715442771, "learning_rate": 6.332781207754605e-06, "loss": 0.9199, "step": 3279 }, { "epoch": 0.6310422779087105, "grad_norm": 1.3671874017743986, "learning_rate": 6.326984023352435e-06, "loss": 1.0156, "step": 3280 }, { "epoch": 0.6312346688470973, "grad_norm": 1.4877072653903074, "learning_rate": 6.321188265507342e-06, "loss": 0.985, "step": 3281 }, { "epoch": 0.6314270597854841, "grad_norm": 1.4237221125291568, "learning_rate": 6.31539393647034e-06, "loss": 1.0685, "step": 3282 }, { "epoch": 0.6316194507238709, "grad_norm": 0.7526316526034873, "learning_rate": 6.309601038491874e-06, "loss": 0.8783, "step": 3283 }, { "epoch": 0.6318118416622577, "grad_norm": 1.4228010142164493, "learning_rate": 6.303809573821842e-06, "loss": 0.9635, "step": 3284 }, { "epoch": 0.6320042326006445, "grad_norm": 1.44803958870596, "learning_rate": 6.298019544709579e-06, "loss": 0.9628, "step": 3285 }, { "epoch": 0.6321966235390313, "grad_norm": 1.497760037149801, "learning_rate": 6.292230953403866e-06, "loss": 0.9583, "step": 3286 }, { "epoch": 0.6323890144774181, "grad_norm": 1.291933592996588, "learning_rate": 6.286443802152926e-06, "loss": 1.0097, "step": 3287 }, { "epoch": 0.6325814054158049, "grad_norm": 1.2942609908202323, "learning_rate": 6.280658093204422e-06, "loss": 1.0417, "step": 3288 }, { "epoch": 0.6327737963541917, "grad_norm": 1.3540143324840994, "learning_rate": 6.274873828805459e-06, "loss": 0.9378, "step": 3289 }, { "epoch": 0.6329661872925785, "grad_norm": 1.641514462855454, "learning_rate": 6.269091011202576e-06, "loss": 1.0057, "step": 3290 }, { "epoch": 0.6331585782309653, "grad_norm": 1.4356664761515996, "learning_rate": 6.263309642641751e-06, "loss": 0.9812, "step": 3291 }, { "epoch": 0.6333509691693521, "grad_norm": 1.4265318802137084, "learning_rate": 6.257529725368405e-06, "loss": 0.9858, "step": 3292 }, { "epoch": 0.6335433601077389, "grad_norm": 1.254385594559538, "learning_rate": 6.251751261627386e-06, "loss": 0.9378, "step": 3293 }, { "epoch": 0.6337357510461257, "grad_norm": 1.3241391436775958, "learning_rate": 6.245974253662988e-06, "loss": 0.9579, "step": 3294 }, { "epoch": 0.6339281419845125, "grad_norm": 1.5381207133139176, "learning_rate": 6.240198703718932e-06, "loss": 0.9498, "step": 3295 }, { "epoch": 0.6341205329228994, "grad_norm": 1.5109326923718507, "learning_rate": 6.234424614038375e-06, "loss": 1.0486, "step": 3296 }, { "epoch": 0.6343129238612861, "grad_norm": 1.3890223959989534, "learning_rate": 6.2286519868639095e-06, "loss": 0.9654, "step": 3297 }, { "epoch": 0.634505314799673, "grad_norm": 1.4415055033621273, "learning_rate": 6.222880824437549e-06, "loss": 0.9774, "step": 3298 }, { "epoch": 0.6346977057380597, "grad_norm": 1.377235099908128, "learning_rate": 6.217111129000759e-06, "loss": 0.9508, "step": 3299 }, { "epoch": 0.6348900966764466, "grad_norm": 1.4326440216650471, "learning_rate": 6.211342902794413e-06, "loss": 0.9193, "step": 3300 }, { "epoch": 0.6350824876148333, "grad_norm": 1.3864112129177093, "learning_rate": 6.205576148058828e-06, "loss": 1.0156, "step": 3301 }, { "epoch": 0.6352748785532202, "grad_norm": 1.4293961902410415, "learning_rate": 6.199810867033745e-06, "loss": 0.9744, "step": 3302 }, { "epoch": 0.6354672694916069, "grad_norm": 1.4403436985229354, "learning_rate": 6.19404706195833e-06, "loss": 1.0343, "step": 3303 }, { "epoch": 0.6356596604299938, "grad_norm": 1.3572731179457007, "learning_rate": 6.188284735071177e-06, "loss": 1.0069, "step": 3304 }, { "epoch": 0.6358520513683805, "grad_norm": 1.3367160784714747, "learning_rate": 6.182523888610316e-06, "loss": 0.9395, "step": 3305 }, { "epoch": 0.6360444423067674, "grad_norm": 1.3644393723711834, "learning_rate": 6.176764524813187e-06, "loss": 1.0253, "step": 3306 }, { "epoch": 0.6362368332451541, "grad_norm": 1.417828643669946, "learning_rate": 6.171006645916662e-06, "loss": 1.0004, "step": 3307 }, { "epoch": 0.636429224183541, "grad_norm": 1.394660560505906, "learning_rate": 6.165250254157032e-06, "loss": 0.9409, "step": 3308 }, { "epoch": 0.6366216151219277, "grad_norm": 1.3384008815670771, "learning_rate": 6.159495351770017e-06, "loss": 0.9433, "step": 3309 }, { "epoch": 0.6368140060603146, "grad_norm": 1.4338019526141297, "learning_rate": 6.153741940990749e-06, "loss": 0.9568, "step": 3310 }, { "epoch": 0.6370063969987013, "grad_norm": 1.311437830651553, "learning_rate": 6.1479900240537956e-06, "loss": 0.9828, "step": 3311 }, { "epoch": 0.6371987879370882, "grad_norm": 1.5116098377798142, "learning_rate": 6.142239603193128e-06, "loss": 0.9327, "step": 3312 }, { "epoch": 0.6373911788754749, "grad_norm": 1.4012925011310677, "learning_rate": 6.136490680642146e-06, "loss": 0.8828, "step": 3313 }, { "epoch": 0.6375835698138618, "grad_norm": 1.4087273106784675, "learning_rate": 6.130743258633667e-06, "loss": 1.0414, "step": 3314 }, { "epoch": 0.6377759607522485, "grad_norm": 1.5006844652354272, "learning_rate": 6.124997339399916e-06, "loss": 0.9915, "step": 3315 }, { "epoch": 0.6379683516906354, "grad_norm": 1.3854392903297756, "learning_rate": 6.119252925172549e-06, "loss": 0.9633, "step": 3316 }, { "epoch": 0.6381607426290222, "grad_norm": 1.4736905704204846, "learning_rate": 6.113510018182628e-06, "loss": 0.9147, "step": 3317 }, { "epoch": 0.638353133567409, "grad_norm": 1.526808800964783, "learning_rate": 6.107768620660633e-06, "loss": 0.9105, "step": 3318 }, { "epoch": 0.6385455245057958, "grad_norm": 1.4542226374932161, "learning_rate": 6.102028734836456e-06, "loss": 0.913, "step": 3319 }, { "epoch": 0.6387379154441826, "grad_norm": 1.2870047459643004, "learning_rate": 6.0962903629394e-06, "loss": 0.8104, "step": 3320 }, { "epoch": 0.6389303063825694, "grad_norm": 0.7232381438390166, "learning_rate": 6.090553507198187e-06, "loss": 0.8144, "step": 3321 }, { "epoch": 0.6391226973209562, "grad_norm": 1.3262946239140365, "learning_rate": 6.0848181698409384e-06, "loss": 0.8943, "step": 3322 }, { "epoch": 0.639315088259343, "grad_norm": 1.3138676292799372, "learning_rate": 6.079084353095202e-06, "loss": 0.9662, "step": 3323 }, { "epoch": 0.6395074791977298, "grad_norm": 1.421343762072421, "learning_rate": 6.07335205918792e-06, "loss": 1.0675, "step": 3324 }, { "epoch": 0.6396998701361166, "grad_norm": 1.4467100328114766, "learning_rate": 6.067621290345455e-06, "loss": 0.9651, "step": 3325 }, { "epoch": 0.6398922610745034, "grad_norm": 1.4540467347596389, "learning_rate": 6.061892048793568e-06, "loss": 0.9927, "step": 3326 }, { "epoch": 0.6400846520128902, "grad_norm": 1.4447378855394784, "learning_rate": 6.056164336757426e-06, "loss": 0.9781, "step": 3327 }, { "epoch": 0.640277042951277, "grad_norm": 1.3345030981451023, "learning_rate": 6.050438156461613e-06, "loss": 0.884, "step": 3328 }, { "epoch": 0.6404694338896638, "grad_norm": 1.5804741791277421, "learning_rate": 6.044713510130108e-06, "loss": 1.0066, "step": 3329 }, { "epoch": 0.6406618248280506, "grad_norm": 1.3849113293342097, "learning_rate": 6.038990399986302e-06, "loss": 0.9358, "step": 3330 }, { "epoch": 0.6408542157664374, "grad_norm": 1.3778024705264587, "learning_rate": 6.03326882825298e-06, "loss": 0.9781, "step": 3331 }, { "epoch": 0.6410466067048242, "grad_norm": 1.4001991568287195, "learning_rate": 6.027548797152336e-06, "loss": 0.9473, "step": 3332 }, { "epoch": 0.641238997643211, "grad_norm": 1.400602210884415, "learning_rate": 6.021830308905963e-06, "loss": 0.9712, "step": 3333 }, { "epoch": 0.6414313885815978, "grad_norm": 1.4265643774454997, "learning_rate": 6.016113365734861e-06, "loss": 0.9885, "step": 3334 }, { "epoch": 0.6416237795199846, "grad_norm": 1.605311416775843, "learning_rate": 6.0103979698594215e-06, "loss": 1.0357, "step": 3335 }, { "epoch": 0.6418161704583714, "grad_norm": 1.3862514881508117, "learning_rate": 6.004684123499436e-06, "loss": 0.8663, "step": 3336 }, { "epoch": 0.6420085613967582, "grad_norm": 1.5005281167408557, "learning_rate": 5.998971828874102e-06, "loss": 1.0677, "step": 3337 }, { "epoch": 0.6422009523351451, "grad_norm": 1.4683630612302845, "learning_rate": 5.993261088202005e-06, "loss": 1.0077, "step": 3338 }, { "epoch": 0.6423933432735318, "grad_norm": 0.7701731545861026, "learning_rate": 5.987551903701128e-06, "loss": 0.9078, "step": 3339 }, { "epoch": 0.6425857342119187, "grad_norm": 1.355628678435174, "learning_rate": 5.9818442775888595e-06, "loss": 1.0444, "step": 3340 }, { "epoch": 0.6427781251503054, "grad_norm": 1.3531010134410864, "learning_rate": 5.97613821208197e-06, "loss": 0.9579, "step": 3341 }, { "epoch": 0.6429705160886923, "grad_norm": 1.5552875635144274, "learning_rate": 5.970433709396635e-06, "loss": 0.974, "step": 3342 }, { "epoch": 0.643162907027079, "grad_norm": 1.253179755310937, "learning_rate": 5.964730771748415e-06, "loss": 0.9464, "step": 3343 }, { "epoch": 0.6433552979654659, "grad_norm": 1.4218251159510114, "learning_rate": 5.959029401352262e-06, "loss": 1.0175, "step": 3344 }, { "epoch": 0.6435476889038526, "grad_norm": 1.5384122093985855, "learning_rate": 5.953329600422524e-06, "loss": 0.9492, "step": 3345 }, { "epoch": 0.6437400798422395, "grad_norm": 1.552303843912779, "learning_rate": 5.947631371172943e-06, "loss": 0.9943, "step": 3346 }, { "epoch": 0.6439324707806262, "grad_norm": 1.3740510525314882, "learning_rate": 5.941934715816642e-06, "loss": 0.9728, "step": 3347 }, { "epoch": 0.6441248617190131, "grad_norm": 1.6123405370749158, "learning_rate": 5.936239636566137e-06, "loss": 1.0751, "step": 3348 }, { "epoch": 0.6443172526573998, "grad_norm": 1.3802285746328196, "learning_rate": 5.930546135633327e-06, "loss": 0.9965, "step": 3349 }, { "epoch": 0.6445096435957867, "grad_norm": 1.3432690843092967, "learning_rate": 5.924854215229509e-06, "loss": 0.895, "step": 3350 }, { "epoch": 0.6447020345341734, "grad_norm": 1.4414715789523347, "learning_rate": 5.919163877565351e-06, "loss": 0.9686, "step": 3351 }, { "epoch": 0.6448944254725603, "grad_norm": 1.2844282546124286, "learning_rate": 5.9134751248509236e-06, "loss": 0.923, "step": 3352 }, { "epoch": 0.645086816410947, "grad_norm": 1.3389152769066173, "learning_rate": 5.9077879592956675e-06, "loss": 0.9394, "step": 3353 }, { "epoch": 0.6452792073493339, "grad_norm": 1.3848865992383712, "learning_rate": 5.902102383108415e-06, "loss": 1.0321, "step": 3354 }, { "epoch": 0.6454715982877206, "grad_norm": 1.3223576109628976, "learning_rate": 5.896418398497377e-06, "loss": 0.9577, "step": 3355 }, { "epoch": 0.6456639892261075, "grad_norm": 1.3282827287545584, "learning_rate": 5.890736007670144e-06, "loss": 0.9723, "step": 3356 }, { "epoch": 0.6458563801644942, "grad_norm": 1.3254988578003237, "learning_rate": 5.885055212833696e-06, "loss": 0.9235, "step": 3357 }, { "epoch": 0.6460487711028811, "grad_norm": 1.3807518856652798, "learning_rate": 5.879376016194387e-06, "loss": 0.9749, "step": 3358 }, { "epoch": 0.6462411620412678, "grad_norm": 1.4327943380888553, "learning_rate": 5.873698419957952e-06, "loss": 0.9406, "step": 3359 }, { "epoch": 0.6464335529796547, "grad_norm": 1.4362185767225668, "learning_rate": 5.8680224263295045e-06, "loss": 1.0362, "step": 3360 }, { "epoch": 0.6466259439180415, "grad_norm": 1.3016172109167186, "learning_rate": 5.862348037513533e-06, "loss": 0.9228, "step": 3361 }, { "epoch": 0.6468183348564283, "grad_norm": 1.446577446408814, "learning_rate": 5.856675255713905e-06, "loss": 0.9159, "step": 3362 }, { "epoch": 0.6470107257948151, "grad_norm": 1.453720359541848, "learning_rate": 5.851004083133862e-06, "loss": 0.948, "step": 3363 }, { "epoch": 0.6472031167332019, "grad_norm": 1.4744462115361927, "learning_rate": 5.8453345219760275e-06, "loss": 0.9609, "step": 3364 }, { "epoch": 0.6473955076715887, "grad_norm": 1.3663037368917685, "learning_rate": 5.839666574442389e-06, "loss": 0.9756, "step": 3365 }, { "epoch": 0.6475878986099755, "grad_norm": 1.5808287223504673, "learning_rate": 5.834000242734317e-06, "loss": 0.979, "step": 3366 }, { "epoch": 0.6477802895483623, "grad_norm": 1.4302831722751557, "learning_rate": 5.828335529052541e-06, "loss": 0.9418, "step": 3367 }, { "epoch": 0.647972680486749, "grad_norm": 1.3288006511006805, "learning_rate": 5.822672435597172e-06, "loss": 0.9662, "step": 3368 }, { "epoch": 0.6481650714251359, "grad_norm": 1.3079506061714126, "learning_rate": 5.817010964567702e-06, "loss": 1.0008, "step": 3369 }, { "epoch": 0.6483574623635227, "grad_norm": 1.5216686528874954, "learning_rate": 5.811351118162969e-06, "loss": 1.0089, "step": 3370 }, { "epoch": 0.6485498533019095, "grad_norm": 1.4157478841416402, "learning_rate": 5.805692898581196e-06, "loss": 0.9218, "step": 3371 }, { "epoch": 0.6487422442402963, "grad_norm": 1.4835371996033866, "learning_rate": 5.800036308019974e-06, "loss": 0.9635, "step": 3372 }, { "epoch": 0.6489346351786831, "grad_norm": 1.271768804002318, "learning_rate": 5.79438134867625e-06, "loss": 1.0096, "step": 3373 }, { "epoch": 0.6491270261170698, "grad_norm": 1.4415635801704274, "learning_rate": 5.788728022746348e-06, "loss": 1.0227, "step": 3374 }, { "epoch": 0.6493194170554567, "grad_norm": 1.4030328560759615, "learning_rate": 5.783076332425957e-06, "loss": 0.9032, "step": 3375 }, { "epoch": 0.6495118079938434, "grad_norm": 1.4343103207090184, "learning_rate": 5.777426279910125e-06, "loss": 0.9561, "step": 3376 }, { "epoch": 0.6497041989322303, "grad_norm": 1.3848682682668252, "learning_rate": 5.771777867393275e-06, "loss": 0.9597, "step": 3377 }, { "epoch": 0.649896589870617, "grad_norm": 1.4806400399811739, "learning_rate": 5.766131097069174e-06, "loss": 0.9563, "step": 3378 }, { "epoch": 0.6500889808090039, "grad_norm": 1.5628770969520636, "learning_rate": 5.760485971130969e-06, "loss": 0.9727, "step": 3379 }, { "epoch": 0.6502813717473906, "grad_norm": 1.544699938284782, "learning_rate": 5.7548424917711596e-06, "loss": 0.9349, "step": 3380 }, { "epoch": 0.6504737626857775, "grad_norm": 1.4099167078686103, "learning_rate": 5.749200661181611e-06, "loss": 0.9702, "step": 3381 }, { "epoch": 0.6506661536241644, "grad_norm": 1.5583356109854094, "learning_rate": 5.7435604815535475e-06, "loss": 0.897, "step": 3382 }, { "epoch": 0.6508585445625511, "grad_norm": 1.3503888893428149, "learning_rate": 5.7379219550775415e-06, "loss": 0.9952, "step": 3383 }, { "epoch": 0.651050935500938, "grad_norm": 1.3866930655261074, "learning_rate": 5.732285083943537e-06, "loss": 0.9215, "step": 3384 }, { "epoch": 0.6512433264393247, "grad_norm": 1.4709994327445663, "learning_rate": 5.726649870340833e-06, "loss": 0.9066, "step": 3385 }, { "epoch": 0.6514357173777116, "grad_norm": 1.3473949209161489, "learning_rate": 5.721016316458068e-06, "loss": 0.9879, "step": 3386 }, { "epoch": 0.6516281083160983, "grad_norm": 1.4053469498411106, "learning_rate": 5.715384424483268e-06, "loss": 1.0009, "step": 3387 }, { "epoch": 0.6518204992544852, "grad_norm": 1.482393723192133, "learning_rate": 5.709754196603781e-06, "loss": 0.9759, "step": 3388 }, { "epoch": 0.6520128901928719, "grad_norm": 1.2876488617472848, "learning_rate": 5.704125635006329e-06, "loss": 0.9142, "step": 3389 }, { "epoch": 0.6522052811312588, "grad_norm": 1.348381435581967, "learning_rate": 5.6984987418769825e-06, "loss": 0.9433, "step": 3390 }, { "epoch": 0.6523976720696455, "grad_norm": 1.358830786664309, "learning_rate": 5.692873519401154e-06, "loss": 1.028, "step": 3391 }, { "epoch": 0.6525900630080324, "grad_norm": 1.4094786482142252, "learning_rate": 5.6872499697636195e-06, "loss": 0.8437, "step": 3392 }, { "epoch": 0.6527824539464191, "grad_norm": 1.2754326257296882, "learning_rate": 5.681628095148502e-06, "loss": 0.9742, "step": 3393 }, { "epoch": 0.652974844884806, "grad_norm": 1.396928998448677, "learning_rate": 5.6760078977392706e-06, "loss": 0.8868, "step": 3394 }, { "epoch": 0.6531672358231927, "grad_norm": 1.3154497797676492, "learning_rate": 5.67038937971875e-06, "loss": 0.9625, "step": 3395 }, { "epoch": 0.6533596267615795, "grad_norm": 1.4973850002712585, "learning_rate": 5.664772543269101e-06, "loss": 0.9806, "step": 3396 }, { "epoch": 0.6535520176999663, "grad_norm": 0.753723142544086, "learning_rate": 5.659157390571842e-06, "loss": 0.8619, "step": 3397 }, { "epoch": 0.6537444086383531, "grad_norm": 1.4521859109003343, "learning_rate": 5.653543923807833e-06, "loss": 0.9453, "step": 3398 }, { "epoch": 0.6539367995767399, "grad_norm": 1.4285235695326495, "learning_rate": 5.6479321451572785e-06, "loss": 0.9847, "step": 3399 }, { "epoch": 0.6541291905151267, "grad_norm": 1.38348972655592, "learning_rate": 5.642322056799732e-06, "loss": 0.9769, "step": 3400 }, { "epoch": 0.6543215814535135, "grad_norm": 1.5127813358795923, "learning_rate": 5.636713660914087e-06, "loss": 0.9024, "step": 3401 }, { "epoch": 0.6545139723919003, "grad_norm": 1.433788351302863, "learning_rate": 5.631106959678575e-06, "loss": 1.0579, "step": 3402 }, { "epoch": 0.6547063633302871, "grad_norm": 1.3844213756057702, "learning_rate": 5.625501955270777e-06, "loss": 0.9289, "step": 3403 }, { "epoch": 0.6548987542686739, "grad_norm": 1.4651361004862478, "learning_rate": 5.619898649867612e-06, "loss": 0.8994, "step": 3404 }, { "epoch": 0.6550911452070608, "grad_norm": 1.518017934702041, "learning_rate": 5.614297045645339e-06, "loss": 0.9876, "step": 3405 }, { "epoch": 0.6552835361454475, "grad_norm": 1.4940799963210412, "learning_rate": 5.6086971447795625e-06, "loss": 1.0079, "step": 3406 }, { "epoch": 0.6554759270838344, "grad_norm": 1.4624307685942435, "learning_rate": 5.603098949445209e-06, "loss": 0.9656, "step": 3407 }, { "epoch": 0.6556683180222211, "grad_norm": 1.3447521676732734, "learning_rate": 5.597502461816557e-06, "loss": 1.0024, "step": 3408 }, { "epoch": 0.655860708960608, "grad_norm": 1.3771585905189274, "learning_rate": 5.5919076840672215e-06, "loss": 0.9024, "step": 3409 }, { "epoch": 0.6560530998989947, "grad_norm": 1.4919507406748125, "learning_rate": 5.5863146183701454e-06, "loss": 0.9723, "step": 3410 }, { "epoch": 0.6562454908373816, "grad_norm": 1.3717980641746792, "learning_rate": 5.580723266897616e-06, "loss": 0.8644, "step": 3411 }, { "epoch": 0.6564378817757683, "grad_norm": 1.4270108940281487, "learning_rate": 5.575133631821243e-06, "loss": 0.9258, "step": 3412 }, { "epoch": 0.6566302727141552, "grad_norm": 1.3946821335680162, "learning_rate": 5.5695457153119806e-06, "loss": 0.9928, "step": 3413 }, { "epoch": 0.6568226636525419, "grad_norm": 1.4133696143357595, "learning_rate": 5.563959519540114e-06, "loss": 0.9872, "step": 3414 }, { "epoch": 0.6570150545909288, "grad_norm": 1.4613126066605184, "learning_rate": 5.558375046675244e-06, "loss": 1.023, "step": 3415 }, { "epoch": 0.6572074455293155, "grad_norm": 1.2830791987011334, "learning_rate": 5.552792298886335e-06, "loss": 0.9414, "step": 3416 }, { "epoch": 0.6573998364677024, "grad_norm": 1.4303803615196495, "learning_rate": 5.547211278341646e-06, "loss": 1.0415, "step": 3417 }, { "epoch": 0.6575922274060891, "grad_norm": 1.4368099422172453, "learning_rate": 5.541631987208789e-06, "loss": 0.9641, "step": 3418 }, { "epoch": 0.657784618344476, "grad_norm": 1.335254446595054, "learning_rate": 5.536054427654698e-06, "loss": 0.9092, "step": 3419 }, { "epoch": 0.6579770092828627, "grad_norm": 1.6413701939116212, "learning_rate": 5.530478601845624e-06, "loss": 0.9329, "step": 3420 }, { "epoch": 0.6581694002212496, "grad_norm": 1.3571827569058512, "learning_rate": 5.52490451194716e-06, "loss": 0.9659, "step": 3421 }, { "epoch": 0.6583617911596363, "grad_norm": 1.429610863325186, "learning_rate": 5.519332160124215e-06, "loss": 1.0132, "step": 3422 }, { "epoch": 0.6585541820980232, "grad_norm": 1.3378729095418571, "learning_rate": 5.513761548541032e-06, "loss": 0.8956, "step": 3423 }, { "epoch": 0.6587465730364099, "grad_norm": 1.4345266016900666, "learning_rate": 5.508192679361169e-06, "loss": 0.8911, "step": 3424 }, { "epoch": 0.6589389639747968, "grad_norm": 1.4861796893935812, "learning_rate": 5.502625554747508e-06, "loss": 0.9092, "step": 3425 }, { "epoch": 0.6591313549131836, "grad_norm": 1.4291799877058935, "learning_rate": 5.497060176862259e-06, "loss": 1.0593, "step": 3426 }, { "epoch": 0.6593237458515704, "grad_norm": 1.420545257175911, "learning_rate": 5.491496547866948e-06, "loss": 0.9596, "step": 3427 }, { "epoch": 0.6595161367899572, "grad_norm": 1.3206318726785147, "learning_rate": 5.485934669922428e-06, "loss": 0.9497, "step": 3428 }, { "epoch": 0.659708527728344, "grad_norm": 1.3353109378005457, "learning_rate": 5.480374545188866e-06, "loss": 0.9512, "step": 3429 }, { "epoch": 0.6599009186667308, "grad_norm": 1.5488704748894888, "learning_rate": 5.474816175825754e-06, "loss": 1.0411, "step": 3430 }, { "epoch": 0.6600933096051176, "grad_norm": 1.4117317201871158, "learning_rate": 5.469259563991894e-06, "loss": 0.9724, "step": 3431 }, { "epoch": 0.6602857005435044, "grad_norm": 1.4071359530683794, "learning_rate": 5.46370471184541e-06, "loss": 0.9709, "step": 3432 }, { "epoch": 0.6604780914818912, "grad_norm": 1.4667233586676522, "learning_rate": 5.458151621543744e-06, "loss": 0.9206, "step": 3433 }, { "epoch": 0.660670482420278, "grad_norm": 1.3153752813785722, "learning_rate": 5.452600295243653e-06, "loss": 0.9971, "step": 3434 }, { "epoch": 0.6608628733586648, "grad_norm": 1.3710587056302221, "learning_rate": 5.4470507351012116e-06, "loss": 0.9606, "step": 3435 }, { "epoch": 0.6610552642970516, "grad_norm": 1.4415071883007984, "learning_rate": 5.441502943271797e-06, "loss": 0.8519, "step": 3436 }, { "epoch": 0.6612476552354384, "grad_norm": 1.384017233446616, "learning_rate": 5.4359569219101115e-06, "loss": 0.9409, "step": 3437 }, { "epoch": 0.6614400461738252, "grad_norm": 1.4056464266981963, "learning_rate": 5.430412673170167e-06, "loss": 0.9917, "step": 3438 }, { "epoch": 0.661632437112212, "grad_norm": 1.41545171826759, "learning_rate": 5.424870199205283e-06, "loss": 0.8683, "step": 3439 }, { "epoch": 0.6618248280505988, "grad_norm": 1.362022898193194, "learning_rate": 5.4193295021681e-06, "loss": 0.997, "step": 3440 }, { "epoch": 0.6620172189889856, "grad_norm": 1.4666286535972015, "learning_rate": 5.413790584210551e-06, "loss": 1.0193, "step": 3441 }, { "epoch": 0.6622096099273724, "grad_norm": 0.7592996060178209, "learning_rate": 5.408253447483892e-06, "loss": 0.9063, "step": 3442 }, { "epoch": 0.6624020008657592, "grad_norm": 1.2722399021852253, "learning_rate": 5.402718094138688e-06, "loss": 0.9638, "step": 3443 }, { "epoch": 0.662594391804146, "grad_norm": 1.5227989465636478, "learning_rate": 5.397184526324792e-06, "loss": 0.9727, "step": 3444 }, { "epoch": 0.6627867827425328, "grad_norm": 1.3912628445474973, "learning_rate": 5.391652746191398e-06, "loss": 0.9973, "step": 3445 }, { "epoch": 0.6629791736809196, "grad_norm": 1.4893390630577648, "learning_rate": 5.38612275588697e-06, "loss": 0.9911, "step": 3446 }, { "epoch": 0.6631715646193065, "grad_norm": 1.3389265074772994, "learning_rate": 5.380594557559298e-06, "loss": 0.99, "step": 3447 }, { "epoch": 0.6633639555576932, "grad_norm": 1.494908978716162, "learning_rate": 5.375068153355474e-06, "loss": 0.9872, "step": 3448 }, { "epoch": 0.6635563464960801, "grad_norm": 1.3838050409110885, "learning_rate": 5.369543545421883e-06, "loss": 0.9469, "step": 3449 }, { "epoch": 0.6637487374344668, "grad_norm": 1.4870022003371, "learning_rate": 5.364020735904223e-06, "loss": 0.9636, "step": 3450 }, { "epoch": 0.6639411283728537, "grad_norm": 1.4677740890845787, "learning_rate": 5.358499726947488e-06, "loss": 0.9591, "step": 3451 }, { "epoch": 0.6641335193112404, "grad_norm": 1.3097868078508759, "learning_rate": 5.352980520695974e-06, "loss": 0.9535, "step": 3452 }, { "epoch": 0.6643259102496273, "grad_norm": 1.3648331483652614, "learning_rate": 5.347463119293283e-06, "loss": 0.9856, "step": 3453 }, { "epoch": 0.664518301188014, "grad_norm": 1.402288902238278, "learning_rate": 5.341947524882301e-06, "loss": 1.026, "step": 3454 }, { "epoch": 0.6647106921264009, "grad_norm": 1.412845000911281, "learning_rate": 5.336433739605227e-06, "loss": 1.0098, "step": 3455 }, { "epoch": 0.6649030830647876, "grad_norm": 1.5364886226614158, "learning_rate": 5.330921765603549e-06, "loss": 1.0606, "step": 3456 }, { "epoch": 0.6650954740031745, "grad_norm": 1.3859498608119059, "learning_rate": 5.325411605018056e-06, "loss": 0.918, "step": 3457 }, { "epoch": 0.6652878649415612, "grad_norm": 1.1992715785802384, "learning_rate": 5.31990325998883e-06, "loss": 0.8898, "step": 3458 }, { "epoch": 0.6654802558799481, "grad_norm": 1.2863448835273144, "learning_rate": 5.314396732655253e-06, "loss": 0.9881, "step": 3459 }, { "epoch": 0.6656726468183348, "grad_norm": 1.4242487708124447, "learning_rate": 5.308892025155989e-06, "loss": 0.9764, "step": 3460 }, { "epoch": 0.6658650377567217, "grad_norm": 1.4114203208013947, "learning_rate": 5.303389139629007e-06, "loss": 1.0162, "step": 3461 }, { "epoch": 0.6660574286951084, "grad_norm": 1.3097994893736502, "learning_rate": 5.297888078211564e-06, "loss": 0.8415, "step": 3462 }, { "epoch": 0.6662498196334953, "grad_norm": 0.7211725413731783, "learning_rate": 5.2923888430402085e-06, "loss": 0.8642, "step": 3463 }, { "epoch": 0.666442210571882, "grad_norm": 1.3874563489790765, "learning_rate": 5.286891436250785e-06, "loss": 1.0042, "step": 3464 }, { "epoch": 0.6666346015102689, "grad_norm": 1.375343261711345, "learning_rate": 5.281395859978414e-06, "loss": 1.0292, "step": 3465 }, { "epoch": 0.6668269924486556, "grad_norm": 1.3820966539818158, "learning_rate": 5.2759021163575184e-06, "loss": 0.9681, "step": 3466 }, { "epoch": 0.6670193833870425, "grad_norm": 1.4534251114817103, "learning_rate": 5.27041020752181e-06, "loss": 0.9785, "step": 3467 }, { "epoch": 0.6672117743254292, "grad_norm": 1.323390836265294, "learning_rate": 5.26492013560427e-06, "loss": 0.9907, "step": 3468 }, { "epoch": 0.6674041652638161, "grad_norm": 1.4078620966220643, "learning_rate": 5.259431902737195e-06, "loss": 0.9322, "step": 3469 }, { "epoch": 0.6675965562022029, "grad_norm": 0.6955053421942821, "learning_rate": 5.2539455110521385e-06, "loss": 0.8347, "step": 3470 }, { "epoch": 0.6677889471405897, "grad_norm": 1.3900814687034566, "learning_rate": 5.248460962679958e-06, "loss": 0.9834, "step": 3471 }, { "epoch": 0.6679813380789765, "grad_norm": 1.3913883570714372, "learning_rate": 5.24297825975079e-06, "loss": 0.9949, "step": 3472 }, { "epoch": 0.6681737290173633, "grad_norm": 1.20634073496212, "learning_rate": 5.237497404394044e-06, "loss": 0.9563, "step": 3473 }, { "epoch": 0.6683661199557501, "grad_norm": 1.3209519206900682, "learning_rate": 5.232018398738436e-06, "loss": 0.9674, "step": 3474 }, { "epoch": 0.6685585108941369, "grad_norm": 1.3219800481356694, "learning_rate": 5.226541244911936e-06, "loss": 0.983, "step": 3475 }, { "epoch": 0.6687509018325237, "grad_norm": 1.3841462165803837, "learning_rate": 5.221065945041811e-06, "loss": 1.0275, "step": 3476 }, { "epoch": 0.6689432927709105, "grad_norm": 1.3871337346867851, "learning_rate": 5.215592501254609e-06, "loss": 0.9138, "step": 3477 }, { "epoch": 0.6691356837092973, "grad_norm": 1.4124606329661535, "learning_rate": 5.210120915676147e-06, "loss": 0.9849, "step": 3478 }, { "epoch": 0.6693280746476841, "grad_norm": 1.2816881188058895, "learning_rate": 5.2046511904315265e-06, "loss": 0.908, "step": 3479 }, { "epoch": 0.6695204655860709, "grad_norm": 1.4374556711301514, "learning_rate": 5.199183327645128e-06, "loss": 0.9844, "step": 3480 }, { "epoch": 0.6697128565244577, "grad_norm": 1.254165666832009, "learning_rate": 5.193717329440604e-06, "loss": 0.9365, "step": 3481 }, { "epoch": 0.6699052474628445, "grad_norm": 1.3340708506589596, "learning_rate": 5.188253197940889e-06, "loss": 0.9639, "step": 3482 }, { "epoch": 0.6700976384012313, "grad_norm": 1.3713201486834348, "learning_rate": 5.182790935268185e-06, "loss": 1.0191, "step": 3483 }, { "epoch": 0.6702900293396181, "grad_norm": 1.3431357929782994, "learning_rate": 5.177330543543971e-06, "loss": 0.9453, "step": 3484 }, { "epoch": 0.6704824202780049, "grad_norm": 1.3174234521957284, "learning_rate": 5.171872024889004e-06, "loss": 0.954, "step": 3485 }, { "epoch": 0.6706748112163917, "grad_norm": 1.4332658338005655, "learning_rate": 5.166415381423306e-06, "loss": 0.9933, "step": 3486 }, { "epoch": 0.6708672021547785, "grad_norm": 1.2685137621961455, "learning_rate": 5.160960615266179e-06, "loss": 0.9041, "step": 3487 }, { "epoch": 0.6710595930931653, "grad_norm": 1.403412417399053, "learning_rate": 5.155507728536191e-06, "loss": 0.9936, "step": 3488 }, { "epoch": 0.6712519840315521, "grad_norm": 1.3289648865033732, "learning_rate": 5.150056723351173e-06, "loss": 0.9651, "step": 3489 }, { "epoch": 0.6714443749699389, "grad_norm": 1.4170796898823124, "learning_rate": 5.14460760182824e-06, "loss": 0.9831, "step": 3490 }, { "epoch": 0.6716367659083258, "grad_norm": 1.4191735267604835, "learning_rate": 5.139160366083765e-06, "loss": 1.0116, "step": 3491 }, { "epoch": 0.6718291568467125, "grad_norm": 1.428858651120377, "learning_rate": 5.133715018233393e-06, "loss": 0.9536, "step": 3492 }, { "epoch": 0.6720215477850994, "grad_norm": 1.4967880878531403, "learning_rate": 5.128271560392037e-06, "loss": 0.9656, "step": 3493 }, { "epoch": 0.6722139387234861, "grad_norm": 1.3128641612970922, "learning_rate": 5.122829994673866e-06, "loss": 0.9923, "step": 3494 }, { "epoch": 0.672406329661873, "grad_norm": 1.3519045188696062, "learning_rate": 5.117390323192326e-06, "loss": 0.9366, "step": 3495 }, { "epoch": 0.6725987206002597, "grad_norm": 1.430697983415817, "learning_rate": 5.111952548060126e-06, "loss": 0.9863, "step": 3496 }, { "epoch": 0.6727911115386466, "grad_norm": 1.444423048099914, "learning_rate": 5.106516671389224e-06, "loss": 0.9479, "step": 3497 }, { "epoch": 0.6729835024770333, "grad_norm": 1.326452632482308, "learning_rate": 5.101082695290866e-06, "loss": 0.9169, "step": 3498 }, { "epoch": 0.6731758934154202, "grad_norm": 1.4770910846638616, "learning_rate": 5.0956506218755344e-06, "loss": 0.9946, "step": 3499 }, { "epoch": 0.6733682843538069, "grad_norm": 1.3274671420222566, "learning_rate": 5.09022045325299e-06, "loss": 0.9941, "step": 3500 }, { "epoch": 0.6735606752921938, "grad_norm": 1.4940613942847116, "learning_rate": 5.0847921915322486e-06, "loss": 0.9149, "step": 3501 }, { "epoch": 0.6737530662305805, "grad_norm": 1.40838413981315, "learning_rate": 5.07936583882158e-06, "loss": 0.945, "step": 3502 }, { "epoch": 0.6739454571689674, "grad_norm": 1.4588176774861663, "learning_rate": 5.073941397228518e-06, "loss": 0.9362, "step": 3503 }, { "epoch": 0.6741378481073541, "grad_norm": 1.4559488095670137, "learning_rate": 5.068518868859854e-06, "loss": 0.9393, "step": 3504 }, { "epoch": 0.674330239045741, "grad_norm": 1.4432898999570603, "learning_rate": 5.063098255821637e-06, "loss": 1.044, "step": 3505 }, { "epoch": 0.6745226299841277, "grad_norm": 1.4996635496437074, "learning_rate": 5.0576795602191734e-06, "loss": 0.9547, "step": 3506 }, { "epoch": 0.6747150209225146, "grad_norm": 1.3942467118235298, "learning_rate": 5.052262784157014e-06, "loss": 0.9859, "step": 3507 }, { "epoch": 0.6749074118609013, "grad_norm": 1.4512829447238578, "learning_rate": 5.046847929738971e-06, "loss": 0.9803, "step": 3508 }, { "epoch": 0.6750998027992882, "grad_norm": 1.404344383864989, "learning_rate": 5.041434999068127e-06, "loss": 0.9179, "step": 3509 }, { "epoch": 0.6752921937376749, "grad_norm": 0.6903088889880176, "learning_rate": 5.036023994246787e-06, "loss": 0.8498, "step": 3510 }, { "epoch": 0.6754845846760618, "grad_norm": 1.4570606744317718, "learning_rate": 5.030614917376532e-06, "loss": 0.984, "step": 3511 }, { "epoch": 0.6756769756144485, "grad_norm": 1.4931374024408275, "learning_rate": 5.025207770558176e-06, "loss": 0.9746, "step": 3512 }, { "epoch": 0.6758693665528354, "grad_norm": 1.4509502304791742, "learning_rate": 5.0198025558917985e-06, "loss": 0.9162, "step": 3513 }, { "epoch": 0.6760617574912222, "grad_norm": 1.6032259652508796, "learning_rate": 5.014399275476721e-06, "loss": 1.0005, "step": 3514 }, { "epoch": 0.676254148429609, "grad_norm": 1.3609764833138196, "learning_rate": 5.008997931411517e-06, "loss": 0.9405, "step": 3515 }, { "epoch": 0.6764465393679958, "grad_norm": 1.4189987450405586, "learning_rate": 5.003598525794002e-06, "loss": 0.9633, "step": 3516 }, { "epoch": 0.6766389303063826, "grad_norm": 1.4264991562769906, "learning_rate": 4.998201060721253e-06, "loss": 0.9283, "step": 3517 }, { "epoch": 0.6768313212447694, "grad_norm": 1.5445298458668386, "learning_rate": 4.992805538289571e-06, "loss": 0.9852, "step": 3518 }, { "epoch": 0.6770237121831562, "grad_norm": 1.5071482832363199, "learning_rate": 4.987411960594521e-06, "loss": 0.9892, "step": 3519 }, { "epoch": 0.677216103121543, "grad_norm": 1.5681409439678686, "learning_rate": 4.982020329730904e-06, "loss": 1.0468, "step": 3520 }, { "epoch": 0.6774084940599298, "grad_norm": 1.4666207986895134, "learning_rate": 4.976630647792771e-06, "loss": 0.9051, "step": 3521 }, { "epoch": 0.6776008849983166, "grad_norm": 1.3753870175846434, "learning_rate": 4.971242916873412e-06, "loss": 0.9235, "step": 3522 }, { "epoch": 0.6777932759367034, "grad_norm": 1.5042259524004948, "learning_rate": 4.965857139065354e-06, "loss": 0.9624, "step": 3523 }, { "epoch": 0.6779856668750902, "grad_norm": 1.212196723451523, "learning_rate": 4.9604733164603755e-06, "loss": 0.8642, "step": 3524 }, { "epoch": 0.678178057813477, "grad_norm": 1.3933365055760314, "learning_rate": 4.955091451149495e-06, "loss": 1.0105, "step": 3525 }, { "epoch": 0.6783704487518638, "grad_norm": 1.410596152255152, "learning_rate": 4.9497115452229535e-06, "loss": 0.9028, "step": 3526 }, { "epoch": 0.6785628396902506, "grad_norm": 1.3539934831102651, "learning_rate": 4.9443336007702614e-06, "loss": 0.9157, "step": 3527 }, { "epoch": 0.6787552306286374, "grad_norm": 1.4960218279342508, "learning_rate": 4.938957619880138e-06, "loss": 1.0228, "step": 3528 }, { "epoch": 0.6789476215670242, "grad_norm": 0.703694175439851, "learning_rate": 4.9335836046405575e-06, "loss": 0.8113, "step": 3529 }, { "epoch": 0.679140012505411, "grad_norm": 1.4263721800567286, "learning_rate": 4.928211557138728e-06, "loss": 0.9927, "step": 3530 }, { "epoch": 0.6793324034437977, "grad_norm": 1.5824197025211209, "learning_rate": 4.922841479461083e-06, "loss": 0.9522, "step": 3531 }, { "epoch": 0.6795247943821846, "grad_norm": 0.7286411080738338, "learning_rate": 4.917473373693305e-06, "loss": 0.8385, "step": 3532 }, { "epoch": 0.6797171853205713, "grad_norm": 1.5535249866749425, "learning_rate": 4.9121072419203016e-06, "loss": 0.9625, "step": 3533 }, { "epoch": 0.6799095762589582, "grad_norm": 1.4075476432835778, "learning_rate": 4.906743086226218e-06, "loss": 0.9504, "step": 3534 }, { "epoch": 0.6801019671973451, "grad_norm": 1.4900483686899748, "learning_rate": 4.901380908694434e-06, "loss": 1.0338, "step": 3535 }, { "epoch": 0.6802943581357318, "grad_norm": 1.4266042862216934, "learning_rate": 4.8960207114075495e-06, "loss": 0.9285, "step": 3536 }, { "epoch": 0.6804867490741187, "grad_norm": 1.443275987884965, "learning_rate": 4.890662496447407e-06, "loss": 1.022, "step": 3537 }, { "epoch": 0.6806791400125054, "grad_norm": 1.3640239392326063, "learning_rate": 4.8853062658950765e-06, "loss": 0.9131, "step": 3538 }, { "epoch": 0.6808715309508923, "grad_norm": 1.3125303361254068, "learning_rate": 4.879952021830856e-06, "loss": 1.015, "step": 3539 }, { "epoch": 0.681063921889279, "grad_norm": 1.5974267969311815, "learning_rate": 4.874599766334276e-06, "loss": 0.9576, "step": 3540 }, { "epoch": 0.6812563128276659, "grad_norm": 1.3072349912464079, "learning_rate": 4.8692495014840825e-06, "loss": 0.9317, "step": 3541 }, { "epoch": 0.6814487037660526, "grad_norm": 1.4095750811938508, "learning_rate": 4.863901229358261e-06, "loss": 1.0039, "step": 3542 }, { "epoch": 0.6816410947044395, "grad_norm": 1.3558686427851467, "learning_rate": 4.858554952034019e-06, "loss": 0.931, "step": 3543 }, { "epoch": 0.6818334856428262, "grad_norm": 1.3547276678380473, "learning_rate": 4.853210671587789e-06, "loss": 1.0346, "step": 3544 }, { "epoch": 0.682025876581213, "grad_norm": 1.41523905858298, "learning_rate": 4.847868390095227e-06, "loss": 0.9443, "step": 3545 }, { "epoch": 0.6822182675195998, "grad_norm": 1.3119969679881025, "learning_rate": 4.842528109631218e-06, "loss": 0.9479, "step": 3546 }, { "epoch": 0.6824106584579867, "grad_norm": 1.449059667357497, "learning_rate": 4.837189832269858e-06, "loss": 0.972, "step": 3547 }, { "epoch": 0.6826030493963734, "grad_norm": 0.7976802705782452, "learning_rate": 4.8318535600844775e-06, "loss": 0.8324, "step": 3548 }, { "epoch": 0.6827954403347603, "grad_norm": 1.400064346028505, "learning_rate": 4.8265192951476206e-06, "loss": 0.9517, "step": 3549 }, { "epoch": 0.682987831273147, "grad_norm": 1.4920466144362912, "learning_rate": 4.8211870395310556e-06, "loss": 0.9827, "step": 3550 }, { "epoch": 0.6831802222115338, "grad_norm": 1.3830441307566628, "learning_rate": 4.815856795305772e-06, "loss": 0.922, "step": 3551 }, { "epoch": 0.6833726131499206, "grad_norm": 1.3524006968369455, "learning_rate": 4.81052856454197e-06, "loss": 0.927, "step": 3552 }, { "epoch": 0.6835650040883074, "grad_norm": 1.3774807021406201, "learning_rate": 4.805202349309074e-06, "loss": 1.035, "step": 3553 }, { "epoch": 0.6837573950266942, "grad_norm": 1.4703476538281295, "learning_rate": 4.7998781516757295e-06, "loss": 1.0298, "step": 3554 }, { "epoch": 0.683949785965081, "grad_norm": 1.4722175968117348, "learning_rate": 4.794555973709783e-06, "loss": 0.9038, "step": 3555 }, { "epoch": 0.6841421769034678, "grad_norm": 1.385832519269715, "learning_rate": 4.789235817478322e-06, "loss": 0.944, "step": 3556 }, { "epoch": 0.6843345678418546, "grad_norm": 1.4801414599212166, "learning_rate": 4.783917685047621e-06, "loss": 0.9939, "step": 3557 }, { "epoch": 0.6845269587802415, "grad_norm": 1.4195599725685086, "learning_rate": 4.778601578483187e-06, "loss": 0.9987, "step": 3558 }, { "epoch": 0.6847193497186282, "grad_norm": 1.4254618737050972, "learning_rate": 4.773287499849737e-06, "loss": 1.0212, "step": 3559 }, { "epoch": 0.6849117406570151, "grad_norm": 1.369290058771687, "learning_rate": 4.767975451211191e-06, "loss": 0.9675, "step": 3560 }, { "epoch": 0.6851041315954018, "grad_norm": 1.5101771655808727, "learning_rate": 4.762665434630692e-06, "loss": 0.9637, "step": 3561 }, { "epoch": 0.6852965225337887, "grad_norm": 1.3985235299416932, "learning_rate": 4.757357452170588e-06, "loss": 0.9341, "step": 3562 }, { "epoch": 0.6854889134721754, "grad_norm": 0.764116939581802, "learning_rate": 4.752051505892438e-06, "loss": 0.9021, "step": 3563 }, { "epoch": 0.6856813044105623, "grad_norm": 1.5050991937884293, "learning_rate": 4.746747597857014e-06, "loss": 0.9533, "step": 3564 }, { "epoch": 0.685873695348949, "grad_norm": 1.4063334753778471, "learning_rate": 4.741445730124287e-06, "loss": 0.9562, "step": 3565 }, { "epoch": 0.6860660862873359, "grad_norm": 1.232042287927378, "learning_rate": 4.736145904753445e-06, "loss": 0.9425, "step": 3566 }, { "epoch": 0.6862584772257226, "grad_norm": 1.3250410086063698, "learning_rate": 4.730848123802877e-06, "loss": 0.9793, "step": 3567 }, { "epoch": 0.6864508681641095, "grad_norm": 1.5546144999467633, "learning_rate": 4.725552389330183e-06, "loss": 1.0042, "step": 3568 }, { "epoch": 0.6866432591024962, "grad_norm": 1.4740349183758619, "learning_rate": 4.720258703392161e-06, "loss": 0.9821, "step": 3569 }, { "epoch": 0.6868356500408831, "grad_norm": 1.6307433545023815, "learning_rate": 4.714967068044826e-06, "loss": 0.9942, "step": 3570 }, { "epoch": 0.6870280409792698, "grad_norm": 1.3383400229403997, "learning_rate": 4.7096774853433765e-06, "loss": 0.9069, "step": 3571 }, { "epoch": 0.6872204319176567, "grad_norm": 1.2860433712124557, "learning_rate": 4.704389957342237e-06, "loss": 0.8997, "step": 3572 }, { "epoch": 0.6874128228560434, "grad_norm": 1.488894964043863, "learning_rate": 4.699104486095008e-06, "loss": 0.938, "step": 3573 }, { "epoch": 0.6876052137944303, "grad_norm": 1.399974083927229, "learning_rate": 4.69382107365452e-06, "loss": 0.8939, "step": 3574 }, { "epoch": 0.687797604732817, "grad_norm": 1.3849596196541956, "learning_rate": 4.6885397220727855e-06, "loss": 0.9601, "step": 3575 }, { "epoch": 0.6879899956712039, "grad_norm": 1.4803610267009804, "learning_rate": 4.683260433401016e-06, "loss": 1.0196, "step": 3576 }, { "epoch": 0.6881823866095906, "grad_norm": 1.4841649392942187, "learning_rate": 4.677983209689631e-06, "loss": 1.0114, "step": 3577 }, { "epoch": 0.6883747775479775, "grad_norm": 1.3078744727919946, "learning_rate": 4.6727080529882394e-06, "loss": 0.8922, "step": 3578 }, { "epoch": 0.6885671684863643, "grad_norm": 1.445841741836568, "learning_rate": 4.667434965345654e-06, "loss": 0.9776, "step": 3579 }, { "epoch": 0.6887595594247511, "grad_norm": 1.3256896610945454, "learning_rate": 4.6621639488098856e-06, "loss": 0.9203, "step": 3580 }, { "epoch": 0.6889519503631379, "grad_norm": 1.2971629531219626, "learning_rate": 4.656895005428127e-06, "loss": 0.9455, "step": 3581 }, { "epoch": 0.6891443413015247, "grad_norm": 1.464049286507804, "learning_rate": 4.651628137246781e-06, "loss": 0.9433, "step": 3582 }, { "epoch": 0.6893367322399115, "grad_norm": 1.3366586602260293, "learning_rate": 4.6463633463114395e-06, "loss": 1.0074, "step": 3583 }, { "epoch": 0.6895291231782983, "grad_norm": 1.3207268033217459, "learning_rate": 4.641100634666877e-06, "loss": 0.9276, "step": 3584 }, { "epoch": 0.6897215141166851, "grad_norm": 1.3021717465040497, "learning_rate": 4.635840004357086e-06, "loss": 0.9063, "step": 3585 }, { "epoch": 0.6899139050550719, "grad_norm": 1.303733247995253, "learning_rate": 4.630581457425222e-06, "loss": 0.9388, "step": 3586 }, { "epoch": 0.6901062959934587, "grad_norm": 1.3384923211268258, "learning_rate": 4.625324995913648e-06, "loss": 0.9603, "step": 3587 }, { "epoch": 0.6902986869318455, "grad_norm": 1.5243234874250529, "learning_rate": 4.620070621863917e-06, "loss": 0.9398, "step": 3588 }, { "epoch": 0.6904910778702323, "grad_norm": 1.419919691807179, "learning_rate": 4.614818337316759e-06, "loss": 0.9783, "step": 3589 }, { "epoch": 0.6906834688086191, "grad_norm": 1.3170251596952032, "learning_rate": 4.609568144312107e-06, "loss": 0.9523, "step": 3590 }, { "epoch": 0.6908758597470059, "grad_norm": 0.6968730156650271, "learning_rate": 4.6043200448890724e-06, "loss": 0.8228, "step": 3591 }, { "epoch": 0.6910682506853927, "grad_norm": 1.322799580150761, "learning_rate": 4.599074041085958e-06, "loss": 0.9337, "step": 3592 }, { "epoch": 0.6912606416237795, "grad_norm": 0.7465740819016171, "learning_rate": 4.593830134940256e-06, "loss": 0.8913, "step": 3593 }, { "epoch": 0.6914530325621663, "grad_norm": 1.513631465045744, "learning_rate": 4.588588328488629e-06, "loss": 0.9666, "step": 3594 }, { "epoch": 0.6916454235005531, "grad_norm": 1.341712995808757, "learning_rate": 4.5833486237669414e-06, "loss": 0.9581, "step": 3595 }, { "epoch": 0.6918378144389399, "grad_norm": 1.3783991131532811, "learning_rate": 4.578111022810231e-06, "loss": 0.9207, "step": 3596 }, { "epoch": 0.6920302053773267, "grad_norm": 1.3839753435319417, "learning_rate": 4.5728755276527225e-06, "loss": 1.0175, "step": 3597 }, { "epoch": 0.6922225963157135, "grad_norm": 1.3862972945517822, "learning_rate": 4.567642140327823e-06, "loss": 0.8714, "step": 3598 }, { "epoch": 0.6924149872541003, "grad_norm": 1.2791822015931686, "learning_rate": 4.562410862868123e-06, "loss": 0.8694, "step": 3599 }, { "epoch": 0.6926073781924872, "grad_norm": 1.4637559825508217, "learning_rate": 4.557181697305383e-06, "loss": 0.9369, "step": 3600 }, { "epoch": 0.6927997691308739, "grad_norm": 1.2717623422767141, "learning_rate": 4.551954645670557e-06, "loss": 0.8905, "step": 3601 }, { "epoch": 0.6929921600692608, "grad_norm": 1.3235520070631355, "learning_rate": 4.546729709993762e-06, "loss": 0.8934, "step": 3602 }, { "epoch": 0.6931845510076475, "grad_norm": 0.8154990598707399, "learning_rate": 4.541506892304314e-06, "loss": 0.8468, "step": 3603 }, { "epoch": 0.6933769419460344, "grad_norm": 0.6727246773656933, "learning_rate": 4.536286194630694e-06, "loss": 0.8015, "step": 3604 }, { "epoch": 0.6935693328844211, "grad_norm": 1.4565592679065407, "learning_rate": 4.531067619000553e-06, "loss": 0.9941, "step": 3605 }, { "epoch": 0.693761723822808, "grad_norm": 1.5245726346733037, "learning_rate": 4.525851167440731e-06, "loss": 1.0231, "step": 3606 }, { "epoch": 0.6939541147611947, "grad_norm": 1.4056346221660243, "learning_rate": 4.52063684197724e-06, "loss": 0.8882, "step": 3607 }, { "epoch": 0.6941465056995816, "grad_norm": 1.3504665523069859, "learning_rate": 4.515424644635254e-06, "loss": 0.9321, "step": 3608 }, { "epoch": 0.6943388966379683, "grad_norm": 1.3458735781943634, "learning_rate": 4.510214577439146e-06, "loss": 0.966, "step": 3609 }, { "epoch": 0.6945312875763552, "grad_norm": 1.519897608975256, "learning_rate": 4.5050066424124324e-06, "loss": 0.9586, "step": 3610 }, { "epoch": 0.6947236785147419, "grad_norm": 1.4866436464779955, "learning_rate": 4.49980084157782e-06, "loss": 0.9887, "step": 3611 }, { "epoch": 0.6949160694531288, "grad_norm": 1.2586401859527, "learning_rate": 4.494597176957186e-06, "loss": 0.8365, "step": 3612 }, { "epoch": 0.6951084603915155, "grad_norm": 1.3947768053463205, "learning_rate": 4.489395650571562e-06, "loss": 1.0166, "step": 3613 }, { "epoch": 0.6953008513299024, "grad_norm": 1.3928404423564544, "learning_rate": 4.4841962644411765e-06, "loss": 0.8995, "step": 3614 }, { "epoch": 0.6954932422682891, "grad_norm": 1.3854200531899779, "learning_rate": 4.4789990205854e-06, "loss": 0.9311, "step": 3615 }, { "epoch": 0.695685633206676, "grad_norm": 1.5079921616777958, "learning_rate": 4.473803921022784e-06, "loss": 0.9744, "step": 3616 }, { "epoch": 0.6958780241450627, "grad_norm": 1.4277922076632816, "learning_rate": 4.468610967771051e-06, "loss": 0.9018, "step": 3617 }, { "epoch": 0.6960704150834496, "grad_norm": 1.345283964005257, "learning_rate": 4.4634201628470766e-06, "loss": 0.9519, "step": 3618 }, { "epoch": 0.6962628060218363, "grad_norm": 1.3812302498908358, "learning_rate": 4.458231508266912e-06, "loss": 1.0323, "step": 3619 }, { "epoch": 0.6964551969602232, "grad_norm": 1.388248140712057, "learning_rate": 4.453045006045773e-06, "loss": 1.0029, "step": 3620 }, { "epoch": 0.6966475878986099, "grad_norm": 1.3292142594525722, "learning_rate": 4.447860658198035e-06, "loss": 0.9782, "step": 3621 }, { "epoch": 0.6968399788369968, "grad_norm": 1.3849854319934267, "learning_rate": 4.442678466737245e-06, "loss": 0.9674, "step": 3622 }, { "epoch": 0.6970323697753836, "grad_norm": 1.4240094913773915, "learning_rate": 4.4374984336760975e-06, "loss": 0.9796, "step": 3623 }, { "epoch": 0.6972247607137704, "grad_norm": 0.7341333473683694, "learning_rate": 4.432320561026461e-06, "loss": 0.8557, "step": 3624 }, { "epoch": 0.6974171516521572, "grad_norm": 1.3535177538431469, "learning_rate": 4.427144850799363e-06, "loss": 0.909, "step": 3625 }, { "epoch": 0.697609542590544, "grad_norm": 1.509393365850683, "learning_rate": 4.421971305004989e-06, "loss": 1.0046, "step": 3626 }, { "epoch": 0.6978019335289308, "grad_norm": 1.4042836308564008, "learning_rate": 4.416799925652684e-06, "loss": 0.9988, "step": 3627 }, { "epoch": 0.6979943244673176, "grad_norm": 1.3692955568364489, "learning_rate": 4.411630714750956e-06, "loss": 0.878, "step": 3628 }, { "epoch": 0.6981867154057044, "grad_norm": 1.4768656642233837, "learning_rate": 4.4064636743074605e-06, "loss": 0.9521, "step": 3629 }, { "epoch": 0.6983791063440912, "grad_norm": 1.3453333514610255, "learning_rate": 4.40129880632902e-06, "loss": 0.9611, "step": 3630 }, { "epoch": 0.698571497282478, "grad_norm": 1.4073277682197218, "learning_rate": 4.396136112821608e-06, "loss": 1.0214, "step": 3631 }, { "epoch": 0.6987638882208648, "grad_norm": 1.431277960938962, "learning_rate": 4.390975595790358e-06, "loss": 1.0111, "step": 3632 }, { "epoch": 0.6989562791592516, "grad_norm": 1.4114646163543312, "learning_rate": 4.385817257239556e-06, "loss": 0.918, "step": 3633 }, { "epoch": 0.6991486700976384, "grad_norm": 1.4151482867852836, "learning_rate": 4.380661099172636e-06, "loss": 0.9828, "step": 3634 }, { "epoch": 0.6993410610360252, "grad_norm": 1.4046921286884195, "learning_rate": 4.375507123592194e-06, "loss": 0.9591, "step": 3635 }, { "epoch": 0.699533451974412, "grad_norm": 1.350010755294927, "learning_rate": 4.370355332499977e-06, "loss": 1.0273, "step": 3636 }, { "epoch": 0.6997258429127988, "grad_norm": 1.3530568462396857, "learning_rate": 4.365205727896872e-06, "loss": 0.9149, "step": 3637 }, { "epoch": 0.6999182338511856, "grad_norm": 1.393765655795396, "learning_rate": 4.36005831178294e-06, "loss": 0.9378, "step": 3638 }, { "epoch": 0.7001106247895724, "grad_norm": 1.3972830144117412, "learning_rate": 4.354913086157367e-06, "loss": 0.9995, "step": 3639 }, { "epoch": 0.7003030157279592, "grad_norm": 1.4569077551021996, "learning_rate": 4.349770053018502e-06, "loss": 1.0099, "step": 3640 }, { "epoch": 0.700495406666346, "grad_norm": 1.473986106803252, "learning_rate": 4.344629214363845e-06, "loss": 1.0298, "step": 3641 }, { "epoch": 0.7006877976047328, "grad_norm": 1.3537414380999317, "learning_rate": 4.339490572190031e-06, "loss": 1.0719, "step": 3642 }, { "epoch": 0.7008801885431196, "grad_norm": 1.4744758570629737, "learning_rate": 4.334354128492851e-06, "loss": 0.9889, "step": 3643 }, { "epoch": 0.7010725794815065, "grad_norm": 1.4969741673758954, "learning_rate": 4.329219885267244e-06, "loss": 0.9552, "step": 3644 }, { "epoch": 0.7012649704198932, "grad_norm": 1.2955803970358037, "learning_rate": 4.324087844507289e-06, "loss": 0.9418, "step": 3645 }, { "epoch": 0.7014573613582801, "grad_norm": 1.522245423401873, "learning_rate": 4.318958008206214e-06, "loss": 0.9701, "step": 3646 }, { "epoch": 0.7016497522966668, "grad_norm": 1.4899208194607956, "learning_rate": 4.313830378356384e-06, "loss": 0.984, "step": 3647 }, { "epoch": 0.7018421432350537, "grad_norm": 1.363463790770594, "learning_rate": 4.3087049569493136e-06, "loss": 0.7994, "step": 3648 }, { "epoch": 0.7020345341734404, "grad_norm": 1.4424228965743406, "learning_rate": 4.303581745975656e-06, "loss": 0.9701, "step": 3649 }, { "epoch": 0.7022269251118273, "grad_norm": 1.3745451488901859, "learning_rate": 4.2984607474252084e-06, "loss": 0.9439, "step": 3650 }, { "epoch": 0.702419316050214, "grad_norm": 1.5294046599268143, "learning_rate": 4.293341963286912e-06, "loss": 0.9085, "step": 3651 }, { "epoch": 0.7026117069886009, "grad_norm": 1.2458325589801573, "learning_rate": 4.288225395548835e-06, "loss": 0.9536, "step": 3652 }, { "epoch": 0.7028040979269876, "grad_norm": 1.5903294767123215, "learning_rate": 4.283111046198198e-06, "loss": 1.0172, "step": 3653 }, { "epoch": 0.7029964888653745, "grad_norm": 1.4152614834065576, "learning_rate": 4.277998917221354e-06, "loss": 0.9017, "step": 3654 }, { "epoch": 0.7031888798037612, "grad_norm": 1.3662481334151826, "learning_rate": 4.272889010603798e-06, "loss": 1.0306, "step": 3655 }, { "epoch": 0.7033812707421481, "grad_norm": 1.3470307584947385, "learning_rate": 4.267781328330155e-06, "loss": 0.9018, "step": 3656 }, { "epoch": 0.7035736616805348, "grad_norm": 1.252616229588948, "learning_rate": 4.262675872384197e-06, "loss": 0.989, "step": 3657 }, { "epoch": 0.7037660526189217, "grad_norm": 1.4443263111034996, "learning_rate": 4.257572644748813e-06, "loss": 0.9931, "step": 3658 }, { "epoch": 0.7039584435573084, "grad_norm": 1.4334000903281956, "learning_rate": 4.252471647406045e-06, "loss": 0.9728, "step": 3659 }, { "epoch": 0.7041508344956953, "grad_norm": 1.481002148542211, "learning_rate": 4.2473728823370605e-06, "loss": 0.9922, "step": 3660 }, { "epoch": 0.704343225434082, "grad_norm": 1.6434514856410458, "learning_rate": 4.242276351522161e-06, "loss": 0.9258, "step": 3661 }, { "epoch": 0.7045356163724689, "grad_norm": 1.306286178959286, "learning_rate": 4.237182056940784e-06, "loss": 1.0325, "step": 3662 }, { "epoch": 0.7047280073108556, "grad_norm": 1.5427163887946493, "learning_rate": 4.232090000571488e-06, "loss": 1.0185, "step": 3663 }, { "epoch": 0.7049203982492425, "grad_norm": 1.3826819743297019, "learning_rate": 4.2270001843919714e-06, "loss": 0.9994, "step": 3664 }, { "epoch": 0.7051127891876292, "grad_norm": 1.4622628708965972, "learning_rate": 4.221912610379065e-06, "loss": 1.0181, "step": 3665 }, { "epoch": 0.7053051801260161, "grad_norm": 0.7446784230784554, "learning_rate": 4.216827280508712e-06, "loss": 0.9231, "step": 3666 }, { "epoch": 0.7054975710644029, "grad_norm": 1.4275491443473132, "learning_rate": 4.211744196756011e-06, "loss": 0.9889, "step": 3667 }, { "epoch": 0.7056899620027897, "grad_norm": 1.4309519785350826, "learning_rate": 4.206663361095164e-06, "loss": 1.0199, "step": 3668 }, { "epoch": 0.7058823529411765, "grad_norm": 1.311354040304256, "learning_rate": 4.201584775499509e-06, "loss": 0.9012, "step": 3669 }, { "epoch": 0.7060747438795633, "grad_norm": 1.5718146509567796, "learning_rate": 4.196508441941516e-06, "loss": 0.9585, "step": 3670 }, { "epoch": 0.7062671348179501, "grad_norm": 1.4499719289251154, "learning_rate": 4.191434362392768e-06, "loss": 0.9735, "step": 3671 }, { "epoch": 0.7064595257563369, "grad_norm": 1.4652955825970244, "learning_rate": 4.186362538823981e-06, "loss": 1.0305, "step": 3672 }, { "epoch": 0.7066519166947237, "grad_norm": 1.5621865203589205, "learning_rate": 4.181292973204992e-06, "loss": 1.0216, "step": 3673 }, { "epoch": 0.7068443076331105, "grad_norm": 1.4519978033149188, "learning_rate": 4.1762256675047655e-06, "loss": 0.9679, "step": 3674 }, { "epoch": 0.7070366985714973, "grad_norm": 1.4528862514587748, "learning_rate": 4.171160623691384e-06, "loss": 0.9233, "step": 3675 }, { "epoch": 0.7072290895098841, "grad_norm": 1.4287541546435645, "learning_rate": 4.166097843732048e-06, "loss": 0.9504, "step": 3676 }, { "epoch": 0.7074214804482709, "grad_norm": 1.349994731940593, "learning_rate": 4.161037329593085e-06, "loss": 0.9931, "step": 3677 }, { "epoch": 0.7076138713866577, "grad_norm": 1.2876398024814384, "learning_rate": 4.155979083239942e-06, "loss": 0.9279, "step": 3678 }, { "epoch": 0.7078062623250445, "grad_norm": 1.4406957186028775, "learning_rate": 4.1509231066371815e-06, "loss": 0.9458, "step": 3679 }, { "epoch": 0.7079986532634313, "grad_norm": 0.6933221437723367, "learning_rate": 4.1458694017484915e-06, "loss": 0.7969, "step": 3680 }, { "epoch": 0.7081910442018181, "grad_norm": 1.2348711439382025, "learning_rate": 4.140817970536664e-06, "loss": 0.8964, "step": 3681 }, { "epoch": 0.7083834351402049, "grad_norm": 1.2660061243757985, "learning_rate": 4.135768814963622e-06, "loss": 0.9193, "step": 3682 }, { "epoch": 0.7085758260785917, "grad_norm": 1.5419595931300045, "learning_rate": 4.130721936990399e-06, "loss": 0.8926, "step": 3683 }, { "epoch": 0.7087682170169785, "grad_norm": 1.3660984032224899, "learning_rate": 4.1256773385771444e-06, "loss": 1.0114, "step": 3684 }, { "epoch": 0.7089606079553653, "grad_norm": 1.306160804355701, "learning_rate": 4.120635021683122e-06, "loss": 0.9901, "step": 3685 }, { "epoch": 0.709152998893752, "grad_norm": 1.500766396296523, "learning_rate": 4.115594988266711e-06, "loss": 0.9975, "step": 3686 }, { "epoch": 0.7093453898321389, "grad_norm": 1.3876158501104812, "learning_rate": 4.1105572402853976e-06, "loss": 0.9554, "step": 3687 }, { "epoch": 0.7095377807705258, "grad_norm": 1.6452547600680831, "learning_rate": 4.1055217796957895e-06, "loss": 1.0389, "step": 3688 }, { "epoch": 0.7097301717089125, "grad_norm": 1.591440209530904, "learning_rate": 4.100488608453599e-06, "loss": 0.9564, "step": 3689 }, { "epoch": 0.7099225626472994, "grad_norm": 1.5046366405534868, "learning_rate": 4.095457728513652e-06, "loss": 0.9715, "step": 3690 }, { "epoch": 0.7101149535856861, "grad_norm": 1.4333437097084163, "learning_rate": 4.09042914182989e-06, "loss": 1.0174, "step": 3691 }, { "epoch": 0.710307344524073, "grad_norm": 1.271153290453865, "learning_rate": 4.08540285035535e-06, "loss": 0.9216, "step": 3692 }, { "epoch": 0.7104997354624597, "grad_norm": 1.3605469234978307, "learning_rate": 4.0803788560421885e-06, "loss": 0.9138, "step": 3693 }, { "epoch": 0.7106921264008466, "grad_norm": 1.353157153155571, "learning_rate": 4.075357160841671e-06, "loss": 0.9067, "step": 3694 }, { "epoch": 0.7108845173392333, "grad_norm": 1.4070048527838466, "learning_rate": 4.070337766704155e-06, "loss": 1.0267, "step": 3695 }, { "epoch": 0.7110769082776202, "grad_norm": 1.3877357064896545, "learning_rate": 4.065320675579132e-06, "loss": 0.9121, "step": 3696 }, { "epoch": 0.7112692992160069, "grad_norm": 1.235903981992918, "learning_rate": 4.0603058894151685e-06, "loss": 0.8962, "step": 3697 }, { "epoch": 0.7114616901543938, "grad_norm": 1.4431262156466669, "learning_rate": 4.055293410159954e-06, "loss": 0.9224, "step": 3698 }, { "epoch": 0.7116540810927805, "grad_norm": 1.488269810460066, "learning_rate": 4.050283239760282e-06, "loss": 0.9844, "step": 3699 }, { "epoch": 0.7118464720311674, "grad_norm": 1.4829767625870076, "learning_rate": 4.045275380162038e-06, "loss": 0.9909, "step": 3700 }, { "epoch": 0.7120388629695541, "grad_norm": 1.4259971134188756, "learning_rate": 4.04026983331022e-06, "loss": 1.0178, "step": 3701 }, { "epoch": 0.712231253907941, "grad_norm": 1.433602500646847, "learning_rate": 4.035266601148924e-06, "loss": 0.9824, "step": 3702 }, { "epoch": 0.7124236448463277, "grad_norm": 0.6955355172421437, "learning_rate": 4.03026568562135e-06, "loss": 0.7937, "step": 3703 }, { "epoch": 0.7126160357847146, "grad_norm": 1.3274285939288397, "learning_rate": 4.025267088669797e-06, "loss": 0.9878, "step": 3704 }, { "epoch": 0.7128084267231013, "grad_norm": 1.5062392586693398, "learning_rate": 4.020270812235656e-06, "loss": 0.9836, "step": 3705 }, { "epoch": 0.7130008176614882, "grad_norm": 1.4174333345365593, "learning_rate": 4.015276858259427e-06, "loss": 0.9495, "step": 3706 }, { "epoch": 0.7131932085998749, "grad_norm": 1.3595526059600815, "learning_rate": 4.010285228680705e-06, "loss": 0.932, "step": 3707 }, { "epoch": 0.7133855995382617, "grad_norm": 1.4243947694011159, "learning_rate": 4.005295925438181e-06, "loss": 0.9449, "step": 3708 }, { "epoch": 0.7135779904766486, "grad_norm": 1.3037034989528198, "learning_rate": 4.000308950469646e-06, "loss": 0.9325, "step": 3709 }, { "epoch": 0.7137703814150353, "grad_norm": 1.3833105742600986, "learning_rate": 3.995324305711976e-06, "loss": 0.9182, "step": 3710 }, { "epoch": 0.7139627723534222, "grad_norm": 1.4432311453905944, "learning_rate": 3.990341993101154e-06, "loss": 0.957, "step": 3711 }, { "epoch": 0.714155163291809, "grad_norm": 1.4592587327387723, "learning_rate": 3.985362014572256e-06, "loss": 1.0285, "step": 3712 }, { "epoch": 0.7143475542301958, "grad_norm": 1.501434792720648, "learning_rate": 3.9803843720594385e-06, "loss": 0.9526, "step": 3713 }, { "epoch": 0.7145399451685825, "grad_norm": 1.4625136232303464, "learning_rate": 3.97540906749597e-06, "loss": 0.9919, "step": 3714 }, { "epoch": 0.7147323361069694, "grad_norm": 1.3871949620774664, "learning_rate": 3.970436102814203e-06, "loss": 0.9653, "step": 3715 }, { "epoch": 0.7149247270453561, "grad_norm": 1.4113831318046712, "learning_rate": 3.965465479945569e-06, "loss": 0.8569, "step": 3716 }, { "epoch": 0.715117117983743, "grad_norm": 1.4031147502515735, "learning_rate": 3.9604972008206085e-06, "loss": 0.9782, "step": 3717 }, { "epoch": 0.7153095089221297, "grad_norm": 0.6576396004005821, "learning_rate": 3.955531267368942e-06, "loss": 0.8039, "step": 3718 }, { "epoch": 0.7155018998605166, "grad_norm": 1.334586203791629, "learning_rate": 3.950567681519279e-06, "loss": 0.9162, "step": 3719 }, { "epoch": 0.7156942907989033, "grad_norm": 1.348962903263139, "learning_rate": 3.945606445199427e-06, "loss": 0.9975, "step": 3720 }, { "epoch": 0.7158866817372902, "grad_norm": 1.3061915036615228, "learning_rate": 3.940647560336262e-06, "loss": 1.0504, "step": 3721 }, { "epoch": 0.7160790726756769, "grad_norm": 1.2908747471493787, "learning_rate": 3.935691028855763e-06, "loss": 0.9231, "step": 3722 }, { "epoch": 0.7162714636140638, "grad_norm": 1.3907157384108102, "learning_rate": 3.930736852682993e-06, "loss": 1.0165, "step": 3723 }, { "epoch": 0.7164638545524505, "grad_norm": 1.3906547167688765, "learning_rate": 3.9257850337420856e-06, "loss": 0.9056, "step": 3724 }, { "epoch": 0.7166562454908374, "grad_norm": 1.4840758149281112, "learning_rate": 3.920835573956285e-06, "loss": 1.0388, "step": 3725 }, { "epoch": 0.7168486364292241, "grad_norm": 0.7081734922831159, "learning_rate": 3.915888475247894e-06, "loss": 0.8066, "step": 3726 }, { "epoch": 0.717041027367611, "grad_norm": 1.2211100364768392, "learning_rate": 3.910943739538313e-06, "loss": 0.8874, "step": 3727 }, { "epoch": 0.7172334183059977, "grad_norm": 1.3581852054195969, "learning_rate": 3.906001368748023e-06, "loss": 0.9976, "step": 3728 }, { "epoch": 0.7174258092443846, "grad_norm": 1.3292088537229179, "learning_rate": 3.901061364796574e-06, "loss": 0.9251, "step": 3729 }, { "epoch": 0.7176182001827713, "grad_norm": 1.4548513388758755, "learning_rate": 3.8961237296026155e-06, "loss": 0.967, "step": 3730 }, { "epoch": 0.7178105911211582, "grad_norm": 1.4537668948257765, "learning_rate": 3.891188465083865e-06, "loss": 0.9519, "step": 3731 }, { "epoch": 0.718002982059545, "grad_norm": 1.4223848396648953, "learning_rate": 3.886255573157121e-06, "loss": 1.0585, "step": 3732 }, { "epoch": 0.7181953729979318, "grad_norm": 1.411557340510305, "learning_rate": 3.88132505573827e-06, "loss": 0.9896, "step": 3733 }, { "epoch": 0.7183877639363186, "grad_norm": 1.4236791449100619, "learning_rate": 3.876396914742258e-06, "loss": 0.9068, "step": 3734 }, { "epoch": 0.7185801548747054, "grad_norm": 1.4408806734616693, "learning_rate": 3.871471152083121e-06, "loss": 0.9292, "step": 3735 }, { "epoch": 0.7187725458130922, "grad_norm": 1.3855259187491138, "learning_rate": 3.866547769673968e-06, "loss": 0.9588, "step": 3736 }, { "epoch": 0.718964936751479, "grad_norm": 1.5581516866280405, "learning_rate": 3.861626769426988e-06, "loss": 1.0082, "step": 3737 }, { "epoch": 0.7191573276898658, "grad_norm": 1.5246976233833682, "learning_rate": 3.8567081532534374e-06, "loss": 1.01, "step": 3738 }, { "epoch": 0.7193497186282526, "grad_norm": 1.4059457943120528, "learning_rate": 3.851791923063655e-06, "loss": 0.9425, "step": 3739 }, { "epoch": 0.7195421095666394, "grad_norm": 1.2825300745918407, "learning_rate": 3.846878080767039e-06, "loss": 0.9178, "step": 3740 }, { "epoch": 0.7197345005050262, "grad_norm": 1.404144780922453, "learning_rate": 3.841966628272079e-06, "loss": 0.9174, "step": 3741 }, { "epoch": 0.719926891443413, "grad_norm": 1.4288639980143345, "learning_rate": 3.837057567486314e-06, "loss": 0.9353, "step": 3742 }, { "epoch": 0.7201192823817998, "grad_norm": 1.3012501677059638, "learning_rate": 3.832150900316377e-06, "loss": 0.9395, "step": 3743 }, { "epoch": 0.7203116733201866, "grad_norm": 1.2769634473752196, "learning_rate": 3.827246628667962e-06, "loss": 0.9944, "step": 3744 }, { "epoch": 0.7205040642585734, "grad_norm": 1.4716861272652404, "learning_rate": 3.822344754445826e-06, "loss": 0.961, "step": 3745 }, { "epoch": 0.7206964551969602, "grad_norm": 1.4527270622416344, "learning_rate": 3.817445279553801e-06, "loss": 0.8968, "step": 3746 }, { "epoch": 0.720888846135347, "grad_norm": 1.369687798917323, "learning_rate": 3.8125482058947905e-06, "loss": 0.9411, "step": 3747 }, { "epoch": 0.7210812370737338, "grad_norm": 1.4106968511177436, "learning_rate": 3.8076535353707523e-06, "loss": 0.9696, "step": 3748 }, { "epoch": 0.7212736280121206, "grad_norm": 0.7704762165401935, "learning_rate": 3.8027612698827344e-06, "loss": 0.8868, "step": 3749 }, { "epoch": 0.7214660189505074, "grad_norm": 1.383223729149388, "learning_rate": 3.7978714113308246e-06, "loss": 0.984, "step": 3750 }, { "epoch": 0.7216584098888942, "grad_norm": 1.3867384781890926, "learning_rate": 3.7929839616141917e-06, "loss": 0.9292, "step": 3751 }, { "epoch": 0.721850800827281, "grad_norm": 1.449254749225528, "learning_rate": 3.788098922631067e-06, "loss": 0.9681, "step": 3752 }, { "epoch": 0.7220431917656679, "grad_norm": 1.3880832076648086, "learning_rate": 3.7832162962787355e-06, "loss": 0.9169, "step": 3753 }, { "epoch": 0.7222355827040546, "grad_norm": 1.5447594255073365, "learning_rate": 3.7783360844535653e-06, "loss": 0.9626, "step": 3754 }, { "epoch": 0.7224279736424415, "grad_norm": 1.332856809137198, "learning_rate": 3.773458289050963e-06, "loss": 0.8167, "step": 3755 }, { "epoch": 0.7226203645808282, "grad_norm": 1.3980832448030172, "learning_rate": 3.768582911965414e-06, "loss": 0.883, "step": 3756 }, { "epoch": 0.7228127555192151, "grad_norm": 1.3569108752212182, "learning_rate": 3.763709955090461e-06, "loss": 0.985, "step": 3757 }, { "epoch": 0.7230051464576018, "grad_norm": 1.3921668838635597, "learning_rate": 3.7588394203186963e-06, "loss": 0.9114, "step": 3758 }, { "epoch": 0.7231975373959887, "grad_norm": 1.2829436255427076, "learning_rate": 3.753971309541784e-06, "loss": 0.8989, "step": 3759 }, { "epoch": 0.7233899283343754, "grad_norm": 1.3439781409700868, "learning_rate": 3.7491056246504433e-06, "loss": 0.827, "step": 3760 }, { "epoch": 0.7235823192727623, "grad_norm": 1.4435455892819038, "learning_rate": 3.7442423675344474e-06, "loss": 1.0104, "step": 3761 }, { "epoch": 0.723774710211149, "grad_norm": 1.3720421514326335, "learning_rate": 3.739381540082635e-06, "loss": 0.9917, "step": 3762 }, { "epoch": 0.7239671011495359, "grad_norm": 1.3029451637302172, "learning_rate": 3.7345231441828876e-06, "loss": 0.9686, "step": 3763 }, { "epoch": 0.7241594920879226, "grad_norm": 1.446088783585899, "learning_rate": 3.729667181722154e-06, "loss": 1.0105, "step": 3764 }, { "epoch": 0.7243518830263095, "grad_norm": 1.4667264896463215, "learning_rate": 3.7248136545864345e-06, "loss": 0.9883, "step": 3765 }, { "epoch": 0.7245442739646962, "grad_norm": 1.5001011273943023, "learning_rate": 3.719962564660783e-06, "loss": 0.9602, "step": 3766 }, { "epoch": 0.7247366649030831, "grad_norm": 1.4050192253626517, "learning_rate": 3.7151139138293056e-06, "loss": 0.9315, "step": 3767 }, { "epoch": 0.7249290558414698, "grad_norm": 1.4273135240884605, "learning_rate": 3.7102677039751667e-06, "loss": 1.0818, "step": 3768 }, { "epoch": 0.7251214467798567, "grad_norm": 1.5966516112334195, "learning_rate": 3.705423936980572e-06, "loss": 0.9301, "step": 3769 }, { "epoch": 0.7253138377182434, "grad_norm": 1.4437533557923934, "learning_rate": 3.700582614726791e-06, "loss": 0.9355, "step": 3770 }, { "epoch": 0.7255062286566303, "grad_norm": 1.460072540712548, "learning_rate": 3.6957437390941274e-06, "loss": 1.0305, "step": 3771 }, { "epoch": 0.725698619595017, "grad_norm": 1.4164722808936356, "learning_rate": 3.6909073119619555e-06, "loss": 0.956, "step": 3772 }, { "epoch": 0.7258910105334039, "grad_norm": 1.2901821287370292, "learning_rate": 3.6860733352086866e-06, "loss": 0.9958, "step": 3773 }, { "epoch": 0.7260834014717906, "grad_norm": 1.333042052925115, "learning_rate": 3.6812418107117765e-06, "loss": 0.9512, "step": 3774 }, { "epoch": 0.7262757924101775, "grad_norm": 1.4806127725830707, "learning_rate": 3.6764127403477347e-06, "loss": 0.9664, "step": 3775 }, { "epoch": 0.7264681833485643, "grad_norm": 1.4630730913838452, "learning_rate": 3.6715861259921226e-06, "loss": 0.9379, "step": 3776 }, { "epoch": 0.7266605742869511, "grad_norm": 1.4122678873939931, "learning_rate": 3.6667619695195287e-06, "loss": 0.9333, "step": 3777 }, { "epoch": 0.7268529652253379, "grad_norm": 1.45813908045724, "learning_rate": 3.6619402728036157e-06, "loss": 0.9376, "step": 3778 }, { "epoch": 0.7270453561637247, "grad_norm": 1.4435402673048106, "learning_rate": 3.657121037717064e-06, "loss": 1.0035, "step": 3779 }, { "epoch": 0.7272377471021115, "grad_norm": 1.363458183750599, "learning_rate": 3.652304266131612e-06, "loss": 1.0473, "step": 3780 }, { "epoch": 0.7274301380404983, "grad_norm": 1.4934486627038672, "learning_rate": 3.6474899599180426e-06, "loss": 1.0713, "step": 3781 }, { "epoch": 0.7276225289788851, "grad_norm": 1.4187070521914011, "learning_rate": 3.642678120946168e-06, "loss": 0.9609, "step": 3782 }, { "epoch": 0.7278149199172719, "grad_norm": 1.4704768757967919, "learning_rate": 3.6378687510848576e-06, "loss": 0.9391, "step": 3783 }, { "epoch": 0.7280073108556587, "grad_norm": 1.4509374379730513, "learning_rate": 3.6330618522020124e-06, "loss": 1.0023, "step": 3784 }, { "epoch": 0.7281997017940455, "grad_norm": 1.429553905736306, "learning_rate": 3.6282574261645776e-06, "loss": 0.8771, "step": 3785 }, { "epoch": 0.7283920927324323, "grad_norm": 1.3790882997834293, "learning_rate": 3.62345547483854e-06, "loss": 0.9418, "step": 3786 }, { "epoch": 0.7285844836708191, "grad_norm": 1.465717261407896, "learning_rate": 3.618656000088916e-06, "loss": 1.0322, "step": 3787 }, { "epoch": 0.7287768746092059, "grad_norm": 1.4803315582859968, "learning_rate": 3.6138590037797695e-06, "loss": 0.9267, "step": 3788 }, { "epoch": 0.7289692655475927, "grad_norm": 1.6021896143985672, "learning_rate": 3.6090644877741986e-06, "loss": 0.9775, "step": 3789 }, { "epoch": 0.7291616564859795, "grad_norm": 1.3805784775941563, "learning_rate": 3.6042724539343378e-06, "loss": 0.9888, "step": 3790 }, { "epoch": 0.7293540474243663, "grad_norm": 1.438326893700135, "learning_rate": 3.599482904121361e-06, "loss": 1.0097, "step": 3791 }, { "epoch": 0.7295464383627531, "grad_norm": 1.4756639740319357, "learning_rate": 3.594695840195468e-06, "loss": 0.9143, "step": 3792 }, { "epoch": 0.7297388293011399, "grad_norm": 1.3668527288959829, "learning_rate": 3.5899112640159017e-06, "loss": 0.9909, "step": 3793 }, { "epoch": 0.7299312202395267, "grad_norm": 1.4200642865017095, "learning_rate": 3.585129177440938e-06, "loss": 0.9789, "step": 3794 }, { "epoch": 0.7301236111779135, "grad_norm": 0.699030947685413, "learning_rate": 3.580349582327882e-06, "loss": 0.8116, "step": 3795 }, { "epoch": 0.7303160021163003, "grad_norm": 1.4342563896034657, "learning_rate": 3.575572480533076e-06, "loss": 0.9871, "step": 3796 }, { "epoch": 0.7305083930546872, "grad_norm": 1.602118683909977, "learning_rate": 3.570797873911892e-06, "loss": 1.0175, "step": 3797 }, { "epoch": 0.7307007839930739, "grad_norm": 1.4384919927171373, "learning_rate": 3.566025764318728e-06, "loss": 0.9797, "step": 3798 }, { "epoch": 0.7308931749314608, "grad_norm": 1.2742605454873546, "learning_rate": 3.5612561536070213e-06, "loss": 0.836, "step": 3799 }, { "epoch": 0.7310855658698475, "grad_norm": 1.4383557853768372, "learning_rate": 3.5564890436292243e-06, "loss": 0.877, "step": 3800 }, { "epoch": 0.7312779568082344, "grad_norm": 1.3930108441598235, "learning_rate": 3.5517244362368363e-06, "loss": 1.0564, "step": 3801 }, { "epoch": 0.7314703477466211, "grad_norm": 1.5215286437890068, "learning_rate": 3.5469623332803795e-06, "loss": 0.9811, "step": 3802 }, { "epoch": 0.731662738685008, "grad_norm": 1.357853889476842, "learning_rate": 3.5422027366093893e-06, "loss": 0.9633, "step": 3803 }, { "epoch": 0.7318551296233947, "grad_norm": 1.4431580332177554, "learning_rate": 3.5374456480724427e-06, "loss": 0.9378, "step": 3804 }, { "epoch": 0.7320475205617816, "grad_norm": 1.4558557915361992, "learning_rate": 3.532691069517142e-06, "loss": 0.9779, "step": 3805 }, { "epoch": 0.7322399115001683, "grad_norm": 1.5481100496262161, "learning_rate": 3.5279390027901004e-06, "loss": 0.9936, "step": 3806 }, { "epoch": 0.7324323024385552, "grad_norm": 0.7093359822508165, "learning_rate": 3.5231894497369802e-06, "loss": 0.8152, "step": 3807 }, { "epoch": 0.7326246933769419, "grad_norm": 1.485115138074731, "learning_rate": 3.5184424122024406e-06, "loss": 0.9642, "step": 3808 }, { "epoch": 0.7328170843153288, "grad_norm": 1.2994244644978743, "learning_rate": 3.5136978920301822e-06, "loss": 0.9214, "step": 3809 }, { "epoch": 0.7330094752537155, "grad_norm": 0.698012344374361, "learning_rate": 3.508955891062924e-06, "loss": 0.8444, "step": 3810 }, { "epoch": 0.7332018661921024, "grad_norm": 1.5328459073414793, "learning_rate": 3.5042164111423983e-06, "loss": 0.9866, "step": 3811 }, { "epoch": 0.7333942571304891, "grad_norm": 1.4146514305402047, "learning_rate": 3.4994794541093667e-06, "loss": 0.9557, "step": 3812 }, { "epoch": 0.733586648068876, "grad_norm": 1.4305842676586114, "learning_rate": 3.4947450218036106e-06, "loss": 0.9374, "step": 3813 }, { "epoch": 0.7337790390072627, "grad_norm": 1.3999912810116428, "learning_rate": 3.4900131160639283e-06, "loss": 0.947, "step": 3814 }, { "epoch": 0.7339714299456496, "grad_norm": 1.331905704075772, "learning_rate": 3.485283738728139e-06, "loss": 0.9618, "step": 3815 }, { "epoch": 0.7341638208840363, "grad_norm": 1.3064185516346194, "learning_rate": 3.4805568916330747e-06, "loss": 0.9817, "step": 3816 }, { "epoch": 0.7343562118224232, "grad_norm": 1.4877865184876216, "learning_rate": 3.4758325766145896e-06, "loss": 0.9706, "step": 3817 }, { "epoch": 0.73454860276081, "grad_norm": 1.4541248216541338, "learning_rate": 3.471110795507554e-06, "loss": 1.0348, "step": 3818 }, { "epoch": 0.7347409936991968, "grad_norm": 1.3850486113762366, "learning_rate": 3.4663915501458523e-06, "loss": 0.9751, "step": 3819 }, { "epoch": 0.7349333846375836, "grad_norm": 1.3017479923435167, "learning_rate": 3.4616748423623893e-06, "loss": 0.9517, "step": 3820 }, { "epoch": 0.7351257755759704, "grad_norm": 1.380036273676804, "learning_rate": 3.4569606739890737e-06, "loss": 0.9818, "step": 3821 }, { "epoch": 0.7353181665143572, "grad_norm": 1.2740984613592259, "learning_rate": 3.452249046856836e-06, "loss": 0.954, "step": 3822 }, { "epoch": 0.735510557452744, "grad_norm": 0.7006521053893858, "learning_rate": 3.4475399627956197e-06, "loss": 0.8655, "step": 3823 }, { "epoch": 0.7357029483911308, "grad_norm": 1.3373506812149154, "learning_rate": 3.4428334236343774e-06, "loss": 0.922, "step": 3824 }, { "epoch": 0.7358953393295176, "grad_norm": 1.4241927490174278, "learning_rate": 3.438129431201075e-06, "loss": 0.943, "step": 3825 }, { "epoch": 0.7360877302679044, "grad_norm": 1.4000250418294875, "learning_rate": 3.433427987322693e-06, "loss": 1.0046, "step": 3826 }, { "epoch": 0.7362801212062912, "grad_norm": 1.3068805150887821, "learning_rate": 3.4287290938252103e-06, "loss": 1.0141, "step": 3827 }, { "epoch": 0.736472512144678, "grad_norm": 1.3655404725780618, "learning_rate": 3.424032752533627e-06, "loss": 0.9876, "step": 3828 }, { "epoch": 0.7366649030830648, "grad_norm": 1.50943453548253, "learning_rate": 3.4193389652719478e-06, "loss": 0.9581, "step": 3829 }, { "epoch": 0.7368572940214516, "grad_norm": 0.7331881589697186, "learning_rate": 3.4146477338631856e-06, "loss": 0.8695, "step": 3830 }, { "epoch": 0.7370496849598384, "grad_norm": 1.43480061481442, "learning_rate": 3.4099590601293632e-06, "loss": 0.9871, "step": 3831 }, { "epoch": 0.7372420758982252, "grad_norm": 1.3838955535732695, "learning_rate": 3.4052729458915024e-06, "loss": 0.9658, "step": 3832 }, { "epoch": 0.737434466836612, "grad_norm": 1.3007249329959272, "learning_rate": 3.4005893929696377e-06, "loss": 0.8797, "step": 3833 }, { "epoch": 0.7376268577749988, "grad_norm": 1.3881689638763424, "learning_rate": 3.3959084031828114e-06, "loss": 0.9658, "step": 3834 }, { "epoch": 0.7378192487133856, "grad_norm": 1.4591874362449293, "learning_rate": 3.3912299783490567e-06, "loss": 0.9286, "step": 3835 }, { "epoch": 0.7380116396517724, "grad_norm": 0.7352397922473509, "learning_rate": 3.3865541202854314e-06, "loss": 0.8647, "step": 3836 }, { "epoch": 0.7382040305901592, "grad_norm": 1.3232933895452152, "learning_rate": 3.3818808308079753e-06, "loss": 0.9686, "step": 3837 }, { "epoch": 0.738396421528546, "grad_norm": 1.2975335327273314, "learning_rate": 3.3772101117317437e-06, "loss": 0.988, "step": 3838 }, { "epoch": 0.7385888124669328, "grad_norm": 1.4992688558763443, "learning_rate": 3.372541964870795e-06, "loss": 0.9295, "step": 3839 }, { "epoch": 0.7387812034053196, "grad_norm": 1.4615120397121588, "learning_rate": 3.367876392038174e-06, "loss": 0.9167, "step": 3840 }, { "epoch": 0.7389735943437065, "grad_norm": 1.4130363469591514, "learning_rate": 3.363213395045941e-06, "loss": 0.9202, "step": 3841 }, { "epoch": 0.7391659852820932, "grad_norm": 1.4791660461903349, "learning_rate": 3.3585529757051504e-06, "loss": 0.9851, "step": 3842 }, { "epoch": 0.7393583762204801, "grad_norm": 1.418304162987552, "learning_rate": 3.353895135825854e-06, "loss": 1.052, "step": 3843 }, { "epoch": 0.7395507671588668, "grad_norm": 1.310710209218927, "learning_rate": 3.3492398772171074e-06, "loss": 0.9714, "step": 3844 }, { "epoch": 0.7397431580972537, "grad_norm": 1.3186874195156533, "learning_rate": 3.344587201686952e-06, "loss": 0.994, "step": 3845 }, { "epoch": 0.7399355490356404, "grad_norm": 1.3372546594867405, "learning_rate": 3.3399371110424372e-06, "loss": 0.9758, "step": 3846 }, { "epoch": 0.7401279399740273, "grad_norm": 1.3626201270056992, "learning_rate": 3.3352896070896057e-06, "loss": 0.973, "step": 3847 }, { "epoch": 0.740320330912414, "grad_norm": 1.5145605556536692, "learning_rate": 3.330644691633492e-06, "loss": 1.0063, "step": 3848 }, { "epoch": 0.7405127218508009, "grad_norm": 1.4836401479488897, "learning_rate": 3.3260023664781326e-06, "loss": 0.9656, "step": 3849 }, { "epoch": 0.7407051127891876, "grad_norm": 1.4489328013299896, "learning_rate": 3.321362633426547e-06, "loss": 1.0073, "step": 3850 }, { "epoch": 0.7408975037275745, "grad_norm": 1.5126719481525779, "learning_rate": 3.316725494280757e-06, "loss": 0.9023, "step": 3851 }, { "epoch": 0.7410898946659612, "grad_norm": 0.7456945883531305, "learning_rate": 3.3120909508417754e-06, "loss": 0.8663, "step": 3852 }, { "epoch": 0.7412822856043481, "grad_norm": 1.3701911617313525, "learning_rate": 3.307459004909599e-06, "loss": 0.9582, "step": 3853 }, { "epoch": 0.7414746765427348, "grad_norm": 1.3917351559276543, "learning_rate": 3.3028296582832285e-06, "loss": 0.9222, "step": 3854 }, { "epoch": 0.7416670674811217, "grad_norm": 1.475797490831793, "learning_rate": 3.2982029127606517e-06, "loss": 0.8585, "step": 3855 }, { "epoch": 0.7418594584195084, "grad_norm": 1.6043734742135958, "learning_rate": 3.2935787701388346e-06, "loss": 0.9912, "step": 3856 }, { "epoch": 0.7420518493578953, "grad_norm": 1.4041340235655964, "learning_rate": 3.2889572322137454e-06, "loss": 0.8949, "step": 3857 }, { "epoch": 0.742244240296282, "grad_norm": 1.5236254375008833, "learning_rate": 3.2843383007803364e-06, "loss": 0.9412, "step": 3858 }, { "epoch": 0.7424366312346689, "grad_norm": 1.3525042567198091, "learning_rate": 3.279721977632546e-06, "loss": 0.9522, "step": 3859 }, { "epoch": 0.7426290221730556, "grad_norm": 1.3460321432227496, "learning_rate": 3.275108264563306e-06, "loss": 0.8886, "step": 3860 }, { "epoch": 0.7428214131114425, "grad_norm": 1.4256666027330467, "learning_rate": 3.270497163364521e-06, "loss": 1.0276, "step": 3861 }, { "epoch": 0.7430138040498293, "grad_norm": 1.3492285410292015, "learning_rate": 3.2658886758270947e-06, "loss": 0.9154, "step": 3862 }, { "epoch": 0.743206194988216, "grad_norm": 1.4230826655557254, "learning_rate": 3.2612828037409116e-06, "loss": 0.9445, "step": 3863 }, { "epoch": 0.7433985859266029, "grad_norm": 1.4046410909390465, "learning_rate": 3.256679548894831e-06, "loss": 0.9348, "step": 3864 }, { "epoch": 0.7435909768649897, "grad_norm": 1.3280145178498135, "learning_rate": 3.252078913076718e-06, "loss": 1.0248, "step": 3865 }, { "epoch": 0.7437833678033765, "grad_norm": 1.535539545466285, "learning_rate": 3.247480898073395e-06, "loss": 0.8637, "step": 3866 }, { "epoch": 0.7439757587417632, "grad_norm": 1.3925075207357007, "learning_rate": 3.242885505670681e-06, "loss": 0.8582, "step": 3867 }, { "epoch": 0.7441681496801501, "grad_norm": 1.3733075362246348, "learning_rate": 3.238292737653379e-06, "loss": 0.9027, "step": 3868 }, { "epoch": 0.7443605406185368, "grad_norm": 1.3360581236445719, "learning_rate": 3.233702595805258e-06, "loss": 0.9468, "step": 3869 }, { "epoch": 0.7445529315569237, "grad_norm": 1.3537364548455424, "learning_rate": 3.229115081909082e-06, "loss": 0.9522, "step": 3870 }, { "epoch": 0.7447453224953104, "grad_norm": 1.4254172574554187, "learning_rate": 3.224530197746587e-06, "loss": 0.9653, "step": 3871 }, { "epoch": 0.7449377134336973, "grad_norm": 1.2761076958811535, "learning_rate": 3.2199479450984892e-06, "loss": 0.8906, "step": 3872 }, { "epoch": 0.745130104372084, "grad_norm": 1.384742592030853, "learning_rate": 3.2153683257444856e-06, "loss": 0.9827, "step": 3873 }, { "epoch": 0.7453224953104709, "grad_norm": 1.6006534093081493, "learning_rate": 3.210791341463243e-06, "loss": 0.9958, "step": 3874 }, { "epoch": 0.7455148862488576, "grad_norm": 1.4642511463423133, "learning_rate": 3.206216994032411e-06, "loss": 0.8966, "step": 3875 }, { "epoch": 0.7457072771872445, "grad_norm": 0.697033863897167, "learning_rate": 3.2016452852286127e-06, "loss": 0.8321, "step": 3876 }, { "epoch": 0.7458996681256312, "grad_norm": 1.4018065744555062, "learning_rate": 3.1970762168274495e-06, "loss": 0.9292, "step": 3877 }, { "epoch": 0.7460920590640181, "grad_norm": 0.6963664955843498, "learning_rate": 3.1925097906034962e-06, "loss": 0.8532, "step": 3878 }, { "epoch": 0.7462844500024048, "grad_norm": 1.504536483156526, "learning_rate": 3.187946008330295e-06, "loss": 0.9399, "step": 3879 }, { "epoch": 0.7464768409407917, "grad_norm": 1.3830440082011775, "learning_rate": 3.1833848717803674e-06, "loss": 0.9199, "step": 3880 }, { "epoch": 0.7466692318791784, "grad_norm": 1.4451308701170884, "learning_rate": 3.178826382725212e-06, "loss": 0.951, "step": 3881 }, { "epoch": 0.7468616228175653, "grad_norm": 1.4172870251595118, "learning_rate": 3.1742705429352827e-06, "loss": 0.9719, "step": 3882 }, { "epoch": 0.747054013755952, "grad_norm": 1.3648431090866266, "learning_rate": 3.1697173541800254e-06, "loss": 0.9388, "step": 3883 }, { "epoch": 0.7472464046943389, "grad_norm": 1.3559291489754774, "learning_rate": 3.165166818227845e-06, "loss": 0.9787, "step": 3884 }, { "epoch": 0.7474387956327257, "grad_norm": 1.3124355967105656, "learning_rate": 3.1606189368461117e-06, "loss": 0.861, "step": 3885 }, { "epoch": 0.7476311865711125, "grad_norm": 1.468346348081907, "learning_rate": 3.156073711801172e-06, "loss": 1.0249, "step": 3886 }, { "epoch": 0.7478235775094993, "grad_norm": 1.3345620461749412, "learning_rate": 3.151531144858344e-06, "loss": 0.9445, "step": 3887 }, { "epoch": 0.7480159684478861, "grad_norm": 1.5309357901639145, "learning_rate": 3.146991237781899e-06, "loss": 0.9306, "step": 3888 }, { "epoch": 0.748208359386273, "grad_norm": 1.4346415107708541, "learning_rate": 3.142453992335096e-06, "loss": 1.0445, "step": 3889 }, { "epoch": 0.7484007503246597, "grad_norm": 1.4292276696181436, "learning_rate": 3.137919410280139e-06, "loss": 0.9607, "step": 3890 }, { "epoch": 0.7485931412630465, "grad_norm": 1.46939835715986, "learning_rate": 3.1333874933782114e-06, "loss": 0.9407, "step": 3891 }, { "epoch": 0.7487855322014333, "grad_norm": 0.7473964537777738, "learning_rate": 3.128858243389461e-06, "loss": 0.8924, "step": 3892 }, { "epoch": 0.7489779231398201, "grad_norm": 1.4387013001639497, "learning_rate": 3.124331662072987e-06, "loss": 0.9824, "step": 3893 }, { "epoch": 0.7491703140782069, "grad_norm": 1.2852169941596832, "learning_rate": 3.119807751186872e-06, "loss": 0.9403, "step": 3894 }, { "epoch": 0.7493627050165937, "grad_norm": 1.427433500497107, "learning_rate": 3.1152865124881436e-06, "loss": 1.0242, "step": 3895 }, { "epoch": 0.7495550959549805, "grad_norm": 1.5498565838185678, "learning_rate": 3.110767947732801e-06, "loss": 0.9348, "step": 3896 }, { "epoch": 0.7497474868933673, "grad_norm": 1.4067921386723494, "learning_rate": 3.106252058675806e-06, "loss": 0.8983, "step": 3897 }, { "epoch": 0.7499398778317541, "grad_norm": 1.505188453938128, "learning_rate": 3.101738847071072e-06, "loss": 1.0117, "step": 3898 }, { "epoch": 0.7501322687701409, "grad_norm": 1.3362948545579123, "learning_rate": 3.097228314671481e-06, "loss": 0.8791, "step": 3899 }, { "epoch": 0.7503246597085277, "grad_norm": 1.5189971569375555, "learning_rate": 3.092720463228872e-06, "loss": 1.0204, "step": 3900 }, { "epoch": 0.7505170506469145, "grad_norm": 1.4461327444781207, "learning_rate": 3.0882152944940423e-06, "loss": 0.9623, "step": 3901 }, { "epoch": 0.7507094415853013, "grad_norm": 1.3436735943195366, "learning_rate": 3.0837128102167514e-06, "loss": 0.8916, "step": 3902 }, { "epoch": 0.7509018325236881, "grad_norm": 1.3939291082377543, "learning_rate": 3.079213012145705e-06, "loss": 0.9432, "step": 3903 }, { "epoch": 0.7510942234620749, "grad_norm": 1.4635661298467166, "learning_rate": 3.0747159020285766e-06, "loss": 0.9259, "step": 3904 }, { "epoch": 0.7512866144004617, "grad_norm": 1.3783001785475777, "learning_rate": 3.0702214816119925e-06, "loss": 0.9783, "step": 3905 }, { "epoch": 0.7514790053388486, "grad_norm": 1.3688847606474464, "learning_rate": 3.065729752641532e-06, "loss": 0.9487, "step": 3906 }, { "epoch": 0.7516713962772353, "grad_norm": 1.4869600794737055, "learning_rate": 3.0612407168617352e-06, "loss": 0.9517, "step": 3907 }, { "epoch": 0.7518637872156222, "grad_norm": 1.321064585762034, "learning_rate": 3.0567543760160866e-06, "loss": 0.8998, "step": 3908 }, { "epoch": 0.7520561781540089, "grad_norm": 1.4278567515574954, "learning_rate": 3.05227073184703e-06, "loss": 0.9916, "step": 3909 }, { "epoch": 0.7522485690923958, "grad_norm": 1.4723959694684203, "learning_rate": 3.047789786095967e-06, "loss": 0.899, "step": 3910 }, { "epoch": 0.7524409600307825, "grad_norm": 1.4527963381227236, "learning_rate": 3.0433115405032334e-06, "loss": 0.9019, "step": 3911 }, { "epoch": 0.7526333509691694, "grad_norm": 1.3619917229621412, "learning_rate": 3.0388359968081395e-06, "loss": 0.8669, "step": 3912 }, { "epoch": 0.7528257419075561, "grad_norm": 1.4670384260668972, "learning_rate": 3.034363156748933e-06, "loss": 1.0018, "step": 3913 }, { "epoch": 0.753018132845943, "grad_norm": 1.4350281532769413, "learning_rate": 3.0298930220628086e-06, "loss": 1.0343, "step": 3914 }, { "epoch": 0.7532105237843297, "grad_norm": 1.4501780906816593, "learning_rate": 3.025425594485919e-06, "loss": 1.0428, "step": 3915 }, { "epoch": 0.7534029147227166, "grad_norm": 1.4418204901859955, "learning_rate": 3.0209608757533626e-06, "loss": 0.9748, "step": 3916 }, { "epoch": 0.7535953056611033, "grad_norm": 1.3200856152701022, "learning_rate": 3.0164988675991768e-06, "loss": 0.8718, "step": 3917 }, { "epoch": 0.7537876965994902, "grad_norm": 1.3747052258041572, "learning_rate": 3.0120395717563655e-06, "loss": 0.9502, "step": 3918 }, { "epoch": 0.7539800875378769, "grad_norm": 1.4788102338187943, "learning_rate": 3.00758298995686e-06, "loss": 0.9342, "step": 3919 }, { "epoch": 0.7541724784762638, "grad_norm": 1.4909789458306442, "learning_rate": 3.0031291239315473e-06, "loss": 0.9007, "step": 3920 }, { "epoch": 0.7543648694146505, "grad_norm": 1.2653531111005116, "learning_rate": 2.9986779754102613e-06, "loss": 0.8537, "step": 3921 }, { "epoch": 0.7545572603530374, "grad_norm": 1.4703210620810856, "learning_rate": 2.9942295461217698e-06, "loss": 0.9275, "step": 3922 }, { "epoch": 0.7547496512914241, "grad_norm": 1.4388983574090792, "learning_rate": 2.9897838377937947e-06, "loss": 0.9503, "step": 3923 }, { "epoch": 0.754942042229811, "grad_norm": 1.2202704474027384, "learning_rate": 2.985340852152999e-06, "loss": 0.8131, "step": 3924 }, { "epoch": 0.7551344331681977, "grad_norm": 1.3923232293788192, "learning_rate": 2.9809005909249866e-06, "loss": 0.9082, "step": 3925 }, { "epoch": 0.7553268241065846, "grad_norm": 1.3345280364182903, "learning_rate": 2.9764630558343064e-06, "loss": 0.9348, "step": 3926 }, { "epoch": 0.7555192150449714, "grad_norm": 0.725846882307082, "learning_rate": 2.9720282486044407e-06, "loss": 0.778, "step": 3927 }, { "epoch": 0.7557116059833582, "grad_norm": 1.4134967649556274, "learning_rate": 2.9675961709578194e-06, "loss": 0.9478, "step": 3928 }, { "epoch": 0.755903996921745, "grad_norm": 1.3308038554625523, "learning_rate": 2.9631668246158105e-06, "loss": 0.9851, "step": 3929 }, { "epoch": 0.7560963878601318, "grad_norm": 1.4696665237114657, "learning_rate": 2.958740211298722e-06, "loss": 0.9268, "step": 3930 }, { "epoch": 0.7562887787985186, "grad_norm": 1.425001198314969, "learning_rate": 2.9543163327258016e-06, "loss": 0.8908, "step": 3931 }, { "epoch": 0.7564811697369054, "grad_norm": 1.35549804756954, "learning_rate": 2.949895190615227e-06, "loss": 0.971, "step": 3932 }, { "epoch": 0.7566735606752922, "grad_norm": 1.3896888748039131, "learning_rate": 2.9454767866841225e-06, "loss": 0.8923, "step": 3933 }, { "epoch": 0.756865951613679, "grad_norm": 1.3808056973685308, "learning_rate": 2.941061122648545e-06, "loss": 0.9305, "step": 3934 }, { "epoch": 0.7570583425520658, "grad_norm": 0.7071972119668947, "learning_rate": 2.9366482002234874e-06, "loss": 0.765, "step": 3935 }, { "epoch": 0.7572507334904526, "grad_norm": 1.5433047037039, "learning_rate": 2.932238021122877e-06, "loss": 0.9014, "step": 3936 }, { "epoch": 0.7574431244288394, "grad_norm": 1.4845090031092052, "learning_rate": 2.9278305870595814e-06, "loss": 1.0115, "step": 3937 }, { "epoch": 0.7576355153672262, "grad_norm": 1.5728280996071937, "learning_rate": 2.92342589974539e-06, "loss": 0.9652, "step": 3938 }, { "epoch": 0.757827906305613, "grad_norm": 1.3956137940966864, "learning_rate": 2.919023960891039e-06, "loss": 0.9931, "step": 3939 }, { "epoch": 0.7580202972439998, "grad_norm": 1.3169012014153676, "learning_rate": 2.9146247722061806e-06, "loss": 0.9745, "step": 3940 }, { "epoch": 0.7582126881823866, "grad_norm": 1.5338708543996311, "learning_rate": 2.910228335399419e-06, "loss": 0.938, "step": 3941 }, { "epoch": 0.7584050791207734, "grad_norm": 1.4792537805126755, "learning_rate": 2.90583465217828e-06, "loss": 0.9223, "step": 3942 }, { "epoch": 0.7585974700591602, "grad_norm": 1.363130373823874, "learning_rate": 2.9014437242492133e-06, "loss": 0.9571, "step": 3943 }, { "epoch": 0.758789860997547, "grad_norm": 1.3201370453298409, "learning_rate": 2.897055553317607e-06, "loss": 0.9553, "step": 3944 }, { "epoch": 0.7589822519359338, "grad_norm": 1.3545508927392331, "learning_rate": 2.89267014108778e-06, "loss": 0.9959, "step": 3945 }, { "epoch": 0.7591746428743206, "grad_norm": 1.4952573895958394, "learning_rate": 2.8882874892629654e-06, "loss": 1.0016, "step": 3946 }, { "epoch": 0.7593670338127074, "grad_norm": 1.2437033350885367, "learning_rate": 2.883907599545348e-06, "loss": 0.9615, "step": 3947 }, { "epoch": 0.7595594247510942, "grad_norm": 0.6974497540679908, "learning_rate": 2.8795304736360184e-06, "loss": 0.8212, "step": 3948 }, { "epoch": 0.759751815689481, "grad_norm": 1.3824105480927105, "learning_rate": 2.8751561132350025e-06, "loss": 0.9353, "step": 3949 }, { "epoch": 0.7599442066278679, "grad_norm": 1.3799523474087454, "learning_rate": 2.8707845200412567e-06, "loss": 0.8658, "step": 3950 }, { "epoch": 0.7601365975662546, "grad_norm": 1.4812710262763773, "learning_rate": 2.866415695752649e-06, "loss": 0.979, "step": 3951 }, { "epoch": 0.7603289885046415, "grad_norm": 0.7074212418696016, "learning_rate": 2.862049642065986e-06, "loss": 0.8253, "step": 3952 }, { "epoch": 0.7605213794430282, "grad_norm": 1.4149925268507966, "learning_rate": 2.857686360676991e-06, "loss": 1.0197, "step": 3953 }, { "epoch": 0.7607137703814151, "grad_norm": 1.341342261082781, "learning_rate": 2.853325853280312e-06, "loss": 0.9829, "step": 3954 }, { "epoch": 0.7609061613198018, "grad_norm": 1.4089212625123408, "learning_rate": 2.8489681215695242e-06, "loss": 0.9716, "step": 3955 }, { "epoch": 0.7610985522581887, "grad_norm": 1.4294912200265621, "learning_rate": 2.8446131672371136e-06, "loss": 0.9387, "step": 3956 }, { "epoch": 0.7612909431965754, "grad_norm": 1.3845759476920236, "learning_rate": 2.840260991974497e-06, "loss": 0.8807, "step": 3957 }, { "epoch": 0.7614833341349623, "grad_norm": 1.584690488582743, "learning_rate": 2.83591159747201e-06, "loss": 0.9207, "step": 3958 }, { "epoch": 0.761675725073349, "grad_norm": 1.4884826645113693, "learning_rate": 2.8315649854189066e-06, "loss": 0.9734, "step": 3959 }, { "epoch": 0.7618681160117359, "grad_norm": 1.3861568538212845, "learning_rate": 2.8272211575033635e-06, "loss": 0.8874, "step": 3960 }, { "epoch": 0.7620605069501226, "grad_norm": 1.5384628332812271, "learning_rate": 2.8228801154124687e-06, "loss": 1.0011, "step": 3961 }, { "epoch": 0.7622528978885095, "grad_norm": 1.48796943582241, "learning_rate": 2.8185418608322344e-06, "loss": 1.0338, "step": 3962 }, { "epoch": 0.7624452888268962, "grad_norm": 1.3636398404017265, "learning_rate": 2.814206395447593e-06, "loss": 0.9639, "step": 3963 }, { "epoch": 0.7626376797652831, "grad_norm": 1.4707308535845995, "learning_rate": 2.80987372094238e-06, "loss": 0.8939, "step": 3964 }, { "epoch": 0.7628300707036698, "grad_norm": 1.5063828822906489, "learning_rate": 2.805543838999364e-06, "loss": 1.0287, "step": 3965 }, { "epoch": 0.7630224616420567, "grad_norm": 1.3625688989988083, "learning_rate": 2.801216751300223e-06, "loss": 1.0562, "step": 3966 }, { "epoch": 0.7632148525804434, "grad_norm": 1.438564427705654, "learning_rate": 2.796892459525541e-06, "loss": 0.9149, "step": 3967 }, { "epoch": 0.7634072435188303, "grad_norm": 1.5897458801847741, "learning_rate": 2.7925709653548295e-06, "loss": 0.9544, "step": 3968 }, { "epoch": 0.763599634457217, "grad_norm": 1.4050473313108094, "learning_rate": 2.788252270466497e-06, "loss": 0.9426, "step": 3969 }, { "epoch": 0.7637920253956039, "grad_norm": 1.5171494069260851, "learning_rate": 2.783936376537886e-06, "loss": 0.9111, "step": 3970 }, { "epoch": 0.7639844163339907, "grad_norm": 0.7320677037496242, "learning_rate": 2.7796232852452378e-06, "loss": 0.812, "step": 3971 }, { "epoch": 0.7641768072723775, "grad_norm": 1.3536319673025363, "learning_rate": 2.775312998263703e-06, "loss": 0.9429, "step": 3972 }, { "epoch": 0.7643691982107643, "grad_norm": 1.2593990142534364, "learning_rate": 2.771005517267349e-06, "loss": 0.9219, "step": 3973 }, { "epoch": 0.7645615891491511, "grad_norm": 1.5136776861325811, "learning_rate": 2.7667008439291552e-06, "loss": 0.9297, "step": 3974 }, { "epoch": 0.7647539800875379, "grad_norm": 1.326581401764026, "learning_rate": 2.762398979920998e-06, "loss": 0.9358, "step": 3975 }, { "epoch": 0.7649463710259247, "grad_norm": 1.4370941497867764, "learning_rate": 2.7580999269136854e-06, "loss": 0.974, "step": 3976 }, { "epoch": 0.7651387619643115, "grad_norm": 1.3622245788414318, "learning_rate": 2.7538036865769093e-06, "loss": 1.005, "step": 3977 }, { "epoch": 0.7653311529026983, "grad_norm": 1.4558964814354078, "learning_rate": 2.749510260579282e-06, "loss": 0.9972, "step": 3978 }, { "epoch": 0.7655235438410851, "grad_norm": 1.4157431391991602, "learning_rate": 2.7452196505883265e-06, "loss": 1.0606, "step": 3979 }, { "epoch": 0.7657159347794719, "grad_norm": 1.2949251744453942, "learning_rate": 2.7409318582704594e-06, "loss": 0.9248, "step": 3980 }, { "epoch": 0.7659083257178587, "grad_norm": 1.447844230836848, "learning_rate": 2.736646885291011e-06, "loss": 0.9221, "step": 3981 }, { "epoch": 0.7661007166562455, "grad_norm": 1.3379098889592245, "learning_rate": 2.7323647333142176e-06, "loss": 0.949, "step": 3982 }, { "epoch": 0.7662931075946323, "grad_norm": 1.4496322360690017, "learning_rate": 2.728085404003217e-06, "loss": 0.9869, "step": 3983 }, { "epoch": 0.7664854985330191, "grad_norm": 1.4577094943402462, "learning_rate": 2.7238088990200538e-06, "loss": 0.9874, "step": 3984 }, { "epoch": 0.7666778894714059, "grad_norm": 1.383509046765144, "learning_rate": 2.7195352200256675e-06, "loss": 0.8711, "step": 3985 }, { "epoch": 0.7668702804097927, "grad_norm": 1.410220800397288, "learning_rate": 2.7152643686799095e-06, "loss": 0.9846, "step": 3986 }, { "epoch": 0.7670626713481795, "grad_norm": 1.4754633638779875, "learning_rate": 2.710996346641528e-06, "loss": 1.0082, "step": 3987 }, { "epoch": 0.7672550622865663, "grad_norm": 0.7467934634007936, "learning_rate": 2.706731155568175e-06, "loss": 0.8582, "step": 3988 }, { "epoch": 0.7674474532249531, "grad_norm": 1.4271106020375621, "learning_rate": 2.7024687971164032e-06, "loss": 0.9871, "step": 3989 }, { "epoch": 0.7676398441633399, "grad_norm": 1.3947218548601072, "learning_rate": 2.698209272941659e-06, "loss": 0.9754, "step": 3990 }, { "epoch": 0.7678322351017267, "grad_norm": 1.5142485857507884, "learning_rate": 2.693952584698294e-06, "loss": 0.9977, "step": 3991 }, { "epoch": 0.7680246260401135, "grad_norm": 1.4052237274309052, "learning_rate": 2.689698734039561e-06, "loss": 0.9679, "step": 3992 }, { "epoch": 0.7682170169785003, "grad_norm": 1.4399642948931468, "learning_rate": 2.685447722617597e-06, "loss": 0.9804, "step": 3993 }, { "epoch": 0.7684094079168872, "grad_norm": 1.388851312824852, "learning_rate": 2.6811995520834543e-06, "loss": 0.9922, "step": 3994 }, { "epoch": 0.7686017988552739, "grad_norm": 1.4805152676356874, "learning_rate": 2.676954224087075e-06, "loss": 0.939, "step": 3995 }, { "epoch": 0.7687941897936608, "grad_norm": 1.4361526865416059, "learning_rate": 2.6727117402772886e-06, "loss": 0.9573, "step": 3996 }, { "epoch": 0.7689865807320475, "grad_norm": 0.6953947048281715, "learning_rate": 2.668472102301829e-06, "loss": 0.803, "step": 3997 }, { "epoch": 0.7691789716704344, "grad_norm": 1.3013875653195246, "learning_rate": 2.664235311807327e-06, "loss": 0.9282, "step": 3998 }, { "epoch": 0.7693713626088211, "grad_norm": 1.458266814306329, "learning_rate": 2.6600013704392946e-06, "loss": 1.0179, "step": 3999 }, { "epoch": 0.769563753547208, "grad_norm": 1.3798328487938551, "learning_rate": 2.655770279842157e-06, "loss": 0.9838, "step": 4000 }, { "epoch": 0.7697561444855947, "grad_norm": 1.5282939222298777, "learning_rate": 2.651542041659211e-06, "loss": 1.0024, "step": 4001 }, { "epoch": 0.7699485354239816, "grad_norm": 1.3132693117961522, "learning_rate": 2.6473166575326603e-06, "loss": 0.9419, "step": 4002 }, { "epoch": 0.7701409263623683, "grad_norm": 1.374683753810175, "learning_rate": 2.6430941291035984e-06, "loss": 0.9577, "step": 4003 }, { "epoch": 0.7703333173007552, "grad_norm": 1.3789181200392275, "learning_rate": 2.6388744580119975e-06, "loss": 0.9738, "step": 4004 }, { "epoch": 0.7705257082391419, "grad_norm": 1.5668447325057049, "learning_rate": 2.6346576458967397e-06, "loss": 0.9904, "step": 4005 }, { "epoch": 0.7707180991775288, "grad_norm": 1.3737788960178687, "learning_rate": 2.630443694395579e-06, "loss": 0.9656, "step": 4006 }, { "epoch": 0.7709104901159155, "grad_norm": 1.5143165317122869, "learning_rate": 2.6262326051451683e-06, "loss": 0.9738, "step": 4007 }, { "epoch": 0.7711028810543024, "grad_norm": 1.4106512425741646, "learning_rate": 2.6220243797810483e-06, "loss": 0.9849, "step": 4008 }, { "epoch": 0.7712952719926891, "grad_norm": 1.3633231632249505, "learning_rate": 2.6178190199376394e-06, "loss": 1.0372, "step": 4009 }, { "epoch": 0.771487662931076, "grad_norm": 1.486379573712632, "learning_rate": 2.6136165272482596e-06, "loss": 1.0234, "step": 4010 }, { "epoch": 0.7716800538694627, "grad_norm": 0.7270892472592038, "learning_rate": 2.6094169033451066e-06, "loss": 0.9031, "step": 4011 }, { "epoch": 0.7718724448078496, "grad_norm": 1.4541131218556065, "learning_rate": 2.605220149859267e-06, "loss": 1.0423, "step": 4012 }, { "epoch": 0.7720648357462363, "grad_norm": 1.492016238403461, "learning_rate": 2.6010262684207134e-06, "loss": 0.9629, "step": 4013 }, { "epoch": 0.7722572266846232, "grad_norm": 1.2910836587779333, "learning_rate": 2.596835260658297e-06, "loss": 0.9424, "step": 4014 }, { "epoch": 0.77244961762301, "grad_norm": 1.3810312953591584, "learning_rate": 2.5926471281997577e-06, "loss": 0.9353, "step": 4015 }, { "epoch": 0.7726420085613968, "grad_norm": 1.2986407398247581, "learning_rate": 2.588461872671719e-06, "loss": 0.9012, "step": 4016 }, { "epoch": 0.7728343994997836, "grad_norm": 1.240921718565644, "learning_rate": 2.5842794956996863e-06, "loss": 0.8227, "step": 4017 }, { "epoch": 0.7730267904381704, "grad_norm": 1.573604660986559, "learning_rate": 2.580099998908049e-06, "loss": 0.9876, "step": 4018 }, { "epoch": 0.7732191813765572, "grad_norm": 1.3585916704085454, "learning_rate": 2.575923383920069e-06, "loss": 0.9066, "step": 4019 }, { "epoch": 0.773411572314944, "grad_norm": 1.455159136265093, "learning_rate": 2.5717496523578998e-06, "loss": 0.9668, "step": 4020 }, { "epoch": 0.7736039632533308, "grad_norm": 1.5509819322113167, "learning_rate": 2.5675788058425723e-06, "loss": 0.9349, "step": 4021 }, { "epoch": 0.7737963541917176, "grad_norm": 1.3201885562441644, "learning_rate": 2.563410845993988e-06, "loss": 1.0067, "step": 4022 }, { "epoch": 0.7739887451301044, "grad_norm": 0.6655875926380953, "learning_rate": 2.5592457744309405e-06, "loss": 0.8233, "step": 4023 }, { "epoch": 0.7741811360684911, "grad_norm": 1.3955962103440827, "learning_rate": 2.5550835927710982e-06, "loss": 0.9287, "step": 4024 }, { "epoch": 0.774373527006878, "grad_norm": 1.3785692240755458, "learning_rate": 2.5509243026309983e-06, "loss": 0.9978, "step": 4025 }, { "epoch": 0.7745659179452647, "grad_norm": 1.2727794026209265, "learning_rate": 2.546767905626063e-06, "loss": 0.9345, "step": 4026 }, { "epoch": 0.7747583088836516, "grad_norm": 1.396353364110735, "learning_rate": 2.5426144033705937e-06, "loss": 0.936, "step": 4027 }, { "epoch": 0.7749506998220383, "grad_norm": 1.4094069243549892, "learning_rate": 2.5384637974777513e-06, "loss": 0.956, "step": 4028 }, { "epoch": 0.7751430907604252, "grad_norm": 1.4619548533604987, "learning_rate": 2.5343160895595977e-06, "loss": 0.969, "step": 4029 }, { "epoch": 0.775335481698812, "grad_norm": 1.4657820000870372, "learning_rate": 2.530171281227044e-06, "loss": 0.964, "step": 4030 }, { "epoch": 0.7755278726371988, "grad_norm": 1.2775277962440057, "learning_rate": 2.52602937408989e-06, "loss": 0.9784, "step": 4031 }, { "epoch": 0.7757202635755855, "grad_norm": 1.4513071872577608, "learning_rate": 2.5218903697568075e-06, "loss": 1.0057, "step": 4032 }, { "epoch": 0.7759126545139724, "grad_norm": 1.4687036970593697, "learning_rate": 2.517754269835332e-06, "loss": 0.9856, "step": 4033 }, { "epoch": 0.7761050454523591, "grad_norm": 1.3242680940797715, "learning_rate": 2.5136210759318814e-06, "loss": 0.8927, "step": 4034 }, { "epoch": 0.776297436390746, "grad_norm": 1.3554522033295944, "learning_rate": 2.5094907896517383e-06, "loss": 0.9102, "step": 4035 }, { "epoch": 0.7764898273291327, "grad_norm": 1.3561480610078154, "learning_rate": 2.505363412599059e-06, "loss": 0.9865, "step": 4036 }, { "epoch": 0.7766822182675196, "grad_norm": 1.3883522401540922, "learning_rate": 2.5012389463768737e-06, "loss": 0.948, "step": 4037 }, { "epoch": 0.7768746092059065, "grad_norm": 1.459558029997432, "learning_rate": 2.4971173925870694e-06, "loss": 0.868, "step": 4038 }, { "epoch": 0.7770670001442932, "grad_norm": 1.3510031466158257, "learning_rate": 2.4929987528304144e-06, "loss": 0.9757, "step": 4039 }, { "epoch": 0.77725939108268, "grad_norm": 1.3727522212329324, "learning_rate": 2.4888830287065414e-06, "loss": 0.9723, "step": 4040 }, { "epoch": 0.7774517820210668, "grad_norm": 1.4638349351513547, "learning_rate": 2.4847702218139493e-06, "loss": 0.8741, "step": 4041 }, { "epoch": 0.7776441729594537, "grad_norm": 1.3643038766090063, "learning_rate": 2.480660333750007e-06, "loss": 0.9047, "step": 4042 }, { "epoch": 0.7778365638978404, "grad_norm": 1.4427932242551873, "learning_rate": 2.476553366110944e-06, "loss": 0.9607, "step": 4043 }, { "epoch": 0.7780289548362272, "grad_norm": 1.3681472133415138, "learning_rate": 2.4724493204918598e-06, "loss": 0.9645, "step": 4044 }, { "epoch": 0.778221345774614, "grad_norm": 1.4146398285143977, "learning_rate": 2.4683481984867207e-06, "loss": 0.944, "step": 4045 }, { "epoch": 0.7784137367130008, "grad_norm": 1.4879367149635956, "learning_rate": 2.4642500016883532e-06, "loss": 0.9602, "step": 4046 }, { "epoch": 0.7786061276513876, "grad_norm": 1.3829198738533512, "learning_rate": 2.4601547316884544e-06, "loss": 0.8551, "step": 4047 }, { "epoch": 0.7787985185897744, "grad_norm": 1.4447830391989103, "learning_rate": 2.4560623900775728e-06, "loss": 0.9033, "step": 4048 }, { "epoch": 0.7789909095281612, "grad_norm": 1.4456184034719048, "learning_rate": 2.45197297844513e-06, "loss": 1.0464, "step": 4049 }, { "epoch": 0.779183300466548, "grad_norm": 1.4465286735543108, "learning_rate": 2.447886498379409e-06, "loss": 0.9503, "step": 4050 }, { "epoch": 0.7793756914049348, "grad_norm": 1.3560882003867527, "learning_rate": 2.4438029514675444e-06, "loss": 0.8469, "step": 4051 }, { "epoch": 0.7795680823433216, "grad_norm": 1.5708056245245194, "learning_rate": 2.439722339295545e-06, "loss": 0.9489, "step": 4052 }, { "epoch": 0.7797604732817084, "grad_norm": 1.3974470160515227, "learning_rate": 2.4356446634482756e-06, "loss": 0.9392, "step": 4053 }, { "epoch": 0.7799528642200952, "grad_norm": 1.4408965510552167, "learning_rate": 2.4315699255094516e-06, "loss": 0.8906, "step": 4054 }, { "epoch": 0.780145255158482, "grad_norm": 1.4176782839445474, "learning_rate": 2.427498127061658e-06, "loss": 0.9775, "step": 4055 }, { "epoch": 0.7803376460968688, "grad_norm": 1.4109191861771164, "learning_rate": 2.423429269686336e-06, "loss": 1.0204, "step": 4056 }, { "epoch": 0.7805300370352556, "grad_norm": 0.6842567431315225, "learning_rate": 2.4193633549637765e-06, "loss": 0.831, "step": 4057 }, { "epoch": 0.7807224279736424, "grad_norm": 1.3986107805814554, "learning_rate": 2.4153003844731425e-06, "loss": 0.9014, "step": 4058 }, { "epoch": 0.7809148189120293, "grad_norm": 0.7199141218293441, "learning_rate": 2.411240359792438e-06, "loss": 0.8273, "step": 4059 }, { "epoch": 0.781107209850416, "grad_norm": 1.3705891303859379, "learning_rate": 2.407183282498534e-06, "loss": 0.9825, "step": 4060 }, { "epoch": 0.7812996007888029, "grad_norm": 1.422896290788282, "learning_rate": 2.403129154167153e-06, "loss": 0.9379, "step": 4061 }, { "epoch": 0.7814919917271896, "grad_norm": 1.363441613902711, "learning_rate": 2.3990779763728666e-06, "loss": 0.9728, "step": 4062 }, { "epoch": 0.7816843826655765, "grad_norm": 1.4075604776388757, "learning_rate": 2.3950297506891084e-06, "loss": 0.9902, "step": 4063 }, { "epoch": 0.7818767736039632, "grad_norm": 1.5385720207474785, "learning_rate": 2.390984478688164e-06, "loss": 0.984, "step": 4064 }, { "epoch": 0.7820691645423501, "grad_norm": 1.4819464786465586, "learning_rate": 2.386942161941169e-06, "loss": 0.9503, "step": 4065 }, { "epoch": 0.7822615554807368, "grad_norm": 1.431325513209866, "learning_rate": 2.3829028020181154e-06, "loss": 1.0152, "step": 4066 }, { "epoch": 0.7824539464191237, "grad_norm": 1.3320261420997785, "learning_rate": 2.3788664004878405e-06, "loss": 0.8852, "step": 4067 }, { "epoch": 0.7826463373575104, "grad_norm": 1.316367848481456, "learning_rate": 2.374832958918035e-06, "loss": 0.9617, "step": 4068 }, { "epoch": 0.7828387282958973, "grad_norm": 1.3576843458626915, "learning_rate": 2.3708024788752448e-06, "loss": 0.8929, "step": 4069 }, { "epoch": 0.783031119234284, "grad_norm": 1.366334367764102, "learning_rate": 2.3667749619248614e-06, "loss": 0.9563, "step": 4070 }, { "epoch": 0.7832235101726709, "grad_norm": 1.445810732517421, "learning_rate": 2.3627504096311273e-06, "loss": 0.9063, "step": 4071 }, { "epoch": 0.7834159011110576, "grad_norm": 1.2807815085068315, "learning_rate": 2.3587288235571258e-06, "loss": 0.9577, "step": 4072 }, { "epoch": 0.7836082920494445, "grad_norm": 1.5288572668501943, "learning_rate": 2.354710205264801e-06, "loss": 0.9701, "step": 4073 }, { "epoch": 0.7838006829878312, "grad_norm": 1.327974545689053, "learning_rate": 2.350694556314934e-06, "loss": 0.9785, "step": 4074 }, { "epoch": 0.7839930739262181, "grad_norm": 1.348848685363446, "learning_rate": 2.3466818782671597e-06, "loss": 0.8591, "step": 4075 }, { "epoch": 0.7841854648646048, "grad_norm": 1.295301916766783, "learning_rate": 2.3426721726799573e-06, "loss": 0.9551, "step": 4076 }, { "epoch": 0.7843778558029917, "grad_norm": 1.37609486528222, "learning_rate": 2.3386654411106446e-06, "loss": 0.8418, "step": 4077 }, { "epoch": 0.7845702467413784, "grad_norm": 1.3925376916331076, "learning_rate": 2.3346616851153935e-06, "loss": 0.939, "step": 4078 }, { "epoch": 0.7847626376797653, "grad_norm": 1.4711148313442768, "learning_rate": 2.330660906249218e-06, "loss": 0.9993, "step": 4079 }, { "epoch": 0.7849550286181521, "grad_norm": 1.4495581834114517, "learning_rate": 2.3266631060659685e-06, "loss": 1.0079, "step": 4080 }, { "epoch": 0.7851474195565389, "grad_norm": 1.5051298431425322, "learning_rate": 2.32266828611835e-06, "loss": 0.905, "step": 4081 }, { "epoch": 0.7853398104949257, "grad_norm": 1.4378258609232912, "learning_rate": 2.318676447957907e-06, "loss": 0.9608, "step": 4082 }, { "epoch": 0.7855322014333125, "grad_norm": 1.4805199175606463, "learning_rate": 2.3146875931350165e-06, "loss": 0.9518, "step": 4083 }, { "epoch": 0.7857245923716993, "grad_norm": 1.2950958151903023, "learning_rate": 2.310701723198908e-06, "loss": 0.9684, "step": 4084 }, { "epoch": 0.7859169833100861, "grad_norm": 1.372387691019901, "learning_rate": 2.3067188396976482e-06, "loss": 0.8463, "step": 4085 }, { "epoch": 0.7861093742484729, "grad_norm": 1.3623181529453978, "learning_rate": 2.3027389441781368e-06, "loss": 0.9269, "step": 4086 }, { "epoch": 0.7863017651868597, "grad_norm": 1.4154768591949258, "learning_rate": 2.2987620381861288e-06, "loss": 1.0014, "step": 4087 }, { "epoch": 0.7864941561252465, "grad_norm": 1.6535152309146672, "learning_rate": 2.2947881232662007e-06, "loss": 0.9345, "step": 4088 }, { "epoch": 0.7866865470636333, "grad_norm": 1.4703169321509066, "learning_rate": 2.290817200961779e-06, "loss": 0.941, "step": 4089 }, { "epoch": 0.7868789380020201, "grad_norm": 1.384636432474177, "learning_rate": 2.286849272815126e-06, "loss": 0.9335, "step": 4090 }, { "epoch": 0.7870713289404069, "grad_norm": 1.4007378533294594, "learning_rate": 2.2828843403673338e-06, "loss": 0.84, "step": 4091 }, { "epoch": 0.7872637198787937, "grad_norm": 1.286410221462768, "learning_rate": 2.2789224051583403e-06, "loss": 0.9614, "step": 4092 }, { "epoch": 0.7874561108171805, "grad_norm": 1.3293811618926679, "learning_rate": 2.274963468726914e-06, "loss": 0.9131, "step": 4093 }, { "epoch": 0.7876485017555673, "grad_norm": 1.3781641038515113, "learning_rate": 2.2710075326106618e-06, "loss": 0.9356, "step": 4094 }, { "epoch": 0.7878408926939541, "grad_norm": 1.3960740081466632, "learning_rate": 2.2670545983460245e-06, "loss": 0.9192, "step": 4095 }, { "epoch": 0.7880332836323409, "grad_norm": 1.6348798630042383, "learning_rate": 2.263104667468272e-06, "loss": 0.9308, "step": 4096 }, { "epoch": 0.7882256745707277, "grad_norm": 1.4767203596481757, "learning_rate": 2.259157741511515e-06, "loss": 0.9883, "step": 4097 }, { "epoch": 0.7884180655091145, "grad_norm": 1.2886938169410418, "learning_rate": 2.2552138220086927e-06, "loss": 0.8457, "step": 4098 }, { "epoch": 0.7886104564475013, "grad_norm": 1.339601793260176, "learning_rate": 2.2512729104915787e-06, "loss": 0.9671, "step": 4099 }, { "epoch": 0.7888028473858881, "grad_norm": 1.4180694168359516, "learning_rate": 2.2473350084907806e-06, "loss": 0.8572, "step": 4100 }, { "epoch": 0.7889952383242749, "grad_norm": 1.3348790366160987, "learning_rate": 2.243400117535729e-06, "loss": 0.9034, "step": 4101 }, { "epoch": 0.7891876292626617, "grad_norm": 1.3699381841621496, "learning_rate": 2.2394682391546928e-06, "loss": 0.983, "step": 4102 }, { "epoch": 0.7893800202010486, "grad_norm": 0.7286024785606151, "learning_rate": 2.2355393748747702e-06, "loss": 0.8255, "step": 4103 }, { "epoch": 0.7895724111394353, "grad_norm": 1.4905576359860924, "learning_rate": 2.2316135262218787e-06, "loss": 0.8924, "step": 4104 }, { "epoch": 0.7897648020778222, "grad_norm": 1.4451397265074977, "learning_rate": 2.227690694720784e-06, "loss": 0.9495, "step": 4105 }, { "epoch": 0.7899571930162089, "grad_norm": 0.705399890518167, "learning_rate": 2.223770881895061e-06, "loss": 0.7956, "step": 4106 }, { "epoch": 0.7901495839545958, "grad_norm": 1.3520536372569614, "learning_rate": 2.2198540892671215e-06, "loss": 0.9943, "step": 4107 }, { "epoch": 0.7903419748929825, "grad_norm": 1.3747247754492542, "learning_rate": 2.215940318358206e-06, "loss": 0.9417, "step": 4108 }, { "epoch": 0.7905343658313694, "grad_norm": 1.399246207652896, "learning_rate": 2.2120295706883698e-06, "loss": 0.9267, "step": 4109 }, { "epoch": 0.7907267567697561, "grad_norm": 1.4498866652419349, "learning_rate": 2.2081218477765097e-06, "loss": 0.9007, "step": 4110 }, { "epoch": 0.790919147708143, "grad_norm": 1.3866325276204463, "learning_rate": 2.204217151140342e-06, "loss": 0.8893, "step": 4111 }, { "epoch": 0.7911115386465297, "grad_norm": 1.4236073126200675, "learning_rate": 2.200315482296399e-06, "loss": 0.9603, "step": 4112 }, { "epoch": 0.7913039295849166, "grad_norm": 1.595692976489546, "learning_rate": 2.1964168427600462e-06, "loss": 0.9638, "step": 4113 }, { "epoch": 0.7914963205233033, "grad_norm": 1.3476113474115667, "learning_rate": 2.1925212340454737e-06, "loss": 0.9404, "step": 4114 }, { "epoch": 0.7916887114616902, "grad_norm": 1.3936932449408002, "learning_rate": 2.1886286576656834e-06, "loss": 0.9618, "step": 4115 }, { "epoch": 0.7918811024000769, "grad_norm": 1.2972791041741847, "learning_rate": 2.184739115132517e-06, "loss": 0.9027, "step": 4116 }, { "epoch": 0.7920734933384638, "grad_norm": 1.4026673201603694, "learning_rate": 2.1808526079566215e-06, "loss": 0.8977, "step": 4117 }, { "epoch": 0.7922658842768505, "grad_norm": 1.3290590608581234, "learning_rate": 2.1769691376474722e-06, "loss": 0.846, "step": 4118 }, { "epoch": 0.7924582752152374, "grad_norm": 1.422813947583318, "learning_rate": 2.1730887057133678e-06, "loss": 1.0059, "step": 4119 }, { "epoch": 0.7926506661536241, "grad_norm": 1.3993087405991498, "learning_rate": 2.1692113136614177e-06, "loss": 1.0498, "step": 4120 }, { "epoch": 0.792843057092011, "grad_norm": 1.4334059123408383, "learning_rate": 2.1653369629975595e-06, "loss": 1.0534, "step": 4121 }, { "epoch": 0.7930354480303977, "grad_norm": 1.388706671379058, "learning_rate": 2.1614656552265457e-06, "loss": 0.8984, "step": 4122 }, { "epoch": 0.7932278389687846, "grad_norm": 1.3849313151415072, "learning_rate": 2.157597391851949e-06, "loss": 1.0268, "step": 4123 }, { "epoch": 0.7934202299071714, "grad_norm": 1.4257997716030468, "learning_rate": 2.1537321743761587e-06, "loss": 0.9785, "step": 4124 }, { "epoch": 0.7936126208455582, "grad_norm": 1.4880838423885188, "learning_rate": 2.1498700043003773e-06, "loss": 0.98, "step": 4125 }, { "epoch": 0.793805011783945, "grad_norm": 1.3200174749336953, "learning_rate": 2.1460108831246295e-06, "loss": 0.9348, "step": 4126 }, { "epoch": 0.7939974027223318, "grad_norm": 1.3842759122068176, "learning_rate": 2.142154812347753e-06, "loss": 0.98, "step": 4127 }, { "epoch": 0.7941897936607186, "grad_norm": 1.3247978808751015, "learning_rate": 2.1383017934674012e-06, "loss": 1.007, "step": 4128 }, { "epoch": 0.7943821845991054, "grad_norm": 1.4025874256231985, "learning_rate": 2.1344518279800454e-06, "loss": 0.9394, "step": 4129 }, { "epoch": 0.7945745755374922, "grad_norm": 1.3815300370827825, "learning_rate": 2.130604917380962e-06, "loss": 0.9984, "step": 4130 }, { "epoch": 0.794766966475879, "grad_norm": 1.2310473552175063, "learning_rate": 2.12676106316425e-06, "loss": 0.894, "step": 4131 }, { "epoch": 0.7949593574142658, "grad_norm": 1.3454904471302553, "learning_rate": 2.1229202668228197e-06, "loss": 0.9383, "step": 4132 }, { "epoch": 0.7951517483526526, "grad_norm": 0.6978462554682643, "learning_rate": 2.1190825298483855e-06, "loss": 0.7238, "step": 4133 }, { "epoch": 0.7953441392910394, "grad_norm": 1.4052277552355175, "learning_rate": 2.115247853731488e-06, "loss": 0.9409, "step": 4134 }, { "epoch": 0.7955365302294262, "grad_norm": 0.7111393947430068, "learning_rate": 2.11141623996147e-06, "loss": 0.8455, "step": 4135 }, { "epoch": 0.795728921167813, "grad_norm": 1.3405478931476407, "learning_rate": 2.107587690026481e-06, "loss": 0.9115, "step": 4136 }, { "epoch": 0.7959213121061998, "grad_norm": 1.2339889583235373, "learning_rate": 2.103762205413493e-06, "loss": 0.9555, "step": 4137 }, { "epoch": 0.7961137030445866, "grad_norm": 1.3428597506518156, "learning_rate": 2.0999397876082726e-06, "loss": 0.9663, "step": 4138 }, { "epoch": 0.7963060939829734, "grad_norm": 1.459974801923927, "learning_rate": 2.096120438095404e-06, "loss": 0.9673, "step": 4139 }, { "epoch": 0.7964984849213602, "grad_norm": 1.4823735918900274, "learning_rate": 2.092304158358286e-06, "loss": 0.928, "step": 4140 }, { "epoch": 0.796690875859747, "grad_norm": 1.244205172796557, "learning_rate": 2.0884909498791106e-06, "loss": 0.9175, "step": 4141 }, { "epoch": 0.7968832667981338, "grad_norm": 1.3730995811778872, "learning_rate": 2.0846808141388852e-06, "loss": 0.936, "step": 4142 }, { "epoch": 0.7970756577365206, "grad_norm": 1.5697434225867544, "learning_rate": 2.080873752617426e-06, "loss": 0.9329, "step": 4143 }, { "epoch": 0.7972680486749074, "grad_norm": 1.355545248125749, "learning_rate": 2.0770697667933436e-06, "loss": 0.9907, "step": 4144 }, { "epoch": 0.7974604396132942, "grad_norm": 1.4382823538644163, "learning_rate": 2.073268858144074e-06, "loss": 1.0245, "step": 4145 }, { "epoch": 0.797652830551681, "grad_norm": 1.42076304563268, "learning_rate": 2.0694710281458372e-06, "loss": 0.9812, "step": 4146 }, { "epoch": 0.7978452214900679, "grad_norm": 1.197827136162647, "learning_rate": 2.0656762782736693e-06, "loss": 0.9556, "step": 4147 }, { "epoch": 0.7980376124284546, "grad_norm": 1.419223036366402, "learning_rate": 2.061884610001411e-06, "loss": 0.9743, "step": 4148 }, { "epoch": 0.7982300033668415, "grad_norm": 1.3807159739272352, "learning_rate": 2.0580960248016966e-06, "loss": 0.9792, "step": 4149 }, { "epoch": 0.7984223943052282, "grad_norm": 1.3296278610728145, "learning_rate": 2.0543105241459713e-06, "loss": 0.9444, "step": 4150 }, { "epoch": 0.7986147852436151, "grad_norm": 1.4977911733945628, "learning_rate": 2.0505281095044804e-06, "loss": 0.9284, "step": 4151 }, { "epoch": 0.7988071761820018, "grad_norm": 1.4217742001354814, "learning_rate": 2.0467487823462696e-06, "loss": 0.9322, "step": 4152 }, { "epoch": 0.7989995671203887, "grad_norm": 1.4809486241081118, "learning_rate": 2.042972544139189e-06, "loss": 0.9553, "step": 4153 }, { "epoch": 0.7991919580587754, "grad_norm": 1.4437834858942007, "learning_rate": 2.039199396349881e-06, "loss": 0.923, "step": 4154 }, { "epoch": 0.7993843489971623, "grad_norm": 1.536627594734291, "learning_rate": 2.0354293404437963e-06, "loss": 0.882, "step": 4155 }, { "epoch": 0.799576739935549, "grad_norm": 1.2804275646875838, "learning_rate": 2.0316623778851784e-06, "loss": 0.999, "step": 4156 }, { "epoch": 0.7997691308739359, "grad_norm": 1.3022888964491681, "learning_rate": 2.0278985101370753e-06, "loss": 0.8348, "step": 4157 }, { "epoch": 0.7999615218123226, "grad_norm": 1.3864991907228008, "learning_rate": 2.024137738661329e-06, "loss": 0.976, "step": 4158 }, { "epoch": 0.8001539127507095, "grad_norm": 1.4153965702717888, "learning_rate": 2.020380064918579e-06, "loss": 0.9276, "step": 4159 }, { "epoch": 0.8003463036890962, "grad_norm": 1.39393251721052, "learning_rate": 2.01662549036826e-06, "loss": 0.9392, "step": 4160 }, { "epoch": 0.8005386946274831, "grad_norm": 1.4012410979213596, "learning_rate": 2.0128740164686134e-06, "loss": 0.9432, "step": 4161 }, { "epoch": 0.8007310855658698, "grad_norm": 1.3749466057987776, "learning_rate": 2.009125644676656e-06, "loss": 0.9417, "step": 4162 }, { "epoch": 0.8009234765042567, "grad_norm": 1.4246196431052454, "learning_rate": 2.0053803764482226e-06, "loss": 0.9075, "step": 4163 }, { "epoch": 0.8011158674426434, "grad_norm": 1.4165824286782116, "learning_rate": 2.001638213237932e-06, "loss": 0.9795, "step": 4164 }, { "epoch": 0.8013082583810303, "grad_norm": 1.5598170706863446, "learning_rate": 1.997899156499191e-06, "loss": 0.9705, "step": 4165 }, { "epoch": 0.801500649319417, "grad_norm": 1.5518990047952497, "learning_rate": 1.994163207684212e-06, "loss": 0.9705, "step": 4166 }, { "epoch": 0.8016930402578039, "grad_norm": 1.3665904810883098, "learning_rate": 1.9904303682439896e-06, "loss": 0.9499, "step": 4167 }, { "epoch": 0.8018854311961907, "grad_norm": 1.539639025323638, "learning_rate": 1.986700639628316e-06, "loss": 0.9126, "step": 4168 }, { "epoch": 0.8020778221345775, "grad_norm": 1.5759091179315816, "learning_rate": 1.9829740232857807e-06, "loss": 0.9257, "step": 4169 }, { "epoch": 0.8022702130729643, "grad_norm": 0.6993117399568651, "learning_rate": 1.9792505206637523e-06, "loss": 0.8546, "step": 4170 }, { "epoch": 0.802462604011351, "grad_norm": 1.4256577139998492, "learning_rate": 1.9755301332083997e-06, "loss": 0.9741, "step": 4171 }, { "epoch": 0.8026549949497379, "grad_norm": 1.3697769296999194, "learning_rate": 1.9718128623646792e-06, "loss": 0.9208, "step": 4172 }, { "epoch": 0.8028473858881247, "grad_norm": 0.7087829435249055, "learning_rate": 1.9680987095763315e-06, "loss": 0.8773, "step": 4173 }, { "epoch": 0.8030397768265115, "grad_norm": 1.4636525365044981, "learning_rate": 1.964387676285894e-06, "loss": 1.0434, "step": 4174 }, { "epoch": 0.8032321677648983, "grad_norm": 1.3470172621173562, "learning_rate": 1.9606797639346874e-06, "loss": 0.8795, "step": 4175 }, { "epoch": 0.8034245587032851, "grad_norm": 1.34728314616772, "learning_rate": 1.9569749739628243e-06, "loss": 1.0061, "step": 4176 }, { "epoch": 0.8036169496416719, "grad_norm": 1.4797354490681627, "learning_rate": 1.9532733078092034e-06, "loss": 0.9551, "step": 4177 }, { "epoch": 0.8038093405800587, "grad_norm": 1.478332133510924, "learning_rate": 1.9495747669115062e-06, "loss": 0.9925, "step": 4178 }, { "epoch": 0.8040017315184455, "grad_norm": 1.342000620703085, "learning_rate": 1.9458793527062035e-06, "loss": 0.9085, "step": 4179 }, { "epoch": 0.8041941224568323, "grad_norm": 0.7187060142249263, "learning_rate": 1.9421870666285523e-06, "loss": 0.7967, "step": 4180 }, { "epoch": 0.804386513395219, "grad_norm": 1.392078331430324, "learning_rate": 1.9384979101125944e-06, "loss": 0.9054, "step": 4181 }, { "epoch": 0.8045789043336059, "grad_norm": 1.3152583224316807, "learning_rate": 1.934811884591159e-06, "loss": 0.9357, "step": 4182 }, { "epoch": 0.8047712952719926, "grad_norm": 1.3181437338191748, "learning_rate": 1.93112899149585e-06, "loss": 0.8668, "step": 4183 }, { "epoch": 0.8049636862103795, "grad_norm": 1.5104222689585909, "learning_rate": 1.9274492322570616e-06, "loss": 0.9227, "step": 4184 }, { "epoch": 0.8051560771487662, "grad_norm": 1.4367595031517675, "learning_rate": 1.923772608303972e-06, "loss": 0.9272, "step": 4185 }, { "epoch": 0.8053484680871531, "grad_norm": 1.4033306986201477, "learning_rate": 1.9200991210645394e-06, "loss": 0.918, "step": 4186 }, { "epoch": 0.8055408590255398, "grad_norm": 1.4387562350190266, "learning_rate": 1.916428771965506e-06, "loss": 0.9656, "step": 4187 }, { "epoch": 0.8057332499639267, "grad_norm": 1.4144502361243674, "learning_rate": 1.912761562432388e-06, "loss": 1.0151, "step": 4188 }, { "epoch": 0.8059256409023136, "grad_norm": 1.3815675951340805, "learning_rate": 1.9090974938894902e-06, "loss": 0.8995, "step": 4189 }, { "epoch": 0.8061180318407003, "grad_norm": 1.3915982370006845, "learning_rate": 1.9054365677598963e-06, "loss": 0.966, "step": 4190 }, { "epoch": 0.8063104227790872, "grad_norm": 1.4142715570870172, "learning_rate": 1.9017787854654613e-06, "loss": 0.9501, "step": 4191 }, { "epoch": 0.8065028137174739, "grad_norm": 1.442650884425548, "learning_rate": 1.898124148426832e-06, "loss": 0.9418, "step": 4192 }, { "epoch": 0.8066952046558608, "grad_norm": 1.404058750259977, "learning_rate": 1.8944726580634287e-06, "loss": 0.9724, "step": 4193 }, { "epoch": 0.8068875955942475, "grad_norm": 1.4272986962790661, "learning_rate": 1.8908243157934424e-06, "loss": 0.9719, "step": 4194 }, { "epoch": 0.8070799865326344, "grad_norm": 1.3081760813102663, "learning_rate": 1.8871791230338499e-06, "loss": 0.956, "step": 4195 }, { "epoch": 0.8072723774710211, "grad_norm": 1.3042658675984282, "learning_rate": 1.883537081200404e-06, "loss": 0.9087, "step": 4196 }, { "epoch": 0.807464768409408, "grad_norm": 1.4341108117093442, "learning_rate": 1.8798981917076254e-06, "loss": 0.9715, "step": 4197 }, { "epoch": 0.8076571593477947, "grad_norm": 1.2956591069530434, "learning_rate": 1.876262455968826e-06, "loss": 0.9386, "step": 4198 }, { "epoch": 0.8078495502861816, "grad_norm": 1.4176247038728933, "learning_rate": 1.872629875396076e-06, "loss": 0.935, "step": 4199 }, { "epoch": 0.8080419412245683, "grad_norm": 1.5091893996499564, "learning_rate": 1.8690004514002314e-06, "loss": 0.9553, "step": 4200 }, { "epoch": 0.8082343321629551, "grad_norm": 1.3602389625332953, "learning_rate": 1.8653741853909201e-06, "loss": 0.9286, "step": 4201 }, { "epoch": 0.8084267231013419, "grad_norm": 1.5540364604533552, "learning_rate": 1.861751078776538e-06, "loss": 0.9121, "step": 4202 }, { "epoch": 0.8086191140397287, "grad_norm": 1.362607895517618, "learning_rate": 1.8581311329642592e-06, "loss": 0.8482, "step": 4203 }, { "epoch": 0.8088115049781155, "grad_norm": 1.4389759994831397, "learning_rate": 1.8545143493600293e-06, "loss": 1.0089, "step": 4204 }, { "epoch": 0.8090038959165023, "grad_norm": 1.303013108299279, "learning_rate": 1.8509007293685666e-06, "loss": 1.0394, "step": 4205 }, { "epoch": 0.8091962868548891, "grad_norm": 1.4840655459449381, "learning_rate": 1.8472902743933608e-06, "loss": 1.0358, "step": 4206 }, { "epoch": 0.809388677793276, "grad_norm": 1.4031961470105867, "learning_rate": 1.8436829858366655e-06, "loss": 0.9176, "step": 4207 }, { "epoch": 0.8095810687316627, "grad_norm": 1.4318028232558344, "learning_rate": 1.8400788650995137e-06, "loss": 0.898, "step": 4208 }, { "epoch": 0.8097734596700495, "grad_norm": 1.3112841394345964, "learning_rate": 1.8364779135817045e-06, "loss": 0.9179, "step": 4209 }, { "epoch": 0.8099658506084363, "grad_norm": 1.4674187154779785, "learning_rate": 1.8328801326818045e-06, "loss": 0.919, "step": 4210 }, { "epoch": 0.8101582415468231, "grad_norm": 1.2723583575971176, "learning_rate": 1.829285523797155e-06, "loss": 0.9141, "step": 4211 }, { "epoch": 0.81035063248521, "grad_norm": 1.459749845284315, "learning_rate": 1.8256940883238538e-06, "loss": 0.9819, "step": 4212 }, { "epoch": 0.8105430234235967, "grad_norm": 1.3899449656755047, "learning_rate": 1.822105827656776e-06, "loss": 1.012, "step": 4213 }, { "epoch": 0.8107354143619836, "grad_norm": 1.4501895474837025, "learning_rate": 1.8185207431895613e-06, "loss": 1.0091, "step": 4214 }, { "epoch": 0.8109278053003703, "grad_norm": 1.5220924310236463, "learning_rate": 1.8149388363146148e-06, "loss": 0.9596, "step": 4215 }, { "epoch": 0.8111201962387572, "grad_norm": 1.4091913041244635, "learning_rate": 1.8113601084231091e-06, "loss": 0.9984, "step": 4216 }, { "epoch": 0.8113125871771439, "grad_norm": 0.752439349391878, "learning_rate": 1.8077845609049782e-06, "loss": 0.8699, "step": 4217 }, { "epoch": 0.8115049781155308, "grad_norm": 1.3445790536172133, "learning_rate": 1.8042121951489254e-06, "loss": 0.9444, "step": 4218 }, { "epoch": 0.8116973690539175, "grad_norm": 1.3656838175572752, "learning_rate": 1.800643012542418e-06, "loss": 0.9411, "step": 4219 }, { "epoch": 0.8118897599923044, "grad_norm": 1.421615202586214, "learning_rate": 1.7970770144716777e-06, "loss": 0.9087, "step": 4220 }, { "epoch": 0.8120821509306911, "grad_norm": 1.404991757395811, "learning_rate": 1.7935142023217056e-06, "loss": 1.0282, "step": 4221 }, { "epoch": 0.812274541869078, "grad_norm": 1.3137797630836243, "learning_rate": 1.7899545774762573e-06, "loss": 0.9547, "step": 4222 }, { "epoch": 0.8124669328074647, "grad_norm": 0.7616393173558095, "learning_rate": 1.7863981413178433e-06, "loss": 0.8621, "step": 4223 }, { "epoch": 0.8126593237458516, "grad_norm": 1.5368193393641478, "learning_rate": 1.7828448952277456e-06, "loss": 0.9109, "step": 4224 }, { "epoch": 0.8128517146842383, "grad_norm": 0.6986967184669085, "learning_rate": 1.7792948405860079e-06, "loss": 0.8457, "step": 4225 }, { "epoch": 0.8130441056226252, "grad_norm": 1.4948972470875548, "learning_rate": 1.7757479787714217e-06, "loss": 1.0462, "step": 4226 }, { "epoch": 0.8132364965610119, "grad_norm": 1.4813268809301985, "learning_rate": 1.7722043111615572e-06, "loss": 0.9257, "step": 4227 }, { "epoch": 0.8134288874993988, "grad_norm": 1.4574209149429886, "learning_rate": 1.768663839132727e-06, "loss": 0.9784, "step": 4228 }, { "epoch": 0.8136212784377855, "grad_norm": 1.3991782039764065, "learning_rate": 1.7651265640600113e-06, "loss": 0.9104, "step": 4229 }, { "epoch": 0.8138136693761724, "grad_norm": 1.3376207463381733, "learning_rate": 1.7615924873172506e-06, "loss": 0.985, "step": 4230 }, { "epoch": 0.8140060603145591, "grad_norm": 1.2247714030218249, "learning_rate": 1.7580616102770353e-06, "loss": 0.8301, "step": 4231 }, { "epoch": 0.814198451252946, "grad_norm": 1.4941811576553583, "learning_rate": 1.754533934310717e-06, "loss": 0.9618, "step": 4232 }, { "epoch": 0.8143908421913328, "grad_norm": 1.6909680758891608, "learning_rate": 1.7510094607884075e-06, "loss": 0.9675, "step": 4233 }, { "epoch": 0.8145832331297196, "grad_norm": 1.4062132248805208, "learning_rate": 1.7474881910789698e-06, "loss": 0.9694, "step": 4234 }, { "epoch": 0.8147756240681064, "grad_norm": 1.443923283218152, "learning_rate": 1.7439701265500274e-06, "loss": 1.0135, "step": 4235 }, { "epoch": 0.8149680150064932, "grad_norm": 1.541160724603967, "learning_rate": 1.740455268567951e-06, "loss": 0.9434, "step": 4236 }, { "epoch": 0.81516040594488, "grad_norm": 1.3727817668337752, "learning_rate": 1.7369436184978738e-06, "loss": 0.9314, "step": 4237 }, { "epoch": 0.8153527968832668, "grad_norm": 0.6978179721978452, "learning_rate": 1.7334351777036807e-06, "loss": 0.831, "step": 4238 }, { "epoch": 0.8155451878216536, "grad_norm": 1.5166476374113642, "learning_rate": 1.729929947548008e-06, "loss": 0.9601, "step": 4239 }, { "epoch": 0.8157375787600404, "grad_norm": 1.5208596821355063, "learning_rate": 1.7264279293922503e-06, "loss": 1.0197, "step": 4240 }, { "epoch": 0.8159299696984272, "grad_norm": 1.4178010334865856, "learning_rate": 1.7229291245965463e-06, "loss": 0.9398, "step": 4241 }, { "epoch": 0.816122360636814, "grad_norm": 1.5271159324522705, "learning_rate": 1.7194335345197933e-06, "loss": 0.976, "step": 4242 }, { "epoch": 0.8163147515752008, "grad_norm": 1.3395547195664943, "learning_rate": 1.7159411605196407e-06, "loss": 0.9221, "step": 4243 }, { "epoch": 0.8165071425135876, "grad_norm": 0.683338062204505, "learning_rate": 1.7124520039524805e-06, "loss": 0.7647, "step": 4244 }, { "epoch": 0.8166995334519744, "grad_norm": 1.5397506883501015, "learning_rate": 1.7089660661734685e-06, "loss": 1.0358, "step": 4245 }, { "epoch": 0.8168919243903612, "grad_norm": 0.6989030011807005, "learning_rate": 1.7054833485364962e-06, "loss": 0.7689, "step": 4246 }, { "epoch": 0.817084315328748, "grad_norm": 1.3086148625292882, "learning_rate": 1.702003852394214e-06, "loss": 0.9832, "step": 4247 }, { "epoch": 0.8172767062671348, "grad_norm": 1.3670580942340413, "learning_rate": 1.6985275790980205e-06, "loss": 0.9789, "step": 4248 }, { "epoch": 0.8174690972055216, "grad_norm": 1.3512042523170014, "learning_rate": 1.6950545299980526e-06, "loss": 0.9167, "step": 4249 }, { "epoch": 0.8176614881439084, "grad_norm": 1.2352235065144281, "learning_rate": 1.691584706443209e-06, "loss": 0.8996, "step": 4250 }, { "epoch": 0.8178538790822952, "grad_norm": 1.5057293155842277, "learning_rate": 1.6881181097811305e-06, "loss": 0.9223, "step": 4251 }, { "epoch": 0.818046270020682, "grad_norm": 1.535727709495079, "learning_rate": 1.6846547413581981e-06, "loss": 1.0146, "step": 4252 }, { "epoch": 0.8182386609590688, "grad_norm": 1.4399523645546304, "learning_rate": 1.6811946025195459e-06, "loss": 0.981, "step": 4253 }, { "epoch": 0.8184310518974556, "grad_norm": 1.352741589607982, "learning_rate": 1.6777376946090552e-06, "loss": 0.9096, "step": 4254 }, { "epoch": 0.8186234428358424, "grad_norm": 1.4581713172896293, "learning_rate": 1.674284018969342e-06, "loss": 0.9229, "step": 4255 }, { "epoch": 0.8188158337742293, "grad_norm": 1.2761418210410076, "learning_rate": 1.6708335769417827e-06, "loss": 0.9377, "step": 4256 }, { "epoch": 0.819008224712616, "grad_norm": 1.4256150048453216, "learning_rate": 1.667386369866484e-06, "loss": 0.9365, "step": 4257 }, { "epoch": 0.8192006156510029, "grad_norm": 0.6827960869744225, "learning_rate": 1.6639423990823011e-06, "loss": 0.7608, "step": 4258 }, { "epoch": 0.8193930065893896, "grad_norm": 1.4752757981304911, "learning_rate": 1.660501665926838e-06, "loss": 1.0028, "step": 4259 }, { "epoch": 0.8195853975277765, "grad_norm": 1.3732621309983193, "learning_rate": 1.6570641717364277e-06, "loss": 0.999, "step": 4260 }, { "epoch": 0.8197777884661632, "grad_norm": 1.4878726724529066, "learning_rate": 1.653629917846159e-06, "loss": 0.9747, "step": 4261 }, { "epoch": 0.8199701794045501, "grad_norm": 1.4711337779187488, "learning_rate": 1.6501989055898537e-06, "loss": 0.9783, "step": 4262 }, { "epoch": 0.8201625703429368, "grad_norm": 1.4485160462529667, "learning_rate": 1.6467711363000794e-06, "loss": 0.9982, "step": 4263 }, { "epoch": 0.8203549612813237, "grad_norm": 1.4374156013314665, "learning_rate": 1.6433466113081442e-06, "loss": 0.9435, "step": 4264 }, { "epoch": 0.8205473522197104, "grad_norm": 1.407449227470131, "learning_rate": 1.6399253319440888e-06, "loss": 0.9198, "step": 4265 }, { "epoch": 0.8207397431580973, "grad_norm": 1.3183121272346023, "learning_rate": 1.6365072995367004e-06, "loss": 0.8436, "step": 4266 }, { "epoch": 0.820932134096484, "grad_norm": 1.2956021641083124, "learning_rate": 1.6330925154135057e-06, "loss": 0.9282, "step": 4267 }, { "epoch": 0.8211245250348709, "grad_norm": 1.2773073971371354, "learning_rate": 1.6296809809007652e-06, "loss": 0.9685, "step": 4268 }, { "epoch": 0.8213169159732576, "grad_norm": 1.350367012369775, "learning_rate": 1.6262726973234844e-06, "loss": 0.939, "step": 4269 }, { "epoch": 0.8215093069116445, "grad_norm": 1.3229343805072253, "learning_rate": 1.6228676660053932e-06, "loss": 0.9923, "step": 4270 }, { "epoch": 0.8217016978500312, "grad_norm": 1.677338134293564, "learning_rate": 1.6194658882689718e-06, "loss": 1.0526, "step": 4271 }, { "epoch": 0.8218940887884181, "grad_norm": 1.4495768655968855, "learning_rate": 1.6160673654354331e-06, "loss": 0.9394, "step": 4272 }, { "epoch": 0.8220864797268048, "grad_norm": 1.3336373797879417, "learning_rate": 1.6126720988247168e-06, "loss": 1.0022, "step": 4273 }, { "epoch": 0.8222788706651917, "grad_norm": 1.59440765742731, "learning_rate": 1.609280089755515e-06, "loss": 1.0471, "step": 4274 }, { "epoch": 0.8224712616035784, "grad_norm": 1.3755624061348115, "learning_rate": 1.605891339545237e-06, "loss": 1.0238, "step": 4275 }, { "epoch": 0.8226636525419653, "grad_norm": 1.41262940858017, "learning_rate": 1.6025058495100388e-06, "loss": 0.93, "step": 4276 }, { "epoch": 0.8228560434803521, "grad_norm": 1.3424012973616206, "learning_rate": 1.5991236209648052e-06, "loss": 0.9272, "step": 4277 }, { "epoch": 0.8230484344187389, "grad_norm": 1.2434066547647855, "learning_rate": 1.5957446552231526e-06, "loss": 0.8912, "step": 4278 }, { "epoch": 0.8232408253571257, "grad_norm": 1.5318459594136302, "learning_rate": 1.5923689535974307e-06, "loss": 0.969, "step": 4279 }, { "epoch": 0.8234332162955125, "grad_norm": 1.1919371682230817, "learning_rate": 1.588996517398731e-06, "loss": 0.83, "step": 4280 }, { "epoch": 0.8236256072338993, "grad_norm": 1.3304521237987246, "learning_rate": 1.5856273479368611e-06, "loss": 0.9413, "step": 4281 }, { "epoch": 0.8238179981722861, "grad_norm": 1.3487623959995536, "learning_rate": 1.582261446520371e-06, "loss": 0.9183, "step": 4282 }, { "epoch": 0.8240103891106729, "grad_norm": 0.6409076553755543, "learning_rate": 1.5788988144565397e-06, "loss": 0.7987, "step": 4283 }, { "epoch": 0.8242027800490597, "grad_norm": 1.3807106424288096, "learning_rate": 1.575539453051369e-06, "loss": 0.9174, "step": 4284 }, { "epoch": 0.8243951709874465, "grad_norm": 0.7293986305794288, "learning_rate": 1.572183363609603e-06, "loss": 0.8671, "step": 4285 }, { "epoch": 0.8245875619258333, "grad_norm": 1.2916701157566712, "learning_rate": 1.568830547434703e-06, "loss": 0.9766, "step": 4286 }, { "epoch": 0.8247799528642201, "grad_norm": 1.331313920076138, "learning_rate": 1.5654810058288661e-06, "loss": 0.8982, "step": 4287 }, { "epoch": 0.8249723438026069, "grad_norm": 1.3327766017917917, "learning_rate": 1.5621347400930176e-06, "loss": 0.9245, "step": 4288 }, { "epoch": 0.8251647347409937, "grad_norm": 1.5059057187826428, "learning_rate": 1.5587917515268048e-06, "loss": 0.991, "step": 4289 }, { "epoch": 0.8253571256793805, "grad_norm": 1.3457998106159708, "learning_rate": 1.5554520414286067e-06, "loss": 0.8923, "step": 4290 }, { "epoch": 0.8255495166177673, "grad_norm": 1.3299988459065037, "learning_rate": 1.5521156110955293e-06, "loss": 1.0064, "step": 4291 }, { "epoch": 0.8257419075561541, "grad_norm": 1.3450077662284168, "learning_rate": 1.5487824618234049e-06, "loss": 0.8853, "step": 4292 }, { "epoch": 0.8259342984945409, "grad_norm": 1.3013831884753864, "learning_rate": 1.54545259490679e-06, "loss": 0.9603, "step": 4293 }, { "epoch": 0.8261266894329277, "grad_norm": 1.3181282136143233, "learning_rate": 1.5421260116389636e-06, "loss": 0.9088, "step": 4294 }, { "epoch": 0.8263190803713145, "grad_norm": 1.442741477818239, "learning_rate": 1.5388027133119343e-06, "loss": 0.9716, "step": 4295 }, { "epoch": 0.8265114713097013, "grad_norm": 1.4079188689737598, "learning_rate": 1.535482701216433e-06, "loss": 0.9594, "step": 4296 }, { "epoch": 0.8267038622480881, "grad_norm": 1.3639612520827245, "learning_rate": 1.5321659766419129e-06, "loss": 0.8675, "step": 4297 }, { "epoch": 0.826896253186475, "grad_norm": 1.4026077256260063, "learning_rate": 1.5288525408765564e-06, "loss": 1.0168, "step": 4298 }, { "epoch": 0.8270886441248617, "grad_norm": 1.359243113577039, "learning_rate": 1.5255423952072567e-06, "loss": 0.9559, "step": 4299 }, { "epoch": 0.8272810350632486, "grad_norm": 1.4799161191156067, "learning_rate": 1.52223554091964e-06, "loss": 1.0118, "step": 4300 }, { "epoch": 0.8274734260016353, "grad_norm": 1.358409516182964, "learning_rate": 1.5189319792980517e-06, "loss": 0.9589, "step": 4301 }, { "epoch": 0.8276658169400222, "grad_norm": 1.4329447506203328, "learning_rate": 1.5156317116255515e-06, "loss": 0.9245, "step": 4302 }, { "epoch": 0.8278582078784089, "grad_norm": 1.3272108722173657, "learning_rate": 1.5123347391839305e-06, "loss": 0.9668, "step": 4303 }, { "epoch": 0.8280505988167958, "grad_norm": 1.2818049164857317, "learning_rate": 1.5090410632536968e-06, "loss": 0.889, "step": 4304 }, { "epoch": 0.8282429897551825, "grad_norm": 1.4577996231356773, "learning_rate": 1.5057506851140701e-06, "loss": 0.9974, "step": 4305 }, { "epoch": 0.8284353806935694, "grad_norm": 1.4369293013284465, "learning_rate": 1.5024636060429998e-06, "loss": 0.9092, "step": 4306 }, { "epoch": 0.8286277716319561, "grad_norm": 1.3165554966686333, "learning_rate": 1.4991798273171465e-06, "loss": 0.9496, "step": 4307 }, { "epoch": 0.828820162570343, "grad_norm": 1.504575012839067, "learning_rate": 1.49589935021189e-06, "loss": 0.9034, "step": 4308 }, { "epoch": 0.8290125535087297, "grad_norm": 1.467608479078948, "learning_rate": 1.4926221760013393e-06, "loss": 0.9569, "step": 4309 }, { "epoch": 0.8292049444471166, "grad_norm": 1.4134953449165601, "learning_rate": 1.4893483059583014e-06, "loss": 0.9446, "step": 4310 }, { "epoch": 0.8293973353855033, "grad_norm": 1.4248085746364147, "learning_rate": 1.4860777413543138e-06, "loss": 0.9659, "step": 4311 }, { "epoch": 0.8295897263238902, "grad_norm": 1.2822099991658746, "learning_rate": 1.4828104834596268e-06, "loss": 0.9673, "step": 4312 }, { "epoch": 0.8297821172622769, "grad_norm": 1.3016081221551081, "learning_rate": 1.4795465335432036e-06, "loss": 0.9182, "step": 4313 }, { "epoch": 0.8299745082006638, "grad_norm": 1.3692065839244518, "learning_rate": 1.4762858928727241e-06, "loss": 0.8866, "step": 4314 }, { "epoch": 0.8301668991390505, "grad_norm": 1.4811148130638137, "learning_rate": 1.4730285627145858e-06, "loss": 0.9826, "step": 4315 }, { "epoch": 0.8303592900774374, "grad_norm": 1.4397341489457418, "learning_rate": 1.4697745443338984e-06, "loss": 1.0197, "step": 4316 }, { "epoch": 0.8305516810158241, "grad_norm": 1.47100130346138, "learning_rate": 1.4665238389944859e-06, "loss": 0.9902, "step": 4317 }, { "epoch": 0.830744071954211, "grad_norm": 1.385592400439522, "learning_rate": 1.46327644795888e-06, "loss": 0.8889, "step": 4318 }, { "epoch": 0.8309364628925977, "grad_norm": 1.5172605582287935, "learning_rate": 1.4600323724883337e-06, "loss": 0.9388, "step": 4319 }, { "epoch": 0.8311288538309846, "grad_norm": 1.4963106802038628, "learning_rate": 1.4567916138428072e-06, "loss": 0.9358, "step": 4320 }, { "epoch": 0.8313212447693714, "grad_norm": 1.4102665874883655, "learning_rate": 1.4535541732809755e-06, "loss": 1.013, "step": 4321 }, { "epoch": 0.8315136357077582, "grad_norm": 1.364794823995081, "learning_rate": 1.4503200520602245e-06, "loss": 0.8676, "step": 4322 }, { "epoch": 0.831706026646145, "grad_norm": 1.4819637245649215, "learning_rate": 1.4470892514366442e-06, "loss": 0.9586, "step": 4323 }, { "epoch": 0.8318984175845318, "grad_norm": 1.3295567328339422, "learning_rate": 1.443861772665044e-06, "loss": 0.9016, "step": 4324 }, { "epoch": 0.8320908085229186, "grad_norm": 1.5988457744087172, "learning_rate": 1.4406376169989389e-06, "loss": 0.9875, "step": 4325 }, { "epoch": 0.8322831994613054, "grad_norm": 1.4675645336362082, "learning_rate": 1.4374167856905542e-06, "loss": 0.9327, "step": 4326 }, { "epoch": 0.8324755903996922, "grad_norm": 1.3551993068155066, "learning_rate": 1.4341992799908255e-06, "loss": 1.017, "step": 4327 }, { "epoch": 0.832667981338079, "grad_norm": 1.3352475709996556, "learning_rate": 1.4309851011493903e-06, "loss": 0.9443, "step": 4328 }, { "epoch": 0.8328603722764658, "grad_norm": 1.3389399767774373, "learning_rate": 1.427774250414601e-06, "loss": 0.9658, "step": 4329 }, { "epoch": 0.8330527632148526, "grad_norm": 1.3680992659193498, "learning_rate": 1.4245667290335175e-06, "loss": 1.0019, "step": 4330 }, { "epoch": 0.8332451541532394, "grad_norm": 1.4153565422856478, "learning_rate": 1.421362538251897e-06, "loss": 1.0247, "step": 4331 }, { "epoch": 0.8334375450916262, "grad_norm": 1.356733857333408, "learning_rate": 1.4181616793142173e-06, "loss": 0.9676, "step": 4332 }, { "epoch": 0.833629936030013, "grad_norm": 1.5156501863866179, "learning_rate": 1.414964153463655e-06, "loss": 0.9425, "step": 4333 }, { "epoch": 0.8338223269683998, "grad_norm": 1.429559806757554, "learning_rate": 1.4117699619420878e-06, "loss": 0.9707, "step": 4334 }, { "epoch": 0.8340147179067866, "grad_norm": 1.4775085521469877, "learning_rate": 1.4085791059901077e-06, "loss": 0.9608, "step": 4335 }, { "epoch": 0.8342071088451734, "grad_norm": 1.3306270780803016, "learning_rate": 1.4053915868470013e-06, "loss": 0.9861, "step": 4336 }, { "epoch": 0.8343994997835602, "grad_norm": 1.5112724601400453, "learning_rate": 1.402207405750765e-06, "loss": 0.9699, "step": 4337 }, { "epoch": 0.834591890721947, "grad_norm": 1.3540849199525316, "learning_rate": 1.399026563938105e-06, "loss": 0.9114, "step": 4338 }, { "epoch": 0.8347842816603338, "grad_norm": 1.4432601751938932, "learning_rate": 1.3958490626444154e-06, "loss": 0.9042, "step": 4339 }, { "epoch": 0.8349766725987205, "grad_norm": 1.3838784940393325, "learning_rate": 1.3926749031038055e-06, "loss": 0.9849, "step": 4340 }, { "epoch": 0.8351690635371074, "grad_norm": 1.4967271658166645, "learning_rate": 1.3895040865490817e-06, "loss": 0.9266, "step": 4341 }, { "epoch": 0.8353614544754943, "grad_norm": 1.3100189370156676, "learning_rate": 1.3863366142117506e-06, "loss": 0.8941, "step": 4342 }, { "epoch": 0.835553845413881, "grad_norm": 1.3025699818253242, "learning_rate": 1.383172487322023e-06, "loss": 0.9181, "step": 4343 }, { "epoch": 0.8357462363522679, "grad_norm": 1.4327584054561417, "learning_rate": 1.3800117071088104e-06, "loss": 0.9384, "step": 4344 }, { "epoch": 0.8359386272906546, "grad_norm": 1.2725102545671605, "learning_rate": 1.3768542747997215e-06, "loss": 0.9781, "step": 4345 }, { "epoch": 0.8361310182290415, "grad_norm": 1.4287752984779143, "learning_rate": 1.3737001916210713e-06, "loss": 0.9485, "step": 4346 }, { "epoch": 0.8363234091674282, "grad_norm": 1.3258347124588774, "learning_rate": 1.370549458797863e-06, "loss": 0.9769, "step": 4347 }, { "epoch": 0.836515800105815, "grad_norm": 1.4452834522041775, "learning_rate": 1.3674020775538078e-06, "loss": 0.9163, "step": 4348 }, { "epoch": 0.8367081910442018, "grad_norm": 1.559020648012419, "learning_rate": 1.3642580491113122e-06, "loss": 1.0129, "step": 4349 }, { "epoch": 0.8369005819825887, "grad_norm": 1.4515639645752914, "learning_rate": 1.3611173746914797e-06, "loss": 0.9901, "step": 4350 }, { "epoch": 0.8370929729209754, "grad_norm": 1.3437349049142333, "learning_rate": 1.3579800555141165e-06, "loss": 0.9915, "step": 4351 }, { "epoch": 0.8372853638593623, "grad_norm": 1.3451602277630492, "learning_rate": 1.3548460927977158e-06, "loss": 0.9499, "step": 4352 }, { "epoch": 0.837477754797749, "grad_norm": 1.3802528174790805, "learning_rate": 1.351715487759474e-06, "loss": 0.9603, "step": 4353 }, { "epoch": 0.8376701457361359, "grad_norm": 1.5269323435134086, "learning_rate": 1.3485882416152819e-06, "loss": 0.9588, "step": 4354 }, { "epoch": 0.8378625366745226, "grad_norm": 1.4827010419022695, "learning_rate": 1.3454643555797276e-06, "loss": 0.9459, "step": 4355 }, { "epoch": 0.8380549276129095, "grad_norm": 1.3481957455265778, "learning_rate": 1.3423438308660929e-06, "loss": 0.963, "step": 4356 }, { "epoch": 0.8382473185512962, "grad_norm": 1.4124927193551677, "learning_rate": 1.3392266686863508e-06, "loss": 1.0513, "step": 4357 }, { "epoch": 0.838439709489683, "grad_norm": 1.4395377856838483, "learning_rate": 1.3361128702511716e-06, "loss": 0.9543, "step": 4358 }, { "epoch": 0.8386321004280698, "grad_norm": 1.4902459115764364, "learning_rate": 1.3330024367699224e-06, "loss": 1.004, "step": 4359 }, { "epoch": 0.8388244913664566, "grad_norm": 1.4648881750105933, "learning_rate": 1.3298953694506522e-06, "loss": 1.0262, "step": 4360 }, { "epoch": 0.8390168823048434, "grad_norm": 0.700626965236324, "learning_rate": 1.3267916695001172e-06, "loss": 0.8006, "step": 4361 }, { "epoch": 0.8392092732432302, "grad_norm": 1.3492948982786563, "learning_rate": 1.3236913381237592e-06, "loss": 0.9873, "step": 4362 }, { "epoch": 0.839401664181617, "grad_norm": 1.3956621707298014, "learning_rate": 1.3205943765257057e-06, "loss": 0.9057, "step": 4363 }, { "epoch": 0.8395940551200038, "grad_norm": 1.364376388616282, "learning_rate": 1.317500785908783e-06, "loss": 0.9701, "step": 4364 }, { "epoch": 0.8397864460583907, "grad_norm": 1.4944934320395704, "learning_rate": 1.31441056747451e-06, "loss": 0.9765, "step": 4365 }, { "epoch": 0.8399788369967774, "grad_norm": 1.3984251729646442, "learning_rate": 1.3113237224230836e-06, "loss": 1.051, "step": 4366 }, { "epoch": 0.8401712279351643, "grad_norm": 1.23738856022619, "learning_rate": 1.3082402519534076e-06, "loss": 0.9462, "step": 4367 }, { "epoch": 0.840363618873551, "grad_norm": 1.441187421775729, "learning_rate": 1.3051601572630611e-06, "loss": 0.9499, "step": 4368 }, { "epoch": 0.8405560098119379, "grad_norm": 1.6139828601022912, "learning_rate": 1.3020834395483195e-06, "loss": 0.9341, "step": 4369 }, { "epoch": 0.8407484007503246, "grad_norm": 1.488245507206915, "learning_rate": 1.2990101000041445e-06, "loss": 1.0311, "step": 4370 }, { "epoch": 0.8409407916887115, "grad_norm": 1.3302618280337213, "learning_rate": 1.2959401398241844e-06, "loss": 0.9301, "step": 4371 }, { "epoch": 0.8411331826270982, "grad_norm": 1.4875394559823332, "learning_rate": 1.2928735602007768e-06, "loss": 0.8918, "step": 4372 }, { "epoch": 0.8413255735654851, "grad_norm": 0.7035295418564976, "learning_rate": 1.2898103623249458e-06, "loss": 0.8582, "step": 4373 }, { "epoch": 0.8415179645038718, "grad_norm": 1.3108401705047066, "learning_rate": 1.2867505473864029e-06, "loss": 0.8983, "step": 4374 }, { "epoch": 0.8417103554422587, "grad_norm": 1.3836644572608934, "learning_rate": 1.283694116573546e-06, "loss": 0.9067, "step": 4375 }, { "epoch": 0.8419027463806454, "grad_norm": 1.5010408908364317, "learning_rate": 1.2806410710734552e-06, "loss": 0.9195, "step": 4376 }, { "epoch": 0.8420951373190323, "grad_norm": 1.5417117465178112, "learning_rate": 1.2775914120718992e-06, "loss": 0.828, "step": 4377 }, { "epoch": 0.842287528257419, "grad_norm": 1.4114718294041708, "learning_rate": 1.2745451407533294e-06, "loss": 0.9933, "step": 4378 }, { "epoch": 0.8424799191958059, "grad_norm": 1.289541820272038, "learning_rate": 1.2715022583008851e-06, "loss": 0.9579, "step": 4379 }, { "epoch": 0.8426723101341926, "grad_norm": 1.4744469449095186, "learning_rate": 1.2684627658963865e-06, "loss": 1.0438, "step": 4380 }, { "epoch": 0.8428647010725795, "grad_norm": 1.5682036711845255, "learning_rate": 1.265426664720334e-06, "loss": 0.9258, "step": 4381 }, { "epoch": 0.8430570920109662, "grad_norm": 1.3249527601060445, "learning_rate": 1.2623939559519161e-06, "loss": 0.9727, "step": 4382 }, { "epoch": 0.8432494829493531, "grad_norm": 1.3476559855179806, "learning_rate": 1.2593646407690051e-06, "loss": 0.9142, "step": 4383 }, { "epoch": 0.8434418738877398, "grad_norm": 1.4674095131137435, "learning_rate": 1.2563387203481447e-06, "loss": 1.0109, "step": 4384 }, { "epoch": 0.8436342648261267, "grad_norm": 1.4548988360644521, "learning_rate": 1.2533161958645755e-06, "loss": 0.9919, "step": 4385 }, { "epoch": 0.8438266557645135, "grad_norm": 1.4064068075431349, "learning_rate": 1.2502970684922067e-06, "loss": 0.9256, "step": 4386 }, { "epoch": 0.8440190467029003, "grad_norm": 0.7460161079034628, "learning_rate": 1.2472813394036344e-06, "loss": 0.8019, "step": 4387 }, { "epoch": 0.8442114376412871, "grad_norm": 1.495225209132626, "learning_rate": 1.2442690097701327e-06, "loss": 0.9412, "step": 4388 }, { "epoch": 0.8444038285796739, "grad_norm": 1.4329166958779045, "learning_rate": 1.2412600807616526e-06, "loss": 0.9323, "step": 4389 }, { "epoch": 0.8445962195180607, "grad_norm": 1.440234845388168, "learning_rate": 1.2382545535468316e-06, "loss": 0.9562, "step": 4390 }, { "epoch": 0.8447886104564475, "grad_norm": 1.3195871811653404, "learning_rate": 1.2352524292929823e-06, "loss": 0.8556, "step": 4391 }, { "epoch": 0.8449810013948343, "grad_norm": 1.3960272051685516, "learning_rate": 1.2322537091660912e-06, "loss": 0.9164, "step": 4392 }, { "epoch": 0.8451733923332211, "grad_norm": 1.4442432053490457, "learning_rate": 1.22925839433083e-06, "loss": 0.8455, "step": 4393 }, { "epoch": 0.8453657832716079, "grad_norm": 1.582746299523643, "learning_rate": 1.2262664859505434e-06, "loss": 0.9593, "step": 4394 }, { "epoch": 0.8455581742099947, "grad_norm": 1.5489875934472586, "learning_rate": 1.2232779851872511e-06, "loss": 0.8233, "step": 4395 }, { "epoch": 0.8457505651483815, "grad_norm": 1.4909496526366746, "learning_rate": 1.2202928932016588e-06, "loss": 0.9941, "step": 4396 }, { "epoch": 0.8459429560867683, "grad_norm": 1.4044572741065575, "learning_rate": 1.217311211153137e-06, "loss": 0.9425, "step": 4397 }, { "epoch": 0.8461353470251551, "grad_norm": 1.3085836906962514, "learning_rate": 1.2143329401997372e-06, "loss": 0.847, "step": 4398 }, { "epoch": 0.8463277379635419, "grad_norm": 1.3244778173214184, "learning_rate": 1.2113580814981884e-06, "loss": 0.9637, "step": 4399 }, { "epoch": 0.8465201289019287, "grad_norm": 0.7656862580471504, "learning_rate": 1.2083866362038865e-06, "loss": 0.8712, "step": 4400 }, { "epoch": 0.8467125198403155, "grad_norm": 1.2806271555955693, "learning_rate": 1.2054186054709105e-06, "loss": 0.9005, "step": 4401 }, { "epoch": 0.8469049107787023, "grad_norm": 1.444461723321088, "learning_rate": 1.2024539904520072e-06, "loss": 0.9928, "step": 4402 }, { "epoch": 0.8470973017170891, "grad_norm": 1.4748088687684633, "learning_rate": 1.1994927922985999e-06, "loss": 1.0205, "step": 4403 }, { "epoch": 0.8472896926554759, "grad_norm": 1.5024636528849673, "learning_rate": 1.1965350121607866e-06, "loss": 0.9769, "step": 4404 }, { "epoch": 0.8474820835938627, "grad_norm": 1.329701736593509, "learning_rate": 1.1935806511873306e-06, "loss": 0.8974, "step": 4405 }, { "epoch": 0.8476744745322495, "grad_norm": 1.4642779180557468, "learning_rate": 1.1906297105256725e-06, "loss": 0.9194, "step": 4406 }, { "epoch": 0.8478668654706364, "grad_norm": 1.3736561173916997, "learning_rate": 1.187682191321925e-06, "loss": 0.9854, "step": 4407 }, { "epoch": 0.8480592564090231, "grad_norm": 1.4851431315909018, "learning_rate": 1.1847380947208697e-06, "loss": 0.9228, "step": 4408 }, { "epoch": 0.84825164734741, "grad_norm": 1.2326320065581893, "learning_rate": 1.1817974218659621e-06, "loss": 0.9319, "step": 4409 }, { "epoch": 0.8484440382857967, "grad_norm": 1.3693643082327782, "learning_rate": 1.178860173899321e-06, "loss": 0.9554, "step": 4410 }, { "epoch": 0.8486364292241836, "grad_norm": 0.7312228606943109, "learning_rate": 1.1759263519617437e-06, "loss": 0.8331, "step": 4411 }, { "epoch": 0.8488288201625703, "grad_norm": 1.4002538635157433, "learning_rate": 1.172995957192693e-06, "loss": 0.8682, "step": 4412 }, { "epoch": 0.8490212111009572, "grad_norm": 1.344797641198136, "learning_rate": 1.1700689907302953e-06, "loss": 0.9023, "step": 4413 }, { "epoch": 0.8492136020393439, "grad_norm": 1.3518301992070725, "learning_rate": 1.167145453711358e-06, "loss": 0.8595, "step": 4414 }, { "epoch": 0.8494059929777308, "grad_norm": 1.4255513430553994, "learning_rate": 1.1642253472713427e-06, "loss": 0.8989, "step": 4415 }, { "epoch": 0.8495983839161175, "grad_norm": 1.440337085661305, "learning_rate": 1.161308672544389e-06, "loss": 0.9614, "step": 4416 }, { "epoch": 0.8497907748545044, "grad_norm": 1.3680980389425408, "learning_rate": 1.1583954306633004e-06, "loss": 0.9397, "step": 4417 }, { "epoch": 0.8499831657928911, "grad_norm": 1.385999408143207, "learning_rate": 1.1554856227595435e-06, "loss": 0.8802, "step": 4418 }, { "epoch": 0.850175556731278, "grad_norm": 1.4409159306086767, "learning_rate": 1.1525792499632526e-06, "loss": 1.0088, "step": 4419 }, { "epoch": 0.8503679476696647, "grad_norm": 1.4238989218854703, "learning_rate": 1.1496763134032363e-06, "loss": 1.0052, "step": 4420 }, { "epoch": 0.8505603386080516, "grad_norm": 1.392652655901913, "learning_rate": 1.1467768142069546e-06, "loss": 0.9116, "step": 4421 }, { "epoch": 0.8507527295464383, "grad_norm": 1.369897012337564, "learning_rate": 1.1438807535005437e-06, "loss": 1.0021, "step": 4422 }, { "epoch": 0.8509451204848252, "grad_norm": 1.50738853484211, "learning_rate": 1.1409881324088013e-06, "loss": 0.9883, "step": 4423 }, { "epoch": 0.8511375114232119, "grad_norm": 1.4026152084065127, "learning_rate": 1.138098952055181e-06, "loss": 0.98, "step": 4424 }, { "epoch": 0.8513299023615988, "grad_norm": 1.5558533543218787, "learning_rate": 1.1352132135618165e-06, "loss": 0.8988, "step": 4425 }, { "epoch": 0.8515222932999855, "grad_norm": 1.3327615831514925, "learning_rate": 1.13233091804949e-06, "loss": 0.8976, "step": 4426 }, { "epoch": 0.8517146842383724, "grad_norm": 1.287802226523741, "learning_rate": 1.1294520666376518e-06, "loss": 0.8784, "step": 4427 }, { "epoch": 0.8519070751767591, "grad_norm": 1.4414239563869573, "learning_rate": 1.1265766604444172e-06, "loss": 0.9502, "step": 4428 }, { "epoch": 0.852099466115146, "grad_norm": 1.4156776464250633, "learning_rate": 1.1237047005865576e-06, "loss": 0.8815, "step": 4429 }, { "epoch": 0.8522918570535328, "grad_norm": 1.5013360983029893, "learning_rate": 1.1208361881795116e-06, "loss": 0.8989, "step": 4430 }, { "epoch": 0.8524842479919196, "grad_norm": 1.3350549344476337, "learning_rate": 1.1179711243373736e-06, "loss": 0.9213, "step": 4431 }, { "epoch": 0.8526766389303064, "grad_norm": 1.6013012292335052, "learning_rate": 1.1151095101729047e-06, "loss": 0.9196, "step": 4432 }, { "epoch": 0.8528690298686932, "grad_norm": 1.356573094536514, "learning_rate": 1.1122513467975237e-06, "loss": 0.9173, "step": 4433 }, { "epoch": 0.85306142080708, "grad_norm": 1.5088719532235433, "learning_rate": 1.1093966353213036e-06, "loss": 0.9664, "step": 4434 }, { "epoch": 0.8532538117454668, "grad_norm": 1.5059682060524449, "learning_rate": 1.1065453768529844e-06, "loss": 0.9824, "step": 4435 }, { "epoch": 0.8534462026838536, "grad_norm": 1.3074250842518, "learning_rate": 1.103697572499961e-06, "loss": 0.9664, "step": 4436 }, { "epoch": 0.8536385936222404, "grad_norm": 1.4866377043821328, "learning_rate": 1.1008532233682878e-06, "loss": 0.8969, "step": 4437 }, { "epoch": 0.8538309845606272, "grad_norm": 1.5876290517817975, "learning_rate": 1.0980123305626812e-06, "loss": 0.9731, "step": 4438 }, { "epoch": 0.854023375499014, "grad_norm": 1.4171118269624856, "learning_rate": 1.0951748951865048e-06, "loss": 0.9924, "step": 4439 }, { "epoch": 0.8542157664374008, "grad_norm": 1.3912527033744757, "learning_rate": 1.0923409183417887e-06, "loss": 0.9354, "step": 4440 }, { "epoch": 0.8544081573757876, "grad_norm": 1.4138853416295571, "learning_rate": 1.0895104011292202e-06, "loss": 0.8899, "step": 4441 }, { "epoch": 0.8546005483141744, "grad_norm": 1.264756763435399, "learning_rate": 1.0866833446481317e-06, "loss": 0.949, "step": 4442 }, { "epoch": 0.8547929392525612, "grad_norm": 1.4100197331413593, "learning_rate": 1.0838597499965276e-06, "loss": 0.9259, "step": 4443 }, { "epoch": 0.854985330190948, "grad_norm": 1.4231320265967178, "learning_rate": 1.0810396182710535e-06, "loss": 0.8819, "step": 4444 }, { "epoch": 0.8551777211293348, "grad_norm": 1.4144368517838668, "learning_rate": 1.0782229505670195e-06, "loss": 0.9169, "step": 4445 }, { "epoch": 0.8553701120677216, "grad_norm": 1.355633902219136, "learning_rate": 1.0754097479783876e-06, "loss": 0.8999, "step": 4446 }, { "epoch": 0.8555625030061084, "grad_norm": 1.4053778782794406, "learning_rate": 1.0726000115977696e-06, "loss": 0.8904, "step": 4447 }, { "epoch": 0.8557548939444952, "grad_norm": 1.3698474313856623, "learning_rate": 1.069793742516435e-06, "loss": 0.8958, "step": 4448 }, { "epoch": 0.855947284882882, "grad_norm": 1.353408272226175, "learning_rate": 1.066990941824312e-06, "loss": 0.9322, "step": 4449 }, { "epoch": 0.8561396758212688, "grad_norm": 1.6543738715865395, "learning_rate": 1.0641916106099691e-06, "loss": 0.982, "step": 4450 }, { "epoch": 0.8563320667596557, "grad_norm": 1.3982991164302292, "learning_rate": 1.0613957499606388e-06, "loss": 0.9871, "step": 4451 }, { "epoch": 0.8565244576980424, "grad_norm": 1.4574164922924346, "learning_rate": 1.0586033609622004e-06, "loss": 0.9172, "step": 4452 }, { "epoch": 0.8567168486364293, "grad_norm": 1.2759032411460463, "learning_rate": 1.0558144446991836e-06, "loss": 0.9725, "step": 4453 }, { "epoch": 0.856909239574816, "grad_norm": 1.3899375015113824, "learning_rate": 1.053029002254773e-06, "loss": 0.9782, "step": 4454 }, { "epoch": 0.8571016305132029, "grad_norm": 1.4646171371578294, "learning_rate": 1.0502470347108017e-06, "loss": 0.9591, "step": 4455 }, { "epoch": 0.8572940214515896, "grad_norm": 1.4942948897248565, "learning_rate": 1.0474685431477537e-06, "loss": 1.0055, "step": 4456 }, { "epoch": 0.8574864123899765, "grad_norm": 1.4261396475166368, "learning_rate": 1.0446935286447657e-06, "loss": 0.98, "step": 4457 }, { "epoch": 0.8576788033283632, "grad_norm": 1.4124725764520982, "learning_rate": 1.0419219922796175e-06, "loss": 0.8881, "step": 4458 }, { "epoch": 0.8578711942667501, "grad_norm": 0.7382025059732576, "learning_rate": 1.039153935128744e-06, "loss": 0.9063, "step": 4459 }, { "epoch": 0.8580635852051368, "grad_norm": 1.409976296687032, "learning_rate": 1.0363893582672246e-06, "loss": 1.0398, "step": 4460 }, { "epoch": 0.8582559761435237, "grad_norm": 1.4575643244641852, "learning_rate": 1.033628262768792e-06, "loss": 0.9003, "step": 4461 }, { "epoch": 0.8584483670819104, "grad_norm": 1.4532172214967018, "learning_rate": 1.0308706497058252e-06, "loss": 0.9528, "step": 4462 }, { "epoch": 0.8586407580202973, "grad_norm": 1.3814571519432017, "learning_rate": 1.0281165201493437e-06, "loss": 0.9954, "step": 4463 }, { "epoch": 0.858833148958684, "grad_norm": 1.3540479070933282, "learning_rate": 1.0253658751690232e-06, "loss": 0.9336, "step": 4464 }, { "epoch": 0.8590255398970709, "grad_norm": 1.3326391456774709, "learning_rate": 1.0226187158331825e-06, "loss": 0.913, "step": 4465 }, { "epoch": 0.8592179308354576, "grad_norm": 1.5334163117524886, "learning_rate": 1.0198750432087855e-06, "loss": 0.9096, "step": 4466 }, { "epoch": 0.8594103217738445, "grad_norm": 1.4731310349133606, "learning_rate": 1.017134858361446e-06, "loss": 0.9406, "step": 4467 }, { "epoch": 0.8596027127122312, "grad_norm": 1.4089824855913748, "learning_rate": 1.0143981623554155e-06, "loss": 0.9777, "step": 4468 }, { "epoch": 0.8597951036506181, "grad_norm": 1.5281867639132878, "learning_rate": 1.0116649562535984e-06, "loss": 0.8997, "step": 4469 }, { "epoch": 0.8599874945890048, "grad_norm": 1.4858827432345574, "learning_rate": 1.0089352411175424e-06, "loss": 0.961, "step": 4470 }, { "epoch": 0.8601798855273917, "grad_norm": 1.4352532792718495, "learning_rate": 1.006209018007429e-06, "loss": 0.9568, "step": 4471 }, { "epoch": 0.8603722764657784, "grad_norm": 1.3531705285889921, "learning_rate": 1.0034862879821029e-06, "loss": 0.9762, "step": 4472 }, { "epoch": 0.8605646674041653, "grad_norm": 1.4183771233810867, "learning_rate": 1.0007670520990331e-06, "loss": 0.8658, "step": 4473 }, { "epoch": 0.8607570583425521, "grad_norm": 1.422113834181057, "learning_rate": 9.98051311414342e-07, "loss": 0.8878, "step": 4474 }, { "epoch": 0.8609494492809389, "grad_norm": 1.3830588600688958, "learning_rate": 9.953390669827944e-07, "loss": 0.9692, "step": 4475 }, { "epoch": 0.8611418402193257, "grad_norm": 1.5608040976910176, "learning_rate": 9.926303198577913e-07, "loss": 0.9309, "step": 4476 }, { "epoch": 0.8613342311577125, "grad_norm": 1.380560867937015, "learning_rate": 9.899250710913767e-07, "loss": 1.0021, "step": 4477 }, { "epoch": 0.8615266220960993, "grad_norm": 1.4253941984516862, "learning_rate": 9.872233217342463e-07, "loss": 0.9684, "step": 4478 }, { "epoch": 0.8617190130344861, "grad_norm": 1.3865068340842652, "learning_rate": 9.845250728357214e-07, "loss": 0.9381, "step": 4479 }, { "epoch": 0.8619114039728729, "grad_norm": 1.5771097306499258, "learning_rate": 9.818303254437723e-07, "loss": 0.9514, "step": 4480 }, { "epoch": 0.8621037949112597, "grad_norm": 1.3434296206441896, "learning_rate": 9.791390806050117e-07, "loss": 0.7816, "step": 4481 }, { "epoch": 0.8622961858496465, "grad_norm": 1.3892198115961176, "learning_rate": 9.764513393646812e-07, "loss": 0.9842, "step": 4482 }, { "epoch": 0.8624885767880333, "grad_norm": 1.3370097108692647, "learning_rate": 9.737671027666728e-07, "loss": 0.9017, "step": 4483 }, { "epoch": 0.8626809677264201, "grad_norm": 1.4188428781120301, "learning_rate": 9.710863718535135e-07, "loss": 0.9735, "step": 4484 }, { "epoch": 0.8628733586648069, "grad_norm": 1.4286200555297357, "learning_rate": 9.684091476663659e-07, "loss": 0.9653, "step": 4485 }, { "epoch": 0.8630657496031937, "grad_norm": 1.281690466834801, "learning_rate": 9.657354312450363e-07, "loss": 0.8839, "step": 4486 }, { "epoch": 0.8632581405415805, "grad_norm": 1.419191110095869, "learning_rate": 9.630652236279626e-07, "loss": 0.8568, "step": 4487 }, { "epoch": 0.8634505314799673, "grad_norm": 1.484645833145966, "learning_rate": 9.603985258522219e-07, "loss": 0.9038, "step": 4488 }, { "epoch": 0.863642922418354, "grad_norm": 1.402812716801383, "learning_rate": 9.577353389535315e-07, "loss": 0.9759, "step": 4489 }, { "epoch": 0.8638353133567409, "grad_norm": 1.4056862317858339, "learning_rate": 9.550756639662417e-07, "loss": 0.9234, "step": 4490 }, { "epoch": 0.8640277042951277, "grad_norm": 1.2223247252584881, "learning_rate": 9.524195019233407e-07, "loss": 0.8842, "step": 4491 }, { "epoch": 0.8642200952335145, "grad_norm": 1.4786807041279584, "learning_rate": 9.497668538564475e-07, "loss": 0.8888, "step": 4492 }, { "epoch": 0.8644124861719013, "grad_norm": 1.3616683856746274, "learning_rate": 9.471177207958238e-07, "loss": 0.9426, "step": 4493 }, { "epoch": 0.8646048771102881, "grad_norm": 1.4307858797799404, "learning_rate": 9.444721037703597e-07, "loss": 0.9581, "step": 4494 }, { "epoch": 0.864797268048675, "grad_norm": 1.6018390072220132, "learning_rate": 9.418300038075845e-07, "loss": 1.0233, "step": 4495 }, { "epoch": 0.8649896589870617, "grad_norm": 1.4277101029349066, "learning_rate": 9.391914219336606e-07, "loss": 0.94, "step": 4496 }, { "epoch": 0.8651820499254486, "grad_norm": 1.4078933389618815, "learning_rate": 9.365563591733784e-07, "loss": 0.8879, "step": 4497 }, { "epoch": 0.8653744408638353, "grad_norm": 1.385766389713634, "learning_rate": 9.33924816550168e-07, "loss": 0.9603, "step": 4498 }, { "epoch": 0.8655668318022222, "grad_norm": 0.7241211633201912, "learning_rate": 9.31296795086093e-07, "loss": 0.8319, "step": 4499 }, { "epoch": 0.8657592227406089, "grad_norm": 1.3327814478209261, "learning_rate": 9.286722958018391e-07, "loss": 0.9725, "step": 4500 }, { "epoch": 0.8659516136789958, "grad_norm": 1.3645105432619922, "learning_rate": 9.260513197167398e-07, "loss": 0.924, "step": 4501 }, { "epoch": 0.8661440046173825, "grad_norm": 1.4197268714250775, "learning_rate": 9.234338678487509e-07, "loss": 1.0348, "step": 4502 }, { "epoch": 0.8663363955557694, "grad_norm": 1.3567330194199196, "learning_rate": 9.208199412144559e-07, "loss": 0.9329, "step": 4503 }, { "epoch": 0.8665287864941561, "grad_norm": 1.3385432403137203, "learning_rate": 9.182095408290781e-07, "loss": 0.9599, "step": 4504 }, { "epoch": 0.866721177432543, "grad_norm": 1.3523817276966437, "learning_rate": 9.156026677064633e-07, "loss": 0.903, "step": 4505 }, { "epoch": 0.8669135683709297, "grad_norm": 1.5437599389348045, "learning_rate": 9.129993228590917e-07, "loss": 0.9283, "step": 4506 }, { "epoch": 0.8671059593093166, "grad_norm": 1.4386845636150467, "learning_rate": 9.103995072980765e-07, "loss": 0.9668, "step": 4507 }, { "epoch": 0.8672983502477033, "grad_norm": 1.2968152506136685, "learning_rate": 9.078032220331523e-07, "loss": 0.873, "step": 4508 }, { "epoch": 0.8674907411860902, "grad_norm": 1.416569347687761, "learning_rate": 9.052104680726859e-07, "loss": 0.9866, "step": 4509 }, { "epoch": 0.8676831321244769, "grad_norm": 1.408733760086765, "learning_rate": 9.026212464236772e-07, "loss": 1.0201, "step": 4510 }, { "epoch": 0.8678755230628638, "grad_norm": 1.5037393418264862, "learning_rate": 9.000355580917464e-07, "loss": 0.9737, "step": 4511 }, { "epoch": 0.8680679140012505, "grad_norm": 1.428177180656253, "learning_rate": 8.974534040811444e-07, "loss": 0.9159, "step": 4512 }, { "epoch": 0.8682603049396374, "grad_norm": 1.3744393781673507, "learning_rate": 8.948747853947526e-07, "loss": 0.9511, "step": 4513 }, { "epoch": 0.8684526958780241, "grad_norm": 1.352512829614826, "learning_rate": 8.922997030340752e-07, "loss": 0.9276, "step": 4514 }, { "epoch": 0.868645086816411, "grad_norm": 1.494632949907623, "learning_rate": 8.897281579992467e-07, "loss": 0.9544, "step": 4515 }, { "epoch": 0.8688374777547978, "grad_norm": 1.4865880558562117, "learning_rate": 8.871601512890238e-07, "loss": 0.9708, "step": 4516 }, { "epoch": 0.8690298686931845, "grad_norm": 1.3659995645871224, "learning_rate": 8.845956839007897e-07, "loss": 1.0012, "step": 4517 }, { "epoch": 0.8692222596315714, "grad_norm": 1.5024024055959468, "learning_rate": 8.820347568305543e-07, "loss": 1.0267, "step": 4518 }, { "epoch": 0.8694146505699581, "grad_norm": 1.373444951575029, "learning_rate": 8.794773710729543e-07, "loss": 0.9324, "step": 4519 }, { "epoch": 0.869607041508345, "grad_norm": 1.359623677048272, "learning_rate": 8.769235276212496e-07, "loss": 0.9204, "step": 4520 }, { "epoch": 0.8697994324467317, "grad_norm": 1.354186644230862, "learning_rate": 8.743732274673189e-07, "loss": 0.9005, "step": 4521 }, { "epoch": 0.8699918233851186, "grad_norm": 1.3052954388305895, "learning_rate": 8.718264716016722e-07, "loss": 0.9466, "step": 4522 }, { "epoch": 0.8701842143235053, "grad_norm": 1.4663323933912866, "learning_rate": 8.692832610134428e-07, "loss": 0.8975, "step": 4523 }, { "epoch": 0.8703766052618922, "grad_norm": 1.3824184395454737, "learning_rate": 8.66743596690377e-07, "loss": 1.0289, "step": 4524 }, { "epoch": 0.8705689962002789, "grad_norm": 1.3228963901498167, "learning_rate": 8.642074796188594e-07, "loss": 0.9736, "step": 4525 }, { "epoch": 0.8707613871386658, "grad_norm": 1.3795125650443258, "learning_rate": 8.61674910783884e-07, "loss": 0.9904, "step": 4526 }, { "epoch": 0.8709537780770525, "grad_norm": 1.3880711866128734, "learning_rate": 8.59145891169072e-07, "loss": 0.9932, "step": 4527 }, { "epoch": 0.8711461690154394, "grad_norm": 1.4025255412507955, "learning_rate": 8.566204217566664e-07, "loss": 0.9293, "step": 4528 }, { "epoch": 0.8713385599538261, "grad_norm": 1.5210690885146465, "learning_rate": 8.540985035275273e-07, "loss": 0.9697, "step": 4529 }, { "epoch": 0.871530950892213, "grad_norm": 1.3300055485336257, "learning_rate": 8.515801374611432e-07, "loss": 0.8398, "step": 4530 }, { "epoch": 0.8717233418305997, "grad_norm": 1.3613057909594168, "learning_rate": 8.490653245356184e-07, "loss": 1.006, "step": 4531 }, { "epoch": 0.8719157327689866, "grad_norm": 1.643957046896751, "learning_rate": 8.465540657276728e-07, "loss": 0.9962, "step": 4532 }, { "epoch": 0.8721081237073733, "grad_norm": 1.3121218856091161, "learning_rate": 8.44046362012656e-07, "loss": 0.9114, "step": 4533 }, { "epoch": 0.8723005146457602, "grad_norm": 0.7227354088081759, "learning_rate": 8.415422143645247e-07, "loss": 0.8423, "step": 4534 }, { "epoch": 0.8724929055841469, "grad_norm": 1.478485008497859, "learning_rate": 8.390416237558641e-07, "loss": 1.0195, "step": 4535 }, { "epoch": 0.8726852965225338, "grad_norm": 1.5199688014828243, "learning_rate": 8.365445911578785e-07, "loss": 0.9336, "step": 4536 }, { "epoch": 0.8728776874609205, "grad_norm": 1.448634738623593, "learning_rate": 8.340511175403809e-07, "loss": 0.9357, "step": 4537 }, { "epoch": 0.8730700783993074, "grad_norm": 1.3924041517963304, "learning_rate": 8.315612038718101e-07, "loss": 0.9267, "step": 4538 }, { "epoch": 0.8732624693376942, "grad_norm": 1.4236691550485734, "learning_rate": 8.290748511192214e-07, "loss": 0.9395, "step": 4539 }, { "epoch": 0.873454860276081, "grad_norm": 1.4302272401075617, "learning_rate": 8.265920602482825e-07, "loss": 0.991, "step": 4540 }, { "epoch": 0.8736472512144678, "grad_norm": 1.3840060310596634, "learning_rate": 8.241128322232816e-07, "loss": 0.9499, "step": 4541 }, { "epoch": 0.8738396421528546, "grad_norm": 1.4634454402452999, "learning_rate": 8.216371680071244e-07, "loss": 0.9483, "step": 4542 }, { "epoch": 0.8740320330912414, "grad_norm": 1.3922565828079219, "learning_rate": 8.191650685613273e-07, "loss": 0.9457, "step": 4543 }, { "epoch": 0.8742244240296282, "grad_norm": 1.529232877594287, "learning_rate": 8.166965348460298e-07, "loss": 0.8898, "step": 4544 }, { "epoch": 0.874416814968015, "grad_norm": 1.4363857117601306, "learning_rate": 8.142315678199764e-07, "loss": 0.9675, "step": 4545 }, { "epoch": 0.8746092059064018, "grad_norm": 1.4847776212144306, "learning_rate": 8.117701684405343e-07, "loss": 0.9772, "step": 4546 }, { "epoch": 0.8748015968447886, "grad_norm": 1.3192372448338017, "learning_rate": 8.093123376636836e-07, "loss": 0.9803, "step": 4547 }, { "epoch": 0.8749939877831754, "grad_norm": 1.4103111461009599, "learning_rate": 8.06858076444017e-07, "loss": 0.9257, "step": 4548 }, { "epoch": 0.8751863787215622, "grad_norm": 1.3894720045802826, "learning_rate": 8.044073857347423e-07, "loss": 0.9224, "step": 4549 }, { "epoch": 0.875378769659949, "grad_norm": 1.3130175311216037, "learning_rate": 8.019602664876758e-07, "loss": 0.9163, "step": 4550 }, { "epoch": 0.8755711605983358, "grad_norm": 1.3016442997468969, "learning_rate": 7.995167196532527e-07, "loss": 0.9425, "step": 4551 }, { "epoch": 0.8757635515367226, "grad_norm": 1.4401287985158902, "learning_rate": 7.970767461805218e-07, "loss": 0.9565, "step": 4552 }, { "epoch": 0.8759559424751094, "grad_norm": 1.478011850984846, "learning_rate": 7.946403470171326e-07, "loss": 0.8751, "step": 4553 }, { "epoch": 0.8761483334134962, "grad_norm": 1.3230167419397532, "learning_rate": 7.922075231093628e-07, "loss": 0.9036, "step": 4554 }, { "epoch": 0.876340724351883, "grad_norm": 1.348987522534048, "learning_rate": 7.897782754020889e-07, "loss": 0.8755, "step": 4555 }, { "epoch": 0.8765331152902698, "grad_norm": 1.3239929323990256, "learning_rate": 7.873526048388025e-07, "loss": 0.9013, "step": 4556 }, { "epoch": 0.8767255062286566, "grad_norm": 1.3361385806971329, "learning_rate": 7.849305123616091e-07, "loss": 0.942, "step": 4557 }, { "epoch": 0.8769178971670434, "grad_norm": 1.3549762907812537, "learning_rate": 7.825119989112173e-07, "loss": 0.8931, "step": 4558 }, { "epoch": 0.8771102881054302, "grad_norm": 1.51740137069564, "learning_rate": 7.800970654269513e-07, "loss": 0.9818, "step": 4559 }, { "epoch": 0.8773026790438171, "grad_norm": 1.3982860058031727, "learning_rate": 7.776857128467464e-07, "loss": 0.9442, "step": 4560 }, { "epoch": 0.8774950699822038, "grad_norm": 1.539913950563114, "learning_rate": 7.75277942107141e-07, "loss": 0.9594, "step": 4561 }, { "epoch": 0.8776874609205907, "grad_norm": 1.3570699910238728, "learning_rate": 7.728737541432862e-07, "loss": 0.9539, "step": 4562 }, { "epoch": 0.8778798518589774, "grad_norm": 1.4303686740155312, "learning_rate": 7.704731498889428e-07, "loss": 0.8893, "step": 4563 }, { "epoch": 0.8780722427973643, "grad_norm": 1.2780492751223225, "learning_rate": 7.680761302764727e-07, "loss": 0.9217, "step": 4564 }, { "epoch": 0.878264633735751, "grad_norm": 1.3375841254510121, "learning_rate": 7.65682696236858e-07, "loss": 0.8805, "step": 4565 }, { "epoch": 0.8784570246741379, "grad_norm": 1.3709116782427548, "learning_rate": 7.632928486996749e-07, "loss": 0.797, "step": 4566 }, { "epoch": 0.8786494156125246, "grad_norm": 1.470353741582708, "learning_rate": 7.609065885931155e-07, "loss": 0.8612, "step": 4567 }, { "epoch": 0.8788418065509115, "grad_norm": 1.2603894161058675, "learning_rate": 7.585239168439762e-07, "loss": 0.8993, "step": 4568 }, { "epoch": 0.8790341974892982, "grad_norm": 1.4195852253306451, "learning_rate": 7.561448343776567e-07, "loss": 1.0171, "step": 4569 }, { "epoch": 0.8792265884276851, "grad_norm": 1.5464325939081496, "learning_rate": 7.537693421181658e-07, "loss": 0.9496, "step": 4570 }, { "epoch": 0.8794189793660718, "grad_norm": 1.2764266713081074, "learning_rate": 7.513974409881186e-07, "loss": 0.8863, "step": 4571 }, { "epoch": 0.8796113703044587, "grad_norm": 1.326044537087708, "learning_rate": 7.49029131908734e-07, "loss": 0.9384, "step": 4572 }, { "epoch": 0.8798037612428454, "grad_norm": 1.5278012180003395, "learning_rate": 7.466644157998371e-07, "loss": 1.0361, "step": 4573 }, { "epoch": 0.8799961521812323, "grad_norm": 1.4107272594901807, "learning_rate": 7.443032935798533e-07, "loss": 0.9974, "step": 4574 }, { "epoch": 0.880188543119619, "grad_norm": 1.3824194493892792, "learning_rate": 7.419457661658169e-07, "loss": 0.9707, "step": 4575 }, { "epoch": 0.8803809340580059, "grad_norm": 1.3911870542025797, "learning_rate": 7.395918344733644e-07, "loss": 0.9933, "step": 4576 }, { "epoch": 0.8805733249963926, "grad_norm": 1.25084698535279, "learning_rate": 7.372414994167354e-07, "loss": 0.9413, "step": 4577 }, { "epoch": 0.8807657159347795, "grad_norm": 1.4136215708409856, "learning_rate": 7.348947619087754e-07, "loss": 0.9366, "step": 4578 }, { "epoch": 0.8809581068731662, "grad_norm": 1.3778808489194978, "learning_rate": 7.325516228609264e-07, "loss": 0.9074, "step": 4579 }, { "epoch": 0.8811504978115531, "grad_norm": 1.3306706134479231, "learning_rate": 7.302120831832382e-07, "loss": 0.8808, "step": 4580 }, { "epoch": 0.8813428887499398, "grad_norm": 1.4512112465420717, "learning_rate": 7.278761437843629e-07, "loss": 0.9529, "step": 4581 }, { "epoch": 0.8815352796883267, "grad_norm": 1.3932374581568154, "learning_rate": 7.255438055715469e-07, "loss": 0.9481, "step": 4582 }, { "epoch": 0.8817276706267135, "grad_norm": 1.3975479875937278, "learning_rate": 7.232150694506512e-07, "loss": 1.0156, "step": 4583 }, { "epoch": 0.8819200615651003, "grad_norm": 1.4240653531058187, "learning_rate": 7.208899363261234e-07, "loss": 0.9181, "step": 4584 }, { "epoch": 0.8821124525034871, "grad_norm": 1.5205324640538562, "learning_rate": 7.185684071010224e-07, "loss": 0.8648, "step": 4585 }, { "epoch": 0.8823048434418739, "grad_norm": 1.426964606599537, "learning_rate": 7.162504826770033e-07, "loss": 0.9064, "step": 4586 }, { "epoch": 0.8824972343802607, "grad_norm": 1.4298108678160142, "learning_rate": 7.139361639543185e-07, "loss": 0.9184, "step": 4587 }, { "epoch": 0.8826896253186475, "grad_norm": 1.3506783036144177, "learning_rate": 7.116254518318222e-07, "loss": 0.9039, "step": 4588 }, { "epoch": 0.8828820162570343, "grad_norm": 1.357258141837241, "learning_rate": 7.093183472069753e-07, "loss": 0.9782, "step": 4589 }, { "epoch": 0.8830744071954211, "grad_norm": 1.3665329074725447, "learning_rate": 7.070148509758223e-07, "loss": 0.9235, "step": 4590 }, { "epoch": 0.8832667981338079, "grad_norm": 1.4054706706361075, "learning_rate": 7.047149640330197e-07, "loss": 0.9188, "step": 4591 }, { "epoch": 0.8834591890721947, "grad_norm": 1.446090210305333, "learning_rate": 7.024186872718164e-07, "loss": 0.9524, "step": 4592 }, { "epoch": 0.8836515800105815, "grad_norm": 1.2980142091299631, "learning_rate": 7.001260215840567e-07, "loss": 0.9129, "step": 4593 }, { "epoch": 0.8838439709489683, "grad_norm": 1.4075546383169144, "learning_rate": 6.978369678601892e-07, "loss": 0.9202, "step": 4594 }, { "epoch": 0.8840363618873551, "grad_norm": 1.595058691018159, "learning_rate": 6.955515269892533e-07, "loss": 0.9008, "step": 4595 }, { "epoch": 0.8842287528257419, "grad_norm": 1.3426184536242778, "learning_rate": 6.932696998588895e-07, "loss": 0.8999, "step": 4596 }, { "epoch": 0.8844211437641287, "grad_norm": 1.3209169223263526, "learning_rate": 6.909914873553347e-07, "loss": 0.9501, "step": 4597 }, { "epoch": 0.8846135347025155, "grad_norm": 1.3506637473399659, "learning_rate": 6.887168903634178e-07, "loss": 0.9552, "step": 4598 }, { "epoch": 0.8848059256409023, "grad_norm": 1.3696177582281455, "learning_rate": 6.864459097665654e-07, "loss": 0.9488, "step": 4599 }, { "epoch": 0.8849983165792891, "grad_norm": 1.3972914921124413, "learning_rate": 6.84178546446802e-07, "loss": 0.9319, "step": 4600 }, { "epoch": 0.8851907075176759, "grad_norm": 1.3646977449350568, "learning_rate": 6.819148012847454e-07, "loss": 0.9731, "step": 4601 }, { "epoch": 0.8853830984560627, "grad_norm": 1.3405343857801741, "learning_rate": 6.796546751596089e-07, "loss": 0.9446, "step": 4602 }, { "epoch": 0.8855754893944495, "grad_norm": 1.354816326958317, "learning_rate": 6.77398168949196e-07, "loss": 0.9248, "step": 4603 }, { "epoch": 0.8857678803328364, "grad_norm": 1.3844158152263415, "learning_rate": 6.751452835299111e-07, "loss": 0.9067, "step": 4604 }, { "epoch": 0.8859602712712231, "grad_norm": 1.3908063113503792, "learning_rate": 6.728960197767475e-07, "loss": 0.9584, "step": 4605 }, { "epoch": 0.88615266220961, "grad_norm": 1.4561172509605382, "learning_rate": 6.706503785632934e-07, "loss": 1.016, "step": 4606 }, { "epoch": 0.8863450531479967, "grad_norm": 1.4193444576626906, "learning_rate": 6.68408360761732e-07, "loss": 0.9254, "step": 4607 }, { "epoch": 0.8865374440863836, "grad_norm": 1.3186589147878685, "learning_rate": 6.661699672428334e-07, "loss": 0.9726, "step": 4608 }, { "epoch": 0.8867298350247703, "grad_norm": 0.7012876503153387, "learning_rate": 6.639351988759657e-07, "loss": 0.8501, "step": 4609 }, { "epoch": 0.8869222259631572, "grad_norm": 1.4367759168199838, "learning_rate": 6.6170405652909e-07, "loss": 0.9534, "step": 4610 }, { "epoch": 0.8871146169015439, "grad_norm": 0.7064546224233452, "learning_rate": 6.594765410687487e-07, "loss": 0.8374, "step": 4611 }, { "epoch": 0.8873070078399308, "grad_norm": 1.3787578733273682, "learning_rate": 6.57252653360092e-07, "loss": 0.861, "step": 4612 }, { "epoch": 0.8874993987783175, "grad_norm": 1.5105498468873122, "learning_rate": 6.550323942668469e-07, "loss": 0.9641, "step": 4613 }, { "epoch": 0.8876917897167044, "grad_norm": 1.4451793726093947, "learning_rate": 6.528157646513378e-07, "loss": 0.9357, "step": 4614 }, { "epoch": 0.8878841806550911, "grad_norm": 1.4027366581872462, "learning_rate": 6.506027653744796e-07, "loss": 0.9742, "step": 4615 }, { "epoch": 0.888076571593478, "grad_norm": 1.6552347046198708, "learning_rate": 6.483933972957734e-07, "loss": 0.9419, "step": 4616 }, { "epoch": 0.8882689625318647, "grad_norm": 1.398464868207296, "learning_rate": 6.461876612733109e-07, "loss": 0.9336, "step": 4617 }, { "epoch": 0.8884613534702516, "grad_norm": 1.568008046963088, "learning_rate": 6.4398555816378e-07, "loss": 0.9618, "step": 4618 }, { "epoch": 0.8886537444086383, "grad_norm": 0.6510987435297237, "learning_rate": 6.417870888224476e-07, "loss": 0.8388, "step": 4619 }, { "epoch": 0.8888461353470252, "grad_norm": 1.3055202384159132, "learning_rate": 6.395922541031741e-07, "loss": 0.9522, "step": 4620 }, { "epoch": 0.8890385262854119, "grad_norm": 1.5045423270786529, "learning_rate": 6.374010548584119e-07, "loss": 0.9613, "step": 4621 }, { "epoch": 0.8892309172237988, "grad_norm": 1.356255407131011, "learning_rate": 6.352134919391928e-07, "loss": 0.9618, "step": 4622 }, { "epoch": 0.8894233081621855, "grad_norm": 1.4014443127398368, "learning_rate": 6.330295661951436e-07, "loss": 0.9528, "step": 4623 }, { "epoch": 0.8896156991005724, "grad_norm": 1.5474926097107928, "learning_rate": 6.308492784744746e-07, "loss": 0.9806, "step": 4624 }, { "epoch": 0.8898080900389591, "grad_norm": 1.4329925288493524, "learning_rate": 6.286726296239854e-07, "loss": 0.9714, "step": 4625 }, { "epoch": 0.890000480977346, "grad_norm": 1.3450134034861967, "learning_rate": 6.264996204890628e-07, "loss": 0.8584, "step": 4626 }, { "epoch": 0.8901928719157328, "grad_norm": 1.3632653911756558, "learning_rate": 6.24330251913674e-07, "loss": 0.9292, "step": 4627 }, { "epoch": 0.8903852628541196, "grad_norm": 1.502344269579054, "learning_rate": 6.221645247403807e-07, "loss": 0.9139, "step": 4628 }, { "epoch": 0.8905776537925064, "grad_norm": 1.3689602554763374, "learning_rate": 6.200024398103255e-07, "loss": 0.9518, "step": 4629 }, { "epoch": 0.8907700447308932, "grad_norm": 1.2998743727943558, "learning_rate": 6.178439979632367e-07, "loss": 0.9398, "step": 4630 }, { "epoch": 0.89096243566928, "grad_norm": 1.3783555383671384, "learning_rate": 6.156892000374293e-07, "loss": 0.9012, "step": 4631 }, { "epoch": 0.8911548266076668, "grad_norm": 1.4183445220727506, "learning_rate": 6.135380468698004e-07, "loss": 1.0012, "step": 4632 }, { "epoch": 0.8913472175460536, "grad_norm": 1.5638355723919972, "learning_rate": 6.113905392958342e-07, "loss": 0.9599, "step": 4633 }, { "epoch": 0.8915396084844404, "grad_norm": 1.3959231309691782, "learning_rate": 6.092466781495976e-07, "loss": 1.0224, "step": 4634 }, { "epoch": 0.8917319994228272, "grad_norm": 1.4341180318950526, "learning_rate": 6.071064642637404e-07, "loss": 0.9278, "step": 4635 }, { "epoch": 0.891924390361214, "grad_norm": 1.3567046159575207, "learning_rate": 6.049698984695007e-07, "loss": 0.8745, "step": 4636 }, { "epoch": 0.8921167812996008, "grad_norm": 1.3663846992116686, "learning_rate": 6.028369815966917e-07, "loss": 0.9542, "step": 4637 }, { "epoch": 0.8923091722379876, "grad_norm": 1.4968399329047668, "learning_rate": 6.007077144737161e-07, "loss": 1.0034, "step": 4638 }, { "epoch": 0.8925015631763744, "grad_norm": 1.4210067667208242, "learning_rate": 5.985820979275569e-07, "loss": 0.9342, "step": 4639 }, { "epoch": 0.8926939541147612, "grad_norm": 1.3240280343078252, "learning_rate": 5.96460132783776e-07, "loss": 0.9542, "step": 4640 }, { "epoch": 0.892886345053148, "grad_norm": 1.4361192695115825, "learning_rate": 5.943418198665251e-07, "loss": 0.9236, "step": 4641 }, { "epoch": 0.8930787359915348, "grad_norm": 0.7382691396028048, "learning_rate": 5.922271599985286e-07, "loss": 0.8556, "step": 4642 }, { "epoch": 0.8932711269299216, "grad_norm": 1.3903248369931103, "learning_rate": 5.901161540010969e-07, "loss": 0.9176, "step": 4643 }, { "epoch": 0.8934635178683084, "grad_norm": 1.3909437914915526, "learning_rate": 5.880088026941233e-07, "loss": 0.9832, "step": 4644 }, { "epoch": 0.8936559088066952, "grad_norm": 1.489853950162255, "learning_rate": 5.859051068960741e-07, "loss": 0.8986, "step": 4645 }, { "epoch": 0.893848299745082, "grad_norm": 1.3728367492885754, "learning_rate": 5.838050674240025e-07, "loss": 0.9465, "step": 4646 }, { "epoch": 0.8940406906834688, "grad_norm": 1.4187955154361205, "learning_rate": 5.817086850935416e-07, "loss": 0.982, "step": 4647 }, { "epoch": 0.8942330816218557, "grad_norm": 1.475859357789376, "learning_rate": 5.796159607189001e-07, "loss": 0.9446, "step": 4648 }, { "epoch": 0.8944254725602424, "grad_norm": 1.30785006275075, "learning_rate": 5.775268951128676e-07, "loss": 0.9203, "step": 4649 }, { "epoch": 0.8946178634986293, "grad_norm": 1.4308629643366328, "learning_rate": 5.754414890868154e-07, "loss": 1.0006, "step": 4650 }, { "epoch": 0.894810254437016, "grad_norm": 1.3473497868549553, "learning_rate": 5.733597434506877e-07, "loss": 0.9554, "step": 4651 }, { "epoch": 0.8950026453754029, "grad_norm": 1.391579966467021, "learning_rate": 5.712816590130133e-07, "loss": 0.9394, "step": 4652 }, { "epoch": 0.8951950363137896, "grad_norm": 1.3561374557185806, "learning_rate": 5.692072365808954e-07, "loss": 0.8975, "step": 4653 }, { "epoch": 0.8953874272521765, "grad_norm": 1.4093848766879964, "learning_rate": 5.671364769600162e-07, "loss": 0.9962, "step": 4654 }, { "epoch": 0.8955798181905632, "grad_norm": 1.3541948629980347, "learning_rate": 5.650693809546348e-07, "loss": 0.9731, "step": 4655 }, { "epoch": 0.8957722091289501, "grad_norm": 1.3988718347058342, "learning_rate": 5.630059493675866e-07, "loss": 0.8743, "step": 4656 }, { "epoch": 0.8959646000673368, "grad_norm": 1.3216783740843747, "learning_rate": 5.60946183000285e-07, "loss": 0.8885, "step": 4657 }, { "epoch": 0.8961569910057237, "grad_norm": 1.341750964174061, "learning_rate": 5.5889008265272e-07, "loss": 0.9158, "step": 4658 }, { "epoch": 0.8963493819441104, "grad_norm": 1.3144820106806265, "learning_rate": 5.568376491234562e-07, "loss": 0.9632, "step": 4659 }, { "epoch": 0.8965417728824973, "grad_norm": 1.3749638183892972, "learning_rate": 5.547888832096382e-07, "loss": 0.9316, "step": 4660 }, { "epoch": 0.896734163820884, "grad_norm": 1.4748284773635754, "learning_rate": 5.527437857069784e-07, "loss": 0.9411, "step": 4661 }, { "epoch": 0.8969265547592709, "grad_norm": 1.4097272108274665, "learning_rate": 5.507023574097725e-07, "loss": 0.9422, "step": 4662 }, { "epoch": 0.8971189456976576, "grad_norm": 1.328987538020718, "learning_rate": 5.486645991108875e-07, "loss": 0.995, "step": 4663 }, { "epoch": 0.8973113366360445, "grad_norm": 1.3934456973148528, "learning_rate": 5.466305116017623e-07, "loss": 0.8355, "step": 4664 }, { "epoch": 0.8975037275744312, "grad_norm": 1.3309658479727315, "learning_rate": 5.446000956724174e-07, "loss": 0.9822, "step": 4665 }, { "epoch": 0.897696118512818, "grad_norm": 1.3552836990649304, "learning_rate": 5.425733521114396e-07, "loss": 0.9991, "step": 4666 }, { "epoch": 0.8978885094512048, "grad_norm": 1.3956729110630406, "learning_rate": 5.405502817059937e-07, "loss": 0.964, "step": 4667 }, { "epoch": 0.8980809003895917, "grad_norm": 1.2556305100521303, "learning_rate": 5.385308852418191e-07, "loss": 0.944, "step": 4668 }, { "epoch": 0.8982732913279785, "grad_norm": 1.494607479486131, "learning_rate": 5.365151635032218e-07, "loss": 0.9678, "step": 4669 }, { "epoch": 0.8984656822663653, "grad_norm": 1.4152751603019937, "learning_rate": 5.345031172730875e-07, "loss": 0.8935, "step": 4670 }, { "epoch": 0.8986580732047521, "grad_norm": 1.3437306579355248, "learning_rate": 5.324947473328735e-07, "loss": 0.889, "step": 4671 }, { "epoch": 0.8988504641431388, "grad_norm": 1.325824889522526, "learning_rate": 5.304900544626046e-07, "loss": 0.89, "step": 4672 }, { "epoch": 0.8990428550815257, "grad_norm": 1.415355265599702, "learning_rate": 5.284890394408826e-07, "loss": 1.004, "step": 4673 }, { "epoch": 0.8992352460199124, "grad_norm": 1.4114342719720794, "learning_rate": 5.264917030448757e-07, "loss": 0.8704, "step": 4674 }, { "epoch": 0.8994276369582993, "grad_norm": 1.3713988502088446, "learning_rate": 5.244980460503268e-07, "loss": 0.9083, "step": 4675 }, { "epoch": 0.899620027896686, "grad_norm": 1.396218693184521, "learning_rate": 5.225080692315532e-07, "loss": 0.9574, "step": 4676 }, { "epoch": 0.8998124188350729, "grad_norm": 1.351467442637634, "learning_rate": 5.205217733614353e-07, "loss": 0.9854, "step": 4677 }, { "epoch": 0.9000048097734596, "grad_norm": 1.3436594401860407, "learning_rate": 5.185391592114286e-07, "loss": 0.94, "step": 4678 }, { "epoch": 0.9001972007118465, "grad_norm": 0.7445669082521276, "learning_rate": 5.165602275515592e-07, "loss": 0.9353, "step": 4679 }, { "epoch": 0.9003895916502332, "grad_norm": 1.4498435949478938, "learning_rate": 5.145849791504187e-07, "loss": 0.9452, "step": 4680 }, { "epoch": 0.9005819825886201, "grad_norm": 1.312126604561809, "learning_rate": 5.126134147751716e-07, "loss": 0.8973, "step": 4681 }, { "epoch": 0.9007743735270068, "grad_norm": 1.4936419053030658, "learning_rate": 5.106455351915507e-07, "loss": 0.9896, "step": 4682 }, { "epoch": 0.9009667644653937, "grad_norm": 1.3724387294216462, "learning_rate": 5.086813411638581e-07, "loss": 0.8929, "step": 4683 }, { "epoch": 0.9011591554037804, "grad_norm": 1.40069733307601, "learning_rate": 5.067208334549656e-07, "loss": 0.9835, "step": 4684 }, { "epoch": 0.9013515463421673, "grad_norm": 1.4069595449911678, "learning_rate": 5.047640128263087e-07, "loss": 0.9505, "step": 4685 }, { "epoch": 0.901543937280554, "grad_norm": 1.302762336548104, "learning_rate": 5.028108800378961e-07, "loss": 0.9474, "step": 4686 }, { "epoch": 0.9017363282189409, "grad_norm": 1.470994418371278, "learning_rate": 5.008614358483021e-07, "loss": 0.9748, "step": 4687 }, { "epoch": 0.9019287191573276, "grad_norm": 1.3039886497398576, "learning_rate": 4.989156810146667e-07, "loss": 1.0104, "step": 4688 }, { "epoch": 0.9021211100957145, "grad_norm": 1.3332814645215059, "learning_rate": 4.969736162927019e-07, "loss": 0.9192, "step": 4689 }, { "epoch": 0.9023135010341012, "grad_norm": 1.326417546919438, "learning_rate": 4.9503524243668e-07, "loss": 0.9457, "step": 4690 }, { "epoch": 0.9025058919724881, "grad_norm": 1.3488842457795274, "learning_rate": 4.931005601994432e-07, "loss": 0.98, "step": 4691 }, { "epoch": 0.902698282910875, "grad_norm": 1.4977768390619752, "learning_rate": 4.911695703324038e-07, "loss": 1.0276, "step": 4692 }, { "epoch": 0.9028906738492617, "grad_norm": 1.4702987747128973, "learning_rate": 4.892422735855284e-07, "loss": 0.8854, "step": 4693 }, { "epoch": 0.9030830647876485, "grad_norm": 1.395095283974757, "learning_rate": 4.873186707073663e-07, "loss": 0.9451, "step": 4694 }, { "epoch": 0.9032754557260353, "grad_norm": 1.485993861521044, "learning_rate": 4.853987624450151e-07, "loss": 0.9608, "step": 4695 }, { "epoch": 0.9034678466644221, "grad_norm": 1.4834881259721504, "learning_rate": 4.834825495441475e-07, "loss": 0.8609, "step": 4696 }, { "epoch": 0.9036602376028089, "grad_norm": 1.4087159968717016, "learning_rate": 4.815700327490014e-07, "loss": 0.954, "step": 4697 }, { "epoch": 0.9038526285411957, "grad_norm": 1.3411357920446463, "learning_rate": 4.796612128023726e-07, "loss": 0.9199, "step": 4698 }, { "epoch": 0.9040450194795825, "grad_norm": 1.5092344483522757, "learning_rate": 4.777560904456236e-07, "loss": 0.9297, "step": 4699 }, { "epoch": 0.9042374104179693, "grad_norm": 1.5361799490867791, "learning_rate": 4.7585466641868696e-07, "loss": 0.9311, "step": 4700 }, { "epoch": 0.9044298013563561, "grad_norm": 1.5301678758166257, "learning_rate": 4.7395694146004976e-07, "loss": 0.9155, "step": 4701 }, { "epoch": 0.9046221922947429, "grad_norm": 0.7187492904913628, "learning_rate": 4.7206291630677024e-07, "loss": 0.776, "step": 4702 }, { "epoch": 0.9048145832331297, "grad_norm": 1.3817742711963468, "learning_rate": 4.7017259169446104e-07, "loss": 0.9491, "step": 4703 }, { "epoch": 0.9050069741715165, "grad_norm": 1.3649123646017973, "learning_rate": 4.6828596835730487e-07, "loss": 0.9101, "step": 4704 }, { "epoch": 0.9051993651099033, "grad_norm": 1.4048812548044518, "learning_rate": 4.664030470280467e-07, "loss": 0.8349, "step": 4705 }, { "epoch": 0.9053917560482901, "grad_norm": 1.381744003582568, "learning_rate": 4.645238284379883e-07, "loss": 0.9054, "step": 4706 }, { "epoch": 0.9055841469866769, "grad_norm": 1.3712113237674781, "learning_rate": 4.626483133169968e-07, "loss": 0.9275, "step": 4707 }, { "epoch": 0.9057765379250637, "grad_norm": 1.2554297709443278, "learning_rate": 4.60776502393504e-07, "loss": 0.9995, "step": 4708 }, { "epoch": 0.9059689288634505, "grad_norm": 1.4015383662014287, "learning_rate": 4.5890839639449514e-07, "loss": 0.9436, "step": 4709 }, { "epoch": 0.9061613198018373, "grad_norm": 1.530710087062251, "learning_rate": 4.5704399604552417e-07, "loss": 0.9286, "step": 4710 }, { "epoch": 0.9063537107402241, "grad_norm": 1.4322418958158334, "learning_rate": 4.551833020707008e-07, "loss": 0.9598, "step": 4711 }, { "epoch": 0.9065461016786109, "grad_norm": 1.510213559646755, "learning_rate": 4.533263151926981e-07, "loss": 0.9338, "step": 4712 }, { "epoch": 0.9067384926169978, "grad_norm": 1.5448756428998378, "learning_rate": 4.514730361327502e-07, "loss": 0.9456, "step": 4713 }, { "epoch": 0.9069308835553845, "grad_norm": 1.3181271485157695, "learning_rate": 4.4962346561064574e-07, "loss": 0.9424, "step": 4714 }, { "epoch": 0.9071232744937714, "grad_norm": 1.3738826090008436, "learning_rate": 4.4777760434473796e-07, "loss": 0.9184, "step": 4715 }, { "epoch": 0.9073156654321581, "grad_norm": 1.406138187392525, "learning_rate": 4.4593545305193774e-07, "loss": 0.8575, "step": 4716 }, { "epoch": 0.907508056370545, "grad_norm": 1.316559981579428, "learning_rate": 4.440970124477173e-07, "loss": 1.0155, "step": 4717 }, { "epoch": 0.9077004473089317, "grad_norm": 1.3905853718446817, "learning_rate": 4.4226228324610544e-07, "loss": 0.9769, "step": 4718 }, { "epoch": 0.9078928382473186, "grad_norm": 1.3198483329051547, "learning_rate": 4.404312661596877e-07, "loss": 0.9821, "step": 4719 }, { "epoch": 0.9080852291857053, "grad_norm": 1.3979225417785113, "learning_rate": 4.386039618996119e-07, "loss": 1.023, "step": 4720 }, { "epoch": 0.9082776201240922, "grad_norm": 1.3836928933348818, "learning_rate": 4.3678037117558135e-07, "loss": 0.9138, "step": 4721 }, { "epoch": 0.9084700110624789, "grad_norm": 1.4033835861586903, "learning_rate": 4.349604946958563e-07, "loss": 0.9742, "step": 4722 }, { "epoch": 0.9086624020008658, "grad_norm": 1.4524845635441666, "learning_rate": 4.331443331672591e-07, "loss": 0.9488, "step": 4723 }, { "epoch": 0.9088547929392525, "grad_norm": 1.4115622938734884, "learning_rate": 4.313318872951633e-07, "loss": 0.9199, "step": 4724 }, { "epoch": 0.9090471838776394, "grad_norm": 1.2986398809783075, "learning_rate": 4.295231577835024e-07, "loss": 0.8776, "step": 4725 }, { "epoch": 0.9092395748160261, "grad_norm": 1.3065017954595104, "learning_rate": 4.2771814533476787e-07, "loss": 0.9606, "step": 4726 }, { "epoch": 0.909431965754413, "grad_norm": 1.373624039108555, "learning_rate": 4.2591685065000223e-07, "loss": 0.9393, "step": 4727 }, { "epoch": 0.9096243566927997, "grad_norm": 1.3695868213591162, "learning_rate": 4.241192744288092e-07, "loss": 0.9246, "step": 4728 }, { "epoch": 0.9098167476311866, "grad_norm": 1.398536016249744, "learning_rate": 4.223254173693492e-07, "loss": 0.9733, "step": 4729 }, { "epoch": 0.9100091385695733, "grad_norm": 1.3618299662105724, "learning_rate": 4.2053528016833267e-07, "loss": 0.8993, "step": 4730 }, { "epoch": 0.9102015295079602, "grad_norm": 1.4630302055401152, "learning_rate": 4.1874886352103015e-07, "loss": 0.9112, "step": 4731 }, { "epoch": 0.9103939204463469, "grad_norm": 1.5055722198410038, "learning_rate": 4.1696616812126333e-07, "loss": 0.9598, "step": 4732 }, { "epoch": 0.9105863113847338, "grad_norm": 1.431656895436711, "learning_rate": 4.1518719466141165e-07, "loss": 0.9001, "step": 4733 }, { "epoch": 0.9107787023231205, "grad_norm": 1.3558395720040128, "learning_rate": 4.13411943832408e-07, "loss": 0.9847, "step": 4734 }, { "epoch": 0.9109710932615074, "grad_norm": 1.479092564625573, "learning_rate": 4.116404163237386e-07, "loss": 0.9903, "step": 4735 }, { "epoch": 0.9111634841998942, "grad_norm": 1.4433981307021635, "learning_rate": 4.0987261282344425e-07, "loss": 0.9761, "step": 4736 }, { "epoch": 0.911355875138281, "grad_norm": 1.4766951000495343, "learning_rate": 4.081085340181223e-07, "loss": 0.9472, "step": 4737 }, { "epoch": 0.9115482660766678, "grad_norm": 1.3568994172578186, "learning_rate": 4.06348180592917e-07, "loss": 0.9282, "step": 4738 }, { "epoch": 0.9117406570150546, "grad_norm": 1.3706729190156093, "learning_rate": 4.0459155323153034e-07, "loss": 0.9708, "step": 4739 }, { "epoch": 0.9119330479534414, "grad_norm": 1.2959820320661246, "learning_rate": 4.028386526162176e-07, "loss": 0.9584, "step": 4740 }, { "epoch": 0.9121254388918282, "grad_norm": 1.3675914799183706, "learning_rate": 4.010894794277831e-07, "loss": 0.9164, "step": 4741 }, { "epoch": 0.912317829830215, "grad_norm": 1.4928270522777876, "learning_rate": 3.993440343455879e-07, "loss": 0.9777, "step": 4742 }, { "epoch": 0.9125102207686018, "grad_norm": 1.5141588107038624, "learning_rate": 3.976023180475397e-07, "loss": 0.9848, "step": 4743 }, { "epoch": 0.9127026117069886, "grad_norm": 1.4155051045705107, "learning_rate": 3.95864331210104e-07, "loss": 0.9534, "step": 4744 }, { "epoch": 0.9128950026453754, "grad_norm": 1.4224850388290906, "learning_rate": 3.941300745082932e-07, "loss": 0.9058, "step": 4745 }, { "epoch": 0.9130873935837622, "grad_norm": 1.4253597309585726, "learning_rate": 3.9239954861567177e-07, "loss": 0.9479, "step": 4746 }, { "epoch": 0.913279784522149, "grad_norm": 1.3722049973323158, "learning_rate": 3.906727542043598e-07, "loss": 0.9566, "step": 4747 }, { "epoch": 0.9134721754605358, "grad_norm": 1.393963908437136, "learning_rate": 3.8894969194502083e-07, "loss": 0.9092, "step": 4748 }, { "epoch": 0.9136645663989226, "grad_norm": 1.3730824487445532, "learning_rate": 3.87230362506873e-07, "loss": 0.9643, "step": 4749 }, { "epoch": 0.9138569573373094, "grad_norm": 1.5380935524031156, "learning_rate": 3.8551476655768527e-07, "loss": 0.9517, "step": 4750 }, { "epoch": 0.9140493482756962, "grad_norm": 1.4112540932045197, "learning_rate": 3.8380290476377255e-07, "loss": 1.0746, "step": 4751 }, { "epoch": 0.914241739214083, "grad_norm": 1.4260156299569728, "learning_rate": 3.8209477779000637e-07, "loss": 0.9879, "step": 4752 }, { "epoch": 0.9144341301524698, "grad_norm": 1.3505464419961968, "learning_rate": 3.803903862998004e-07, "loss": 0.8836, "step": 4753 }, { "epoch": 0.9146265210908566, "grad_norm": 1.4303586163617772, "learning_rate": 3.7868973095512185e-07, "loss": 0.9379, "step": 4754 }, { "epoch": 0.9148189120292434, "grad_norm": 1.3063519310642613, "learning_rate": 3.7699281241648565e-07, "loss": 0.9027, "step": 4755 }, { "epoch": 0.9150113029676302, "grad_norm": 1.3248929557570224, "learning_rate": 3.7529963134295466e-07, "loss": 0.9825, "step": 4756 }, { "epoch": 0.9152036939060171, "grad_norm": 1.3548302118379087, "learning_rate": 3.736101883921406e-07, "loss": 0.9511, "step": 4757 }, { "epoch": 0.9153960848444038, "grad_norm": 1.2747728893068335, "learning_rate": 3.719244842202074e-07, "loss": 0.9443, "step": 4758 }, { "epoch": 0.9155884757827907, "grad_norm": 1.3773414246070785, "learning_rate": 3.702425194818582e-07, "loss": 0.964, "step": 4759 }, { "epoch": 0.9157808667211774, "grad_norm": 0.7122962891869052, "learning_rate": 3.685642948303503e-07, "loss": 0.8086, "step": 4760 }, { "epoch": 0.9159732576595643, "grad_norm": 0.7576501645081201, "learning_rate": 3.66889810917489e-07, "loss": 0.8408, "step": 4761 }, { "epoch": 0.916165648597951, "grad_norm": 1.3634490744890138, "learning_rate": 3.6521906839362187e-07, "loss": 0.9316, "step": 4762 }, { "epoch": 0.9163580395363379, "grad_norm": 1.4072952813975534, "learning_rate": 3.635520679076465e-07, "loss": 0.9307, "step": 4763 }, { "epoch": 0.9165504304747246, "grad_norm": 1.4532643798252538, "learning_rate": 3.6188881010700725e-07, "loss": 1.0167, "step": 4764 }, { "epoch": 0.9167428214131115, "grad_norm": 1.3871588553860112, "learning_rate": 3.6022929563769513e-07, "loss": 0.9375, "step": 4765 }, { "epoch": 0.9169352123514982, "grad_norm": 1.3159530070494454, "learning_rate": 3.5857352514424573e-07, "loss": 0.9794, "step": 4766 }, { "epoch": 0.9171276032898851, "grad_norm": 1.4225073408691105, "learning_rate": 3.5692149926974006e-07, "loss": 0.9911, "step": 4767 }, { "epoch": 0.9173199942282718, "grad_norm": 1.3766320102690293, "learning_rate": 3.552732186558072e-07, "loss": 0.9062, "step": 4768 }, { "epoch": 0.9175123851666587, "grad_norm": 0.6814693254874394, "learning_rate": 3.536286839426195e-07, "loss": 0.8602, "step": 4769 }, { "epoch": 0.9177047761050454, "grad_norm": 1.3687648097480607, "learning_rate": 3.51987895768896e-07, "loss": 0.9665, "step": 4770 }, { "epoch": 0.9178971670434323, "grad_norm": 1.4149686001721664, "learning_rate": 3.5035085477190143e-07, "loss": 0.9975, "step": 4771 }, { "epoch": 0.918089557981819, "grad_norm": 1.4575431748271601, "learning_rate": 3.4871756158744054e-07, "loss": 0.9696, "step": 4772 }, { "epoch": 0.9182819489202059, "grad_norm": 1.3372628476806334, "learning_rate": 3.4708801684986693e-07, "loss": 1.0075, "step": 4773 }, { "epoch": 0.9184743398585926, "grad_norm": 1.2977644662635128, "learning_rate": 3.454622211920766e-07, "loss": 0.9368, "step": 4774 }, { "epoch": 0.9186667307969795, "grad_norm": 1.3066905421232633, "learning_rate": 3.4384017524551116e-07, "loss": 0.9987, "step": 4775 }, { "epoch": 0.9188591217353662, "grad_norm": 1.3811046958524065, "learning_rate": 3.422218796401544e-07, "loss": 0.8776, "step": 4776 }, { "epoch": 0.9190515126737531, "grad_norm": 1.376263656133721, "learning_rate": 3.4060733500453247e-07, "loss": 0.9871, "step": 4777 }, { "epoch": 0.9192439036121399, "grad_norm": 1.5022295293814951, "learning_rate": 3.38996541965716e-07, "loss": 0.9175, "step": 4778 }, { "epoch": 0.9194362945505267, "grad_norm": 1.263773034291039, "learning_rate": 3.3738950114932e-07, "loss": 0.8896, "step": 4779 }, { "epoch": 0.9196286854889135, "grad_norm": 0.7300500689924706, "learning_rate": 3.3578621317949755e-07, "loss": 0.8385, "step": 4780 }, { "epoch": 0.9198210764273003, "grad_norm": 1.4096648864616024, "learning_rate": 3.341866786789505e-07, "loss": 0.9043, "step": 4781 }, { "epoch": 0.9200134673656871, "grad_norm": 1.4401124718494585, "learning_rate": 3.325908982689185e-07, "loss": 0.9501, "step": 4782 }, { "epoch": 0.9202058583040739, "grad_norm": 1.308333817819037, "learning_rate": 3.309988725691837e-07, "loss": 0.8861, "step": 4783 }, { "epoch": 0.9203982492424607, "grad_norm": 1.5203208671216843, "learning_rate": 3.294106021980714e-07, "loss": 1.0031, "step": 4784 }, { "epoch": 0.9205906401808475, "grad_norm": 1.359592866733618, "learning_rate": 3.278260877724471e-07, "loss": 0.966, "step": 4785 }, { "epoch": 0.9207830311192343, "grad_norm": 1.406467261722654, "learning_rate": 3.2624532990771507e-07, "loss": 0.8877, "step": 4786 }, { "epoch": 0.9209754220576211, "grad_norm": 1.4681032388049189, "learning_rate": 3.2466832921782986e-07, "loss": 0.9256, "step": 4787 }, { "epoch": 0.9211678129960079, "grad_norm": 1.41128453460613, "learning_rate": 3.2309508631527486e-07, "loss": 0.9592, "step": 4788 }, { "epoch": 0.9213602039343947, "grad_norm": 1.3611120356146946, "learning_rate": 3.215256018110824e-07, "loss": 0.9432, "step": 4789 }, { "epoch": 0.9215525948727815, "grad_norm": 1.385124030364162, "learning_rate": 3.199598763148215e-07, "loss": 0.884, "step": 4790 }, { "epoch": 0.9217449858111683, "grad_norm": 0.7160300642500588, "learning_rate": 3.183979104346002e-07, "loss": 0.8409, "step": 4791 }, { "epoch": 0.9219373767495551, "grad_norm": 1.328539402027435, "learning_rate": 3.1683970477706994e-07, "loss": 0.8718, "step": 4792 }, { "epoch": 0.9221297676879419, "grad_norm": 1.4227940188556845, "learning_rate": 3.1528525994741876e-07, "loss": 0.9129, "step": 4793 }, { "epoch": 0.9223221586263287, "grad_norm": 1.4872880911317483, "learning_rate": 3.13734576549376e-07, "loss": 0.9784, "step": 4794 }, { "epoch": 0.9225145495647155, "grad_norm": 1.4568393049881205, "learning_rate": 3.121876551852099e-07, "loss": 0.9461, "step": 4795 }, { "epoch": 0.9227069405031023, "grad_norm": 1.251609911164816, "learning_rate": 3.1064449645572536e-07, "loss": 0.8652, "step": 4796 }, { "epoch": 0.9228993314414891, "grad_norm": 1.3882172015257088, "learning_rate": 3.091051009602675e-07, "loss": 1.0036, "step": 4797 }, { "epoch": 0.9230917223798759, "grad_norm": 1.3219072640104175, "learning_rate": 3.0756946929672017e-07, "loss": 0.9879, "step": 4798 }, { "epoch": 0.9232841133182627, "grad_norm": 1.360777553420641, "learning_rate": 3.060376020615052e-07, "loss": 0.8908, "step": 4799 }, { "epoch": 0.9234765042566495, "grad_norm": 1.3280289998138173, "learning_rate": 3.0450949984958347e-07, "loss": 0.9621, "step": 4800 }, { "epoch": 0.9236688951950364, "grad_norm": 1.3029565439285191, "learning_rate": 3.0298516325444893e-07, "loss": 0.8485, "step": 4801 }, { "epoch": 0.9238612861334231, "grad_norm": 1.4340723819596468, "learning_rate": 3.0146459286813924e-07, "loss": 0.9871, "step": 4802 }, { "epoch": 0.92405367707181, "grad_norm": 1.457177107585417, "learning_rate": 2.999477892812264e-07, "loss": 0.9342, "step": 4803 }, { "epoch": 0.9242460680101967, "grad_norm": 1.3007100299593615, "learning_rate": 2.984347530828158e-07, "loss": 0.8166, "step": 4804 }, { "epoch": 0.9244384589485836, "grad_norm": 1.3729141363183224, "learning_rate": 2.969254848605585e-07, "loss": 0.9118, "step": 4805 }, { "epoch": 0.9246308498869703, "grad_norm": 1.3736728782976757, "learning_rate": 2.9541998520063344e-07, "loss": 1.0645, "step": 4806 }, { "epoch": 0.9248232408253572, "grad_norm": 1.314308440413224, "learning_rate": 2.9391825468775946e-07, "loss": 0.8776, "step": 4807 }, { "epoch": 0.9250156317637439, "grad_norm": 1.2454453996199177, "learning_rate": 2.9242029390519454e-07, "loss": 0.9928, "step": 4808 }, { "epoch": 0.9252080227021308, "grad_norm": 1.4912930055954625, "learning_rate": 2.909261034347255e-07, "loss": 0.988, "step": 4809 }, { "epoch": 0.9254004136405175, "grad_norm": 1.444636599063678, "learning_rate": 2.894356838566792e-07, "loss": 0.9227, "step": 4810 }, { "epoch": 0.9255928045789044, "grad_norm": 0.7259810836947121, "learning_rate": 2.879490357499204e-07, "loss": 0.8279, "step": 4811 }, { "epoch": 0.9257851955172911, "grad_norm": 1.5039702251619993, "learning_rate": 2.864661596918428e-07, "loss": 0.95, "step": 4812 }, { "epoch": 0.925977586455678, "grad_norm": 1.5263513608198562, "learning_rate": 2.849870562583812e-07, "loss": 0.9799, "step": 4813 }, { "epoch": 0.9261699773940647, "grad_norm": 1.363824125813841, "learning_rate": 2.8351172602399945e-07, "loss": 0.895, "step": 4814 }, { "epoch": 0.9263623683324516, "grad_norm": 1.4113280507429407, "learning_rate": 2.8204016956169924e-07, "loss": 0.97, "step": 4815 }, { "epoch": 0.9265547592708383, "grad_norm": 1.3382174421651056, "learning_rate": 2.8057238744301994e-07, "loss": 0.9458, "step": 4816 }, { "epoch": 0.9267471502092252, "grad_norm": 1.4693337909012028, "learning_rate": 2.7910838023802676e-07, "loss": 0.9792, "step": 4817 }, { "epoch": 0.9269395411476119, "grad_norm": 1.380989108645725, "learning_rate": 2.7764814851532485e-07, "loss": 0.9542, "step": 4818 }, { "epoch": 0.9271319320859988, "grad_norm": 1.4592640716967586, "learning_rate": 2.761916928420527e-07, "loss": 0.9845, "step": 4819 }, { "epoch": 0.9273243230243855, "grad_norm": 1.4474173192264377, "learning_rate": 2.74739013783879e-07, "loss": 1.0045, "step": 4820 }, { "epoch": 0.9275167139627724, "grad_norm": 1.3820505954805773, "learning_rate": 2.7329011190500797e-07, "loss": 0.9004, "step": 4821 }, { "epoch": 0.9277091049011592, "grad_norm": 0.7352665697597479, "learning_rate": 2.7184498776817615e-07, "loss": 0.9021, "step": 4822 }, { "epoch": 0.927901495839546, "grad_norm": 1.386293604043885, "learning_rate": 2.704036419346534e-07, "loss": 0.9549, "step": 4823 }, { "epoch": 0.9280938867779328, "grad_norm": 1.4054876186284846, "learning_rate": 2.689660749642442e-07, "loss": 0.8494, "step": 4824 }, { "epoch": 0.9282862777163196, "grad_norm": 1.4270475875515902, "learning_rate": 2.675322874152786e-07, "loss": 0.9535, "step": 4825 }, { "epoch": 0.9284786686547064, "grad_norm": 1.3739235825066791, "learning_rate": 2.6610227984462556e-07, "loss": 0.9477, "step": 4826 }, { "epoch": 0.9286710595930932, "grad_norm": 1.420815339977204, "learning_rate": 2.646760528076842e-07, "loss": 0.9945, "step": 4827 }, { "epoch": 0.92886345053148, "grad_norm": 1.2900695568261065, "learning_rate": 2.6325360685838243e-07, "loss": 0.9164, "step": 4828 }, { "epoch": 0.9290558414698667, "grad_norm": 1.3781610689207862, "learning_rate": 2.618349425491851e-07, "loss": 0.9855, "step": 4829 }, { "epoch": 0.9292482324082536, "grad_norm": 0.6694450572589221, "learning_rate": 2.604200604310825e-07, "loss": 0.7708, "step": 4830 }, { "epoch": 0.9294406233466403, "grad_norm": 0.7170888714986191, "learning_rate": 2.590089610535984e-07, "loss": 0.8172, "step": 4831 }, { "epoch": 0.9296330142850272, "grad_norm": 1.5557497467615358, "learning_rate": 2.5760164496479e-07, "loss": 0.9888, "step": 4832 }, { "epoch": 0.929825405223414, "grad_norm": 1.503231368790645, "learning_rate": 2.5619811271123897e-07, "loss": 0.9383, "step": 4833 }, { "epoch": 0.9300177961618008, "grad_norm": 1.2855707543743276, "learning_rate": 2.5479836483806586e-07, "loss": 0.9192, "step": 4834 }, { "epoch": 0.9302101871001875, "grad_norm": 1.3523014357232925, "learning_rate": 2.5340240188891143e-07, "loss": 0.9649, "step": 4835 }, { "epoch": 0.9304025780385744, "grad_norm": 0.6664863441335636, "learning_rate": 2.520102244059552e-07, "loss": 0.8525, "step": 4836 }, { "epoch": 0.9305949689769611, "grad_norm": 1.409422911729024, "learning_rate": 2.506218329299026e-07, "loss": 0.9439, "step": 4837 }, { "epoch": 0.930787359915348, "grad_norm": 1.4921465502650995, "learning_rate": 2.4923722799998664e-07, "loss": 0.9417, "step": 4838 }, { "epoch": 0.9309797508537347, "grad_norm": 1.4074180670828444, "learning_rate": 2.4785641015397375e-07, "loss": 1.0078, "step": 4839 }, { "epoch": 0.9311721417921216, "grad_norm": 1.3030741988069425, "learning_rate": 2.464793799281573e-07, "loss": 0.8889, "step": 4840 }, { "epoch": 0.9313645327305083, "grad_norm": 1.3898372962666985, "learning_rate": 2.4510613785735936e-07, "loss": 0.9515, "step": 4841 }, { "epoch": 0.9315569236688952, "grad_norm": 1.5361847790859429, "learning_rate": 2.4373668447493225e-07, "loss": 0.9117, "step": 4842 }, { "epoch": 0.9317493146072819, "grad_norm": 1.4972333144555818, "learning_rate": 2.423710203127561e-07, "loss": 0.9026, "step": 4843 }, { "epoch": 0.9319417055456688, "grad_norm": 1.437379558132554, "learning_rate": 2.410091459012376e-07, "loss": 0.8986, "step": 4844 }, { "epoch": 0.9321340964840557, "grad_norm": 1.394015091508429, "learning_rate": 2.3965106176931375e-07, "loss": 0.9313, "step": 4845 }, { "epoch": 0.9323264874224424, "grad_norm": 1.30903465451443, "learning_rate": 2.3829676844444926e-07, "loss": 0.9478, "step": 4846 }, { "epoch": 0.9325188783608293, "grad_norm": 1.483680483237721, "learning_rate": 2.3694626645263675e-07, "loss": 0.9719, "step": 4847 }, { "epoch": 0.932711269299216, "grad_norm": 1.4488700041011884, "learning_rate": 2.3559955631839436e-07, "loss": 0.9802, "step": 4848 }, { "epoch": 0.9329036602376028, "grad_norm": 1.355669465540639, "learning_rate": 2.3425663856476932e-07, "loss": 0.9213, "step": 4849 }, { "epoch": 0.9330960511759896, "grad_norm": 1.480814215842873, "learning_rate": 2.3291751371333438e-07, "loss": 0.9693, "step": 4850 }, { "epoch": 0.9332884421143764, "grad_norm": 1.3898066208420832, "learning_rate": 2.3158218228419127e-07, "loss": 1.0005, "step": 4851 }, { "epoch": 0.9334808330527632, "grad_norm": 1.4246409870126284, "learning_rate": 2.3025064479596625e-07, "loss": 0.9339, "step": 4852 }, { "epoch": 0.93367322399115, "grad_norm": 1.2219485621061503, "learning_rate": 2.2892290176581678e-07, "loss": 0.9428, "step": 4853 }, { "epoch": 0.9338656149295368, "grad_norm": 0.7555545089450715, "learning_rate": 2.2759895370941809e-07, "loss": 0.8394, "step": 4854 }, { "epoch": 0.9340580058679236, "grad_norm": 1.2796939786016568, "learning_rate": 2.2627880114097779e-07, "loss": 0.9563, "step": 4855 }, { "epoch": 0.9342503968063104, "grad_norm": 1.4056323237845414, "learning_rate": 2.2496244457323013e-07, "loss": 0.8915, "step": 4856 }, { "epoch": 0.9344427877446972, "grad_norm": 1.3454167355783866, "learning_rate": 2.2364988451742953e-07, "loss": 0.9928, "step": 4857 }, { "epoch": 0.934635178683084, "grad_norm": 1.3392722792247163, "learning_rate": 2.2234112148336373e-07, "loss": 0.9335, "step": 4858 }, { "epoch": 0.9348275696214708, "grad_norm": 0.7277224211387331, "learning_rate": 2.2103615597933613e-07, "loss": 0.8172, "step": 4859 }, { "epoch": 0.9350199605598576, "grad_norm": 1.3092636871031593, "learning_rate": 2.1973498851218244e-07, "loss": 0.9423, "step": 4860 }, { "epoch": 0.9352123514982444, "grad_norm": 1.6277634386951998, "learning_rate": 2.1843761958726283e-07, "loss": 0.8773, "step": 4861 }, { "epoch": 0.9354047424366312, "grad_norm": 1.4760882048018042, "learning_rate": 2.1714404970845647e-07, "loss": 1.0058, "step": 4862 }, { "epoch": 0.935597133375018, "grad_norm": 1.3418133996974047, "learning_rate": 2.1585427937817594e-07, "loss": 0.9734, "step": 4863 }, { "epoch": 0.9357895243134048, "grad_norm": 1.359108950508893, "learning_rate": 2.145683090973494e-07, "loss": 0.889, "step": 4864 }, { "epoch": 0.9359819152517916, "grad_norm": 1.5676749879131062, "learning_rate": 2.1328613936543396e-07, "loss": 0.898, "step": 4865 }, { "epoch": 0.9361743061901785, "grad_norm": 1.4670284986989188, "learning_rate": 2.1200777068041134e-07, "loss": 0.9993, "step": 4866 }, { "epoch": 0.9363666971285652, "grad_norm": 1.3401481581739176, "learning_rate": 2.1073320353878102e-07, "loss": 0.9608, "step": 4867 }, { "epoch": 0.9365590880669521, "grad_norm": 0.7425317080204544, "learning_rate": 2.0946243843557367e-07, "loss": 0.856, "step": 4868 }, { "epoch": 0.9367514790053388, "grad_norm": 1.422915502715271, "learning_rate": 2.0819547586434008e-07, "loss": 0.9243, "step": 4869 }, { "epoch": 0.9369438699437257, "grad_norm": 1.4414241786500435, "learning_rate": 2.0693231631715105e-07, "loss": 0.9633, "step": 4870 }, { "epoch": 0.9371362608821124, "grad_norm": 1.3105441422450284, "learning_rate": 2.0567296028460638e-07, "loss": 0.8965, "step": 4871 }, { "epoch": 0.9373286518204993, "grad_norm": 1.2971513469578642, "learning_rate": 2.0441740825582258e-07, "loss": 0.9024, "step": 4872 }, { "epoch": 0.937521042758886, "grad_norm": 1.4084902091400855, "learning_rate": 2.0316566071844402e-07, "loss": 0.9607, "step": 4873 }, { "epoch": 0.9377134336972729, "grad_norm": 1.4839665992035265, "learning_rate": 2.0191771815863292e-07, "loss": 0.9588, "step": 4874 }, { "epoch": 0.9379058246356596, "grad_norm": 1.5375739015301377, "learning_rate": 2.0067358106107714e-07, "loss": 0.9579, "step": 4875 }, { "epoch": 0.9380982155740465, "grad_norm": 1.364669560067481, "learning_rate": 1.994332499089846e-07, "loss": 0.9141, "step": 4876 }, { "epoch": 0.9382906065124332, "grad_norm": 1.3988321735307787, "learning_rate": 1.9819672518408662e-07, "loss": 1.0083, "step": 4877 }, { "epoch": 0.9384829974508201, "grad_norm": 1.423412479033358, "learning_rate": 1.9696400736663457e-07, "loss": 0.9214, "step": 4878 }, { "epoch": 0.9386753883892068, "grad_norm": 1.4689961402854406, "learning_rate": 1.9573509693540104e-07, "loss": 0.9338, "step": 4879 }, { "epoch": 0.9388677793275937, "grad_norm": 1.458181953705896, "learning_rate": 1.9450999436768093e-07, "loss": 0.9158, "step": 4880 }, { "epoch": 0.9390601702659804, "grad_norm": 1.325270064234449, "learning_rate": 1.9328870013929134e-07, "loss": 0.9333, "step": 4881 }, { "epoch": 0.9392525612043673, "grad_norm": 1.4013221819981683, "learning_rate": 1.9207121472456846e-07, "loss": 0.9, "step": 4882 }, { "epoch": 0.939444952142754, "grad_norm": 1.4751738950147377, "learning_rate": 1.9085753859636736e-07, "loss": 0.9237, "step": 4883 }, { "epoch": 0.9396373430811409, "grad_norm": 0.7506700050861405, "learning_rate": 1.8964767222606873e-07, "loss": 0.8487, "step": 4884 }, { "epoch": 0.9398297340195276, "grad_norm": 1.4364357812292352, "learning_rate": 1.8844161608356782e-07, "loss": 0.884, "step": 4885 }, { "epoch": 0.9400221249579145, "grad_norm": 1.4250946336830461, "learning_rate": 1.872393706372866e-07, "loss": 0.9604, "step": 4886 }, { "epoch": 0.9402145158963013, "grad_norm": 0.7078508101266772, "learning_rate": 1.8604093635416155e-07, "loss": 0.8523, "step": 4887 }, { "epoch": 0.9404069068346881, "grad_norm": 1.3040096605611842, "learning_rate": 1.8484631369964922e-07, "loss": 1.0101, "step": 4888 }, { "epoch": 0.9405992977730749, "grad_norm": 0.7056447334157225, "learning_rate": 1.8365550313772852e-07, "loss": 0.7774, "step": 4889 }, { "epoch": 0.9407916887114617, "grad_norm": 1.4432123114428104, "learning_rate": 1.8246850513089832e-07, "loss": 0.977, "step": 4890 }, { "epoch": 0.9409840796498485, "grad_norm": 1.2968926097188802, "learning_rate": 1.8128532014017098e-07, "loss": 0.8844, "step": 4891 }, { "epoch": 0.9411764705882353, "grad_norm": 0.7234206006887813, "learning_rate": 1.8010594862508669e-07, "loss": 0.8554, "step": 4892 }, { "epoch": 0.9413688615266221, "grad_norm": 1.491748801358031, "learning_rate": 1.789303910436968e-07, "loss": 0.9095, "step": 4893 }, { "epoch": 0.9415612524650089, "grad_norm": 1.3914806735597054, "learning_rate": 1.777586478525739e-07, "loss": 0.9401, "step": 4894 }, { "epoch": 0.9417536434033957, "grad_norm": 1.2459858861347146, "learning_rate": 1.7659071950681172e-07, "loss": 0.9197, "step": 4895 }, { "epoch": 0.9419460343417825, "grad_norm": 1.4476420192307096, "learning_rate": 1.754266064600174e-07, "loss": 1.0132, "step": 4896 }, { "epoch": 0.9421384252801693, "grad_norm": 1.367745372597132, "learning_rate": 1.742663091643204e-07, "loss": 0.9407, "step": 4897 }, { "epoch": 0.9423308162185561, "grad_norm": 1.5165233837881584, "learning_rate": 1.7310982807036915e-07, "loss": 1.0271, "step": 4898 }, { "epoch": 0.9425232071569429, "grad_norm": 1.325746377992132, "learning_rate": 1.719571636273243e-07, "loss": 0.9747, "step": 4899 }, { "epoch": 0.9427155980953297, "grad_norm": 1.4826369916397708, "learning_rate": 1.7080831628286886e-07, "loss": 0.9293, "step": 4900 }, { "epoch": 0.9429079890337165, "grad_norm": 1.4394695460444458, "learning_rate": 1.6966328648320152e-07, "loss": 0.8796, "step": 4901 }, { "epoch": 0.9431003799721033, "grad_norm": 1.4715777339269027, "learning_rate": 1.685220746730387e-07, "loss": 0.9566, "step": 4902 }, { "epoch": 0.9432927709104901, "grad_norm": 1.4291181041455305, "learning_rate": 1.673846812956137e-07, "loss": 0.9142, "step": 4903 }, { "epoch": 0.9434851618488769, "grad_norm": 1.4843415901581687, "learning_rate": 1.6625110679267642e-07, "loss": 0.9332, "step": 4904 }, { "epoch": 0.9436775527872637, "grad_norm": 1.3647889407622946, "learning_rate": 1.6512135160449583e-07, "loss": 0.9634, "step": 4905 }, { "epoch": 0.9438699437256505, "grad_norm": 1.2846953087076187, "learning_rate": 1.6399541616985648e-07, "loss": 1.009, "step": 4906 }, { "epoch": 0.9440623346640373, "grad_norm": 1.4023558645047423, "learning_rate": 1.6287330092605525e-07, "loss": 0.9346, "step": 4907 }, { "epoch": 0.9442547256024241, "grad_norm": 1.3506026919398129, "learning_rate": 1.6175500630891128e-07, "loss": 0.8862, "step": 4908 }, { "epoch": 0.9444471165408109, "grad_norm": 1.4030755124958119, "learning_rate": 1.6064053275275716e-07, "loss": 0.9438, "step": 4909 }, { "epoch": 0.9446395074791978, "grad_norm": 1.3929382407944937, "learning_rate": 1.5952988069044105e-07, "loss": 0.9789, "step": 4910 }, { "epoch": 0.9448318984175845, "grad_norm": 1.461364203239357, "learning_rate": 1.5842305055332796e-07, "loss": 0.9524, "step": 4911 }, { "epoch": 0.9450242893559714, "grad_norm": 1.4294897107981266, "learning_rate": 1.573200427712973e-07, "loss": 0.91, "step": 4912 }, { "epoch": 0.9452166802943581, "grad_norm": 0.6569334715349631, "learning_rate": 1.562208577727442e-07, "loss": 0.829, "step": 4913 }, { "epoch": 0.945409071232745, "grad_norm": 1.401575287636099, "learning_rate": 1.551254959845805e-07, "loss": 1.0602, "step": 4914 }, { "epoch": 0.9456014621711317, "grad_norm": 0.7202555876592122, "learning_rate": 1.540339578322314e-07, "loss": 0.8824, "step": 4915 }, { "epoch": 0.9457938531095186, "grad_norm": 1.3723293602394784, "learning_rate": 1.5294624373963894e-07, "loss": 0.9684, "step": 4916 }, { "epoch": 0.9459862440479053, "grad_norm": 1.344802341467013, "learning_rate": 1.5186235412925744e-07, "loss": 0.9644, "step": 4917 }, { "epoch": 0.9461786349862922, "grad_norm": 1.339038189255991, "learning_rate": 1.5078228942205674e-07, "loss": 0.9325, "step": 4918 }, { "epoch": 0.9463710259246789, "grad_norm": 1.462778500119026, "learning_rate": 1.4970605003752359e-07, "loss": 0.8882, "step": 4919 }, { "epoch": 0.9465634168630658, "grad_norm": 1.5888555546395142, "learning_rate": 1.4863363639365357e-07, "loss": 0.9607, "step": 4920 }, { "epoch": 0.9467558078014525, "grad_norm": 1.5901647809745358, "learning_rate": 1.4756504890696466e-07, "loss": 1.014, "step": 4921 }, { "epoch": 0.9469481987398394, "grad_norm": 1.473067267609338, "learning_rate": 1.4650028799247928e-07, "loss": 1.0364, "step": 4922 }, { "epoch": 0.9471405896782261, "grad_norm": 1.406158874671065, "learning_rate": 1.454393540637411e-07, "loss": 0.8895, "step": 4923 }, { "epoch": 0.947332980616613, "grad_norm": 1.3396889364991142, "learning_rate": 1.4438224753280384e-07, "loss": 0.9541, "step": 4924 }, { "epoch": 0.9475253715549997, "grad_norm": 1.3389031423134408, "learning_rate": 1.4332896881023462e-07, "loss": 0.9191, "step": 4925 }, { "epoch": 0.9477177624933866, "grad_norm": 1.5166677014940364, "learning_rate": 1.422795183051151e-07, "loss": 0.9995, "step": 4926 }, { "epoch": 0.9479101534317733, "grad_norm": 1.4889272310147441, "learning_rate": 1.4123389642504148e-07, "loss": 0.9252, "step": 4927 }, { "epoch": 0.9481025443701602, "grad_norm": 1.3902942269152845, "learning_rate": 1.401921035761189e-07, "loss": 0.8973, "step": 4928 }, { "epoch": 0.9482949353085469, "grad_norm": 1.4240719517761404, "learning_rate": 1.3915414016296925e-07, "loss": 0.9871, "step": 4929 }, { "epoch": 0.9484873262469338, "grad_norm": 1.4440965380689772, "learning_rate": 1.381200065887256e-07, "loss": 0.9238, "step": 4930 }, { "epoch": 0.9486797171853206, "grad_norm": 1.3745378015115275, "learning_rate": 1.3708970325503222e-07, "loss": 0.9274, "step": 4931 }, { "epoch": 0.9488721081237074, "grad_norm": 1.5333573341795785, "learning_rate": 1.3606323056204795e-07, "loss": 1.0231, "step": 4932 }, { "epoch": 0.9490644990620942, "grad_norm": 1.5513726214184451, "learning_rate": 1.3504058890844274e-07, "loss": 1.0239, "step": 4933 }, { "epoch": 0.949256890000481, "grad_norm": 1.2619891580867617, "learning_rate": 1.3402177869139887e-07, "loss": 0.9204, "step": 4934 }, { "epoch": 0.9494492809388678, "grad_norm": 1.3525345205957966, "learning_rate": 1.3300680030661096e-07, "loss": 0.9803, "step": 4935 }, { "epoch": 0.9496416718772546, "grad_norm": 1.3886119264636012, "learning_rate": 1.3199565414828363e-07, "loss": 0.9607, "step": 4936 }, { "epoch": 0.9498340628156414, "grad_norm": 1.3817218774802473, "learning_rate": 1.3098834060913612e-07, "loss": 0.9482, "step": 4937 }, { "epoch": 0.9500264537540282, "grad_norm": 1.3616056351221895, "learning_rate": 1.2998486008039545e-07, "loss": 1.0068, "step": 4938 }, { "epoch": 0.950218844692415, "grad_norm": 1.3740833831756585, "learning_rate": 1.289852129518032e-07, "loss": 0.9507, "step": 4939 }, { "epoch": 0.9504112356308018, "grad_norm": 1.4272797774825885, "learning_rate": 1.2798939961161217e-07, "loss": 0.9336, "step": 4940 }, { "epoch": 0.9506036265691886, "grad_norm": 1.2993677196766298, "learning_rate": 1.269974204465818e-07, "loss": 0.956, "step": 4941 }, { "epoch": 0.9507960175075754, "grad_norm": 1.3654332759112426, "learning_rate": 1.2600927584198618e-07, "loss": 0.9429, "step": 4942 }, { "epoch": 0.9509884084459622, "grad_norm": 1.3937732733763826, "learning_rate": 1.2502496618161165e-07, "loss": 0.9221, "step": 4943 }, { "epoch": 0.951180799384349, "grad_norm": 1.3639110322814794, "learning_rate": 1.2404449184774904e-07, "loss": 0.9637, "step": 4944 }, { "epoch": 0.9513731903227358, "grad_norm": 1.2670754080636224, "learning_rate": 1.2306785322120596e-07, "loss": 0.9348, "step": 4945 }, { "epoch": 0.9515655812611226, "grad_norm": 1.5826604701508589, "learning_rate": 1.220950506812968e-07, "loss": 0.9473, "step": 4946 }, { "epoch": 0.9517579721995094, "grad_norm": 1.363099862516401, "learning_rate": 1.2112608460584707e-07, "loss": 0.9109, "step": 4947 }, { "epoch": 0.9519503631378962, "grad_norm": 1.438180046570719, "learning_rate": 1.2016095537119242e-07, "loss": 0.8589, "step": 4948 }, { "epoch": 0.952142754076283, "grad_norm": 1.3367795519815089, "learning_rate": 1.1919966335217636e-07, "loss": 0.9237, "step": 4949 }, { "epoch": 0.9523351450146698, "grad_norm": 1.3520216741572786, "learning_rate": 1.1824220892215465e-07, "loss": 1.0045, "step": 4950 }, { "epoch": 0.9525275359530566, "grad_norm": 1.3708258651531617, "learning_rate": 1.1728859245299207e-07, "loss": 0.9198, "step": 4951 }, { "epoch": 0.9527199268914434, "grad_norm": 1.4997798390539845, "learning_rate": 1.1633881431506122e-07, "loss": 0.9523, "step": 4952 }, { "epoch": 0.9529123178298302, "grad_norm": 1.5084628780753568, "learning_rate": 1.1539287487724594e-07, "loss": 0.9549, "step": 4953 }, { "epoch": 0.9531047087682171, "grad_norm": 1.3782503882983843, "learning_rate": 1.1445077450693786e-07, "loss": 1.0038, "step": 4954 }, { "epoch": 0.9532970997066038, "grad_norm": 1.4178389494597023, "learning_rate": 1.1351251357003656e-07, "loss": 0.95, "step": 4955 }, { "epoch": 0.9534894906449907, "grad_norm": 1.4083997592422888, "learning_rate": 1.1257809243095385e-07, "loss": 0.9617, "step": 4956 }, { "epoch": 0.9536818815833774, "grad_norm": 1.3709217639376872, "learning_rate": 1.1164751145260722e-07, "loss": 0.9386, "step": 4957 }, { "epoch": 0.9538742725217643, "grad_norm": 1.3631193657599345, "learning_rate": 1.107207709964242e-07, "loss": 0.9636, "step": 4958 }, { "epoch": 0.954066663460151, "grad_norm": 1.2219897373960522, "learning_rate": 1.0979787142233911e-07, "loss": 0.91, "step": 4959 }, { "epoch": 0.9542590543985379, "grad_norm": 1.3466732377989226, "learning_rate": 1.0887881308879633e-07, "loss": 0.9577, "step": 4960 }, { "epoch": 0.9544514453369246, "grad_norm": 1.4160224053835893, "learning_rate": 1.0796359635274701e-07, "loss": 0.9454, "step": 4961 }, { "epoch": 0.9546438362753115, "grad_norm": 1.4460489774402483, "learning_rate": 1.0705222156965011e-07, "loss": 0.98, "step": 4962 }, { "epoch": 0.9548362272136982, "grad_norm": 1.4176414768160286, "learning_rate": 1.0614468909347476e-07, "loss": 0.8947, "step": 4963 }, { "epoch": 0.9550286181520851, "grad_norm": 1.4381937322195062, "learning_rate": 1.0524099927669563e-07, "loss": 0.9495, "step": 4964 }, { "epoch": 0.9552210090904718, "grad_norm": 1.4115167458704507, "learning_rate": 1.0434115247029419e-07, "loss": 0.9557, "step": 4965 }, { "epoch": 0.9554134000288587, "grad_norm": 1.345150237955829, "learning_rate": 1.0344514902376201e-07, "loss": 0.9616, "step": 4966 }, { "epoch": 0.9556057909672454, "grad_norm": 1.5373192110085032, "learning_rate": 1.0255298928509627e-07, "loss": 0.9591, "step": 4967 }, { "epoch": 0.9557981819056323, "grad_norm": 1.4667258734744282, "learning_rate": 1.0166467360079979e-07, "loss": 0.9753, "step": 4968 }, { "epoch": 0.955990572844019, "grad_norm": 1.2667965087662116, "learning_rate": 1.007802023158877e-07, "loss": 0.9703, "step": 4969 }, { "epoch": 0.9561829637824059, "grad_norm": 1.4057302720587301, "learning_rate": 9.989957577387521e-08, "loss": 0.9267, "step": 4970 }, { "epoch": 0.9563753547207926, "grad_norm": 1.373777226465329, "learning_rate": 9.902279431678874e-08, "loss": 0.9008, "step": 4971 }, { "epoch": 0.9565677456591795, "grad_norm": 0.7458377353644671, "learning_rate": 9.814985828516033e-08, "loss": 0.8611, "step": 4972 }, { "epoch": 0.9567601365975662, "grad_norm": 1.3240618855550035, "learning_rate": 9.728076801802656e-08, "loss": 0.9081, "step": 4973 }, { "epoch": 0.9569525275359531, "grad_norm": 1.4847120138934131, "learning_rate": 9.641552385293518e-08, "loss": 0.9041, "step": 4974 }, { "epoch": 0.9571449184743399, "grad_norm": 1.2884297797701845, "learning_rate": 9.555412612593518e-08, "loss": 0.887, "step": 4975 }, { "epoch": 0.9573373094127267, "grad_norm": 1.5468776981938448, "learning_rate": 9.469657517158226e-08, "loss": 0.8944, "step": 4976 }, { "epoch": 0.9575297003511135, "grad_norm": 1.485779774924814, "learning_rate": 9.384287132294223e-08, "loss": 0.9732, "step": 4977 }, { "epoch": 0.9577220912895003, "grad_norm": 0.7298737855523182, "learning_rate": 9.299301491158207e-08, "loss": 0.8818, "step": 4978 }, { "epoch": 0.9579144822278871, "grad_norm": 1.3490716320328178, "learning_rate": 9.214700626757667e-08, "loss": 1.0116, "step": 4979 }, { "epoch": 0.9581068731662739, "grad_norm": 1.487810572154167, "learning_rate": 9.130484571950538e-08, "loss": 0.9718, "step": 4980 }, { "epoch": 0.9582992641046607, "grad_norm": 1.444702844455353, "learning_rate": 9.046653359445323e-08, "loss": 0.9413, "step": 4981 }, { "epoch": 0.9584916550430475, "grad_norm": 1.3185400531368876, "learning_rate": 8.963207021801423e-08, "loss": 0.9156, "step": 4982 }, { "epoch": 0.9586840459814343, "grad_norm": 1.3763665520694792, "learning_rate": 8.880145591428024e-08, "loss": 0.9522, "step": 4983 }, { "epoch": 0.958876436919821, "grad_norm": 1.4783197619709878, "learning_rate": 8.797469100585432e-08, "loss": 0.9751, "step": 4984 }, { "epoch": 0.9590688278582079, "grad_norm": 1.2847819630305584, "learning_rate": 8.715177581384182e-08, "loss": 0.8958, "step": 4985 }, { "epoch": 0.9592612187965947, "grad_norm": 1.4771356652785292, "learning_rate": 8.633271065785486e-08, "loss": 0.9449, "step": 4986 }, { "epoch": 0.9594536097349815, "grad_norm": 1.390934580048981, "learning_rate": 8.551749585600678e-08, "loss": 0.8551, "step": 4987 }, { "epoch": 0.9596460006733682, "grad_norm": 1.3656828216123778, "learning_rate": 8.470613172491981e-08, "loss": 0.9041, "step": 4988 }, { "epoch": 0.9598383916117551, "grad_norm": 1.4569291561666327, "learning_rate": 8.389861857971748e-08, "loss": 0.8864, "step": 4989 }, { "epoch": 0.9600307825501418, "grad_norm": 1.573878515882926, "learning_rate": 8.309495673402778e-08, "loss": 1.0273, "step": 4990 }, { "epoch": 0.9602231734885287, "grad_norm": 1.4728546218572967, "learning_rate": 8.229514649998438e-08, "loss": 0.9295, "step": 4991 }, { "epoch": 0.9604155644269154, "grad_norm": 1.2800494225224903, "learning_rate": 8.149918818822433e-08, "loss": 0.9132, "step": 4992 }, { "epoch": 0.9606079553653023, "grad_norm": 1.3413786191691683, "learning_rate": 8.070708210788925e-08, "loss": 0.9678, "step": 4993 }, { "epoch": 0.960800346303689, "grad_norm": 1.3585055716653953, "learning_rate": 7.991882856662303e-08, "loss": 1.0014, "step": 4994 }, { "epoch": 0.9609927372420759, "grad_norm": 1.3154735039552965, "learning_rate": 7.913442787057523e-08, "loss": 0.8835, "step": 4995 }, { "epoch": 0.9611851281804628, "grad_norm": 1.423755230912406, "learning_rate": 7.835388032439661e-08, "loss": 0.9894, "step": 4996 }, { "epoch": 0.9613775191188495, "grad_norm": 1.5016504446827075, "learning_rate": 7.757718623124466e-08, "loss": 0.9526, "step": 4997 }, { "epoch": 0.9615699100572364, "grad_norm": 1.3447100091606372, "learning_rate": 7.680434589277696e-08, "loss": 0.8938, "step": 4998 }, { "epoch": 0.9617623009956231, "grad_norm": 1.3191957411321749, "learning_rate": 7.603535960915675e-08, "loss": 0.8443, "step": 4999 }, { "epoch": 0.96195469193401, "grad_norm": 1.3820732294977651, "learning_rate": 7.527022767904957e-08, "loss": 0.9201, "step": 5000 }, { "epoch": 0.9621470828723967, "grad_norm": 1.384337784977436, "learning_rate": 7.450895039962214e-08, "loss": 0.857, "step": 5001 }, { "epoch": 0.9623394738107836, "grad_norm": 1.4927298844471002, "learning_rate": 7.375152806654685e-08, "loss": 0.9775, "step": 5002 }, { "epoch": 0.9625318647491703, "grad_norm": 1.3731220792607624, "learning_rate": 7.299796097399947e-08, "loss": 0.9204, "step": 5003 }, { "epoch": 0.9627242556875572, "grad_norm": 1.561040938623386, "learning_rate": 7.224824941465369e-08, "loss": 0.9752, "step": 5004 }, { "epoch": 0.9629166466259439, "grad_norm": 1.4463641607522713, "learning_rate": 7.150239367969102e-08, "loss": 0.9772, "step": 5005 }, { "epoch": 0.9631090375643307, "grad_norm": 1.3729326402511057, "learning_rate": 7.076039405879309e-08, "loss": 0.886, "step": 5006 }, { "epoch": 0.9633014285027175, "grad_norm": 1.413519844392199, "learning_rate": 7.002225084014269e-08, "loss": 0.9459, "step": 5007 }, { "epoch": 0.9634938194411043, "grad_norm": 1.347850389366427, "learning_rate": 6.928796431042717e-08, "loss": 0.926, "step": 5008 }, { "epoch": 0.9636862103794911, "grad_norm": 1.4387264351088351, "learning_rate": 6.855753475483507e-08, "loss": 0.9992, "step": 5009 }, { "epoch": 0.963878601317878, "grad_norm": 1.4671215086054659, "learning_rate": 6.783096245705612e-08, "loss": 0.8874, "step": 5010 }, { "epoch": 0.9640709922562647, "grad_norm": 1.4297412057871808, "learning_rate": 6.710824769928349e-08, "loss": 0.9764, "step": 5011 }, { "epoch": 0.9642633831946515, "grad_norm": 1.4427175773959193, "learning_rate": 6.63893907622104e-08, "loss": 1.0203, "step": 5012 }, { "epoch": 0.9644557741330383, "grad_norm": 1.4067315234194224, "learning_rate": 6.567439192503244e-08, "loss": 0.9966, "step": 5013 }, { "epoch": 0.9646481650714251, "grad_norm": 1.2844094517773847, "learning_rate": 6.496325146544746e-08, "loss": 0.8864, "step": 5014 }, { "epoch": 0.9648405560098119, "grad_norm": 1.3927363543185904, "learning_rate": 6.425596965965453e-08, "loss": 0.9795, "step": 5015 }, { "epoch": 0.9650329469481987, "grad_norm": 1.4833249500046566, "learning_rate": 6.35525467823539e-08, "loss": 0.9974, "step": 5016 }, { "epoch": 0.9652253378865855, "grad_norm": 1.4171349341270818, "learning_rate": 6.285298310674703e-08, "loss": 0.9575, "step": 5017 }, { "epoch": 0.9654177288249723, "grad_norm": 1.3322849974044912, "learning_rate": 6.215727890453438e-08, "loss": 0.9065, "step": 5018 }, { "epoch": 0.9656101197633592, "grad_norm": 1.3761887685305738, "learning_rate": 6.14654344459209e-08, "loss": 0.9879, "step": 5019 }, { "epoch": 0.9658025107017459, "grad_norm": 1.3189786356999136, "learning_rate": 6.077744999961166e-08, "loss": 0.9416, "step": 5020 }, { "epoch": 0.9659949016401328, "grad_norm": 1.4741771876864835, "learning_rate": 6.009332583281069e-08, "loss": 0.9618, "step": 5021 }, { "epoch": 0.9661872925785195, "grad_norm": 0.7460096572907514, "learning_rate": 5.941306221122545e-08, "loss": 0.8515, "step": 5022 }, { "epoch": 0.9663796835169064, "grad_norm": 1.460689163953328, "learning_rate": 5.873665939906015e-08, "loss": 0.9116, "step": 5023 }, { "epoch": 0.9665720744552931, "grad_norm": 1.4967383757795982, "learning_rate": 5.806411765902353e-08, "loss": 1.0012, "step": 5024 }, { "epoch": 0.96676446539368, "grad_norm": 1.5289232126242003, "learning_rate": 5.739543725232333e-08, "loss": 0.9431, "step": 5025 }, { "epoch": 0.9669568563320667, "grad_norm": 1.2708387170596336, "learning_rate": 5.673061843866623e-08, "loss": 0.9841, "step": 5026 }, { "epoch": 0.9671492472704536, "grad_norm": 1.4270368302892886, "learning_rate": 5.606966147626125e-08, "loss": 0.8371, "step": 5027 }, { "epoch": 0.9673416382088403, "grad_norm": 1.5320571108589764, "learning_rate": 5.5412566621815266e-08, "loss": 0.948, "step": 5028 }, { "epoch": 0.9675340291472272, "grad_norm": 1.3808949255526, "learning_rate": 5.4759334130536355e-08, "loss": 0.9462, "step": 5029 }, { "epoch": 0.9677264200856139, "grad_norm": 1.3755261716872929, "learning_rate": 5.41099642561338e-08, "loss": 0.9558, "step": 5030 }, { "epoch": 0.9679188110240008, "grad_norm": 0.6905240386259744, "learning_rate": 5.3464457250814726e-08, "loss": 0.8081, "step": 5031 }, { "epoch": 0.9681112019623875, "grad_norm": 1.4649257727672458, "learning_rate": 5.2822813365286386e-08, "loss": 0.8593, "step": 5032 }, { "epoch": 0.9683035929007744, "grad_norm": 0.7016778880064947, "learning_rate": 5.218503284875609e-08, "loss": 0.8138, "step": 5033 }, { "epoch": 0.9684959838391611, "grad_norm": 1.3728681759108028, "learning_rate": 5.155111594893014e-08, "loss": 0.9063, "step": 5034 }, { "epoch": 0.968688374777548, "grad_norm": 1.393594561603896, "learning_rate": 5.092106291201604e-08, "loss": 0.9787, "step": 5035 }, { "epoch": 0.9688807657159347, "grad_norm": 1.5293145446542569, "learning_rate": 5.029487398271693e-08, "loss": 0.9823, "step": 5036 }, { "epoch": 0.9690731566543216, "grad_norm": 1.388978152451303, "learning_rate": 4.967254940423716e-08, "loss": 0.9171, "step": 5037 }, { "epoch": 0.9692655475927083, "grad_norm": 1.4121388667297898, "learning_rate": 4.905408941828338e-08, "loss": 0.8872, "step": 5038 }, { "epoch": 0.9694579385310952, "grad_norm": 1.5335356773764293, "learning_rate": 4.8439494265055674e-08, "loss": 0.895, "step": 5039 }, { "epoch": 0.969650329469482, "grad_norm": 1.489095902232824, "learning_rate": 4.7828764183257545e-08, "loss": 0.9731, "step": 5040 }, { "epoch": 0.9698427204078688, "grad_norm": 1.5074285241535772, "learning_rate": 4.722189941008703e-08, "loss": 1.0107, "step": 5041 }, { "epoch": 0.9700351113462556, "grad_norm": 1.5007570157366978, "learning_rate": 4.66189001812456e-08, "loss": 1.0202, "step": 5042 }, { "epoch": 0.9702275022846424, "grad_norm": 1.3753695135554473, "learning_rate": 4.6019766730930336e-08, "loss": 0.9237, "step": 5043 }, { "epoch": 0.9704198932230292, "grad_norm": 1.3925551057282834, "learning_rate": 4.542449929183845e-08, "loss": 0.8641, "step": 5044 }, { "epoch": 0.970612284161416, "grad_norm": 1.3607761768426565, "learning_rate": 4.4833098095165005e-08, "loss": 0.9152, "step": 5045 }, { "epoch": 0.9708046750998028, "grad_norm": 1.2883178521048662, "learning_rate": 4.424556337060182e-08, "loss": 0.932, "step": 5046 }, { "epoch": 0.9709970660381896, "grad_norm": 1.439212768941228, "learning_rate": 4.366189534634191e-08, "loss": 0.9119, "step": 5047 }, { "epoch": 0.9711894569765764, "grad_norm": 1.5641156487454169, "learning_rate": 4.308209424907506e-08, "loss": 0.9281, "step": 5048 }, { "epoch": 0.9713818479149632, "grad_norm": 1.360960147000317, "learning_rate": 4.250616030399002e-08, "loss": 0.9952, "step": 5049 }, { "epoch": 0.97157423885335, "grad_norm": 1.3992320162197236, "learning_rate": 4.1934093734771194e-08, "loss": 0.9394, "step": 5050 }, { "epoch": 0.9717666297917368, "grad_norm": 1.380053983710579, "learning_rate": 4.13658947636042e-08, "loss": 0.8921, "step": 5051 }, { "epoch": 0.9719590207301236, "grad_norm": 1.3311226352206151, "learning_rate": 4.080156361117027e-08, "loss": 0.8983, "step": 5052 }, { "epoch": 0.9721514116685104, "grad_norm": 1.3366461177276387, "learning_rate": 4.024110049664853e-08, "loss": 0.9131, "step": 5053 }, { "epoch": 0.9723438026068972, "grad_norm": 1.4629929556956887, "learning_rate": 3.9684505637718194e-08, "loss": 0.9194, "step": 5054 }, { "epoch": 0.972536193545284, "grad_norm": 1.4229228625140462, "learning_rate": 3.913177925055189e-08, "loss": 0.8984, "step": 5055 }, { "epoch": 0.9727285844836708, "grad_norm": 1.5315583328761844, "learning_rate": 3.858292154982457e-08, "loss": 0.9688, "step": 5056 }, { "epoch": 0.9729209754220576, "grad_norm": 1.426260024589343, "learning_rate": 3.8037932748704596e-08, "loss": 0.9875, "step": 5057 }, { "epoch": 0.9731133663604444, "grad_norm": 1.3270300769000924, "learning_rate": 3.7496813058859325e-08, "loss": 0.9524, "step": 5058 }, { "epoch": 0.9733057572988312, "grad_norm": 1.3500471130928928, "learning_rate": 3.695956269045509e-08, "loss": 0.8855, "step": 5059 }, { "epoch": 0.973498148237218, "grad_norm": 1.3423508473193793, "learning_rate": 3.642618185215163e-08, "loss": 0.8916, "step": 5060 }, { "epoch": 0.9736905391756048, "grad_norm": 1.3756650790512708, "learning_rate": 3.589667075110992e-08, "loss": 0.914, "step": 5061 }, { "epoch": 0.9738829301139916, "grad_norm": 1.5198668684340144, "learning_rate": 3.537102959298322e-08, "loss": 0.9811, "step": 5062 }, { "epoch": 0.9740753210523785, "grad_norm": 0.7237943868948775, "learning_rate": 3.4849258581928224e-08, "loss": 0.8336, "step": 5063 }, { "epoch": 0.9742677119907652, "grad_norm": 0.7064868448104464, "learning_rate": 3.4331357920591726e-08, "loss": 0.9094, "step": 5064 }, { "epoch": 0.9744601029291521, "grad_norm": 1.2963058065834467, "learning_rate": 3.3817327810121706e-08, "loss": 0.8816, "step": 5065 }, { "epoch": 0.9746524938675388, "grad_norm": 1.3215275780404938, "learning_rate": 3.3307168450160685e-08, "loss": 0.8726, "step": 5066 }, { "epoch": 0.9748448848059257, "grad_norm": 1.446534059301206, "learning_rate": 3.280088003885018e-08, "loss": 0.9626, "step": 5067 }, { "epoch": 0.9750372757443124, "grad_norm": 1.4988712358191338, "learning_rate": 3.229846277282511e-08, "loss": 0.8688, "step": 5068 }, { "epoch": 0.9752296666826993, "grad_norm": 1.4752109941121339, "learning_rate": 3.1799916847220505e-08, "loss": 0.9621, "step": 5069 }, { "epoch": 0.975422057621086, "grad_norm": 1.3919101349540584, "learning_rate": 3.130524245566369e-08, "loss": 0.9761, "step": 5070 }, { "epoch": 0.9756144485594729, "grad_norm": 1.358558948740083, "learning_rate": 3.0814439790280984e-08, "loss": 0.9809, "step": 5071 }, { "epoch": 0.9758068394978596, "grad_norm": 1.5606105310097833, "learning_rate": 3.032750904169546e-08, "loss": 0.9212, "step": 5072 }, { "epoch": 0.9759992304362465, "grad_norm": 1.44135781371916, "learning_rate": 2.9844450399024725e-08, "loss": 0.9517, "step": 5073 }, { "epoch": 0.9761916213746332, "grad_norm": 1.5228981648348492, "learning_rate": 2.936526404988427e-08, "loss": 0.9643, "step": 5074 }, { "epoch": 0.9763840123130201, "grad_norm": 1.4912067708304284, "learning_rate": 2.8889950180382985e-08, "loss": 0.9169, "step": 5075 }, { "epoch": 0.9765764032514068, "grad_norm": 1.3498005356665428, "learning_rate": 2.8418508975127655e-08, "loss": 0.9087, "step": 5076 }, { "epoch": 0.9767687941897937, "grad_norm": 1.5270859994903687, "learning_rate": 2.7950940617221812e-08, "loss": 1.0112, "step": 5077 }, { "epoch": 0.9769611851281804, "grad_norm": 1.3421827991304958, "learning_rate": 2.7487245288261298e-08, "loss": 0.9739, "step": 5078 }, { "epoch": 0.9771535760665673, "grad_norm": 1.4119167640619883, "learning_rate": 2.702742316834206e-08, "loss": 0.9245, "step": 5079 }, { "epoch": 0.977345967004954, "grad_norm": 1.3939365839270192, "learning_rate": 2.657147443605457e-08, "loss": 0.9417, "step": 5080 }, { "epoch": 0.9775383579433409, "grad_norm": 1.473370315728262, "learning_rate": 2.6119399268480505e-08, "loss": 0.9985, "step": 5081 }, { "epoch": 0.9777307488817276, "grad_norm": 0.6747270802951717, "learning_rate": 2.5671197841203867e-08, "loss": 0.8506, "step": 5082 }, { "epoch": 0.9779231398201145, "grad_norm": 1.583720577232094, "learning_rate": 2.5226870328299845e-08, "loss": 0.9604, "step": 5083 }, { "epoch": 0.9781155307585013, "grad_norm": 1.423682981760227, "learning_rate": 2.4786416902339293e-08, "loss": 0.9363, "step": 5084 }, { "epoch": 0.9783079216968881, "grad_norm": 1.4690156320111105, "learning_rate": 2.4349837734390923e-08, "loss": 0.968, "step": 5085 }, { "epoch": 0.9785003126352749, "grad_norm": 1.406677966056924, "learning_rate": 2.3917132994016878e-08, "loss": 0.9391, "step": 5086 }, { "epoch": 0.9786927035736617, "grad_norm": 1.3607519120737293, "learning_rate": 2.3488302849272726e-08, "loss": 0.9138, "step": 5087 }, { "epoch": 0.9788850945120485, "grad_norm": 1.4318593840554186, "learning_rate": 2.3063347466713014e-08, "loss": 0.9208, "step": 5088 }, { "epoch": 0.9790774854504353, "grad_norm": 1.4100792388643044, "learning_rate": 2.264226701138461e-08, "loss": 0.9735, "step": 5089 }, { "epoch": 0.9792698763888221, "grad_norm": 1.509982936636022, "learning_rate": 2.2225061646830025e-08, "loss": 0.994, "step": 5090 }, { "epoch": 0.9794622673272089, "grad_norm": 1.4456807853430587, "learning_rate": 2.181173153508853e-08, "loss": 0.9913, "step": 5091 }, { "epoch": 0.9796546582655957, "grad_norm": 1.4323565213772291, "learning_rate": 2.1402276836691715e-08, "loss": 0.9937, "step": 5092 }, { "epoch": 0.9798470492039825, "grad_norm": 1.4755258807147673, "learning_rate": 2.0996697710666812e-08, "loss": 0.9812, "step": 5093 }, { "epoch": 0.9800394401423693, "grad_norm": 1.3498305754676214, "learning_rate": 2.059499431453671e-08, "loss": 0.9579, "step": 5094 }, { "epoch": 0.9802318310807561, "grad_norm": 1.3153992771136838, "learning_rate": 2.0197166804317713e-08, "loss": 0.9996, "step": 5095 }, { "epoch": 0.9804242220191429, "grad_norm": 1.3389394158003227, "learning_rate": 1.9803215334522895e-08, "loss": 0.9363, "step": 5096 }, { "epoch": 0.9806166129575297, "grad_norm": 1.3586517995271397, "learning_rate": 1.9413140058156533e-08, "loss": 0.9144, "step": 5097 }, { "epoch": 0.9808090038959165, "grad_norm": 1.3087566742789416, "learning_rate": 1.9026941126721876e-08, "loss": 0.8157, "step": 5098 }, { "epoch": 0.9810013948343033, "grad_norm": 1.502969281740921, "learning_rate": 1.864461869021117e-08, "loss": 0.9097, "step": 5099 }, { "epoch": 0.9811937857726901, "grad_norm": 1.3817950531806193, "learning_rate": 1.826617289711563e-08, "loss": 0.9185, "step": 5100 }, { "epoch": 0.9813861767110769, "grad_norm": 1.4147569474134172, "learning_rate": 1.7891603894418797e-08, "loss": 0.9316, "step": 5101 }, { "epoch": 0.9815785676494637, "grad_norm": 1.4441839412658861, "learning_rate": 1.7520911827598742e-08, "loss": 0.98, "step": 5102 }, { "epoch": 0.9817709585878505, "grad_norm": 1.3792087894527232, "learning_rate": 1.7154096840629186e-08, "loss": 0.9455, "step": 5103 }, { "epoch": 0.9819633495262373, "grad_norm": 1.3852277696574224, "learning_rate": 1.679115907597617e-08, "loss": 0.9552, "step": 5104 }, { "epoch": 0.9821557404646241, "grad_norm": 1.3398612617401986, "learning_rate": 1.6432098674600272e-08, "loss": 1.0054, "step": 5105 }, { "epoch": 0.9823481314030109, "grad_norm": 1.4258119538558, "learning_rate": 1.6076915775956604e-08, "loss": 0.9345, "step": 5106 }, { "epoch": 0.9825405223413978, "grad_norm": 1.432237550153133, "learning_rate": 1.5725610517994816e-08, "loss": 0.852, "step": 5107 }, { "epoch": 0.9827329132797845, "grad_norm": 1.3013043538726496, "learning_rate": 1.537818303715688e-08, "loss": 0.9394, "step": 5108 }, { "epoch": 0.9829253042181714, "grad_norm": 1.3631560598472774, "learning_rate": 1.50346334683793e-08, "loss": 0.99, "step": 5109 }, { "epoch": 0.9831176951565581, "grad_norm": 1.337969508628868, "learning_rate": 1.4694961945093122e-08, "loss": 0.9619, "step": 5110 }, { "epoch": 0.983310086094945, "grad_norm": 1.471848639095627, "learning_rate": 1.4359168599223926e-08, "loss": 0.9754, "step": 5111 }, { "epoch": 0.9835024770333317, "grad_norm": 1.4238736044818043, "learning_rate": 1.4027253561188502e-08, "loss": 0.9398, "step": 5112 }, { "epoch": 0.9836948679717186, "grad_norm": 1.3849557064396467, "learning_rate": 1.3699216959899287e-08, "loss": 0.9066, "step": 5113 }, { "epoch": 0.9838872589101053, "grad_norm": 1.496679676492152, "learning_rate": 1.337505892276325e-08, "loss": 0.8544, "step": 5114 }, { "epoch": 0.9840796498484922, "grad_norm": 1.3073102023118044, "learning_rate": 1.3054779575677468e-08, "loss": 0.9038, "step": 5115 }, { "epoch": 0.9842720407868789, "grad_norm": 1.3946550245794, "learning_rate": 1.2738379043035765e-08, "loss": 1.0039, "step": 5116 }, { "epoch": 0.9844644317252658, "grad_norm": 1.3587283668249097, "learning_rate": 1.2425857447725397e-08, "loss": 0.9756, "step": 5117 }, { "epoch": 0.9846568226636525, "grad_norm": 0.7006649800176623, "learning_rate": 1.211721491112372e-08, "loss": 0.8366, "step": 5118 }, { "epoch": 0.9848492136020394, "grad_norm": 1.550650555779471, "learning_rate": 1.1812451553107063e-08, "loss": 0.9683, "step": 5119 }, { "epoch": 0.9850416045404261, "grad_norm": 1.4064981643438033, "learning_rate": 1.1511567492038522e-08, "loss": 0.951, "step": 5120 }, { "epoch": 0.985233995478813, "grad_norm": 1.3633144671189943, "learning_rate": 1.1214562844781285e-08, "loss": 0.9238, "step": 5121 }, { "epoch": 0.9854263864171997, "grad_norm": 1.3267520075634271, "learning_rate": 1.0921437726686413e-08, "loss": 0.8737, "step": 5122 }, { "epoch": 0.9856187773555866, "grad_norm": 1.3155858050173166, "learning_rate": 1.0632192251601726e-08, "loss": 0.8643, "step": 5123 }, { "epoch": 0.9858111682939733, "grad_norm": 1.4039239750682921, "learning_rate": 1.0346826531865139e-08, "loss": 0.9696, "step": 5124 }, { "epoch": 0.9860035592323602, "grad_norm": 1.352962134216269, "learning_rate": 1.006534067831022e-08, "loss": 0.9556, "step": 5125 }, { "epoch": 0.9861959501707469, "grad_norm": 1.3752042721058844, "learning_rate": 9.78773480026396e-09, "loss": 0.9219, "step": 5126 }, { "epoch": 0.9863883411091338, "grad_norm": 1.3819075556197686, "learning_rate": 9.514009005543445e-09, "loss": 0.9823, "step": 5127 }, { "epoch": 0.9865807320475206, "grad_norm": 0.6944880691310507, "learning_rate": 9.244163400462525e-09, "loss": 0.8581, "step": 5128 }, { "epoch": 0.9867731229859074, "grad_norm": 1.2815469087516898, "learning_rate": 8.97819808982403e-09, "loss": 0.9009, "step": 5129 }, { "epoch": 0.9869655139242942, "grad_norm": 1.5025201110127613, "learning_rate": 8.716113176927554e-09, "loss": 1.0048, "step": 5130 }, { "epoch": 0.987157904862681, "grad_norm": 1.445979437677411, "learning_rate": 8.457908763562783e-09, "loss": 0.9488, "step": 5131 }, { "epoch": 0.9873502958010678, "grad_norm": 1.4735653195137435, "learning_rate": 8.203584950013942e-09, "loss": 0.9041, "step": 5132 }, { "epoch": 0.9875426867394546, "grad_norm": 1.4388458986062564, "learning_rate": 7.953141835057576e-09, "loss": 1.0052, "step": 5133 }, { "epoch": 0.9877350776778414, "grad_norm": 1.4100215358999102, "learning_rate": 7.706579515962542e-09, "loss": 0.9554, "step": 5134 }, { "epoch": 0.9879274686162282, "grad_norm": 0.7288352472109959, "learning_rate": 7.463898088490019e-09, "loss": 0.8466, "step": 5135 }, { "epoch": 0.988119859554615, "grad_norm": 1.364711978484702, "learning_rate": 7.225097646895718e-09, "loss": 0.949, "step": 5136 }, { "epoch": 0.9883122504930018, "grad_norm": 1.3145910394429992, "learning_rate": 6.990178283927673e-09, "loss": 0.9951, "step": 5137 }, { "epoch": 0.9885046414313886, "grad_norm": 1.4022530524564618, "learning_rate": 6.759140090824012e-09, "loss": 0.9477, "step": 5138 }, { "epoch": 0.9886970323697754, "grad_norm": 1.483539486906842, "learning_rate": 6.531983157318511e-09, "loss": 0.9957, "step": 5139 }, { "epoch": 0.9888894233081622, "grad_norm": 1.5511547981203528, "learning_rate": 6.308707571636152e-09, "loss": 1.0689, "step": 5140 }, { "epoch": 0.989081814246549, "grad_norm": 1.427054842152277, "learning_rate": 6.089313420494236e-09, "loss": 0.9319, "step": 5141 }, { "epoch": 0.9892742051849358, "grad_norm": 1.4452250264714523, "learning_rate": 5.87380078910349e-09, "loss": 0.947, "step": 5142 }, { "epoch": 0.9894665961233226, "grad_norm": 1.3713999394770573, "learning_rate": 5.662169761165848e-09, "loss": 0.984, "step": 5143 }, { "epoch": 0.9896589870617094, "grad_norm": 1.4616598186038259, "learning_rate": 5.4544204188777815e-09, "loss": 0.9746, "step": 5144 }, { "epoch": 0.9898513780000961, "grad_norm": 1.4211883584484213, "learning_rate": 5.25055284292475e-09, "loss": 0.9468, "step": 5145 }, { "epoch": 0.990043768938483, "grad_norm": 0.7690030082440504, "learning_rate": 5.050567112487858e-09, "loss": 0.8563, "step": 5146 }, { "epoch": 0.9902361598768697, "grad_norm": 1.44708745036852, "learning_rate": 4.85446330523831e-09, "loss": 1.0196, "step": 5147 }, { "epoch": 0.9904285508152566, "grad_norm": 1.5061114903537625, "learning_rate": 4.662241497341846e-09, "loss": 0.9692, "step": 5148 }, { "epoch": 0.9906209417536435, "grad_norm": 1.4040094532309244, "learning_rate": 4.473901763454302e-09, "loss": 0.9276, "step": 5149 }, { "epoch": 0.9908133326920302, "grad_norm": 1.5030675474603743, "learning_rate": 4.289444176724944e-09, "loss": 0.9421, "step": 5150 }, { "epoch": 0.9910057236304171, "grad_norm": 1.5537500554269024, "learning_rate": 4.108868808794242e-09, "loss": 0.9372, "step": 5151 }, { "epoch": 0.9911981145688038, "grad_norm": 1.3102959757194124, "learning_rate": 3.932175729797205e-09, "loss": 0.8971, "step": 5152 }, { "epoch": 0.9913905055071907, "grad_norm": 1.380068288832521, "learning_rate": 3.759365008357829e-09, "loss": 0.9474, "step": 5153 }, { "epoch": 0.9915828964455774, "grad_norm": 1.3287463935580646, "learning_rate": 3.590436711594647e-09, "loss": 1.0035, "step": 5154 }, { "epoch": 0.9917752873839643, "grad_norm": 1.4346067484009708, "learning_rate": 3.4253909051173985e-09, "loss": 0.9804, "step": 5155 }, { "epoch": 0.991967678322351, "grad_norm": 1.3428881330475226, "learning_rate": 3.2642276530281404e-09, "loss": 0.9185, "step": 5156 }, { "epoch": 0.9921600692607379, "grad_norm": 1.3512450331933994, "learning_rate": 3.1069470179201365e-09, "loss": 0.9653, "step": 5157 }, { "epoch": 0.9923524601991246, "grad_norm": 1.426567713826575, "learning_rate": 2.9535490608789687e-09, "loss": 0.8986, "step": 5158 }, { "epoch": 0.9925448511375115, "grad_norm": 1.4029044211303594, "learning_rate": 2.8040338414847545e-09, "loss": 0.9387, "step": 5159 }, { "epoch": 0.9927372420758982, "grad_norm": 1.4103845754300433, "learning_rate": 2.6584014178054894e-09, "loss": 0.952, "step": 5160 }, { "epoch": 0.992929633014285, "grad_norm": 1.435615036562948, "learning_rate": 2.516651846403706e-09, "loss": 0.9557, "step": 5161 }, { "epoch": 0.9931220239526718, "grad_norm": 1.3873848026265683, "learning_rate": 2.378785182333143e-09, "loss": 0.9836, "step": 5162 }, { "epoch": 0.9933144148910587, "grad_norm": 1.4945784552929717, "learning_rate": 2.2448014791398574e-09, "loss": 1.034, "step": 5163 }, { "epoch": 0.9935068058294454, "grad_norm": 1.351954231354965, "learning_rate": 2.1147007888622227e-09, "loss": 0.966, "step": 5164 }, { "epoch": 0.9936991967678322, "grad_norm": 1.486251343989117, "learning_rate": 1.9884831620287094e-09, "loss": 1.034, "step": 5165 }, { "epoch": 0.993891587706219, "grad_norm": 1.2911985317888388, "learning_rate": 1.8661486476612144e-09, "loss": 0.9392, "step": 5166 }, { "epoch": 0.9940839786446058, "grad_norm": 1.388622824436206, "learning_rate": 1.747697293272843e-09, "loss": 0.9271, "step": 5167 }, { "epoch": 0.9942763695829926, "grad_norm": 1.4602613090739442, "learning_rate": 1.633129144870127e-09, "loss": 0.9264, "step": 5168 }, { "epoch": 0.9944687605213794, "grad_norm": 1.2460932022873696, "learning_rate": 1.5224442469474743e-09, "loss": 0.9426, "step": 5169 }, { "epoch": 0.9946611514597662, "grad_norm": 1.3684449342568972, "learning_rate": 1.4156426424960513e-09, "loss": 0.9496, "step": 5170 }, { "epoch": 0.994853542398153, "grad_norm": 1.4951204592310565, "learning_rate": 1.3127243729949003e-09, "loss": 0.9466, "step": 5171 }, { "epoch": 0.9950459333365399, "grad_norm": 1.4533484772555003, "learning_rate": 1.2136894784176011e-09, "loss": 0.9926, "step": 5172 }, { "epoch": 0.9952383242749266, "grad_norm": 1.3801837617594062, "learning_rate": 1.1185379972256105e-09, "loss": 0.995, "step": 5173 }, { "epoch": 0.9954307152133135, "grad_norm": 1.4994007031148089, "learning_rate": 1.0272699663782525e-09, "loss": 0.9558, "step": 5174 }, { "epoch": 0.9956231061517002, "grad_norm": 1.4768801365996402, "learning_rate": 9.398854213193976e-10, "loss": 0.9664, "step": 5175 }, { "epoch": 0.9958154970900871, "grad_norm": 1.4493573810938807, "learning_rate": 8.563843959907836e-10, "loss": 0.901, "step": 5176 }, { "epoch": 0.9960078880284738, "grad_norm": 1.4363192234998523, "learning_rate": 7.767669228231356e-10, "loss": 0.9207, "step": 5177 }, { "epoch": 0.9962002789668607, "grad_norm": 1.2786370013905628, "learning_rate": 7.010330327372749e-10, "loss": 0.9407, "step": 5178 }, { "epoch": 0.9963926699052474, "grad_norm": 1.2707720135625795, "learning_rate": 6.291827551474505e-10, "loss": 0.8378, "step": 5179 }, { "epoch": 0.9965850608436343, "grad_norm": 1.6216487273072284, "learning_rate": 5.612161179613385e-10, "loss": 1.0265, "step": 5180 }, { "epoch": 0.996777451782021, "grad_norm": 1.3220134268064936, "learning_rate": 4.971331475756013e-10, "loss": 0.9619, "step": 5181 }, { "epoch": 0.9969698427204079, "grad_norm": 1.4878582720778657, "learning_rate": 4.369338688781088e-10, "loss": 0.8947, "step": 5182 }, { "epoch": 0.9971622336587946, "grad_norm": 1.4763929179527515, "learning_rate": 3.8061830525126797e-10, "loss": 0.9259, "step": 5183 }, { "epoch": 0.9973546245971815, "grad_norm": 1.3609623944295164, "learning_rate": 3.2818647856647234e-10, "loss": 0.9817, "step": 5184 }, { "epoch": 0.9975470155355682, "grad_norm": 1.2967767510083112, "learning_rate": 2.796384091885429e-10, "loss": 0.9311, "step": 5185 }, { "epoch": 0.9977394064739551, "grad_norm": 1.2245276774446305, "learning_rate": 2.3497411597128705e-10, "loss": 0.8707, "step": 5186 }, { "epoch": 0.9979317974123418, "grad_norm": 1.3991475275472827, "learning_rate": 1.9419361626416e-10, "loss": 0.9407, "step": 5187 }, { "epoch": 0.9981241883507287, "grad_norm": 1.2949318135256185, "learning_rate": 1.5729692590338296e-10, "loss": 0.9973, "step": 5188 }, { "epoch": 0.9983165792891154, "grad_norm": 1.379494475495635, "learning_rate": 1.2428405921971476e-10, "loss": 1.0525, "step": 5189 }, { "epoch": 0.9985089702275023, "grad_norm": 1.3364146817741347, "learning_rate": 9.515502903734153e-11, "loss": 0.8836, "step": 5190 }, { "epoch": 0.998701361165889, "grad_norm": 1.4030434227220674, "learning_rate": 6.990984666610523e-11, "loss": 0.9635, "step": 5191 }, { "epoch": 0.9988937521042759, "grad_norm": 1.4231496382608066, "learning_rate": 4.854852191371606e-11, "loss": 1.0192, "step": 5192 }, { "epoch": 0.9990861430426627, "grad_norm": 1.530131259436536, "learning_rate": 3.1071063075760466e-11, "loss": 1.0372, "step": 5193 }, { "epoch": 0.9992785339810495, "grad_norm": 1.4092999456094846, "learning_rate": 1.7477476940142013e-11, "loss": 1.0007, "step": 5194 }, { "epoch": 0.9994709249194363, "grad_norm": 1.5526179356534073, "learning_rate": 7.767768785971185e-12, "loss": 0.8964, "step": 5195 }, { "epoch": 0.9996633158578231, "grad_norm": 1.3825254506583187, "learning_rate": 1.9419423846755993e-12, "loss": 0.9312, "step": 5196 }, { "epoch": 0.9998557067962099, "grad_norm": 1.7878963056040313, "learning_rate": 0.0, "loss": 0.7753, "step": 5197 }, { "epoch": 0.9998557067962099, "step": 5197, "total_flos": 4043527175667712.0, "train_loss": 0.9949393240401257, "train_runtime": 49260.169, "train_samples_per_second": 13.506, "train_steps_per_second": 0.106 } ], "logging_steps": 1.0, "max_steps": 5197, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 50000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 4043527175667712.0, "train_batch_size": 8, "trial_name": null, "trial_params": null }