{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 500, "global_step": 3062, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0, "grad_norm": 8.492539248778657, "learning_rate": 3.2608695652173915e-07, "loss": 0.9859, "step": 1 }, { "epoch": 0.0, "grad_norm": 11.476061044441659, "learning_rate": 6.521739130434783e-07, "loss": 1.4288, "step": 2 }, { "epoch": 0.0, "grad_norm": 8.67232394820012, "learning_rate": 9.782608695652173e-07, "loss": 1.187, "step": 3 }, { "epoch": 0.0, "grad_norm": 8.558007162286211, "learning_rate": 1.3043478260869566e-06, "loss": 1.1621, "step": 4 }, { "epoch": 0.0, "grad_norm": 12.732280458685807, "learning_rate": 1.6304347826086957e-06, "loss": 1.404, "step": 5 }, { "epoch": 0.0, "grad_norm": 7.854455706716011, "learning_rate": 1.9565217391304346e-06, "loss": 1.1296, "step": 6 }, { "epoch": 0.0, "grad_norm": 5.6736617556468145, "learning_rate": 2.282608695652174e-06, "loss": 1.0772, "step": 7 }, { "epoch": 0.0, "grad_norm": 4.664147317983734, "learning_rate": 2.6086956521739132e-06, "loss": 1.0553, "step": 8 }, { "epoch": 0.0, "grad_norm": 2.04869163174705, "learning_rate": 2.9347826086956523e-06, "loss": 1.1684, "step": 9 }, { "epoch": 0.0, "grad_norm": 2.2287707412803015, "learning_rate": 3.2608695652173914e-06, "loss": 0.9392, "step": 10 }, { "epoch": 0.0, "grad_norm": 2.094345154741896, "learning_rate": 3.5869565217391305e-06, "loss": 0.9431, "step": 11 }, { "epoch": 0.0, "grad_norm": 2.0827991804615866, "learning_rate": 3.913043478260869e-06, "loss": 0.8942, "step": 12 }, { "epoch": 0.0, "grad_norm": 1.9506743289523296, "learning_rate": 4.239130434782608e-06, "loss": 1.1225, "step": 13 }, { "epoch": 0.0, "grad_norm": 2.0894108064137504, "learning_rate": 4.565217391304348e-06, "loss": 0.9203, "step": 14 }, { "epoch": 0.0, "grad_norm": 1.6968063615079527, "learning_rate": 4.891304347826087e-06, "loss": 0.8766, "step": 15 }, { "epoch": 0.01, "grad_norm": 1.1839495699712022, "learning_rate": 5.2173913043478265e-06, "loss": 1.0905, "step": 16 }, { "epoch": 0.01, "grad_norm": 1.7116620341031912, "learning_rate": 5.543478260869565e-06, "loss": 0.8675, "step": 17 }, { "epoch": 0.01, "grad_norm": 1.567685788018572, "learning_rate": 5.869565217391305e-06, "loss": 0.8395, "step": 18 }, { "epoch": 0.01, "grad_norm": 1.2173638836782414, "learning_rate": 6.195652173913044e-06, "loss": 0.8614, "step": 19 }, { "epoch": 0.01, "grad_norm": 0.9671769966950868, "learning_rate": 6.521739130434783e-06, "loss": 1.0723, "step": 20 }, { "epoch": 0.01, "grad_norm": 1.081558714522479, "learning_rate": 6.847826086956521e-06, "loss": 0.8369, "step": 21 }, { "epoch": 0.01, "grad_norm": 0.9861804453391124, "learning_rate": 7.173913043478261e-06, "loss": 0.8445, "step": 22 }, { "epoch": 0.01, "grad_norm": 0.8085936417731859, "learning_rate": 7.5e-06, "loss": 1.0399, "step": 23 }, { "epoch": 0.01, "grad_norm": 0.9485784523636834, "learning_rate": 7.826086956521738e-06, "loss": 0.8112, "step": 24 }, { "epoch": 0.01, "grad_norm": 0.8238602295413909, "learning_rate": 8.152173913043478e-06, "loss": 0.8136, "step": 25 }, { "epoch": 0.01, "grad_norm": 0.656547621699397, "learning_rate": 8.478260869565217e-06, "loss": 1.0657, "step": 26 }, { "epoch": 0.01, "grad_norm": 0.754228591517229, "learning_rate": 8.804347826086957e-06, "loss": 0.8014, "step": 27 }, { "epoch": 0.01, "grad_norm": 0.7391721166511256, "learning_rate": 9.130434782608697e-06, "loss": 0.8332, "step": 28 }, { "epoch": 0.01, "grad_norm": 0.7134964251116312, "learning_rate": 9.456521739130436e-06, "loss": 0.7815, "step": 29 }, { "epoch": 0.01, "grad_norm": 0.6647718870624, "learning_rate": 9.782608695652175e-06, "loss": 0.7745, "step": 30 }, { "epoch": 0.01, "grad_norm": 0.6007452129077422, "learning_rate": 1.0108695652173914e-05, "loss": 1.0519, "step": 31 }, { "epoch": 0.01, "grad_norm": 0.6441207134242788, "learning_rate": 1.0434782608695653e-05, "loss": 0.7888, "step": 32 }, { "epoch": 0.01, "grad_norm": 0.6322175832298551, "learning_rate": 1.076086956521739e-05, "loss": 0.7743, "step": 33 }, { "epoch": 0.01, "grad_norm": 0.5340104302895209, "learning_rate": 1.108695652173913e-05, "loss": 1.0035, "step": 34 }, { "epoch": 0.01, "grad_norm": 0.6939518098634568, "learning_rate": 1.141304347826087e-05, "loss": 0.7646, "step": 35 }, { "epoch": 0.01, "grad_norm": 0.5599980790610366, "learning_rate": 1.173913043478261e-05, "loss": 0.7541, "step": 36 }, { "epoch": 0.01, "grad_norm": 0.5719145275820134, "learning_rate": 1.2065217391304348e-05, "loss": 1.042, "step": 37 }, { "epoch": 0.01, "grad_norm": 0.6266709198402662, "learning_rate": 1.2391304347826088e-05, "loss": 0.7693, "step": 38 }, { "epoch": 0.01, "grad_norm": 0.5665013077102612, "learning_rate": 1.2717391304347827e-05, "loss": 0.799, "step": 39 }, { "epoch": 0.01, "grad_norm": 0.5627580504575114, "learning_rate": 1.3043478260869566e-05, "loss": 0.7446, "step": 40 }, { "epoch": 0.01, "grad_norm": 0.5215423912657596, "learning_rate": 1.3369565217391305e-05, "loss": 0.7591, "step": 41 }, { "epoch": 0.01, "grad_norm": 0.5030914140580001, "learning_rate": 1.3695652173913042e-05, "loss": 1.0031, "step": 42 }, { "epoch": 0.01, "grad_norm": 0.5548865582115394, "learning_rate": 1.4021739130434783e-05, "loss": 0.768, "step": 43 }, { "epoch": 0.01, "grad_norm": 0.4835550137431244, "learning_rate": 1.4347826086956522e-05, "loss": 0.7552, "step": 44 }, { "epoch": 0.01, "grad_norm": 0.5070824550913681, "learning_rate": 1.4673913043478261e-05, "loss": 0.9716, "step": 45 }, { "epoch": 0.02, "grad_norm": 0.5527511484060249, "learning_rate": 1.5e-05, "loss": 0.7572, "step": 46 }, { "epoch": 0.02, "grad_norm": 0.496420345158628, "learning_rate": 1.532608695652174e-05, "loss": 0.7341, "step": 47 }, { "epoch": 0.02, "grad_norm": 0.5174998563797033, "learning_rate": 1.5652173913043477e-05, "loss": 1.0142, "step": 48 }, { "epoch": 0.02, "grad_norm": 0.5099482858678012, "learning_rate": 1.597826086956522e-05, "loss": 0.7606, "step": 49 }, { "epoch": 0.02, "grad_norm": 0.5190872254267888, "learning_rate": 1.6304347826086955e-05, "loss": 0.809, "step": 50 }, { "epoch": 0.02, "grad_norm": 0.4769707635101255, "learning_rate": 1.6630434782608698e-05, "loss": 0.733, "step": 51 }, { "epoch": 0.02, "grad_norm": 0.5464905004049824, "learning_rate": 1.6956521739130433e-05, "loss": 0.9779, "step": 52 }, { "epoch": 0.02, "grad_norm": 0.49098992315156725, "learning_rate": 1.7282608695652176e-05, "loss": 0.7439, "step": 53 }, { "epoch": 0.02, "grad_norm": 0.5712062790584733, "learning_rate": 1.7608695652173915e-05, "loss": 0.7448, "step": 54 }, { "epoch": 0.02, "grad_norm": 0.5203639378744831, "learning_rate": 1.7934782608695654e-05, "loss": 0.744, "step": 55 }, { "epoch": 0.02, "grad_norm": 0.5026371863360279, "learning_rate": 1.8260869565217393e-05, "loss": 0.9843, "step": 56 }, { "epoch": 0.02, "grad_norm": 0.5177452969972849, "learning_rate": 1.858695652173913e-05, "loss": 0.7525, "step": 57 }, { "epoch": 0.02, "grad_norm": 0.5628678527942377, "learning_rate": 1.891304347826087e-05, "loss": 0.7437, "step": 58 }, { "epoch": 0.02, "grad_norm": 0.472698825198783, "learning_rate": 1.9239130434782607e-05, "loss": 0.991, "step": 59 }, { "epoch": 0.02, "grad_norm": 0.544264417232749, "learning_rate": 1.956521739130435e-05, "loss": 0.7235, "step": 60 }, { "epoch": 0.02, "grad_norm": 0.4549792349205737, "learning_rate": 1.9891304347826085e-05, "loss": 0.7411, "step": 61 }, { "epoch": 0.02, "grad_norm": 0.5072210414125832, "learning_rate": 2.0217391304347828e-05, "loss": 0.7172, "step": 62 }, { "epoch": 0.02, "grad_norm": 0.5011680342824683, "learning_rate": 2.0543478260869567e-05, "loss": 0.9834, "step": 63 }, { "epoch": 0.02, "grad_norm": 0.5563394503360095, "learning_rate": 2.0869565217391306e-05, "loss": 0.7213, "step": 64 }, { "epoch": 0.02, "grad_norm": 0.595575455565999, "learning_rate": 2.1195652173913045e-05, "loss": 0.7104, "step": 65 }, { "epoch": 0.02, "grad_norm": 0.553502531612874, "learning_rate": 2.152173913043478e-05, "loss": 0.776, "step": 66 }, { "epoch": 0.02, "grad_norm": 0.5296639798328564, "learning_rate": 2.1847826086956523e-05, "loss": 0.9165, "step": 67 }, { "epoch": 0.02, "grad_norm": 0.5510930106218987, "learning_rate": 2.217391304347826e-05, "loss": 0.7131, "step": 68 }, { "epoch": 0.02, "grad_norm": 0.5276717262461303, "learning_rate": 2.25e-05, "loss": 1.0245, "step": 69 }, { "epoch": 0.02, "grad_norm": 0.5768687719897174, "learning_rate": 2.282608695652174e-05, "loss": 0.723, "step": 70 }, { "epoch": 0.02, "grad_norm": 0.663184713343694, "learning_rate": 2.315217391304348e-05, "loss": 0.7089, "step": 71 }, { "epoch": 0.02, "grad_norm": 0.6904319421229019, "learning_rate": 2.347826086956522e-05, "loss": 0.7492, "step": 72 }, { "epoch": 0.02, "grad_norm": 0.5336944383237118, "learning_rate": 2.3804347826086958e-05, "loss": 0.7047, "step": 73 }, { "epoch": 0.02, "grad_norm": 0.6274981293193478, "learning_rate": 2.4130434782608697e-05, "loss": 0.9708, "step": 74 }, { "epoch": 0.02, "grad_norm": 0.53385596166341, "learning_rate": 2.4456521739130433e-05, "loss": 0.7115, "step": 75 }, { "epoch": 0.02, "grad_norm": 0.6038077661317958, "learning_rate": 2.4782608695652175e-05, "loss": 0.728, "step": 76 }, { "epoch": 0.03, "grad_norm": 0.8274809784072898, "learning_rate": 2.5108695652173914e-05, "loss": 0.785, "step": 77 }, { "epoch": 0.03, "grad_norm": 0.5770570624918475, "learning_rate": 2.5434782608695653e-05, "loss": 0.9664, "step": 78 }, { "epoch": 0.03, "grad_norm": 0.5927452250661398, "learning_rate": 2.5760869565217392e-05, "loss": 0.7325, "step": 79 }, { "epoch": 0.03, "grad_norm": 0.8527087087780119, "learning_rate": 2.608695652173913e-05, "loss": 0.7039, "step": 80 }, { "epoch": 0.03, "grad_norm": 0.9414203384785111, "learning_rate": 2.641304347826087e-05, "loss": 0.7283, "step": 81 }, { "epoch": 0.03, "grad_norm": 0.7899927566452062, "learning_rate": 2.673913043478261e-05, "loss": 0.943, "step": 82 }, { "epoch": 0.03, "grad_norm": 0.683110430990533, "learning_rate": 2.706521739130435e-05, "loss": 0.7068, "step": 83 }, { "epoch": 0.03, "grad_norm": 1.1162193818935942, "learning_rate": 2.7391304347826085e-05, "loss": 0.7153, "step": 84 }, { "epoch": 0.03, "grad_norm": 0.9467487928003426, "learning_rate": 2.7717391304347827e-05, "loss": 0.9658, "step": 85 }, { "epoch": 0.03, "grad_norm": 0.7221254820396366, "learning_rate": 2.8043478260869566e-05, "loss": 0.7215, "step": 86 }, { "epoch": 0.03, "grad_norm": 1.3271671286815554, "learning_rate": 2.8369565217391305e-05, "loss": 0.7067, "step": 87 }, { "epoch": 0.03, "grad_norm": 1.0334914152752352, "learning_rate": 2.8695652173913044e-05, "loss": 0.7564, "step": 88 }, { "epoch": 0.03, "grad_norm": 0.7270508055548834, "learning_rate": 2.9021739130434783e-05, "loss": 0.9762, "step": 89 }, { "epoch": 0.03, "grad_norm": 1.0051483262599508, "learning_rate": 2.9347826086956523e-05, "loss": 0.6967, "step": 90 }, { "epoch": 0.03, "grad_norm": 1.0541472044963887, "learning_rate": 2.967391304347826e-05, "loss": 0.7038, "step": 91 }, { "epoch": 0.03, "grad_norm": 0.694323130147114, "learning_rate": 3e-05, "loss": 0.9751, "step": 92 }, { "epoch": 0.03, "grad_norm": 0.9075741447999567, "learning_rate": 2.9999991608336328e-05, "loss": 0.7213, "step": 93 }, { "epoch": 0.03, "grad_norm": 1.2008198951909994, "learning_rate": 2.999996643335469e-05, "loss": 0.7151, "step": 94 }, { "epoch": 0.03, "grad_norm": 0.7858031733006479, "learning_rate": 2.9999924475083266e-05, "loss": 0.9989, "step": 95 }, { "epoch": 0.03, "grad_norm": 1.0066900206043783, "learning_rate": 2.9999865733568992e-05, "loss": 0.7211, "step": 96 }, { "epoch": 0.03, "grad_norm": 0.855480653208801, "learning_rate": 2.9999790208877606e-05, "loss": 0.7179, "step": 97 }, { "epoch": 0.03, "grad_norm": 0.7101193551794752, "learning_rate": 2.9999697901093597e-05, "loss": 0.9925, "step": 98 }, { "epoch": 0.03, "grad_norm": 0.9139587166710175, "learning_rate": 2.9999588810320256e-05, "loss": 0.522, "step": 99 }, { "epoch": 0.03, "grad_norm": 0.7215656771513655, "learning_rate": 2.9999462936679644e-05, "loss": 0.9493, "step": 100 }, { "epoch": 0.03, "grad_norm": 0.8421690952621826, "learning_rate": 2.999932028031259e-05, "loss": 0.7163, "step": 101 }, { "epoch": 0.03, "grad_norm": 0.7878571428138801, "learning_rate": 2.9999160841378727e-05, "loss": 0.6852, "step": 102 }, { "epoch": 0.03, "grad_norm": 0.6104338895179253, "learning_rate": 2.9998984620056437e-05, "loss": 0.9779, "step": 103 }, { "epoch": 0.03, "grad_norm": 0.656457051510017, "learning_rate": 2.9998791616542892e-05, "loss": 0.7116, "step": 104 }, { "epoch": 0.03, "grad_norm": 0.5220908756469721, "learning_rate": 2.9998581831054042e-05, "loss": 0.6986, "step": 105 }, { "epoch": 0.03, "grad_norm": 0.5234074646427421, "learning_rate": 2.999835526382462e-05, "loss": 0.9454, "step": 106 }, { "epoch": 0.03, "grad_norm": 0.5722797361624514, "learning_rate": 2.9998111915108128e-05, "loss": 0.7146, "step": 107 }, { "epoch": 0.04, "grad_norm": 0.5385725650012017, "learning_rate": 2.999785178517684e-05, "loss": 0.7075, "step": 108 }, { "epoch": 0.04, "grad_norm": 0.5364695185426929, "learning_rate": 2.9997574874321812e-05, "loss": 0.966, "step": 109 }, { "epoch": 0.04, "grad_norm": 0.5895416054572736, "learning_rate": 2.9997281182852888e-05, "loss": 0.5121, "step": 110 }, { "epoch": 0.04, "grad_norm": 0.5992947222474976, "learning_rate": 2.9996970711098663e-05, "loss": 0.9672, "step": 111 }, { "epoch": 0.04, "grad_norm": 0.6606724542128426, "learning_rate": 2.9996643459406528e-05, "loss": 0.7147, "step": 112 }, { "epoch": 0.04, "grad_norm": 0.684766456632734, "learning_rate": 2.999629942814264e-05, "loss": 0.7142, "step": 113 }, { "epoch": 0.04, "grad_norm": 0.5587835779177058, "learning_rate": 2.9995938617691925e-05, "loss": 0.9517, "step": 114 }, { "epoch": 0.04, "grad_norm": 0.5826113627174914, "learning_rate": 2.99955610284581e-05, "loss": 0.682, "step": 115 }, { "epoch": 0.04, "grad_norm": 0.6569633191384018, "learning_rate": 2.9995166660863637e-05, "loss": 0.6933, "step": 116 }, { "epoch": 0.04, "grad_norm": 0.5739967833021188, "learning_rate": 2.9994755515349793e-05, "loss": 0.9379, "step": 117 }, { "epoch": 0.04, "grad_norm": 0.5528933112177005, "learning_rate": 2.9994327592376594e-05, "loss": 0.6937, "step": 118 }, { "epoch": 0.04, "grad_norm": 0.49717271520302747, "learning_rate": 2.999388289242284e-05, "loss": 0.6956, "step": 119 }, { "epoch": 0.04, "grad_norm": 0.5116181528595268, "learning_rate": 2.9993421415986094e-05, "loss": 0.9499, "step": 120 }, { "epoch": 0.04, "grad_norm": 0.5371256715585858, "learning_rate": 2.9992943163582702e-05, "loss": 0.7253, "step": 121 }, { "epoch": 0.04, "grad_norm": 0.4682374678745955, "learning_rate": 2.999244813574778e-05, "loss": 0.6991, "step": 122 }, { "epoch": 0.04, "grad_norm": 0.5435669651695999, "learning_rate": 2.9991936333035196e-05, "loss": 0.7055, "step": 123 }, { "epoch": 0.04, "grad_norm": 0.48626457707527665, "learning_rate": 2.999140775601761e-05, "loss": 0.9461, "step": 124 }, { "epoch": 0.04, "grad_norm": 0.5060401531541404, "learning_rate": 2.9990862405286438e-05, "loss": 0.7013, "step": 125 }, { "epoch": 0.04, "grad_norm": 0.5324535882084562, "learning_rate": 2.9990300281451867e-05, "loss": 0.7071, "step": 126 }, { "epoch": 0.04, "grad_norm": 0.5917786578806833, "learning_rate": 2.9989721385142848e-05, "loss": 0.7128, "step": 127 }, { "epoch": 0.04, "grad_norm": 0.4955422114296115, "learning_rate": 2.9989125717007107e-05, "loss": 0.9565, "step": 128 }, { "epoch": 0.04, "grad_norm": 0.5625354228735476, "learning_rate": 2.9988513277711124e-05, "loss": 0.6857, "step": 129 }, { "epoch": 0.04, "grad_norm": 0.5946683867484927, "learning_rate": 2.9987884067940154e-05, "loss": 0.683, "step": 130 }, { "epoch": 0.04, "grad_norm": 0.49626627352524244, "learning_rate": 2.9987238088398212e-05, "loss": 0.9686, "step": 131 }, { "epoch": 0.04, "grad_norm": 0.549730814055932, "learning_rate": 2.9986575339808077e-05, "loss": 0.7305, "step": 132 }, { "epoch": 0.04, "grad_norm": 0.6424753394168775, "learning_rate": 2.9985895822911292e-05, "loss": 0.6913, "step": 133 }, { "epoch": 0.04, "grad_norm": 0.5334730971511613, "learning_rate": 2.9985199538468158e-05, "loss": 0.6987, "step": 134 }, { "epoch": 0.04, "grad_norm": 0.5150816225695143, "learning_rate": 2.998448648725774e-05, "loss": 0.964, "step": 135 }, { "epoch": 0.04, "grad_norm": 0.5188019628945361, "learning_rate": 2.998375667007787e-05, "loss": 0.679, "step": 136 }, { "epoch": 0.04, "grad_norm": 0.5535282101890473, "learning_rate": 2.998301008774512e-05, "loss": 0.6914, "step": 137 }, { "epoch": 0.05, "grad_norm": 0.5053442838699246, "learning_rate": 2.998224674109484e-05, "loss": 0.9222, "step": 138 }, { "epoch": 0.05, "grad_norm": 0.5689067492309814, "learning_rate": 2.998146663098113e-05, "loss": 0.6771, "step": 139 }, { "epoch": 0.05, "grad_norm": 0.5352140533531281, "learning_rate": 2.998066975827684e-05, "loss": 0.6834, "step": 140 }, { "epoch": 0.05, "grad_norm": 0.4775162389420813, "learning_rate": 2.9979856123873592e-05, "loss": 0.94, "step": 141 }, { "epoch": 0.05, "grad_norm": 0.5590250192304753, "learning_rate": 2.997902572868174e-05, "loss": 0.6862, "step": 142 }, { "epoch": 0.05, "grad_norm": 0.5361167649483943, "learning_rate": 2.9978178573630414e-05, "loss": 0.7251, "step": 143 }, { "epoch": 0.05, "grad_norm": 0.5521592900264157, "learning_rate": 2.9977314659667477e-05, "loss": 0.698, "step": 144 }, { "epoch": 0.05, "grad_norm": 0.5982002244577341, "learning_rate": 2.997643398775956e-05, "loss": 0.6797, "step": 145 }, { "epoch": 0.05, "grad_norm": 0.48488683266442256, "learning_rate": 2.9975536558892034e-05, "loss": 0.9537, "step": 146 }, { "epoch": 0.05, "grad_norm": 1.0384277126335768, "learning_rate": 2.9974622374069024e-05, "loss": 0.6973, "step": 147 }, { "epoch": 0.05, "grad_norm": 0.529047815827236, "learning_rate": 2.9973691434313395e-05, "loss": 0.6995, "step": 148 }, { "epoch": 0.05, "grad_norm": 0.5489851183983747, "learning_rate": 2.997274374066677e-05, "loss": 0.7544, "step": 149 }, { "epoch": 0.05, "grad_norm": 0.5217974394099686, "learning_rate": 2.997177929418951e-05, "loss": 0.8331, "step": 150 }, { "epoch": 0.05, "grad_norm": 1.7244533266791393, "learning_rate": 2.9970798095960727e-05, "loss": 0.6744, "step": 151 }, { "epoch": 0.05, "grad_norm": 0.4998205426990884, "learning_rate": 2.9969800147078265e-05, "loss": 0.9387, "step": 152 }, { "epoch": 0.05, "grad_norm": 0.502809405375461, "learning_rate": 2.9968785448658725e-05, "loss": 0.6862, "step": 153 }, { "epoch": 0.05, "grad_norm": 0.6078076691211733, "learning_rate": 2.9967754001837437e-05, "loss": 0.7376, "step": 154 }, { "epoch": 0.05, "grad_norm": 0.4632396675124097, "learning_rate": 2.9966705807768477e-05, "loss": 0.6777, "step": 155 }, { "epoch": 0.05, "grad_norm": 0.4958090454937399, "learning_rate": 2.9965640867624656e-05, "loss": 0.6858, "step": 156 }, { "epoch": 0.05, "grad_norm": 0.4882992304028967, "learning_rate": 2.9964559182597524e-05, "loss": 0.9687, "step": 157 }, { "epoch": 0.05, "grad_norm": 0.4737083519874242, "learning_rate": 2.9963460753897364e-05, "loss": 0.7042, "step": 158 }, { "epoch": 0.05, "grad_norm": 0.4912204747095427, "learning_rate": 2.99623455827532e-05, "loss": 0.6731, "step": 159 }, { "epoch": 0.05, "grad_norm": 0.5429716795349292, "learning_rate": 2.9961213670412777e-05, "loss": 0.7302, "step": 160 }, { "epoch": 0.05, "grad_norm": 0.4860238000272987, "learning_rate": 2.9960065018142584e-05, "loss": 0.9289, "step": 161 }, { "epoch": 0.05, "grad_norm": 0.5067557467340627, "learning_rate": 2.995889962722784e-05, "loss": 0.6849, "step": 162 }, { "epoch": 0.05, "grad_norm": 0.6169634227004159, "learning_rate": 2.9957717498972475e-05, "loss": 0.6843, "step": 163 }, { "epoch": 0.05, "grad_norm": 0.511643765739591, "learning_rate": 2.995651863469916e-05, "loss": 0.9217, "step": 164 }, { "epoch": 0.05, "grad_norm": 0.5565080169825648, "learning_rate": 2.9955303035749303e-05, "loss": 0.6818, "step": 165 }, { "epoch": 0.05, "grad_norm": 0.7355290882751241, "learning_rate": 2.9954070703483014e-05, "loss": 0.6713, "step": 166 }, { "epoch": 0.05, "grad_norm": 0.6265896304191569, "learning_rate": 2.9952821639279136e-05, "loss": 0.9166, "step": 167 }, { "epoch": 0.05, "grad_norm": 0.5979555464638101, "learning_rate": 2.9951555844535233e-05, "loss": 0.6859, "step": 168 }, { "epoch": 0.06, "grad_norm": 0.9057391920343535, "learning_rate": 2.995027332066759e-05, "loss": 0.6884, "step": 169 }, { "epoch": 0.06, "grad_norm": 0.6860241641323211, "learning_rate": 2.994897406911121e-05, "loss": 0.6594, "step": 170 }, { "epoch": 0.06, "grad_norm": 0.5213453460834583, "learning_rate": 2.9947658091319804e-05, "loss": 0.7293, "step": 171 }, { "epoch": 0.06, "grad_norm": 0.6080416857400462, "learning_rate": 2.9946325388765812e-05, "loss": 0.9602, "step": 172 }, { "epoch": 0.06, "grad_norm": 0.47132440220690924, "learning_rate": 2.994497596294037e-05, "loss": 0.683, "step": 173 }, { "epoch": 0.06, "grad_norm": 0.6981401403562657, "learning_rate": 2.9943609815353346e-05, "loss": 0.6884, "step": 174 }, { "epoch": 0.06, "grad_norm": 0.5581456468496492, "learning_rate": 2.99422269475333e-05, "loss": 0.941, "step": 175 }, { "epoch": 0.06, "grad_norm": 0.6818073889816805, "learning_rate": 2.994082736102751e-05, "loss": 0.6722, "step": 176 }, { "epoch": 0.06, "grad_norm": 0.946573403917522, "learning_rate": 2.9939411057401952e-05, "loss": 0.67, "step": 177 }, { "epoch": 0.06, "grad_norm": 0.4919641964857162, "learning_rate": 2.9937978038241315e-05, "loss": 0.9431, "step": 178 }, { "epoch": 0.06, "grad_norm": 0.7070406071928622, "learning_rate": 2.9936528305148993e-05, "loss": 0.693, "step": 179 }, { "epoch": 0.06, "grad_norm": 0.708873069214759, "learning_rate": 2.9935061859747065e-05, "loss": 0.6921, "step": 180 }, { "epoch": 0.06, "grad_norm": 0.49168861773719313, "learning_rate": 2.993357870367633e-05, "loss": 0.9282, "step": 181 }, { "epoch": 0.06, "grad_norm": 0.6815823807489155, "learning_rate": 2.993207883859627e-05, "loss": 0.4865, "step": 182 }, { "epoch": 0.06, "grad_norm": 0.7258908883398096, "learning_rate": 2.9930562266185068e-05, "loss": 0.9302, "step": 183 }, { "epoch": 0.06, "grad_norm": 0.6370408679096319, "learning_rate": 2.9929028988139594e-05, "loss": 0.6812, "step": 184 }, { "epoch": 0.06, "grad_norm": 0.9101468779032295, "learning_rate": 2.992747900617542e-05, "loss": 0.6856, "step": 185 }, { "epoch": 0.06, "grad_norm": 0.46613561435821954, "learning_rate": 2.9925912322026802e-05, "loss": 0.933, "step": 186 }, { "epoch": 0.06, "grad_norm": 0.8919234343405715, "learning_rate": 2.992432893744669e-05, "loss": 0.6778, "step": 187 }, { "epoch": 0.06, "grad_norm": 0.7083491816753998, "learning_rate": 2.9922728854206704e-05, "loss": 0.6877, "step": 188 }, { "epoch": 0.06, "grad_norm": 0.5930678621455631, "learning_rate": 2.9921112074097165e-05, "loss": 0.9308, "step": 189 }, { "epoch": 0.06, "grad_norm": 0.723899732353784, "learning_rate": 2.9919478598927073e-05, "loss": 0.6697, "step": 190 }, { "epoch": 0.06, "grad_norm": 0.5273217828258369, "learning_rate": 2.99178284305241e-05, "loss": 0.6589, "step": 191 }, { "epoch": 0.06, "grad_norm": 0.6016186748614608, "learning_rate": 2.99161615707346e-05, "loss": 0.9316, "step": 192 }, { "epoch": 0.06, "grad_norm": 0.6397591918972924, "learning_rate": 2.9914478021423607e-05, "loss": 0.7189, "step": 193 }, { "epoch": 0.06, "grad_norm": 0.740900654664896, "learning_rate": 2.991277778447482e-05, "loss": 0.6813, "step": 194 }, { "epoch": 0.06, "grad_norm": 0.5538100220421596, "learning_rate": 2.991106086179062e-05, "loss": 0.6617, "step": 195 }, { "epoch": 0.06, "grad_norm": 0.7929391660695634, "learning_rate": 2.9909327255292043e-05, "loss": 0.9129, "step": 196 }, { "epoch": 0.06, "grad_norm": 0.558505798735182, "learning_rate": 2.990757696691881e-05, "loss": 0.6703, "step": 197 }, { "epoch": 0.06, "grad_norm": 0.6617542741809586, "learning_rate": 2.9905809998629302e-05, "loss": 0.6537, "step": 198 }, { "epoch": 0.06, "grad_norm": 0.4771905390601857, "learning_rate": 2.9904026352400548e-05, "loss": 0.6597, "step": 199 }, { "epoch": 0.07, "grad_norm": 0.5640426925062384, "learning_rate": 2.9902226030228252e-05, "loss": 0.9155, "step": 200 }, { "epoch": 0.07, "grad_norm": 0.5056354219654766, "learning_rate": 2.9900409034126782e-05, "loss": 0.68, "step": 201 }, { "epoch": 0.07, "grad_norm": 0.6054234387435584, "learning_rate": 2.9898575366129145e-05, "loss": 0.6762, "step": 202 }, { "epoch": 0.07, "grad_norm": 0.5000274436045503, "learning_rate": 2.9896725028287017e-05, "loss": 0.9324, "step": 203 }, { "epoch": 0.07, "grad_norm": 0.6026696037140091, "learning_rate": 2.9894858022670714e-05, "loss": 0.7157, "step": 204 }, { "epoch": 0.07, "grad_norm": 0.4990655685232816, "learning_rate": 2.9892974351369215e-05, "loss": 0.6804, "step": 205 }, { "epoch": 0.07, "grad_norm": 0.47546888035484375, "learning_rate": 2.989107401649013e-05, "loss": 0.665, "step": 206 }, { "epoch": 0.07, "grad_norm": 0.5295569009643406, "learning_rate": 2.9889157020159724e-05, "loss": 0.9413, "step": 207 }, { "epoch": 0.07, "grad_norm": 0.5092808578751946, "learning_rate": 2.9887223364522903e-05, "loss": 0.6866, "step": 208 }, { "epoch": 0.07, "grad_norm": 0.61871703674558, "learning_rate": 2.9885273051743216e-05, "loss": 0.6948, "step": 209 }, { "epoch": 0.07, "grad_norm": 0.4886170033655132, "learning_rate": 2.988330608400284e-05, "loss": 0.8962, "step": 210 }, { "epoch": 0.07, "grad_norm": 0.5620432416016636, "learning_rate": 2.9881322463502593e-05, "loss": 0.6691, "step": 211 }, { "epoch": 0.07, "grad_norm": 0.5116021133975324, "learning_rate": 2.9879322192461932e-05, "loss": 0.6428, "step": 212 }, { "epoch": 0.07, "grad_norm": 0.45250358176337413, "learning_rate": 2.9877305273118923e-05, "loss": 0.9063, "step": 213 }, { "epoch": 0.07, "grad_norm": 0.5336819371418162, "learning_rate": 2.987527170773029e-05, "loss": 0.6826, "step": 214 }, { "epoch": 0.07, "grad_norm": 0.5908194710911264, "learning_rate": 2.987322149857136e-05, "loss": 0.7284, "step": 215 }, { "epoch": 0.07, "grad_norm": 0.45965284698786457, "learning_rate": 2.987115464793608e-05, "loss": 0.6605, "step": 216 }, { "epoch": 0.07, "grad_norm": 0.415304945804936, "learning_rate": 2.986907115813704e-05, "loss": 0.66, "step": 217 }, { "epoch": 0.07, "grad_norm": 0.4920913270742246, "learning_rate": 2.986697103150542e-05, "loss": 0.9093, "step": 218 }, { "epoch": 0.07, "grad_norm": 0.44078225050842557, "learning_rate": 2.9864854270391033e-05, "loss": 0.7072, "step": 219 }, { "epoch": 0.07, "grad_norm": 0.4885646557761786, "learning_rate": 2.9862720877162298e-05, "loss": 0.6808, "step": 220 }, { "epoch": 0.07, "grad_norm": 0.5071395883327944, "learning_rate": 2.9860570854206247e-05, "loss": 0.9563, "step": 221 }, { "epoch": 0.07, "grad_norm": 0.47204157409881053, "learning_rate": 2.985840420392851e-05, "loss": 0.6621, "step": 222 }, { "epoch": 0.07, "grad_norm": 0.48639545494476005, "learning_rate": 2.9856220928753336e-05, "loss": 0.6729, "step": 223 }, { "epoch": 0.07, "grad_norm": 0.45713568052450837, "learning_rate": 2.9854021031123555e-05, "loss": 0.9298, "step": 224 }, { "epoch": 0.07, "grad_norm": 0.4988959724883191, "learning_rate": 2.9851804513500614e-05, "loss": 0.4856, "step": 225 }, { "epoch": 0.07, "grad_norm": 0.47313286490420803, "learning_rate": 2.9849571378364553e-05, "loss": 0.9376, "step": 226 }, { "epoch": 0.07, "grad_norm": 0.5137125589154344, "learning_rate": 2.984732162821399e-05, "loss": 0.6777, "step": 227 }, { "epoch": 0.07, "grad_norm": 0.633609776571836, "learning_rate": 2.9845055265566154e-05, "loss": 0.6476, "step": 228 }, { "epoch": 0.07, "grad_norm": 0.5078784747863635, "learning_rate": 2.984277229295685e-05, "loss": 0.9508, "step": 229 }, { "epoch": 0.08, "grad_norm": 0.4586694870295808, "learning_rate": 2.984047271294047e-05, "loss": 0.6612, "step": 230 }, { "epoch": 0.08, "grad_norm": 0.606299609395451, "learning_rate": 2.983815652808998e-05, "loss": 0.7056, "step": 231 }, { "epoch": 0.08, "grad_norm": 0.4804043609000068, "learning_rate": 2.9835823740996944e-05, "loss": 0.6515, "step": 232 }, { "epoch": 0.08, "grad_norm": 0.4824205230985662, "learning_rate": 2.9833474354271487e-05, "loss": 0.9405, "step": 233 }, { "epoch": 0.08, "grad_norm": 0.5206923248046585, "learning_rate": 2.9831108370542306e-05, "loss": 0.6469, "step": 234 }, { "epoch": 0.08, "grad_norm": 0.5086392932281077, "learning_rate": 2.9828725792456676e-05, "loss": 0.9299, "step": 235 }, { "epoch": 0.08, "grad_norm": 0.5244723986822152, "learning_rate": 2.9826326622680436e-05, "loss": 0.727, "step": 236 }, { "epoch": 0.08, "grad_norm": 0.5205012610292201, "learning_rate": 2.9823910863897995e-05, "loss": 0.6723, "step": 237 }, { "epoch": 0.08, "grad_norm": 0.48915272238115887, "learning_rate": 2.9821478518812307e-05, "loss": 0.6579, "step": 238 }, { "epoch": 0.08, "grad_norm": 0.585286167980275, "learning_rate": 2.9819029590144903e-05, "loss": 0.9418, "step": 239 }, { "epoch": 0.08, "grad_norm": 0.5283293520261767, "learning_rate": 2.9816564080635856e-05, "loss": 0.6597, "step": 240 }, { "epoch": 0.08, "grad_norm": 0.5385097749246471, "learning_rate": 2.9814081993043803e-05, "loss": 0.6701, "step": 241 }, { "epoch": 0.08, "grad_norm": 0.6119953476648748, "learning_rate": 2.9811583330145915e-05, "loss": 0.72, "step": 242 }, { "epoch": 0.08, "grad_norm": 0.4768573777971515, "learning_rate": 2.9809068094737923e-05, "loss": 0.6529, "step": 243 }, { "epoch": 0.08, "grad_norm": 0.4846234598303821, "learning_rate": 2.9806536289634095e-05, "loss": 0.9218, "step": 244 }, { "epoch": 0.08, "grad_norm": 0.5022275088248602, "learning_rate": 2.9803987917667234e-05, "loss": 0.684, "step": 245 }, { "epoch": 0.08, "grad_norm": 0.5092158652328147, "learning_rate": 2.980142298168869e-05, "loss": 0.6763, "step": 246 }, { "epoch": 0.08, "grad_norm": 0.4795898172288634, "learning_rate": 2.9798841484568334e-05, "loss": 0.9027, "step": 247 }, { "epoch": 0.08, "grad_norm": 0.4981423768370757, "learning_rate": 2.9796243429194578e-05, "loss": 0.6856, "step": 248 }, { "epoch": 0.08, "grad_norm": 0.42960609665232213, "learning_rate": 2.9793628818474356e-05, "loss": 0.6807, "step": 249 }, { "epoch": 0.08, "grad_norm": 0.47592376151994314, "learning_rate": 2.9790997655333123e-05, "loss": 0.9015, "step": 250 }, { "epoch": 0.08, "grad_norm": 0.4789767721993736, "learning_rate": 2.9788349942714858e-05, "loss": 0.6647, "step": 251 }, { "epoch": 0.08, "grad_norm": 0.4092345973168466, "learning_rate": 2.9785685683582057e-05, "loss": 0.668, "step": 252 }, { "epoch": 0.08, "grad_norm": 0.4522893853826958, "learning_rate": 2.978300488091573e-05, "loss": 0.7078, "step": 253 }, { "epoch": 0.08, "grad_norm": 0.4139133401092401, "learning_rate": 2.9780307537715396e-05, "loss": 0.6646, "step": 254 }, { "epoch": 0.08, "grad_norm": 0.45990768222881845, "learning_rate": 2.9777593656999078e-05, "loss": 0.9122, "step": 255 }, { "epoch": 0.08, "grad_norm": 0.4498140033377025, "learning_rate": 2.9774863241803313e-05, "loss": 0.6521, "step": 256 }, { "epoch": 0.08, "grad_norm": 0.4523669081179706, "learning_rate": 2.977211629518312e-05, "loss": 0.6554, "step": 257 }, { "epoch": 0.08, "grad_norm": 0.521604623604145, "learning_rate": 2.976935282021204e-05, "loss": 0.9226, "step": 258 }, { "epoch": 0.08, "grad_norm": 0.45521261733853774, "learning_rate": 2.976657281998208e-05, "loss": 0.6745, "step": 259 }, { "epoch": 0.08, "grad_norm": 0.46935656768565637, "learning_rate": 2.9763776297603758e-05, "loss": 0.667, "step": 260 }, { "epoch": 0.09, "grad_norm": 0.45470984090990396, "learning_rate": 2.9760963256206068e-05, "loss": 0.8913, "step": 261 }, { "epoch": 0.09, "grad_norm": 0.4902568388729428, "learning_rate": 2.975813369893649e-05, "loss": 0.6602, "step": 262 }, { "epoch": 0.09, "grad_norm": 0.4605295994695376, "learning_rate": 2.9755287628960982e-05, "loss": 0.6627, "step": 263 }, { "epoch": 0.09, "grad_norm": 0.5321259633757994, "learning_rate": 2.975242504946398e-05, "loss": 0.7027, "step": 264 }, { "epoch": 0.09, "grad_norm": 0.4979239391347357, "learning_rate": 2.9749545963648395e-05, "loss": 0.902, "step": 265 }, { "epoch": 0.09, "grad_norm": 0.5192912207746857, "learning_rate": 2.97466503747356e-05, "loss": 0.6378, "step": 266 }, { "epoch": 0.09, "grad_norm": 0.47159308402665284, "learning_rate": 2.9743738285965433e-05, "loss": 0.6432, "step": 267 }, { "epoch": 0.09, "grad_norm": 0.45714385339764774, "learning_rate": 2.97408097005962e-05, "loss": 0.6629, "step": 268 }, { "epoch": 0.09, "grad_norm": 0.5110892903888681, "learning_rate": 2.973786462190466e-05, "loss": 0.9487, "step": 269 }, { "epoch": 0.09, "grad_norm": 0.39652653176174923, "learning_rate": 2.973490305318603e-05, "loss": 0.6596, "step": 270 }, { "epoch": 0.09, "grad_norm": 0.4836582217275633, "learning_rate": 2.9731924997753974e-05, "loss": 0.6553, "step": 271 }, { "epoch": 0.09, "grad_norm": 0.5023980553366248, "learning_rate": 2.97289304589406e-05, "loss": 0.9363, "step": 272 }, { "epoch": 0.09, "grad_norm": 0.5629785430222661, "learning_rate": 2.972591944009647e-05, "loss": 0.6685, "step": 273 }, { "epoch": 0.09, "grad_norm": 0.49863614081625, "learning_rate": 2.9722891944590572e-05, "loss": 0.6649, "step": 274 }, { "epoch": 0.09, "grad_norm": 0.44597559600086834, "learning_rate": 2.9719847975810343e-05, "loss": 0.6973, "step": 275 }, { "epoch": 0.09, "grad_norm": 0.577039121113811, "learning_rate": 2.9716787537161635e-05, "loss": 0.9201, "step": 276 }, { "epoch": 0.09, "grad_norm": 0.4191145741298294, "learning_rate": 2.971371063206875e-05, "loss": 0.667, "step": 277 }, { "epoch": 0.09, "grad_norm": 0.5210148955159346, "learning_rate": 2.9710617263974385e-05, "loss": 0.6605, "step": 278 }, { "epoch": 0.09, "grad_norm": 0.5532420831128856, "learning_rate": 2.970750743633969e-05, "loss": 0.9424, "step": 279 }, { "epoch": 0.09, "grad_norm": 0.4735415455039284, "learning_rate": 2.9704381152644205e-05, "loss": 0.6597, "step": 280 }, { "epoch": 0.09, "grad_norm": 0.5412438977736135, "learning_rate": 2.97012384163859e-05, "loss": 0.6476, "step": 281 }, { "epoch": 0.09, "grad_norm": 0.44597547954304856, "learning_rate": 2.9698079231081144e-05, "loss": 0.943, "step": 282 }, { "epoch": 0.09, "grad_norm": 0.5141368959996211, "learning_rate": 2.9694903600264712e-05, "loss": 0.6465, "step": 283 }, { "epoch": 0.09, "grad_norm": 0.49668760963966974, "learning_rate": 2.9691711527489777e-05, "loss": 0.6532, "step": 284 }, { "epoch": 0.09, "grad_norm": 0.38392784465631064, "learning_rate": 2.968850301632792e-05, "loss": 0.6857, "step": 285 }, { "epoch": 0.09, "grad_norm": 0.5819251083944892, "learning_rate": 2.9685278070369104e-05, "loss": 0.7356, "step": 286 }, { "epoch": 0.09, "grad_norm": 0.4606415939187071, "learning_rate": 2.9682036693221684e-05, "loss": 0.9079, "step": 287 }, { "epoch": 0.09, "grad_norm": 0.4162991214065934, "learning_rate": 2.9678778888512396e-05, "loss": 0.6482, "step": 288 }, { "epoch": 0.09, "grad_norm": 0.43800537612651697, "learning_rate": 2.9675504659886362e-05, "loss": 0.6662, "step": 289 }, { "epoch": 0.09, "grad_norm": 0.42880183232398883, "learning_rate": 2.9672214011007087e-05, "loss": 0.8873, "step": 290 }, { "epoch": 0.1, "grad_norm": 0.47413915177190985, "learning_rate": 2.9668906945556428e-05, "loss": 0.66, "step": 291 }, { "epoch": 0.1, "grad_norm": 0.4383582595292897, "learning_rate": 2.966558346723463e-05, "loss": 0.6514, "step": 292 }, { "epoch": 0.1, "grad_norm": 0.6851912628717889, "learning_rate": 2.966224357976029e-05, "loss": 0.9211, "step": 293 }, { "epoch": 0.1, "grad_norm": 0.43195700274155796, "learning_rate": 2.9658887286870377e-05, "loss": 0.6715, "step": 294 }, { "epoch": 0.1, "grad_norm": 0.5576997704702761, "learning_rate": 2.9655514592320195e-05, "loss": 0.6574, "step": 295 }, { "epoch": 0.1, "grad_norm": 0.4594805763389257, "learning_rate": 2.9652125499883428e-05, "loss": 0.9259, "step": 296 }, { "epoch": 0.1, "grad_norm": 0.49046088478625127, "learning_rate": 2.9648720013352086e-05, "loss": 0.4688, "step": 297 }, { "epoch": 0.1, "grad_norm": 0.5224554915504589, "learning_rate": 2.9645298136536524e-05, "loss": 0.9096, "step": 298 }, { "epoch": 0.1, "grad_norm": 0.4290640247261897, "learning_rate": 2.9641859873265452e-05, "loss": 0.6559, "step": 299 }, { "epoch": 0.1, "grad_norm": 0.521725510708576, "learning_rate": 2.9638405227385894e-05, "loss": 0.6548, "step": 300 }, { "epoch": 0.1, "grad_norm": 0.4727364565431979, "learning_rate": 2.9634934202763214e-05, "loss": 0.907, "step": 301 }, { "epoch": 0.1, "grad_norm": 0.5660073448505623, "learning_rate": 2.963144680328111e-05, "loss": 0.6737, "step": 302 }, { "epoch": 0.1, "grad_norm": 0.5660444915856627, "learning_rate": 2.9627943032841577e-05, "loss": 0.6239, "step": 303 }, { "epoch": 0.1, "grad_norm": 0.4658963535526167, "learning_rate": 2.9624422895364963e-05, "loss": 0.8996, "step": 304 }, { "epoch": 0.1, "grad_norm": 0.5359862051372392, "learning_rate": 2.962088639478989e-05, "loss": 0.6712, "step": 305 }, { "epoch": 0.1, "grad_norm": 0.5362762488003882, "learning_rate": 2.9617333535073324e-05, "loss": 0.658, "step": 306 }, { "epoch": 0.1, "grad_norm": 2.0567698725224504, "learning_rate": 2.9613764320190508e-05, "loss": 0.9695, "step": 307 }, { "epoch": 0.1, "grad_norm": 0.6335593333363224, "learning_rate": 2.9610178754135005e-05, "loss": 0.72, "step": 308 }, { "epoch": 0.1, "grad_norm": 0.5023763982308612, "learning_rate": 2.9606576840918657e-05, "loss": 0.6469, "step": 309 }, { "epoch": 0.1, "grad_norm": 0.6390190759173342, "learning_rate": 2.96029585845716e-05, "loss": 0.6525, "step": 310 }, { "epoch": 0.1, "grad_norm": 0.51068047709934, "learning_rate": 2.9599323989142266e-05, "loss": 0.8944, "step": 311 }, { "epoch": 0.1, "grad_norm": 0.6241225375761374, "learning_rate": 2.959567305869736e-05, "loss": 0.6617, "step": 312 }, { "epoch": 0.1, "grad_norm": 0.544966425333211, "learning_rate": 2.9592005797321862e-05, "loss": 0.7175, "step": 313 }, { "epoch": 0.1, "grad_norm": 0.6146560021637698, "learning_rate": 2.9588322209119037e-05, "loss": 0.6343, "step": 314 }, { "epoch": 0.1, "grad_norm": 0.6190749562797465, "learning_rate": 2.9584622298210405e-05, "loss": 0.6582, "step": 315 }, { "epoch": 0.1, "grad_norm": 0.540657734651325, "learning_rate": 2.9580906068735754e-05, "loss": 0.9333, "step": 316 }, { "epoch": 0.1, "grad_norm": 0.6603917705687957, "learning_rate": 2.957717352485313e-05, "loss": 0.6229, "step": 317 }, { "epoch": 0.1, "grad_norm": 0.5536718518456261, "learning_rate": 2.9573424670738827e-05, "loss": 0.9048, "step": 318 }, { "epoch": 0.1, "grad_norm": 0.5513956182933006, "learning_rate": 2.9569659510587405e-05, "loss": 0.69, "step": 319 }, { "epoch": 0.1, "grad_norm": 0.5481974079097557, "learning_rate": 2.9565878048611655e-05, "loss": 0.6906, "step": 320 }, { "epoch": 0.1, "grad_norm": 0.6045497749398104, "learning_rate": 2.956208028904261e-05, "loss": 0.6596, "step": 321 }, { "epoch": 0.11, "grad_norm": 0.48284976004055286, "learning_rate": 2.955826623612954e-05, "loss": 0.9227, "step": 322 }, { "epoch": 0.11, "grad_norm": 0.5204407652025896, "learning_rate": 2.9554435894139945e-05, "loss": 0.6538, "step": 323 }, { "epoch": 0.11, "grad_norm": 0.5629398975884938, "learning_rate": 2.9550589267359554e-05, "loss": 0.6822, "step": 324 }, { "epoch": 0.11, "grad_norm": 0.48248943041761116, "learning_rate": 2.9546726360092307e-05, "loss": 0.6703, "step": 325 }, { "epoch": 0.11, "grad_norm": 0.5433763116338408, "learning_rate": 2.954284717666037e-05, "loss": 0.6638, "step": 326 }, { "epoch": 0.11, "grad_norm": 0.47209569907446614, "learning_rate": 2.9538951721404116e-05, "loss": 0.9173, "step": 327 }, { "epoch": 0.11, "grad_norm": 0.5365926071713311, "learning_rate": 2.9535039998682126e-05, "loss": 0.6638, "step": 328 }, { "epoch": 0.11, "grad_norm": 0.4577873015121729, "learning_rate": 2.9531112012871176e-05, "loss": 0.6421, "step": 329 }, { "epoch": 0.11, "grad_norm": 0.5353648573371762, "learning_rate": 2.952716776836626e-05, "loss": 0.9121, "step": 330 }, { "epoch": 0.11, "grad_norm": 0.47451546964622954, "learning_rate": 2.9523207269580526e-05, "loss": 0.6544, "step": 331 }, { "epoch": 0.11, "grad_norm": 0.5022118236997233, "learning_rate": 2.9519230520945346e-05, "loss": 0.6817, "step": 332 }, { "epoch": 0.11, "grad_norm": 0.5513198893348137, "learning_rate": 2.951523752691025e-05, "loss": 0.9065, "step": 333 }, { "epoch": 0.11, "grad_norm": 0.4563112350075268, "learning_rate": 2.951122829194296e-05, "loss": 0.6469, "step": 334 }, { "epoch": 0.11, "grad_norm": 0.6076373470118571, "learning_rate": 2.950720282052936e-05, "loss": 0.6822, "step": 335 }, { "epoch": 0.11, "grad_norm": 0.4394628640370711, "learning_rate": 2.95031611171735e-05, "loss": 0.6383, "step": 336 }, { "epoch": 0.11, "grad_norm": 0.5301986359697761, "learning_rate": 2.9499103186397598e-05, "loss": 0.9036, "step": 337 }, { "epoch": 0.11, "grad_norm": 0.49815405306783966, "learning_rate": 2.9495029032742025e-05, "loss": 0.6724, "step": 338 }, { "epoch": 0.11, "grad_norm": 0.493479116882354, "learning_rate": 2.9490938660765313e-05, "loss": 0.6337, "step": 339 }, { "epoch": 0.11, "grad_norm": 0.4912024133355778, "learning_rate": 2.948683207504412e-05, "loss": 0.6627, "step": 340 }, { "epoch": 0.11, "grad_norm": 0.46707486147343974, "learning_rate": 2.9482709280173267e-05, "loss": 0.9112, "step": 341 }, { "epoch": 0.11, "grad_norm": 0.47249180825322595, "learning_rate": 2.947857028076569e-05, "loss": 0.6673, "step": 342 }, { "epoch": 0.11, "grad_norm": 0.4398499549916723, "learning_rate": 2.9474415081452482e-05, "loss": 0.6613, "step": 343 }, { "epoch": 0.11, "grad_norm": 0.46817381598852387, "learning_rate": 2.9470243686882838e-05, "loss": 0.8975, "step": 344 }, { "epoch": 0.11, "grad_norm": 0.4874306094501961, "learning_rate": 2.9466056101724085e-05, "loss": 0.6362, "step": 345 }, { "epoch": 0.11, "grad_norm": 0.4703998231181428, "learning_rate": 2.9461852330661664e-05, "loss": 0.69, "step": 346 }, { "epoch": 0.11, "grad_norm": 0.4758210624328601, "learning_rate": 2.9457632378399133e-05, "loss": 0.6664, "step": 347 }, { "epoch": 0.11, "grad_norm": 0.5863947105452997, "learning_rate": 2.945339624965814e-05, "loss": 0.9173, "step": 348 }, { "epoch": 0.11, "grad_norm": 0.5032538137088288, "learning_rate": 2.944914394917844e-05, "loss": 0.6376, "step": 349 }, { "epoch": 0.11, "grad_norm": 0.40994548782830353, "learning_rate": 2.9444875481717888e-05, "loss": 0.6622, "step": 350 }, { "epoch": 0.11, "grad_norm": 0.48396312433639777, "learning_rate": 2.944059085205242e-05, "loss": 0.9311, "step": 351 }, { "epoch": 0.11, "grad_norm": 0.4155610305414481, "learning_rate": 2.943629006497606e-05, "loss": 0.6458, "step": 352 }, { "epoch": 0.12, "grad_norm": 0.5911686478438456, "learning_rate": 2.943197312530091e-05, "loss": 0.6427, "step": 353 }, { "epoch": 0.12, "grad_norm": 0.4459862094115797, "learning_rate": 2.942764003785714e-05, "loss": 0.901, "step": 354 }, { "epoch": 0.12, "grad_norm": 0.5372117570890818, "learning_rate": 2.9423290807492998e-05, "loss": 0.6689, "step": 355 }, { "epoch": 0.12, "grad_norm": 0.47884812143650035, "learning_rate": 2.9418925439074784e-05, "loss": 0.6444, "step": 356 }, { "epoch": 0.12, "grad_norm": 0.45367842302439737, "learning_rate": 2.941454393748686e-05, "loss": 0.71, "step": 357 }, { "epoch": 0.12, "grad_norm": 0.46179606709638155, "learning_rate": 2.9410146307631637e-05, "loss": 0.6565, "step": 358 }, { "epoch": 0.12, "grad_norm": 0.44634964420030593, "learning_rate": 2.940573255442957e-05, "loss": 0.9262, "step": 359 }, { "epoch": 0.12, "grad_norm": 0.4746299667412623, "learning_rate": 2.9401302682819156e-05, "loss": 0.6431, "step": 360 }, { "epoch": 0.12, "grad_norm": 0.46815760672123685, "learning_rate": 2.9396856697756935e-05, "loss": 0.6765, "step": 361 }, { "epoch": 0.12, "grad_norm": 0.48891223033758197, "learning_rate": 2.939239460421746e-05, "loss": 0.9126, "step": 362 }, { "epoch": 0.12, "grad_norm": 0.4216127835202362, "learning_rate": 2.9387916407193326e-05, "loss": 0.6389, "step": 363 }, { "epoch": 0.12, "grad_norm": 0.5010225823198274, "learning_rate": 2.9383422111695122e-05, "loss": 0.6605, "step": 364 }, { "epoch": 0.12, "grad_norm": 0.47983585694274267, "learning_rate": 2.9378911722751477e-05, "loss": 0.9293, "step": 365 }, { "epoch": 0.12, "grad_norm": 0.4941054040917964, "learning_rate": 2.9374385245409003e-05, "loss": 0.6531, "step": 366 }, { "epoch": 0.12, "grad_norm": 0.5553377959786, "learning_rate": 2.9369842684732334e-05, "loss": 0.6332, "step": 367 }, { "epoch": 0.12, "grad_norm": 0.45810236705302193, "learning_rate": 2.936528404580408e-05, "loss": 0.7043, "step": 368 }, { "epoch": 0.12, "grad_norm": 0.5050341467500221, "learning_rate": 2.936070933372485e-05, "loss": 0.6643, "step": 369 }, { "epoch": 0.12, "grad_norm": 0.4832702416502093, "learning_rate": 2.9356118553613244e-05, "loss": 0.9129, "step": 370 }, { "epoch": 0.12, "grad_norm": 0.5180768428141715, "learning_rate": 2.9351511710605828e-05, "loss": 0.6381, "step": 371 }, { "epoch": 0.12, "grad_norm": 0.4984169012194238, "learning_rate": 2.934688880985714e-05, "loss": 0.6246, "step": 372 }, { "epoch": 0.12, "grad_norm": 2.2975022615104015, "learning_rate": 2.9342249856539705e-05, "loss": 0.9514, "step": 373 }, { "epoch": 0.12, "grad_norm": 0.8563175740058667, "learning_rate": 2.9337594855843976e-05, "loss": 0.6451, "step": 374 }, { "epoch": 0.12, "grad_norm": 0.44222161068768506, "learning_rate": 2.9332923812978395e-05, "loss": 0.6515, "step": 375 }, { "epoch": 0.12, "grad_norm": 0.9417356471129726, "learning_rate": 2.9328236733169337e-05, "loss": 0.9057, "step": 376 }, { "epoch": 0.12, "grad_norm": 0.43453250953669187, "learning_rate": 2.9323533621661107e-05, "loss": 0.6436, "step": 377 }, { "epoch": 0.12, "grad_norm": 0.4108712188780906, "learning_rate": 2.9318814483715982e-05, "loss": 0.6378, "step": 378 }, { "epoch": 0.12, "grad_norm": 0.4706527571387842, "learning_rate": 2.9314079324614132e-05, "loss": 0.6916, "step": 379 }, { "epoch": 0.12, "grad_norm": 0.4394220292103483, "learning_rate": 2.930932814965369e-05, "loss": 0.8963, "step": 380 }, { "epoch": 0.12, "grad_norm": 0.41402927526719646, "learning_rate": 2.9304560964150683e-05, "loss": 0.6529, "step": 381 }, { "epoch": 0.12, "grad_norm": 0.42360567963609386, "learning_rate": 2.9299777773439056e-05, "loss": 0.6566, "step": 382 }, { "epoch": 0.13, "grad_norm": 0.4479397662356841, "learning_rate": 2.929497858287067e-05, "loss": 0.9348, "step": 383 }, { "epoch": 0.13, "grad_norm": 0.4402327589824743, "learning_rate": 2.9290163397815283e-05, "loss": 0.6405, "step": 384 }, { "epoch": 0.13, "grad_norm": 0.46548867247958, "learning_rate": 2.928533222366055e-05, "loss": 0.6461, "step": 385 }, { "epoch": 0.13, "grad_norm": 0.41592194339423216, "learning_rate": 2.9280485065812025e-05, "loss": 0.6545, "step": 386 }, { "epoch": 0.13, "grad_norm": 0.4599824931010298, "learning_rate": 2.927562192969312e-05, "loss": 0.9508, "step": 387 }, { "epoch": 0.13, "grad_norm": 0.4353243656201792, "learning_rate": 2.9270742820745157e-05, "loss": 0.6578, "step": 388 }, { "epoch": 0.13, "grad_norm": 0.42810394775712785, "learning_rate": 2.9265847744427305e-05, "loss": 0.6309, "step": 389 }, { "epoch": 0.13, "grad_norm": 0.45836144403996704, "learning_rate": 2.9260936706216615e-05, "loss": 0.6784, "step": 390 }, { "epoch": 0.13, "grad_norm": 0.4370219399249084, "learning_rate": 2.925600971160799e-05, "loss": 0.891, "step": 391 }, { "epoch": 0.13, "grad_norm": 0.5965109385413148, "learning_rate": 2.925106676611418e-05, "loss": 0.6577, "step": 392 }, { "epoch": 0.13, "grad_norm": 0.4240381198175698, "learning_rate": 2.9246107875265802e-05, "loss": 0.6358, "step": 393 }, { "epoch": 0.13, "grad_norm": 0.4836637342405808, "learning_rate": 2.9241133044611295e-05, "loss": 0.934, "step": 394 }, { "epoch": 0.13, "grad_norm": 0.44731977526756017, "learning_rate": 2.9236142279716938e-05, "loss": 0.6518, "step": 395 }, { "epoch": 0.13, "grad_norm": 0.5045241458118127, "learning_rate": 2.9231135586166845e-05, "loss": 0.7071, "step": 396 }, { "epoch": 0.13, "grad_norm": 0.4298098268418705, "learning_rate": 2.9226112969562952e-05, "loss": 0.6405, "step": 397 }, { "epoch": 0.13, "grad_norm": 0.4776280870150111, "learning_rate": 2.9221074435524995e-05, "loss": 0.6564, "step": 398 }, { "epoch": 0.13, "grad_norm": 0.49364269118662174, "learning_rate": 2.9216019989690536e-05, "loss": 0.928, "step": 399 }, { "epoch": 0.13, "grad_norm": 0.4877886832371823, "learning_rate": 2.921094963771494e-05, "loss": 0.813, "step": 400 }, { "epoch": 0.13, "grad_norm": 0.4853940811342239, "learning_rate": 2.9205863385271364e-05, "loss": 0.4655, "step": 401 }, { "epoch": 0.13, "grad_norm": 0.4332161696845459, "learning_rate": 2.9200761238050756e-05, "loss": 0.9244, "step": 402 }, { "epoch": 0.13, "grad_norm": 0.4417217924925487, "learning_rate": 2.919564320176185e-05, "loss": 0.6365, "step": 403 }, { "epoch": 0.13, "grad_norm": 0.44834839034826507, "learning_rate": 2.9190509282131156e-05, "loss": 0.6506, "step": 404 }, { "epoch": 0.13, "grad_norm": 3.648948760416141, "learning_rate": 2.9185359484902965e-05, "loss": 1.0201, "step": 405 }, { "epoch": 0.13, "grad_norm": 0.45841615426514204, "learning_rate": 2.9180193815839322e-05, "loss": 0.6409, "step": 406 }, { "epoch": 0.13, "grad_norm": 0.4783866513887179, "learning_rate": 2.9175012280720024e-05, "loss": 0.7154, "step": 407 }, { "epoch": 0.13, "grad_norm": 0.3889915573337204, "learning_rate": 2.916981488534265e-05, "loss": 0.6397, "step": 408 }, { "epoch": 0.13, "grad_norm": 0.5177734869047784, "learning_rate": 2.9164601635522493e-05, "loss": 0.9076, "step": 409 }, { "epoch": 0.13, "grad_norm": 0.4277082871334709, "learning_rate": 2.91593725370926e-05, "loss": 0.6369, "step": 410 }, { "epoch": 0.13, "grad_norm": 0.4360132283230203, "learning_rate": 2.9154127595903755e-05, "loss": 0.6445, "step": 411 }, { "epoch": 0.13, "grad_norm": 0.5760306684471718, "learning_rate": 2.9148866817824454e-05, "loss": 0.6542, "step": 412 }, { "epoch": 0.13, "grad_norm": 0.4299463219083792, "learning_rate": 2.9143590208740925e-05, "loss": 0.8959, "step": 413 }, { "epoch": 0.14, "grad_norm": 0.42220515460074765, "learning_rate": 2.913829777455711e-05, "loss": 0.6499, "step": 414 }, { "epoch": 0.14, "grad_norm": 0.412101128433801, "learning_rate": 2.913298952119464e-05, "loss": 0.6609, "step": 415 }, { "epoch": 0.14, "grad_norm": 0.4200264796070036, "learning_rate": 2.9127665454592872e-05, "loss": 0.8941, "step": 416 }, { "epoch": 0.14, "grad_norm": 0.3909426739205423, "learning_rate": 2.9122325580708833e-05, "loss": 0.6624, "step": 417 }, { "epoch": 0.14, "grad_norm": 0.4499735907880299, "learning_rate": 2.9116969905517256e-05, "loss": 0.7168, "step": 418 }, { "epoch": 0.14, "grad_norm": 0.40101377095545854, "learning_rate": 2.9111598435010533e-05, "loss": 0.6675, "step": 419 }, { "epoch": 0.14, "grad_norm": 0.46519822186202325, "learning_rate": 2.9106211175198753e-05, "loss": 0.8914, "step": 420 }, { "epoch": 0.14, "grad_norm": 0.4829635653238317, "learning_rate": 2.9100808132109648e-05, "loss": 0.6492, "step": 421 }, { "epoch": 0.14, "grad_norm": 0.39528699537240736, "learning_rate": 2.9095389311788626e-05, "loss": 0.6405, "step": 422 }, { "epoch": 0.14, "grad_norm": 0.4860723642916951, "learning_rate": 2.908995472029874e-05, "loss": 0.9024, "step": 423 }, { "epoch": 0.14, "grad_norm": 0.37760462895818925, "learning_rate": 2.90845043637207e-05, "loss": 0.6372, "step": 424 }, { "epoch": 0.14, "grad_norm": 0.8162990565750706, "learning_rate": 2.9079038248152838e-05, "loss": 0.6486, "step": 425 }, { "epoch": 0.14, "grad_norm": 0.4379923280446208, "learning_rate": 2.907355637971113e-05, "loss": 0.9416, "step": 426 }, { "epoch": 0.14, "grad_norm": 0.45561186429996225, "learning_rate": 2.9068058764529172e-05, "loss": 0.6683, "step": 427 }, { "epoch": 0.14, "grad_norm": 0.42592583185547805, "learning_rate": 2.9062545408758193e-05, "loss": 0.6564, "step": 428 }, { "epoch": 0.14, "grad_norm": 0.5749829007567769, "learning_rate": 2.9057016318567015e-05, "loss": 0.704, "step": 429 }, { "epoch": 0.14, "grad_norm": 0.4364697417940015, "learning_rate": 2.9051471500142075e-05, "loss": 0.6375, "step": 430 }, { "epoch": 0.14, "grad_norm": 0.4882202597538823, "learning_rate": 2.904591095968741e-05, "loss": 0.9, "step": 431 }, { "epoch": 0.14, "grad_norm": 0.4332696292128969, "learning_rate": 2.9040334703424637e-05, "loss": 0.6292, "step": 432 }, { "epoch": 0.14, "grad_norm": 0.5540748189984196, "learning_rate": 2.9034742737592973e-05, "loss": 0.6539, "step": 433 }, { "epoch": 0.14, "grad_norm": 0.4431404727662558, "learning_rate": 2.90291350684492e-05, "loss": 0.8727, "step": 434 }, { "epoch": 0.14, "grad_norm": 0.47454857876889833, "learning_rate": 2.9023511702267672e-05, "loss": 0.6487, "step": 435 }, { "epoch": 0.14, "grad_norm": 0.526980118331581, "learning_rate": 2.9017872645340317e-05, "loss": 0.6506, "step": 436 }, { "epoch": 0.14, "grad_norm": 0.4072812171475702, "learning_rate": 2.9012217903976604e-05, "loss": 0.893, "step": 437 }, { "epoch": 0.14, "grad_norm": 0.49159634329042184, "learning_rate": 2.9006547484503557e-05, "loss": 0.6486, "step": 438 }, { "epoch": 0.14, "grad_norm": 0.44367393286624723, "learning_rate": 2.9000861393265748e-05, "loss": 0.6758, "step": 439 }, { "epoch": 0.14, "grad_norm": 0.4184686361024677, "learning_rate": 2.899515963662528e-05, "loss": 0.6334, "step": 440 }, { "epoch": 0.14, "grad_norm": 0.48413819250666285, "learning_rate": 2.8989442220961774e-05, "loss": 0.6388, "step": 441 }, { "epoch": 0.14, "grad_norm": 0.4206800782538987, "learning_rate": 2.8983709152672386e-05, "loss": 0.9098, "step": 442 }, { "epoch": 0.14, "grad_norm": 0.5718636110140484, "learning_rate": 2.8977960438171787e-05, "loss": 0.6432, "step": 443 }, { "epoch": 0.15, "grad_norm": 0.45692835451279196, "learning_rate": 2.8972196083892138e-05, "loss": 0.6449, "step": 444 }, { "epoch": 0.15, "grad_norm": 0.4843879013286376, "learning_rate": 2.896641609628311e-05, "loss": 0.8947, "step": 445 }, { "epoch": 0.15, "grad_norm": 0.5161351219819171, "learning_rate": 2.8960620481811866e-05, "loss": 0.6628, "step": 446 }, { "epoch": 0.15, "grad_norm": 0.4203967944928605, "learning_rate": 2.8954809246963058e-05, "loss": 0.6384, "step": 447 }, { "epoch": 0.15, "grad_norm": 0.45797649923830086, "learning_rate": 2.89489823982388e-05, "loss": 0.8821, "step": 448 }, { "epoch": 0.15, "grad_norm": 0.4654715477464315, "learning_rate": 2.894313994215869e-05, "loss": 0.6602, "step": 449 }, { "epoch": 0.15, "grad_norm": 0.5169683518764391, "learning_rate": 2.8937281885259784e-05, "loss": 0.6927, "step": 450 }, { "epoch": 0.15, "grad_norm": 0.4308887388881215, "learning_rate": 2.89314082340966e-05, "loss": 0.6463, "step": 451 }, { "epoch": 0.15, "grad_norm": 0.5073794265922195, "learning_rate": 2.892551899524109e-05, "loss": 0.9077, "step": 452 }, { "epoch": 0.15, "grad_norm": 0.40823761012348275, "learning_rate": 2.8919614175282662e-05, "loss": 0.6445, "step": 453 }, { "epoch": 0.15, "grad_norm": 0.4748031391747549, "learning_rate": 2.891369378082815e-05, "loss": 0.6441, "step": 454 }, { "epoch": 0.15, "grad_norm": 0.41740672319506367, "learning_rate": 2.890775781850181e-05, "loss": 0.6234, "step": 455 }, { "epoch": 0.15, "grad_norm": 0.5366537152878331, "learning_rate": 2.890180629494533e-05, "loss": 0.9208, "step": 456 }, { "epoch": 0.15, "grad_norm": 0.4066117200963689, "learning_rate": 2.8895839216817797e-05, "loss": 0.6426, "step": 457 }, { "epoch": 0.15, "grad_norm": 0.35110280780669384, "learning_rate": 2.8889856590795705e-05, "loss": 0.671, "step": 458 }, { "epoch": 0.15, "grad_norm": 0.45296524363812957, "learning_rate": 2.888385842357295e-05, "loss": 0.933, "step": 459 }, { "epoch": 0.15, "grad_norm": 0.37373894282109377, "learning_rate": 2.8877844721860804e-05, "loss": 0.6556, "step": 460 }, { "epoch": 0.15, "grad_norm": 0.4290088924076341, "learning_rate": 2.8871815492387938e-05, "loss": 0.6653, "step": 461 }, { "epoch": 0.15, "grad_norm": 0.39377732452218644, "learning_rate": 2.8865770741900382e-05, "loss": 0.6665, "step": 462 }, { "epoch": 0.15, "grad_norm": 0.4665538449785029, "learning_rate": 2.8859710477161535e-05, "loss": 0.9186, "step": 463 }, { "epoch": 0.15, "grad_norm": 0.42168233074526906, "learning_rate": 2.8853634704952167e-05, "loss": 0.6323, "step": 464 }, { "epoch": 0.15, "grad_norm": 0.41500586643604526, "learning_rate": 2.884754343207038e-05, "loss": 0.6474, "step": 465 }, { "epoch": 0.15, "grad_norm": 0.45855779407457153, "learning_rate": 2.8841436665331634e-05, "loss": 0.9257, "step": 466 }, { "epoch": 0.15, "grad_norm": 0.4376230820806902, "learning_rate": 2.883531441156872e-05, "loss": 0.641, "step": 467 }, { "epoch": 0.15, "grad_norm": 0.41106164975651804, "learning_rate": 2.8829176677631758e-05, "loss": 0.6532, "step": 468 }, { "epoch": 0.15, "grad_norm": 0.44380226347185603, "learning_rate": 2.8823023470388185e-05, "loss": 0.8802, "step": 469 }, { "epoch": 0.15, "grad_norm": 0.42301106527551297, "learning_rate": 2.8816854796722754e-05, "loss": 0.623, "step": 470 }, { "epoch": 0.15, "grad_norm": 0.45524654004853865, "learning_rate": 2.8810670663537523e-05, "loss": 0.6674, "step": 471 }, { "epoch": 0.15, "grad_norm": 0.45966465374401083, "learning_rate": 2.8804471077751847e-05, "loss": 0.7029, "step": 472 }, { "epoch": 0.15, "grad_norm": 0.38889047296670676, "learning_rate": 2.8798256046302376e-05, "loss": 0.6538, "step": 473 }, { "epoch": 0.15, "grad_norm": 0.5238902268443127, "learning_rate": 2.8792025576143032e-05, "loss": 0.9276, "step": 474 }, { "epoch": 0.16, "grad_norm": 0.37969637405918605, "learning_rate": 2.878577967424502e-05, "loss": 0.6407, "step": 475 }, { "epoch": 0.16, "grad_norm": 0.4362413268972957, "learning_rate": 2.8779518347596803e-05, "loss": 0.6204, "step": 476 }, { "epoch": 0.16, "grad_norm": 0.4339756058034954, "learning_rate": 2.877324160320411e-05, "loss": 0.9098, "step": 477 }, { "epoch": 0.16, "grad_norm": 0.545218647519162, "learning_rate": 2.8766949448089917e-05, "loss": 0.6623, "step": 478 }, { "epoch": 0.16, "grad_norm": 0.42822620377471166, "learning_rate": 2.8760641889294453e-05, "loss": 0.6377, "step": 479 }, { "epoch": 0.16, "grad_norm": 0.4161711166634251, "learning_rate": 2.875431893387516e-05, "loss": 0.6454, "step": 480 }, { "epoch": 0.16, "grad_norm": 0.4179496071349973, "learning_rate": 2.874798058890672e-05, "loss": 0.9152, "step": 481 }, { "epoch": 0.16, "grad_norm": 0.42566660183162713, "learning_rate": 2.8741626861481043e-05, "loss": 0.6272, "step": 482 }, { "epoch": 0.16, "grad_norm": 0.38882721317690294, "learning_rate": 2.873525775870724e-05, "loss": 0.6273, "step": 483 }, { "epoch": 0.16, "grad_norm": 0.38320802678932875, "learning_rate": 2.8728873287711622e-05, "loss": 0.6383, "step": 484 }, { "epoch": 0.16, "grad_norm": 0.5094243063088904, "learning_rate": 2.8722473455637702e-05, "loss": 0.8724, "step": 485 }, { "epoch": 0.16, "grad_norm": 0.3688382342507443, "learning_rate": 2.8716058269646182e-05, "loss": 0.6203, "step": 486 }, { "epoch": 0.16, "grad_norm": 0.47364848694246114, "learning_rate": 2.8709627736914938e-05, "loss": 0.6493, "step": 487 }, { "epoch": 0.16, "grad_norm": 0.42897235801740585, "learning_rate": 2.8703181864639013e-05, "loss": 0.9061, "step": 488 }, { "epoch": 0.16, "grad_norm": 0.4803849911476859, "learning_rate": 2.8696720660030628e-05, "loss": 0.7073, "step": 489 }, { "epoch": 0.16, "grad_norm": 0.4762116354435143, "learning_rate": 2.8690244130319145e-05, "loss": 0.6542, "step": 490 }, { "epoch": 0.16, "grad_norm": 0.3807043750384292, "learning_rate": 2.868375228275108e-05, "loss": 0.6348, "step": 491 }, { "epoch": 0.16, "grad_norm": 0.42237052621735177, "learning_rate": 2.8677245124590087e-05, "loss": 0.8782, "step": 492 }, { "epoch": 0.16, "grad_norm": 0.3739349383934636, "learning_rate": 2.867072266311695e-05, "loss": 0.6435, "step": 493 }, { "epoch": 0.16, "grad_norm": 0.5032988633011936, "learning_rate": 2.8664184905629577e-05, "loss": 0.6417, "step": 494 }, { "epoch": 0.16, "grad_norm": 0.4286711101017065, "learning_rate": 2.865763185944299e-05, "loss": 0.8952, "step": 495 }, { "epoch": 0.16, "grad_norm": 0.3677267340927617, "learning_rate": 2.865106353188931e-05, "loss": 0.6035, "step": 496 }, { "epoch": 0.16, "grad_norm": 0.45636072223170315, "learning_rate": 2.8644479930317776e-05, "loss": 0.6306, "step": 497 }, { "epoch": 0.16, "grad_norm": 0.40855865528120483, "learning_rate": 2.863788106209469e-05, "loss": 0.9089, "step": 498 }, { "epoch": 0.16, "grad_norm": 0.45165859181470097, "learning_rate": 2.8631266934603467e-05, "loss": 0.6305, "step": 499 }, { "epoch": 0.16, "grad_norm": 0.5495855496581937, "learning_rate": 2.8624637555244556e-05, "loss": 0.7096, "step": 500 }, { "epoch": 0.16, "grad_norm": 0.3708244699621869, "learning_rate": 2.8617992931435512e-05, "loss": 0.647, "step": 501 }, { "epoch": 0.16, "grad_norm": 0.4588068723669083, "learning_rate": 2.8611333070610918e-05, "loss": 0.6354, "step": 502 }, { "epoch": 0.16, "grad_norm": 0.40063728360176754, "learning_rate": 2.860465798022242e-05, "loss": 0.8946, "step": 503 }, { "epoch": 0.16, "grad_norm": 0.4527565039834011, "learning_rate": 2.85979676677387e-05, "loss": 0.6218, "step": 504 }, { "epoch": 0.16, "grad_norm": 0.5066153786835709, "learning_rate": 2.8591262140645465e-05, "loss": 0.6544, "step": 505 }, { "epoch": 0.17, "grad_norm": 0.4448467235297304, "learning_rate": 2.8584541406445462e-05, "loss": 0.9152, "step": 506 }, { "epoch": 0.17, "grad_norm": 0.47495446658220275, "learning_rate": 2.8577805472658435e-05, "loss": 0.6335, "step": 507 }, { "epoch": 0.17, "grad_norm": 0.39786575254721246, "learning_rate": 2.8571054346821153e-05, "loss": 0.635, "step": 508 }, { "epoch": 0.17, "grad_norm": 0.47823610613167383, "learning_rate": 2.856428803648736e-05, "loss": 0.9018, "step": 509 }, { "epoch": 0.17, "grad_norm": 0.41455938783060314, "learning_rate": 2.855750654922781e-05, "loss": 0.6363, "step": 510 }, { "epoch": 0.17, "grad_norm": 0.4200962343345511, "learning_rate": 2.8550709892630227e-05, "loss": 0.6427, "step": 511 }, { "epoch": 0.17, "grad_norm": 0.40593767474446596, "learning_rate": 2.8543898074299322e-05, "loss": 0.6505, "step": 512 }, { "epoch": 0.17, "grad_norm": 0.408473999989803, "learning_rate": 2.8537071101856744e-05, "loss": 0.6502, "step": 513 }, { "epoch": 0.17, "grad_norm": 0.4150321572323016, "learning_rate": 2.853022898294113e-05, "loss": 0.8814, "step": 514 }, { "epoch": 0.17, "grad_norm": 0.5451569292059286, "learning_rate": 2.852337172520804e-05, "loss": 0.6267, "step": 515 }, { "epoch": 0.17, "grad_norm": 0.4092894427224817, "learning_rate": 2.8516499336329974e-05, "loss": 0.623, "step": 516 }, { "epoch": 0.17, "grad_norm": 0.4378351402894465, "learning_rate": 2.8509611823996384e-05, "loss": 0.9064, "step": 517 }, { "epoch": 0.17, "grad_norm": 0.4102731295478398, "learning_rate": 2.8502709195913617e-05, "loss": 0.6459, "step": 518 }, { "epoch": 0.17, "grad_norm": 0.38188227137307235, "learning_rate": 2.8495791459804952e-05, "loss": 0.6161, "step": 519 }, { "epoch": 0.17, "grad_norm": 0.45685512278087587, "learning_rate": 2.848885862341056e-05, "loss": 0.882, "step": 520 }, { "epoch": 0.17, "grad_norm": 0.5985408933430514, "learning_rate": 2.8481910694487507e-05, "loss": 0.6383, "step": 521 }, { "epoch": 0.17, "grad_norm": 0.463600224761772, "learning_rate": 2.8474947680809754e-05, "loss": 0.6807, "step": 522 }, { "epoch": 0.17, "grad_norm": 0.3828734005797913, "learning_rate": 2.8467969590168144e-05, "loss": 0.6579, "step": 523 }, { "epoch": 0.17, "grad_norm": 0.411140136149198, "learning_rate": 2.8460976430370375e-05, "loss": 0.8869, "step": 524 }, { "epoch": 0.17, "grad_norm": 0.3806441813500094, "learning_rate": 2.845396820924102e-05, "loss": 0.6222, "step": 525 }, { "epoch": 0.17, "grad_norm": 0.3850543528981122, "learning_rate": 2.844694493462148e-05, "loss": 0.6112, "step": 526 }, { "epoch": 0.17, "grad_norm": 0.5204941375381967, "learning_rate": 2.8439906614370037e-05, "loss": 0.6288, "step": 527 }, { "epoch": 0.17, "grad_norm": 0.4942822908100749, "learning_rate": 2.843285325636178e-05, "loss": 0.9311, "step": 528 }, { "epoch": 0.17, "grad_norm": 0.4112172723684616, "learning_rate": 2.8425784868488618e-05, "loss": 0.6379, "step": 529 }, { "epoch": 0.17, "grad_norm": 0.39653217806105395, "learning_rate": 2.8418701458659307e-05, "loss": 0.6168, "step": 530 }, { "epoch": 0.17, "grad_norm": 0.4182272526017336, "learning_rate": 2.841160303479938e-05, "loss": 0.8952, "step": 531 }, { "epoch": 0.17, "grad_norm": 0.4047375803682781, "learning_rate": 2.8404489604851186e-05, "loss": 0.6376, "step": 532 }, { "epoch": 0.17, "grad_norm": 0.45944482758325844, "learning_rate": 2.839736117677386e-05, "loss": 0.6827, "step": 533 }, { "epoch": 0.17, "grad_norm": 0.4126803649153416, "learning_rate": 2.8390217758543315e-05, "loss": 0.6566, "step": 534 }, { "epoch": 0.17, "grad_norm": 0.554871963799865, "learning_rate": 2.8383059358152246e-05, "loss": 0.9044, "step": 535 }, { "epoch": 0.18, "grad_norm": 0.4298320285054984, "learning_rate": 2.83758859836101e-05, "loss": 0.6197, "step": 536 }, { "epoch": 0.18, "grad_norm": 0.41590018113511795, "learning_rate": 2.8368697642943083e-05, "loss": 0.6541, "step": 537 }, { "epoch": 0.18, "grad_norm": 0.47270251418481024, "learning_rate": 2.8361494344194152e-05, "loss": 0.8943, "step": 538 }, { "epoch": 0.18, "grad_norm": 0.38431596183925654, "learning_rate": 2.8354276095422986e-05, "loss": 0.6382, "step": 539 }, { "epoch": 0.18, "grad_norm": 0.5526987731465783, "learning_rate": 2.834704290470601e-05, "loss": 0.6403, "step": 540 }, { "epoch": 0.18, "grad_norm": 0.4243534537765687, "learning_rate": 2.8339794780136348e-05, "loss": 0.8928, "step": 541 }, { "epoch": 0.18, "grad_norm": 0.3787257004630748, "learning_rate": 2.8332531729823853e-05, "loss": 0.6443, "step": 542 }, { "epoch": 0.18, "grad_norm": 0.43645410057487805, "learning_rate": 2.8325253761895063e-05, "loss": 0.6508, "step": 543 }, { "epoch": 0.18, "grad_norm": 0.40984827908785904, "learning_rate": 2.8317960884493213e-05, "loss": 0.6699, "step": 544 }, { "epoch": 0.18, "grad_norm": 0.3903190903824307, "learning_rate": 2.8310653105778218e-05, "loss": 0.6395, "step": 545 }, { "epoch": 0.18, "grad_norm": 0.3869810328579648, "learning_rate": 2.830333043392667e-05, "loss": 0.9189, "step": 546 }, { "epoch": 0.18, "grad_norm": 0.3750522028733386, "learning_rate": 2.8295992877131826e-05, "loss": 0.6499, "step": 547 }, { "epoch": 0.18, "grad_norm": 0.4128264072860223, "learning_rate": 2.8288640443603587e-05, "loss": 0.6406, "step": 548 }, { "epoch": 0.18, "grad_norm": 0.3943291470917676, "learning_rate": 2.8281273141568505e-05, "loss": 0.6838, "step": 549 }, { "epoch": 0.18, "grad_norm": 0.4102240004100346, "learning_rate": 2.827389097926978e-05, "loss": 0.7975, "step": 550 }, { "epoch": 0.18, "grad_norm": 0.40487965346999183, "learning_rate": 2.8266493964967213e-05, "loss": 0.6458, "step": 551 }, { "epoch": 0.18, "grad_norm": 0.4050345942903766, "learning_rate": 2.8259082106937255e-05, "loss": 0.8951, "step": 552 }, { "epoch": 0.18, "grad_norm": 0.3962830619964766, "learning_rate": 2.825165541347294e-05, "loss": 0.6654, "step": 553 }, { "epoch": 0.18, "grad_norm": 0.5051884041976252, "learning_rate": 2.8244213892883907e-05, "loss": 0.6472, "step": 554 }, { "epoch": 0.18, "grad_norm": 0.4108314761956005, "learning_rate": 2.823675755349639e-05, "loss": 0.6808, "step": 555 }, { "epoch": 0.18, "grad_norm": 0.4474446134404235, "learning_rate": 2.82292864036532e-05, "loss": 0.6249, "step": 556 }, { "epoch": 0.18, "grad_norm": 0.4432194339248852, "learning_rate": 2.822180045171373e-05, "loss": 0.8686, "step": 557 }, { "epoch": 0.18, "grad_norm": 0.48098525582934654, "learning_rate": 2.8214299706053918e-05, "loss": 0.6529, "step": 558 }, { "epoch": 0.18, "grad_norm": 0.5282283048600401, "learning_rate": 2.820678417506626e-05, "loss": 0.6531, "step": 559 }, { "epoch": 0.18, "grad_norm": 0.4209585216782718, "learning_rate": 2.81992538671598e-05, "loss": 0.704, "step": 560 }, { "epoch": 0.18, "grad_norm": 0.4526190787349011, "learning_rate": 2.8191708790760112e-05, "loss": 0.8942, "step": 561 }, { "epoch": 0.18, "grad_norm": 0.4256726501657445, "learning_rate": 2.8184148954309295e-05, "loss": 0.623, "step": 562 }, { "epoch": 0.18, "grad_norm": 0.3916404649617502, "learning_rate": 2.8176574366265963e-05, "loss": 0.6373, "step": 563 }, { "epoch": 0.18, "grad_norm": 0.4762397266757786, "learning_rate": 2.8168985035105236e-05, "loss": 0.9082, "step": 564 }, { "epoch": 0.18, "grad_norm": 0.38092462849036834, "learning_rate": 2.8161380969318728e-05, "loss": 0.6303, "step": 565 }, { "epoch": 0.18, "grad_norm": 0.5660056286500637, "learning_rate": 2.8153762177414545e-05, "loss": 0.6261, "step": 566 }, { "epoch": 0.19, "grad_norm": 0.45577713270287157, "learning_rate": 2.8146128667917258e-05, "loss": 0.9081, "step": 567 }, { "epoch": 0.19, "grad_norm": 0.4954672534798452, "learning_rate": 2.8138480449367923e-05, "loss": 0.6552, "step": 568 }, { "epoch": 0.19, "grad_norm": 0.5343867095964945, "learning_rate": 2.8130817530324033e-05, "loss": 0.639, "step": 569 }, { "epoch": 0.19, "grad_norm": 0.3841053908353787, "learning_rate": 2.812313991935955e-05, "loss": 0.6371, "step": 570 }, { "epoch": 0.19, "grad_norm": 0.46943893436889167, "learning_rate": 2.8115447625064857e-05, "loss": 0.6771, "step": 571 }, { "epoch": 0.19, "grad_norm": 0.43228939094307395, "learning_rate": 2.8107740656046775e-05, "loss": 0.9028, "step": 572 }, { "epoch": 0.19, "grad_norm": 0.5464684294544326, "learning_rate": 2.810001902092854e-05, "loss": 0.6405, "step": 573 }, { "epoch": 0.19, "grad_norm": 0.4762009598466699, "learning_rate": 2.8092282728349817e-05, "loss": 0.6461, "step": 574 }, { "epoch": 0.19, "grad_norm": 0.4220077471846214, "learning_rate": 2.8084531786966632e-05, "loss": 0.9197, "step": 575 }, { "epoch": 0.19, "grad_norm": 0.4207425645490711, "learning_rate": 2.8076766205451435e-05, "loss": 0.6376, "step": 576 }, { "epoch": 0.19, "grad_norm": 0.39841546888581675, "learning_rate": 2.806898599249305e-05, "loss": 0.6265, "step": 577 }, { "epoch": 0.19, "grad_norm": 0.418575088332761, "learning_rate": 2.8061191156796658e-05, "loss": 0.9159, "step": 578 }, { "epoch": 0.19, "grad_norm": 0.38180714613831823, "learning_rate": 2.805338170708382e-05, "loss": 0.6326, "step": 579 }, { "epoch": 0.19, "grad_norm": 0.408679644533415, "learning_rate": 2.804555765209243e-05, "loss": 0.6376, "step": 580 }, { "epoch": 0.19, "grad_norm": 0.4058244933329194, "learning_rate": 2.8037719000576744e-05, "loss": 0.893, "step": 581 }, { "epoch": 0.19, "grad_norm": 0.43257064486873414, "learning_rate": 2.802986576130733e-05, "loss": 0.4491, "step": 582 }, { "epoch": 0.19, "grad_norm": 0.4493022364930085, "learning_rate": 2.802199794307109e-05, "loss": 0.8842, "step": 583 }, { "epoch": 0.19, "grad_norm": 0.4719106605724196, "learning_rate": 2.801411555467124e-05, "loss": 0.642, "step": 584 }, { "epoch": 0.19, "grad_norm": 0.4895803268394636, "learning_rate": 2.8006218604927284e-05, "loss": 0.6384, "step": 585 }, { "epoch": 0.19, "grad_norm": 0.5505377387135137, "learning_rate": 2.7998307102675035e-05, "loss": 0.9044, "step": 586 }, { "epoch": 0.19, "grad_norm": 0.47964174370655516, "learning_rate": 2.7990381056766583e-05, "loss": 0.6304, "step": 587 }, { "epoch": 0.19, "grad_norm": 0.40407867618439774, "learning_rate": 2.798244047607028e-05, "loss": 0.6147, "step": 588 }, { "epoch": 0.19, "grad_norm": 0.5849967733428494, "learning_rate": 2.7974485369470764e-05, "loss": 0.8614, "step": 589 }, { "epoch": 0.19, "grad_norm": 0.4646422605512494, "learning_rate": 2.7966515745868904e-05, "loss": 0.619, "step": 590 }, { "epoch": 0.19, "grad_norm": 0.4351191711557609, "learning_rate": 2.795853161418182e-05, "loss": 0.6375, "step": 591 }, { "epoch": 0.19, "grad_norm": 0.47575304222703524, "learning_rate": 2.7950532983342863e-05, "loss": 0.8998, "step": 592 }, { "epoch": 0.19, "grad_norm": 0.5352707221394369, "learning_rate": 2.7942519862301618e-05, "loss": 0.6872, "step": 593 }, { "epoch": 0.19, "grad_norm": 0.3913669721527101, "learning_rate": 2.7934492260023865e-05, "loss": 0.6284, "step": 594 }, { "epoch": 0.19, "grad_norm": 0.42277338060816, "learning_rate": 2.79264501854916e-05, "loss": 0.6318, "step": 595 }, { "epoch": 0.19, "grad_norm": 0.44352479056458044, "learning_rate": 2.7918393647703013e-05, "loss": 0.9112, "step": 596 }, { "epoch": 0.19, "grad_norm": 0.4148469723095011, "learning_rate": 2.7910322655672456e-05, "loss": 0.6354, "step": 597 }, { "epoch": 0.2, "grad_norm": 0.4210615103936188, "learning_rate": 2.7902237218430485e-05, "loss": 0.6292, "step": 598 }, { "epoch": 0.2, "grad_norm": 0.4144251980286665, "learning_rate": 2.7894137345023786e-05, "loss": 0.6213, "step": 599 }, { "epoch": 0.2, "grad_norm": 0.4910461081283438, "learning_rate": 2.7886023044515234e-05, "loss": 0.8709, "step": 600 }, { "epoch": 0.2, "grad_norm": 0.36234214323717867, "learning_rate": 2.7877894325983814e-05, "loss": 0.6213, "step": 601 }, { "epoch": 0.2, "grad_norm": 1.2169705101386594, "learning_rate": 2.7869751198524656e-05, "loss": 0.6359, "step": 602 }, { "epoch": 0.2, "grad_norm": 0.4504305495984645, "learning_rate": 2.7861593671249018e-05, "loss": 0.9053, "step": 603 }, { "epoch": 0.2, "grad_norm": 0.42941898351719776, "learning_rate": 2.7853421753284253e-05, "loss": 0.6723, "step": 604 }, { "epoch": 0.2, "grad_norm": 0.3980402692972693, "learning_rate": 2.7845235453773837e-05, "loss": 0.6475, "step": 605 }, { "epoch": 0.2, "grad_norm": 0.3436123286173719, "learning_rate": 2.7837034781877317e-05, "loss": 0.629, "step": 606 }, { "epoch": 0.2, "grad_norm": 0.43095096597775956, "learning_rate": 2.7828819746770337e-05, "loss": 0.8882, "step": 607 }, { "epoch": 0.2, "grad_norm": 0.3752657273131738, "learning_rate": 2.7820590357644604e-05, "loss": 0.6296, "step": 608 }, { "epoch": 0.2, "grad_norm": 0.43222723386912415, "learning_rate": 2.7812346623707887e-05, "loss": 0.635, "step": 609 }, { "epoch": 0.2, "grad_norm": 0.8336642135656587, "learning_rate": 2.7804088554184002e-05, "loss": 0.8762, "step": 610 }, { "epoch": 0.2, "grad_norm": 0.4623730580869914, "learning_rate": 2.7795816158312805e-05, "loss": 0.6266, "step": 611 }, { "epoch": 0.2, "grad_norm": 0.417728305624967, "learning_rate": 2.7787529445350192e-05, "loss": 0.6292, "step": 612 }, { "epoch": 0.2, "grad_norm": 0.40520305061617656, "learning_rate": 2.777922842456807e-05, "loss": 0.8553, "step": 613 }, { "epoch": 0.2, "grad_norm": 0.437202074399463, "learning_rate": 2.7770913105254355e-05, "loss": 0.6152, "step": 614 }, { "epoch": 0.2, "grad_norm": 0.40485729539777116, "learning_rate": 2.776258349671296e-05, "loss": 0.7027, "step": 615 }, { "epoch": 0.2, "grad_norm": 0.43958041615771876, "learning_rate": 2.7754239608263785e-05, "loss": 0.6247, "step": 616 }, { "epoch": 0.2, "grad_norm": 0.440145950347491, "learning_rate": 2.7745881449242717e-05, "loss": 0.6355, "step": 617 }, { "epoch": 0.2, "grad_norm": 0.4440697999662624, "learning_rate": 2.77375090290016e-05, "loss": 0.9176, "step": 618 }, { "epoch": 0.2, "grad_norm": 0.41771368593007563, "learning_rate": 2.7729122356908244e-05, "loss": 0.6295, "step": 619 }, { "epoch": 0.2, "grad_norm": 0.4379977397533082, "learning_rate": 2.772072144234639e-05, "loss": 0.6272, "step": 620 }, { "epoch": 0.2, "grad_norm": 0.42836163067626637, "learning_rate": 2.7712306294715734e-05, "loss": 0.932, "step": 621 }, { "epoch": 0.2, "grad_norm": 0.4394409777771525, "learning_rate": 2.7703876923431882e-05, "loss": 0.6293, "step": 622 }, { "epoch": 0.2, "grad_norm": 0.35111563292685327, "learning_rate": 2.7695433337926362e-05, "loss": 0.6503, "step": 623 }, { "epoch": 0.2, "grad_norm": 0.39826235372841107, "learning_rate": 2.7686975547646614e-05, "loss": 0.8764, "step": 624 }, { "epoch": 0.2, "grad_norm": 0.3698676116700673, "learning_rate": 2.7678503562055947e-05, "loss": 0.6179, "step": 625 }, { "epoch": 0.2, "grad_norm": 0.4685847476973362, "learning_rate": 2.7670017390633578e-05, "loss": 0.6908, "step": 626 }, { "epoch": 0.2, "grad_norm": 0.3647973146958208, "learning_rate": 2.766151704287458e-05, "loss": 0.6428, "step": 627 }, { "epoch": 0.21, "grad_norm": 0.3767603760583791, "learning_rate": 2.7653002528289895e-05, "loss": 0.6323, "step": 628 }, { "epoch": 0.21, "grad_norm": 0.4026962137754299, "learning_rate": 2.7644473856406324e-05, "loss": 0.8795, "step": 629 }, { "epoch": 0.21, "grad_norm": 1.7612396095860157, "learning_rate": 2.763593103676649e-05, "loss": 0.6511, "step": 630 }, { "epoch": 0.21, "grad_norm": 0.38981157630131313, "learning_rate": 2.7627374078928862e-05, "loss": 0.6268, "step": 631 }, { "epoch": 0.21, "grad_norm": 0.39791452704998903, "learning_rate": 2.7618802992467718e-05, "loss": 0.6618, "step": 632 }, { "epoch": 0.21, "grad_norm": 0.3988229391485032, "learning_rate": 2.761021778697315e-05, "loss": 0.835, "step": 633 }, { "epoch": 0.21, "grad_norm": 0.37871530747429055, "learning_rate": 2.7601618472051044e-05, "loss": 0.6284, "step": 634 }, { "epoch": 0.21, "grad_norm": 0.4253524596324212, "learning_rate": 2.7593005057323072e-05, "loss": 0.9342, "step": 635 }, { "epoch": 0.21, "grad_norm": 0.5303452899595816, "learning_rate": 2.7584377552426692e-05, "loss": 0.624, "step": 636 }, { "epoch": 0.21, "grad_norm": 0.3984405794439977, "learning_rate": 2.7575735967015113e-05, "loss": 0.6987, "step": 637 }, { "epoch": 0.21, "grad_norm": 0.39616164406530896, "learning_rate": 2.756708031075731e-05, "loss": 0.6194, "step": 638 }, { "epoch": 0.21, "grad_norm": 0.46974687971676377, "learning_rate": 2.755841059333799e-05, "loss": 0.9078, "step": 639 }, { "epoch": 0.21, "grad_norm": 0.3971680710514997, "learning_rate": 2.7549726824457614e-05, "loss": 0.6444, "step": 640 }, { "epoch": 0.21, "grad_norm": 0.4101678560397066, "learning_rate": 2.7541029013832338e-05, "loss": 0.6298, "step": 641 }, { "epoch": 0.21, "grad_norm": 0.402283718779502, "learning_rate": 2.753231717119405e-05, "loss": 0.6416, "step": 642 }, { "epoch": 0.21, "grad_norm": 0.40406497219707127, "learning_rate": 2.752359130629032e-05, "loss": 0.6725, "step": 643 }, { "epoch": 0.21, "grad_norm": 0.46628523465104504, "learning_rate": 2.7514851428884432e-05, "loss": 0.902, "step": 644 }, { "epoch": 0.21, "grad_norm": 0.367765039356584, "learning_rate": 2.7506097548755322e-05, "loss": 0.627, "step": 645 }, { "epoch": 0.21, "grad_norm": 0.4638008773558133, "learning_rate": 2.749732967569761e-05, "loss": 0.6238, "step": 646 }, { "epoch": 0.21, "grad_norm": 0.43791893044788804, "learning_rate": 2.748854781952157e-05, "loss": 0.9151, "step": 647 }, { "epoch": 0.21, "grad_norm": 0.448967080894844, "learning_rate": 2.747975199005312e-05, "loss": 0.6419, "step": 648 }, { "epoch": 0.21, "grad_norm": 0.6431376990160488, "learning_rate": 2.7470942197133813e-05, "loss": 0.6284, "step": 649 }, { "epoch": 0.21, "grad_norm": 0.4256844678943661, "learning_rate": 2.7462118450620818e-05, "loss": 0.8757, "step": 650 }, { "epoch": 0.21, "grad_norm": 0.47417492137159173, "learning_rate": 2.7453280760386934e-05, "loss": 0.6486, "step": 651 }, { "epoch": 0.21, "grad_norm": 0.42990520601826226, "learning_rate": 2.744442913632054e-05, "loss": 0.6317, "step": 652 }, { "epoch": 0.21, "grad_norm": 0.4018477887643915, "learning_rate": 2.7435563588325627e-05, "loss": 0.8664, "step": 653 }, { "epoch": 0.21, "grad_norm": 0.4510309670821486, "learning_rate": 2.7426684126321743e-05, "loss": 0.4416, "step": 654 }, { "epoch": 0.21, "grad_norm": 0.4580823302828529, "learning_rate": 2.741779076024403e-05, "loss": 0.8991, "step": 655 }, { "epoch": 0.21, "grad_norm": 0.4345294521404785, "learning_rate": 2.7408883500043157e-05, "loss": 0.6281, "step": 656 }, { "epoch": 0.21, "grad_norm": 0.48117715079103457, "learning_rate": 2.739996235568537e-05, "loss": 0.6195, "step": 657 }, { "epoch": 0.21, "grad_norm": 0.4176774358860328, "learning_rate": 2.7391027337152425e-05, "loss": 0.8817, "step": 658 }, { "epoch": 0.22, "grad_norm": 0.4464922252688219, "learning_rate": 2.738207845444161e-05, "loss": 0.6312, "step": 659 }, { "epoch": 0.22, "grad_norm": 0.4306437682432233, "learning_rate": 2.7373115717565735e-05, "loss": 0.6501, "step": 660 }, { "epoch": 0.22, "grad_norm": 0.43758729261006385, "learning_rate": 2.7364139136553096e-05, "loss": 0.8954, "step": 661 }, { "epoch": 0.22, "grad_norm": 0.40008132738862084, "learning_rate": 2.7355148721447492e-05, "loss": 0.622, "step": 662 }, { "epoch": 0.22, "grad_norm": 0.3885324303659026, "learning_rate": 2.734614448230819e-05, "loss": 0.6213, "step": 663 }, { "epoch": 0.22, "grad_norm": 0.4431412594594135, "learning_rate": 2.7337126429209935e-05, "loss": 0.6737, "step": 664 }, { "epoch": 0.22, "grad_norm": 0.4606344390351968, "learning_rate": 2.732809457224292e-05, "loss": 0.9034, "step": 665 }, { "epoch": 0.22, "grad_norm": 0.3878833376351238, "learning_rate": 2.731904892151278e-05, "loss": 0.6253, "step": 666 }, { "epoch": 0.22, "grad_norm": 0.3741553150704403, "learning_rate": 2.73099894871406e-05, "loss": 0.6259, "step": 667 }, { "epoch": 0.22, "grad_norm": 0.37389782312433234, "learning_rate": 2.7300916279262868e-05, "loss": 0.6122, "step": 668 }, { "epoch": 0.22, "grad_norm": 0.4217996547535171, "learning_rate": 2.7291829308031495e-05, "loss": 0.9028, "step": 669 }, { "epoch": 0.22, "grad_norm": 0.4382684932656845, "learning_rate": 2.7282728583613794e-05, "loss": 0.6405, "step": 670 }, { "epoch": 0.22, "grad_norm": 0.41599273244864077, "learning_rate": 2.727361411619245e-05, "loss": 0.632, "step": 671 }, { "epoch": 0.22, "grad_norm": 0.4198254448750625, "learning_rate": 2.7264485915965548e-05, "loss": 0.8618, "step": 672 }, { "epoch": 0.22, "grad_norm": 0.3970547990319387, "learning_rate": 2.725534399314651e-05, "loss": 0.6303, "step": 673 }, { "epoch": 0.22, "grad_norm": 0.34257120272875896, "learning_rate": 2.7246188357964147e-05, "loss": 0.6138, "step": 674 }, { "epoch": 0.22, "grad_norm": 0.4205961573418774, "learning_rate": 2.7237019020662585e-05, "loss": 0.6673, "step": 675 }, { "epoch": 0.22, "grad_norm": 0.43888113404954576, "learning_rate": 2.7227835991501284e-05, "loss": 0.8911, "step": 676 }, { "epoch": 0.22, "grad_norm": 0.3829281172195052, "learning_rate": 2.7218639280755037e-05, "loss": 0.6321, "step": 677 }, { "epoch": 0.22, "grad_norm": 0.3871490530689638, "learning_rate": 2.7209428898713935e-05, "loss": 0.5995, "step": 678 }, { "epoch": 0.22, "grad_norm": 0.40639441717895425, "learning_rate": 2.720020485568337e-05, "loss": 0.9092, "step": 679 }, { "epoch": 0.22, "grad_norm": 0.39174337438978485, "learning_rate": 2.719096716198402e-05, "loss": 0.6466, "step": 680 }, { "epoch": 0.22, "grad_norm": 0.35016986725307203, "learning_rate": 2.718171582795183e-05, "loss": 0.6125, "step": 681 }, { "epoch": 0.22, "grad_norm": 0.42655222096032175, "learning_rate": 2.717245086393801e-05, "loss": 0.9087, "step": 682 }, { "epoch": 0.22, "grad_norm": 0.35461476443038265, "learning_rate": 2.7163172280309028e-05, "loss": 0.6093, "step": 683 }, { "epoch": 0.22, "grad_norm": 0.38564093356550855, "learning_rate": 2.715388008744658e-05, "loss": 0.645, "step": 684 }, { "epoch": 0.22, "grad_norm": 0.3897838052243748, "learning_rate": 2.7144574295747594e-05, "loss": 0.8858, "step": 685 }, { "epoch": 0.22, "grad_norm": 0.4019096810104857, "learning_rate": 2.7135254915624213e-05, "loss": 0.4302, "step": 686 }, { "epoch": 0.22, "grad_norm": 0.3894344456275822, "learning_rate": 2.7125921957503785e-05, "loss": 0.866, "step": 687 }, { "epoch": 0.22, "grad_norm": 0.3919833305862916, "learning_rate": 2.7116575431828844e-05, "loss": 0.6361, "step": 688 }, { "epoch": 0.23, "grad_norm": 0.36010842873621013, "learning_rate": 2.710721534905712e-05, "loss": 0.6009, "step": 689 }, { "epoch": 0.23, "grad_norm": 0.3925010704394055, "learning_rate": 2.7097841719661498e-05, "loss": 0.8998, "step": 690 }, { "epoch": 0.23, "grad_norm": 0.3703827714267066, "learning_rate": 2.708845455413002e-05, "loss": 0.6327, "step": 691 }, { "epoch": 0.23, "grad_norm": 0.4000679796662479, "learning_rate": 2.707905386296588e-05, "loss": 0.6029, "step": 692 }, { "epoch": 0.23, "grad_norm": 0.40682891977349317, "learning_rate": 2.70696396566874e-05, "loss": 0.8945, "step": 693 }, { "epoch": 0.23, "grad_norm": 0.39895869166731196, "learning_rate": 2.706021194582804e-05, "loss": 0.601, "step": 694 }, { "epoch": 0.23, "grad_norm": 0.4139780175017376, "learning_rate": 2.7050770740936338e-05, "loss": 0.6231, "step": 695 }, { "epoch": 0.23, "grad_norm": 0.4160913759064976, "learning_rate": 2.7041316052575967e-05, "loss": 0.8742, "step": 696 }, { "epoch": 0.23, "grad_norm": 0.4139135333780499, "learning_rate": 2.7031847891325658e-05, "loss": 0.4381, "step": 697 }, { "epoch": 0.23, "grad_norm": 0.4311492634828156, "learning_rate": 2.702236626777923e-05, "loss": 0.8857, "step": 698 }, { "epoch": 0.23, "grad_norm": 0.5189788570420679, "learning_rate": 2.7012871192545565e-05, "loss": 0.642, "step": 699 }, { "epoch": 0.23, "grad_norm": 0.4414794941822095, "learning_rate": 2.700336267624859e-05, "loss": 0.6136, "step": 700 }, { "epoch": 0.23, "grad_norm": 0.4406469836754748, "learning_rate": 2.6993840729527273e-05, "loss": 0.8923, "step": 701 }, { "epoch": 0.23, "grad_norm": 0.3962143615563923, "learning_rate": 2.6984305363035616e-05, "loss": 0.6394, "step": 702 }, { "epoch": 0.23, "grad_norm": 0.42714576825197087, "learning_rate": 2.6974756587442636e-05, "loss": 0.6258, "step": 703 }, { "epoch": 0.23, "grad_norm": 0.4056618564655204, "learning_rate": 2.696519441343233e-05, "loss": 0.9036, "step": 704 }, { "epoch": 0.23, "grad_norm": 0.4047360594005986, "learning_rate": 2.695561885170372e-05, "loss": 0.6026, "step": 705 }, { "epoch": 0.23, "grad_norm": 0.4140062519647793, "learning_rate": 2.6946029912970785e-05, "loss": 0.6284, "step": 706 }, { "epoch": 0.23, "grad_norm": 0.3933680886320087, "learning_rate": 2.6936427607962484e-05, "loss": 0.8939, "step": 707 }, { "epoch": 0.23, "grad_norm": 0.4107352512079901, "learning_rate": 2.6926811947422717e-05, "loss": 0.6574, "step": 708 }, { "epoch": 0.23, "grad_norm": 0.35289004541043384, "learning_rate": 2.691718294211034e-05, "loss": 0.628, "step": 709 }, { "epoch": 0.23, "grad_norm": 0.3856911146026206, "learning_rate": 2.690754060279914e-05, "loss": 0.6215, "step": 710 }, { "epoch": 0.23, "grad_norm": 0.4386806804368987, "learning_rate": 2.6897884940277815e-05, "loss": 0.8967, "step": 711 }, { "epoch": 0.23, "grad_norm": 0.43495130326757514, "learning_rate": 2.6888215965349974e-05, "loss": 0.6295, "step": 712 }, { "epoch": 0.23, "grad_norm": 0.39945849595093325, "learning_rate": 2.6878533688834125e-05, "loss": 0.6062, "step": 713 }, { "epoch": 0.23, "grad_norm": 0.43264936502424023, "learning_rate": 2.6868838121563658e-05, "loss": 0.6712, "step": 714 }, { "epoch": 0.23, "grad_norm": 0.4338526452961556, "learning_rate": 2.685912927438682e-05, "loss": 0.6307, "step": 715 }, { "epoch": 0.23, "grad_norm": 0.3969824055295609, "learning_rate": 2.6849407158166743e-05, "loss": 0.7821, "step": 716 }, { "epoch": 0.23, "grad_norm": 0.4078727506698432, "learning_rate": 2.6839671783781382e-05, "loss": 0.6289, "step": 717 }, { "epoch": 0.23, "grad_norm": 0.4772741371195383, "learning_rate": 2.6829923162123534e-05, "loss": 0.908, "step": 718 }, { "epoch": 0.23, "grad_norm": 0.44513086203394453, "learning_rate": 2.6820161304100828e-05, "loss": 0.6862, "step": 719 }, { "epoch": 0.24, "grad_norm": 0.42922492172015847, "learning_rate": 2.6810386220635693e-05, "loss": 0.614, "step": 720 }, { "epoch": 0.24, "grad_norm": 0.3895197973968205, "learning_rate": 2.6800597922665346e-05, "loss": 0.616, "step": 721 }, { "epoch": 0.24, "grad_norm": 0.44575212694227223, "learning_rate": 2.6790796421141813e-05, "loss": 0.8689, "step": 722 }, { "epoch": 0.24, "grad_norm": 0.4411257998473039, "learning_rate": 2.6780981727031877e-05, "loss": 0.6468, "step": 723 }, { "epoch": 0.24, "grad_norm": 0.4347266563935388, "learning_rate": 2.6771153851317088e-05, "loss": 0.6216, "step": 724 }, { "epoch": 0.24, "grad_norm": 0.4490835758061524, "learning_rate": 2.6761312804993737e-05, "loss": 0.6899, "step": 725 }, { "epoch": 0.24, "grad_norm": 0.5157736734823923, "learning_rate": 2.6751458599072862e-05, "loss": 0.6406, "step": 726 }, { "epoch": 0.24, "grad_norm": 0.4344424119135314, "learning_rate": 2.674159124458023e-05, "loss": 0.8839, "step": 727 }, { "epoch": 0.24, "grad_norm": 1.5183795689983295, "learning_rate": 2.6731710752556293e-05, "loss": 0.6345, "step": 728 }, { "epoch": 0.24, "grad_norm": 0.5493393946643435, "learning_rate": 2.6721817134056227e-05, "loss": 0.6387, "step": 729 }, { "epoch": 0.24, "grad_norm": 0.42694326383900205, "learning_rate": 2.671191040014989e-05, "loss": 0.883, "step": 730 }, { "epoch": 0.24, "grad_norm": 0.46759007693085886, "learning_rate": 2.6701990561921812e-05, "loss": 0.6376, "step": 731 }, { "epoch": 0.24, "grad_norm": 0.4416137379018691, "learning_rate": 2.6692057630471184e-05, "loss": 0.6324, "step": 732 }, { "epoch": 0.24, "grad_norm": 0.42463676523792915, "learning_rate": 2.668211161691185e-05, "loss": 0.8711, "step": 733 }, { "epoch": 0.24, "grad_norm": 0.4879795046646651, "learning_rate": 2.667215253237229e-05, "loss": 0.6211, "step": 734 }, { "epoch": 0.24, "grad_norm": 0.36021581218628024, "learning_rate": 2.666218038799561e-05, "loss": 0.6248, "step": 735 }, { "epoch": 0.24, "grad_norm": 0.502023986952676, "learning_rate": 2.6652195194939525e-05, "loss": 0.6557, "step": 736 }, { "epoch": 0.24, "grad_norm": 0.3881361122852573, "learning_rate": 2.6642196964376355e-05, "loss": 0.8977, "step": 737 }, { "epoch": 0.24, "grad_norm": 0.5526347377704214, "learning_rate": 2.6632185707493e-05, "loss": 0.6185, "step": 738 }, { "epoch": 0.24, "grad_norm": 0.34224888072184717, "learning_rate": 2.662216143549095e-05, "loss": 0.6307, "step": 739 }, { "epoch": 0.24, "grad_norm": 0.42001073537764266, "learning_rate": 2.661212415958624e-05, "loss": 0.6196, "step": 740 }, { "epoch": 0.24, "grad_norm": 0.4443540975682335, "learning_rate": 2.6602073891009458e-05, "loss": 0.896, "step": 741 }, { "epoch": 0.24, "grad_norm": 0.41399645272687097, "learning_rate": 2.6592010641005745e-05, "loss": 0.6256, "step": 742 }, { "epoch": 0.24, "grad_norm": 0.5278146686391838, "learning_rate": 2.6581934420834754e-05, "loss": 0.6301, "step": 743 }, { "epoch": 0.24, "grad_norm": 0.4110781163980199, "learning_rate": 2.6571845241770645e-05, "loss": 0.9052, "step": 744 }, { "epoch": 0.24, "grad_norm": 0.38172694633426824, "learning_rate": 2.6561743115102085e-05, "loss": 0.6377, "step": 745 }, { "epoch": 0.24, "grad_norm": 0.4177791150800698, "learning_rate": 2.6551628052132237e-05, "loss": 0.6316, "step": 746 }, { "epoch": 0.24, "grad_norm": 0.3960591142290854, "learning_rate": 2.6541500064178714e-05, "loss": 0.6545, "step": 747 }, { "epoch": 0.24, "grad_norm": 0.4854583309413615, "learning_rate": 2.6531359162573618e-05, "loss": 0.8832, "step": 748 }, { "epoch": 0.24, "grad_norm": 0.4404074633733206, "learning_rate": 2.6521205358663477e-05, "loss": 0.6254, "step": 749 }, { "epoch": 0.24, "grad_norm": 0.38800682948774895, "learning_rate": 2.6511038663809277e-05, "loss": 0.6381, "step": 750 }, { "epoch": 0.25, "grad_norm": 0.47120785288860323, "learning_rate": 2.65008590893864e-05, "loss": 0.8904, "step": 751 }, { "epoch": 0.25, "grad_norm": 0.4748243345673216, "learning_rate": 2.649066664678467e-05, "loss": 0.635, "step": 752 }, { "epoch": 0.25, "grad_norm": 0.36545657174639434, "learning_rate": 2.6480461347408285e-05, "loss": 0.6208, "step": 753 }, { "epoch": 0.25, "grad_norm": 0.4360973723160973, "learning_rate": 2.6470243202675842e-05, "loss": 0.877, "step": 754 }, { "epoch": 0.25, "grad_norm": 0.3876749560932101, "learning_rate": 2.64600122240203e-05, "loss": 0.6206, "step": 755 }, { "epoch": 0.25, "grad_norm": 0.3905956791874913, "learning_rate": 2.644976842288899e-05, "loss": 0.6318, "step": 756 }, { "epoch": 0.25, "grad_norm": 0.36709023374994254, "learning_rate": 2.6439511810743572e-05, "loss": 0.6107, "step": 757 }, { "epoch": 0.25, "grad_norm": 0.4504933170333573, "learning_rate": 2.6429242399060063e-05, "loss": 0.6871, "step": 758 }, { "epoch": 0.25, "grad_norm": 0.3863346721980486, "learning_rate": 2.6418960199328783e-05, "loss": 0.9171, "step": 759 }, { "epoch": 0.25, "grad_norm": 0.3742567014774438, "learning_rate": 2.640866522305437e-05, "loss": 0.6253, "step": 760 }, { "epoch": 0.25, "grad_norm": 0.3657541074366722, "learning_rate": 2.6398357481755756e-05, "loss": 0.6138, "step": 761 }, { "epoch": 0.25, "grad_norm": 0.41047913246995366, "learning_rate": 2.638803698696615e-05, "loss": 0.9286, "step": 762 }, { "epoch": 0.25, "grad_norm": 0.3883655386677752, "learning_rate": 2.6377703750233034e-05, "loss": 0.635, "step": 763 }, { "epoch": 0.25, "grad_norm": 0.35860091243809494, "learning_rate": 2.6367357783118152e-05, "loss": 0.6133, "step": 764 }, { "epoch": 0.25, "grad_norm": 0.41393104677322046, "learning_rate": 2.6356999097197487e-05, "loss": 0.9106, "step": 765 }, { "epoch": 0.25, "grad_norm": 0.3977788525592455, "learning_rate": 2.634662770406125e-05, "loss": 0.6166, "step": 766 }, { "epoch": 0.25, "grad_norm": 0.403348728386824, "learning_rate": 2.6336243615313876e-05, "loss": 0.6305, "step": 767 }, { "epoch": 0.25, "grad_norm": 0.4552363507796755, "learning_rate": 2.6325846842574e-05, "loss": 0.8608, "step": 768 }, { "epoch": 0.25, "grad_norm": 0.4321009528982105, "learning_rate": 2.6315437397474452e-05, "loss": 0.4423, "step": 769 }, { "epoch": 0.25, "grad_norm": 0.5081917156374297, "learning_rate": 2.6305015291662247e-05, "loss": 0.8695, "step": 770 }, { "epoch": 0.25, "grad_norm": 0.3976509586239672, "learning_rate": 2.629458053679855e-05, "loss": 0.6068, "step": 771 }, { "epoch": 0.25, "grad_norm": 0.4121731632221818, "learning_rate": 2.6284133144558697e-05, "loss": 0.636, "step": 772 }, { "epoch": 0.25, "grad_norm": 0.4379002735120939, "learning_rate": 2.627367312663214e-05, "loss": 0.8626, "step": 773 }, { "epoch": 0.25, "grad_norm": 0.44434650380527607, "learning_rate": 2.626320049472249e-05, "loss": 0.6525, "step": 774 }, { "epoch": 0.25, "grad_norm": 0.4536746866571549, "learning_rate": 2.6252715260547444e-05, "loss": 0.6239, "step": 775 }, { "epoch": 0.25, "grad_norm": 0.40040598878696493, "learning_rate": 2.624221743583881e-05, "loss": 0.8907, "step": 776 }, { "epoch": 0.25, "grad_norm": 0.439256353626243, "learning_rate": 2.623170703234249e-05, "loss": 0.6246, "step": 777 }, { "epoch": 0.25, "grad_norm": 0.38811353109241986, "learning_rate": 2.622118406181845e-05, "loss": 0.6186, "step": 778 }, { "epoch": 0.25, "grad_norm": 0.47449540440049387, "learning_rate": 2.6210648536040715e-05, "loss": 0.9218, "step": 779 }, { "epoch": 0.25, "grad_norm": 0.4081324175113107, "learning_rate": 2.6200100466797367e-05, "loss": 0.672, "step": 780 }, { "epoch": 0.26, "grad_norm": 0.37658734064277755, "learning_rate": 2.6189539865890527e-05, "loss": 0.6152, "step": 781 }, { "epoch": 0.26, "grad_norm": 0.3978122032226821, "learning_rate": 2.6178966745136322e-05, "loss": 0.6093, "step": 782 }, { "epoch": 0.26, "grad_norm": 0.38664017486992763, "learning_rate": 2.6168381116364898e-05, "loss": 0.9154, "step": 783 }, { "epoch": 0.26, "grad_norm": 0.3969960780605676, "learning_rate": 2.6157782991420385e-05, "loss": 0.631, "step": 784 }, { "epoch": 0.26, "grad_norm": 0.3586949185885609, "learning_rate": 2.6147172382160913e-05, "loss": 0.62, "step": 785 }, { "epoch": 0.26, "grad_norm": 0.376970501242231, "learning_rate": 2.6136549300458572e-05, "loss": 0.6206, "step": 786 }, { "epoch": 0.26, "grad_norm": 0.40482540305595904, "learning_rate": 2.6125913758199393e-05, "loss": 0.9073, "step": 787 }, { "epoch": 0.26, "grad_norm": 0.3644120443323625, "learning_rate": 2.6115265767283377e-05, "loss": 0.6184, "step": 788 }, { "epoch": 0.26, "grad_norm": 0.34964748389644446, "learning_rate": 2.610460533962443e-05, "loss": 0.6124, "step": 789 }, { "epoch": 0.26, "grad_norm": 0.418830952442657, "learning_rate": 2.609393248715038e-05, "loss": 0.9309, "step": 790 }, { "epoch": 0.26, "grad_norm": 0.4045093117451661, "learning_rate": 2.6083247221802967e-05, "loss": 0.6887, "step": 791 }, { "epoch": 0.26, "grad_norm": 0.6270879658106631, "learning_rate": 2.60725495555378e-05, "loss": 0.6271, "step": 792 }, { "epoch": 0.26, "grad_norm": 0.5142021840435012, "learning_rate": 2.6061839500324386e-05, "loss": 0.6417, "step": 793 }, { "epoch": 0.26, "grad_norm": 0.4213948699985123, "learning_rate": 2.6051117068146073e-05, "loss": 0.8875, "step": 794 }, { "epoch": 0.26, "grad_norm": 0.38145594780585257, "learning_rate": 2.604038227100008e-05, "loss": 0.6062, "step": 795 }, { "epoch": 0.26, "grad_norm": 0.3926202499155095, "learning_rate": 2.6029635120897434e-05, "loss": 0.6771, "step": 796 }, { "epoch": 0.26, "grad_norm": 0.3870986273096716, "learning_rate": 2.6018875629862997e-05, "loss": 0.6333, "step": 797 }, { "epoch": 0.26, "grad_norm": 0.4248755414104529, "learning_rate": 2.6008103809935456e-05, "loss": 0.6237, "step": 798 }, { "epoch": 0.26, "grad_norm": 0.46876571735543454, "learning_rate": 2.599731967316726e-05, "loss": 0.8026, "step": 799 }, { "epoch": 0.26, "grad_norm": 0.40636161155239586, "learning_rate": 2.5986523231624665e-05, "loss": 0.9065, "step": 800 }, { "epoch": 0.26, "grad_norm": 0.38781841097682096, "learning_rate": 2.5975714497387678e-05, "loss": 0.6115, "step": 801 }, { "epoch": 0.26, "grad_norm": 0.3919407594130036, "learning_rate": 2.5964893482550076e-05, "loss": 0.6746, "step": 802 }, { "epoch": 0.26, "grad_norm": 0.4222219123197301, "learning_rate": 2.5954060199219364e-05, "loss": 0.6175, "step": 803 }, { "epoch": 0.26, "grad_norm": 0.3456531151925001, "learning_rate": 2.594321465951677e-05, "loss": 0.6006, "step": 804 }, { "epoch": 0.26, "grad_norm": 0.4492629126261837, "learning_rate": 2.5932356875577255e-05, "loss": 0.8832, "step": 805 }, { "epoch": 0.26, "grad_norm": 0.38773755348839184, "learning_rate": 2.5921486859549462e-05, "loss": 0.636, "step": 806 }, { "epoch": 0.26, "grad_norm": 0.44083906736901274, "learning_rate": 2.5910604623595732e-05, "loss": 0.6665, "step": 807 }, { "epoch": 0.26, "grad_norm": 0.402575228448972, "learning_rate": 2.5899710179892067e-05, "loss": 0.6068, "step": 808 }, { "epoch": 0.26, "grad_norm": 0.4313243251042865, "learning_rate": 2.5888803540628143e-05, "loss": 0.8831, "step": 809 }, { "epoch": 0.26, "grad_norm": 0.4192874620735385, "learning_rate": 2.5877884718007262e-05, "loss": 0.6175, "step": 810 }, { "epoch": 0.26, "grad_norm": 1.4794232293514378, "learning_rate": 2.586695372424638e-05, "loss": 0.621, "step": 811 }, { "epoch": 0.27, "grad_norm": 0.48232272389983616, "learning_rate": 2.5856010571576052e-05, "loss": 0.6264, "step": 812 }, { "epoch": 0.27, "grad_norm": 0.4560076626628479, "learning_rate": 2.584505527224045e-05, "loss": 0.8798, "step": 813 }, { "epoch": 0.27, "grad_norm": 0.48365395566457015, "learning_rate": 2.5834087838497337e-05, "loss": 0.6278, "step": 814 }, { "epoch": 0.27, "grad_norm": 0.4164754993607037, "learning_rate": 2.5823108282618034e-05, "loss": 0.6181, "step": 815 }, { "epoch": 0.27, "grad_norm": 0.41044037536636774, "learning_rate": 2.5812116616887448e-05, "loss": 0.8439, "step": 816 }, { "epoch": 0.27, "grad_norm": 0.4242214965310601, "learning_rate": 2.5801112853604035e-05, "loss": 0.6201, "step": 817 }, { "epoch": 0.27, "grad_norm": 0.44705785214515037, "learning_rate": 2.5790097005079766e-05, "loss": 0.6757, "step": 818 }, { "epoch": 0.27, "grad_norm": 0.42372974166875005, "learning_rate": 2.5779069083640154e-05, "loss": 0.6056, "step": 819 }, { "epoch": 0.27, "grad_norm": 0.4204574475767701, "learning_rate": 2.576802910162422e-05, "loss": 0.8773, "step": 820 }, { "epoch": 0.27, "grad_norm": 0.7704812098221359, "learning_rate": 2.5756977071384455e-05, "loss": 0.6394, "step": 821 }, { "epoch": 0.27, "grad_norm": 0.4248099472005494, "learning_rate": 2.574591300528686e-05, "loss": 0.6352, "step": 822 }, { "epoch": 0.27, "grad_norm": 0.4430724790234615, "learning_rate": 2.5734836915710892e-05, "loss": 0.9108, "step": 823 }, { "epoch": 0.27, "grad_norm": 0.49373921596747045, "learning_rate": 2.572374881504945e-05, "loss": 0.6249, "step": 824 }, { "epoch": 0.27, "grad_norm": 0.39453752156039074, "learning_rate": 2.57126487157089e-05, "loss": 0.6242, "step": 825 }, { "epoch": 0.27, "grad_norm": 0.4515135020043287, "learning_rate": 2.5701536630109005e-05, "loss": 0.895, "step": 826 }, { "epoch": 0.27, "grad_norm": 0.42192867696938496, "learning_rate": 2.5690412570682946e-05, "loss": 0.6219, "step": 827 }, { "epoch": 0.27, "grad_norm": 0.5604102106590716, "learning_rate": 2.5679276549877307e-05, "loss": 0.603, "step": 828 }, { "epoch": 0.27, "grad_norm": 0.4811911818847427, "learning_rate": 2.566812858015206e-05, "loss": 0.655, "step": 829 }, { "epoch": 0.27, "grad_norm": 0.3833237047635833, "learning_rate": 2.5656968673980536e-05, "loss": 0.6169, "step": 830 }, { "epoch": 0.27, "grad_norm": 0.3842906058052285, "learning_rate": 2.5645796843849418e-05, "loss": 0.8897, "step": 831 }, { "epoch": 0.27, "grad_norm": 0.4519809922709654, "learning_rate": 2.563461310225875e-05, "loss": 0.634, "step": 832 }, { "epoch": 0.27, "grad_norm": 0.33324680900045445, "learning_rate": 2.5623417461721887e-05, "loss": 0.6095, "step": 833 }, { "epoch": 0.27, "grad_norm": 0.41032030991509333, "learning_rate": 2.5612209934765503e-05, "loss": 0.9067, "step": 834 }, { "epoch": 0.27, "grad_norm": 0.4300852144855238, "learning_rate": 2.560099053392957e-05, "loss": 0.6233, "step": 835 }, { "epoch": 0.27, "grad_norm": 0.33308407683440083, "learning_rate": 2.5589759271767344e-05, "loss": 0.6238, "step": 836 }, { "epoch": 0.27, "grad_norm": 0.4092219704793837, "learning_rate": 2.5578516160845363e-05, "loss": 0.9013, "step": 837 }, { "epoch": 0.27, "grad_norm": 0.37367802846546017, "learning_rate": 2.5567261213743406e-05, "loss": 0.6087, "step": 838 }, { "epoch": 0.27, "grad_norm": 0.362358055472123, "learning_rate": 2.5555994443054506e-05, "loss": 0.6159, "step": 839 }, { "epoch": 0.27, "grad_norm": 0.3965310458953332, "learning_rate": 2.554471586138493e-05, "loss": 0.6527, "step": 840 }, { "epoch": 0.27, "grad_norm": 0.37430214511255033, "learning_rate": 2.553342548135414e-05, "loss": 0.6151, "step": 841 }, { "epoch": 0.27, "grad_norm": 0.41254008539816506, "learning_rate": 2.552212331559482e-05, "loss": 0.8827, "step": 842 }, { "epoch": 0.28, "grad_norm": 0.35591219257701084, "learning_rate": 2.5510809376752836e-05, "loss": 0.6211, "step": 843 }, { "epoch": 0.28, "grad_norm": 0.39860770720900085, "learning_rate": 2.549948367748722e-05, "loss": 0.6064, "step": 844 }, { "epoch": 0.28, "grad_norm": 0.36618549850572785, "learning_rate": 2.548814623047016e-05, "loss": 0.8323, "step": 845 }, { "epoch": 0.28, "grad_norm": 0.36336640034055673, "learning_rate": 2.547679704838701e-05, "loss": 0.6186, "step": 846 }, { "epoch": 0.28, "grad_norm": 0.32065412031902146, "learning_rate": 2.5465436143936228e-05, "loss": 0.629, "step": 847 }, { "epoch": 0.28, "grad_norm": 0.39645276779920635, "learning_rate": 2.5454063529829405e-05, "loss": 0.8763, "step": 848 }, { "epoch": 0.28, "grad_norm": 0.3481942799803621, "learning_rate": 2.544267921879122e-05, "loss": 0.6135, "step": 849 }, { "epoch": 0.28, "grad_norm": 0.32489878463707744, "learning_rate": 2.5431283223559457e-05, "loss": 0.633, "step": 850 }, { "epoch": 0.28, "grad_norm": 0.3854190290662365, "learning_rate": 2.541987555688496e-05, "loss": 0.6731, "step": 851 }, { "epoch": 0.28, "grad_norm": 0.3764376267813591, "learning_rate": 2.5408456231531634e-05, "loss": 0.8799, "step": 852 }, { "epoch": 0.28, "grad_norm": 0.33455220275466546, "learning_rate": 2.5397025260276436e-05, "loss": 0.6106, "step": 853 }, { "epoch": 0.28, "grad_norm": 0.35112710725768864, "learning_rate": 2.538558265590934e-05, "loss": 0.6297, "step": 854 }, { "epoch": 0.28, "grad_norm": 0.3307344878458334, "learning_rate": 2.5374128431233352e-05, "loss": 0.6135, "step": 855 }, { "epoch": 0.28, "grad_norm": 0.400063699725538, "learning_rate": 2.5362662599064475e-05, "loss": 0.8849, "step": 856 }, { "epoch": 0.28, "grad_norm": 0.45932137968983405, "learning_rate": 2.5351185172231683e-05, "loss": 0.6039, "step": 857 }, { "epoch": 0.28, "grad_norm": 0.3868240487140226, "learning_rate": 2.5339696163576943e-05, "loss": 0.611, "step": 858 }, { "epoch": 0.28, "grad_norm": 0.40937872034452244, "learning_rate": 2.532819558595518e-05, "loss": 0.8718, "step": 859 }, { "epoch": 0.28, "grad_norm": 0.3691307206232266, "learning_rate": 2.5316683452234256e-05, "loss": 0.6094, "step": 860 }, { "epoch": 0.28, "grad_norm": 0.36481566676646365, "learning_rate": 2.530515977529496e-05, "loss": 0.6394, "step": 861 }, { "epoch": 0.28, "grad_norm": 0.40030600146490236, "learning_rate": 2.5293624568031008e-05, "loss": 0.6516, "step": 862 }, { "epoch": 0.28, "grad_norm": 0.4401853386939005, "learning_rate": 2.5282077843349e-05, "loss": 0.8942, "step": 863 }, { "epoch": 0.28, "grad_norm": 0.5038065145469045, "learning_rate": 2.527051961416845e-05, "loss": 0.6283, "step": 864 }, { "epoch": 0.28, "grad_norm": 0.3865877214204246, "learning_rate": 2.525894989342171e-05, "loss": 0.6134, "step": 865 }, { "epoch": 0.28, "grad_norm": 0.41743565978140607, "learning_rate": 2.5247368694054017e-05, "loss": 0.9137, "step": 866 }, { "epoch": 0.28, "grad_norm": 0.36894980648484943, "learning_rate": 2.523577602902344e-05, "loss": 0.6105, "step": 867 }, { "epoch": 0.28, "grad_norm": 0.37955715398734663, "learning_rate": 2.5224171911300883e-05, "loss": 0.635, "step": 868 }, { "epoch": 0.28, "grad_norm": 0.451431793154599, "learning_rate": 2.5212556353870053e-05, "loss": 0.8994, "step": 869 }, { "epoch": 0.28, "grad_norm": 0.39905956446132657, "learning_rate": 2.5200929369727466e-05, "loss": 0.623, "step": 870 }, { "epoch": 0.28, "grad_norm": 0.3452600264854442, "learning_rate": 2.5189290971882418e-05, "loss": 0.6159, "step": 871 }, { "epoch": 0.28, "grad_norm": 0.46486790183144866, "learning_rate": 2.5177641173356985e-05, "loss": 0.9051, "step": 872 }, { "epoch": 0.29, "grad_norm": 0.3592752800244049, "learning_rate": 2.516597998718599e-05, "loss": 0.4294, "step": 873 }, { "epoch": 0.29, "grad_norm": 0.38871586354073817, "learning_rate": 2.5154307426416992e-05, "loss": 0.8981, "step": 874 }, { "epoch": 0.29, "grad_norm": 0.37885882213506306, "learning_rate": 2.514262350411029e-05, "loss": 0.6444, "step": 875 }, { "epoch": 0.29, "grad_norm": 0.340060392302459, "learning_rate": 2.51309282333389e-05, "loss": 0.6167, "step": 876 }, { "epoch": 0.29, "grad_norm": 0.4608881302373129, "learning_rate": 2.5119221627188513e-05, "loss": 0.8869, "step": 877 }, { "epoch": 0.29, "grad_norm": 2.7724192074728786, "learning_rate": 2.510750369875752e-05, "loss": 0.701, "step": 878 }, { "epoch": 0.29, "grad_norm": 0.3758187722696669, "learning_rate": 2.5095774461156975e-05, "loss": 0.6074, "step": 879 }, { "epoch": 0.29, "grad_norm": 0.373972588054626, "learning_rate": 2.508403392751059e-05, "loss": 0.605, "step": 880 }, { "epoch": 0.29, "grad_norm": 0.4220732310643281, "learning_rate": 2.5072282110954714e-05, "loss": 0.8664, "step": 881 }, { "epoch": 0.29, "grad_norm": 0.4118502241499798, "learning_rate": 2.5060519024638312e-05, "loss": 0.6298, "step": 882 }, { "epoch": 0.29, "grad_norm": 0.41413096338584027, "learning_rate": 2.5048744681722975e-05, "loss": 0.7854, "step": 883 }, { "epoch": 0.29, "grad_norm": 0.4862566455338483, "learning_rate": 2.5036959095382875e-05, "loss": 0.4298, "step": 884 }, { "epoch": 0.29, "grad_norm": 0.4164822614720324, "learning_rate": 2.5025162278804765e-05, "loss": 0.8932, "step": 885 }, { "epoch": 0.29, "grad_norm": 0.9280214502099174, "learning_rate": 2.501335424518798e-05, "loss": 0.6016, "step": 886 }, { "epoch": 0.29, "grad_norm": 0.3859566096360625, "learning_rate": 2.5001535007744375e-05, "loss": 0.6249, "step": 887 }, { "epoch": 0.29, "grad_norm": 0.4865615680098826, "learning_rate": 2.498970457969837e-05, "loss": 0.8948, "step": 888 }, { "epoch": 0.29, "grad_norm": 0.43814884018468503, "learning_rate": 2.4977862974286898e-05, "loss": 0.6822, "step": 889 }, { "epoch": 0.29, "grad_norm": 0.43572747928568895, "learning_rate": 2.4966010204759385e-05, "loss": 0.6193, "step": 890 }, { "epoch": 0.29, "grad_norm": 0.37091301275267163, "learning_rate": 2.4954146284377768e-05, "loss": 0.5985, "step": 891 }, { "epoch": 0.29, "grad_norm": 0.5812682855335701, "learning_rate": 2.4942271226416444e-05, "loss": 0.9191, "step": 892 }, { "epoch": 0.29, "grad_norm": 0.44706899668474365, "learning_rate": 2.4930385044162282e-05, "loss": 0.6244, "step": 893 }, { "epoch": 0.29, "grad_norm": 0.4069475836961546, "learning_rate": 2.4918487750914592e-05, "loss": 0.599, "step": 894 }, { "epoch": 0.29, "grad_norm": 0.4187949038092134, "learning_rate": 2.4906579359985115e-05, "loss": 0.8851, "step": 895 }, { "epoch": 0.29, "grad_norm": 0.39899403184030735, "learning_rate": 2.4894659884698024e-05, "loss": 0.617, "step": 896 }, { "epoch": 0.29, "grad_norm": 0.4688365185742953, "learning_rate": 2.4882729338389873e-05, "loss": 0.6258, "step": 897 }, { "epoch": 0.29, "grad_norm": 0.3844569198687274, "learning_rate": 2.487078773440962e-05, "loss": 0.898, "step": 898 }, { "epoch": 0.29, "grad_norm": 0.41467293036331826, "learning_rate": 2.4858835086118584e-05, "loss": 0.6274, "step": 899 }, { "epoch": 0.29, "grad_norm": 0.36691618199289244, "learning_rate": 2.484687140689044e-05, "loss": 0.6398, "step": 900 }, { "epoch": 0.29, "grad_norm": 0.38725696755050343, "learning_rate": 2.483489671011122e-05, "loss": 0.6078, "step": 901 }, { "epoch": 0.29, "grad_norm": 0.3858561738552613, "learning_rate": 2.482291100917928e-05, "loss": 0.6168, "step": 902 }, { "epoch": 0.29, "grad_norm": 0.38901014118985117, "learning_rate": 2.4810914317505267e-05, "loss": 0.8769, "step": 903 }, { "epoch": 0.3, "grad_norm": 0.415187689081323, "learning_rate": 2.4798906648512145e-05, "loss": 0.6052, "step": 904 }, { "epoch": 0.3, "grad_norm": 0.3310758418819622, "learning_rate": 2.4786888015635163e-05, "loss": 0.6209, "step": 905 }, { "epoch": 0.3, "grad_norm": 0.4069438073027696, "learning_rate": 2.477485843232183e-05, "loss": 0.8656, "step": 906 }, { "epoch": 0.3, "grad_norm": 0.3445588644780669, "learning_rate": 2.4762817912031904e-05, "loss": 0.608, "step": 907 }, { "epoch": 0.3, "grad_norm": 0.3476864192240196, "learning_rate": 2.4750766468237388e-05, "loss": 0.6391, "step": 908 }, { "epoch": 0.3, "grad_norm": 0.4588387009856667, "learning_rate": 2.47387041144225e-05, "loss": 0.8697, "step": 909 }, { "epoch": 0.3, "grad_norm": 0.38536534667230943, "learning_rate": 2.4726630864083677e-05, "loss": 0.6177, "step": 910 }, { "epoch": 0.3, "grad_norm": 0.4552587275790139, "learning_rate": 2.471454673072953e-05, "loss": 0.6858, "step": 911 }, { "epoch": 0.3, "grad_norm": 0.3543906909936889, "learning_rate": 2.4702451727880862e-05, "loss": 0.6253, "step": 912 }, { "epoch": 0.3, "grad_norm": 0.4847691013416135, "learning_rate": 2.4690345869070633e-05, "loss": 0.6148, "step": 913 }, { "epoch": 0.3, "grad_norm": 0.39226064104954966, "learning_rate": 2.4678229167843945e-05, "loss": 0.8949, "step": 914 }, { "epoch": 0.3, "grad_norm": 0.4677048981798642, "learning_rate": 2.466610163775804e-05, "loss": 0.6149, "step": 915 }, { "epoch": 0.3, "grad_norm": 0.3849059365665163, "learning_rate": 2.4653963292382272e-05, "loss": 0.6141, "step": 916 }, { "epoch": 0.3, "grad_norm": 0.42874918418232066, "learning_rate": 2.464181414529809e-05, "loss": 0.8896, "step": 917 }, { "epoch": 0.3, "grad_norm": 0.3795747802605821, "learning_rate": 2.462965421009904e-05, "loss": 0.6215, "step": 918 }, { "epoch": 0.3, "grad_norm": 0.3527692900583213, "learning_rate": 2.461748350039073e-05, "loss": 0.6052, "step": 919 }, { "epoch": 0.3, "grad_norm": 0.40284104780631386, "learning_rate": 2.4605302029790835e-05, "loss": 0.8799, "step": 920 }, { "epoch": 0.3, "grad_norm": 0.40601276379482215, "learning_rate": 2.4593109811929054e-05, "loss": 0.6339, "step": 921 }, { "epoch": 0.3, "grad_norm": 0.40988411380003914, "learning_rate": 2.458090686044712e-05, "loss": 0.6465, "step": 922 }, { "epoch": 0.3, "grad_norm": 0.3843751430780029, "learning_rate": 2.456869318899878e-05, "loss": 0.6101, "step": 923 }, { "epoch": 0.3, "grad_norm": 0.4386543588964841, "learning_rate": 2.455646881124977e-05, "loss": 0.877, "step": 924 }, { "epoch": 0.3, "grad_norm": 0.3919587835683279, "learning_rate": 2.4544233740877796e-05, "loss": 0.629, "step": 925 }, { "epoch": 0.3, "grad_norm": 0.5297734497186023, "learning_rate": 2.4531987991572543e-05, "loss": 0.6031, "step": 926 }, { "epoch": 0.3, "grad_norm": 0.33194025772910507, "learning_rate": 2.4519731577035642e-05, "loss": 0.6056, "step": 927 }, { "epoch": 0.3, "grad_norm": 0.520575791692142, "learning_rate": 2.4507464510980652e-05, "loss": 0.8915, "step": 928 }, { "epoch": 0.3, "grad_norm": 0.3408176887820626, "learning_rate": 2.4495186807133058e-05, "loss": 0.6291, "step": 929 }, { "epoch": 0.3, "grad_norm": 0.4025917000305877, "learning_rate": 2.4482898479230232e-05, "loss": 0.6262, "step": 930 }, { "epoch": 0.3, "grad_norm": 0.4014962533015661, "learning_rate": 2.4470599541021444e-05, "loss": 0.8594, "step": 931 }, { "epoch": 0.3, "grad_norm": 0.3804835025415987, "learning_rate": 2.445829000626784e-05, "loss": 0.6152, "step": 932 }, { "epoch": 0.3, "grad_norm": 0.3927906846352127, "learning_rate": 2.444596988874241e-05, "loss": 0.6593, "step": 933 }, { "epoch": 0.31, "grad_norm": 0.35439558078319455, "learning_rate": 2.443363920223e-05, "loss": 0.645, "step": 934 }, { "epoch": 0.31, "grad_norm": 0.4029399102561676, "learning_rate": 2.4421297960527262e-05, "loss": 0.8803, "step": 935 }, { "epoch": 0.31, "grad_norm": 0.3738595756275533, "learning_rate": 2.4408946177442678e-05, "loss": 0.6222, "step": 936 }, { "epoch": 0.31, "grad_norm": 0.37126661833512664, "learning_rate": 2.439658386679652e-05, "loss": 0.6153, "step": 937 }, { "epoch": 0.31, "grad_norm": 0.39729349364210553, "learning_rate": 2.4384211042420826e-05, "loss": 0.84, "step": 938 }, { "epoch": 0.31, "grad_norm": 0.402692384397272, "learning_rate": 2.43718277181594e-05, "loss": 0.6184, "step": 939 }, { "epoch": 0.31, "grad_norm": 0.35212502532441864, "learning_rate": 2.4359433907867822e-05, "loss": 0.6243, "step": 940 }, { "epoch": 0.31, "grad_norm": 0.38323150785841364, "learning_rate": 2.4347029625413365e-05, "loss": 0.9067, "step": 941 }, { "epoch": 0.31, "grad_norm": 0.35229198608081713, "learning_rate": 2.433461488467505e-05, "loss": 0.5896, "step": 942 }, { "epoch": 0.31, "grad_norm": 0.3745710854325177, "learning_rate": 2.4322189699543585e-05, "loss": 0.6439, "step": 943 }, { "epoch": 0.31, "grad_norm": 0.3604877435770622, "learning_rate": 2.4309754083921358e-05, "loss": 0.6639, "step": 944 }, { "epoch": 0.31, "grad_norm": 0.38490400120340723, "learning_rate": 2.4297308051722448e-05, "loss": 0.6226, "step": 945 }, { "epoch": 0.31, "grad_norm": 0.39024776241775383, "learning_rate": 2.4284851616872565e-05, "loss": 0.8715, "step": 946 }, { "epoch": 0.31, "grad_norm": 0.35190833639344093, "learning_rate": 2.4272384793309077e-05, "loss": 0.6285, "step": 947 }, { "epoch": 0.31, "grad_norm": 0.33521075766407504, "learning_rate": 2.4259907594980977e-05, "loss": 0.6176, "step": 948 }, { "epoch": 0.31, "grad_norm": 0.40539225585723815, "learning_rate": 2.4247420035848846e-05, "loss": 0.8844, "step": 949 }, { "epoch": 0.31, "grad_norm": 0.3626971037390911, "learning_rate": 2.4234922129884873e-05, "loss": 0.612, "step": 950 }, { "epoch": 0.31, "grad_norm": 0.3619859238965943, "learning_rate": 2.422241389107283e-05, "loss": 0.6416, "step": 951 }, { "epoch": 0.31, "grad_norm": 0.4186146950207437, "learning_rate": 2.4209895333408028e-05, "loss": 0.8786, "step": 952 }, { "epoch": 0.31, "grad_norm": 0.36521986985976274, "learning_rate": 2.4197366470897352e-05, "loss": 0.6046, "step": 953 }, { "epoch": 0.31, "grad_norm": 0.34884044705476985, "learning_rate": 2.4184827317559187e-05, "loss": 0.6252, "step": 954 }, { "epoch": 0.31, "grad_norm": 0.4226100438329558, "learning_rate": 2.4172277887423455e-05, "loss": 0.6431, "step": 955 }, { "epoch": 0.31, "grad_norm": 1.501173711942295, "learning_rate": 2.4159718194531573e-05, "loss": 0.6856, "step": 956 }, { "epoch": 0.31, "grad_norm": 0.3813515880730873, "learning_rate": 2.414714825293643e-05, "loss": 0.875, "step": 957 }, { "epoch": 0.31, "grad_norm": 0.3931832848792613, "learning_rate": 2.4134568076702388e-05, "loss": 0.6242, "step": 958 }, { "epoch": 0.31, "grad_norm": 0.3613983002099049, "learning_rate": 2.412197767990527e-05, "loss": 0.6135, "step": 959 }, { "epoch": 0.31, "grad_norm": 0.4145616026924644, "learning_rate": 2.4109377076632315e-05, "loss": 0.862, "step": 960 }, { "epoch": 0.31, "grad_norm": 0.396547805059063, "learning_rate": 2.4096766280982204e-05, "loss": 0.6499, "step": 961 }, { "epoch": 0.31, "grad_norm": 0.3660907629822963, "learning_rate": 2.4084145307065e-05, "loss": 0.6206, "step": 962 }, { "epoch": 0.31, "grad_norm": 0.39411078473702904, "learning_rate": 2.4071514169002175e-05, "loss": 0.6207, "step": 963 }, { "epoch": 0.31, "grad_norm": 0.42979193322703085, "learning_rate": 2.405887288092656e-05, "loss": 0.857, "step": 964 }, { "epoch": 0.32, "grad_norm": 0.4032071917016939, "learning_rate": 2.4046221456982343e-05, "loss": 0.6014, "step": 965 }, { "epoch": 0.32, "grad_norm": 0.34646260534118034, "learning_rate": 2.4033559911325063e-05, "loss": 0.6033, "step": 966 }, { "epoch": 0.32, "grad_norm": 0.40496195793612577, "learning_rate": 2.4020888258121577e-05, "loss": 0.8756, "step": 967 }, { "epoch": 0.32, "grad_norm": 0.3648499001555182, "learning_rate": 2.400820651155005e-05, "loss": 0.6005, "step": 968 }, { "epoch": 0.32, "grad_norm": 0.34556154602192374, "learning_rate": 2.3995514685799937e-05, "loss": 0.6207, "step": 969 }, { "epoch": 0.32, "grad_norm": 0.3945568327607841, "learning_rate": 2.3982812795071985e-05, "loss": 0.8745, "step": 970 }, { "epoch": 0.32, "grad_norm": 0.34702996802253155, "learning_rate": 2.3970100853578187e-05, "loss": 0.6239, "step": 971 }, { "epoch": 0.32, "grad_norm": 0.37334367675500685, "learning_rate": 2.3957378875541795e-05, "loss": 0.6902, "step": 972 }, { "epoch": 0.32, "grad_norm": 0.3675516696511365, "learning_rate": 2.3944646875197274e-05, "loss": 0.6394, "step": 973 }, { "epoch": 0.32, "grad_norm": 0.3411330170913808, "learning_rate": 2.393190486679032e-05, "loss": 0.6045, "step": 974 }, { "epoch": 0.32, "grad_norm": 0.3968061949588955, "learning_rate": 2.391915286457782e-05, "loss": 0.8918, "step": 975 }, { "epoch": 0.32, "grad_norm": 0.3448125924741946, "learning_rate": 2.390639088282784e-05, "loss": 0.6251, "step": 976 }, { "epoch": 0.32, "grad_norm": 0.5849272738979492, "learning_rate": 2.389361893581961e-05, "loss": 0.6115, "step": 977 }, { "epoch": 0.32, "grad_norm": 0.3914527440737505, "learning_rate": 2.388083703784352e-05, "loss": 0.8904, "step": 978 }, { "epoch": 0.32, "grad_norm": 0.355703842334043, "learning_rate": 2.3868045203201085e-05, "loss": 0.5878, "step": 979 }, { "epoch": 0.32, "grad_norm": 0.34882120776400805, "learning_rate": 2.3855243446204948e-05, "loss": 0.5895, "step": 980 }, { "epoch": 0.32, "grad_norm": 0.3916389020244703, "learning_rate": 2.3842431781178845e-05, "loss": 0.8908, "step": 981 }, { "epoch": 0.32, "grad_norm": 0.3504769781286879, "learning_rate": 2.382961022245759e-05, "loss": 0.6212, "step": 982 }, { "epoch": 0.32, "grad_norm": 0.39392565314271044, "learning_rate": 2.3816778784387097e-05, "loss": 0.6644, "step": 983 }, { "epoch": 0.32, "grad_norm": 0.37181421950316856, "learning_rate": 2.38039374813243e-05, "loss": 0.629, "step": 984 }, { "epoch": 0.32, "grad_norm": 0.3450166749132028, "learning_rate": 2.379108632763719e-05, "loss": 0.6194, "step": 985 }, { "epoch": 0.32, "grad_norm": 0.39199612084696994, "learning_rate": 2.3778225337704776e-05, "loss": 0.905, "step": 986 }, { "epoch": 0.32, "grad_norm": 0.3682660843325854, "learning_rate": 2.3765354525917064e-05, "loss": 0.6038, "step": 987 }, { "epoch": 0.32, "grad_norm": 0.3744936224213074, "learning_rate": 2.3752473906675064e-05, "loss": 0.5998, "step": 988 }, { "epoch": 0.32, "grad_norm": 0.6028829027531976, "learning_rate": 2.3739583494390755e-05, "loss": 0.8793, "step": 989 }, { "epoch": 0.32, "grad_norm": 0.3994772784182351, "learning_rate": 2.3726683303487065e-05, "loss": 0.5961, "step": 990 }, { "epoch": 0.32, "grad_norm": 0.3767164903525304, "learning_rate": 2.3713773348397874e-05, "loss": 0.607, "step": 991 }, { "epoch": 0.32, "grad_norm": 0.3906011582333362, "learning_rate": 2.3700853643567973e-05, "loss": 0.8935, "step": 992 }, { "epoch": 0.32, "grad_norm": 0.38286253977743345, "learning_rate": 2.3687924203453085e-05, "loss": 0.6263, "step": 993 }, { "epoch": 0.32, "grad_norm": 0.3775955315923773, "learning_rate": 2.3674985042519795e-05, "loss": 0.6626, "step": 994 }, { "epoch": 0.32, "grad_norm": 0.3786470754461334, "learning_rate": 2.3662036175245597e-05, "loss": 0.6161, "step": 995 }, { "epoch": 0.33, "grad_norm": 0.3897773715932928, "learning_rate": 2.3649077616118827e-05, "loss": 0.8845, "step": 996 }, { "epoch": 0.33, "grad_norm": 0.3569734793146045, "learning_rate": 2.363610937963866e-05, "loss": 0.6378, "step": 997 }, { "epoch": 0.33, "grad_norm": 0.3434089122134322, "learning_rate": 2.3623131480315107e-05, "loss": 0.6285, "step": 998 }, { "epoch": 0.33, "grad_norm": 0.33367771962971604, "learning_rate": 2.3610143932669002e-05, "loss": 0.6155, "step": 999 }, { "epoch": 0.33, "grad_norm": 0.3723470866876824, "learning_rate": 2.3597146751231946e-05, "loss": 0.8992, "step": 1000 }, { "epoch": 0.33, "grad_norm": 0.3455131209117335, "learning_rate": 2.3584139950546347e-05, "loss": 0.6081, "step": 1001 }, { "epoch": 0.33, "grad_norm": 0.43106175673995323, "learning_rate": 2.3571123545165362e-05, "loss": 0.6068, "step": 1002 }, { "epoch": 0.33, "grad_norm": 0.37472454626737156, "learning_rate": 2.3558097549652892e-05, "loss": 0.8792, "step": 1003 }, { "epoch": 0.33, "grad_norm": 0.3593289453350586, "learning_rate": 2.3545061978583584e-05, "loss": 0.6108, "step": 1004 }, { "epoch": 0.33, "grad_norm": 0.464669649063245, "learning_rate": 2.3532016846542778e-05, "loss": 0.6235, "step": 1005 }, { "epoch": 0.33, "grad_norm": 0.36298119952179103, "learning_rate": 2.351896216812653e-05, "loss": 0.6142, "step": 1006 }, { "epoch": 0.33, "grad_norm": 0.37440890861722476, "learning_rate": 2.350589795794156e-05, "loss": 0.8545, "step": 1007 }, { "epoch": 0.33, "grad_norm": 1.3023440530087702, "learning_rate": 2.349282423060527e-05, "loss": 0.6512, "step": 1008 }, { "epoch": 0.33, "grad_norm": 0.3415034558764394, "learning_rate": 2.3479741000745708e-05, "loss": 0.6092, "step": 1009 }, { "epoch": 0.33, "grad_norm": 0.46191994706930006, "learning_rate": 2.3466648283001542e-05, "loss": 0.8788, "step": 1010 }, { "epoch": 0.33, "grad_norm": 0.4154963937610332, "learning_rate": 2.3453546092022057e-05, "loss": 0.6296, "step": 1011 }, { "epoch": 0.33, "grad_norm": 0.32207190465132174, "learning_rate": 2.3440434442467155e-05, "loss": 0.6048, "step": 1012 }, { "epoch": 0.33, "grad_norm": 0.40581621262231854, "learning_rate": 2.3427313349007302e-05, "loss": 0.8763, "step": 1013 }, { "epoch": 0.33, "grad_norm": 0.3369296795931145, "learning_rate": 2.341418282632355e-05, "loss": 0.5996, "step": 1014 }, { "epoch": 0.33, "grad_norm": 0.43040172453176456, "learning_rate": 2.3401042889107473e-05, "loss": 0.6203, "step": 1015 }, { "epoch": 0.33, "grad_norm": 0.3821999474634015, "learning_rate": 2.3387893552061202e-05, "loss": 0.676, "step": 1016 }, { "epoch": 0.33, "grad_norm": 0.3453250092175418, "learning_rate": 2.3374734829897384e-05, "loss": 0.6116, "step": 1017 }, { "epoch": 0.33, "grad_norm": 0.3749455015163414, "learning_rate": 2.3361566737339166e-05, "loss": 0.849, "step": 1018 }, { "epoch": 0.33, "grad_norm": 0.48817081391022454, "learning_rate": 2.334838928912016e-05, "loss": 0.6193, "step": 1019 }, { "epoch": 0.33, "grad_norm": 0.33725730086409894, "learning_rate": 2.333520249998447e-05, "loss": 0.6152, "step": 1020 }, { "epoch": 0.33, "grad_norm": 0.39886012296660206, "learning_rate": 2.3322006384686638e-05, "loss": 0.8863, "step": 1021 }, { "epoch": 0.33, "grad_norm": 0.331753503926136, "learning_rate": 2.3308800957991657e-05, "loss": 0.6071, "step": 1022 }, { "epoch": 0.33, "grad_norm": 0.5137009100438659, "learning_rate": 2.3295586234674912e-05, "loss": 0.6045, "step": 1023 }, { "epoch": 0.33, "grad_norm": 0.381569119720948, "learning_rate": 2.3282362229522217e-05, "loss": 0.8683, "step": 1024 }, { "epoch": 0.33, "grad_norm": 0.3310775721688144, "learning_rate": 2.326912895732975e-05, "loss": 0.5853, "step": 1025 }, { "epoch": 0.34, "grad_norm": 0.3602658015902071, "learning_rate": 2.3255886432904072e-05, "loss": 0.5983, "step": 1026 }, { "epoch": 0.34, "grad_norm": 0.35869394648623465, "learning_rate": 2.324263467106209e-05, "loss": 0.6478, "step": 1027 }, { "epoch": 0.34, "grad_norm": 0.40168192661050794, "learning_rate": 2.322937368663105e-05, "loss": 0.6141, "step": 1028 }, { "epoch": 0.34, "grad_norm": 0.413716833659025, "learning_rate": 2.321610349444851e-05, "loss": 0.9, "step": 1029 }, { "epoch": 0.34, "grad_norm": 0.369811725664155, "learning_rate": 2.3202824109362338e-05, "loss": 0.6195, "step": 1030 }, { "epoch": 0.34, "grad_norm": 0.3531863993712593, "learning_rate": 2.3189535546230688e-05, "loss": 0.6001, "step": 1031 }, { "epoch": 0.34, "grad_norm": 0.40307245450414636, "learning_rate": 2.3176237819921975e-05, "loss": 0.9016, "step": 1032 }, { "epoch": 0.34, "grad_norm": 0.35994850256483096, "learning_rate": 2.3162930945314878e-05, "loss": 0.6252, "step": 1033 }, { "epoch": 0.34, "grad_norm": 1.264129967464423, "learning_rate": 2.31496149372983e-05, "loss": 0.6116, "step": 1034 }, { "epoch": 0.34, "grad_norm": 0.37353433360963534, "learning_rate": 2.3136289810771366e-05, "loss": 0.8816, "step": 1035 }, { "epoch": 0.34, "grad_norm": 0.3515447630884832, "learning_rate": 2.312295558064342e-05, "loss": 0.6071, "step": 1036 }, { "epoch": 0.34, "grad_norm": 0.35294767465457233, "learning_rate": 2.3109612261833967e-05, "loss": 0.6093, "step": 1037 }, { "epoch": 0.34, "grad_norm": 0.3836577492977392, "learning_rate": 2.3096259869272694e-05, "loss": 0.6461, "step": 1038 }, { "epoch": 0.34, "grad_norm": 0.39974078030519916, "learning_rate": 2.3082898417899446e-05, "loss": 0.8532, "step": 1039 }, { "epoch": 0.34, "grad_norm": 0.36463745652850826, "learning_rate": 2.3069527922664188e-05, "loss": 0.6353, "step": 1040 }, { "epoch": 0.34, "grad_norm": 0.33607474732806636, "learning_rate": 2.305614839852702e-05, "loss": 0.6142, "step": 1041 }, { "epoch": 0.34, "grad_norm": 0.3181789293267713, "learning_rate": 2.3042759860458142e-05, "loss": 0.6063, "step": 1042 }, { "epoch": 0.34, "grad_norm": 0.40099681080450766, "learning_rate": 2.3029362323437823e-05, "loss": 0.6615, "step": 1043 }, { "epoch": 0.34, "grad_norm": 0.39092811271570554, "learning_rate": 2.301595580245642e-05, "loss": 0.8684, "step": 1044 }, { "epoch": 0.34, "grad_norm": 0.35654957062905157, "learning_rate": 2.3002540312514332e-05, "loss": 0.6279, "step": 1045 }, { "epoch": 0.34, "grad_norm": 0.3687036068571383, "learning_rate": 2.2989115868622e-05, "loss": 0.6071, "step": 1046 }, { "epoch": 0.34, "grad_norm": 0.40454313729261265, "learning_rate": 2.297568248579988e-05, "loss": 0.9145, "step": 1047 }, { "epoch": 0.34, "grad_norm": 0.34623772914530787, "learning_rate": 2.2962240179078423e-05, "loss": 0.5986, "step": 1048 }, { "epoch": 0.34, "grad_norm": 0.3800618127394188, "learning_rate": 2.2948788963498073e-05, "loss": 0.598, "step": 1049 }, { "epoch": 0.34, "grad_norm": 0.384838380952753, "learning_rate": 2.293532885410925e-05, "loss": 0.8573, "step": 1050 }, { "epoch": 0.34, "grad_norm": 0.3577361411001462, "learning_rate": 2.2921859865972305e-05, "loss": 0.6013, "step": 1051 }, { "epoch": 0.34, "grad_norm": 0.35012125487741225, "learning_rate": 2.2908382014157536e-05, "loss": 0.6216, "step": 1052 }, { "epoch": 0.34, "grad_norm": 0.4142384486116444, "learning_rate": 2.289489531374516e-05, "loss": 0.8768, "step": 1053 }, { "epoch": 0.34, "grad_norm": 0.35838705599388254, "learning_rate": 2.2881399779825286e-05, "loss": 0.4281, "step": 1054 }, { "epoch": 0.34, "grad_norm": 0.4221471497605399, "learning_rate": 2.2867895427497916e-05, "loss": 0.8805, "step": 1055 }, { "epoch": 0.34, "grad_norm": 0.3503251638735252, "learning_rate": 2.285438227187291e-05, "loss": 0.6078, "step": 1056 }, { "epoch": 0.35, "grad_norm": 0.37646107256402916, "learning_rate": 2.2840860328069986e-05, "loss": 0.6052, "step": 1057 }, { "epoch": 0.35, "grad_norm": 0.41878920209755677, "learning_rate": 2.282732961121869e-05, "loss": 0.8924, "step": 1058 }, { "epoch": 0.35, "grad_norm": 0.3656233971157038, "learning_rate": 2.2813790136458384e-05, "loss": 0.6102, "step": 1059 }, { "epoch": 0.35, "grad_norm": 0.3438423427650855, "learning_rate": 2.280024191893823e-05, "loss": 0.5973, "step": 1060 }, { "epoch": 0.35, "grad_norm": 0.39479306289919724, "learning_rate": 2.2786684973817184e-05, "loss": 0.8857, "step": 1061 }, { "epoch": 0.35, "grad_norm": 0.37383306550583556, "learning_rate": 2.2773119316263935e-05, "loss": 0.6021, "step": 1062 }, { "epoch": 0.35, "grad_norm": 0.31483043600694766, "learning_rate": 2.2759544961456953e-05, "loss": 0.6123, "step": 1063 }, { "epoch": 0.35, "grad_norm": 0.42420705264594644, "learning_rate": 2.274596192458443e-05, "loss": 0.8895, "step": 1064 }, { "epoch": 0.35, "grad_norm": 0.3846389406936204, "learning_rate": 2.2732370220844265e-05, "loss": 0.6391, "step": 1065 }, { "epoch": 0.35, "grad_norm": 0.4005854420287183, "learning_rate": 2.2718769865444064e-05, "loss": 0.61, "step": 1066 }, { "epoch": 0.35, "grad_norm": 0.3585045431491058, "learning_rate": 2.27051608736011e-05, "loss": 0.6191, "step": 1067 }, { "epoch": 0.35, "grad_norm": 0.5879423958429079, "learning_rate": 2.269154326054232e-05, "loss": 0.8865, "step": 1068 }, { "epoch": 0.35, "grad_norm": 0.4046138884088457, "learning_rate": 2.2677917041504315e-05, "loss": 0.6002, "step": 1069 }, { "epoch": 0.35, "grad_norm": 0.37954671465468576, "learning_rate": 2.266428223173331e-05, "loss": 0.6134, "step": 1070 }, { "epoch": 0.35, "grad_norm": 0.3966747781434648, "learning_rate": 2.265063884648513e-05, "loss": 0.6344, "step": 1071 }, { "epoch": 0.35, "grad_norm": 0.4381683851167879, "learning_rate": 2.2636986901025208e-05, "loss": 0.8414, "step": 1072 }, { "epoch": 0.35, "grad_norm": 0.36030557099206234, "learning_rate": 2.2623326410628537e-05, "loss": 0.6213, "step": 1073 }, { "epoch": 0.35, "grad_norm": 0.3611879153918909, "learning_rate": 2.2609657390579693e-05, "loss": 0.6179, "step": 1074 }, { "epoch": 0.35, "grad_norm": 0.4083762708847462, "learning_rate": 2.2595979856172786e-05, "loss": 0.8623, "step": 1075 }, { "epoch": 0.35, "grad_norm": 0.40354222028325804, "learning_rate": 2.2582293822711446e-05, "loss": 0.6611, "step": 1076 }, { "epoch": 0.35, "grad_norm": 0.3552576524601911, "learning_rate": 2.256859930550882e-05, "loss": 0.6153, "step": 1077 }, { "epoch": 0.35, "grad_norm": 0.323214005953647, "learning_rate": 2.2554896319887545e-05, "loss": 0.6155, "step": 1078 }, { "epoch": 0.35, "grad_norm": 0.38621861119048695, "learning_rate": 2.254118488117974e-05, "loss": 0.875, "step": 1079 }, { "epoch": 0.35, "grad_norm": 0.35108073051266436, "learning_rate": 2.2527465004726967e-05, "loss": 0.6322, "step": 1080 }, { "epoch": 0.35, "grad_norm": 0.30793335458073656, "learning_rate": 2.2513736705880242e-05, "loss": 0.6234, "step": 1081 }, { "epoch": 0.35, "grad_norm": 0.3917795740500186, "learning_rate": 2.25e-05, "loss": 0.8521, "step": 1082 }, { "epoch": 0.35, "grad_norm": 0.33570420389154665, "learning_rate": 2.2486254902456084e-05, "loss": 0.6008, "step": 1083 }, { "epoch": 0.35, "grad_norm": 0.328178910886788, "learning_rate": 2.2472501428627726e-05, "loss": 0.617, "step": 1084 }, { "epoch": 0.35, "grad_norm": 0.39346064453339, "learning_rate": 2.2458739593903533e-05, "loss": 0.8886, "step": 1085 }, { "epoch": 0.35, "grad_norm": 0.3355160060298116, "learning_rate": 2.244496941368145e-05, "loss": 0.6038, "step": 1086 }, { "epoch": 0.35, "grad_norm": 0.37418727035154403, "learning_rate": 2.243119090336879e-05, "loss": 0.6839, "step": 1087 }, { "epoch": 0.36, "grad_norm": 0.3416524192764799, "learning_rate": 2.2417404078382153e-05, "loss": 0.6239, "step": 1088 }, { "epoch": 0.36, "grad_norm": 0.3300651774083945, "learning_rate": 2.2403608954147472e-05, "loss": 0.6198, "step": 1089 }, { "epoch": 0.36, "grad_norm": 0.3932219468730837, "learning_rate": 2.238980554609995e-05, "loss": 0.8797, "step": 1090 }, { "epoch": 0.36, "grad_norm": 0.3468734515845758, "learning_rate": 2.237599386968406e-05, "loss": 0.6101, "step": 1091 }, { "epoch": 0.36, "grad_norm": 0.31970215384777706, "learning_rate": 2.2362173940353522e-05, "loss": 0.6229, "step": 1092 }, { "epoch": 0.36, "grad_norm": 0.3860022079753528, "learning_rate": 2.2348345773571305e-05, "loss": 0.8763, "step": 1093 }, { "epoch": 0.36, "grad_norm": 0.3417262927612519, "learning_rate": 2.2334509384809584e-05, "loss": 0.6199, "step": 1094 }, { "epoch": 0.36, "grad_norm": 0.32180863066220894, "learning_rate": 2.2320664789549733e-05, "loss": 0.6063, "step": 1095 }, { "epoch": 0.36, "grad_norm": 0.388180233045591, "learning_rate": 2.2306812003282304e-05, "loss": 0.9469, "step": 1096 }, { "epoch": 0.36, "grad_norm": 0.3413807652922741, "learning_rate": 2.229295104150703e-05, "loss": 0.6237, "step": 1097 }, { "epoch": 0.36, "grad_norm": 2.2261736118882953, "learning_rate": 2.227908191973278e-05, "loss": 0.6872, "step": 1098 }, { "epoch": 0.36, "grad_norm": 0.35623978515777416, "learning_rate": 2.2265204653477553e-05, "loss": 0.6194, "step": 1099 }, { "epoch": 0.36, "grad_norm": 0.3462022272056582, "learning_rate": 2.2251319258268457e-05, "loss": 0.6003, "step": 1100 }, { "epoch": 0.36, "grad_norm": 0.7881501105492802, "learning_rate": 2.2237425749641712e-05, "loss": 0.8838, "step": 1101 }, { "epoch": 0.36, "grad_norm": 0.34254004038236563, "learning_rate": 2.2223524143142595e-05, "loss": 0.6188, "step": 1102 }, { "epoch": 0.36, "grad_norm": 1.3657039866259058, "learning_rate": 2.220961445432546e-05, "loss": 0.6254, "step": 1103 }, { "epoch": 0.36, "grad_norm": 0.5001044893885837, "learning_rate": 2.2195696698753695e-05, "loss": 0.8861, "step": 1104 }, { "epoch": 0.36, "grad_norm": 0.361780393150336, "learning_rate": 2.2181770891999714e-05, "loss": 0.6161, "step": 1105 }, { "epoch": 0.36, "grad_norm": 0.5403129787200968, "learning_rate": 2.216783704964495e-05, "loss": 0.5997, "step": 1106 }, { "epoch": 0.36, "grad_norm": 0.4804607487671906, "learning_rate": 2.215389518727981e-05, "loss": 0.8836, "step": 1107 }, { "epoch": 0.36, "grad_norm": 0.4386453790628535, "learning_rate": 2.2139945320503685e-05, "loss": 0.6165, "step": 1108 }, { "epoch": 0.36, "grad_norm": 0.44102322288592033, "learning_rate": 2.212598746492493e-05, "loss": 0.641, "step": 1109 }, { "epoch": 0.36, "grad_norm": 0.41125417641356143, "learning_rate": 2.211202163616081e-05, "loss": 0.6216, "step": 1110 }, { "epoch": 0.36, "grad_norm": 0.48701498427589, "learning_rate": 2.2098047849837544e-05, "loss": 0.8676, "step": 1111 }, { "epoch": 0.36, "grad_norm": 0.3713249269967001, "learning_rate": 2.2084066121590242e-05, "loss": 0.6075, "step": 1112 }, { "epoch": 0.36, "grad_norm": 0.44008950826843063, "learning_rate": 2.2070076467062893e-05, "loss": 0.6087, "step": 1113 }, { "epoch": 0.36, "grad_norm": 0.3931238757779468, "learning_rate": 2.2056078901908363e-05, "loss": 0.6469, "step": 1114 }, { "epoch": 0.36, "grad_norm": 0.42748004087672153, "learning_rate": 2.2042073441788363e-05, "loss": 0.606, "step": 1115 }, { "epoch": 0.36, "grad_norm": 0.41505510694255593, "learning_rate": 2.2028060102373443e-05, "loss": 0.7966, "step": 1116 }, { "epoch": 0.36, "grad_norm": 0.3969802766197224, "learning_rate": 2.201403889934297e-05, "loss": 0.627, "step": 1117 }, { "epoch": 0.37, "grad_norm": 0.5091397677531126, "learning_rate": 2.2000009848385107e-05, "loss": 0.8821, "step": 1118 }, { "epoch": 0.37, "grad_norm": 0.379793409404289, "learning_rate": 2.198597296519679e-05, "loss": 0.6181, "step": 1119 }, { "epoch": 0.37, "grad_norm": 0.37362679602226895, "learning_rate": 2.197192826548374e-05, "loss": 0.6456, "step": 1120 }, { "epoch": 0.37, "grad_norm": 0.41456620617579815, "learning_rate": 2.195787576496039e-05, "loss": 0.5932, "step": 1121 }, { "epoch": 0.37, "grad_norm": 0.4084008890524961, "learning_rate": 2.194381547934994e-05, "loss": 0.9192, "step": 1122 }, { "epoch": 0.37, "grad_norm": 0.4001284748085814, "learning_rate": 2.1929747424384272e-05, "loss": 0.6288, "step": 1123 }, { "epoch": 0.37, "grad_norm": 0.33212629068373495, "learning_rate": 2.191567161580397e-05, "loss": 0.6203, "step": 1124 }, { "epoch": 0.37, "grad_norm": 0.5414625327570132, "learning_rate": 2.19015880693583e-05, "loss": 0.6725, "step": 1125 }, { "epoch": 0.37, "grad_norm": 0.3341482835704232, "learning_rate": 2.1887496800805175e-05, "loss": 0.6081, "step": 1126 }, { "epoch": 0.37, "grad_norm": 0.4185091632294238, "learning_rate": 2.1873397825911156e-05, "loss": 0.8974, "step": 1127 }, { "epoch": 0.37, "grad_norm": 0.3404749603655738, "learning_rate": 2.1859291160451426e-05, "loss": 0.6074, "step": 1128 }, { "epoch": 0.37, "grad_norm": 0.36423429004376295, "learning_rate": 2.1845176820209765e-05, "loss": 0.6149, "step": 1129 }, { "epoch": 0.37, "grad_norm": 0.3748689287648968, "learning_rate": 2.183105482097855e-05, "loss": 0.9234, "step": 1130 }, { "epoch": 0.37, "grad_norm": 0.34799794711820675, "learning_rate": 2.1816925178558712e-05, "loss": 0.5949, "step": 1131 }, { "epoch": 0.37, "grad_norm": 0.32915216817379417, "learning_rate": 2.1802787908759767e-05, "loss": 0.5956, "step": 1132 }, { "epoch": 0.37, "grad_norm": 0.376720535468962, "learning_rate": 2.1788643027399725e-05, "loss": 0.8716, "step": 1133 }, { "epoch": 0.37, "grad_norm": 0.33518271897246477, "learning_rate": 2.177449055030514e-05, "loss": 0.6105, "step": 1134 }, { "epoch": 0.37, "grad_norm": 0.3497208603973855, "learning_rate": 2.176033049331105e-05, "loss": 0.6033, "step": 1135 }, { "epoch": 0.37, "grad_norm": 0.36014327036108057, "learning_rate": 2.1746162872260988e-05, "loss": 0.6733, "step": 1136 }, { "epoch": 0.37, "grad_norm": 0.373186626949208, "learning_rate": 2.1731987703006933e-05, "loss": 0.8309, "step": 1137 }, { "epoch": 0.37, "grad_norm": 0.3592626330462923, "learning_rate": 2.1717805001409325e-05, "loss": 0.602, "step": 1138 }, { "epoch": 0.37, "grad_norm": 0.37968888318592586, "learning_rate": 2.1703614783337023e-05, "loss": 0.613, "step": 1139 }, { "epoch": 0.37, "grad_norm": 0.34970025623419954, "learning_rate": 2.1689417064667295e-05, "loss": 0.6043, "step": 1140 }, { "epoch": 0.37, "grad_norm": 0.3965178394830245, "learning_rate": 2.1675211861285814e-05, "loss": 0.8766, "step": 1141 }, { "epoch": 0.37, "grad_norm": 0.38739992903973164, "learning_rate": 2.1660999189086613e-05, "loss": 0.5842, "step": 1142 }, { "epoch": 0.37, "grad_norm": 0.3541649283637698, "learning_rate": 2.1646779063972085e-05, "loss": 0.621, "step": 1143 }, { "epoch": 0.37, "grad_norm": 0.394596772514014, "learning_rate": 2.163255150185297e-05, "loss": 0.8752, "step": 1144 }, { "epoch": 0.37, "grad_norm": 0.33688700952689854, "learning_rate": 2.161831651864832e-05, "loss": 0.6067, "step": 1145 }, { "epoch": 0.37, "grad_norm": 0.3552564738066645, "learning_rate": 2.1604074130285496e-05, "loss": 0.6066, "step": 1146 }, { "epoch": 0.37, "grad_norm": 0.3800414582322943, "learning_rate": 2.158982435270014e-05, "loss": 0.6292, "step": 1147 }, { "epoch": 0.37, "grad_norm": 0.4264006797429244, "learning_rate": 2.157556720183616e-05, "loss": 0.877, "step": 1148 }, { "epoch": 0.38, "grad_norm": 0.3694144134230337, "learning_rate": 2.156130269364573e-05, "loss": 0.6128, "step": 1149 }, { "epoch": 0.38, "grad_norm": 0.37455656115057917, "learning_rate": 2.1547030844089234e-05, "loss": 0.6077, "step": 1150 }, { "epoch": 0.38, "grad_norm": 0.4162709437580814, "learning_rate": 2.1532751669135287e-05, "loss": 0.8627, "step": 1151 }, { "epoch": 0.38, "grad_norm": 0.9498544386675492, "learning_rate": 2.1518465184760686e-05, "loss": 0.6328, "step": 1152 }, { "epoch": 0.38, "grad_norm": 0.3422988043835213, "learning_rate": 2.1504171406950417e-05, "loss": 0.5964, "step": 1153 }, { "epoch": 0.38, "grad_norm": 24.7803102762461, "learning_rate": 2.1489870351697624e-05, "loss": 0.9314, "step": 1154 }, { "epoch": 0.38, "grad_norm": 1.1035401576973478, "learning_rate": 2.1475562035003597e-05, "loss": 0.5894, "step": 1155 }, { "epoch": 0.38, "grad_norm": 2.7582640417317235, "learning_rate": 2.1461246472877742e-05, "loss": 0.6255, "step": 1156 }, { "epoch": 0.38, "grad_norm": 0.4397777269717639, "learning_rate": 2.1446923681337578e-05, "loss": 0.9238, "step": 1157 }, { "epoch": 0.38, "grad_norm": 0.5940926921940687, "learning_rate": 2.1432593676408707e-05, "loss": 0.426, "step": 1158 }, { "epoch": 0.38, "grad_norm": 0.5852787429318937, "learning_rate": 2.1418256474124813e-05, "loss": 0.8745, "step": 1159 }, { "epoch": 0.38, "grad_norm": 0.40167316878353443, "learning_rate": 2.1403912090527625e-05, "loss": 0.6196, "step": 1160 }, { "epoch": 0.38, "grad_norm": 0.4185494287023016, "learning_rate": 2.1389560541666914e-05, "loss": 0.6135, "step": 1161 }, { "epoch": 0.38, "grad_norm": 0.4422107046364257, "learning_rate": 2.1375201843600452e-05, "loss": 0.8556, "step": 1162 }, { "epoch": 0.38, "grad_norm": 0.3909780466374, "learning_rate": 2.136083601239403e-05, "loss": 0.6228, "step": 1163 }, { "epoch": 0.38, "grad_norm": 1.3568423295620657, "learning_rate": 2.1346463064121402e-05, "loss": 0.6265, "step": 1164 }, { "epoch": 0.38, "grad_norm": 0.4142651489302172, "learning_rate": 2.1332083014864302e-05, "loss": 0.8786, "step": 1165 }, { "epoch": 0.38, "grad_norm": 0.3764666370625969, "learning_rate": 2.13176958807124e-05, "loss": 0.6259, "step": 1166 }, { "epoch": 0.38, "grad_norm": 0.40180679910153805, "learning_rate": 2.1303301677763292e-05, "loss": 0.5966, "step": 1167 }, { "epoch": 0.38, "grad_norm": 0.39839812286105936, "learning_rate": 2.128890042212249e-05, "loss": 0.8728, "step": 1168 }, { "epoch": 0.38, "grad_norm": 0.43421535921424415, "learning_rate": 2.127449212990339e-05, "loss": 0.4276, "step": 1169 }, { "epoch": 0.38, "grad_norm": 0.4346252670012633, "learning_rate": 2.1260076817227268e-05, "loss": 0.881, "step": 1170 }, { "epoch": 0.38, "grad_norm": 0.3866293445239999, "learning_rate": 2.1245654500223252e-05, "loss": 0.6149, "step": 1171 }, { "epoch": 0.38, "grad_norm": 0.3638394388774601, "learning_rate": 2.12312251950283e-05, "loss": 0.6123, "step": 1172 }, { "epoch": 0.38, "grad_norm": 0.40276888005977324, "learning_rate": 2.1216788917787195e-05, "loss": 0.8847, "step": 1173 }, { "epoch": 0.38, "grad_norm": 0.36115389907358064, "learning_rate": 2.120234568465253e-05, "loss": 0.6109, "step": 1174 }, { "epoch": 0.38, "grad_norm": 0.3680048248356503, "learning_rate": 2.118789551178467e-05, "loss": 0.6351, "step": 1175 }, { "epoch": 0.38, "grad_norm": 0.39661980978650574, "learning_rate": 2.1173438415351744e-05, "loss": 0.8872, "step": 1176 }, { "epoch": 0.38, "grad_norm": 0.4049378733791182, "learning_rate": 2.1158974411529627e-05, "loss": 0.6007, "step": 1177 }, { "epoch": 0.38, "grad_norm": 0.33825437878626274, "learning_rate": 2.114450351650193e-05, "loss": 0.6036, "step": 1178 }, { "epoch": 0.39, "grad_norm": 0.4097859364944901, "learning_rate": 2.113002574645998e-05, "loss": 0.8573, "step": 1179 }, { "epoch": 0.39, "grad_norm": 0.479782424172194, "learning_rate": 2.1115541117602776e-05, "loss": 0.6338, "step": 1180 }, { "epoch": 0.39, "grad_norm": 0.362624936577896, "learning_rate": 2.1101049646137008e-05, "loss": 0.6184, "step": 1181 }, { "epoch": 0.39, "grad_norm": 0.3402376747164177, "learning_rate": 2.108655134827701e-05, "loss": 0.5999, "step": 1182 }, { "epoch": 0.39, "grad_norm": 0.37809735176020404, "learning_rate": 2.1072046240244765e-05, "loss": 0.828, "step": 1183 }, { "epoch": 0.39, "grad_norm": 0.34830973192011194, "learning_rate": 2.1057534338269874e-05, "loss": 0.6135, "step": 1184 }, { "epoch": 0.39, "grad_norm": 0.325221860201287, "learning_rate": 2.1043015658589532e-05, "loss": 0.6036, "step": 1185 }, { "epoch": 0.39, "grad_norm": 0.3511929912183037, "learning_rate": 2.1028490217448526e-05, "loss": 0.5947, "step": 1186 }, { "epoch": 0.39, "grad_norm": 0.3724104097124004, "learning_rate": 2.1013958031099205e-05, "loss": 0.8763, "step": 1187 }, { "epoch": 0.39, "grad_norm": 0.3672812677268421, "learning_rate": 2.0999419115801465e-05, "loss": 0.616, "step": 1188 }, { "epoch": 0.39, "grad_norm": 0.9172329377837002, "learning_rate": 2.098487348782273e-05, "loss": 0.607, "step": 1189 }, { "epoch": 0.39, "grad_norm": 0.3890054081380562, "learning_rate": 2.0970321163437937e-05, "loss": 0.8727, "step": 1190 }, { "epoch": 0.39, "grad_norm": 0.4058109138664689, "learning_rate": 2.0955762158929506e-05, "loss": 0.6715, "step": 1191 }, { "epoch": 0.39, "grad_norm": 0.36436043942016616, "learning_rate": 2.0941196490587352e-05, "loss": 0.603, "step": 1192 }, { "epoch": 0.39, "grad_norm": 1.3655610681014105, "learning_rate": 2.092662417470883e-05, "loss": 0.6317, "step": 1193 }, { "epoch": 0.39, "grad_norm": 0.40261458948574314, "learning_rate": 2.091204522759873e-05, "loss": 0.8459, "step": 1194 }, { "epoch": 0.39, "grad_norm": 0.3463241866012169, "learning_rate": 2.089745966556928e-05, "loss": 0.5884, "step": 1195 }, { "epoch": 0.39, "grad_norm": 1.2844262991437323, "learning_rate": 2.0882867504940083e-05, "loss": 0.6285, "step": 1196 }, { "epoch": 0.39, "grad_norm": 0.37611311944197456, "learning_rate": 2.086826876203815e-05, "loss": 0.6719, "step": 1197 }, { "epoch": 0.39, "grad_norm": 0.3470129855369759, "learning_rate": 2.0853663453197842e-05, "loss": 0.6163, "step": 1198 }, { "epoch": 0.39, "grad_norm": 0.3606918595434996, "learning_rate": 2.0839051594760875e-05, "loss": 0.782, "step": 1199 }, { "epoch": 0.39, "grad_norm": 0.47762899863647235, "learning_rate": 2.0824433203076285e-05, "loss": 0.9022, "step": 1200 }, { "epoch": 0.39, "grad_norm": 0.36846524894738275, "learning_rate": 2.080980829450042e-05, "loss": 0.6149, "step": 1201 }, { "epoch": 0.39, "grad_norm": 0.38720317324104786, "learning_rate": 2.0795176885396928e-05, "loss": 0.6947, "step": 1202 }, { "epoch": 0.39, "grad_norm": 0.35721036543250334, "learning_rate": 2.0780538992136728e-05, "loss": 0.6148, "step": 1203 }, { "epoch": 0.39, "grad_norm": 0.36442254521947903, "learning_rate": 2.0765894631097978e-05, "loss": 0.6119, "step": 1204 }, { "epoch": 0.39, "grad_norm": 0.4255738951253742, "learning_rate": 2.075124381866609e-05, "loss": 0.9109, "step": 1205 }, { "epoch": 0.39, "grad_norm": 0.35977636419873, "learning_rate": 2.0736586571233692e-05, "loss": 0.5884, "step": 1206 }, { "epoch": 0.39, "grad_norm": 6.768041030626724, "learning_rate": 2.072192290520061e-05, "loss": 0.7128, "step": 1207 }, { "epoch": 0.39, "grad_norm": 1.0722658301041765, "learning_rate": 2.0707252836973844e-05, "loss": 0.6511, "step": 1208 }, { "epoch": 0.39, "grad_norm": 0.5236116649966505, "learning_rate": 2.0692576382967577e-05, "loss": 0.9075, "step": 1209 }, { "epoch": 0.4, "grad_norm": 0.45705724853747015, "learning_rate": 2.0677893559603113e-05, "loss": 0.6171, "step": 1210 }, { "epoch": 0.4, "grad_norm": 0.4243403699698115, "learning_rate": 2.06632043833089e-05, "loss": 0.6313, "step": 1211 }, { "epoch": 0.4, "grad_norm": 0.4522044234104238, "learning_rate": 2.064850887052048e-05, "loss": 0.5911, "step": 1212 }, { "epoch": 0.4, "grad_norm": 0.43669864634400735, "learning_rate": 2.0633807037680506e-05, "loss": 0.9091, "step": 1213 }, { "epoch": 0.4, "grad_norm": 0.40927584685500074, "learning_rate": 2.0619098901238684e-05, "loss": 0.5967, "step": 1214 }, { "epoch": 0.4, "grad_norm": 0.40768905626402807, "learning_rate": 2.060438447765177e-05, "loss": 0.6042, "step": 1215 }, { "epoch": 0.4, "grad_norm": 0.4996733462209911, "learning_rate": 2.058966378338358e-05, "loss": 0.8747, "step": 1216 }, { "epoch": 0.4, "grad_norm": 0.3810469569059986, "learning_rate": 2.0574936834904913e-05, "loss": 0.6062, "step": 1217 }, { "epoch": 0.4, "grad_norm": 0.5453113415015737, "learning_rate": 2.0560203648693594e-05, "loss": 0.6372, "step": 1218 }, { "epoch": 0.4, "grad_norm": 0.3981966785882236, "learning_rate": 2.054546424123441e-05, "loss": 0.663, "step": 1219 }, { "epoch": 0.4, "grad_norm": 0.42295117244296965, "learning_rate": 2.0530718629019114e-05, "loss": 0.8798, "step": 1220 }, { "epoch": 0.4, "grad_norm": 0.4506359273464277, "learning_rate": 2.0515966828546397e-05, "loss": 0.6248, "step": 1221 }, { "epoch": 0.4, "grad_norm": 0.3971035585584009, "learning_rate": 2.0501208856321895e-05, "loss": 0.6254, "step": 1222 }, { "epoch": 0.4, "grad_norm": 0.4550548888001102, "learning_rate": 2.048644472885812e-05, "loss": 0.8891, "step": 1223 }, { "epoch": 0.4, "grad_norm": 0.36254653415038124, "learning_rate": 2.0471674462674486e-05, "loss": 0.6209, "step": 1224 }, { "epoch": 0.4, "grad_norm": 0.3987009384584952, "learning_rate": 2.0456898074297273e-05, "loss": 0.6364, "step": 1225 }, { "epoch": 0.4, "grad_norm": 0.4845365573973938, "learning_rate": 2.0442115580259615e-05, "loss": 0.8715, "step": 1226 }, { "epoch": 0.4, "grad_norm": 0.4286698375537526, "learning_rate": 2.0427326997101474e-05, "loss": 0.6188, "step": 1227 }, { "epoch": 0.4, "grad_norm": 0.33462475857691865, "learning_rate": 2.041253234136963e-05, "loss": 0.5988, "step": 1228 }, { "epoch": 0.4, "grad_norm": 0.37069470192786436, "learning_rate": 2.039773162961764e-05, "loss": 0.6104, "step": 1229 }, { "epoch": 0.4, "grad_norm": 0.367885960055262, "learning_rate": 2.0382924878405862e-05, "loss": 0.64, "step": 1230 }, { "epoch": 0.4, "grad_norm": 0.38199970075080836, "learning_rate": 2.036811210430139e-05, "loss": 0.8733, "step": 1231 }, { "epoch": 0.4, "grad_norm": 0.3523951063468357, "learning_rate": 2.035329332387808e-05, "loss": 0.6228, "step": 1232 }, { "epoch": 0.4, "grad_norm": 0.326121969202778, "learning_rate": 2.033846855371648e-05, "loss": 0.6268, "step": 1233 }, { "epoch": 0.4, "grad_norm": 1.6931891671077222, "learning_rate": 2.0323637810403866e-05, "loss": 0.9103, "step": 1234 }, { "epoch": 0.4, "grad_norm": 0.3357968536785172, "learning_rate": 2.0308801110534178e-05, "loss": 0.6153, "step": 1235 }, { "epoch": 0.4, "grad_norm": 0.3548666100378999, "learning_rate": 2.029395847070803e-05, "loss": 0.6085, "step": 1236 }, { "epoch": 0.4, "grad_norm": 0.44138472411832264, "learning_rate": 2.0279109907532694e-05, "loss": 0.8984, "step": 1237 }, { "epoch": 0.4, "grad_norm": 0.3626725939448954, "learning_rate": 2.0264255437622037e-05, "loss": 0.6216, "step": 1238 }, { "epoch": 0.4, "grad_norm": 0.35236002911740416, "learning_rate": 2.024939507759656e-05, "loss": 0.5941, "step": 1239 }, { "epoch": 0.4, "grad_norm": 0.38640729415279795, "learning_rate": 2.023452884408336e-05, "loss": 0.8451, "step": 1240 }, { "epoch": 0.41, "grad_norm": 0.3371297155267744, "learning_rate": 2.0219656753716075e-05, "loss": 0.4236, "step": 1241 }, { "epoch": 0.41, "grad_norm": 0.4268464651091601, "learning_rate": 2.0204778823134936e-05, "loss": 0.8715, "step": 1242 }, { "epoch": 0.41, "grad_norm": 0.33728631649203683, "learning_rate": 2.018989506898668e-05, "loss": 0.6378, "step": 1243 }, { "epoch": 0.41, "grad_norm": 0.3368935682251005, "learning_rate": 2.017500550792456e-05, "loss": 0.5999, "step": 1244 }, { "epoch": 0.41, "grad_norm": 0.3917123037788826, "learning_rate": 2.0160110156608346e-05, "loss": 0.8463, "step": 1245 }, { "epoch": 0.41, "grad_norm": 0.3460336518116496, "learning_rate": 2.0145209031704273e-05, "loss": 0.6185, "step": 1246 }, { "epoch": 0.41, "grad_norm": 0.32287906888974044, "learning_rate": 2.0130302149885033e-05, "loss": 0.6204, "step": 1247 }, { "epoch": 0.41, "grad_norm": 0.3851387795296635, "learning_rate": 2.0115389527829767e-05, "loss": 0.8826, "step": 1248 }, { "epoch": 0.41, "grad_norm": 0.3320327238016165, "learning_rate": 2.010047118222404e-05, "loss": 0.5942, "step": 1249 }, { "epoch": 0.41, "grad_norm": 0.5563228758312904, "learning_rate": 2.0085547129759808e-05, "loss": 0.6122, "step": 1250 }, { "epoch": 0.41, "grad_norm": 0.4437629653806828, "learning_rate": 2.0070617387135433e-05, "loss": 0.8714, "step": 1251 }, { "epoch": 0.41, "grad_norm": 1.4897639287798499, "learning_rate": 2.0055681971055626e-05, "loss": 0.7084, "step": 1252 }, { "epoch": 0.41, "grad_norm": 0.36813932795147714, "learning_rate": 2.004074089823145e-05, "loss": 0.6291, "step": 1253 }, { "epoch": 0.41, "grad_norm": 0.33216706597018547, "learning_rate": 2.0025794185380307e-05, "loss": 0.5986, "step": 1254 }, { "epoch": 0.41, "grad_norm": 0.37990961059276424, "learning_rate": 2.0010841849225896e-05, "loss": 0.8664, "step": 1255 }, { "epoch": 0.41, "grad_norm": 0.3470208908568236, "learning_rate": 1.999588390649821e-05, "loss": 0.6284, "step": 1256 }, { "epoch": 0.41, "grad_norm": 0.3308755342517641, "learning_rate": 1.998092037393353e-05, "loss": 0.6211, "step": 1257 }, { "epoch": 0.41, "grad_norm": 0.32659592388449504, "learning_rate": 1.9965951268274373e-05, "loss": 0.5764, "step": 1258 }, { "epoch": 0.41, "grad_norm": 0.38590680587805254, "learning_rate": 1.9950976606269497e-05, "loss": 0.8673, "step": 1259 }, { "epoch": 0.41, "grad_norm": 0.33403395273580444, "learning_rate": 1.9935996404673886e-05, "loss": 0.5765, "step": 1260 }, { "epoch": 0.41, "grad_norm": 0.32579966913092834, "learning_rate": 1.992101068024871e-05, "loss": 0.6046, "step": 1261 }, { "epoch": 0.41, "grad_norm": 0.37845914997440677, "learning_rate": 1.990601944976133e-05, "loss": 0.8998, "step": 1262 }, { "epoch": 0.41, "grad_norm": 0.3621214538338608, "learning_rate": 1.989102272998525e-05, "loss": 0.6638, "step": 1263 }, { "epoch": 0.41, "grad_norm": 0.34199366591032004, "learning_rate": 1.987602053770013e-05, "loss": 0.6294, "step": 1264 }, { "epoch": 0.41, "grad_norm": 0.35805245973979405, "learning_rate": 1.9861012889691757e-05, "loss": 0.6139, "step": 1265 }, { "epoch": 0.41, "grad_norm": 0.4254811934860485, "learning_rate": 1.984599980275202e-05, "loss": 0.8452, "step": 1266 }, { "epoch": 0.41, "grad_norm": 0.34080778142823936, "learning_rate": 1.9830981293678878e-05, "loss": 0.5985, "step": 1267 }, { "epoch": 0.41, "grad_norm": 0.38058030572806, "learning_rate": 1.9815957379276363e-05, "loss": 0.6044, "step": 1268 }, { "epoch": 0.41, "grad_norm": 0.4224456466380447, "learning_rate": 1.980092807635457e-05, "loss": 0.8571, "step": 1269 }, { "epoch": 0.41, "grad_norm": 0.3456591599676722, "learning_rate": 1.9785893401729618e-05, "loss": 0.6031, "step": 1270 }, { "epoch": 0.42, "grad_norm": 0.3223654901962104, "learning_rate": 1.9770853372223627e-05, "loss": 0.6084, "step": 1271 }, { "epoch": 0.42, "grad_norm": 0.3954354036219136, "learning_rate": 1.9755808004664702e-05, "loss": 0.8892, "step": 1272 }, { "epoch": 0.42, "grad_norm": 0.3338258664193315, "learning_rate": 1.9740757315886946e-05, "loss": 0.6175, "step": 1273 }, { "epoch": 0.42, "grad_norm": 0.39523598105475033, "learning_rate": 1.9725701322730394e-05, "loss": 0.6575, "step": 1274 }, { "epoch": 0.42, "grad_norm": 0.3309988116908458, "learning_rate": 1.9710640042041023e-05, "loss": 0.5977, "step": 1275 }, { "epoch": 0.42, "grad_norm": 0.33817097391188417, "learning_rate": 1.9695573490670734e-05, "loss": 0.6275, "step": 1276 }, { "epoch": 0.42, "grad_norm": 0.37926990243870373, "learning_rate": 1.9680501685477305e-05, "loss": 0.8871, "step": 1277 }, { "epoch": 0.42, "grad_norm": 0.36045296163006496, "learning_rate": 1.966542464332442e-05, "loss": 0.5997, "step": 1278 }, { "epoch": 0.42, "grad_norm": 0.39617598607197946, "learning_rate": 1.96503423810816e-05, "loss": 0.6667, "step": 1279 }, { "epoch": 0.42, "grad_norm": 0.32688074396952915, "learning_rate": 1.963525491562421e-05, "loss": 0.5889, "step": 1280 }, { "epoch": 0.42, "grad_norm": 0.3813057628277453, "learning_rate": 1.9620162263833455e-05, "loss": 0.9061, "step": 1281 }, { "epoch": 0.42, "grad_norm": 0.32731061663727695, "learning_rate": 1.9605064442596316e-05, "loss": 0.5744, "step": 1282 }, { "epoch": 0.42, "grad_norm": 0.3589512314946074, "learning_rate": 1.958996146880558e-05, "loss": 0.8789, "step": 1283 }, { "epoch": 0.42, "grad_norm": 0.35828565576562493, "learning_rate": 1.9574853359359782e-05, "loss": 0.5979, "step": 1284 }, { "epoch": 0.42, "grad_norm": 0.33731022374881986, "learning_rate": 1.9559740131163223e-05, "loss": 0.6354, "step": 1285 }, { "epoch": 0.42, "grad_norm": 0.33017157418314047, "learning_rate": 1.9544621801125913e-05, "loss": 0.598, "step": 1286 }, { "epoch": 0.42, "grad_norm": 0.30274012866544403, "learning_rate": 1.9529498386163572e-05, "loss": 0.5916, "step": 1287 }, { "epoch": 0.42, "grad_norm": 0.3586101927044782, "learning_rate": 1.9514369903197628e-05, "loss": 0.8782, "step": 1288 }, { "epoch": 0.42, "grad_norm": 0.3352492093856682, "learning_rate": 1.949923636915516e-05, "loss": 0.6065, "step": 1289 }, { "epoch": 0.42, "grad_norm": 0.3800339067711002, "learning_rate": 1.9484097800968907e-05, "loss": 0.6261, "step": 1290 }, { "epoch": 0.42, "grad_norm": 0.31224155020139677, "learning_rate": 1.9468954215577227e-05, "loss": 0.5941, "step": 1291 }, { "epoch": 0.42, "grad_norm": 0.3656066391806644, "learning_rate": 1.9453805629924126e-05, "loss": 0.858, "step": 1292 }, { "epoch": 0.42, "grad_norm": 0.34320082385072354, "learning_rate": 1.9438652060959158e-05, "loss": 0.5881, "step": 1293 }, { "epoch": 0.42, "grad_norm": 0.36740622367454473, "learning_rate": 1.9423493525637495e-05, "loss": 0.6171, "step": 1294 }, { "epoch": 0.42, "grad_norm": 0.4151312583846056, "learning_rate": 1.940833004091984e-05, "loss": 0.8816, "step": 1295 }, { "epoch": 0.42, "grad_norm": 0.3554126251212465, "learning_rate": 1.9393161623772443e-05, "loss": 0.5891, "step": 1296 }, { "epoch": 0.42, "grad_norm": 0.30423658673208875, "learning_rate": 1.9377988291167068e-05, "loss": 0.6117, "step": 1297 }, { "epoch": 0.42, "grad_norm": 0.40354875845587707, "learning_rate": 1.9362810060080987e-05, "loss": 0.8352, "step": 1298 }, { "epoch": 0.42, "grad_norm": 0.3167799939178324, "learning_rate": 1.9347626947496945e-05, "loss": 0.6089, "step": 1299 }, { "epoch": 0.42, "grad_norm": 0.35002065855800757, "learning_rate": 1.9332438970403156e-05, "loss": 0.5843, "step": 1300 }, { "epoch": 0.42, "grad_norm": 0.3541477757825453, "learning_rate": 1.9317246145793267e-05, "loss": 0.6604, "step": 1301 }, { "epoch": 0.43, "grad_norm": 0.3477261765446466, "learning_rate": 1.9302048490666356e-05, "loss": 0.627, "step": 1302 }, { "epoch": 0.43, "grad_norm": 0.36520788522687847, "learning_rate": 1.9286846022026904e-05, "loss": 0.9191, "step": 1303 }, { "epoch": 0.43, "grad_norm": 0.3597262402831081, "learning_rate": 1.9271638756884787e-05, "loss": 0.6031, "step": 1304 }, { "epoch": 0.43, "grad_norm": 0.29984408209777863, "learning_rate": 1.9256426712255228e-05, "loss": 0.6035, "step": 1305 }, { "epoch": 0.43, "grad_norm": 0.38039950717511695, "learning_rate": 1.9241209905158807e-05, "loss": 0.8576, "step": 1306 }, { "epoch": 0.43, "grad_norm": 0.3317143864946089, "learning_rate": 1.9225988352621445e-05, "loss": 0.6023, "step": 1307 }, { "epoch": 0.43, "grad_norm": 0.3560928796097481, "learning_rate": 1.9210762071674358e-05, "loss": 0.6122, "step": 1308 }, { "epoch": 0.43, "grad_norm": 0.3815490704553757, "learning_rate": 1.9195531079354056e-05, "loss": 0.8445, "step": 1309 }, { "epoch": 0.43, "grad_norm": 0.31706964470538757, "learning_rate": 1.918029539270232e-05, "loss": 0.597, "step": 1310 }, { "epoch": 0.43, "grad_norm": 0.7961490197294379, "learning_rate": 1.9165055028766178e-05, "loss": 0.6347, "step": 1311 }, { "epoch": 0.43, "grad_norm": 0.3589320255500845, "learning_rate": 1.9149810004597906e-05, "loss": 0.6435, "step": 1312 }, { "epoch": 0.43, "grad_norm": 0.3254262159889682, "learning_rate": 1.9134560337254986e-05, "loss": 0.6018, "step": 1313 }, { "epoch": 0.43, "grad_norm": 0.39168386544663747, "learning_rate": 1.91193060438001e-05, "loss": 0.867, "step": 1314 }, { "epoch": 0.43, "grad_norm": 0.3172006464058323, "learning_rate": 1.9104047141301088e-05, "loss": 0.6088, "step": 1315 }, { "epoch": 0.43, "grad_norm": 0.3349382064098492, "learning_rate": 1.908878364683097e-05, "loss": 0.5974, "step": 1316 }, { "epoch": 0.43, "grad_norm": 0.4078228793783182, "learning_rate": 1.9073515577467887e-05, "loss": 0.8788, "step": 1317 }, { "epoch": 0.43, "grad_norm": 0.33273136860095404, "learning_rate": 1.9058242950295118e-05, "loss": 0.613, "step": 1318 }, { "epoch": 0.43, "grad_norm": 0.3405960055261859, "learning_rate": 1.904296578240102e-05, "loss": 0.6167, "step": 1319 }, { "epoch": 0.43, "grad_norm": 0.3737766491570266, "learning_rate": 1.902768409087904e-05, "loss": 0.8506, "step": 1320 }, { "epoch": 0.43, "grad_norm": 0.36429340193223053, "learning_rate": 1.9012397892827695e-05, "loss": 0.6044, "step": 1321 }, { "epoch": 0.43, "grad_norm": 0.33556328743918723, "learning_rate": 1.8997107205350525e-05, "loss": 0.6045, "step": 1322 }, { "epoch": 0.43, "grad_norm": 0.3503407972684935, "learning_rate": 1.8981812045556113e-05, "loss": 0.6535, "step": 1323 }, { "epoch": 0.43, "grad_norm": 0.3948225031042218, "learning_rate": 1.8966512430558036e-05, "loss": 0.8689, "step": 1324 }, { "epoch": 0.43, "grad_norm": 0.31547418795028315, "learning_rate": 1.895120837747485e-05, "loss": 0.583, "step": 1325 }, { "epoch": 0.43, "grad_norm": 0.5731058585510378, "learning_rate": 1.8935899903430095e-05, "loss": 0.6055, "step": 1326 }, { "epoch": 0.43, "grad_norm": 0.47518936022044866, "learning_rate": 1.8920587025552237e-05, "loss": 0.6218, "step": 1327 }, { "epoch": 0.43, "grad_norm": 0.4029252773642662, "learning_rate": 1.8905269760974683e-05, "loss": 0.8595, "step": 1328 }, { "epoch": 0.43, "grad_norm": 0.3426789993067246, "learning_rate": 1.8889948126835745e-05, "loss": 0.6249, "step": 1329 }, { "epoch": 0.43, "grad_norm": 0.4135155792425198, "learning_rate": 1.8874622140278618e-05, "loss": 0.5915, "step": 1330 }, { "epoch": 0.43, "grad_norm": 0.3730431703607662, "learning_rate": 1.8859291818451375e-05, "loss": 0.8999, "step": 1331 }, { "epoch": 0.44, "grad_norm": 0.4141030493171409, "learning_rate": 1.884395717850694e-05, "loss": 0.5864, "step": 1332 }, { "epoch": 0.44, "grad_norm": 0.3133927930655968, "learning_rate": 1.8828618237603062e-05, "loss": 0.6148, "step": 1333 }, { "epoch": 0.44, "grad_norm": 0.4112841267274145, "learning_rate": 1.8813275012902307e-05, "loss": 0.6556, "step": 1334 }, { "epoch": 0.44, "grad_norm": 0.36522144061047906, "learning_rate": 1.879792752157203e-05, "loss": 0.8846, "step": 1335 }, { "epoch": 0.44, "grad_norm": 0.3700227846095818, "learning_rate": 1.878257578078436e-05, "loss": 0.6048, "step": 1336 }, { "epoch": 0.44, "grad_norm": 0.3148433922160107, "learning_rate": 1.8767219807716186e-05, "loss": 0.6222, "step": 1337 }, { "epoch": 0.44, "grad_norm": 0.4061363731261511, "learning_rate": 1.8751859619549138e-05, "loss": 0.8856, "step": 1338 }, { "epoch": 0.44, "grad_norm": 0.3293175167776505, "learning_rate": 1.8736495233469537e-05, "loss": 0.6054, "step": 1339 }, { "epoch": 0.44, "grad_norm": 0.323619769223986, "learning_rate": 1.8721126666668435e-05, "loss": 0.6094, "step": 1340 }, { "epoch": 0.44, "grad_norm": 0.6554879702965036, "learning_rate": 1.870575393634153e-05, "loss": 0.8864, "step": 1341 }, { "epoch": 0.44, "grad_norm": 0.33658824530882625, "learning_rate": 1.8690377059689202e-05, "loss": 0.5979, "step": 1342 }, { "epoch": 0.44, "grad_norm": 0.3447222598197727, "learning_rate": 1.8674996053916458e-05, "loss": 0.5973, "step": 1343 }, { "epoch": 0.44, "grad_norm": 0.3885135271212663, "learning_rate": 1.8659610936232927e-05, "loss": 0.8655, "step": 1344 }, { "epoch": 0.44, "grad_norm": 0.4967509595051533, "learning_rate": 1.8644221723852843e-05, "loss": 0.422, "step": 1345 }, { "epoch": 0.44, "grad_norm": 0.40071042870857587, "learning_rate": 1.8628828433995013e-05, "loss": 0.8402, "step": 1346 }, { "epoch": 0.44, "grad_norm": 0.38113247262047223, "learning_rate": 1.8613431083882827e-05, "loss": 0.603, "step": 1347 }, { "epoch": 0.44, "grad_norm": 0.33457999571818353, "learning_rate": 1.8598029690744184e-05, "loss": 0.5875, "step": 1348 }, { "epoch": 0.44, "grad_norm": 0.41679747928210237, "learning_rate": 1.8582624271811534e-05, "loss": 0.8654, "step": 1349 }, { "epoch": 0.44, "grad_norm": 0.3399894108949177, "learning_rate": 1.856721484432182e-05, "loss": 0.588, "step": 1350 }, { "epoch": 0.44, "grad_norm": 0.3843620280610406, "learning_rate": 1.855180142551648e-05, "loss": 0.6143, "step": 1351 }, { "epoch": 0.44, "grad_norm": 0.40420708275229705, "learning_rate": 1.853638403264141e-05, "loss": 0.8765, "step": 1352 }, { "epoch": 0.44, "grad_norm": 0.3627903057682197, "learning_rate": 1.8520962682946948e-05, "loss": 0.5984, "step": 1353 }, { "epoch": 0.44, "grad_norm": 0.30181385785927384, "learning_rate": 1.850553739368786e-05, "loss": 0.5942, "step": 1354 }, { "epoch": 0.44, "grad_norm": 0.4201508309564684, "learning_rate": 1.8490108182123335e-05, "loss": 0.8794, "step": 1355 }, { "epoch": 0.44, "grad_norm": 0.3464491124609798, "learning_rate": 1.847467506551694e-05, "loss": 0.4091, "step": 1356 }, { "epoch": 0.44, "grad_norm": 0.4599583095205838, "learning_rate": 1.8459238061136604e-05, "loss": 0.8742, "step": 1357 }, { "epoch": 0.44, "grad_norm": 0.3800725660780907, "learning_rate": 1.8443797186254616e-05, "loss": 0.5858, "step": 1358 }, { "epoch": 0.44, "grad_norm": 0.4007185021969695, "learning_rate": 1.8428352458147602e-05, "loss": 0.6222, "step": 1359 }, { "epoch": 0.44, "grad_norm": 0.4029035708965362, "learning_rate": 1.8412903894096472e-05, "loss": 0.8403, "step": 1360 }, { "epoch": 0.44, "grad_norm": 0.37552624651102245, "learning_rate": 1.839745151138647e-05, "loss": 0.6536, "step": 1361 }, { "epoch": 0.44, "grad_norm": 1.7903001664416578, "learning_rate": 1.8381995327307067e-05, "loss": 0.62, "step": 1362 }, { "epoch": 0.45, "grad_norm": 0.3380829017859619, "learning_rate": 1.8366535359152022e-05, "loss": 0.5815, "step": 1363 }, { "epoch": 0.45, "grad_norm": 0.41692187703376354, "learning_rate": 1.8351071624219313e-05, "loss": 0.8503, "step": 1364 }, { "epoch": 0.45, "grad_norm": 0.37137520959947445, "learning_rate": 1.8335604139811135e-05, "loss": 0.5676, "step": 1365 }, { "epoch": 0.45, "grad_norm": 0.4545320266302253, "learning_rate": 1.832013292323388e-05, "loss": 0.8625, "step": 1366 }, { "epoch": 0.45, "grad_norm": 0.4002715610756697, "learning_rate": 1.8304657991798112e-05, "loss": 0.6417, "step": 1367 }, { "epoch": 0.45, "grad_norm": 0.3798338026885802, "learning_rate": 1.828917936281855e-05, "loss": 0.5834, "step": 1368 }, { "epoch": 0.45, "grad_norm": 0.3806045235228948, "learning_rate": 1.827369705361406e-05, "loss": 0.5927, "step": 1369 }, { "epoch": 0.45, "grad_norm": 0.40816800427771344, "learning_rate": 1.8258211081507616e-05, "loss": 0.84, "step": 1370 }, { "epoch": 0.45, "grad_norm": 0.4114234802300347, "learning_rate": 1.82427214638263e-05, "loss": 0.6156, "step": 1371 }, { "epoch": 0.45, "grad_norm": 0.39920241315531824, "learning_rate": 1.822722821790126e-05, "loss": 0.6481, "step": 1372 }, { "epoch": 0.45, "grad_norm": 0.40654563336649735, "learning_rate": 1.8211731361067708e-05, "loss": 0.606, "step": 1373 }, { "epoch": 0.45, "grad_norm": 0.32534699636294584, "learning_rate": 1.8196230910664904e-05, "loss": 0.6174, "step": 1374 }, { "epoch": 0.45, "grad_norm": 0.4248751616483441, "learning_rate": 1.818072688403613e-05, "loss": 0.8517, "step": 1375 }, { "epoch": 0.45, "grad_norm": 0.32093211578202624, "learning_rate": 1.8165219298528646e-05, "loss": 0.5986, "step": 1376 }, { "epoch": 0.45, "grad_norm": 0.3538109404769192, "learning_rate": 1.8149708171493725e-05, "loss": 0.5869, "step": 1377 }, { "epoch": 0.45, "grad_norm": 0.38368325574568846, "learning_rate": 1.8134193520286587e-05, "loss": 0.8769, "step": 1378 }, { "epoch": 0.45, "grad_norm": 0.32859018355423353, "learning_rate": 1.8118675362266388e-05, "loss": 0.6142, "step": 1379 }, { "epoch": 0.45, "grad_norm": 0.3453919757726037, "learning_rate": 1.810315371479623e-05, "loss": 0.5842, "step": 1380 }, { "epoch": 0.45, "grad_norm": 0.3726746607593359, "learning_rate": 1.8087628595243096e-05, "loss": 0.8731, "step": 1381 }, { "epoch": 0.45, "grad_norm": 0.38132632229738495, "learning_rate": 1.807210002097786e-05, "loss": 0.5984, "step": 1382 }, { "epoch": 0.45, "grad_norm": 0.37463253661826906, "learning_rate": 1.8056568009375274e-05, "loss": 0.6256, "step": 1383 }, { "epoch": 0.45, "grad_norm": 0.38420406964651094, "learning_rate": 1.8041032577813926e-05, "loss": 0.5858, "step": 1384 }, { "epoch": 0.45, "grad_norm": 0.31868998085883105, "learning_rate": 1.802549374367623e-05, "loss": 0.5882, "step": 1385 }, { "epoch": 0.45, "grad_norm": 0.4039432913760993, "learning_rate": 1.8009951524348402e-05, "loss": 0.8823, "step": 1386 }, { "epoch": 0.45, "grad_norm": 0.32250463267547996, "learning_rate": 1.799440593722046e-05, "loss": 0.6235, "step": 1387 }, { "epoch": 0.45, "grad_norm": 1.0347623742035659, "learning_rate": 1.7978856999686182e-05, "loss": 0.62, "step": 1388 }, { "epoch": 0.45, "grad_norm": 0.41261136891348105, "learning_rate": 1.7963304729143094e-05, "loss": 0.8405, "step": 1389 }, { "epoch": 0.45, "grad_norm": 0.3478211554930761, "learning_rate": 1.794774914299245e-05, "loss": 0.6119, "step": 1390 }, { "epoch": 0.45, "grad_norm": 0.31384344097227496, "learning_rate": 1.7932190258639224e-05, "loss": 0.616, "step": 1391 }, { "epoch": 0.45, "grad_norm": 0.38876186635007565, "learning_rate": 1.791662809349206e-05, "loss": 0.8492, "step": 1392 }, { "epoch": 0.45, "grad_norm": 0.33276713235271027, "learning_rate": 1.79010626649633e-05, "loss": 0.6034, "step": 1393 }, { "epoch": 0.46, "grad_norm": 0.33518023440604555, "learning_rate": 1.7885493990468918e-05, "loss": 0.6383, "step": 1394 }, { "epoch": 0.46, "grad_norm": 0.37111944575411915, "learning_rate": 1.786992208742851e-05, "loss": 0.6002, "step": 1395 }, { "epoch": 0.46, "grad_norm": 0.4139238213334895, "learning_rate": 1.785434697326532e-05, "loss": 0.8756, "step": 1396 }, { "epoch": 0.46, "grad_norm": 0.3248975370026262, "learning_rate": 1.7838768665406154e-05, "loss": 0.5881, "step": 1397 }, { "epoch": 0.46, "grad_norm": 0.3240534357317988, "learning_rate": 1.7823187181281402e-05, "loss": 0.6037, "step": 1398 }, { "epoch": 0.46, "grad_norm": 0.34004978062510255, "learning_rate": 1.780760253832501e-05, "loss": 0.6075, "step": 1399 }, { "epoch": 0.46, "grad_norm": 0.37301262917345973, "learning_rate": 1.7792014753974452e-05, "loss": 0.8828, "step": 1400 }, { "epoch": 0.46, "grad_norm": 0.3347287908124473, "learning_rate": 1.777642384567072e-05, "loss": 0.6033, "step": 1401 }, { "epoch": 0.46, "grad_norm": 0.32876363636695977, "learning_rate": 1.7760829830858305e-05, "loss": 0.6077, "step": 1402 }, { "epoch": 0.46, "grad_norm": 0.538256297735946, "learning_rate": 1.7745232726985168e-05, "loss": 0.8907, "step": 1403 }, { "epoch": 0.46, "grad_norm": 0.33461592958304914, "learning_rate": 1.772963255150274e-05, "loss": 0.6041, "step": 1404 }, { "epoch": 0.46, "grad_norm": 0.3455691138423074, "learning_rate": 1.771402932186586e-05, "loss": 0.6432, "step": 1405 }, { "epoch": 0.46, "grad_norm": 0.33243108885021, "learning_rate": 1.7698423055532808e-05, "loss": 0.6019, "step": 1406 }, { "epoch": 0.46, "grad_norm": 0.35372893206764155, "learning_rate": 1.7682813769965263e-05, "loss": 0.8597, "step": 1407 }, { "epoch": 0.46, "grad_norm": 0.36016016845887555, "learning_rate": 1.7667201482628272e-05, "loss": 0.6044, "step": 1408 }, { "epoch": 0.46, "grad_norm": 0.3035435523595379, "learning_rate": 1.7651586210990236e-05, "loss": 0.6079, "step": 1409 }, { "epoch": 0.46, "grad_norm": 0.3735959641846319, "learning_rate": 1.7635967972522913e-05, "loss": 0.8721, "step": 1410 }, { "epoch": 0.46, "grad_norm": 0.3142068656476621, "learning_rate": 1.762034678470136e-05, "loss": 0.6082, "step": 1411 }, { "epoch": 0.46, "grad_norm": 0.31124330101281794, "learning_rate": 1.760472266500396e-05, "loss": 0.572, "step": 1412 }, { "epoch": 0.46, "grad_norm": 0.37900114849089184, "learning_rate": 1.7589095630912345e-05, "loss": 0.8404, "step": 1413 }, { "epoch": 0.46, "grad_norm": 0.3096748950858377, "learning_rate": 1.7573465699911423e-05, "loss": 0.5891, "step": 1414 }, { "epoch": 0.46, "grad_norm": 0.3266952291946601, "learning_rate": 1.755783288948936e-05, "loss": 0.5945, "step": 1415 }, { "epoch": 0.46, "grad_norm": 0.4045208938244444, "learning_rate": 1.754219721713751e-05, "loss": 0.6729, "step": 1416 }, { "epoch": 0.46, "grad_norm": 0.3340903007058062, "learning_rate": 1.752655870035046e-05, "loss": 0.6071, "step": 1417 }, { "epoch": 0.46, "grad_norm": 0.34799671071233396, "learning_rate": 1.7510917356625962e-05, "loss": 0.8693, "step": 1418 }, { "epoch": 0.46, "grad_norm": 0.34324082597669253, "learning_rate": 1.7495273203464928e-05, "loss": 0.6339, "step": 1419 }, { "epoch": 0.46, "grad_norm": 0.32278054619899094, "learning_rate": 1.7479626258371435e-05, "loss": 0.6039, "step": 1420 }, { "epoch": 0.46, "grad_norm": 0.3530623822225169, "learning_rate": 1.7463976538852656e-05, "loss": 0.8617, "step": 1421 }, { "epoch": 0.46, "grad_norm": 0.339001494343573, "learning_rate": 1.744832406241889e-05, "loss": 0.6048, "step": 1422 }, { "epoch": 0.46, "grad_norm": 0.29358542188736575, "learning_rate": 1.7432668846583514e-05, "loss": 0.6033, "step": 1423 }, { "epoch": 0.47, "grad_norm": 0.3740625366616867, "learning_rate": 1.7417010908862964e-05, "loss": 0.8569, "step": 1424 }, { "epoch": 0.47, "grad_norm": 0.341332683009136, "learning_rate": 1.7401350266776726e-05, "loss": 0.5846, "step": 1425 }, { "epoch": 0.47, "grad_norm": 0.3272328005349161, "learning_rate": 1.7385686937847317e-05, "loss": 0.6104, "step": 1426 }, { "epoch": 0.47, "grad_norm": 0.35832660402298144, "learning_rate": 1.737002093960025e-05, "loss": 0.6635, "step": 1427 }, { "epoch": 0.47, "grad_norm": 0.3158133457209271, "learning_rate": 1.7354352289564034e-05, "loss": 0.6179, "step": 1428 }, { "epoch": 0.47, "grad_norm": 0.3717870052917457, "learning_rate": 1.7338681005270143e-05, "loss": 0.8859, "step": 1429 }, { "epoch": 0.47, "grad_norm": 0.5003991894857002, "learning_rate": 1.7323007104252987e-05, "loss": 0.6351, "step": 1430 }, { "epoch": 0.47, "grad_norm": 0.3012569361744865, "learning_rate": 1.7307330604049927e-05, "loss": 0.6127, "step": 1431 }, { "epoch": 0.47, "grad_norm": 0.37686482242732633, "learning_rate": 1.7291651522201208e-05, "loss": 0.8192, "step": 1432 }, { "epoch": 0.47, "grad_norm": 0.34615190662924594, "learning_rate": 1.7275969876249974e-05, "loss": 0.5992, "step": 1433 }, { "epoch": 0.47, "grad_norm": 0.32415020260798977, "learning_rate": 1.7260285683742248e-05, "loss": 0.6005, "step": 1434 }, { "epoch": 0.47, "grad_norm": 0.6168483497507949, "learning_rate": 1.7244598962226882e-05, "loss": 0.8538, "step": 1435 }, { "epoch": 0.47, "grad_norm": 0.323139070968343, "learning_rate": 1.7228909729255576e-05, "loss": 0.6023, "step": 1436 }, { "epoch": 0.47, "grad_norm": 0.30008117736083717, "learning_rate": 1.721321800238283e-05, "loss": 0.6113, "step": 1437 }, { "epoch": 0.47, "grad_norm": 0.36242335977456136, "learning_rate": 1.7197523799165932e-05, "loss": 0.6386, "step": 1438 }, { "epoch": 0.47, "grad_norm": 0.3660742212013189, "learning_rate": 1.7181827137164954e-05, "loss": 0.8501, "step": 1439 }, { "epoch": 0.47, "grad_norm": 0.714604806408132, "learning_rate": 1.71661280339427e-05, "loss": 0.608, "step": 1440 }, { "epoch": 0.47, "grad_norm": 0.3460833996983384, "learning_rate": 1.7150426507064722e-05, "loss": 0.6079, "step": 1441 }, { "epoch": 0.47, "grad_norm": 0.3701290569288406, "learning_rate": 1.713472257409928e-05, "loss": 0.8231, "step": 1442 }, { "epoch": 0.47, "grad_norm": 0.3255899542461123, "learning_rate": 1.7119016252617314e-05, "loss": 0.5822, "step": 1443 }, { "epoch": 0.47, "grad_norm": 0.3504222100266835, "learning_rate": 1.7103307560192458e-05, "loss": 0.6606, "step": 1444 }, { "epoch": 0.47, "grad_norm": 0.3133589873512018, "learning_rate": 1.7087596514400982e-05, "loss": 0.581, "step": 1445 }, { "epoch": 0.47, "grad_norm": 0.3668108483599513, "learning_rate": 1.70718831328218e-05, "loss": 0.6205, "step": 1446 }, { "epoch": 0.47, "grad_norm": 0.36909223482932557, "learning_rate": 1.705616743303642e-05, "loss": 0.8429, "step": 1447 }, { "epoch": 0.47, "grad_norm": 0.30620630586533687, "learning_rate": 1.7040449432628964e-05, "loss": 0.5652, "step": 1448 }, { "epoch": 0.47, "grad_norm": 0.3661299580557353, "learning_rate": 1.7024729149186127e-05, "loss": 0.6358, "step": 1449 }, { "epoch": 0.47, "grad_norm": 0.4009644893843763, "learning_rate": 1.700900660029715e-05, "loss": 0.8699, "step": 1450 }, { "epoch": 0.47, "grad_norm": 0.31472218946089797, "learning_rate": 1.6993281803553814e-05, "loss": 0.5844, "step": 1451 }, { "epoch": 0.47, "grad_norm": 0.3382957822986683, "learning_rate": 1.6977554776550403e-05, "loss": 0.5846, "step": 1452 }, { "epoch": 0.47, "grad_norm": 0.38056776572741313, "learning_rate": 1.6961825536883725e-05, "loss": 0.8582, "step": 1453 }, { "epoch": 0.47, "grad_norm": 0.3285342462452088, "learning_rate": 1.6946094102153024e-05, "loss": 0.5825, "step": 1454 }, { "epoch": 0.48, "grad_norm": 0.3442662185843514, "learning_rate": 1.6930360489960042e-05, "loss": 0.6459, "step": 1455 }, { "epoch": 0.48, "grad_norm": 0.31321697737210474, "learning_rate": 1.6914624717908922e-05, "loss": 0.5941, "step": 1456 }, { "epoch": 0.48, "grad_norm": 0.310141919576954, "learning_rate": 1.689888680360624e-05, "loss": 0.5985, "step": 1457 }, { "epoch": 0.48, "grad_norm": 0.47495406234633036, "learning_rate": 1.6883146764660975e-05, "loss": 0.8321, "step": 1458 }, { "epoch": 0.48, "grad_norm": 0.3228467495774374, "learning_rate": 1.6867404618684464e-05, "loss": 0.6047, "step": 1459 }, { "epoch": 0.48, "grad_norm": 0.31235132546705435, "learning_rate": 1.6851660383290422e-05, "loss": 0.5947, "step": 1460 }, { "epoch": 0.48, "grad_norm": 0.37850678785654945, "learning_rate": 1.6835914076094893e-05, "loss": 0.8707, "step": 1461 }, { "epoch": 0.48, "grad_norm": 0.3277601884691968, "learning_rate": 1.682016571471623e-05, "loss": 0.583, "step": 1462 }, { "epoch": 0.48, "grad_norm": 0.3789469157916823, "learning_rate": 1.6804415316775102e-05, "loss": 0.6156, "step": 1463 }, { "epoch": 0.48, "grad_norm": 0.36806070758572107, "learning_rate": 1.6788662899894444e-05, "loss": 0.8773, "step": 1464 }, { "epoch": 0.48, "grad_norm": 0.3581304080961864, "learning_rate": 1.6772908481699457e-05, "loss": 0.6113, "step": 1465 }, { "epoch": 0.48, "grad_norm": 0.3272201104067181, "learning_rate": 1.6757152079817576e-05, "loss": 0.6407, "step": 1466 }, { "epoch": 0.48, "grad_norm": 0.329789025754953, "learning_rate": 1.6741393711878455e-05, "loss": 0.5985, "step": 1467 }, { "epoch": 0.48, "grad_norm": 0.3494990245112794, "learning_rate": 1.6725633395513952e-05, "loss": 0.8449, "step": 1468 }, { "epoch": 0.48, "grad_norm": 0.34736118090009876, "learning_rate": 1.670987114835811e-05, "loss": 0.6337, "step": 1469 }, { "epoch": 0.48, "grad_norm": 0.29448553728639665, "learning_rate": 1.669410698804712e-05, "loss": 0.5952, "step": 1470 }, { "epoch": 0.48, "grad_norm": 0.32082346546565504, "learning_rate": 1.667834093221932e-05, "loss": 0.5938, "step": 1471 }, { "epoch": 0.48, "grad_norm": 0.372130595565937, "learning_rate": 1.6662572998515166e-05, "loss": 0.8723, "step": 1472 }, { "epoch": 0.48, "grad_norm": 0.31011166373405336, "learning_rate": 1.664680320457722e-05, "loss": 0.5979, "step": 1473 }, { "epoch": 0.48, "grad_norm": 0.30639616155650834, "learning_rate": 1.663103156805013e-05, "loss": 0.6046, "step": 1474 }, { "epoch": 0.48, "grad_norm": 3.642112127653775, "learning_rate": 1.661525810658059e-05, "loss": 0.8922, "step": 1475 }, { "epoch": 0.48, "grad_norm": 0.31466587699289883, "learning_rate": 1.659948283781734e-05, "loss": 0.6127, "step": 1476 }, { "epoch": 0.48, "grad_norm": 0.3477214934632274, "learning_rate": 1.6583705779411158e-05, "loss": 0.6579, "step": 1477 }, { "epoch": 0.48, "grad_norm": 0.3110451096853916, "learning_rate": 1.6567926949014805e-05, "loss": 0.6011, "step": 1478 }, { "epoch": 0.48, "grad_norm": 0.36428235920596574, "learning_rate": 1.6552146364283036e-05, "loss": 0.8974, "step": 1479 }, { "epoch": 0.48, "grad_norm": 0.9856314906336292, "learning_rate": 1.6536364042872567e-05, "loss": 0.637, "step": 1480 }, { "epoch": 0.48, "grad_norm": 0.32239204856338005, "learning_rate": 1.652058000244205e-05, "loss": 0.5707, "step": 1481 }, { "epoch": 0.48, "grad_norm": 0.38131404204190184, "learning_rate": 1.6504794260652077e-05, "loss": 0.8773, "step": 1482 }, { "epoch": 0.48, "grad_norm": 0.3094808681689915, "learning_rate": 1.6489006835165122e-05, "loss": 0.6091, "step": 1483 }, { "epoch": 0.48, "grad_norm": 0.3234141946607025, "learning_rate": 1.647321774364556e-05, "loss": 0.5944, "step": 1484 }, { "epoch": 0.48, "grad_norm": 0.3577060837216909, "learning_rate": 1.645742700375962e-05, "loss": 0.8224, "step": 1485 }, { "epoch": 0.49, "grad_norm": 0.3086244404985481, "learning_rate": 1.6441634633175378e-05, "loss": 0.6096, "step": 1486 }, { "epoch": 0.49, "grad_norm": 0.2997315478798499, "learning_rate": 1.6425840649562737e-05, "loss": 0.6047, "step": 1487 }, { "epoch": 0.49, "grad_norm": 0.35058411312201043, "learning_rate": 1.6410045070593413e-05, "loss": 0.6661, "step": 1488 }, { "epoch": 0.49, "grad_norm": 0.30439941245763547, "learning_rate": 1.6394247913940888e-05, "loss": 0.6034, "step": 1489 }, { "epoch": 0.49, "grad_norm": 0.379518703418073, "learning_rate": 1.6378449197280414e-05, "loss": 0.8768, "step": 1490 }, { "epoch": 0.49, "grad_norm": 0.31603343672979417, "learning_rate": 1.6362648938289004e-05, "loss": 0.5973, "step": 1491 }, { "epoch": 0.49, "grad_norm": 0.304359290507754, "learning_rate": 1.6346847154645376e-05, "loss": 0.5667, "step": 1492 }, { "epoch": 0.49, "grad_norm": 0.36002103568230015, "learning_rate": 1.6331043864029972e-05, "loss": 0.8527, "step": 1493 }, { "epoch": 0.49, "grad_norm": 0.3112668458962097, "learning_rate": 1.631523908412491e-05, "loss": 0.5897, "step": 1494 }, { "epoch": 0.49, "grad_norm": 0.31056784510829016, "learning_rate": 1.629943283261397e-05, "loss": 0.6118, "step": 1495 }, { "epoch": 0.49, "grad_norm": 0.38357594866725103, "learning_rate": 1.6283625127182596e-05, "loss": 0.8848, "step": 1496 }, { "epoch": 0.49, "grad_norm": 0.3401756701247233, "learning_rate": 1.6267815985517838e-05, "loss": 0.611, "step": 1497 }, { "epoch": 0.49, "grad_norm": 0.30289299980892365, "learning_rate": 1.6252005425308372e-05, "loss": 0.6041, "step": 1498 }, { "epoch": 0.49, "grad_norm": 0.36954668456791445, "learning_rate": 1.6236193464244447e-05, "loss": 0.6304, "step": 1499 }, { "epoch": 0.49, "grad_norm": 0.323186691998162, "learning_rate": 1.6220380120017874e-05, "loss": 0.6048, "step": 1500 }, { "epoch": 0.49, "grad_norm": 0.4247479472604402, "learning_rate": 1.620456541032204e-05, "loss": 0.8293, "step": 1501 }, { "epoch": 0.49, "grad_norm": 0.33029832578165425, "learning_rate": 1.6188749352851825e-05, "loss": 0.6282, "step": 1502 }, { "epoch": 0.49, "grad_norm": 0.3683360925866696, "learning_rate": 1.617293196530365e-05, "loss": 0.5783, "step": 1503 }, { "epoch": 0.49, "grad_norm": 0.4439254779247088, "learning_rate": 1.61571132653754e-05, "loss": 0.8416, "step": 1504 }, { "epoch": 0.49, "grad_norm": 0.31320974122850753, "learning_rate": 1.6141293270766426e-05, "loss": 0.6062, "step": 1505 }, { "epoch": 0.49, "grad_norm": 0.33788666218031926, "learning_rate": 1.6125471999177555e-05, "loss": 0.5867, "step": 1506 }, { "epoch": 0.49, "grad_norm": 0.3912696053537718, "learning_rate": 1.6109649468311015e-05, "loss": 0.8962, "step": 1507 }, { "epoch": 0.49, "grad_norm": 0.3252517034801483, "learning_rate": 1.6093825695870465e-05, "loss": 0.6098, "step": 1508 }, { "epoch": 0.49, "grad_norm": 0.3203935205451462, "learning_rate": 1.6078000699560927e-05, "loss": 0.6202, "step": 1509 }, { "epoch": 0.49, "grad_norm": 0.34821675468909413, "learning_rate": 1.6062174497088813e-05, "loss": 0.6511, "step": 1510 }, { "epoch": 0.49, "grad_norm": 0.3520930369857388, "learning_rate": 1.604634710616188e-05, "loss": 0.8172, "step": 1511 }, { "epoch": 0.49, "grad_norm": 0.39909327941443595, "learning_rate": 1.6030518544489215e-05, "loss": 0.6195, "step": 1512 }, { "epoch": 0.49, "grad_norm": 0.29164395597543785, "learning_rate": 1.6014688829781212e-05, "loss": 0.6206, "step": 1513 }, { "epoch": 0.49, "grad_norm": 0.30825992539520325, "learning_rate": 1.5998857979749562e-05, "loss": 0.5976, "step": 1514 }, { "epoch": 0.49, "grad_norm": 0.3777189628290599, "learning_rate": 1.5983026012107208e-05, "loss": 0.8836, "step": 1515 }, { "epoch": 0.5, "grad_norm": 0.3149139637636734, "learning_rate": 1.5967192944568364e-05, "loss": 0.5929, "step": 1516 }, { "epoch": 0.5, "grad_norm": 0.30098592187240053, "learning_rate": 1.5951358794848467e-05, "loss": 0.6059, "step": 1517 }, { "epoch": 0.5, "grad_norm": 0.34173103954322454, "learning_rate": 1.593552358066417e-05, "loss": 0.854, "step": 1518 }, { "epoch": 0.5, "grad_norm": 0.3535368732709004, "learning_rate": 1.59196873197333e-05, "loss": 0.6303, "step": 1519 }, { "epoch": 0.5, "grad_norm": 0.3058866292929864, "learning_rate": 1.590385002977488e-05, "loss": 0.5826, "step": 1520 }, { "epoch": 0.5, "grad_norm": 0.3442323914049554, "learning_rate": 1.5888011728509058e-05, "loss": 0.6282, "step": 1521 }, { "epoch": 0.5, "grad_norm": 0.3615222739855465, "learning_rate": 1.587217243365714e-05, "loss": 0.864, "step": 1522 }, { "epoch": 0.5, "grad_norm": 0.2965437700461475, "learning_rate": 1.5856332162941523e-05, "loss": 0.6028, "step": 1523 }, { "epoch": 0.5, "grad_norm": 0.30792998174406344, "learning_rate": 1.5840490934085703e-05, "loss": 0.6108, "step": 1524 }, { "epoch": 0.5, "grad_norm": 0.36154727079559795, "learning_rate": 1.5824648764814253e-05, "loss": 0.8591, "step": 1525 }, { "epoch": 0.5, "grad_norm": 0.29579322152422527, "learning_rate": 1.5808805672852792e-05, "loss": 0.414, "step": 1526 }, { "epoch": 0.5, "grad_norm": 0.34956252877380517, "learning_rate": 1.579296167592797e-05, "loss": 0.8716, "step": 1527 }, { "epoch": 0.5, "grad_norm": 0.3193624389518217, "learning_rate": 1.577711679176746e-05, "loss": 0.5911, "step": 1528 }, { "epoch": 0.5, "grad_norm": 0.2982157380086912, "learning_rate": 1.5761271038099913e-05, "loss": 0.5899, "step": 1529 }, { "epoch": 0.5, "grad_norm": 0.38949904513178973, "learning_rate": 1.5745424432654962e-05, "loss": 0.8512, "step": 1530 }, { "epoch": 0.5, "grad_norm": 0.3204187013857067, "learning_rate": 1.57295769931632e-05, "loss": 0.7348, "step": 1531 }, { "epoch": 0.5, "grad_norm": 0.3903903263854228, "learning_rate": 1.5713728737356138e-05, "loss": 0.4014, "step": 1532 }, { "epoch": 0.5, "grad_norm": 0.42108166725189017, "learning_rate": 1.5697879682966203e-05, "loss": 0.8581, "step": 1533 }, { "epoch": 0.5, "grad_norm": 0.34576115826707754, "learning_rate": 1.568202984772672e-05, "loss": 0.6006, "step": 1534 }, { "epoch": 0.5, "grad_norm": 0.32194053069547535, "learning_rate": 1.5666179249371892e-05, "loss": 0.5897, "step": 1535 }, { "epoch": 0.5, "grad_norm": 0.3694958336333634, "learning_rate": 1.5650327905636774e-05, "loss": 0.9253, "step": 1536 }, { "epoch": 0.5, "grad_norm": 0.35893451524708475, "learning_rate": 1.5634475834257247e-05, "loss": 0.6157, "step": 1537 }, { "epoch": 0.5, "grad_norm": 0.31660810264282246, "learning_rate": 1.5618623052970008e-05, "loss": 0.5966, "step": 1538 }, { "epoch": 0.5, "grad_norm": 0.3209280047735517, "learning_rate": 1.560276957951256e-05, "loss": 0.5874, "step": 1539 }, { "epoch": 0.5, "grad_norm": 0.3815459226381016, "learning_rate": 1.5586915431623164e-05, "loss": 0.8589, "step": 1540 }, { "epoch": 0.5, "grad_norm": 0.3043854762952063, "learning_rate": 1.557106062704085e-05, "loss": 0.5928, "step": 1541 }, { "epoch": 0.5, "grad_norm": 0.3106732866553624, "learning_rate": 1.555520518350537e-05, "loss": 0.6057, "step": 1542 }, { "epoch": 0.5, "grad_norm": 0.3002445423418336, "learning_rate": 1.5539349118757204e-05, "loss": 0.596, "step": 1543 }, { "epoch": 0.5, "grad_norm": 0.3428544907442446, "learning_rate": 1.552349245053752e-05, "loss": 0.8609, "step": 1544 }, { "epoch": 0.5, "grad_norm": 0.3070820897810376, "learning_rate": 1.550763519658815e-05, "loss": 0.6021, "step": 1545 }, { "epoch": 0.5, "grad_norm": 0.31873591205919793, "learning_rate": 1.5491777374651607e-05, "loss": 0.5905, "step": 1546 }, { "epoch": 0.51, "grad_norm": 0.3657859720644803, "learning_rate": 1.5475919002471016e-05, "loss": 0.8613, "step": 1547 }, { "epoch": 0.51, "grad_norm": 0.3716133523292356, "learning_rate": 1.5460060097790135e-05, "loss": 0.6469, "step": 1548 }, { "epoch": 0.51, "grad_norm": 0.29894033628443983, "learning_rate": 1.5444200678353302e-05, "loss": 0.5896, "step": 1549 }, { "epoch": 0.51, "grad_norm": 0.3107273938859716, "learning_rate": 1.5428340761905444e-05, "loss": 0.5773, "step": 1550 }, { "epoch": 0.51, "grad_norm": 0.344697837542585, "learning_rate": 1.541248036619204e-05, "loss": 0.8532, "step": 1551 }, { "epoch": 0.51, "grad_norm": 0.33072411336346563, "learning_rate": 1.5396619508959102e-05, "loss": 0.599, "step": 1552 }, { "epoch": 0.51, "grad_norm": 0.3071677346170479, "learning_rate": 1.5380758207953157e-05, "loss": 0.6009, "step": 1553 }, { "epoch": 0.51, "grad_norm": 0.3713782773180895, "learning_rate": 1.5364896480921233e-05, "loss": 0.84, "step": 1554 }, { "epoch": 0.51, "grad_norm": 0.3440973141265517, "learning_rate": 1.5349034345610843e-05, "loss": 0.6126, "step": 1555 }, { "epoch": 0.51, "grad_norm": 0.3124708513070531, "learning_rate": 1.5333171819769943e-05, "loss": 0.6096, "step": 1556 }, { "epoch": 0.51, "grad_norm": 0.3732479791325691, "learning_rate": 1.5317308921146926e-05, "loss": 0.8542, "step": 1557 }, { "epoch": 0.51, "grad_norm": 0.513133517182216, "learning_rate": 1.5301445667490604e-05, "loss": 0.6048, "step": 1558 }, { "epoch": 0.51, "grad_norm": 0.34668772101324274, "learning_rate": 1.52855820765502e-05, "loss": 0.6601, "step": 1559 }, { "epoch": 0.51, "grad_norm": 0.3190910464578522, "learning_rate": 1.5269718166075297e-05, "loss": 0.6035, "step": 1560 }, { "epoch": 0.51, "grad_norm": 0.30712445926313336, "learning_rate": 1.5253853953815843e-05, "loss": 0.6033, "step": 1561 }, { "epoch": 0.51, "grad_norm": 0.3674385075911983, "learning_rate": 1.523798945752212e-05, "loss": 0.8402, "step": 1562 }, { "epoch": 0.51, "grad_norm": 0.32540043319435685, "learning_rate": 1.5222124694944733e-05, "loss": 0.5896, "step": 1563 }, { "epoch": 0.51, "grad_norm": 0.3179169275748932, "learning_rate": 1.520625968383458e-05, "loss": 0.6058, "step": 1564 }, { "epoch": 0.51, "grad_norm": 0.3672848674889762, "learning_rate": 1.5190394441942845e-05, "loss": 0.8817, "step": 1565 }, { "epoch": 0.51, "grad_norm": 0.3153453660581932, "learning_rate": 1.5174528987020958e-05, "loss": 0.5999, "step": 1566 }, { "epoch": 0.51, "grad_norm": 0.2992562079214196, "learning_rate": 1.5158663336820595e-05, "loss": 0.5895, "step": 1567 }, { "epoch": 0.51, "grad_norm": 0.3563327146891787, "learning_rate": 1.5142797509093652e-05, "loss": 0.8431, "step": 1568 }, { "epoch": 0.51, "grad_norm": 0.3361056557079907, "learning_rate": 1.5126931521592215e-05, "loss": 0.6092, "step": 1569 }, { "epoch": 0.51, "grad_norm": 0.33361095345213876, "learning_rate": 1.5111065392068569e-05, "loss": 0.6538, "step": 1570 }, { "epoch": 0.51, "grad_norm": 0.3405311018141979, "learning_rate": 1.509519913827513e-05, "loss": 0.5968, "step": 1571 }, { "epoch": 0.51, "grad_norm": 0.3188219019360358, "learning_rate": 1.5079332777964467e-05, "loss": 0.5903, "step": 1572 }, { "epoch": 0.51, "grad_norm": 0.37227002799828995, "learning_rate": 1.5063466328889278e-05, "loss": 0.8473, "step": 1573 }, { "epoch": 0.51, "grad_norm": 0.3835981957410638, "learning_rate": 1.5047599808802334e-05, "loss": 0.5855, "step": 1574 }, { "epoch": 0.51, "grad_norm": 0.327447825625473, "learning_rate": 1.5031733235456522e-05, "loss": 0.606, "step": 1575 }, { "epoch": 0.51, "grad_norm": 1.3294052145208513, "learning_rate": 1.5015866626604757e-05, "loss": 0.8533, "step": 1576 }, { "epoch": 0.52, "grad_norm": 0.41147347015678953, "learning_rate": 1.5e-05, "loss": 0.6064, "step": 1577 }, { "epoch": 0.52, "grad_norm": 0.34475250800522556, "learning_rate": 1.498413337339525e-05, "loss": 0.5882, "step": 1578 }, { "epoch": 0.52, "grad_norm": 0.36125438170155183, "learning_rate": 1.496826676454348e-05, "loss": 0.851, "step": 1579 }, { "epoch": 0.52, "grad_norm": 0.3411754448234949, "learning_rate": 1.4952400191197667e-05, "loss": 0.6143, "step": 1580 }, { "epoch": 0.52, "grad_norm": 0.6520240471889689, "learning_rate": 1.493653367111073e-05, "loss": 0.6182, "step": 1581 }, { "epoch": 0.52, "grad_norm": 0.33480491617418523, "learning_rate": 1.4920667222035532e-05, "loss": 0.608, "step": 1582 }, { "epoch": 0.52, "grad_norm": 0.4785632176743418, "learning_rate": 1.4904800861724874e-05, "loss": 0.8624, "step": 1583 }, { "epoch": 0.52, "grad_norm": 0.3161199001875394, "learning_rate": 1.4888934607931435e-05, "loss": 0.6097, "step": 1584 }, { "epoch": 0.52, "grad_norm": 0.3441700043378194, "learning_rate": 1.4873068478407782e-05, "loss": 0.6196, "step": 1585 }, { "epoch": 0.52, "grad_norm": 0.33306651039868257, "learning_rate": 1.485720249090635e-05, "loss": 0.58, "step": 1586 }, { "epoch": 0.52, "grad_norm": 0.39373575233908753, "learning_rate": 1.484133666317941e-05, "loss": 0.868, "step": 1587 }, { "epoch": 0.52, "grad_norm": 0.3520443244568107, "learning_rate": 1.4825471012979048e-05, "loss": 0.5911, "step": 1588 }, { "epoch": 0.52, "grad_norm": 0.32967713581827746, "learning_rate": 1.4809605558057157e-05, "loss": 0.6067, "step": 1589 }, { "epoch": 0.52, "grad_norm": 0.37730343083110274, "learning_rate": 1.4793740316165422e-05, "loss": 0.873, "step": 1590 }, { "epoch": 0.52, "grad_norm": 0.35196387942421264, "learning_rate": 1.4777875305055272e-05, "loss": 0.6223, "step": 1591 }, { "epoch": 0.52, "grad_norm": 0.3892952579083961, "learning_rate": 1.4762010542477881e-05, "loss": 0.6777, "step": 1592 }, { "epoch": 0.52, "grad_norm": 0.32271673429031894, "learning_rate": 1.474614604618416e-05, "loss": 0.6054, "step": 1593 }, { "epoch": 0.52, "grad_norm": 0.36670638190746335, "learning_rate": 1.4730281833924704e-05, "loss": 0.8488, "step": 1594 }, { "epoch": 0.52, "grad_norm": 0.307784860592243, "learning_rate": 1.4714417923449799e-05, "loss": 0.6032, "step": 1595 }, { "epoch": 0.52, "grad_norm": 0.31940274994816786, "learning_rate": 1.4698554332509395e-05, "loss": 0.6051, "step": 1596 }, { "epoch": 0.52, "grad_norm": 0.4925129190310338, "learning_rate": 1.4682691078853078e-05, "loss": 0.6445, "step": 1597 }, { "epoch": 0.52, "grad_norm": 0.3898432994873403, "learning_rate": 1.466682818023006e-05, "loss": 0.7489, "step": 1598 }, { "epoch": 0.52, "grad_norm": 0.3114288977200067, "learning_rate": 1.4650965654389158e-05, "loss": 0.6007, "step": 1599 }, { "epoch": 0.52, "grad_norm": 0.37576160338581105, "learning_rate": 1.4635103519078766e-05, "loss": 0.8527, "step": 1600 }, { "epoch": 0.52, "grad_norm": 0.29960165241054365, "learning_rate": 1.4619241792046842e-05, "loss": 0.5877, "step": 1601 }, { "epoch": 0.52, "grad_norm": 0.30402689421254425, "learning_rate": 1.46033804910409e-05, "loss": 0.5814, "step": 1602 }, { "epoch": 0.52, "grad_norm": 0.3441719948081735, "learning_rate": 1.4587519633807962e-05, "loss": 0.632, "step": 1603 }, { "epoch": 0.52, "grad_norm": 0.3113086500336864, "learning_rate": 1.4571659238094557e-05, "loss": 0.6007, "step": 1604 }, { "epoch": 0.52, "grad_norm": 0.36473479289724675, "learning_rate": 1.4555799321646699e-05, "loss": 0.8389, "step": 1605 }, { "epoch": 0.52, "grad_norm": 0.3160816786004416, "learning_rate": 1.453993990220987e-05, "loss": 0.5947, "step": 1606 }, { "epoch": 0.52, "grad_norm": 0.3098881028489543, "learning_rate": 1.4524080997528988e-05, "loss": 0.5956, "step": 1607 }, { "epoch": 0.53, "grad_norm": 0.3663787507694858, "learning_rate": 1.4508222625348397e-05, "loss": 0.6614, "step": 1608 }, { "epoch": 0.53, "grad_norm": 0.3484545464790408, "learning_rate": 1.4492364803411855e-05, "loss": 0.8502, "step": 1609 }, { "epoch": 0.53, "grad_norm": 0.3277254000535004, "learning_rate": 1.447650754946249e-05, "loss": 0.6042, "step": 1610 }, { "epoch": 0.53, "grad_norm": 0.289335987659079, "learning_rate": 1.4460650881242794e-05, "loss": 0.5918, "step": 1611 }, { "epoch": 0.53, "grad_norm": 0.30100613563510775, "learning_rate": 1.4444794816494629e-05, "loss": 0.5694, "step": 1612 }, { "epoch": 0.53, "grad_norm": 0.34715120943113575, "learning_rate": 1.4428939372959152e-05, "loss": 0.8661, "step": 1613 }, { "epoch": 0.53, "grad_norm": 0.32708877714809526, "learning_rate": 1.4413084568376835e-05, "loss": 0.5794, "step": 1614 }, { "epoch": 0.53, "grad_norm": 0.30220105754887655, "learning_rate": 1.439723042048744e-05, "loss": 0.5901, "step": 1615 }, { "epoch": 0.53, "grad_norm": 0.3529069351252895, "learning_rate": 1.4381376947029993e-05, "loss": 0.8713, "step": 1616 }, { "epoch": 0.53, "grad_norm": 0.30948025634218435, "learning_rate": 1.4365524165742754e-05, "loss": 0.5936, "step": 1617 }, { "epoch": 0.53, "grad_norm": 0.29197827666595144, "learning_rate": 1.4349672094363227e-05, "loss": 0.608, "step": 1618 }, { "epoch": 0.53, "grad_norm": 0.3502209540070997, "learning_rate": 1.4333820750628109e-05, "loss": 0.6298, "step": 1619 }, { "epoch": 0.53, "grad_norm": 0.6109367818897147, "learning_rate": 1.4317970152273283e-05, "loss": 0.8639, "step": 1620 }, { "epoch": 0.53, "grad_norm": 0.30211257802744595, "learning_rate": 1.43021203170338e-05, "loss": 0.5824, "step": 1621 }, { "epoch": 0.53, "grad_norm": 0.29622019920613657, "learning_rate": 1.4286271262643866e-05, "loss": 0.5888, "step": 1622 }, { "epoch": 0.53, "grad_norm": 0.37194980078197315, "learning_rate": 1.4270423006836802e-05, "loss": 0.9001, "step": 1623 }, { "epoch": 0.53, "grad_norm": 0.2889306651945915, "learning_rate": 1.4254575567345034e-05, "loss": 0.5872, "step": 1624 }, { "epoch": 0.53, "grad_norm": 0.3135516111053132, "learning_rate": 1.423872896190009e-05, "loss": 0.5886, "step": 1625 }, { "epoch": 0.53, "grad_norm": 0.3958452445488332, "learning_rate": 1.4222883208232544e-05, "loss": 0.8635, "step": 1626 }, { "epoch": 0.53, "grad_norm": 0.3036159356118533, "learning_rate": 1.4207038324072031e-05, "loss": 0.5867, "step": 1627 }, { "epoch": 0.53, "grad_norm": 0.3882495322827077, "learning_rate": 1.4191194327147214e-05, "loss": 0.6015, "step": 1628 }, { "epoch": 0.53, "grad_norm": 0.3762848923190769, "learning_rate": 1.4175351235185753e-05, "loss": 0.8821, "step": 1629 }, { "epoch": 0.53, "grad_norm": 0.3137649216386956, "learning_rate": 1.4159509065914297e-05, "loss": 0.4032, "step": 1630 }, { "epoch": 0.53, "grad_norm": 0.34305294909692025, "learning_rate": 1.414366783705848e-05, "loss": 0.8491, "step": 1631 }, { "epoch": 0.53, "grad_norm": 0.31525749662666536, "learning_rate": 1.4127827566342864e-05, "loss": 0.5816, "step": 1632 }, { "epoch": 0.53, "grad_norm": 0.2908181260257311, "learning_rate": 1.4111988271490941e-05, "loss": 0.5939, "step": 1633 }, { "epoch": 0.53, "grad_norm": 0.3621156579791383, "learning_rate": 1.4096149970225125e-05, "loss": 0.8548, "step": 1634 }, { "epoch": 0.53, "grad_norm": 0.3146142998070681, "learning_rate": 1.4080312680266703e-05, "loss": 0.6156, "step": 1635 }, { "epoch": 0.53, "grad_norm": 0.3029206998796662, "learning_rate": 1.4064476419335834e-05, "loss": 0.5806, "step": 1636 }, { "epoch": 0.53, "grad_norm": 0.3630032635409512, "learning_rate": 1.4048641205151534e-05, "loss": 0.8458, "step": 1637 }, { "epoch": 0.53, "grad_norm": 0.3298037658720421, "learning_rate": 1.403280705543164e-05, "loss": 0.5888, "step": 1638 }, { "epoch": 0.54, "grad_norm": 0.2954114492362546, "learning_rate": 1.40169739878928e-05, "loss": 0.5876, "step": 1639 }, { "epoch": 0.54, "grad_norm": 0.3741979271790974, "learning_rate": 1.4001142020250442e-05, "loss": 0.847, "step": 1640 }, { "epoch": 0.54, "grad_norm": 4.046828697447697, "learning_rate": 1.3985311170218787e-05, "loss": 0.4106, "step": 1641 }, { "epoch": 0.54, "grad_norm": 0.3785520659115123, "learning_rate": 1.3969481455510787e-05, "loss": 0.8807, "step": 1642 }, { "epoch": 0.54, "grad_norm": 0.3918056318895997, "learning_rate": 1.3953652893838121e-05, "loss": 0.6087, "step": 1643 }, { "epoch": 0.54, "grad_norm": 0.3016618776639696, "learning_rate": 1.393782550291119e-05, "loss": 0.5987, "step": 1644 }, { "epoch": 0.54, "grad_norm": 0.3904102374238184, "learning_rate": 1.392199930043908e-05, "loss": 0.8187, "step": 1645 }, { "epoch": 0.54, "grad_norm": 0.30789552312172563, "learning_rate": 1.3906174304129541e-05, "loss": 0.5959, "step": 1646 }, { "epoch": 0.54, "grad_norm": 0.3134583827457704, "learning_rate": 1.3890350531688986e-05, "loss": 0.5926, "step": 1647 }, { "epoch": 0.54, "grad_norm": 0.36697521157176693, "learning_rate": 1.387452800082245e-05, "loss": 0.862, "step": 1648 }, { "epoch": 0.54, "grad_norm": 0.33027547983530886, "learning_rate": 1.3858706729233571e-05, "loss": 0.597, "step": 1649 }, { "epoch": 0.54, "grad_norm": 0.29216606894747277, "learning_rate": 1.3842886734624605e-05, "loss": 0.5906, "step": 1650 }, { "epoch": 0.54, "grad_norm": 0.5048217249121248, "learning_rate": 1.3827068034696353e-05, "loss": 0.8579, "step": 1651 }, { "epoch": 0.54, "grad_norm": 0.36436693028134526, "learning_rate": 1.3811250647148172e-05, "loss": 0.6308, "step": 1652 }, { "epoch": 0.54, "grad_norm": 0.30823926236501825, "learning_rate": 1.3795434589677964e-05, "loss": 0.5742, "step": 1653 }, { "epoch": 0.54, "grad_norm": 1.0195893913271346, "learning_rate": 1.3779619879982127e-05, "loss": 0.603, "step": 1654 }, { "epoch": 0.54, "grad_norm": 0.5303019888755408, "learning_rate": 1.3763806535755563e-05, "loss": 0.8412, "step": 1655 }, { "epoch": 0.54, "grad_norm": 0.3196786171559462, "learning_rate": 1.374799457469163e-05, "loss": 0.5949, "step": 1656 }, { "epoch": 0.54, "grad_norm": 0.27615675059499256, "learning_rate": 1.3732184014482163e-05, "loss": 0.5745, "step": 1657 }, { "epoch": 0.54, "grad_norm": 0.3432281772713932, "learning_rate": 1.3716374872817408e-05, "loss": 0.5919, "step": 1658 }, { "epoch": 0.54, "grad_norm": 0.3841993516778034, "learning_rate": 1.3700567167386027e-05, "loss": 0.8651, "step": 1659 }, { "epoch": 0.54, "grad_norm": 0.3341824638718494, "learning_rate": 1.3684760915875093e-05, "loss": 0.6063, "step": 1660 }, { "epoch": 0.54, "grad_norm": 0.29185633439677344, "learning_rate": 1.366895613597003e-05, "loss": 0.5863, "step": 1661 }, { "epoch": 0.54, "grad_norm": 0.3904905766704826, "learning_rate": 1.3653152845354625e-05, "loss": 0.8155, "step": 1662 }, { "epoch": 0.54, "grad_norm": 0.35090994173141116, "learning_rate": 1.3637351061711002e-05, "loss": 0.625, "step": 1663 }, { "epoch": 0.54, "grad_norm": 0.3181665894327707, "learning_rate": 1.362155080271959e-05, "loss": 0.5996, "step": 1664 }, { "epoch": 0.54, "grad_norm": 0.29651214560849887, "learning_rate": 1.3605752086059118e-05, "loss": 0.6187, "step": 1665 }, { "epoch": 0.54, "grad_norm": 0.3756441140529517, "learning_rate": 1.358995492940659e-05, "loss": 0.873, "step": 1666 }, { "epoch": 0.54, "grad_norm": 0.3102948854734104, "learning_rate": 1.3574159350437264e-05, "loss": 0.6127, "step": 1667 }, { "epoch": 0.54, "grad_norm": 0.31292019476624444, "learning_rate": 1.3558365366824621e-05, "loss": 0.5909, "step": 1668 }, { "epoch": 0.55, "grad_norm": 0.3720822046610206, "learning_rate": 1.3542572996240382e-05, "loss": 0.8485, "step": 1669 }, { "epoch": 0.55, "grad_norm": 0.3023555282054726, "learning_rate": 1.3526782256354443e-05, "loss": 0.5634, "step": 1670 }, { "epoch": 0.55, "grad_norm": 0.3064075886413698, "learning_rate": 1.3510993164834886e-05, "loss": 0.6047, "step": 1671 }, { "epoch": 0.55, "grad_norm": 0.3558166296362969, "learning_rate": 1.3495205739347925e-05, "loss": 0.8379, "step": 1672 }, { "epoch": 0.55, "grad_norm": 0.3170849317547557, "learning_rate": 1.347941999755795e-05, "loss": 0.5789, "step": 1673 }, { "epoch": 0.55, "grad_norm": 0.34299757327344427, "learning_rate": 1.3463635957127439e-05, "loss": 0.633, "step": 1674 }, { "epoch": 0.55, "grad_norm": 0.3173043133126753, "learning_rate": 1.3447853635716967e-05, "loss": 0.6042, "step": 1675 }, { "epoch": 0.55, "grad_norm": 0.29415652941741555, "learning_rate": 1.3432073050985201e-05, "loss": 0.5886, "step": 1676 }, { "epoch": 0.55, "grad_norm": 0.37544692349403214, "learning_rate": 1.3416294220588851e-05, "loss": 0.8284, "step": 1677 }, { "epoch": 0.55, "grad_norm": 0.3245015959297751, "learning_rate": 1.3400517162182662e-05, "loss": 0.6025, "step": 1678 }, { "epoch": 0.55, "grad_norm": 0.30996085072234253, "learning_rate": 1.3384741893419417e-05, "loss": 0.6066, "step": 1679 }, { "epoch": 0.55, "grad_norm": 0.34539434304708583, "learning_rate": 1.3368968431949879e-05, "loss": 0.6256, "step": 1680 }, { "epoch": 0.55, "grad_norm": 0.3352062014770533, "learning_rate": 1.3353196795422778e-05, "loss": 0.7703, "step": 1681 }, { "epoch": 0.55, "grad_norm": 0.29786246292713503, "learning_rate": 1.3337427001484836e-05, "loss": 0.5852, "step": 1682 }, { "epoch": 0.55, "grad_norm": 0.3609861585319644, "learning_rate": 1.3321659067780684e-05, "loss": 0.8579, "step": 1683 }, { "epoch": 0.55, "grad_norm": 0.3144676765308442, "learning_rate": 1.3305893011952882e-05, "loss": 0.5903, "step": 1684 }, { "epoch": 0.55, "grad_norm": 0.3443867553832964, "learning_rate": 1.3290128851641892e-05, "loss": 0.6396, "step": 1685 }, { "epoch": 0.55, "grad_norm": 0.3163839743185421, "learning_rate": 1.3274366604486049e-05, "loss": 0.5951, "step": 1686 }, { "epoch": 0.55, "grad_norm": 0.3096493549547529, "learning_rate": 1.3258606288121545e-05, "loss": 0.5937, "step": 1687 }, { "epoch": 0.55, "grad_norm": 0.3635542451058479, "learning_rate": 1.3242847920182426e-05, "loss": 0.8615, "step": 1688 }, { "epoch": 0.55, "grad_norm": 0.3089114335127662, "learning_rate": 1.3227091518300546e-05, "loss": 0.5859, "step": 1689 }, { "epoch": 0.55, "grad_norm": 0.3012731663443421, "learning_rate": 1.3211337100105562e-05, "loss": 0.5894, "step": 1690 }, { "epoch": 0.55, "grad_norm": 0.3459244389188917, "learning_rate": 1.31955846832249e-05, "loss": 0.6449, "step": 1691 }, { "epoch": 0.55, "grad_norm": 0.3553440070001605, "learning_rate": 1.3179834285283773e-05, "loss": 0.8381, "step": 1692 }, { "epoch": 0.55, "grad_norm": 0.3069673072893443, "learning_rate": 1.3164085923905113e-05, "loss": 0.5884, "step": 1693 }, { "epoch": 0.55, "grad_norm": 0.3224194671449196, "learning_rate": 1.3148339616709579e-05, "loss": 0.5813, "step": 1694 }, { "epoch": 0.55, "grad_norm": 0.3777293343106343, "learning_rate": 1.3132595381315539e-05, "loss": 0.8633, "step": 1695 }, { "epoch": 0.55, "grad_norm": 0.32817163077487027, "learning_rate": 1.3116853235339034e-05, "loss": 0.6059, "step": 1696 }, { "epoch": 0.55, "grad_norm": 0.3145970594875752, "learning_rate": 1.310111319639376e-05, "loss": 0.6003, "step": 1697 }, { "epoch": 0.55, "grad_norm": 0.36760105617672645, "learning_rate": 1.308537528209108e-05, "loss": 0.8312, "step": 1698 }, { "epoch": 0.55, "grad_norm": 0.30913015832347146, "learning_rate": 1.3069639510039965e-05, "loss": 0.5973, "step": 1699 }, { "epoch": 0.56, "grad_norm": 0.3169401753018231, "learning_rate": 1.3053905897846973e-05, "loss": 0.6224, "step": 1700 }, { "epoch": 0.56, "grad_norm": 0.31050429729659246, "learning_rate": 1.303817446311628e-05, "loss": 0.61, "step": 1701 }, { "epoch": 0.56, "grad_norm": 0.340912977768504, "learning_rate": 1.3022445223449596e-05, "loss": 0.6372, "step": 1702 }, { "epoch": 0.56, "grad_norm": 0.368403565758628, "learning_rate": 1.300671819644619e-05, "loss": 0.8628, "step": 1703 }, { "epoch": 0.56, "grad_norm": 0.3136798965833182, "learning_rate": 1.2990993399702853e-05, "loss": 0.6169, "step": 1704 }, { "epoch": 0.56, "grad_norm": 0.30030572676066125, "learning_rate": 1.2975270850813878e-05, "loss": 0.5736, "step": 1705 }, { "epoch": 0.56, "grad_norm": 0.3562893731150036, "learning_rate": 1.295955056737104e-05, "loss": 0.8683, "step": 1706 }, { "epoch": 0.56, "grad_norm": 0.31669062379743634, "learning_rate": 1.2943832566963582e-05, "loss": 0.572, "step": 1707 }, { "epoch": 0.56, "grad_norm": 0.30352590344479363, "learning_rate": 1.2928116867178208e-05, "loss": 0.5901, "step": 1708 }, { "epoch": 0.56, "grad_norm": 0.351436841090328, "learning_rate": 1.2912403485599022e-05, "loss": 0.8565, "step": 1709 }, { "epoch": 0.56, "grad_norm": 0.32623283602598274, "learning_rate": 1.289669243980754e-05, "loss": 0.6047, "step": 1710 }, { "epoch": 0.56, "grad_norm": 0.29836994319957827, "learning_rate": 1.2880983747382685e-05, "loss": 0.5866, "step": 1711 }, { "epoch": 0.56, "grad_norm": 0.5403304346212752, "learning_rate": 1.2865277425900725e-05, "loss": 0.8819, "step": 1712 }, { "epoch": 0.56, "grad_norm": 0.2999652792641207, "learning_rate": 1.2849573492935278e-05, "loss": 0.4022, "step": 1713 }, { "epoch": 0.56, "grad_norm": 0.346298242502661, "learning_rate": 1.2833871966057302e-05, "loss": 0.8256, "step": 1714 }, { "epoch": 0.56, "grad_norm": 0.29277874205232296, "learning_rate": 1.2818172862835052e-05, "loss": 0.5942, "step": 1715 }, { "epoch": 0.56, "grad_norm": 0.29875644529070433, "learning_rate": 1.2802476200834064e-05, "loss": 0.5834, "step": 1716 }, { "epoch": 0.56, "grad_norm": 0.33393636602288, "learning_rate": 1.278678199761717e-05, "loss": 0.8443, "step": 1717 }, { "epoch": 0.56, "grad_norm": 0.30838503353839425, "learning_rate": 1.2771090270744427e-05, "loss": 0.6063, "step": 1718 }, { "epoch": 0.56, "grad_norm": 0.29874893341214487, "learning_rate": 1.2755401037773115e-05, "loss": 0.5792, "step": 1719 }, { "epoch": 0.56, "grad_norm": 0.34470177794811685, "learning_rate": 1.2739714316257753e-05, "loss": 0.8796, "step": 1720 }, { "epoch": 0.56, "grad_norm": 0.3040743782130052, "learning_rate": 1.2724030123750027e-05, "loss": 0.5779, "step": 1721 }, { "epoch": 0.56, "grad_norm": 0.2891355951788101, "learning_rate": 1.2708348477798795e-05, "loss": 0.5862, "step": 1722 }, { "epoch": 0.56, "grad_norm": 0.36058411254889255, "learning_rate": 1.2692669395950078e-05, "loss": 0.8578, "step": 1723 }, { "epoch": 0.56, "grad_norm": 0.34216280680786787, "learning_rate": 1.2676992895747017e-05, "loss": 0.6392, "step": 1724 }, { "epoch": 0.56, "grad_norm": 0.30420985023415525, "learning_rate": 1.2661318994729867e-05, "loss": 0.5786, "step": 1725 }, { "epoch": 0.56, "grad_norm": 0.30143145065343036, "learning_rate": 1.2645647710435965e-05, "loss": 0.5887, "step": 1726 }, { "epoch": 0.56, "grad_norm": 0.3748576467342296, "learning_rate": 1.2629979060399751e-05, "loss": 0.8615, "step": 1727 }, { "epoch": 0.56, "grad_norm": 0.4382563153550041, "learning_rate": 1.2614313062152692e-05, "loss": 0.5844, "step": 1728 }, { "epoch": 0.56, "grad_norm": 0.2973947980130506, "learning_rate": 1.2598649733223275e-05, "loss": 0.574, "step": 1729 }, { "epoch": 0.56, "grad_norm": 0.3224046673568471, "learning_rate": 1.258298909113704e-05, "loss": 0.5938, "step": 1730 }, { "epoch": 0.57, "grad_norm": 0.36684945482524617, "learning_rate": 1.256733115341649e-05, "loss": 0.8546, "step": 1731 }, { "epoch": 0.57, "grad_norm": 0.3036960512557954, "learning_rate": 1.255167593758111e-05, "loss": 0.6002, "step": 1732 }, { "epoch": 0.57, "grad_norm": 0.31450747659565664, "learning_rate": 1.2536023461147347e-05, "loss": 0.5824, "step": 1733 }, { "epoch": 0.57, "grad_norm": 0.4003585658165809, "learning_rate": 1.2520373741628573e-05, "loss": 0.8378, "step": 1734 }, { "epoch": 0.57, "grad_norm": 0.3556712818996085, "learning_rate": 1.2504726796535069e-05, "loss": 0.6252, "step": 1735 }, { "epoch": 0.57, "grad_norm": 0.32646889137246154, "learning_rate": 1.248908264337404e-05, "loss": 0.5828, "step": 1736 }, { "epoch": 0.57, "grad_norm": 0.3008658658459421, "learning_rate": 1.2473441299649545e-05, "loss": 0.5713, "step": 1737 }, { "epoch": 0.57, "grad_norm": 0.43297311816099604, "learning_rate": 1.2457802782862488e-05, "loss": 0.8462, "step": 1738 }, { "epoch": 0.57, "grad_norm": 0.3144639717691245, "learning_rate": 1.2442167110510643e-05, "loss": 0.5822, "step": 1739 }, { "epoch": 0.57, "grad_norm": 0.31154539580303875, "learning_rate": 1.2426534300088577e-05, "loss": 0.5772, "step": 1740 }, { "epoch": 0.57, "grad_norm": 0.3727207723397832, "learning_rate": 1.2410904369087664e-05, "loss": 0.8304, "step": 1741 }, { "epoch": 0.57, "grad_norm": 0.3175758263771139, "learning_rate": 1.2395277334996045e-05, "loss": 0.6013, "step": 1742 }, { "epoch": 0.57, "grad_norm": 0.30391852721689755, "learning_rate": 1.237965321529864e-05, "loss": 0.6002, "step": 1743 }, { "epoch": 0.57, "grad_norm": 0.34444617144136547, "learning_rate": 1.2364032027477093e-05, "loss": 0.8621, "step": 1744 }, { "epoch": 0.57, "grad_norm": 0.3763189455685237, "learning_rate": 1.2348413789009763e-05, "loss": 0.5891, "step": 1745 }, { "epoch": 0.57, "grad_norm": 0.44344266648710334, "learning_rate": 1.2332798517371732e-05, "loss": 0.6213, "step": 1746 }, { "epoch": 0.57, "grad_norm": 0.30698982132558744, "learning_rate": 1.2317186230034743e-05, "loss": 0.5935, "step": 1747 }, { "epoch": 0.57, "grad_norm": 0.3168602516975504, "learning_rate": 1.2301576944467192e-05, "loss": 0.5921, "step": 1748 }, { "epoch": 0.57, "grad_norm": 0.3552567746061379, "learning_rate": 1.2285970678134144e-05, "loss": 0.8654, "step": 1749 }, { "epoch": 0.57, "grad_norm": 0.3132590584417623, "learning_rate": 1.2270367448497268e-05, "loss": 0.5941, "step": 1750 }, { "epoch": 0.57, "grad_norm": 0.31476061037357145, "learning_rate": 1.225476727301483e-05, "loss": 0.589, "step": 1751 }, { "epoch": 0.57, "grad_norm": 0.3611498317790471, "learning_rate": 1.2239170169141696e-05, "loss": 0.8601, "step": 1752 }, { "epoch": 0.57, "grad_norm": 0.320105568482908, "learning_rate": 1.2223576154329285e-05, "loss": 0.5828, "step": 1753 }, { "epoch": 0.57, "grad_norm": 0.34433062466614967, "learning_rate": 1.2207985246025549e-05, "loss": 0.595, "step": 1754 }, { "epoch": 0.57, "grad_norm": 0.345160104626171, "learning_rate": 1.2192397461674992e-05, "loss": 0.8516, "step": 1755 }, { "epoch": 0.57, "grad_norm": 0.31200065938702154, "learning_rate": 1.2176812818718604e-05, "loss": 0.3974, "step": 1756 }, { "epoch": 0.57, "grad_norm": 0.3494893486332444, "learning_rate": 1.2161231334593853e-05, "loss": 0.861, "step": 1757 }, { "epoch": 0.57, "grad_norm": 0.3242345083407676, "learning_rate": 1.214565302673468e-05, "loss": 0.5837, "step": 1758 }, { "epoch": 0.57, "grad_norm": 0.32272294458536027, "learning_rate": 1.213007791257149e-05, "loss": 0.5862, "step": 1759 }, { "epoch": 0.57, "grad_norm": 0.3672161058930586, "learning_rate": 1.2114506009531091e-05, "loss": 0.8562, "step": 1760 }, { "epoch": 0.58, "grad_norm": 0.3200243849110376, "learning_rate": 1.2098937335036701e-05, "loss": 0.5926, "step": 1761 }, { "epoch": 0.58, "grad_norm": 0.35856934646963284, "learning_rate": 1.2083371906507939e-05, "loss": 0.6351, "step": 1762 }, { "epoch": 0.58, "grad_norm": 0.3329893921350503, "learning_rate": 1.206780974136078e-05, "loss": 0.5956, "step": 1763 }, { "epoch": 0.58, "grad_norm": 0.3523676537760977, "learning_rate": 1.2052250857007548e-05, "loss": 0.8105, "step": 1764 }, { "epoch": 0.58, "grad_norm": 0.3104384447086769, "learning_rate": 1.2036695270856909e-05, "loss": 0.5563, "step": 1765 }, { "epoch": 0.58, "grad_norm": 0.34837313318026014, "learning_rate": 1.2021143000313822e-05, "loss": 0.8691, "step": 1766 }, { "epoch": 0.58, "grad_norm": 0.3724470610944217, "learning_rate": 1.2005594062779539e-05, "loss": 0.6188, "step": 1767 }, { "epoch": 0.58, "grad_norm": 0.29451063402520805, "learning_rate": 1.19900484756516e-05, "loss": 0.5981, "step": 1768 }, { "epoch": 0.58, "grad_norm": 0.3138341600958812, "learning_rate": 1.1974506256323776e-05, "loss": 0.607, "step": 1769 }, { "epoch": 0.58, "grad_norm": 0.37484475876658646, "learning_rate": 1.1958967422186076e-05, "loss": 0.8537, "step": 1770 }, { "epoch": 0.58, "grad_norm": 0.29448630375684315, "learning_rate": 1.1943431990624725e-05, "loss": 0.5833, "step": 1771 }, { "epoch": 0.58, "grad_norm": 0.3126973740694464, "learning_rate": 1.1927899979022143e-05, "loss": 0.5805, "step": 1772 }, { "epoch": 0.58, "grad_norm": 0.3624575938834935, "learning_rate": 1.1912371404756905e-05, "loss": 0.6486, "step": 1773 }, { "epoch": 0.58, "grad_norm": 0.3027830882565532, "learning_rate": 1.1896846285203771e-05, "loss": 0.5873, "step": 1774 }, { "epoch": 0.58, "grad_norm": 0.35631693142317933, "learning_rate": 1.1881324637733613e-05, "loss": 0.8534, "step": 1775 }, { "epoch": 0.58, "grad_norm": 1.955525207200316, "learning_rate": 1.186580647971342e-05, "loss": 0.5993, "step": 1776 }, { "epoch": 0.58, "grad_norm": 0.30829759602550477, "learning_rate": 1.1850291828506274e-05, "loss": 0.5981, "step": 1777 }, { "epoch": 0.58, "grad_norm": 0.3668422334730386, "learning_rate": 1.1834780701471356e-05, "loss": 0.8554, "step": 1778 }, { "epoch": 0.58, "grad_norm": 0.34079890159858817, "learning_rate": 1.1819273115963877e-05, "loss": 0.6111, "step": 1779 }, { "epoch": 0.58, "grad_norm": 0.3132860458535729, "learning_rate": 1.1803769089335097e-05, "loss": 0.5858, "step": 1780 }, { "epoch": 0.58, "grad_norm": 0.3661294435335552, "learning_rate": 1.1788268638932296e-05, "loss": 0.856, "step": 1781 }, { "epoch": 0.58, "grad_norm": 0.32002564906508274, "learning_rate": 1.1772771782098748e-05, "loss": 0.6032, "step": 1782 }, { "epoch": 0.58, "grad_norm": 0.32874983825036175, "learning_rate": 1.1757278536173702e-05, "loss": 0.5803, "step": 1783 }, { "epoch": 0.58, "grad_norm": 0.5928965900847469, "learning_rate": 1.1741788918492386e-05, "loss": 0.596, "step": 1784 }, { "epoch": 0.58, "grad_norm": 0.32695294724697477, "learning_rate": 1.1726302946385945e-05, "loss": 0.6118, "step": 1785 }, { "epoch": 0.58, "grad_norm": 0.4837302819554191, "learning_rate": 1.1710820637181449e-05, "loss": 0.8757, "step": 1786 }, { "epoch": 0.58, "grad_norm": 0.31612571007548834, "learning_rate": 1.169534200820189e-05, "loss": 0.5727, "step": 1787 }, { "epoch": 0.58, "grad_norm": 0.3090801568869898, "learning_rate": 1.1679867076766123e-05, "loss": 0.5814, "step": 1788 }, { "epoch": 0.58, "grad_norm": 0.35001161003659803, "learning_rate": 1.1664395860188864e-05, "loss": 0.8302, "step": 1789 }, { "epoch": 0.58, "grad_norm": 0.31153891825027313, "learning_rate": 1.1648928375780688e-05, "loss": 0.5785, "step": 1790 }, { "epoch": 0.58, "grad_norm": 0.2979860754014292, "learning_rate": 1.163346464084798e-05, "loss": 0.6069, "step": 1791 }, { "epoch": 0.59, "grad_norm": 0.36929072790666956, "learning_rate": 1.1618004672692937e-05, "loss": 0.85, "step": 1792 }, { "epoch": 0.59, "grad_norm": 0.2975344484415256, "learning_rate": 1.1602548488613533e-05, "loss": 0.579, "step": 1793 }, { "epoch": 0.59, "grad_norm": 0.29721478945283736, "learning_rate": 1.1587096105903529e-05, "loss": 0.5945, "step": 1794 }, { "epoch": 0.59, "grad_norm": 0.34421043397209744, "learning_rate": 1.1571647541852407e-05, "loss": 0.6537, "step": 1795 }, { "epoch": 0.59, "grad_norm": 0.3545041658069107, "learning_rate": 1.1556202813745383e-05, "loss": 0.8784, "step": 1796 }, { "epoch": 0.59, "grad_norm": 0.3224003697835103, "learning_rate": 1.1540761938863397e-05, "loss": 0.5877, "step": 1797 }, { "epoch": 0.59, "grad_norm": 0.2964101199556964, "learning_rate": 1.1525324934483063e-05, "loss": 0.5863, "step": 1798 }, { "epoch": 0.59, "grad_norm": 0.30496213512739734, "learning_rate": 1.1509891817876664e-05, "loss": 0.5906, "step": 1799 }, { "epoch": 0.59, "grad_norm": 0.36570564659851595, "learning_rate": 1.1494462606312141e-05, "loss": 0.8759, "step": 1800 }, { "epoch": 0.59, "grad_norm": 0.299733085592644, "learning_rate": 1.147903731705306e-05, "loss": 0.5925, "step": 1801 }, { "epoch": 0.59, "grad_norm": 0.30098415303612286, "learning_rate": 1.146361596735859e-05, "loss": 0.5847, "step": 1802 }, { "epoch": 0.59, "grad_norm": 0.37564593938129154, "learning_rate": 1.144819857448352e-05, "loss": 0.8776, "step": 1803 }, { "epoch": 0.59, "grad_norm": 0.36480125223762494, "learning_rate": 1.143278515567818e-05, "loss": 0.6027, "step": 1804 }, { "epoch": 0.59, "grad_norm": 0.29620050044384966, "learning_rate": 1.1417375728188465e-05, "loss": 0.5678, "step": 1805 }, { "epoch": 0.59, "grad_norm": 0.36015328465507296, "learning_rate": 1.1401970309255817e-05, "loss": 0.616, "step": 1806 }, { "epoch": 0.59, "grad_norm": 0.36404016156687197, "learning_rate": 1.1386568916117178e-05, "loss": 0.8346, "step": 1807 }, { "epoch": 0.59, "grad_norm": 0.29677577354585394, "learning_rate": 1.1371171566004986e-05, "loss": 0.5787, "step": 1808 }, { "epoch": 0.59, "grad_norm": 0.32086794659888535, "learning_rate": 1.135577827614716e-05, "loss": 0.6057, "step": 1809 }, { "epoch": 0.59, "grad_norm": 0.3610142585281715, "learning_rate": 1.1340389063767077e-05, "loss": 0.879, "step": 1810 }, { "epoch": 0.59, "grad_norm": 0.30185133984622964, "learning_rate": 1.132500394608355e-05, "loss": 0.5961, "step": 1811 }, { "epoch": 0.59, "grad_norm": 0.30805464642757185, "learning_rate": 1.1309622940310798e-05, "loss": 0.5816, "step": 1812 }, { "epoch": 0.59, "grad_norm": 0.3405938758732022, "learning_rate": 1.1294246063658475e-05, "loss": 0.8457, "step": 1813 }, { "epoch": 0.59, "grad_norm": 0.29689560386302616, "learning_rate": 1.1278873333331572e-05, "loss": 0.6117, "step": 1814 }, { "epoch": 0.59, "grad_norm": 0.4727178383963052, "learning_rate": 1.126350476653046e-05, "loss": 0.6209, "step": 1815 }, { "epoch": 0.59, "grad_norm": 0.2855859680401549, "learning_rate": 1.1248140380450865e-05, "loss": 0.5967, "step": 1816 }, { "epoch": 0.59, "grad_norm": 0.3394945696594637, "learning_rate": 1.1232780192283813e-05, "loss": 0.6175, "step": 1817 }, { "epoch": 0.59, "grad_norm": 0.3495658141519358, "learning_rate": 1.1217424219215641e-05, "loss": 0.835, "step": 1818 }, { "epoch": 0.59, "grad_norm": 0.3004236459956035, "learning_rate": 1.1202072478427975e-05, "loss": 0.5844, "step": 1819 }, { "epoch": 0.59, "grad_norm": 0.2928291480204314, "learning_rate": 1.11867249870977e-05, "loss": 0.5793, "step": 1820 }, { "epoch": 0.59, "grad_norm": 0.8211722234074684, "learning_rate": 1.1171381762396937e-05, "loss": 0.8825, "step": 1821 }, { "epoch": 0.6, "grad_norm": 0.32389950545931157, "learning_rate": 1.1156042821493062e-05, "loss": 0.5944, "step": 1822 }, { "epoch": 0.6, "grad_norm": 0.28394242550479526, "learning_rate": 1.1140708181548629e-05, "loss": 0.5872, "step": 1823 }, { "epoch": 0.6, "grad_norm": 0.36512689787558755, "learning_rate": 1.1125377859721383e-05, "loss": 0.8517, "step": 1824 }, { "epoch": 0.6, "grad_norm": 0.30643524038596576, "learning_rate": 1.1110051873164259e-05, "loss": 0.5825, "step": 1825 }, { "epoch": 0.6, "grad_norm": 0.28444610323039715, "learning_rate": 1.109473023902532e-05, "loss": 0.6138, "step": 1826 }, { "epoch": 0.6, "grad_norm": 0.3589175740377994, "learning_rate": 1.1079412974447769e-05, "loss": 0.8704, "step": 1827 }, { "epoch": 0.6, "grad_norm": 0.31805055468419136, "learning_rate": 1.106410009656991e-05, "loss": 0.3966, "step": 1828 }, { "epoch": 0.6, "grad_norm": 0.3522581067102376, "learning_rate": 1.104879162252515e-05, "loss": 0.8452, "step": 1829 }, { "epoch": 0.6, "grad_norm": 0.31611790917015786, "learning_rate": 1.1033487569441971e-05, "loss": 0.5782, "step": 1830 }, { "epoch": 0.6, "grad_norm": 0.2887579413089533, "learning_rate": 1.1018187954443884e-05, "loss": 0.5951, "step": 1831 }, { "epoch": 0.6, "grad_norm": 0.3617493655424364, "learning_rate": 1.1002892794649478e-05, "loss": 0.8779, "step": 1832 }, { "epoch": 0.6, "grad_norm": 0.321662348823602, "learning_rate": 1.098760210717231e-05, "loss": 0.5809, "step": 1833 }, { "epoch": 0.6, "grad_norm": 0.29109277533746936, "learning_rate": 1.0972315909120958e-05, "loss": 0.5915, "step": 1834 }, { "epoch": 0.6, "grad_norm": 0.3538029910427498, "learning_rate": 1.0957034217598983e-05, "loss": 0.8774, "step": 1835 }, { "epoch": 0.6, "grad_norm": 0.4399935559552116, "learning_rate": 1.0941757049704886e-05, "loss": 0.5999, "step": 1836 }, { "epoch": 0.6, "grad_norm": 0.2823330282041912, "learning_rate": 1.092648442253211e-05, "loss": 0.6008, "step": 1837 }, { "epoch": 0.6, "grad_norm": 0.35636843354346265, "learning_rate": 1.0911216353169034e-05, "loss": 0.8734, "step": 1838 }, { "epoch": 0.6, "grad_norm": 0.34069978302261567, "learning_rate": 1.0895952858698916e-05, "loss": 0.614, "step": 1839 }, { "epoch": 0.6, "grad_norm": 0.28531192065682004, "learning_rate": 1.0880693956199902e-05, "loss": 0.576, "step": 1840 }, { "epoch": 0.6, "grad_norm": 0.29410987488815404, "learning_rate": 1.0865439662745013e-05, "loss": 0.5726, "step": 1841 }, { "epoch": 0.6, "grad_norm": 0.364104724831153, "learning_rate": 1.0850189995402096e-05, "loss": 0.8353, "step": 1842 }, { "epoch": 0.6, "grad_norm": 0.30282936177640146, "learning_rate": 1.083494497123383e-05, "loss": 0.5656, "step": 1843 }, { "epoch": 0.6, "grad_norm": 0.3257991996351362, "learning_rate": 1.0819704607297687e-05, "loss": 0.6293, "step": 1844 }, { "epoch": 0.6, "grad_norm": 0.29386960558675224, "learning_rate": 1.0804468920645949e-05, "loss": 0.5711, "step": 1845 }, { "epoch": 0.6, "grad_norm": 0.3081565957651193, "learning_rate": 1.0789237928325646e-05, "loss": 0.5977, "step": 1846 }, { "epoch": 0.6, "grad_norm": 0.3582771353492958, "learning_rate": 1.0774011647378554e-05, "loss": 0.8589, "step": 1847 }, { "epoch": 0.6, "grad_norm": 0.27388578104684963, "learning_rate": 1.0758790094841194e-05, "loss": 0.5501, "step": 1848 }, { "epoch": 0.6, "grad_norm": 0.3448820206497615, "learning_rate": 1.074357328774478e-05, "loss": 0.8458, "step": 1849 }, { "epoch": 0.6, "grad_norm": 0.3525428692936725, "learning_rate": 1.0728361243115216e-05, "loss": 0.6341, "step": 1850 }, { "epoch": 0.6, "grad_norm": 0.29596003453641306, "learning_rate": 1.0713153977973098e-05, "loss": 0.597, "step": 1851 }, { "epoch": 0.6, "grad_norm": 0.30643139502096506, "learning_rate": 1.069795150933365e-05, "loss": 0.5905, "step": 1852 }, { "epoch": 0.61, "grad_norm": 0.33944280521196735, "learning_rate": 1.0682753854206735e-05, "loss": 0.8442, "step": 1853 }, { "epoch": 0.61, "grad_norm": 0.31109239948547945, "learning_rate": 1.0667561029596847e-05, "loss": 0.5833, "step": 1854 }, { "epoch": 0.61, "grad_norm": 0.334691968669734, "learning_rate": 1.0652373052503059e-05, "loss": 0.641, "step": 1855 }, { "epoch": 0.61, "grad_norm": 0.2896905753069529, "learning_rate": 1.0637189939919014e-05, "loss": 0.5993, "step": 1856 }, { "epoch": 0.61, "grad_norm": 0.31159094114091307, "learning_rate": 1.0622011708832933e-05, "loss": 0.5768, "step": 1857 }, { "epoch": 0.61, "grad_norm": 0.351644414255032, "learning_rate": 1.060683837622756e-05, "loss": 0.8605, "step": 1858 }, { "epoch": 0.61, "grad_norm": 0.29815178128080444, "learning_rate": 1.0591669959080164e-05, "loss": 0.5761, "step": 1859 }, { "epoch": 0.61, "grad_norm": 0.291201158836381, "learning_rate": 1.0576506474362507e-05, "loss": 0.597, "step": 1860 }, { "epoch": 0.61, "grad_norm": 0.3880087738528173, "learning_rate": 1.0561347939040843e-05, "loss": 0.9003, "step": 1861 }, { "epoch": 0.61, "grad_norm": 0.29057737375329346, "learning_rate": 1.0546194370075882e-05, "loss": 0.599, "step": 1862 }, { "epoch": 0.61, "grad_norm": 0.2865613054646457, "learning_rate": 1.053104578442277e-05, "loss": 0.5994, "step": 1863 }, { "epoch": 0.61, "grad_norm": 0.4604915978453122, "learning_rate": 1.0515902199031098e-05, "loss": 0.8658, "step": 1864 }, { "epoch": 0.61, "grad_norm": 0.2879899049800068, "learning_rate": 1.0500763630844842e-05, "loss": 0.5966, "step": 1865 }, { "epoch": 0.61, "grad_norm": 0.33939430307694807, "learning_rate": 1.0485630096802371e-05, "loss": 0.6177, "step": 1866 }, { "epoch": 0.61, "grad_norm": 0.3034828396539909, "learning_rate": 1.0470501613836427e-05, "loss": 0.6129, "step": 1867 }, { "epoch": 0.61, "grad_norm": 0.3410214613948706, "learning_rate": 1.0455378198874093e-05, "loss": 0.8742, "step": 1868 }, { "epoch": 0.61, "grad_norm": 0.3019362882685936, "learning_rate": 1.044025986883678e-05, "loss": 0.5949, "step": 1869 }, { "epoch": 0.61, "grad_norm": 0.3933030253981285, "learning_rate": 1.042514664064022e-05, "loss": 0.5889, "step": 1870 }, { "epoch": 0.61, "grad_norm": 0.2957013045402267, "learning_rate": 1.0410038531194428e-05, "loss": 0.5954, "step": 1871 }, { "epoch": 0.61, "grad_norm": 0.36462251254838884, "learning_rate": 1.0394935557403684e-05, "loss": 0.8403, "step": 1872 }, { "epoch": 0.61, "grad_norm": 0.2847199696333977, "learning_rate": 1.0379837736166548e-05, "loss": 0.5801, "step": 1873 }, { "epoch": 0.61, "grad_norm": 0.313222546773386, "learning_rate": 1.036474508437579e-05, "loss": 0.5954, "step": 1874 }, { "epoch": 0.61, "grad_norm": 0.33593512737418546, "learning_rate": 1.0349657618918402e-05, "loss": 0.8307, "step": 1875 }, { "epoch": 0.61, "grad_norm": 0.2978922793325932, "learning_rate": 1.033457535667558e-05, "loss": 0.6057, "step": 1876 }, { "epoch": 0.61, "grad_norm": 0.34769850849999556, "learning_rate": 1.0319498314522694e-05, "loss": 0.6425, "step": 1877 }, { "epoch": 0.61, "grad_norm": 0.28893093103923034, "learning_rate": 1.0304426509329272e-05, "loss": 0.571, "step": 1878 }, { "epoch": 0.61, "grad_norm": 0.35299702262295896, "learning_rate": 1.0289359957958979e-05, "loss": 0.8477, "step": 1879 }, { "epoch": 0.61, "grad_norm": 0.30480114662048924, "learning_rate": 1.0274298677269612e-05, "loss": 0.5714, "step": 1880 }, { "epoch": 0.61, "grad_norm": 0.2883346278808315, "learning_rate": 1.025924268411306e-05, "loss": 0.606, "step": 1881 }, { "epoch": 0.61, "grad_norm": 0.35669095472134543, "learning_rate": 1.0244191995335299e-05, "loss": 0.8644, "step": 1882 }, { "epoch": 0.61, "grad_norm": 0.32004242673061073, "learning_rate": 1.0229146627776378e-05, "loss": 0.5917, "step": 1883 }, { "epoch": 0.62, "grad_norm": 0.2914346997832967, "learning_rate": 1.0214106598270381e-05, "loss": 0.5749, "step": 1884 }, { "epoch": 0.62, "grad_norm": 0.3653612727173593, "learning_rate": 1.0199071923645425e-05, "loss": 0.8588, "step": 1885 }, { "epoch": 0.62, "grad_norm": 1.4391963479280796, "learning_rate": 1.0184042620723637e-05, "loss": 0.6131, "step": 1886 }, { "epoch": 0.62, "grad_norm": 0.296224419040508, "learning_rate": 1.0169018706321132e-05, "loss": 0.5996, "step": 1887 }, { "epoch": 0.62, "grad_norm": 0.38342512701080184, "learning_rate": 1.0154000197247986e-05, "loss": 0.6335, "step": 1888 }, { "epoch": 0.62, "grad_norm": 0.30632475566218254, "learning_rate": 1.0138987110308242e-05, "loss": 0.5756, "step": 1889 }, { "epoch": 0.62, "grad_norm": 0.3634615521938822, "learning_rate": 1.0123979462299874e-05, "loss": 0.8616, "step": 1890 }, { "epoch": 0.62, "grad_norm": 0.30620204497826337, "learning_rate": 1.0108977270014752e-05, "loss": 0.5967, "step": 1891 }, { "epoch": 0.62, "grad_norm": 0.2897115578623286, "learning_rate": 1.0093980550238676e-05, "loss": 0.5767, "step": 1892 }, { "epoch": 0.62, "grad_norm": 0.45017777573770845, "learning_rate": 1.007898931975129e-05, "loss": 0.865, "step": 1893 }, { "epoch": 0.62, "grad_norm": 0.32644378397294876, "learning_rate": 1.0064003595326118e-05, "loss": 0.593, "step": 1894 }, { "epoch": 0.62, "grad_norm": 0.2870611527590156, "learning_rate": 1.0049023393730502e-05, "loss": 0.5809, "step": 1895 }, { "epoch": 0.62, "grad_norm": 0.35571461466162746, "learning_rate": 1.0034048731725631e-05, "loss": 0.856, "step": 1896 }, { "epoch": 0.62, "grad_norm": 0.29883023200455716, "learning_rate": 1.0019079626066473e-05, "loss": 0.5954, "step": 1897 }, { "epoch": 0.62, "grad_norm": 0.28943647313616094, "learning_rate": 1.000411609350179e-05, "loss": 0.5894, "step": 1898 }, { "epoch": 0.62, "grad_norm": 0.34131874228716585, "learning_rate": 9.989158150774109e-06, "loss": 0.6162, "step": 1899 }, { "epoch": 0.62, "grad_norm": 0.29333209741514543, "learning_rate": 9.974205814619697e-06, "loss": 0.5989, "step": 1900 }, { "epoch": 0.62, "grad_norm": 0.33977458923275783, "learning_rate": 9.959259101768546e-06, "loss": 0.818, "step": 1901 }, { "epoch": 0.62, "grad_norm": 0.29214852094755717, "learning_rate": 9.944318028944374e-06, "loss": 0.5953, "step": 1902 }, { "epoch": 0.62, "grad_norm": 0.3049372108435585, "learning_rate": 9.929382612864568e-06, "loss": 0.5848, "step": 1903 }, { "epoch": 0.62, "grad_norm": 0.3413166083244434, "learning_rate": 9.91445287024019e-06, "loss": 0.8607, "step": 1904 }, { "epoch": 0.62, "grad_norm": 0.3122309161043523, "learning_rate": 9.899528817775964e-06, "loss": 0.5744, "step": 1905 }, { "epoch": 0.62, "grad_norm": 0.28163287835629636, "learning_rate": 9.884610472170235e-06, "loss": 0.592, "step": 1906 }, { "epoch": 0.62, "grad_norm": 0.3572000765999966, "learning_rate": 9.86969785011497e-06, "loss": 0.8666, "step": 1907 }, { "epoch": 0.62, "grad_norm": 0.3050498134491161, "learning_rate": 9.854790968295731e-06, "loss": 0.5696, "step": 1908 }, { "epoch": 0.62, "grad_norm": 0.3061751188940535, "learning_rate": 9.839889843391658e-06, "loss": 0.5882, "step": 1909 }, { "epoch": 0.62, "grad_norm": 0.3638617921526853, "learning_rate": 9.824994492075444e-06, "loss": 0.6465, "step": 1910 }, { "epoch": 0.62, "grad_norm": 0.3566731659905755, "learning_rate": 9.810104931013324e-06, "loss": 0.8619, "step": 1911 }, { "epoch": 0.62, "grad_norm": 0.3023931660324482, "learning_rate": 9.795221176865065e-06, "loss": 0.6076, "step": 1912 }, { "epoch": 0.62, "grad_norm": 0.2932828879896059, "learning_rate": 9.780343246283924e-06, "loss": 0.5714, "step": 1913 }, { "epoch": 0.63, "grad_norm": 0.37649001756496125, "learning_rate": 9.765471155916645e-06, "loss": 0.8714, "step": 1914 }, { "epoch": 0.63, "grad_norm": 0.3076009936612092, "learning_rate": 9.75060492240344e-06, "loss": 0.5661, "step": 1915 }, { "epoch": 0.63, "grad_norm": 0.30876821180530245, "learning_rate": 9.73574456237797e-06, "loss": 0.5775, "step": 1916 }, { "epoch": 0.63, "grad_norm": 0.31612621414213743, "learning_rate": 9.72089009246731e-06, "loss": 0.5894, "step": 1917 }, { "epoch": 0.63, "grad_norm": 0.35928799333984857, "learning_rate": 9.70604152929197e-06, "loss": 0.8481, "step": 1918 }, { "epoch": 0.63, "grad_norm": 0.3130888139504024, "learning_rate": 9.691198889465826e-06, "loss": 0.5897, "step": 1919 }, { "epoch": 0.63, "grad_norm": 0.2867731881757575, "learning_rate": 9.676362189596135e-06, "loss": 0.5774, "step": 1920 }, { "epoch": 0.63, "grad_norm": 0.36587777755373285, "learning_rate": 9.661531446283519e-06, "loss": 0.6381, "step": 1921 }, { "epoch": 0.63, "grad_norm": 0.3569795530844735, "learning_rate": 9.646706676121924e-06, "loss": 0.8366, "step": 1922 }, { "epoch": 0.63, "grad_norm": 0.2995102072181223, "learning_rate": 9.631887895698608e-06, "loss": 0.5891, "step": 1923 }, { "epoch": 0.63, "grad_norm": 0.2946312549465398, "learning_rate": 9.617075121594142e-06, "loss": 0.5895, "step": 1924 }, { "epoch": 0.63, "grad_norm": 0.3500296340522303, "learning_rate": 9.602268370382363e-06, "loss": 0.8448, "step": 1925 }, { "epoch": 0.63, "grad_norm": 0.3014611117520793, "learning_rate": 9.587467658630374e-06, "loss": 0.5586, "step": 1926 }, { "epoch": 0.63, "grad_norm": 0.32293520350603167, "learning_rate": 9.572673002898527e-06, "loss": 0.6099, "step": 1927 }, { "epoch": 0.63, "grad_norm": 0.29724304570651083, "learning_rate": 9.557884419740387e-06, "loss": 0.5881, "step": 1928 }, { "epoch": 0.63, "grad_norm": 0.2879289464960066, "learning_rate": 9.543101925702733e-06, "loss": 0.6083, "step": 1929 }, { "epoch": 0.63, "grad_norm": 0.3425507344863715, "learning_rate": 9.528325537325516e-06, "loss": 0.8381, "step": 1930 }, { "epoch": 0.63, "grad_norm": 0.3125844517873092, "learning_rate": 9.513555271141884e-06, "loss": 0.7569, "step": 1931 }, { "epoch": 0.63, "grad_norm": 0.2723116038746969, "learning_rate": 9.49879114367811e-06, "loss": 0.3957, "step": 1932 }, { "epoch": 0.63, "grad_norm": 0.35245876601810777, "learning_rate": 9.4840331714536e-06, "loss": 0.8685, "step": 1933 }, { "epoch": 0.63, "grad_norm": 0.2855255605239495, "learning_rate": 9.46928137098089e-06, "loss": 0.6132, "step": 1934 }, { "epoch": 0.63, "grad_norm": 0.2893369829221107, "learning_rate": 9.454535758765597e-06, "loss": 0.5955, "step": 1935 }, { "epoch": 0.63, "grad_norm": 0.3471385381230962, "learning_rate": 9.439796351306408e-06, "loss": 0.871, "step": 1936 }, { "epoch": 0.63, "grad_norm": 0.3017435340834715, "learning_rate": 9.42506316509509e-06, "loss": 0.5975, "step": 1937 }, { "epoch": 0.63, "grad_norm": 0.32290726949001897, "learning_rate": 9.410336216616427e-06, "loss": 0.6005, "step": 1938 }, { "epoch": 0.63, "grad_norm": 0.30050593060742653, "learning_rate": 9.395615522348227e-06, "loss": 0.5838, "step": 1939 }, { "epoch": 0.63, "grad_norm": 0.34519254176658903, "learning_rate": 9.380901098761319e-06, "loss": 0.8335, "step": 1940 }, { "epoch": 0.63, "grad_norm": 0.3146459371868473, "learning_rate": 9.366192962319496e-06, "loss": 0.5897, "step": 1941 }, { "epoch": 0.63, "grad_norm": 0.29230209001007434, "learning_rate": 9.351491129479519e-06, "loss": 0.5832, "step": 1942 }, { "epoch": 0.63, "grad_norm": 0.342805548692944, "learning_rate": 9.336795616691104e-06, "loss": 0.5803, "step": 1943 }, { "epoch": 0.63, "grad_norm": 0.35402289578129537, "learning_rate": 9.322106440396891e-06, "loss": 0.8365, "step": 1944 }, { "epoch": 0.64, "grad_norm": 0.3045796455907794, "learning_rate": 9.307423617032429e-06, "loss": 0.5626, "step": 1945 }, { "epoch": 0.64, "grad_norm": 0.38085973322998046, "learning_rate": 9.292747163026155e-06, "loss": 0.6013, "step": 1946 }, { "epoch": 0.64, "grad_norm": 0.376693734396948, "learning_rate": 9.278077094799396e-06, "loss": 0.8874, "step": 1947 }, { "epoch": 0.64, "grad_norm": 0.29599697273516185, "learning_rate": 9.263413428766312e-06, "loss": 0.5848, "step": 1948 }, { "epoch": 0.64, "grad_norm": 0.33522444207420843, "learning_rate": 9.248756181333911e-06, "loss": 0.6252, "step": 1949 }, { "epoch": 0.64, "grad_norm": 0.34065286631727115, "learning_rate": 9.234105368902026e-06, "loss": 0.5998, "step": 1950 }, { "epoch": 0.64, "grad_norm": 0.42874696574863935, "learning_rate": 9.219461007863278e-06, "loss": 0.8791, "step": 1951 }, { "epoch": 0.64, "grad_norm": 0.29624366710176697, "learning_rate": 9.20482311460307e-06, "loss": 0.5968, "step": 1952 }, { "epoch": 0.64, "grad_norm": 0.2998459600156505, "learning_rate": 9.19019170549958e-06, "loss": 0.586, "step": 1953 }, { "epoch": 0.64, "grad_norm": 0.39440752308431754, "learning_rate": 9.175566796923721e-06, "loss": 0.8593, "step": 1954 }, { "epoch": 0.64, "grad_norm": 0.2998339573015956, "learning_rate": 9.16094840523913e-06, "loss": 0.588, "step": 1955 }, { "epoch": 0.64, "grad_norm": 0.28340024013044757, "learning_rate": 9.146336546802162e-06, "loss": 0.5814, "step": 1956 }, { "epoch": 0.64, "grad_norm": 0.3545936735923079, "learning_rate": 9.131731237961856e-06, "loss": 0.8436, "step": 1957 }, { "epoch": 0.64, "grad_norm": 0.31307281618866944, "learning_rate": 9.117132495059918e-06, "loss": 0.5945, "step": 1958 }, { "epoch": 0.64, "grad_norm": 0.43981862151280166, "learning_rate": 9.102540334430724e-06, "loss": 0.5826, "step": 1959 }, { "epoch": 0.64, "grad_norm": 0.3329769064780792, "learning_rate": 9.08795477240127e-06, "loss": 0.6307, "step": 1960 }, { "epoch": 0.64, "grad_norm": 0.2825309057857243, "learning_rate": 9.073375825291172e-06, "loss": 0.5927, "step": 1961 }, { "epoch": 0.64, "grad_norm": 0.35337922806443967, "learning_rate": 9.058803509412647e-06, "loss": 0.8726, "step": 1962 }, { "epoch": 0.64, "grad_norm": 0.29236178223250986, "learning_rate": 9.044237841070494e-06, "loss": 0.5684, "step": 1963 }, { "epoch": 0.64, "grad_norm": 0.2886287978716837, "learning_rate": 9.02967883656207e-06, "loss": 0.581, "step": 1964 }, { "epoch": 0.64, "grad_norm": 0.37267779841450793, "learning_rate": 9.015126512177275e-06, "loss": 0.8528, "step": 1965 }, { "epoch": 0.64, "grad_norm": 0.2938696394587754, "learning_rate": 9.000580884198537e-06, "loss": 0.5883, "step": 1966 }, { "epoch": 0.64, "grad_norm": 0.373479412406044, "learning_rate": 8.986041968900797e-06, "loss": 0.5965, "step": 1967 }, { "epoch": 0.64, "grad_norm": 0.566744462159, "learning_rate": 8.971509782551473e-06, "loss": 0.8799, "step": 1968 }, { "epoch": 0.64, "grad_norm": 0.2745267390983921, "learning_rate": 8.956984341410469e-06, "loss": 0.5527, "step": 1969 }, { "epoch": 0.64, "grad_norm": 0.3211469772703663, "learning_rate": 8.94246566173013e-06, "loss": 0.6253, "step": 1970 }, { "epoch": 0.64, "grad_norm": 0.2993753037841634, "learning_rate": 8.927953759755236e-06, "loss": 0.5902, "step": 1971 }, { "epoch": 0.64, "grad_norm": 0.2876956767764448, "learning_rate": 8.913448651722994e-06, "loss": 0.5773, "step": 1972 }, { "epoch": 0.64, "grad_norm": 0.3587407140419981, "learning_rate": 8.898950353863e-06, "loss": 0.8151, "step": 1973 }, { "epoch": 0.64, "grad_norm": 0.2873804655036646, "learning_rate": 8.884458882397226e-06, "loss": 0.5741, "step": 1974 }, { "epoch": 0.65, "grad_norm": 0.28272779312286694, "learning_rate": 8.869974253540023e-06, "loss": 0.5728, "step": 1975 }, { "epoch": 0.65, "grad_norm": 0.34772920467447355, "learning_rate": 8.85549648349807e-06, "loss": 0.8348, "step": 1976 }, { "epoch": 0.65, "grad_norm": 0.3079149172090381, "learning_rate": 8.841025588470374e-06, "loss": 0.5771, "step": 1977 }, { "epoch": 0.65, "grad_norm": 0.2963992052316419, "learning_rate": 8.826561584648262e-06, "loss": 0.5889, "step": 1978 }, { "epoch": 0.65, "grad_norm": 0.35100839662622835, "learning_rate": 8.812104488215334e-06, "loss": 0.8806, "step": 1979 }, { "epoch": 0.65, "grad_norm": 0.2961596613035231, "learning_rate": 8.797654315347472e-06, "loss": 0.5812, "step": 1980 }, { "epoch": 0.65, "grad_norm": 0.3361071127329587, "learning_rate": 8.783211082212804e-06, "loss": 0.6382, "step": 1981 }, { "epoch": 0.65, "grad_norm": 0.3054939182350643, "learning_rate": 8.768774804971705e-06, "loss": 0.5899, "step": 1982 }, { "epoch": 0.65, "grad_norm": 0.35090791260010035, "learning_rate": 8.754345499776757e-06, "loss": 0.8749, "step": 1983 }, { "epoch": 0.65, "grad_norm": 0.2903944820097756, "learning_rate": 8.739923182772732e-06, "loss": 0.5869, "step": 1984 }, { "epoch": 0.65, "grad_norm": 0.28950866932252517, "learning_rate": 8.72550787009661e-06, "loss": 0.5682, "step": 1985 }, { "epoch": 0.65, "grad_norm": 0.3078326116800906, "learning_rate": 8.711099577877514e-06, "loss": 0.5872, "step": 1986 }, { "epoch": 0.65, "grad_norm": 0.3994800699990002, "learning_rate": 8.696698322236706e-06, "loss": 0.8118, "step": 1987 }, { "epoch": 0.65, "grad_norm": 0.28442211711856596, "learning_rate": 8.682304119287602e-06, "loss": 0.565, "step": 1988 }, { "epoch": 0.65, "grad_norm": 0.2840531453646093, "learning_rate": 8.6679169851357e-06, "loss": 0.584, "step": 1989 }, { "epoch": 0.65, "grad_norm": 0.36349554721682414, "learning_rate": 8.653536935878595e-06, "loss": 0.8813, "step": 1990 }, { "epoch": 0.65, "grad_norm": 0.3113491729985897, "learning_rate": 8.639163987605977e-06, "loss": 0.5819, "step": 1991 }, { "epoch": 0.65, "grad_norm": 0.3316453200435214, "learning_rate": 8.624798156399554e-06, "loss": 0.6672, "step": 1992 }, { "epoch": 0.65, "grad_norm": 0.2849188169848693, "learning_rate": 8.610439458333088e-06, "loss": 0.5676, "step": 1993 }, { "epoch": 0.65, "grad_norm": 0.35002919972313595, "learning_rate": 8.596087909472374e-06, "loss": 0.8354, "step": 1994 }, { "epoch": 0.65, "grad_norm": 0.29074693075593494, "learning_rate": 8.58174352587519e-06, "loss": 0.607, "step": 1995 }, { "epoch": 0.65, "grad_norm": 0.2846125788381585, "learning_rate": 8.567406323591297e-06, "loss": 0.58, "step": 1996 }, { "epoch": 0.65, "grad_norm": 0.3572136233994831, "learning_rate": 8.553076318662425e-06, "loss": 0.8609, "step": 1997 }, { "epoch": 0.65, "grad_norm": 0.2874974171559899, "learning_rate": 8.538753527122262e-06, "loss": 0.5725, "step": 1998 }, { "epoch": 0.65, "grad_norm": 0.28194436968584763, "learning_rate": 8.524437964996407e-06, "loss": 0.5862, "step": 1999 }, { "epoch": 0.65, "grad_norm": 0.42020725421391497, "learning_rate": 8.510129648302373e-06, "loss": 0.8505, "step": 2000 }, { "epoch": 0.65, "grad_norm": 0.31025579646421314, "learning_rate": 8.495828593049589e-06, "loss": 0.6138, "step": 2001 }, { "epoch": 0.65, "grad_norm": 0.29552674280325025, "learning_rate": 8.481534815239323e-06, "loss": 0.5873, "step": 2002 }, { "epoch": 0.65, "grad_norm": 0.3298066448216788, "learning_rate": 8.467248330864719e-06, "loss": 0.6199, "step": 2003 }, { "epoch": 0.65, "grad_norm": 0.2904753625671434, "learning_rate": 8.45296915591077e-06, "loss": 0.6173, "step": 2004 }, { "epoch": 0.65, "grad_norm": 0.3792879407248594, "learning_rate": 8.438697306354274e-06, "loss": 0.8611, "step": 2005 }, { "epoch": 0.66, "grad_norm": 0.2980951927549134, "learning_rate": 8.424432798163838e-06, "loss": 0.5897, "step": 2006 }, { "epoch": 0.66, "grad_norm": 0.29850123614878205, "learning_rate": 8.410175647299864e-06, "loss": 0.5548, "step": 2007 }, { "epoch": 0.66, "grad_norm": 0.3414656962371591, "learning_rate": 8.395925869714508e-06, "loss": 0.8506, "step": 2008 }, { "epoch": 0.66, "grad_norm": 0.33358910841590406, "learning_rate": 8.381683481351676e-06, "loss": 0.6352, "step": 2009 }, { "epoch": 0.66, "grad_norm": 0.29215066516987975, "learning_rate": 8.367448498147034e-06, "loss": 0.5858, "step": 2010 }, { "epoch": 0.66, "grad_norm": 0.2884920283558209, "learning_rate": 8.353220936027919e-06, "loss": 0.6068, "step": 2011 }, { "epoch": 0.66, "grad_norm": 0.3595200782262894, "learning_rate": 8.339000810913388e-06, "loss": 0.8484, "step": 2012 }, { "epoch": 0.66, "grad_norm": 0.2955571214344184, "learning_rate": 8.324788138714187e-06, "loss": 0.5732, "step": 2013 }, { "epoch": 0.66, "grad_norm": 0.28960056153439523, "learning_rate": 8.310582935332705e-06, "loss": 0.5878, "step": 2014 }, { "epoch": 0.66, "grad_norm": 0.28159910216255013, "learning_rate": 8.296385216662984e-06, "loss": 0.5779, "step": 2015 }, { "epoch": 0.66, "grad_norm": 0.34544784548747104, "learning_rate": 8.282194998590676e-06, "loss": 0.835, "step": 2016 }, { "epoch": 0.66, "grad_norm": 0.2979023714049358, "learning_rate": 8.268012296993067e-06, "loss": 0.6093, "step": 2017 }, { "epoch": 0.66, "grad_norm": 0.42162088369096534, "learning_rate": 8.253837127739016e-06, "loss": 0.5937, "step": 2018 }, { "epoch": 0.66, "grad_norm": 0.3578508091370182, "learning_rate": 8.239669506688946e-06, "loss": 0.8604, "step": 2019 }, { "epoch": 0.66, "grad_norm": 0.344273487928932, "learning_rate": 8.225509449694866e-06, "loss": 0.6578, "step": 2020 }, { "epoch": 0.66, "grad_norm": 0.2898603772071304, "learning_rate": 8.211356972600281e-06, "loss": 0.5949, "step": 2021 }, { "epoch": 0.66, "grad_norm": 0.2946711603547827, "learning_rate": 8.197212091240237e-06, "loss": 0.5642, "step": 2022 }, { "epoch": 0.66, "grad_norm": 0.3414100721699114, "learning_rate": 8.183074821441287e-06, "loss": 0.838, "step": 2023 }, { "epoch": 0.66, "grad_norm": 0.2809970525292798, "learning_rate": 8.168945179021459e-06, "loss": 0.5884, "step": 2024 }, { "epoch": 0.66, "grad_norm": 0.29357984339953197, "learning_rate": 8.154823179790237e-06, "loss": 0.5915, "step": 2025 }, { "epoch": 0.66, "grad_norm": 0.3632412690671305, "learning_rate": 8.140708839548578e-06, "loss": 0.8528, "step": 2026 }, { "epoch": 0.66, "grad_norm": 0.2900664764684141, "learning_rate": 8.126602174088843e-06, "loss": 0.5989, "step": 2027 }, { "epoch": 0.66, "grad_norm": 0.2994592643426804, "learning_rate": 8.112503199194821e-06, "loss": 0.5675, "step": 2028 }, { "epoch": 0.66, "grad_norm": 0.33953264512318365, "learning_rate": 8.098411930641705e-06, "loss": 0.8604, "step": 2029 }, { "epoch": 0.66, "grad_norm": 0.33707174170936854, "learning_rate": 8.084328384196035e-06, "loss": 0.5949, "step": 2030 }, { "epoch": 0.66, "grad_norm": 0.33961612091870724, "learning_rate": 8.070252575615737e-06, "loss": 0.6256, "step": 2031 }, { "epoch": 0.66, "grad_norm": 0.291944319762439, "learning_rate": 8.056184520650064e-06, "loss": 0.6057, "step": 2032 }, { "epoch": 0.66, "grad_norm": 0.496612022573627, "learning_rate": 8.042124235039613e-06, "loss": 0.5844, "step": 2033 }, { "epoch": 0.66, "grad_norm": 0.35610230815292193, "learning_rate": 8.02807173451627e-06, "loss": 0.8212, "step": 2034 }, { "epoch": 0.66, "grad_norm": 0.28728591575005846, "learning_rate": 8.014027034803209e-06, "loss": 0.6015, "step": 2035 }, { "epoch": 0.66, "grad_norm": 0.28998587161178463, "learning_rate": 7.999990151614895e-06, "loss": 0.5855, "step": 2036 }, { "epoch": 0.67, "grad_norm": 0.3547755202390015, "learning_rate": 7.98596110065703e-06, "loss": 0.8575, "step": 2037 }, { "epoch": 0.67, "grad_norm": 0.2973996595314882, "learning_rate": 7.97193989762656e-06, "loss": 0.5936, "step": 2038 }, { "epoch": 0.67, "grad_norm": 0.28110051529973196, "learning_rate": 7.957926558211643e-06, "loss": 0.5965, "step": 2039 }, { "epoch": 0.67, "grad_norm": 0.3485024861995234, "learning_rate": 7.943921098091644e-06, "loss": 0.8412, "step": 2040 }, { "epoch": 0.67, "grad_norm": 0.31667613162191577, "learning_rate": 7.929923532937108e-06, "loss": 0.5942, "step": 2041 }, { "epoch": 0.67, "grad_norm": 0.3320143447646092, "learning_rate": 7.915933878409762e-06, "loss": 0.6191, "step": 2042 }, { "epoch": 0.67, "grad_norm": 0.295449982013471, "learning_rate": 7.901952150162459e-06, "loss": 0.589, "step": 2043 }, { "epoch": 0.67, "grad_norm": 0.2969006762018296, "learning_rate": 7.88797836383919e-06, "loss": 0.6, "step": 2044 }, { "epoch": 0.67, "grad_norm": 0.34546100931747303, "learning_rate": 7.874012535075077e-06, "loss": 0.8288, "step": 2045 }, { "epoch": 0.67, "grad_norm": 0.29597869875889216, "learning_rate": 7.860054679496316e-06, "loss": 0.5938, "step": 2046 }, { "epoch": 0.67, "grad_norm": 0.3089482153754184, "learning_rate": 7.846104812720195e-06, "loss": 0.5869, "step": 2047 }, { "epoch": 0.67, "grad_norm": 0.36433287687667815, "learning_rate": 7.832162950355057e-06, "loss": 0.8423, "step": 2048 }, { "epoch": 0.67, "grad_norm": 0.30750164156224075, "learning_rate": 7.81822910800029e-06, "loss": 0.5808, "step": 2049 }, { "epoch": 0.67, "grad_norm": 0.2834041218856756, "learning_rate": 7.804303301246311e-06, "loss": 0.5802, "step": 2050 }, { "epoch": 0.67, "grad_norm": 0.3588649017408353, "learning_rate": 7.79038554567454e-06, "loss": 0.8339, "step": 2051 }, { "epoch": 0.67, "grad_norm": 0.3229090567183305, "learning_rate": 7.776475856857409e-06, "loss": 0.5773, "step": 2052 }, { "epoch": 0.67, "grad_norm": 0.33401509193615897, "learning_rate": 7.762574250358294e-06, "loss": 0.6517, "step": 2053 }, { "epoch": 0.67, "grad_norm": 0.2876633153073723, "learning_rate": 7.74868074173154e-06, "loss": 0.6028, "step": 2054 }, { "epoch": 0.67, "grad_norm": 0.35370813609111623, "learning_rate": 7.73479534652245e-06, "loss": 0.8448, "step": 2055 }, { "epoch": 0.67, "grad_norm": 0.2964622276698857, "learning_rate": 7.720918080267222e-06, "loss": 0.5991, "step": 2056 }, { "epoch": 0.67, "grad_norm": 0.29932205649540894, "learning_rate": 7.707048958492971e-06, "loss": 0.5802, "step": 2057 }, { "epoch": 0.67, "grad_norm": 0.2987836295441598, "learning_rate": 7.693187996717698e-06, "loss": 0.558, "step": 2058 }, { "epoch": 0.67, "grad_norm": 0.3332542862615514, "learning_rate": 7.679335210450276e-06, "loss": 0.8295, "step": 2059 }, { "epoch": 0.67, "grad_norm": 0.29839197823025715, "learning_rate": 7.66549061519042e-06, "loss": 0.6021, "step": 2060 }, { "epoch": 0.67, "grad_norm": 0.30131371127317347, "learning_rate": 7.651654226428696e-06, "loss": 0.6169, "step": 2061 }, { "epoch": 0.67, "grad_norm": 0.34945215836212223, "learning_rate": 7.63782605964648e-06, "loss": 0.8486, "step": 2062 }, { "epoch": 0.67, "grad_norm": 0.28812961134871934, "learning_rate": 7.624006130315942e-06, "loss": 0.5772, "step": 2063 }, { "epoch": 0.67, "grad_norm": 0.3457546324044742, "learning_rate": 7.61019445390005e-06, "loss": 0.6418, "step": 2064 }, { "epoch": 0.67, "grad_norm": 0.2927994871451805, "learning_rate": 7.596391045852529e-06, "loss": 0.5927, "step": 2065 }, { "epoch": 0.67, "grad_norm": 0.3537319692050098, "learning_rate": 7.5825959216178495e-06, "loss": 0.8756, "step": 2066 }, { "epoch": 0.68, "grad_norm": 0.3257795549440655, "learning_rate": 7.568809096631218e-06, "loss": 0.5893, "step": 2067 }, { "epoch": 0.68, "grad_norm": 0.2800910526417546, "learning_rate": 7.555030586318555e-06, "loss": 0.5664, "step": 2068 }, { "epoch": 0.68, "grad_norm": 0.35060706467989144, "learning_rate": 7.541260406096479e-06, "loss": 0.8403, "step": 2069 }, { "epoch": 0.68, "grad_norm": 0.31480550388991096, "learning_rate": 7.527498571372275e-06, "loss": 0.5698, "step": 2070 }, { "epoch": 0.68, "grad_norm": 0.2865617558252989, "learning_rate": 7.513745097543917e-06, "loss": 0.5859, "step": 2071 }, { "epoch": 0.68, "grad_norm": 0.3320322578331628, "learning_rate": 7.500000000000004e-06, "loss": 0.879, "step": 2072 }, { "epoch": 0.68, "grad_norm": 0.2899494798389621, "learning_rate": 7.486263294119757e-06, "loss": 0.5857, "step": 2073 }, { "epoch": 0.68, "grad_norm": 0.28613511194400953, "learning_rate": 7.472534995273034e-06, "loss": 0.5946, "step": 2074 }, { "epoch": 0.68, "grad_norm": 0.32443886898420526, "learning_rate": 7.458815118820262e-06, "loss": 0.6404, "step": 2075 }, { "epoch": 0.68, "grad_norm": 0.5287170012178777, "learning_rate": 7.445103680112457e-06, "loss": 0.5998, "step": 2076 }, { "epoch": 0.68, "grad_norm": 0.3535306343503661, "learning_rate": 7.431400694491183e-06, "loss": 0.8604, "step": 2077 }, { "epoch": 0.68, "grad_norm": 0.28374856890746186, "learning_rate": 7.4177061772885596e-06, "loss": 0.5833, "step": 2078 }, { "epoch": 0.68, "grad_norm": 0.2767719536433052, "learning_rate": 7.4040201438272154e-06, "loss": 0.5801, "step": 2079 }, { "epoch": 0.68, "grad_norm": 0.3255907062034998, "learning_rate": 7.390342609420306e-06, "loss": 0.6276, "step": 2080 }, { "epoch": 0.68, "grad_norm": 0.3018817785976429, "learning_rate": 7.376673589371466e-06, "loss": 0.7287, "step": 2081 }, { "epoch": 0.68, "grad_norm": 0.29435800273483953, "learning_rate": 7.3630130989748e-06, "loss": 0.5769, "step": 2082 }, { "epoch": 0.68, "grad_norm": 0.35876637094270236, "learning_rate": 7.349361153514871e-06, "loss": 0.8896, "step": 2083 }, { "epoch": 0.68, "grad_norm": 0.29453963445194814, "learning_rate": 7.3357177682666926e-06, "loss": 0.5926, "step": 2084 }, { "epoch": 0.68, "grad_norm": 0.29073388128831323, "learning_rate": 7.322082958495686e-06, "loss": 0.5877, "step": 2085 }, { "epoch": 0.68, "grad_norm": 0.33538979014223047, "learning_rate": 7.308456739457686e-06, "loss": 0.6289, "step": 2086 }, { "epoch": 0.68, "grad_norm": 0.3753896305674978, "learning_rate": 7.294839126398909e-06, "loss": 0.5923, "step": 2087 }, { "epoch": 0.68, "grad_norm": 0.3575342061858606, "learning_rate": 7.281230134555946e-06, "loss": 0.8485, "step": 2088 }, { "epoch": 0.68, "grad_norm": 0.28410943801587896, "learning_rate": 7.267629779155736e-06, "loss": 0.5868, "step": 2089 }, { "epoch": 0.68, "grad_norm": 0.2785850788579286, "learning_rate": 7.254038075415573e-06, "loss": 0.5868, "step": 2090 }, { "epoch": 0.68, "grad_norm": 0.32124258166258907, "learning_rate": 7.24045503854305e-06, "loss": 0.6019, "step": 2091 }, { "epoch": 0.68, "grad_norm": 0.3772124543106557, "learning_rate": 7.226880683736066e-06, "loss": 0.8551, "step": 2092 }, { "epoch": 0.68, "grad_norm": 0.28507096813804284, "learning_rate": 7.213315026182822e-06, "loss": 0.562, "step": 2093 }, { "epoch": 0.68, "grad_norm": 0.2874882034556249, "learning_rate": 7.19975808106177e-06, "loss": 0.5968, "step": 2094 }, { "epoch": 0.68, "grad_norm": 0.3439580512297067, "learning_rate": 7.186209863541618e-06, "loss": 0.8219, "step": 2095 }, { "epoch": 0.68, "grad_norm": 0.29721589058017434, "learning_rate": 7.172670388781314e-06, "loss": 0.5803, "step": 2096 }, { "epoch": 0.68, "grad_norm": 1.168503796810182, "learning_rate": 7.159139671930018e-06, "loss": 0.5964, "step": 2097 }, { "epoch": 0.69, "grad_norm": 0.3410158543144548, "learning_rate": 7.1456177281270945e-06, "loss": 0.8512, "step": 2098 }, { "epoch": 0.69, "grad_norm": 0.28483184835317343, "learning_rate": 7.132104572502085e-06, "loss": 0.5991, "step": 2099 }, { "epoch": 0.69, "grad_norm": 0.6100311635068413, "learning_rate": 7.118600220174716e-06, "loss": 0.6018, "step": 2100 }, { "epoch": 0.69, "grad_norm": 0.3958089307113638, "learning_rate": 7.105104686254844e-06, "loss": 0.5793, "step": 2101 }, { "epoch": 0.69, "grad_norm": 0.3260228212414272, "learning_rate": 7.091617985842463e-06, "loss": 0.6476, "step": 2102 }, { "epoch": 0.69, "grad_norm": 0.35878752777022055, "learning_rate": 7.078140134027697e-06, "loss": 0.8572, "step": 2103 }, { "epoch": 0.69, "grad_norm": 0.29459227059632964, "learning_rate": 7.064671145890753e-06, "loss": 0.5774, "step": 2104 }, { "epoch": 0.69, "grad_norm": 0.29486303533392305, "learning_rate": 7.051211036501928e-06, "loss": 0.5957, "step": 2105 }, { "epoch": 0.69, "grad_norm": 0.35155954668396777, "learning_rate": 7.037759820921584e-06, "loss": 0.8687, "step": 2106 }, { "epoch": 0.69, "grad_norm": 0.32786542854681344, "learning_rate": 7.02431751420013e-06, "loss": 0.5943, "step": 2107 }, { "epoch": 0.69, "grad_norm": 0.29473701874015956, "learning_rate": 7.010884131378002e-06, "loss": 0.5878, "step": 2108 }, { "epoch": 0.69, "grad_norm": 0.35429532527685886, "learning_rate": 6.99745968748567e-06, "loss": 0.8536, "step": 2109 }, { "epoch": 0.69, "grad_norm": 0.33214756768393056, "learning_rate": 6.984044197543586e-06, "loss": 0.5892, "step": 2110 }, { "epoch": 0.69, "grad_norm": 0.2978482018036655, "learning_rate": 6.970637676562178e-06, "loss": 0.5823, "step": 2111 }, { "epoch": 0.69, "grad_norm": 0.3629517886016944, "learning_rate": 6.9572401395418615e-06, "loss": 0.8355, "step": 2112 }, { "epoch": 0.69, "grad_norm": 0.29596186249393125, "learning_rate": 6.94385160147298e-06, "loss": 0.3903, "step": 2113 }, { "epoch": 0.69, "grad_norm": 0.3389039118363388, "learning_rate": 6.930472077335814e-06, "loss": 0.8609, "step": 2114 }, { "epoch": 0.69, "grad_norm": 0.2935929363242291, "learning_rate": 6.917101582100559e-06, "loss": 0.5704, "step": 2115 }, { "epoch": 0.69, "grad_norm": 0.30497833151717135, "learning_rate": 6.903740130727312e-06, "loss": 0.5909, "step": 2116 }, { "epoch": 0.69, "grad_norm": 0.3420146491863299, "learning_rate": 6.890387738166042e-06, "loss": 0.8366, "step": 2117 }, { "epoch": 0.69, "grad_norm": 0.2899907161832135, "learning_rate": 6.877044419356584e-06, "loss": 0.592, "step": 2118 }, { "epoch": 0.69, "grad_norm": 0.29807583115749897, "learning_rate": 6.863710189228636e-06, "loss": 0.5801, "step": 2119 }, { "epoch": 0.69, "grad_norm": 0.3386889676466844, "learning_rate": 6.850385062701708e-06, "loss": 0.8378, "step": 2120 }, { "epoch": 0.69, "grad_norm": 0.28092440593666373, "learning_rate": 6.837069054685123e-06, "loss": 0.5711, "step": 2121 }, { "epoch": 0.69, "grad_norm": 0.30304967246156145, "learning_rate": 6.8237621800780255e-06, "loss": 0.5926, "step": 2122 }, { "epoch": 0.69, "grad_norm": 0.33487088606821885, "learning_rate": 6.810464453769314e-06, "loss": 0.8168, "step": 2123 }, { "epoch": 0.69, "grad_norm": 0.33118997687714496, "learning_rate": 6.797175890637663e-06, "loss": 0.6316, "step": 2124 }, { "epoch": 0.69, "grad_norm": 0.28969530765945223, "learning_rate": 6.783896505551493e-06, "loss": 0.5588, "step": 2125 }, { "epoch": 0.69, "grad_norm": 0.2859575682960512, "learning_rate": 6.770626313368955e-06, "loss": 0.5824, "step": 2126 }, { "epoch": 0.69, "grad_norm": 0.3619971278878983, "learning_rate": 6.75736532893791e-06, "loss": 0.849, "step": 2127 }, { "epoch": 0.69, "grad_norm": 0.2980387124642338, "learning_rate": 6.744113567095929e-06, "loss": 0.5757, "step": 2128 }, { "epoch": 0.7, "grad_norm": 0.29647208760348404, "learning_rate": 6.730871042670254e-06, "loss": 0.5631, "step": 2129 }, { "epoch": 0.7, "grad_norm": 0.28263843504623304, "learning_rate": 6.7176377704777845e-06, "loss": 0.5726, "step": 2130 }, { "epoch": 0.7, "grad_norm": 0.36262953429201156, "learning_rate": 6.7044137653250866e-06, "loss": 0.8653, "step": 2131 }, { "epoch": 0.7, "grad_norm": 0.3095145819050988, "learning_rate": 6.691199042008346e-06, "loss": 0.5789, "step": 2132 }, { "epoch": 0.7, "grad_norm": 0.28707111086821, "learning_rate": 6.677993615313363e-06, "loss": 0.5747, "step": 2133 }, { "epoch": 0.7, "grad_norm": 0.34148725551677, "learning_rate": 6.6647975000155345e-06, "loss": 0.8279, "step": 2134 }, { "epoch": 0.7, "grad_norm": 0.3266146254747807, "learning_rate": 6.651610710879847e-06, "loss": 0.6066, "step": 2135 }, { "epoch": 0.7, "grad_norm": 0.28282403906096604, "learning_rate": 6.638433262660844e-06, "loss": 0.591, "step": 2136 }, { "epoch": 0.7, "grad_norm": 0.28802036486792876, "learning_rate": 6.625265170102616e-06, "loss": 0.6085, "step": 2137 }, { "epoch": 0.7, "grad_norm": 0.35484772489408106, "learning_rate": 6.6121064479388e-06, "loss": 0.8583, "step": 2138 }, { "epoch": 0.7, "grad_norm": 0.2843254911971729, "learning_rate": 6.598957110892534e-06, "loss": 0.5635, "step": 2139 }, { "epoch": 0.7, "grad_norm": 0.2864988343973829, "learning_rate": 6.585817173676456e-06, "loss": 0.581, "step": 2140 }, { "epoch": 0.7, "grad_norm": 0.37058098982849763, "learning_rate": 6.5726866509926975e-06, "loss": 0.8265, "step": 2141 }, { "epoch": 0.7, "grad_norm": 0.2797147199892006, "learning_rate": 6.559565557532847e-06, "loss": 0.5726, "step": 2142 }, { "epoch": 0.7, "grad_norm": 0.2789349993801664, "learning_rate": 6.546453907977946e-06, "loss": 0.6002, "step": 2143 }, { "epoch": 0.7, "grad_norm": 0.359190352370393, "learning_rate": 6.533351716998466e-06, "loss": 0.8364, "step": 2144 }, { "epoch": 0.7, "grad_norm": 0.29032022579564537, "learning_rate": 6.520258999254296e-06, "loss": 0.5756, "step": 2145 }, { "epoch": 0.7, "grad_norm": 2.4598002916741017, "learning_rate": 6.507175769394727e-06, "loss": 0.6334, "step": 2146 }, { "epoch": 0.7, "grad_norm": 0.2843578596069952, "learning_rate": 6.494102042058441e-06, "loss": 0.592, "step": 2147 }, { "epoch": 0.7, "grad_norm": 0.29979204917730784, "learning_rate": 6.481037831873475e-06, "loss": 0.5757, "step": 2148 }, { "epoch": 0.7, "grad_norm": 0.3507742246122821, "learning_rate": 6.467983153457227e-06, "loss": 0.845, "step": 2149 }, { "epoch": 0.7, "grad_norm": 0.2974166602450591, "learning_rate": 6.454938021416418e-06, "loss": 0.5602, "step": 2150 }, { "epoch": 0.7, "grad_norm": 0.6759322518740117, "learning_rate": 6.441902450347108e-06, "loss": 0.6034, "step": 2151 }, { "epoch": 0.7, "grad_norm": 0.3602371388567657, "learning_rate": 6.428876454834643e-06, "loss": 0.8789, "step": 2152 }, { "epoch": 0.7, "grad_norm": 0.3144956148447381, "learning_rate": 6.415860049453658e-06, "loss": 0.6088, "step": 2153 }, { "epoch": 0.7, "grad_norm": 0.2924887661360178, "learning_rate": 6.40285324876806e-06, "loss": 0.5629, "step": 2154 }, { "epoch": 0.7, "grad_norm": 0.3414832852486675, "learning_rate": 6.389856067331007e-06, "loss": 0.8698, "step": 2155 }, { "epoch": 0.7, "grad_norm": 0.362871535687849, "learning_rate": 6.376868519684892e-06, "loss": 0.5909, "step": 2156 }, { "epoch": 0.7, "grad_norm": 0.3279905775337495, "learning_rate": 6.363890620361344e-06, "loss": 0.624, "step": 2157 }, { "epoch": 0.7, "grad_norm": 0.2853284619375382, "learning_rate": 6.350922383881178e-06, "loss": 0.5817, "step": 2158 }, { "epoch": 0.71, "grad_norm": 0.29557332385549306, "learning_rate": 6.337963824754399e-06, "loss": 0.5654, "step": 2159 }, { "epoch": 0.71, "grad_norm": 0.341422238757185, "learning_rate": 6.325014957480203e-06, "loss": 0.8692, "step": 2160 }, { "epoch": 0.71, "grad_norm": 0.2894908275046884, "learning_rate": 6.31207579654692e-06, "loss": 0.5785, "step": 2161 }, { "epoch": 0.71, "grad_norm": 0.2892519988213584, "learning_rate": 6.2991463564320296e-06, "loss": 0.5887, "step": 2162 }, { "epoch": 0.71, "grad_norm": 0.3122378258209536, "learning_rate": 6.286226651602134e-06, "loss": 0.6346, "step": 2163 }, { "epoch": 0.71, "grad_norm": 0.3147463494732522, "learning_rate": 6.273316696512941e-06, "loss": 0.7528, "step": 2164 }, { "epoch": 0.71, "grad_norm": 0.29737633953537324, "learning_rate": 6.260416505609247e-06, "loss": 0.5869, "step": 2165 }, { "epoch": 0.71, "grad_norm": 0.3457355332520332, "learning_rate": 6.2475260933249375e-06, "loss": 0.8555, "step": 2166 }, { "epoch": 0.71, "grad_norm": 0.32224447362464237, "learning_rate": 6.23464547408294e-06, "loss": 0.5966, "step": 2167 }, { "epoch": 0.71, "grad_norm": 0.32259282004837775, "learning_rate": 6.221774662295232e-06, "loss": 0.6277, "step": 2168 }, { "epoch": 0.71, "grad_norm": 0.28050647335414775, "learning_rate": 6.20891367236281e-06, "loss": 0.5763, "step": 2169 }, { "epoch": 0.71, "grad_norm": 0.34345114383177394, "learning_rate": 6.196062518675701e-06, "loss": 0.8393, "step": 2170 }, { "epoch": 0.71, "grad_norm": 0.29461150948826526, "learning_rate": 6.1832212156129045e-06, "loss": 0.5849, "step": 2171 }, { "epoch": 0.71, "grad_norm": 0.2809636620103479, "learning_rate": 6.170389777542409e-06, "loss": 0.5822, "step": 2172 }, { "epoch": 0.71, "grad_norm": 0.28941729451245307, "learning_rate": 6.157568218821162e-06, "loss": 0.5813, "step": 2173 }, { "epoch": 0.71, "grad_norm": 0.32217473476426395, "learning_rate": 6.144756553795056e-06, "loss": 0.6127, "step": 2174 }, { "epoch": 0.71, "grad_norm": 0.34495760125237834, "learning_rate": 6.131954796798915e-06, "loss": 0.8437, "step": 2175 }, { "epoch": 0.71, "grad_norm": 0.28480764191911573, "learning_rate": 6.119162962156483e-06, "loss": 0.5669, "step": 2176 }, { "epoch": 0.71, "grad_norm": 0.2819996362603305, "learning_rate": 6.106381064180396e-06, "loss": 0.5899, "step": 2177 }, { "epoch": 0.71, "grad_norm": 0.3448693482783814, "learning_rate": 6.093609117172163e-06, "loss": 0.8405, "step": 2178 }, { "epoch": 0.71, "grad_norm": 0.30038315836350726, "learning_rate": 6.080847135422182e-06, "loss": 0.5946, "step": 2179 }, { "epoch": 0.71, "grad_norm": 0.2989718601039672, "learning_rate": 6.068095133209681e-06, "loss": 0.5852, "step": 2180 }, { "epoch": 0.71, "grad_norm": 0.33349494490624815, "learning_rate": 6.055353124802728e-06, "loss": 0.8412, "step": 2181 }, { "epoch": 0.71, "grad_norm": 0.31598052685438527, "learning_rate": 6.0426211244582105e-06, "loss": 0.5938, "step": 2182 }, { "epoch": 0.71, "grad_norm": 0.2735829618862708, "learning_rate": 6.029899146421817e-06, "loss": 0.5685, "step": 2183 }, { "epoch": 0.71, "grad_norm": 0.3382239265159851, "learning_rate": 6.0171872049280215e-06, "loss": 0.8971, "step": 2184 }, { "epoch": 0.71, "grad_norm": 0.3012814893554373, "learning_rate": 6.004485314200064e-06, "loss": 0.3937, "step": 2185 }, { "epoch": 0.71, "grad_norm": 0.3305217549367917, "learning_rate": 5.991793488449956e-06, "loss": 0.8455, "step": 2186 }, { "epoch": 0.71, "grad_norm": 0.2962129690027876, "learning_rate": 5.979111741878428e-06, "loss": 0.5862, "step": 2187 }, { "epoch": 0.71, "grad_norm": 0.2995958405777306, "learning_rate": 5.966440088674937e-06, "loss": 0.5996, "step": 2188 }, { "epoch": 0.71, "grad_norm": 0.3330597439377167, "learning_rate": 5.953778543017657e-06, "loss": 0.8636, "step": 2189 }, { "epoch": 0.72, "grad_norm": 0.31013367262355734, "learning_rate": 5.941127119073443e-06, "loss": 0.5873, "step": 2190 }, { "epoch": 0.72, "grad_norm": 0.29091150108452385, "learning_rate": 5.928485830997828e-06, "loss": 0.6061, "step": 2191 }, { "epoch": 0.72, "grad_norm": 0.3330743389740455, "learning_rate": 5.915854692935002e-06, "loss": 0.8559, "step": 2192 }, { "epoch": 0.72, "grad_norm": 0.28598598843809486, "learning_rate": 5.9032337190178035e-06, "loss": 0.5825, "step": 2193 }, { "epoch": 0.72, "grad_norm": 0.3006139058879728, "learning_rate": 5.890622923367685e-06, "loss": 0.5952, "step": 2194 }, { "epoch": 0.72, "grad_norm": 0.3419351506994452, "learning_rate": 5.878022320094733e-06, "loss": 0.623, "step": 2195 }, { "epoch": 0.72, "grad_norm": 0.3311504329985719, "learning_rate": 5.865431923297613e-06, "loss": 0.8699, "step": 2196 }, { "epoch": 0.72, "grad_norm": 0.30024222041153203, "learning_rate": 5.85285174706357e-06, "loss": 0.5573, "step": 2197 }, { "epoch": 0.72, "grad_norm": 0.28763409603947987, "learning_rate": 5.840281805468427e-06, "loss": 0.5833, "step": 2198 }, { "epoch": 0.72, "grad_norm": 0.29049543131474775, "learning_rate": 5.827722112576544e-06, "loss": 0.591, "step": 2199 }, { "epoch": 0.72, "grad_norm": 0.3416121093708232, "learning_rate": 5.815172682440814e-06, "loss": 0.8467, "step": 2200 }, { "epoch": 0.72, "grad_norm": 0.295620622357414, "learning_rate": 5.802633529102654e-06, "loss": 0.5657, "step": 2201 }, { "epoch": 0.72, "grad_norm": 0.269122929540225, "learning_rate": 5.790104666591974e-06, "loss": 0.5736, "step": 2202 }, { "epoch": 0.72, "grad_norm": 0.3501533381112726, "learning_rate": 5.777586108927177e-06, "loss": 0.8484, "step": 2203 }, { "epoch": 0.72, "grad_norm": 0.30382110727125167, "learning_rate": 5.765077870115126e-06, "loss": 0.5638, "step": 2204 }, { "epoch": 0.72, "grad_norm": 0.28620726591620754, "learning_rate": 5.7525799641511585e-06, "loss": 0.5869, "step": 2205 }, { "epoch": 0.72, "grad_norm": 0.33042425194749664, "learning_rate": 5.740092405019029e-06, "loss": 0.6216, "step": 2206 }, { "epoch": 0.72, "grad_norm": 0.35455452463326453, "learning_rate": 5.727615206690921e-06, "loss": 0.8287, "step": 2207 }, { "epoch": 0.72, "grad_norm": 0.29061841880144246, "learning_rate": 5.715148383127437e-06, "loss": 0.5626, "step": 2208 }, { "epoch": 0.72, "grad_norm": 0.27720784184397057, "learning_rate": 5.702691948277557e-06, "loss": 0.5829, "step": 2209 }, { "epoch": 0.72, "grad_norm": 3.0765847052665185, "learning_rate": 5.690245916078646e-06, "loss": 0.8954, "step": 2210 }, { "epoch": 0.72, "grad_norm": 0.2986521916420762, "learning_rate": 5.677810300456423e-06, "loss": 0.6057, "step": 2211 }, { "epoch": 0.72, "grad_norm": 0.286113785411175, "learning_rate": 5.665385115324954e-06, "loss": 0.5755, "step": 2212 }, { "epoch": 0.72, "grad_norm": 0.3570389772092904, "learning_rate": 5.652970374586633e-06, "loss": 0.8606, "step": 2213 }, { "epoch": 0.72, "grad_norm": 0.2929935115211417, "learning_rate": 5.64056609213218e-06, "loss": 0.5858, "step": 2214 }, { "epoch": 0.72, "grad_norm": 0.28594890226772857, "learning_rate": 5.6281722818406e-06, "loss": 0.5723, "step": 2215 }, { "epoch": 0.72, "grad_norm": 0.34241228653393413, "learning_rate": 5.615788957579177e-06, "loss": 0.8614, "step": 2216 }, { "epoch": 0.72, "grad_norm": 0.3662984285736406, "learning_rate": 5.6034161332034814e-06, "loss": 0.3961, "step": 2217 }, { "epoch": 0.72, "grad_norm": 0.37058842964404426, "learning_rate": 5.59105382255732e-06, "loss": 0.8153, "step": 2218 }, { "epoch": 0.72, "grad_norm": 0.2828057950754528, "learning_rate": 5.578702039472739e-06, "loss": 0.556, "step": 2219 }, { "epoch": 0.73, "grad_norm": 0.2985472144748764, "learning_rate": 5.566360797770005e-06, "loss": 0.5633, "step": 2220 }, { "epoch": 0.73, "grad_norm": 0.332313220938094, "learning_rate": 5.554030111257594e-06, "loss": 0.8362, "step": 2221 }, { "epoch": 0.73, "grad_norm": 0.28685325658184657, "learning_rate": 5.541709993732168e-06, "loss": 0.5933, "step": 2222 }, { "epoch": 0.73, "grad_norm": 0.2856573680002102, "learning_rate": 5.5294004589785565e-06, "loss": 0.5751, "step": 2223 }, { "epoch": 0.73, "grad_norm": 0.3310457590704881, "learning_rate": 5.517101520769773e-06, "loss": 0.8369, "step": 2224 }, { "epoch": 0.73, "grad_norm": 0.28464735296980176, "learning_rate": 5.504813192866946e-06, "loss": 0.5682, "step": 2225 }, { "epoch": 0.73, "grad_norm": 0.2726420600822048, "learning_rate": 5.492535489019344e-06, "loss": 0.5774, "step": 2226 }, { "epoch": 0.73, "grad_norm": 0.33823781787360807, "learning_rate": 5.480268422964357e-06, "loss": 0.8602, "step": 2227 }, { "epoch": 0.73, "grad_norm": 0.2679948542049774, "learning_rate": 5.468012008427458e-06, "loss": 0.3993, "step": 2228 }, { "epoch": 0.73, "grad_norm": 0.3403382759314743, "learning_rate": 5.455766259122209e-06, "loss": 0.8409, "step": 2229 }, { "epoch": 0.73, "grad_norm": 0.28854221440051736, "learning_rate": 5.443531188750238e-06, "loss": 0.5777, "step": 2230 }, { "epoch": 0.73, "grad_norm": 0.2946960544135847, "learning_rate": 5.431306811001225e-06, "loss": 0.5758, "step": 2231 }, { "epoch": 0.73, "grad_norm": 0.36083927920173564, "learning_rate": 5.419093139552878e-06, "loss": 0.8793, "step": 2232 }, { "epoch": 0.73, "grad_norm": 0.2747478340999522, "learning_rate": 5.406890188070948e-06, "loss": 0.5556, "step": 2233 }, { "epoch": 0.73, "grad_norm": 0.28280116610990147, "learning_rate": 5.3946979702091666e-06, "loss": 0.5756, "step": 2234 }, { "epoch": 0.73, "grad_norm": 0.34854547247397377, "learning_rate": 5.3825164996092695e-06, "loss": 0.8686, "step": 2235 }, { "epoch": 0.73, "grad_norm": 0.28833752592034106, "learning_rate": 5.370345789900958e-06, "loss": 0.5869, "step": 2236 }, { "epoch": 0.73, "grad_norm": 0.2800421307851358, "learning_rate": 5.3581858547019095e-06, "loss": 0.6, "step": 2237 }, { "epoch": 0.73, "grad_norm": 0.3433444245300005, "learning_rate": 5.346036707617731e-06, "loss": 0.8036, "step": 2238 }, { "epoch": 0.73, "grad_norm": 0.3416688410858311, "learning_rate": 5.333898362241962e-06, "loss": 0.6391, "step": 2239 }, { "epoch": 0.73, "grad_norm": 0.28810775236847025, "learning_rate": 5.321770832156057e-06, "loss": 0.567, "step": 2240 }, { "epoch": 0.73, "grad_norm": 0.2859219856729592, "learning_rate": 5.309654130929373e-06, "loss": 0.5983, "step": 2241 }, { "epoch": 0.73, "grad_norm": 0.38269831978429497, "learning_rate": 5.297548272119138e-06, "loss": 0.8523, "step": 2242 }, { "epoch": 0.73, "grad_norm": 0.28285779458349414, "learning_rate": 5.2854532692704714e-06, "loss": 0.5723, "step": 2243 }, { "epoch": 0.73, "grad_norm": 0.2883191940862153, "learning_rate": 5.273369135916328e-06, "loss": 0.5803, "step": 2244 }, { "epoch": 0.73, "grad_norm": 0.33385500425483255, "learning_rate": 5.261295885577498e-06, "loss": 0.6174, "step": 2245 }, { "epoch": 0.73, "grad_norm": 0.2982014003210629, "learning_rate": 5.249233531762613e-06, "loss": 0.5819, "step": 2246 }, { "epoch": 0.73, "grad_norm": 0.32984337145617704, "learning_rate": 5.237182087968098e-06, "loss": 0.75, "step": 2247 }, { "epoch": 0.73, "grad_norm": 0.2898127856990067, "learning_rate": 5.2251415676781726e-06, "loss": 0.5784, "step": 2248 }, { "epoch": 0.73, "grad_norm": 0.3524525412177361, "learning_rate": 5.21311198436484e-06, "loss": 0.8614, "step": 2249 }, { "epoch": 0.73, "grad_norm": 0.3394810312565484, "learning_rate": 5.20109335148786e-06, "loss": 0.6291, "step": 2250 }, { "epoch": 0.74, "grad_norm": 0.2856685722436317, "learning_rate": 5.189085682494737e-06, "loss": 0.5723, "step": 2251 }, { "epoch": 0.74, "grad_norm": 0.27545512478885414, "learning_rate": 5.177088990820725e-06, "loss": 0.5797, "step": 2252 }, { "epoch": 0.74, "grad_norm": 0.36482739789315866, "learning_rate": 5.1651032898887776e-06, "loss": 0.8764, "step": 2253 }, { "epoch": 0.74, "grad_norm": 0.29452976965498157, "learning_rate": 5.153128593109561e-06, "loss": 0.5924, "step": 2254 }, { "epoch": 0.74, "grad_norm": 0.3007319482012876, "learning_rate": 5.1411649138814175e-06, "loss": 0.5882, "step": 2255 }, { "epoch": 0.74, "grad_norm": 0.31582938256178417, "learning_rate": 5.129212265590381e-06, "loss": 0.6095, "step": 2256 }, { "epoch": 0.74, "grad_norm": 0.2676382893202364, "learning_rate": 5.1172706616101275e-06, "loss": 0.5868, "step": 2257 }, { "epoch": 0.74, "grad_norm": 0.33048533953031745, "learning_rate": 5.105340115301978e-06, "loss": 0.8117, "step": 2258 }, { "epoch": 0.74, "grad_norm": 0.3015987400718788, "learning_rate": 5.093420640014887e-06, "loss": 0.585, "step": 2259 }, { "epoch": 0.74, "grad_norm": 0.28658786739186964, "learning_rate": 5.081512249085416e-06, "loss": 0.572, "step": 2260 }, { "epoch": 0.74, "grad_norm": 0.36051114639419635, "learning_rate": 5.069614955837722e-06, "loss": 0.8404, "step": 2261 }, { "epoch": 0.74, "grad_norm": 0.29496320155520006, "learning_rate": 5.05772877358356e-06, "loss": 0.6227, "step": 2262 }, { "epoch": 0.74, "grad_norm": 0.28328452668589776, "learning_rate": 5.0458537156222365e-06, "loss": 0.5814, "step": 2263 }, { "epoch": 0.74, "grad_norm": 0.3633775279723979, "learning_rate": 5.033989795240612e-06, "loss": 0.8576, "step": 2264 }, { "epoch": 0.74, "grad_norm": 0.2908153632720878, "learning_rate": 5.022137025713102e-06, "loss": 0.5702, "step": 2265 }, { "epoch": 0.74, "grad_norm": 0.2801901466848527, "learning_rate": 5.010295420301629e-06, "loss": 0.5779, "step": 2266 }, { "epoch": 0.74, "grad_norm": 0.3419532882525435, "learning_rate": 4.9984649922556275e-06, "loss": 0.6239, "step": 2267 }, { "epoch": 0.74, "grad_norm": 0.3568822941272129, "learning_rate": 4.986645754812027e-06, "loss": 0.8507, "step": 2268 }, { "epoch": 0.74, "grad_norm": 0.2718864012518564, "learning_rate": 4.974837721195238e-06, "loss": 0.5731, "step": 2269 }, { "epoch": 0.74, "grad_norm": 0.27763544764237397, "learning_rate": 4.963040904617131e-06, "loss": 0.5746, "step": 2270 }, { "epoch": 0.74, "grad_norm": 0.2836174680911735, "learning_rate": 4.951255318277027e-06, "loss": 0.583, "step": 2271 }, { "epoch": 0.74, "grad_norm": 0.3366547510863037, "learning_rate": 4.939480975361687e-06, "loss": 0.824, "step": 2272 }, { "epoch": 0.74, "grad_norm": 0.27889742164407394, "learning_rate": 4.92771788904529e-06, "loss": 0.571, "step": 2273 }, { "epoch": 0.74, "grad_norm": 0.27832915184261675, "learning_rate": 4.915966072489407e-06, "loss": 0.5781, "step": 2274 }, { "epoch": 0.74, "grad_norm": 0.34161368538731585, "learning_rate": 4.9042255388430245e-06, "loss": 0.8507, "step": 2275 }, { "epoch": 0.74, "grad_norm": 0.28116441211603993, "learning_rate": 4.892496301242482e-06, "loss": 0.5786, "step": 2276 }, { "epoch": 0.74, "grad_norm": 0.27210930461417165, "learning_rate": 4.88077837281149e-06, "loss": 0.5772, "step": 2277 }, { "epoch": 0.74, "grad_norm": 0.3380538232268532, "learning_rate": 4.869071766661106e-06, "loss": 0.6399, "step": 2278 }, { "epoch": 0.74, "grad_norm": 0.3241450558714518, "learning_rate": 4.857376495889713e-06, "loss": 0.8257, "step": 2279 }, { "epoch": 0.74, "grad_norm": 0.29688134480492034, "learning_rate": 4.845692573583011e-06, "loss": 0.6009, "step": 2280 }, { "epoch": 0.74, "grad_norm": 0.28365204854473286, "learning_rate": 4.834020012814016e-06, "loss": 0.5993, "step": 2281 }, { "epoch": 0.75, "grad_norm": 0.33551252583666535, "learning_rate": 4.822358826643019e-06, "loss": 0.8457, "step": 2282 }, { "epoch": 0.75, "grad_norm": 0.8899802196447385, "learning_rate": 4.81070902811758e-06, "loss": 0.5985, "step": 2283 }, { "epoch": 0.75, "grad_norm": 0.2784912301983082, "learning_rate": 4.799070630272535e-06, "loss": 0.5738, "step": 2284 }, { "epoch": 0.75, "grad_norm": 0.33627031178951156, "learning_rate": 4.787443646129949e-06, "loss": 0.8001, "step": 2285 }, { "epoch": 0.75, "grad_norm": 0.28015386442159856, "learning_rate": 4.775828088699119e-06, "loss": 0.5937, "step": 2286 }, { "epoch": 0.75, "grad_norm": 0.27270574254249275, "learning_rate": 4.76422397097656e-06, "loss": 0.5886, "step": 2287 }, { "epoch": 0.75, "grad_norm": 0.27183807651596276, "learning_rate": 4.7526313059459854e-06, "loss": 0.5733, "step": 2288 }, { "epoch": 0.75, "grad_norm": 0.32779794848555943, "learning_rate": 4.741050106578296e-06, "loss": 0.6374, "step": 2289 }, { "epoch": 0.75, "grad_norm": 0.3457320184080609, "learning_rate": 4.729480385831555e-06, "loss": 0.8782, "step": 2290 }, { "epoch": 0.75, "grad_norm": 0.27853034453019326, "learning_rate": 4.717922156651002e-06, "loss": 0.5638, "step": 2291 }, { "epoch": 0.75, "grad_norm": 0.3191664291794741, "learning_rate": 4.706375431968998e-06, "loss": 0.5896, "step": 2292 }, { "epoch": 0.75, "grad_norm": 0.33002104352966083, "learning_rate": 4.6948402247050395e-06, "loss": 0.8571, "step": 2293 }, { "epoch": 0.75, "grad_norm": 0.2773504922260191, "learning_rate": 4.6833165477657454e-06, "loss": 0.5645, "step": 2294 }, { "epoch": 0.75, "grad_norm": 0.33638385983636243, "learning_rate": 4.67180441404482e-06, "loss": 0.5851, "step": 2295 }, { "epoch": 0.75, "grad_norm": 0.33202073514171687, "learning_rate": 4.660303836423057e-06, "loss": 0.8426, "step": 2296 }, { "epoch": 0.75, "grad_norm": 0.28358302151847137, "learning_rate": 4.648814827768323e-06, "loss": 0.5857, "step": 2297 }, { "epoch": 0.75, "grad_norm": 0.2827074191931312, "learning_rate": 4.637337400935533e-06, "loss": 0.5718, "step": 2298 }, { "epoch": 0.75, "grad_norm": 0.3367383181019836, "learning_rate": 4.625871568766647e-06, "loss": 0.8281, "step": 2299 }, { "epoch": 0.75, "grad_norm": 0.2621702540841223, "learning_rate": 4.614417344090659e-06, "loss": 0.404, "step": 2300 }, { "epoch": 0.75, "grad_norm": 0.3398781631626797, "learning_rate": 4.6029747397235675e-06, "loss": 0.8964, "step": 2301 }, { "epoch": 0.75, "grad_norm": 0.28197880494910255, "learning_rate": 4.591543768468364e-06, "loss": 0.5824, "step": 2302 }, { "epoch": 0.75, "grad_norm": 0.26922929297039916, "learning_rate": 4.58012444311504e-06, "loss": 0.5587, "step": 2303 }, { "epoch": 0.75, "grad_norm": 0.34197613783707376, "learning_rate": 4.568716776440543e-06, "loss": 0.8153, "step": 2304 }, { "epoch": 0.75, "grad_norm": 0.28163906078843576, "learning_rate": 4.55732078120878e-06, "loss": 0.5632, "step": 2305 }, { "epoch": 0.75, "grad_norm": 0.28468309474195497, "learning_rate": 4.5459364701706e-06, "loss": 0.5732, "step": 2306 }, { "epoch": 0.75, "grad_norm": 0.3377922940251202, "learning_rate": 4.534563856063777e-06, "loss": 0.8263, "step": 2307 }, { "epoch": 0.75, "grad_norm": 0.29287165338113336, "learning_rate": 4.523202951612996e-06, "loss": 0.6105, "step": 2308 }, { "epoch": 0.75, "grad_norm": 0.2786273005654294, "learning_rate": 4.51185376952984e-06, "loss": 0.5786, "step": 2309 }, { "epoch": 0.75, "grad_norm": 0.33806316517451945, "learning_rate": 4.500516322512785e-06, "loss": 0.8402, "step": 2310 }, { "epoch": 0.75, "grad_norm": 0.3234811705585942, "learning_rate": 4.489190623247169e-06, "loss": 0.6263, "step": 2311 }, { "epoch": 0.76, "grad_norm": 0.27951244823072147, "learning_rate": 4.4778766844051795e-06, "loss": 0.5936, "step": 2312 }, { "epoch": 0.76, "grad_norm": 0.2720265286455213, "learning_rate": 4.466574518645862e-06, "loss": 0.5721, "step": 2313 }, { "epoch": 0.76, "grad_norm": 0.32733924274488413, "learning_rate": 4.455284138615074e-06, "loss": 0.8107, "step": 2314 }, { "epoch": 0.76, "grad_norm": 0.27474909163541844, "learning_rate": 4.444005556945494e-06, "loss": 0.5806, "step": 2315 }, { "epoch": 0.76, "grad_norm": 0.28223561574182493, "learning_rate": 4.432738786256596e-06, "loss": 0.567, "step": 2316 }, { "epoch": 0.76, "grad_norm": 0.280219005399105, "learning_rate": 4.4214838391546405e-06, "loss": 0.5924, "step": 2317 }, { "epoch": 0.76, "grad_norm": 0.33212718626174625, "learning_rate": 4.410240728232654e-06, "loss": 0.8459, "step": 2318 }, { "epoch": 0.76, "grad_norm": 0.28824629742386915, "learning_rate": 4.399009466070431e-06, "loss": 0.5623, "step": 2319 }, { "epoch": 0.76, "grad_norm": 0.28206270797560845, "learning_rate": 4.387790065234498e-06, "loss": 0.5623, "step": 2320 }, { "epoch": 0.76, "grad_norm": 0.32920698856534597, "learning_rate": 4.376582538278115e-06, "loss": 0.8133, "step": 2321 }, { "epoch": 0.76, "grad_norm": 0.3230249216735367, "learning_rate": 4.365386897741249e-06, "loss": 0.6279, "step": 2322 }, { "epoch": 0.76, "grad_norm": 0.28766341979220267, "learning_rate": 4.354203156150581e-06, "loss": 0.5754, "step": 2323 }, { "epoch": 0.76, "grad_norm": 0.278876066474603, "learning_rate": 4.34303132601947e-06, "loss": 0.5783, "step": 2324 }, { "epoch": 0.76, "grad_norm": 0.3707652126274623, "learning_rate": 4.331871419847943e-06, "loss": 0.8197, "step": 2325 }, { "epoch": 0.76, "grad_norm": 0.2856946782593199, "learning_rate": 4.320723450122695e-06, "loss": 0.5768, "step": 2326 }, { "epoch": 0.76, "grad_norm": 0.3343817578167749, "learning_rate": 4.309587429317061e-06, "loss": 0.6326, "step": 2327 }, { "epoch": 0.76, "grad_norm": 0.2713772139592782, "learning_rate": 4.298463369890999e-06, "loss": 0.5706, "step": 2328 }, { "epoch": 0.76, "grad_norm": 0.28006336601277976, "learning_rate": 4.287351284291102e-06, "loss": 0.5828, "step": 2329 }, { "epoch": 0.76, "grad_norm": 0.2843554756145523, "learning_rate": 4.276251184950548e-06, "loss": 0.7312, "step": 2330 }, { "epoch": 0.76, "grad_norm": 0.3459860955726458, "learning_rate": 4.26516308428911e-06, "loss": 0.8376, "step": 2331 }, { "epoch": 0.76, "grad_norm": 0.2863700313360134, "learning_rate": 4.254086994713141e-06, "loss": 0.5976, "step": 2332 }, { "epoch": 0.76, "grad_norm": 0.32069997487784846, "learning_rate": 4.243022928615549e-06, "loss": 0.6296, "step": 2333 }, { "epoch": 0.76, "grad_norm": 0.27803019694352044, "learning_rate": 4.2319708983757866e-06, "loss": 0.5855, "step": 2334 }, { "epoch": 0.76, "grad_norm": 0.28711378835408197, "learning_rate": 4.220930916359846e-06, "loss": 0.5987, "step": 2335 }, { "epoch": 0.76, "grad_norm": 0.3387930999489222, "learning_rate": 4.209902994920236e-06, "loss": 0.8355, "step": 2336 }, { "epoch": 0.76, "grad_norm": 0.2678571156391979, "learning_rate": 4.1988871463959704e-06, "loss": 0.5759, "step": 2337 }, { "epoch": 0.76, "grad_norm": 0.33463994732534236, "learning_rate": 4.18788338311255e-06, "loss": 0.6124, "step": 2338 }, { "epoch": 0.76, "grad_norm": 0.2787860267096995, "learning_rate": 4.176891717381968e-06, "loss": 0.5966, "step": 2339 }, { "epoch": 0.76, "grad_norm": 0.34649466842010135, "learning_rate": 4.16591216150267e-06, "loss": 0.8612, "step": 2340 }, { "epoch": 0.76, "grad_norm": 0.28711029930657656, "learning_rate": 4.154944727759548e-06, "loss": 0.5772, "step": 2341 }, { "epoch": 0.76, "grad_norm": 0.2727800165543625, "learning_rate": 4.1439894284239474e-06, "loss": 0.5638, "step": 2342 }, { "epoch": 0.77, "grad_norm": 0.2649075970201966, "learning_rate": 4.133046275753622e-06, "loss": 0.5707, "step": 2343 }, { "epoch": 0.77, "grad_norm": 0.3516156297304568, "learning_rate": 4.122115281992739e-06, "loss": 0.851, "step": 2344 }, { "epoch": 0.77, "grad_norm": 0.27214666988742503, "learning_rate": 4.111196459371862e-06, "loss": 0.584, "step": 2345 }, { "epoch": 0.77, "grad_norm": 0.26683910813338213, "learning_rate": 4.100289820107936e-06, "loss": 0.5679, "step": 2346 }, { "epoch": 0.77, "grad_norm": 0.33961417746490635, "learning_rate": 4.089395376404269e-06, "loss": 0.856, "step": 2347 }, { "epoch": 0.77, "grad_norm": 0.27433065138634866, "learning_rate": 4.078513140450538e-06, "loss": 0.573, "step": 2348 }, { "epoch": 0.77, "grad_norm": 0.3214845758950185, "learning_rate": 4.067643124422748e-06, "loss": 0.6122, "step": 2349 }, { "epoch": 0.77, "grad_norm": 0.27362192895937804, "learning_rate": 4.056785340483229e-06, "loss": 0.5954, "step": 2350 }, { "epoch": 0.77, "grad_norm": 0.33595863410691185, "learning_rate": 4.04593980078064e-06, "loss": 0.8603, "step": 2351 }, { "epoch": 0.77, "grad_norm": 0.2746582363064567, "learning_rate": 4.035106517449926e-06, "loss": 0.5831, "step": 2352 }, { "epoch": 0.77, "grad_norm": 0.26889674572978595, "learning_rate": 4.024285502612322e-06, "loss": 0.5735, "step": 2353 }, { "epoch": 0.77, "grad_norm": 0.34973254502447565, "learning_rate": 4.013476768375339e-06, "loss": 0.8406, "step": 2354 }, { "epoch": 0.77, "grad_norm": 0.29490916717152377, "learning_rate": 4.002680326832743e-06, "loss": 0.5857, "step": 2355 }, { "epoch": 0.77, "grad_norm": 0.26637934226491305, "learning_rate": 3.9918961900645505e-06, "loss": 0.5587, "step": 2356 }, { "epoch": 0.77, "grad_norm": 0.34334142211862034, "learning_rate": 3.981124370137002e-06, "loss": 0.8268, "step": 2357 }, { "epoch": 0.77, "grad_norm": 0.27436513787786726, "learning_rate": 3.970364879102572e-06, "loss": 0.5928, "step": 2358 }, { "epoch": 0.77, "grad_norm": 0.2758101258048933, "learning_rate": 3.959617728999928e-06, "loss": 0.5638, "step": 2359 }, { "epoch": 0.77, "grad_norm": 0.3223031223519255, "learning_rate": 3.948882931853924e-06, "loss": 0.6335, "step": 2360 }, { "epoch": 0.77, "grad_norm": 0.2655973495628547, "learning_rate": 3.938160499675616e-06, "loss": 0.5838, "step": 2361 }, { "epoch": 0.77, "grad_norm": 0.32954290489960847, "learning_rate": 3.9274504444622025e-06, "loss": 0.841, "step": 2362 }, { "epoch": 0.77, "grad_norm": 0.274915787941311, "learning_rate": 3.916752778197039e-06, "loss": 0.5768, "step": 2363 }, { "epoch": 0.77, "grad_norm": 0.27859838138856524, "learning_rate": 3.906067512849624e-06, "loss": 0.5759, "step": 2364 }, { "epoch": 0.77, "grad_norm": 0.3287701346320898, "learning_rate": 3.895394660375576e-06, "loss": 0.83, "step": 2365 }, { "epoch": 0.77, "grad_norm": 0.2895124521198522, "learning_rate": 3.8847342327166245e-06, "loss": 0.5824, "step": 2366 }, { "epoch": 0.77, "grad_norm": 0.2674558916673328, "learning_rate": 3.874086241800607e-06, "loss": 0.5603, "step": 2367 }, { "epoch": 0.77, "grad_norm": 0.37471704381979104, "learning_rate": 3.863450699541434e-06, "loss": 0.8805, "step": 2368 }, { "epoch": 0.77, "grad_norm": 0.2689271176402253, "learning_rate": 3.852827617839085e-06, "loss": 0.589, "step": 2369 }, { "epoch": 0.77, "grad_norm": 0.2784385141465303, "learning_rate": 3.842217008579615e-06, "loss": 0.5799, "step": 2370 }, { "epoch": 0.77, "grad_norm": 0.32072541500616175, "learning_rate": 3.831618883635106e-06, "loss": 0.6131, "step": 2371 }, { "epoch": 0.77, "grad_norm": 0.2695056318108863, "learning_rate": 3.82103325486368e-06, "loss": 0.5581, "step": 2372 }, { "epoch": 0.77, "grad_norm": 0.3282898384053917, "learning_rate": 3.8104601341094746e-06, "loss": 0.8583, "step": 2373 }, { "epoch": 0.78, "grad_norm": 0.2695714299788949, "learning_rate": 3.7998995332026333e-06, "loss": 0.5792, "step": 2374 }, { "epoch": 0.78, "grad_norm": 0.29133839780028437, "learning_rate": 3.78935146395929e-06, "loss": 0.5762, "step": 2375 }, { "epoch": 0.78, "grad_norm": 0.3370841271110733, "learning_rate": 3.7788159381815536e-06, "loss": 0.8288, "step": 2376 }, { "epoch": 0.78, "grad_norm": 0.29039782092845773, "learning_rate": 3.768292967657512e-06, "loss": 0.5879, "step": 2377 }, { "epoch": 0.78, "grad_norm": 0.28214954133213854, "learning_rate": 3.7577825641611918e-06, "loss": 0.5793, "step": 2378 }, { "epoch": 0.78, "grad_norm": 0.33628529119919104, "learning_rate": 3.7472847394525565e-06, "loss": 0.8587, "step": 2379 }, { "epoch": 0.78, "grad_norm": 0.30458836907682435, "learning_rate": 3.7367995052775123e-06, "loss": 0.5908, "step": 2380 }, { "epoch": 0.78, "grad_norm": 0.2715656612228924, "learning_rate": 3.7263268733678607e-06, "loss": 0.5867, "step": 2381 }, { "epoch": 0.78, "grad_norm": 0.3202799946027206, "learning_rate": 3.715866855441309e-06, "loss": 0.6422, "step": 2382 }, { "epoch": 0.78, "grad_norm": 0.33523736652254515, "learning_rate": 3.7054194632014517e-06, "loss": 0.8346, "step": 2383 }, { "epoch": 0.78, "grad_norm": 0.28064340793793596, "learning_rate": 3.6949847083377563e-06, "loss": 0.588, "step": 2384 }, { "epoch": 0.78, "grad_norm": 0.2605950890613309, "learning_rate": 3.6845626025255457e-06, "loss": 0.5722, "step": 2385 }, { "epoch": 0.78, "grad_norm": 0.26342931926997626, "learning_rate": 3.674153157426e-06, "loss": 0.5756, "step": 2386 }, { "epoch": 0.78, "grad_norm": 0.3337145013459757, "learning_rate": 3.6637563846861278e-06, "loss": 0.8532, "step": 2387 }, { "epoch": 0.78, "grad_norm": 0.27680742781388495, "learning_rate": 3.6533722959387535e-06, "loss": 0.5927, "step": 2388 }, { "epoch": 0.78, "grad_norm": 0.26506879592833726, "learning_rate": 3.643000902802513e-06, "loss": 0.571, "step": 2389 }, { "epoch": 0.78, "grad_norm": 0.3497980739060594, "learning_rate": 3.6326422168818473e-06, "loss": 0.8285, "step": 2390 }, { "epoch": 0.78, "grad_norm": 0.26867283288647326, "learning_rate": 3.622296249766967e-06, "loss": 0.5902, "step": 2391 }, { "epoch": 0.78, "grad_norm": 0.27281889510200613, "learning_rate": 3.6119630130338537e-06, "loss": 0.5847, "step": 2392 }, { "epoch": 0.78, "grad_norm": 0.3286792356897914, "learning_rate": 3.6016425182442477e-06, "loss": 0.6267, "step": 2393 }, { "epoch": 0.78, "grad_norm": 0.3539981637560091, "learning_rate": 3.591334776945633e-06, "loss": 0.8405, "step": 2394 }, { "epoch": 0.78, "grad_norm": 0.28812683150491186, "learning_rate": 3.5810398006712156e-06, "loss": 0.5964, "step": 2395 }, { "epoch": 0.78, "grad_norm": 0.31573598106604156, "learning_rate": 3.570757600939939e-06, "loss": 0.5945, "step": 2396 }, { "epoch": 0.78, "grad_norm": 0.3293769927207748, "learning_rate": 3.5604881892564294e-06, "loss": 0.8138, "step": 2397 }, { "epoch": 0.78, "grad_norm": 0.2822336428714121, "learning_rate": 3.5502315771110126e-06, "loss": 0.5731, "step": 2398 }, { "epoch": 0.78, "grad_norm": 0.2700515951074531, "learning_rate": 3.5399877759797005e-06, "loss": 0.5673, "step": 2399 }, { "epoch": 0.78, "grad_norm": 0.32643400241923665, "learning_rate": 3.5297567973241595e-06, "loss": 0.8482, "step": 2400 }, { "epoch": 0.78, "grad_norm": 0.26440433260154766, "learning_rate": 3.5195386525917143e-06, "loss": 0.5681, "step": 2401 }, { "epoch": 0.78, "grad_norm": 0.29062856457350716, "learning_rate": 3.5093333532153316e-06, "loss": 0.5921, "step": 2402 }, { "epoch": 0.78, "grad_norm": 0.3381626538143637, "learning_rate": 3.4991409106136e-06, "loss": 0.8642, "step": 2403 }, { "epoch": 0.79, "grad_norm": 0.2670910163666645, "learning_rate": 3.4889613361907246e-06, "loss": 0.389, "step": 2404 }, { "epoch": 0.79, "grad_norm": 0.34737503611288206, "learning_rate": 3.4787946413365207e-06, "loss": 0.8596, "step": 2405 }, { "epoch": 0.79, "grad_norm": 0.26982917281961644, "learning_rate": 3.468640837426384e-06, "loss": 0.5729, "step": 2406 }, { "epoch": 0.79, "grad_norm": 0.2782438474365323, "learning_rate": 3.4584999358212864e-06, "loss": 0.6024, "step": 2407 }, { "epoch": 0.79, "grad_norm": 0.33201302402776106, "learning_rate": 3.448371947867764e-06, "loss": 0.8336, "step": 2408 }, { "epoch": 0.79, "grad_norm": 0.3257119818921927, "learning_rate": 3.438256884897913e-06, "loss": 0.6356, "step": 2409 }, { "epoch": 0.79, "grad_norm": 0.2732729096810846, "learning_rate": 3.428154758229357e-06, "loss": 0.5736, "step": 2410 }, { "epoch": 0.79, "grad_norm": 0.2833096266602842, "learning_rate": 3.418065579165248e-06, "loss": 0.593, "step": 2411 }, { "epoch": 0.79, "grad_norm": 0.33200550081988595, "learning_rate": 3.4079893589942544e-06, "loss": 0.86, "step": 2412 }, { "epoch": 0.79, "grad_norm": 0.27616245136410045, "learning_rate": 3.3979261089905427e-06, "loss": 0.5631, "step": 2413 }, { "epoch": 0.79, "grad_norm": 0.30044241949174133, "learning_rate": 3.3878758404137627e-06, "loss": 0.7415, "step": 2414 }, { "epoch": 0.79, "grad_norm": 0.26161158955795405, "learning_rate": 3.3778385645090526e-06, "loss": 0.395, "step": 2415 }, { "epoch": 0.79, "grad_norm": 0.3373136623215282, "learning_rate": 3.3678142925070004e-06, "loss": 0.8495, "step": 2416 }, { "epoch": 0.79, "grad_norm": 0.2775429921292862, "learning_rate": 3.357803035623646e-06, "loss": 0.5687, "step": 2417 }, { "epoch": 0.79, "grad_norm": 0.2750356205881716, "learning_rate": 3.347804805060476e-06, "loss": 0.5691, "step": 2418 }, { "epoch": 0.79, "grad_norm": 0.33108663686892503, "learning_rate": 3.3378196120043915e-06, "loss": 0.8506, "step": 2419 }, { "epoch": 0.79, "grad_norm": 0.319386953141291, "learning_rate": 3.327847467627712e-06, "loss": 0.6235, "step": 2420 }, { "epoch": 0.79, "grad_norm": 0.26573548118284374, "learning_rate": 3.3178883830881528e-06, "loss": 0.5638, "step": 2421 }, { "epoch": 0.79, "grad_norm": 0.2661093585645269, "learning_rate": 3.3079423695288204e-06, "loss": 0.5706, "step": 2422 }, { "epoch": 0.79, "grad_norm": 0.33532029795881785, "learning_rate": 3.298009438078194e-06, "loss": 0.8502, "step": 2423 }, { "epoch": 0.79, "grad_norm": 0.2906376107722037, "learning_rate": 3.288089599850112e-06, "loss": 0.579, "step": 2424 }, { "epoch": 0.79, "grad_norm": 0.27185064284304333, "learning_rate": 3.278182865943776e-06, "loss": 0.6019, "step": 2425 }, { "epoch": 0.79, "grad_norm": 0.3401144340952734, "learning_rate": 3.2682892474437133e-06, "loss": 0.8409, "step": 2426 }, { "epoch": 0.79, "grad_norm": 0.28263118604090137, "learning_rate": 3.2584087554197744e-06, "loss": 0.5781, "step": 2427 }, { "epoch": 0.79, "grad_norm": 0.28470130536581256, "learning_rate": 3.2485414009271353e-06, "loss": 0.5689, "step": 2428 }, { "epoch": 0.79, "grad_norm": 0.3283148282740768, "learning_rate": 3.238687195006264e-06, "loss": 0.8351, "step": 2429 }, { "epoch": 0.79, "grad_norm": 0.27930227163710186, "learning_rate": 3.228846148682916e-06, "loss": 0.5755, "step": 2430 }, { "epoch": 0.79, "grad_norm": 0.3361203193079879, "learning_rate": 3.219018272968127e-06, "loss": 0.6813, "step": 2431 }, { "epoch": 0.79, "grad_norm": 0.27261688377074345, "learning_rate": 3.2092035788581907e-06, "loss": 0.5978, "step": 2432 }, { "epoch": 0.79, "grad_norm": 0.27423142191541333, "learning_rate": 3.1994020773346543e-06, "loss": 0.5687, "step": 2433 }, { "epoch": 0.79, "grad_norm": 0.32918227939402633, "learning_rate": 3.1896137793643106e-06, "loss": 0.808, "step": 2434 }, { "epoch": 0.8, "grad_norm": 0.2792876844071531, "learning_rate": 3.1798386958991715e-06, "loss": 0.5718, "step": 2435 }, { "epoch": 0.8, "grad_norm": 0.27301672732423254, "learning_rate": 3.1700768378764607e-06, "loss": 0.5667, "step": 2436 }, { "epoch": 0.8, "grad_norm": 0.34230333096401694, "learning_rate": 3.160328216218617e-06, "loss": 0.8681, "step": 2437 }, { "epoch": 0.8, "grad_norm": 0.2698303697394687, "learning_rate": 3.1505928418332574e-06, "loss": 0.5741, "step": 2438 }, { "epoch": 0.8, "grad_norm": 0.4241613075454975, "learning_rate": 3.1408707256131784e-06, "loss": 0.578, "step": 2439 }, { "epoch": 0.8, "grad_norm": 0.33786645937848464, "learning_rate": 3.1311618784363454e-06, "loss": 0.8472, "step": 2440 }, { "epoch": 0.8, "grad_norm": 0.27193316719699795, "learning_rate": 3.1214663111658753e-06, "loss": 0.5895, "step": 2441 }, { "epoch": 0.8, "grad_norm": 0.3130488277096351, "learning_rate": 3.1117840346500287e-06, "loss": 0.6138, "step": 2442 }, { "epoch": 0.8, "grad_norm": 0.272305672007606, "learning_rate": 3.102115059722185e-06, "loss": 0.6045, "step": 2443 }, { "epoch": 0.8, "grad_norm": 0.2834522060401969, "learning_rate": 3.092459397200861e-06, "loss": 0.5678, "step": 2444 }, { "epoch": 0.8, "grad_norm": 0.34177365047654895, "learning_rate": 3.0828170578896598e-06, "loss": 0.8584, "step": 2445 }, { "epoch": 0.8, "grad_norm": 0.27459071378791294, "learning_rate": 3.0731880525772817e-06, "loss": 0.5794, "step": 2446 }, { "epoch": 0.8, "grad_norm": 0.27827283488344456, "learning_rate": 3.0635723920375168e-06, "loss": 0.565, "step": 2447 }, { "epoch": 0.8, "grad_norm": 0.3451104287735247, "learning_rate": 3.053970087029215e-06, "loss": 0.8721, "step": 2448 }, { "epoch": 0.8, "grad_norm": 0.26905586250699326, "learning_rate": 3.044381148296281e-06, "loss": 0.5781, "step": 2449 }, { "epoch": 0.8, "grad_norm": 0.2742091938087031, "learning_rate": 3.034805586567671e-06, "loss": 0.5846, "step": 2450 }, { "epoch": 0.8, "grad_norm": 0.33066251396749197, "learning_rate": 3.0252434125573703e-06, "loss": 0.8223, "step": 2451 }, { "epoch": 0.8, "grad_norm": 0.28151542131646157, "learning_rate": 3.0156946369643803e-06, "loss": 0.5957, "step": 2452 }, { "epoch": 0.8, "grad_norm": 0.3176563409483738, "learning_rate": 3.0061592704727257e-06, "loss": 0.6584, "step": 2453 }, { "epoch": 0.8, "grad_norm": 0.2793541418796504, "learning_rate": 2.9966373237514133e-06, "loss": 0.5854, "step": 2454 }, { "epoch": 0.8, "grad_norm": 0.33327183272115135, "learning_rate": 2.987128807454435e-06, "loss": 0.8411, "step": 2455 }, { "epoch": 0.8, "grad_norm": 0.2747489202760835, "learning_rate": 2.977633732220769e-06, "loss": 0.5736, "step": 2456 }, { "epoch": 0.8, "grad_norm": 0.2900335434253571, "learning_rate": 2.9681521086743426e-06, "loss": 0.5842, "step": 2457 }, { "epoch": 0.8, "grad_norm": 0.27751036532018036, "learning_rate": 2.9586839474240353e-06, "loss": 0.59, "step": 2458 }, { "epoch": 0.8, "grad_norm": 0.3502296089269977, "learning_rate": 2.9492292590636617e-06, "loss": 0.8725, "step": 2459 }, { "epoch": 0.8, "grad_norm": 0.26575377189932, "learning_rate": 2.9397880541719645e-06, "loss": 0.5632, "step": 2460 }, { "epoch": 0.8, "grad_norm": 0.27198682258568263, "learning_rate": 2.930360343312599e-06, "loss": 0.5754, "step": 2461 }, { "epoch": 0.8, "grad_norm": 0.34044615278846, "learning_rate": 2.9209461370341204e-06, "loss": 0.8595, "step": 2462 }, { "epoch": 0.8, "grad_norm": 0.27181990954807345, "learning_rate": 2.911545445869983e-06, "loss": 0.5859, "step": 2463 }, { "epoch": 0.8, "grad_norm": 0.3188181490298025, "learning_rate": 2.902158280338506e-06, "loss": 0.6271, "step": 2464 }, { "epoch": 0.81, "grad_norm": 0.2860095544618378, "learning_rate": 2.8927846509428778e-06, "loss": 0.6061, "step": 2465 }, { "epoch": 0.81, "grad_norm": 0.34035176240581744, "learning_rate": 2.883424568171154e-06, "loss": 0.8342, "step": 2466 }, { "epoch": 0.81, "grad_norm": 0.2802223110127395, "learning_rate": 2.8740780424962172e-06, "loss": 0.5741, "step": 2467 }, { "epoch": 0.81, "grad_norm": 0.27287587292901555, "learning_rate": 2.86474508437579e-06, "loss": 0.5791, "step": 2468 }, { "epoch": 0.81, "grad_norm": 0.32773167734607955, "learning_rate": 2.8554257042524098e-06, "loss": 0.8277, "step": 2469 }, { "epoch": 0.81, "grad_norm": 0.38294807513530105, "learning_rate": 2.8461199125534233e-06, "loss": 0.563, "step": 2470 }, { "epoch": 0.81, "grad_norm": 0.27548995325975756, "learning_rate": 2.8368277196909705e-06, "loss": 0.5883, "step": 2471 }, { "epoch": 0.81, "grad_norm": 0.33674722392864814, "learning_rate": 2.8275491360619875e-06, "loss": 0.8375, "step": 2472 }, { "epoch": 0.81, "grad_norm": 0.2687409087767182, "learning_rate": 2.8182841720481707e-06, "loss": 0.5693, "step": 2473 }, { "epoch": 0.81, "grad_norm": 0.2655069519987435, "learning_rate": 2.809032838015982e-06, "loss": 0.5781, "step": 2474 }, { "epoch": 0.81, "grad_norm": 0.3138613489592893, "learning_rate": 2.7997951443166265e-06, "loss": 0.6435, "step": 2475 }, { "epoch": 0.81, "grad_norm": 0.2654839922009079, "learning_rate": 2.7905711012860647e-06, "loss": 0.5647, "step": 2476 }, { "epoch": 0.81, "grad_norm": 0.3324658093266811, "learning_rate": 2.781360719244964e-06, "loss": 0.8582, "step": 2477 }, { "epoch": 0.81, "grad_norm": 0.2707823999504809, "learning_rate": 2.77216400849872e-06, "loss": 0.598, "step": 2478 }, { "epoch": 0.81, "grad_norm": 0.28939395365866905, "learning_rate": 2.762980979337422e-06, "loss": 0.6015, "step": 2479 }, { "epoch": 0.81, "grad_norm": 0.33528780602994956, "learning_rate": 2.7538116420358573e-06, "loss": 0.858, "step": 2480 }, { "epoch": 0.81, "grad_norm": 0.2638474324303721, "learning_rate": 2.7446560068534882e-06, "loss": 0.5643, "step": 2481 }, { "epoch": 0.81, "grad_norm": 0.2619208207904478, "learning_rate": 2.735514084034457e-06, "loss": 0.5832, "step": 2482 }, { "epoch": 0.81, "grad_norm": 0.3299312187488621, "learning_rate": 2.7263858838075523e-06, "loss": 0.8332, "step": 2483 }, { "epoch": 0.81, "grad_norm": 0.26353461831260155, "learning_rate": 2.7172714163862067e-06, "loss": 0.5837, "step": 2484 }, { "epoch": 0.81, "grad_norm": 0.26185231943348525, "learning_rate": 2.7081706919685045e-06, "loss": 0.5803, "step": 2485 }, { "epoch": 0.81, "grad_norm": 0.32613138984263607, "learning_rate": 2.6990837207371338e-06, "loss": 0.646, "step": 2486 }, { "epoch": 0.81, "grad_norm": 0.38904161067347065, "learning_rate": 2.690010512859403e-06, "loss": 0.5853, "step": 2487 }, { "epoch": 0.81, "grad_norm": 0.3287043354378399, "learning_rate": 2.6809510784872223e-06, "loss": 0.8206, "step": 2488 }, { "epoch": 0.81, "grad_norm": 0.2651043845223994, "learning_rate": 2.6719054277570853e-06, "loss": 0.5873, "step": 2489 }, { "epoch": 0.81, "grad_norm": 0.26279913687603496, "learning_rate": 2.6628735707900653e-06, "loss": 0.572, "step": 2490 }, { "epoch": 0.81, "grad_norm": 0.33258937999884935, "learning_rate": 2.65385551769181e-06, "loss": 0.8354, "step": 2491 }, { "epoch": 0.81, "grad_norm": 0.3213858708799489, "learning_rate": 2.64485127855251e-06, "loss": 0.6469, "step": 2492 }, { "epoch": 0.81, "grad_norm": 0.2718270389389457, "learning_rate": 2.6358608634469056e-06, "loss": 0.5998, "step": 2493 }, { "epoch": 0.81, "grad_norm": 0.28387203520702303, "learning_rate": 2.626884282434266e-06, "loss": 0.5838, "step": 2494 }, { "epoch": 0.81, "grad_norm": 0.3271493347425697, "learning_rate": 2.61792154555839e-06, "loss": 0.8599, "step": 2495 }, { "epoch": 0.82, "grad_norm": 0.2628650787046218, "learning_rate": 2.608972662847579e-06, "loss": 0.5721, "step": 2496 }, { "epoch": 0.82, "grad_norm": 0.2613724339008286, "learning_rate": 2.600037644314634e-06, "loss": 0.566, "step": 2497 }, { "epoch": 0.82, "grad_norm": 0.3366390703355518, "learning_rate": 2.5911164999568443e-06, "loss": 0.8424, "step": 2498 }, { "epoch": 0.82, "grad_norm": 0.27017791529457374, "learning_rate": 2.5822092397559754e-06, "loss": 0.5609, "step": 2499 }, { "epoch": 0.82, "grad_norm": 0.2628789713091475, "learning_rate": 2.573315873678255e-06, "loss": 0.5748, "step": 2500 }, { "epoch": 0.82, "grad_norm": 0.33426020990341226, "learning_rate": 2.564436411674376e-06, "loss": 0.8523, "step": 2501 }, { "epoch": 0.82, "grad_norm": 0.2625651080759352, "learning_rate": 2.5555708636794594e-06, "loss": 0.5724, "step": 2502 }, { "epoch": 0.82, "grad_norm": 0.3190424219653086, "learning_rate": 2.546719239613065e-06, "loss": 0.6298, "step": 2503 }, { "epoch": 0.82, "grad_norm": 0.2609451659532888, "learning_rate": 2.53788154937918e-06, "loss": 0.5687, "step": 2504 }, { "epoch": 0.82, "grad_norm": 0.2658093637807093, "learning_rate": 2.5290578028661883e-06, "loss": 0.5861, "step": 2505 }, { "epoch": 0.82, "grad_norm": 0.3280177881060458, "learning_rate": 2.52024800994688e-06, "loss": 0.8685, "step": 2506 }, { "epoch": 0.82, "grad_norm": 0.26905068600420456, "learning_rate": 2.5114521804784307e-06, "loss": 0.5875, "step": 2507 }, { "epoch": 0.82, "grad_norm": 0.3280241140173291, "learning_rate": 2.502670324302393e-06, "loss": 0.5847, "step": 2508 }, { "epoch": 0.82, "grad_norm": 0.33096971580238266, "learning_rate": 2.4939024512446833e-06, "loss": 0.8181, "step": 2509 }, { "epoch": 0.82, "grad_norm": 0.27645991987711355, "learning_rate": 2.4851485711155716e-06, "loss": 0.5921, "step": 2510 }, { "epoch": 0.82, "grad_norm": 0.27047355341737306, "learning_rate": 2.4764086937096804e-06, "loss": 0.605, "step": 2511 }, { "epoch": 0.82, "grad_norm": 0.34454753897559504, "learning_rate": 2.467682828805956e-06, "loss": 0.8571, "step": 2512 }, { "epoch": 0.82, "grad_norm": 0.25945054536434414, "learning_rate": 2.4589709861676605e-06, "loss": 0.5625, "step": 2513 }, { "epoch": 0.82, "grad_norm": 2.1210358595227565, "learning_rate": 2.450273175542385e-06, "loss": 0.6314, "step": 2514 }, { "epoch": 0.82, "grad_norm": 0.26292841198270117, "learning_rate": 2.4415894066620066e-06, "loss": 0.5883, "step": 2515 }, { "epoch": 0.82, "grad_norm": 0.26854538222027735, "learning_rate": 2.432919689242691e-06, "loss": 0.5893, "step": 2516 }, { "epoch": 0.82, "grad_norm": 0.3363512302856185, "learning_rate": 2.4242640329848886e-06, "loss": 0.8694, "step": 2517 }, { "epoch": 0.82, "grad_norm": 0.26648215354783034, "learning_rate": 2.4156224475733113e-06, "loss": 0.5874, "step": 2518 }, { "epoch": 0.82, "grad_norm": 0.2830748763337651, "learning_rate": 2.406994942676926e-06, "loss": 0.5919, "step": 2519 }, { "epoch": 0.82, "grad_norm": 0.34127430232957173, "learning_rate": 2.398381527948958e-06, "loss": 0.8675, "step": 2520 }, { "epoch": 0.82, "grad_norm": 0.43150674919330495, "learning_rate": 2.3897822130268528e-06, "loss": 0.5868, "step": 2521 }, { "epoch": 0.82, "grad_norm": 0.2574303472632027, "learning_rate": 2.38119700753228e-06, "loss": 0.5756, "step": 2522 }, { "epoch": 0.82, "grad_norm": 0.3389375021878937, "learning_rate": 2.372625921071137e-06, "loss": 0.8147, "step": 2523 }, { "epoch": 0.82, "grad_norm": 0.2676426391919968, "learning_rate": 2.36406896323351e-06, "loss": 0.5715, "step": 2524 }, { "epoch": 0.82, "grad_norm": 0.3130790624507854, "learning_rate": 2.355526143593677e-06, "loss": 0.6159, "step": 2525 }, { "epoch": 0.82, "grad_norm": 0.2628981389052458, "learning_rate": 2.3469974717101054e-06, "loss": 0.5929, "step": 2526 }, { "epoch": 0.83, "grad_norm": 0.3254136794030351, "learning_rate": 2.338482957125424e-06, "loss": 0.8092, "step": 2527 }, { "epoch": 0.83, "grad_norm": 0.2819412795350119, "learning_rate": 2.329982609366428e-06, "loss": 0.587, "step": 2528 }, { "epoch": 0.83, "grad_norm": 0.31748737564259266, "learning_rate": 2.321496437944054e-06, "loss": 0.5833, "step": 2529 }, { "epoch": 0.83, "grad_norm": 0.27461095666911195, "learning_rate": 2.313024452353389e-06, "loss": 0.5877, "step": 2530 }, { "epoch": 0.83, "grad_norm": 0.3401087489606479, "learning_rate": 2.3045666620736354e-06, "loss": 0.8383, "step": 2531 }, { "epoch": 0.83, "grad_norm": 0.26397723056145, "learning_rate": 2.2961230765681158e-06, "loss": 0.5777, "step": 2532 }, { "epoch": 0.83, "grad_norm": 0.2684122804834667, "learning_rate": 2.2876937052842656e-06, "loss": 0.5836, "step": 2533 }, { "epoch": 0.83, "grad_norm": 0.33448012592155946, "learning_rate": 2.279278557653611e-06, "loss": 0.8387, "step": 2534 }, { "epoch": 0.83, "grad_norm": 0.2731375691844368, "learning_rate": 2.2708776430917595e-06, "loss": 0.5853, "step": 2535 }, { "epoch": 0.83, "grad_norm": 0.4062267542898622, "learning_rate": 2.262490970998403e-06, "loss": 0.5996, "step": 2536 }, { "epoch": 0.83, "grad_norm": 0.2727137084775787, "learning_rate": 2.254118550757286e-06, "loss": 0.5586, "step": 2537 }, { "epoch": 0.83, "grad_norm": 0.32877269561134365, "learning_rate": 2.245760391736216e-06, "loss": 0.8415, "step": 2538 }, { "epoch": 0.83, "grad_norm": 0.2668521716375718, "learning_rate": 2.237416503287043e-06, "loss": 0.5559, "step": 2539 }, { "epoch": 0.83, "grad_norm": 0.2676356619420125, "learning_rate": 2.2290868947456466e-06, "loss": 0.5773, "step": 2540 }, { "epoch": 0.83, "grad_norm": 0.328751084237004, "learning_rate": 2.220771575431928e-06, "loss": 0.8347, "step": 2541 }, { "epoch": 0.83, "grad_norm": 0.27344417384841696, "learning_rate": 2.212470554649805e-06, "loss": 0.5979, "step": 2542 }, { "epoch": 0.83, "grad_norm": 0.2664164273903956, "learning_rate": 2.204183841687194e-06, "loss": 0.5773, "step": 2543 }, { "epoch": 0.83, "grad_norm": 0.3239753287738445, "learning_rate": 2.1959114458160022e-06, "loss": 0.8218, "step": 2544 }, { "epoch": 0.83, "grad_norm": 0.2679459366477199, "learning_rate": 2.1876533762921176e-06, "loss": 0.5651, "step": 2545 }, { "epoch": 0.83, "grad_norm": 0.2734280983208972, "learning_rate": 2.1794096423553984e-06, "loss": 0.5953, "step": 2546 }, { "epoch": 0.83, "grad_norm": 0.3133465109017861, "learning_rate": 2.171180253229665e-06, "loss": 0.6333, "step": 2547 }, { "epoch": 0.83, "grad_norm": 0.26427815170731406, "learning_rate": 2.1629652181226835e-06, "loss": 0.5741, "step": 2548 }, { "epoch": 0.83, "grad_norm": 0.33468837297731807, "learning_rate": 2.1547645462261663e-06, "loss": 0.8382, "step": 2549 }, { "epoch": 0.83, "grad_norm": 0.2634790687279652, "learning_rate": 2.1465782467157483e-06, "loss": 0.5869, "step": 2550 }, { "epoch": 0.83, "grad_norm": 0.2809392096757428, "learning_rate": 2.1384063287509835e-06, "loss": 0.5803, "step": 2551 }, { "epoch": 0.83, "grad_norm": 0.34023071036186364, "learning_rate": 2.130248801475344e-06, "loss": 0.882, "step": 2552 }, { "epoch": 0.83, "grad_norm": 0.26461675332150986, "learning_rate": 2.1221056740161872e-06, "loss": 0.5763, "step": 2553 }, { "epoch": 0.83, "grad_norm": 0.2668601998383651, "learning_rate": 2.1139769554847675e-06, "loss": 0.5875, "step": 2554 }, { "epoch": 0.83, "grad_norm": 0.33636504705081277, "learning_rate": 2.105862654976214e-06, "loss": 0.8338, "step": 2555 }, { "epoch": 0.83, "grad_norm": 0.28449115370825856, "learning_rate": 2.0977627815695217e-06, "loss": 0.5966, "step": 2556 }, { "epoch": 0.84, "grad_norm": 0.27170666642893077, "learning_rate": 2.0896773443275447e-06, "loss": 0.5891, "step": 2557 }, { "epoch": 0.84, "grad_norm": 0.3199318876000917, "learning_rate": 2.081606352296991e-06, "loss": 0.6291, "step": 2558 }, { "epoch": 0.84, "grad_norm": 0.2666663668967057, "learning_rate": 2.0735498145083995e-06, "loss": 0.6012, "step": 2559 }, { "epoch": 0.84, "grad_norm": 0.33704134779160355, "learning_rate": 2.0655077399761358e-06, "loss": 0.8567, "step": 2560 }, { "epoch": 0.84, "grad_norm": 0.2758791105743512, "learning_rate": 2.0574801376983817e-06, "loss": 0.5884, "step": 2561 }, { "epoch": 0.84, "grad_norm": 0.4161522695709421, "learning_rate": 2.0494670166571356e-06, "loss": 0.5845, "step": 2562 }, { "epoch": 0.84, "grad_norm": 0.33146033568503763, "learning_rate": 2.0414683858181824e-06, "loss": 0.8201, "step": 2563 }, { "epoch": 0.84, "grad_norm": 0.2680614276275781, "learning_rate": 2.0334842541311003e-06, "loss": 0.5722, "step": 2564 }, { "epoch": 0.84, "grad_norm": 0.267800734288013, "learning_rate": 2.02551463052924e-06, "loss": 0.5733, "step": 2565 }, { "epoch": 0.84, "grad_norm": 0.33299184186946174, "learning_rate": 2.0175595239297213e-06, "loss": 0.8341, "step": 2566 }, { "epoch": 0.84, "grad_norm": 0.26957390765805656, "learning_rate": 2.0096189432334194e-06, "loss": 0.6103, "step": 2567 }, { "epoch": 0.84, "grad_norm": 0.27256756179562075, "learning_rate": 2.001692897324966e-06, "loss": 0.599, "step": 2568 }, { "epoch": 0.84, "grad_norm": 0.31622087030370144, "learning_rate": 1.9937813950727184e-06, "loss": 0.6327, "step": 2569 }, { "epoch": 0.84, "grad_norm": 0.3287861417697612, "learning_rate": 1.985884445328761e-06, "loss": 0.8247, "step": 2570 }, { "epoch": 0.84, "grad_norm": 0.2694946063737444, "learning_rate": 1.978002056928909e-06, "loss": 0.5766, "step": 2571 }, { "epoch": 0.84, "grad_norm": 0.45881641704463944, "learning_rate": 1.9701342386926712e-06, "loss": 0.5742, "step": 2572 }, { "epoch": 0.84, "grad_norm": 0.2664210089560569, "learning_rate": 1.9622809994232586e-06, "loss": 0.6061, "step": 2573 }, { "epoch": 0.84, "grad_norm": 0.31988429993097994, "learning_rate": 1.9544423479075694e-06, "loss": 0.6477, "step": 2574 }, { "epoch": 0.84, "grad_norm": 0.33816466402348255, "learning_rate": 1.9466182929161847e-06, "loss": 0.8626, "step": 2575 }, { "epoch": 0.84, "grad_norm": 0.26524307219714843, "learning_rate": 1.9388088432033446e-06, "loss": 0.5711, "step": 2576 }, { "epoch": 0.84, "grad_norm": 0.2743659688883386, "learning_rate": 1.9310140075069517e-06, "loss": 0.5998, "step": 2577 }, { "epoch": 0.84, "grad_norm": 0.3274848820646191, "learning_rate": 1.9232337945485657e-06, "loss": 0.8516, "step": 2578 }, { "epoch": 0.84, "grad_norm": 0.2653324204739662, "learning_rate": 1.91546821303337e-06, "loss": 0.5728, "step": 2579 }, { "epoch": 0.84, "grad_norm": 0.27157587850820913, "learning_rate": 1.9077172716501863e-06, "loss": 0.5849, "step": 2580 }, { "epoch": 0.84, "grad_norm": 0.33397181823618466, "learning_rate": 1.899980979071455e-06, "loss": 0.8472, "step": 2581 }, { "epoch": 0.84, "grad_norm": 0.27773230229347107, "learning_rate": 1.892259343953226e-06, "loss": 0.5708, "step": 2582 }, { "epoch": 0.84, "grad_norm": 0.2756757784253298, "learning_rate": 1.8845523749351444e-06, "loss": 0.5895, "step": 2583 }, { "epoch": 0.84, "grad_norm": 0.3293688942331363, "learning_rate": 1.8768600806404528e-06, "loss": 0.8375, "step": 2584 }, { "epoch": 0.84, "grad_norm": 0.2509559199666771, "learning_rate": 1.869182469675968e-06, "loss": 0.3916, "step": 2585 }, { "epoch": 0.84, "grad_norm": 0.32005641256825434, "learning_rate": 1.8615195506320764e-06, "loss": 0.8323, "step": 2586 }, { "epoch": 0.84, "grad_norm": 0.2657915590874912, "learning_rate": 1.85387133208274e-06, "loss": 0.5745, "step": 2587 }, { "epoch": 0.85, "grad_norm": 0.2653223566479628, "learning_rate": 1.8462378225854564e-06, "loss": 0.5723, "step": 2588 }, { "epoch": 0.85, "grad_norm": 0.32277633081624774, "learning_rate": 1.8386190306812689e-06, "loss": 0.8552, "step": 2589 }, { "epoch": 0.85, "grad_norm": 0.26716630163581967, "learning_rate": 1.831014964894765e-06, "loss": 0.5585, "step": 2590 }, { "epoch": 0.85, "grad_norm": 0.2685676635744349, "learning_rate": 1.8234256337340389e-06, "loss": 0.5647, "step": 2591 }, { "epoch": 0.85, "grad_norm": 0.34068963719054085, "learning_rate": 1.815851045690708e-06, "loss": 0.8417, "step": 2592 }, { "epoch": 0.85, "grad_norm": 0.26360271436201865, "learning_rate": 1.8082912092398913e-06, "loss": 0.5678, "step": 2593 }, { "epoch": 0.85, "grad_norm": 0.27426989880516583, "learning_rate": 1.8007461328402042e-06, "loss": 0.5881, "step": 2594 }, { "epoch": 0.85, "grad_norm": 0.329733140298438, "learning_rate": 1.7932158249337438e-06, "loss": 0.8408, "step": 2595 }, { "epoch": 0.85, "grad_norm": 0.31779838630100876, "learning_rate": 1.7857002939460837e-06, "loss": 0.6168, "step": 2596 }, { "epoch": 0.85, "grad_norm": 0.26338226924608765, "learning_rate": 1.7781995482862707e-06, "loss": 0.5672, "step": 2597 }, { "epoch": 0.85, "grad_norm": 0.2697405056600169, "learning_rate": 1.770713596346798e-06, "loss": 0.5788, "step": 2598 }, { "epoch": 0.85, "grad_norm": 0.3324886863271867, "learning_rate": 1.763242446503609e-06, "loss": 0.8396, "step": 2599 }, { "epoch": 0.85, "grad_norm": 0.2644610039865245, "learning_rate": 1.7557861071160953e-06, "loss": 0.5615, "step": 2600 }, { "epoch": 0.85, "grad_norm": 0.6356418421468402, "learning_rate": 1.748344586527063e-06, "loss": 0.5956, "step": 2601 }, { "epoch": 0.85, "grad_norm": 0.2687156602711916, "learning_rate": 1.7409178930627473e-06, "loss": 0.5708, "step": 2602 }, { "epoch": 0.85, "grad_norm": 0.32602232729781944, "learning_rate": 1.7335060350327859e-06, "loss": 0.8577, "step": 2603 }, { "epoch": 0.85, "grad_norm": 0.26717841047070084, "learning_rate": 1.726109020730225e-06, "loss": 0.5759, "step": 2604 }, { "epoch": 0.85, "grad_norm": 0.8828056133407399, "learning_rate": 1.7187268584314942e-06, "loss": 0.5642, "step": 2605 }, { "epoch": 0.85, "grad_norm": 0.3593867924153656, "learning_rate": 1.7113595563964164e-06, "loss": 0.8466, "step": 2606 }, { "epoch": 0.85, "grad_norm": 0.3067648635569012, "learning_rate": 1.7040071228681775e-06, "loss": 0.6023, "step": 2607 }, { "epoch": 0.85, "grad_norm": 0.26764508974635026, "learning_rate": 1.6966695660733271e-06, "loss": 0.5755, "step": 2608 }, { "epoch": 0.85, "grad_norm": 0.2623721067608887, "learning_rate": 1.6893468942217815e-06, "loss": 0.5711, "step": 2609 }, { "epoch": 0.85, "grad_norm": 0.32485758030460765, "learning_rate": 1.6820391155067881e-06, "loss": 0.8173, "step": 2610 }, { "epoch": 0.85, "grad_norm": 0.2696901291713876, "learning_rate": 1.6747462381049417e-06, "loss": 0.5824, "step": 2611 }, { "epoch": 0.85, "grad_norm": 0.2644517467891722, "learning_rate": 1.6674682701761496e-06, "loss": 0.5684, "step": 2612 }, { "epoch": 0.85, "grad_norm": 0.3463694833394926, "learning_rate": 1.660205219863653e-06, "loss": 0.8277, "step": 2613 }, { "epoch": 0.85, "grad_norm": 0.27291273644659814, "learning_rate": 1.6529570952939948e-06, "loss": 0.6051, "step": 2614 }, { "epoch": 0.85, "grad_norm": 0.26974179517048513, "learning_rate": 1.6457239045770146e-06, "loss": 0.5661, "step": 2615 }, { "epoch": 0.85, "grad_norm": 0.3366547781275709, "learning_rate": 1.6385056558058514e-06, "loss": 0.8484, "step": 2616 }, { "epoch": 0.85, "grad_norm": 0.26587906956983715, "learning_rate": 1.631302357056917e-06, "loss": 0.5553, "step": 2617 }, { "epoch": 0.85, "grad_norm": 0.3189144464515054, "learning_rate": 1.6241140163898989e-06, "loss": 0.6275, "step": 2618 }, { "epoch": 0.86, "grad_norm": 0.26128297139237605, "learning_rate": 1.6169406418477551e-06, "loss": 0.6023, "step": 2619 }, { "epoch": 0.86, "grad_norm": 0.2787173618509744, "learning_rate": 1.609782241456687e-06, "loss": 0.567, "step": 2620 }, { "epoch": 0.86, "grad_norm": 0.3553156465371861, "learning_rate": 1.602638823226143e-06, "loss": 0.8708, "step": 2621 }, { "epoch": 0.86, "grad_norm": 0.26862698236268706, "learning_rate": 1.5955103951488177e-06, "loss": 0.5907, "step": 2622 }, { "epoch": 0.86, "grad_norm": 0.26634020050229185, "learning_rate": 1.5883969652006242e-06, "loss": 0.5896, "step": 2623 }, { "epoch": 0.86, "grad_norm": 0.34490934219724717, "learning_rate": 1.5812985413406943e-06, "loss": 0.8232, "step": 2624 }, { "epoch": 0.86, "grad_norm": 0.2676106389250192, "learning_rate": 1.57421513151138e-06, "loss": 0.5854, "step": 2625 }, { "epoch": 0.86, "grad_norm": 0.2581069582519352, "learning_rate": 1.5671467436382248e-06, "loss": 0.5756, "step": 2626 }, { "epoch": 0.86, "grad_norm": 0.4532907989312693, "learning_rate": 1.5600933856299637e-06, "loss": 0.8407, "step": 2627 }, { "epoch": 0.86, "grad_norm": 0.26160511445370344, "learning_rate": 1.5530550653785159e-06, "loss": 0.5791, "step": 2628 }, { "epoch": 0.86, "grad_norm": 0.3161455143686558, "learning_rate": 1.5460317907589833e-06, "loss": 0.6312, "step": 2629 }, { "epoch": 0.86, "grad_norm": 0.2656278813249736, "learning_rate": 1.5390235696296268e-06, "loss": 0.5745, "step": 2630 }, { "epoch": 0.86, "grad_norm": 0.2643458022900077, "learning_rate": 1.5320304098318556e-06, "loss": 0.5697, "step": 2631 }, { "epoch": 0.86, "grad_norm": 0.3236300214261658, "learning_rate": 1.5250523191902455e-06, "loss": 0.8755, "step": 2632 }, { "epoch": 0.86, "grad_norm": 0.2914930969069279, "learning_rate": 1.518089305512498e-06, "loss": 0.5833, "step": 2633 }, { "epoch": 0.86, "grad_norm": 0.2660383646468742, "learning_rate": 1.5111413765894439e-06, "loss": 0.575, "step": 2634 }, { "epoch": 0.86, "grad_norm": 0.32476843742691497, "learning_rate": 1.50420854019505e-06, "loss": 0.8431, "step": 2635 }, { "epoch": 0.86, "grad_norm": 0.2719114957397966, "learning_rate": 1.4972908040863831e-06, "loss": 0.5912, "step": 2636 }, { "epoch": 0.86, "grad_norm": 0.26503955377432936, "learning_rate": 1.4903881760036165e-06, "loss": 0.5692, "step": 2637 }, { "epoch": 0.86, "grad_norm": 0.32817554810819616, "learning_rate": 1.4835006636700239e-06, "loss": 0.8255, "step": 2638 }, { "epoch": 0.86, "grad_norm": 0.2570759160338691, "learning_rate": 1.4766282747919674e-06, "loss": 0.5644, "step": 2639 }, { "epoch": 0.86, "grad_norm": 0.3145145562001213, "learning_rate": 1.469771017058873e-06, "loss": 0.6003, "step": 2640 }, { "epoch": 0.86, "grad_norm": 0.2844315891313427, "learning_rate": 1.4629288981432566e-06, "loss": 0.5836, "step": 2641 }, { "epoch": 0.86, "grad_norm": 0.3341631359251498, "learning_rate": 1.456101925700684e-06, "loss": 0.8669, "step": 2642 }, { "epoch": 0.86, "grad_norm": 0.26474314468818233, "learning_rate": 1.4492901073697723e-06, "loss": 0.5795, "step": 2643 }, { "epoch": 0.86, "grad_norm": 0.25813821209302096, "learning_rate": 1.4424934507721926e-06, "loss": 0.5509, "step": 2644 }, { "epoch": 0.86, "grad_norm": 0.31443036664288926, "learning_rate": 1.435711963512642e-06, "loss": 0.6475, "step": 2645 }, { "epoch": 0.86, "grad_norm": 0.26809169524124776, "learning_rate": 1.428945653178851e-06, "loss": 0.5835, "step": 2646 }, { "epoch": 0.86, "grad_norm": 0.2766383184290038, "learning_rate": 1.4221945273415614e-06, "loss": 0.7301, "step": 2647 }, { "epoch": 0.86, "grad_norm": 0.26626778989713257, "learning_rate": 1.4154585935545367e-06, "loss": 0.5728, "step": 2648 }, { "epoch": 0.87, "grad_norm": 0.32850987500701867, "learning_rate": 1.4087378593545358e-06, "loss": 0.8237, "step": 2649 }, { "epoch": 0.87, "grad_norm": 0.2676572807946131, "learning_rate": 1.402032332261302e-06, "loss": 0.5682, "step": 2650 }, { "epoch": 0.87, "grad_norm": 0.31347660825587254, "learning_rate": 1.3953420197775808e-06, "loss": 0.6161, "step": 2651 }, { "epoch": 0.87, "grad_norm": 0.2588549014831032, "learning_rate": 1.3886669293890837e-06, "loss": 0.5547, "step": 2652 }, { "epoch": 0.87, "grad_norm": 0.3332777464295004, "learning_rate": 1.382007068564488e-06, "loss": 0.8626, "step": 2653 }, { "epoch": 0.87, "grad_norm": 0.2663714540389272, "learning_rate": 1.375362444755442e-06, "loss": 0.5788, "step": 2654 }, { "epoch": 0.87, "grad_norm": 0.2647407544778368, "learning_rate": 1.3687330653965363e-06, "loss": 0.5743, "step": 2655 }, { "epoch": 0.87, "grad_norm": 0.31675916235068646, "learning_rate": 1.3621189379053046e-06, "loss": 0.6427, "step": 2656 }, { "epoch": 0.87, "grad_norm": 0.26733759357870684, "learning_rate": 1.3555200696822235e-06, "loss": 0.5611, "step": 2657 }, { "epoch": 0.87, "grad_norm": 0.3433328528449939, "learning_rate": 1.3489364681106903e-06, "loss": 0.8443, "step": 2658 }, { "epoch": 0.87, "grad_norm": 0.45579509313585165, "learning_rate": 1.3423681405570126e-06, "loss": 0.5787, "step": 2659 }, { "epoch": 0.87, "grad_norm": 0.26565395750424164, "learning_rate": 1.3358150943704233e-06, "loss": 0.5776, "step": 2660 }, { "epoch": 0.87, "grad_norm": 0.32673566492493517, "learning_rate": 1.3292773368830508e-06, "loss": 0.8452, "step": 2661 }, { "epoch": 0.87, "grad_norm": 0.26365738650972786, "learning_rate": 1.322754875409915e-06, "loss": 0.5659, "step": 2662 }, { "epoch": 0.87, "grad_norm": 0.2664528802385754, "learning_rate": 1.31624771724892e-06, "loss": 0.5897, "step": 2663 }, { "epoch": 0.87, "grad_norm": 0.3320631451104972, "learning_rate": 1.3097558696808553e-06, "loss": 0.8248, "step": 2664 }, { "epoch": 0.87, "grad_norm": 0.264890722781542, "learning_rate": 1.3032793399693737e-06, "loss": 0.5609, "step": 2665 }, { "epoch": 0.87, "grad_norm": 0.2931352467415526, "learning_rate": 1.2968181353609854e-06, "loss": 0.5908, "step": 2666 }, { "epoch": 0.87, "grad_norm": 0.30549021339743265, "learning_rate": 1.290372263085064e-06, "loss": 0.6178, "step": 2667 }, { "epoch": 0.87, "grad_norm": 0.31953934392926797, "learning_rate": 1.2839417303538203e-06, "loss": 0.8315, "step": 2668 }, { "epoch": 0.87, "grad_norm": 0.26517239195000214, "learning_rate": 1.2775265443622974e-06, "loss": 0.5661, "step": 2669 }, { "epoch": 0.87, "grad_norm": 0.26577683925378087, "learning_rate": 1.2711267122883784e-06, "loss": 0.5839, "step": 2670 }, { "epoch": 0.87, "grad_norm": 0.2560559916816502, "learning_rate": 1.2647422412927623e-06, "loss": 0.5917, "step": 2671 }, { "epoch": 0.87, "grad_norm": 0.5357688687256464, "learning_rate": 1.2583731385189562e-06, "loss": 0.8593, "step": 2672 }, { "epoch": 0.87, "grad_norm": 0.266031688224961, "learning_rate": 1.2520194110932798e-06, "loss": 0.5561, "step": 2673 }, { "epoch": 0.87, "grad_norm": 0.2590847348042691, "learning_rate": 1.2456810661248447e-06, "loss": 0.5638, "step": 2674 }, { "epoch": 0.87, "grad_norm": 0.32348386175755844, "learning_rate": 1.239358110705548e-06, "loss": 0.8585, "step": 2675 }, { "epoch": 0.87, "grad_norm": 0.2666589103945411, "learning_rate": 1.2330505519100803e-06, "loss": 0.5787, "step": 2676 }, { "epoch": 0.87, "grad_norm": 0.2662869157826582, "learning_rate": 1.2267583967958918e-06, "loss": 0.5807, "step": 2677 }, { "epoch": 0.87, "grad_norm": 0.3304113761833091, "learning_rate": 1.2204816524032003e-06, "loss": 0.605, "step": 2678 }, { "epoch": 0.87, "grad_norm": 0.32933202356376334, "learning_rate": 1.2142203257549832e-06, "loss": 0.8238, "step": 2679 }, { "epoch": 0.88, "grad_norm": 0.26688486301851955, "learning_rate": 1.2079744238569701e-06, "loss": 0.5912, "step": 2680 }, { "epoch": 0.88, "grad_norm": 0.2634716046387676, "learning_rate": 1.201743953697627e-06, "loss": 0.5796, "step": 2681 }, { "epoch": 0.88, "grad_norm": 0.317323583917811, "learning_rate": 1.1955289222481513e-06, "loss": 0.812, "step": 2682 }, { "epoch": 0.88, "grad_norm": 0.26271433290792884, "learning_rate": 1.1893293364624786e-06, "loss": 0.5804, "step": 2683 }, { "epoch": 0.88, "grad_norm": 0.29626959130766806, "learning_rate": 1.1831452032772499e-06, "loss": 0.5657, "step": 2684 }, { "epoch": 0.88, "grad_norm": 0.32457114265360326, "learning_rate": 1.1769765296118157e-06, "loss": 0.8283, "step": 2685 }, { "epoch": 0.88, "grad_norm": 0.2639665048082359, "learning_rate": 1.1708233223682429e-06, "loss": 0.5719, "step": 2686 }, { "epoch": 0.88, "grad_norm": 0.2620503275377508, "learning_rate": 1.1646855884312813e-06, "loss": 0.567, "step": 2687 }, { "epoch": 0.88, "grad_norm": 0.3298057178986045, "learning_rate": 1.1585633346683655e-06, "loss": 0.8289, "step": 2688 }, { "epoch": 0.88, "grad_norm": 0.242289370416065, "learning_rate": 1.1524565679296196e-06, "loss": 0.3847, "step": 2689 }, { "epoch": 0.88, "grad_norm": 0.3306562292580629, "learning_rate": 1.146365295047836e-06, "loss": 0.8131, "step": 2690 }, { "epoch": 0.88, "grad_norm": 0.26450684256224555, "learning_rate": 1.140289522838463e-06, "loss": 0.5958, "step": 2691 }, { "epoch": 0.88, "grad_norm": 0.26476066415205307, "learning_rate": 1.1342292580996195e-06, "loss": 0.5748, "step": 2692 }, { "epoch": 0.88, "grad_norm": 0.33178534974787127, "learning_rate": 1.128184507612063e-06, "loss": 0.8376, "step": 2693 }, { "epoch": 0.88, "grad_norm": 0.26770691677070574, "learning_rate": 1.1221552781391931e-06, "loss": 0.5709, "step": 2694 }, { "epoch": 0.88, "grad_norm": 0.2635928242812847, "learning_rate": 1.1161415764270505e-06, "loss": 0.5838, "step": 2695 }, { "epoch": 0.88, "grad_norm": 0.34617672006982325, "learning_rate": 1.110143409204296e-06, "loss": 0.8403, "step": 2696 }, { "epoch": 0.88, "grad_norm": 0.2664872158180326, "learning_rate": 1.1041607831822064e-06, "loss": 0.5751, "step": 2697 }, { "epoch": 0.88, "grad_norm": 0.26173173477862427, "learning_rate": 1.0981937050546703e-06, "loss": 0.5945, "step": 2698 }, { "epoch": 0.88, "grad_norm": 0.321029957154801, "learning_rate": 1.0922421814981904e-06, "loss": 0.8394, "step": 2699 }, { "epoch": 0.88, "grad_norm": 0.2474690620583462, "learning_rate": 1.0863062191718531e-06, "loss": 0.3878, "step": 2700 }, { "epoch": 0.88, "grad_norm": 0.3234801310041259, "learning_rate": 1.0803858247173392e-06, "loss": 0.8626, "step": 2701 }, { "epoch": 0.88, "grad_norm": 0.25991294761289585, "learning_rate": 1.0744810047589116e-06, "loss": 0.587, "step": 2702 }, { "epoch": 0.88, "grad_norm": 0.25985771380792944, "learning_rate": 1.0685917659034033e-06, "loss": 0.5978, "step": 2703 }, { "epoch": 0.88, "grad_norm": 0.32846613918254675, "learning_rate": 1.0627181147402154e-06, "loss": 0.8455, "step": 2704 }, { "epoch": 0.88, "grad_norm": 0.2578802535310665, "learning_rate": 1.056860057841314e-06, "loss": 0.5708, "step": 2705 }, { "epoch": 0.88, "grad_norm": 0.2761134192764629, "learning_rate": 1.051017601761206e-06, "loss": 0.5818, "step": 2706 }, { "epoch": 0.88, "grad_norm": 0.31692317866671066, "learning_rate": 1.0451907530369459e-06, "loss": 0.7885, "step": 2707 }, { "epoch": 0.88, "grad_norm": 0.26505816625341777, "learning_rate": 1.0393795181881349e-06, "loss": 0.59, "step": 2708 }, { "epoch": 0.88, "grad_norm": 0.2728357486351304, "learning_rate": 1.033583903716892e-06, "loss": 0.589, "step": 2709 }, { "epoch": 0.89, "grad_norm": 0.32911878393649924, "learning_rate": 1.0278039161078634e-06, "loss": 0.8302, "step": 2710 }, { "epoch": 0.89, "grad_norm": 0.3077204813310996, "learning_rate": 1.0220395618282159e-06, "loss": 0.6036, "step": 2711 }, { "epoch": 0.89, "grad_norm": 0.27013196354836766, "learning_rate": 1.0162908473276133e-06, "loss": 0.5858, "step": 2712 }, { "epoch": 0.89, "grad_norm": 0.31165399224072304, "learning_rate": 1.0105577790382293e-06, "loss": 0.5741, "step": 2713 }, { "epoch": 0.89, "grad_norm": 0.3289974501227926, "learning_rate": 1.0048403633747239e-06, "loss": 0.8546, "step": 2714 }, { "epoch": 0.89, "grad_norm": 0.26053278201019087, "learning_rate": 9.991386067342551e-07, "loss": 0.5776, "step": 2715 }, { "epoch": 0.89, "grad_norm": 0.2712965395250634, "learning_rate": 9.934525154964463e-07, "loss": 0.5803, "step": 2716 }, { "epoch": 0.89, "grad_norm": 0.26447125069534805, "learning_rate": 9.877820960234003e-07, "loss": 0.5638, "step": 2717 }, { "epoch": 0.89, "grad_norm": 0.3227754225165935, "learning_rate": 9.821273546596864e-07, "loss": 0.8374, "step": 2718 }, { "epoch": 0.89, "grad_norm": 0.26062523798391796, "learning_rate": 9.764882977323291e-07, "loss": 0.5942, "step": 2719 }, { "epoch": 0.89, "grad_norm": 0.26295656477390994, "learning_rate": 9.708649315508012e-07, "loss": 0.5921, "step": 2720 }, { "epoch": 0.89, "grad_norm": 0.3323071504520017, "learning_rate": 9.652572624070295e-07, "loss": 0.832, "step": 2721 }, { "epoch": 0.89, "grad_norm": 0.30656908866437305, "learning_rate": 9.596652965753632e-07, "loss": 0.6326, "step": 2722 }, { "epoch": 0.89, "grad_norm": 0.2605653144762455, "learning_rate": 9.540890403125913e-07, "loss": 0.5873, "step": 2723 }, { "epoch": 0.89, "grad_norm": 0.26116092760304804, "learning_rate": 9.485284998579257e-07, "loss": 0.5894, "step": 2724 }, { "epoch": 0.89, "grad_norm": 1.6998453636916648, "learning_rate": 9.429836814329867e-07, "loss": 0.8651, "step": 2725 }, { "epoch": 0.89, "grad_norm": 0.26530775720608774, "learning_rate": 9.374545912418065e-07, "loss": 0.5795, "step": 2726 }, { "epoch": 0.89, "grad_norm": 0.2632262275684422, "learning_rate": 9.319412354708262e-07, "loss": 0.5922, "step": 2727 }, { "epoch": 0.89, "grad_norm": 0.3305569652295422, "learning_rate": 9.264436202888732e-07, "loss": 0.6188, "step": 2728 }, { "epoch": 0.89, "grad_norm": 0.2633844349816537, "learning_rate": 9.209617518471636e-07, "loss": 0.5903, "step": 2729 }, { "epoch": 0.89, "grad_norm": 0.27524650675152035, "learning_rate": 9.15495636279301e-07, "loss": 0.7495, "step": 2730 }, { "epoch": 0.89, "grad_norm": 0.3285527726216361, "learning_rate": 9.100452797012582e-07, "loss": 0.8439, "step": 2731 }, { "epoch": 0.89, "grad_norm": 0.2689593396169757, "learning_rate": 9.046106882113753e-07, "loss": 0.5742, "step": 2732 }, { "epoch": 0.89, "grad_norm": 0.31114196482799483, "learning_rate": 8.991918678903515e-07, "loss": 0.6143, "step": 2733 }, { "epoch": 0.89, "grad_norm": 0.2697086198717114, "learning_rate": 8.937888248012504e-07, "loss": 0.5693, "step": 2734 }, { "epoch": 0.89, "grad_norm": 0.48314378700441635, "learning_rate": 8.884015649894678e-07, "loss": 0.5715, "step": 2735 }, { "epoch": 0.89, "grad_norm": 0.34096075020268046, "learning_rate": 8.830300944827469e-07, "loss": 0.8466, "step": 2736 }, { "epoch": 0.89, "grad_norm": 0.2652223252543921, "learning_rate": 8.776744192911668e-07, "loss": 0.6024, "step": 2737 }, { "epoch": 0.89, "grad_norm": 0.2648608359266809, "learning_rate": 8.723345454071308e-07, "loss": 0.5756, "step": 2738 }, { "epoch": 0.89, "grad_norm": 0.30902468606973715, "learning_rate": 8.670104788053596e-07, "loss": 0.6456, "step": 2739 }, { "epoch": 0.89, "grad_norm": 0.33456170989941736, "learning_rate": 8.617022254428931e-07, "loss": 0.8632, "step": 2740 }, { "epoch": 0.9, "grad_norm": 0.26648695685488816, "learning_rate": 8.564097912590751e-07, "loss": 0.6012, "step": 2741 }, { "epoch": 0.9, "grad_norm": 0.2625291453895999, "learning_rate": 8.511331821755459e-07, "loss": 0.5834, "step": 2742 }, { "epoch": 0.9, "grad_norm": 1.087288032713873, "learning_rate": 8.458724040962474e-07, "loss": 0.5815, "step": 2743 }, { "epoch": 0.9, "grad_norm": 0.33430896502438484, "learning_rate": 8.406274629074012e-07, "loss": 0.8387, "step": 2744 }, { "epoch": 0.9, "grad_norm": 0.26007562113356003, "learning_rate": 8.353983644775081e-07, "loss": 0.5774, "step": 2745 }, { "epoch": 0.9, "grad_norm": 0.26612700952852997, "learning_rate": 8.301851146573514e-07, "loss": 0.5757, "step": 2746 }, { "epoch": 0.9, "grad_norm": 0.32877886222422226, "learning_rate": 8.249877192799732e-07, "loss": 0.8403, "step": 2747 }, { "epoch": 0.9, "grad_norm": 0.269195266430102, "learning_rate": 8.198061841606819e-07, "loss": 0.5933, "step": 2748 }, { "epoch": 0.9, "grad_norm": 0.2621405219059068, "learning_rate": 8.146405150970348e-07, "loss": 0.5765, "step": 2749 }, { "epoch": 0.9, "grad_norm": 0.3152852180870192, "learning_rate": 8.094907178688421e-07, "loss": 0.6055, "step": 2750 }, { "epoch": 0.9, "grad_norm": 0.32702484834447015, "learning_rate": 8.043567982381511e-07, "loss": 0.8266, "step": 2751 }, { "epoch": 0.9, "grad_norm": 0.26639657343677253, "learning_rate": 7.992387619492436e-07, "loss": 0.574, "step": 2752 }, { "epoch": 0.9, "grad_norm": 0.25991184650358556, "learning_rate": 7.941366147286389e-07, "loss": 0.5715, "step": 2753 }, { "epoch": 0.9, "grad_norm": 0.3317453620931051, "learning_rate": 7.89050362285062e-07, "loss": 0.8271, "step": 2754 }, { "epoch": 0.9, "grad_norm": 0.26217038158548356, "learning_rate": 7.839800103094641e-07, "loss": 0.5719, "step": 2755 }, { "epoch": 0.9, "grad_norm": 0.26677583574417024, "learning_rate": 7.789255644750087e-07, "loss": 0.5808, "step": 2756 }, { "epoch": 0.9, "grad_norm": 0.3214916054754722, "learning_rate": 7.738870304370521e-07, "loss": 0.8636, "step": 2757 }, { "epoch": 0.9, "grad_norm": 0.2775324284968548, "learning_rate": 7.68864413833153e-07, "loss": 0.5855, "step": 2758 }, { "epoch": 0.9, "grad_norm": 0.2617466211144747, "learning_rate": 7.63857720283061e-07, "loss": 0.5615, "step": 2759 }, { "epoch": 0.9, "grad_norm": 0.2630103307068116, "learning_rate": 7.588669553887067e-07, "loss": 0.5756, "step": 2760 }, { "epoch": 0.9, "grad_norm": 0.31613833136547914, "learning_rate": 7.538921247341967e-07, "loss": 0.6397, "step": 2761 }, { "epoch": 0.9, "grad_norm": 0.3246388052640007, "learning_rate": 7.489332338858202e-07, "loss": 0.8221, "step": 2762 }, { "epoch": 0.9, "grad_norm": 0.2616189697284056, "learning_rate": 7.439902883920152e-07, "loss": 0.5976, "step": 2763 }, { "epoch": 0.9, "grad_norm": 0.26204430046235755, "learning_rate": 7.390632937833896e-07, "loss": 0.5909, "step": 2764 }, { "epoch": 0.9, "grad_norm": 0.33166257971098523, "learning_rate": 7.341522555726971e-07, "loss": 0.8065, "step": 2765 }, { "epoch": 0.9, "grad_norm": 0.2664148639424498, "learning_rate": 7.29257179254847e-07, "loss": 0.573, "step": 2766 }, { "epoch": 0.9, "grad_norm": 0.26638156598433405, "learning_rate": 7.243780703068831e-07, "loss": 0.5761, "step": 2767 }, { "epoch": 0.9, "grad_norm": 0.3402122457497822, "learning_rate": 7.195149341879787e-07, "loss": 0.8505, "step": 2768 }, { "epoch": 0.9, "grad_norm": 0.29508367073499014, "learning_rate": 7.146677763394493e-07, "loss": 0.5631, "step": 2769 }, { "epoch": 0.9, "grad_norm": 0.27204944119166735, "learning_rate": 7.098366021847186e-07, "loss": 0.6009, "step": 2770 }, { "epoch": 0.9, "grad_norm": 0.3222970068490483, "learning_rate": 7.050214171293307e-07, "loss": 0.8392, "step": 2771 }, { "epoch": 0.91, "grad_norm": 0.2448202058485172, "learning_rate": 7.002222265609476e-07, "loss": 0.3858, "step": 2772 }, { "epoch": 0.91, "grad_norm": 0.31928925951762666, "learning_rate": 6.954390358493224e-07, "loss": 0.84, "step": 2773 }, { "epoch": 0.91, "grad_norm": 0.2632388438023451, "learning_rate": 6.906718503463105e-07, "loss": 0.5796, "step": 2774 }, { "epoch": 0.91, "grad_norm": 0.2589485534624188, "learning_rate": 6.859206753858654e-07, "loss": 0.5922, "step": 2775 }, { "epoch": 0.91, "grad_norm": 0.3334053075458717, "learning_rate": 6.811855162840214e-07, "loss": 0.864, "step": 2776 }, { "epoch": 0.91, "grad_norm": 0.2676389360745008, "learning_rate": 6.764663783388919e-07, "loss": 0.5714, "step": 2777 }, { "epoch": 0.91, "grad_norm": 0.2629706169696958, "learning_rate": 6.717632668306667e-07, "loss": 0.5892, "step": 2778 }, { "epoch": 0.91, "grad_norm": 0.3365254966667942, "learning_rate": 6.670761870216047e-07, "loss": 0.8253, "step": 2779 }, { "epoch": 0.91, "grad_norm": 0.26015380384375575, "learning_rate": 6.624051441560208e-07, "loss": 0.5756, "step": 2780 }, { "epoch": 0.91, "grad_norm": 0.26848238630945503, "learning_rate": 6.577501434602995e-07, "loss": 0.5903, "step": 2781 }, { "epoch": 0.91, "grad_norm": 0.3727427737781057, "learning_rate": 6.53111190142861e-07, "loss": 0.8325, "step": 2782 }, { "epoch": 0.91, "grad_norm": 0.3096313543579887, "learning_rate": 6.484882893941768e-07, "loss": 0.6086, "step": 2783 }, { "epoch": 0.91, "grad_norm": 0.25651489220323664, "learning_rate": 6.438814463867576e-07, "loss": 0.5478, "step": 2784 }, { "epoch": 0.91, "grad_norm": 0.2695894089183635, "learning_rate": 6.392906662751486e-07, "loss": 0.5964, "step": 2785 }, { "epoch": 0.91, "grad_norm": 0.3276953938010369, "learning_rate": 6.347159541959224e-07, "loss": 0.8193, "step": 2786 }, { "epoch": 0.91, "grad_norm": 0.2597905583476498, "learning_rate": 6.301573152676665e-07, "loss": 0.5986, "step": 2787 }, { "epoch": 0.91, "grad_norm": 0.26392722387589124, "learning_rate": 6.256147545909957e-07, "loss": 0.5854, "step": 2788 }, { "epoch": 0.91, "grad_norm": 0.2622163202746446, "learning_rate": 6.210882772485243e-07, "loss": 0.5591, "step": 2789 }, { "epoch": 0.91, "grad_norm": 0.326297861615749, "learning_rate": 6.165778883048779e-07, "loss": 0.8412, "step": 2790 }, { "epoch": 0.91, "grad_norm": 0.2583429669139558, "learning_rate": 6.120835928066793e-07, "loss": 0.5798, "step": 2791 }, { "epoch": 0.91, "grad_norm": 0.26296692769766494, "learning_rate": 6.076053957825411e-07, "loss": 0.5754, "step": 2792 }, { "epoch": 0.91, "grad_norm": 0.3342982162493739, "learning_rate": 6.031433022430672e-07, "loss": 0.8565, "step": 2793 }, { "epoch": 0.91, "grad_norm": 0.30906389952130703, "learning_rate": 5.986973171808452e-07, "loss": 0.6343, "step": 2794 }, { "epoch": 0.91, "grad_norm": 0.26774843775558027, "learning_rate": 5.942674455704345e-07, "loss": 0.5908, "step": 2795 }, { "epoch": 0.91, "grad_norm": 0.2664956126194552, "learning_rate": 5.898536923683667e-07, "loss": 0.5789, "step": 2796 }, { "epoch": 0.91, "grad_norm": 0.32822922283116956, "learning_rate": 5.854560625131417e-07, "loss": 0.8524, "step": 2797 }, { "epoch": 0.91, "grad_norm": 0.25971186237924104, "learning_rate": 5.810745609252166e-07, "loss": 0.5809, "step": 2798 }, { "epoch": 0.91, "grad_norm": 0.2615281218368077, "learning_rate": 5.767091925070033e-07, "loss": 0.5576, "step": 2799 }, { "epoch": 0.91, "grad_norm": 0.31903655884056753, "learning_rate": 5.723599621428616e-07, "loss": 0.8256, "step": 2800 }, { "epoch": 0.91, "grad_norm": 0.25877334764922255, "learning_rate": 5.680268746990935e-07, "loss": 0.5826, "step": 2801 }, { "epoch": 0.92, "grad_norm": 0.26266182653644343, "learning_rate": 5.637099350239427e-07, "loss": 0.578, "step": 2802 }, { "epoch": 0.92, "grad_norm": 0.33349186811561515, "learning_rate": 5.594091479475821e-07, "loss": 0.8517, "step": 2803 }, { "epoch": 0.92, "grad_norm": 0.2624089530102057, "learning_rate": 5.551245182821157e-07, "loss": 0.592, "step": 2804 }, { "epoch": 0.92, "grad_norm": 0.309916799286011, "learning_rate": 5.50856050821562e-07, "loss": 0.6182, "step": 2805 }, { "epoch": 0.92, "grad_norm": 0.2622136263589409, "learning_rate": 5.466037503418625e-07, "loss": 0.5931, "step": 2806 }, { "epoch": 0.92, "grad_norm": 1.5462617458917376, "learning_rate": 5.423676216008693e-07, "loss": 0.5925, "step": 2807 }, { "epoch": 0.92, "grad_norm": 0.32591387960670487, "learning_rate": 5.381476693383347e-07, "loss": 0.7977, "step": 2808 }, { "epoch": 0.92, "grad_norm": 0.262409588792802, "learning_rate": 5.339438982759165e-07, "loss": 0.5575, "step": 2809 }, { "epoch": 0.92, "grad_norm": 0.30963267747415046, "learning_rate": 5.297563131171656e-07, "loss": 0.5933, "step": 2810 }, { "epoch": 0.92, "grad_norm": 0.25928252196135765, "learning_rate": 5.255849185475225e-07, "loss": 0.5643, "step": 2811 }, { "epoch": 0.92, "grad_norm": 0.332548681860434, "learning_rate": 5.214297192343104e-07, "loss": 0.8387, "step": 2812 }, { "epoch": 0.92, "grad_norm": 0.22980121973573106, "learning_rate": 5.172907198267368e-07, "loss": 0.5483, "step": 2813 }, { "epoch": 0.92, "grad_norm": 0.3224785471851996, "learning_rate": 5.131679249558807e-07, "loss": 0.8107, "step": 2814 }, { "epoch": 0.92, "grad_norm": 0.2574436769048619, "learning_rate": 5.090613392346905e-07, "loss": 0.5803, "step": 2815 }, { "epoch": 0.92, "grad_norm": 0.3084239263571586, "learning_rate": 5.049709672579722e-07, "loss": 0.6109, "step": 2816 }, { "epoch": 0.92, "grad_norm": 0.26223421930329927, "learning_rate": 5.008968136024034e-07, "loss": 0.5906, "step": 2817 }, { "epoch": 0.92, "grad_norm": 0.25571401482331413, "learning_rate": 4.968388828265025e-07, "loss": 0.5883, "step": 2818 }, { "epoch": 0.92, "grad_norm": 0.33712717854075713, "learning_rate": 4.927971794706443e-07, "loss": 0.8612, "step": 2819 }, { "epoch": 0.92, "grad_norm": 0.2865760681824552, "learning_rate": 4.887717080570431e-07, "loss": 0.5508, "step": 2820 }, { "epoch": 0.92, "grad_norm": 0.3132437017341409, "learning_rate": 4.84762473089751e-07, "loss": 0.6305, "step": 2821 }, { "epoch": 0.92, "grad_norm": 0.2677467684731647, "learning_rate": 4.807694790546563e-07, "loss": 0.5837, "step": 2822 }, { "epoch": 0.92, "grad_norm": 0.32922724993541813, "learning_rate": 4.7679273041947525e-07, "loss": 0.7944, "step": 2823 }, { "epoch": 0.92, "grad_norm": 0.2591626017436989, "learning_rate": 4.728322316337452e-07, "loss": 0.5702, "step": 2824 }, { "epoch": 0.92, "grad_norm": 0.26545979830492317, "learning_rate": 4.6888798712882155e-07, "loss": 0.5735, "step": 2825 }, { "epoch": 0.92, "grad_norm": 0.3244974955142281, "learning_rate": 4.6496000131787564e-07, "loss": 0.843, "step": 2826 }, { "epoch": 0.92, "grad_norm": 0.2754265233270991, "learning_rate": 4.6104827859588526e-07, "loss": 0.5633, "step": 2827 }, { "epoch": 0.92, "grad_norm": 0.2610716021697861, "learning_rate": 4.5715282333963105e-07, "loss": 0.5836, "step": 2828 }, { "epoch": 0.92, "grad_norm": 0.3369890522338348, "learning_rate": 4.5327363990769475e-07, "loss": 0.8593, "step": 2829 }, { "epoch": 0.92, "grad_norm": 0.27697416378165163, "learning_rate": 4.4941073264044787e-07, "loss": 0.5897, "step": 2830 }, { "epoch": 0.92, "grad_norm": 0.2596887929357789, "learning_rate": 4.455641058600529e-07, "loss": 0.5726, "step": 2831 }, { "epoch": 0.92, "grad_norm": 0.30766392056395914, "learning_rate": 4.417337638704588e-07, "loss": 0.6281, "step": 2832 }, { "epoch": 0.93, "grad_norm": 0.2653500520663439, "learning_rate": 4.379197109573907e-07, "loss": 0.5739, "step": 2833 }, { "epoch": 0.93, "grad_norm": 0.3189641426382484, "learning_rate": 4.3412195138834656e-07, "loss": 0.8339, "step": 2834 }, { "epoch": 0.93, "grad_norm": 0.26218666237119825, "learning_rate": 4.30340489412594e-07, "loss": 0.5739, "step": 2835 }, { "epoch": 0.93, "grad_norm": 0.26077768834007914, "learning_rate": 4.265753292611735e-07, "loss": 0.5579, "step": 2836 }, { "epoch": 0.93, "grad_norm": 0.3237645904034168, "learning_rate": 4.228264751468752e-07, "loss": 0.8226, "step": 2837 }, { "epoch": 0.93, "grad_norm": 0.3467727619211216, "learning_rate": 4.1909393126424886e-07, "loss": 0.5823, "step": 2838 }, { "epoch": 0.93, "grad_norm": 0.25697019059156195, "learning_rate": 4.153777017895971e-07, "loss": 0.5538, "step": 2839 }, { "epoch": 0.93, "grad_norm": 0.32752659445793086, "learning_rate": 4.116777908809638e-07, "loss": 0.8452, "step": 2840 }, { "epoch": 0.93, "grad_norm": 0.26022298185395176, "learning_rate": 4.079942026781375e-07, "loss": 0.571, "step": 2841 }, { "epoch": 0.93, "grad_norm": 0.25490861493279454, "learning_rate": 4.043269413026429e-07, "loss": 0.5515, "step": 2842 }, { "epoch": 0.93, "grad_norm": 0.3146774875829905, "learning_rate": 4.006760108577362e-07, "loss": 0.6242, "step": 2843 }, { "epoch": 0.93, "grad_norm": 0.2615182020627904, "learning_rate": 3.970414154284013e-07, "loss": 0.574, "step": 2844 }, { "epoch": 0.93, "grad_norm": 0.3228176658626438, "learning_rate": 3.9342315908134694e-07, "loss": 0.8661, "step": 2845 }, { "epoch": 0.93, "grad_norm": 0.2618556529607253, "learning_rate": 3.8982124586499804e-07, "loss": 0.5722, "step": 2846 }, { "epoch": 0.93, "grad_norm": 0.26716636486349926, "learning_rate": 3.862356798094924e-07, "loss": 0.5767, "step": 2847 }, { "epoch": 0.93, "grad_norm": 0.32767093983670176, "learning_rate": 3.8266646492667756e-07, "loss": 0.8314, "step": 2848 }, { "epoch": 0.93, "grad_norm": 0.3715343457475811, "learning_rate": 3.791136052101107e-07, "loss": 0.5941, "step": 2849 }, { "epoch": 0.93, "grad_norm": 0.2608191625206612, "learning_rate": 3.7557710463504205e-07, "loss": 0.5621, "step": 2850 }, { "epoch": 0.93, "grad_norm": 0.3203949259590308, "learning_rate": 3.7205696715842287e-07, "loss": 0.8221, "step": 2851 }, { "epoch": 0.93, "grad_norm": 0.2750332689834594, "learning_rate": 3.6855319671889433e-07, "loss": 0.603, "step": 2852 }, { "epoch": 0.93, "grad_norm": 0.2566727055210367, "learning_rate": 3.6506579723678713e-07, "loss": 0.5727, "step": 2853 }, { "epoch": 0.93, "grad_norm": 0.31490610187205387, "learning_rate": 3.6159477261410667e-07, "loss": 0.6311, "step": 2854 }, { "epoch": 0.93, "grad_norm": 0.3297986132088133, "learning_rate": 3.5814012673454966e-07, "loss": 0.8181, "step": 2855 }, { "epoch": 0.93, "grad_norm": 0.2512023761665641, "learning_rate": 3.547018634634741e-07, "loss": 0.5409, "step": 2856 }, { "epoch": 0.93, "grad_norm": 0.26087715556844926, "learning_rate": 3.5127998664791605e-07, "loss": 0.5913, "step": 2857 }, { "epoch": 0.93, "grad_norm": 0.25892098440899125, "learning_rate": 3.478745001165728e-07, "loss": 0.5533, "step": 2858 }, { "epoch": 0.93, "grad_norm": 0.3307953226489057, "learning_rate": 3.4448540767980487e-07, "loss": 0.8796, "step": 2859 }, { "epoch": 0.93, "grad_norm": 0.25343600583401077, "learning_rate": 3.411127131296271e-07, "loss": 0.5642, "step": 2860 }, { "epoch": 0.93, "grad_norm": 0.25788629691177695, "learning_rate": 3.3775642023971097e-07, "loss": 0.5824, "step": 2861 }, { "epoch": 0.93, "grad_norm": 0.3240865026009953, "learning_rate": 3.3441653276537253e-07, "loss": 0.8476, "step": 2862 }, { "epoch": 0.94, "grad_norm": 0.26093735529988066, "learning_rate": 3.310930544435725e-07, "loss": 0.5706, "step": 2863 }, { "epoch": 0.94, "grad_norm": 0.2698236615737785, "learning_rate": 3.277859889929147e-07, "loss": 0.5756, "step": 2864 }, { "epoch": 0.94, "grad_norm": 0.31025324369991314, "learning_rate": 3.244953401136375e-07, "loss": 0.631, "step": 2865 }, { "epoch": 0.94, "grad_norm": 0.3258837190328951, "learning_rate": 3.212211114876057e-07, "loss": 0.8208, "step": 2866 }, { "epoch": 0.94, "grad_norm": 0.26063613018329085, "learning_rate": 3.179633067783205e-07, "loss": 0.5918, "step": 2867 }, { "epoch": 0.94, "grad_norm": 0.2609349349300113, "learning_rate": 3.1472192963089784e-07, "loss": 0.5831, "step": 2868 }, { "epoch": 0.94, "grad_norm": 0.33151627472546213, "learning_rate": 3.114969836720816e-07, "loss": 0.8307, "step": 2869 }, { "epoch": 0.94, "grad_norm": 0.26069054388784935, "learning_rate": 3.08288472510222e-07, "loss": 0.5694, "step": 2870 }, { "epoch": 0.94, "grad_norm": 0.26054350723011604, "learning_rate": 3.050963997352907e-07, "loss": 0.5871, "step": 2871 }, { "epoch": 0.94, "grad_norm": 0.3306044740849455, "learning_rate": 3.0192076891885745e-07, "loss": 0.8548, "step": 2872 }, { "epoch": 0.94, "grad_norm": 0.25829537165496425, "learning_rate": 2.9876158361409834e-07, "loss": 0.5745, "step": 2873 }, { "epoch": 0.94, "grad_norm": 0.26287172688880617, "learning_rate": 2.956188473557925e-07, "loss": 0.6006, "step": 2874 }, { "epoch": 0.94, "grad_norm": 0.3286699618406529, "learning_rate": 2.924925636603104e-07, "loss": 0.8224, "step": 2875 }, { "epoch": 0.94, "grad_norm": 0.24219235648230453, "learning_rate": 2.8938273602561405e-07, "loss": 0.3834, "step": 2876 }, { "epoch": 0.94, "grad_norm": 0.32589393679068623, "learning_rate": 2.8628936793125506e-07, "loss": 0.8516, "step": 2877 }, { "epoch": 0.94, "grad_norm": 0.2557250288944782, "learning_rate": 2.832124628383664e-07, "loss": 0.5692, "step": 2878 }, { "epoch": 0.94, "grad_norm": 0.26027962618247524, "learning_rate": 2.8015202418965937e-07, "loss": 0.5633, "step": 2879 }, { "epoch": 0.94, "grad_norm": 0.32893239525463314, "learning_rate": 2.7710805540942794e-07, "loss": 0.8732, "step": 2880 }, { "epoch": 0.94, "grad_norm": 0.26219893169005815, "learning_rate": 2.740805599035312e-07, "loss": 0.5885, "step": 2881 }, { "epoch": 0.94, "grad_norm": 0.2548218388535008, "learning_rate": 2.710695410593994e-07, "loss": 0.554, "step": 2882 }, { "epoch": 0.94, "grad_norm": 0.33026906595940314, "learning_rate": 2.6807500224602763e-07, "loss": 0.8581, "step": 2883 }, { "epoch": 0.94, "grad_norm": 0.2569397194369273, "learning_rate": 2.6509694681397076e-07, "loss": 0.5646, "step": 2884 }, { "epoch": 0.94, "grad_norm": 0.253411261875266, "learning_rate": 2.621353780953401e-07, "loss": 0.5742, "step": 2885 }, { "epoch": 0.94, "grad_norm": 0.3239315651942836, "learning_rate": 2.5919029940380147e-07, "loss": 0.8562, "step": 2886 }, { "epoch": 0.94, "grad_norm": 0.24618798582225956, "learning_rate": 2.562617140345691e-07, "loss": 0.384, "step": 2887 }, { "epoch": 0.94, "grad_norm": 0.3281984827297079, "learning_rate": 2.5334962526440355e-07, "loss": 0.8569, "step": 2888 }, { "epoch": 0.94, "grad_norm": 0.25673353610034194, "learning_rate": 2.50454036351605e-07, "loss": 0.563, "step": 2889 }, { "epoch": 0.94, "grad_norm": 0.25379357476843994, "learning_rate": 2.4757495053601864e-07, "loss": 0.5485, "step": 2890 }, { "epoch": 0.94, "grad_norm": 0.3260778468400225, "learning_rate": 2.447123710390192e-07, "loss": 0.8278, "step": 2891 }, { "epoch": 0.94, "grad_norm": 0.315100936110326, "learning_rate": 2.418663010635114e-07, "loss": 0.6453, "step": 2892 }, { "epoch": 0.94, "grad_norm": 0.2846042985960293, "learning_rate": 2.3903674379393446e-07, "loss": 0.5931, "step": 2893 }, { "epoch": 0.95, "grad_norm": 0.25933395969368284, "learning_rate": 2.3622370239624436e-07, "loss": 0.5459, "step": 2894 }, { "epoch": 0.95, "grad_norm": 0.3127625699418851, "learning_rate": 2.334271800179233e-07, "loss": 0.8291, "step": 2895 }, { "epoch": 0.95, "grad_norm": 0.2305137035568352, "learning_rate": 2.3064717978796336e-07, "loss": 0.535, "step": 2896 }, { "epoch": 0.95, "grad_norm": 0.32033856644820397, "learning_rate": 2.278837048168797e-07, "loss": 0.8338, "step": 2897 }, { "epoch": 0.95, "grad_norm": 0.3099080136467589, "learning_rate": 2.25136758196689e-07, "loss": 0.607, "step": 2898 }, { "epoch": 0.95, "grad_norm": 0.26306643859618356, "learning_rate": 2.22406343000921e-07, "loss": 0.5807, "step": 2899 }, { "epoch": 0.95, "grad_norm": 0.26045900337245903, "learning_rate": 2.1969246228460526e-07, "loss": 0.5689, "step": 2900 }, { "epoch": 0.95, "grad_norm": 0.3225510972220235, "learning_rate": 2.1699511908427118e-07, "loss": 0.8189, "step": 2901 }, { "epoch": 0.95, "grad_norm": 0.2697709891256223, "learning_rate": 2.1431431641794287e-07, "loss": 0.5888, "step": 2902 }, { "epoch": 0.95, "grad_norm": 0.3061276599026969, "learning_rate": 2.1165005728514264e-07, "loss": 0.6476, "step": 2903 }, { "epoch": 0.95, "grad_norm": 0.25773285362155396, "learning_rate": 2.090023446668793e-07, "loss": 0.5853, "step": 2904 }, { "epoch": 0.95, "grad_norm": 0.2626016545953547, "learning_rate": 2.0637118152564637e-07, "loss": 0.5774, "step": 2905 }, { "epoch": 0.95, "grad_norm": 0.3334211944985216, "learning_rate": 2.0375657080542398e-07, "loss": 0.8516, "step": 2906 }, { "epoch": 0.95, "grad_norm": 0.2553467811145329, "learning_rate": 2.0115851543166863e-07, "loss": 0.555, "step": 2907 }, { "epoch": 0.95, "grad_norm": 0.2566488942697536, "learning_rate": 1.985770183113117e-07, "loss": 0.5659, "step": 2908 }, { "epoch": 0.95, "grad_norm": 0.3282725040255461, "learning_rate": 1.9601208233276603e-07, "loss": 0.8151, "step": 2909 }, { "epoch": 0.95, "grad_norm": 0.25766658278048843, "learning_rate": 1.934637103659076e-07, "loss": 0.5673, "step": 2910 }, { "epoch": 0.95, "grad_norm": 0.2640264032843368, "learning_rate": 1.9093190526207727e-07, "loss": 0.5982, "step": 2911 }, { "epoch": 0.95, "grad_norm": 0.3259331043991793, "learning_rate": 1.8841666985408566e-07, "loss": 0.839, "step": 2912 }, { "epoch": 0.95, "grad_norm": 0.2568924860675369, "learning_rate": 1.8591800695619832e-07, "loss": 0.5774, "step": 2913 }, { "epoch": 0.95, "grad_norm": 0.31012228688236243, "learning_rate": 1.8343591936414383e-07, "loss": 0.6092, "step": 2914 }, { "epoch": 0.95, "grad_norm": 0.2580693536983779, "learning_rate": 1.8097040985509904e-07, "loss": 0.5625, "step": 2915 }, { "epoch": 0.95, "grad_norm": 0.25370702985313826, "learning_rate": 1.7852148118769563e-07, "loss": 0.5419, "step": 2916 }, { "epoch": 0.95, "grad_norm": 0.32827060681343334, "learning_rate": 1.7608913610201006e-07, "loss": 0.8438, "step": 2917 }, { "epoch": 0.95, "grad_norm": 0.2571119811902067, "learning_rate": 1.7367337731956368e-07, "loss": 0.5721, "step": 2918 }, { "epoch": 0.95, "grad_norm": 0.2604705663623059, "learning_rate": 1.7127420754332601e-07, "loss": 0.5828, "step": 2919 }, { "epoch": 0.95, "grad_norm": 0.33973667688003417, "learning_rate": 1.6889162945769643e-07, "loss": 0.8196, "step": 2920 }, { "epoch": 0.95, "grad_norm": 0.26419705101019875, "learning_rate": 1.665256457285158e-07, "loss": 0.6137, "step": 2921 }, { "epoch": 0.95, "grad_norm": 0.2590956748927298, "learning_rate": 1.6417625900305656e-07, "loss": 0.5785, "step": 2922 }, { "epoch": 0.95, "grad_norm": 0.32572613659039235, "learning_rate": 1.6184347191001926e-07, "loss": 0.8677, "step": 2923 }, { "epoch": 0.95, "grad_norm": 0.2545566373722254, "learning_rate": 1.5952728705953268e-07, "loss": 0.562, "step": 2924 }, { "epoch": 0.96, "grad_norm": 0.3145098144624954, "learning_rate": 1.5722770704315048e-07, "loss": 0.6163, "step": 2925 }, { "epoch": 0.96, "grad_norm": 0.259165507674804, "learning_rate": 1.5494473443384615e-07, "loss": 0.5747, "step": 2926 }, { "epoch": 0.96, "grad_norm": 0.33330713868961487, "learning_rate": 1.5267837178600974e-07, "loss": 0.8528, "step": 2927 }, { "epoch": 0.96, "grad_norm": 0.26153899589654706, "learning_rate": 1.5042862163544947e-07, "loss": 0.5715, "step": 2928 }, { "epoch": 0.96, "grad_norm": 0.2565233574558806, "learning_rate": 1.4819548649938508e-07, "loss": 0.5649, "step": 2929 }, { "epoch": 0.96, "grad_norm": 0.2677235667052446, "learning_rate": 1.4597896887644458e-07, "loss": 0.5909, "step": 2930 }, { "epoch": 0.96, "grad_norm": 0.3363233966857075, "learning_rate": 1.4377907124666746e-07, "loss": 0.853, "step": 2931 }, { "epoch": 0.96, "grad_norm": 0.2603771054605109, "learning_rate": 1.4159579607148976e-07, "loss": 0.5758, "step": 2932 }, { "epoch": 0.96, "grad_norm": 0.26437883003434604, "learning_rate": 1.394291457937541e-07, "loss": 0.5859, "step": 2933 }, { "epoch": 0.96, "grad_norm": 0.3374249056968785, "learning_rate": 1.3727912283770295e-07, "loss": 0.867, "step": 2934 }, { "epoch": 0.96, "grad_norm": 0.2702393705926067, "learning_rate": 1.3514572960896865e-07, "loss": 0.5741, "step": 2935 }, { "epoch": 0.96, "grad_norm": 0.3099831800134539, "learning_rate": 1.3302896849458345e-07, "loss": 0.5884, "step": 2936 }, { "epoch": 0.96, "grad_norm": 0.25808209557775474, "learning_rate": 1.3092884186296283e-07, "loss": 0.5957, "step": 2937 }, { "epoch": 0.96, "grad_norm": 0.3311692173675386, "learning_rate": 1.2884535206392046e-07, "loss": 0.8297, "step": 2938 }, { "epoch": 0.96, "grad_norm": 0.26188855093649627, "learning_rate": 1.2677850142864324e-07, "loss": 0.6023, "step": 2939 }, { "epoch": 0.96, "grad_norm": 0.26242762174843265, "learning_rate": 1.2472829226970961e-07, "loss": 0.5849, "step": 2940 }, { "epoch": 0.96, "grad_norm": 0.3211670614294282, "learning_rate": 1.2269472688107463e-07, "loss": 0.8438, "step": 2941 }, { "epoch": 0.96, "grad_norm": 0.266880611892333, "learning_rate": 1.206778075380699e-07, "loss": 0.5817, "step": 2942 }, { "epoch": 0.96, "grad_norm": 0.26040431861252494, "learning_rate": 1.1867753649740687e-07, "loss": 0.581, "step": 2943 }, { "epoch": 0.96, "grad_norm": 0.33391053681386706, "learning_rate": 1.1669391599716028e-07, "loss": 0.8809, "step": 2944 }, { "epoch": 0.96, "grad_norm": 0.25432418410891133, "learning_rate": 1.1472694825678476e-07, "loss": 0.5778, "step": 2945 }, { "epoch": 0.96, "grad_norm": 0.2606458406065084, "learning_rate": 1.1277663547709649e-07, "loss": 0.5813, "step": 2946 }, { "epoch": 0.96, "grad_norm": 0.30716475499099377, "learning_rate": 1.1084297984027659e-07, "loss": 0.6319, "step": 2947 }, { "epoch": 0.96, "grad_norm": 0.3007574481220254, "learning_rate": 1.0892598350987271e-07, "loss": 0.5905, "step": 2948 }, { "epoch": 0.96, "grad_norm": 0.3255861153525361, "learning_rate": 1.0702564863078745e-07, "loss": 0.8595, "step": 2949 }, { "epoch": 0.96, "grad_norm": 0.2570199819820148, "learning_rate": 1.0514197732928665e-07, "loss": 0.587, "step": 2950 }, { "epoch": 0.96, "grad_norm": 0.2696853967509405, "learning_rate": 1.0327497171298439e-07, "loss": 0.5845, "step": 2951 }, { "epoch": 0.96, "grad_norm": 0.3275405086292251, "learning_rate": 1.0142463387085465e-07, "loss": 0.8062, "step": 2952 }, { "epoch": 0.96, "grad_norm": 0.26123349814131896, "learning_rate": 9.95909658732197e-08, "loss": 0.5702, "step": 2953 }, { "epoch": 0.96, "grad_norm": 0.2634422734997597, "learning_rate": 9.777396977174668e-08, "loss": 0.5855, "step": 2954 }, { "epoch": 0.97, "grad_norm": 0.3315929483548951, "learning_rate": 9.597364759945437e-08, "loss": 0.8604, "step": 2955 }, { "epoch": 0.97, "grad_norm": 0.25455056621034416, "learning_rate": 9.419000137070144e-08, "loss": 0.5576, "step": 2956 }, { "epoch": 0.97, "grad_norm": 0.2564055391213792, "learning_rate": 9.242303308118815e-08, "loss": 0.5741, "step": 2957 }, { "epoch": 0.97, "grad_norm": 0.31511829456947726, "learning_rate": 9.067274470795639e-08, "loss": 0.6165, "step": 2958 }, { "epoch": 0.97, "grad_norm": 0.2578086694354369, "learning_rate": 8.893913820938293e-08, "loss": 0.5854, "step": 2959 }, { "epoch": 0.97, "grad_norm": 0.32222405511421603, "learning_rate": 8.722221552518116e-08, "loss": 0.8283, "step": 2960 }, { "epoch": 0.97, "grad_norm": 0.2557396435140683, "learning_rate": 8.552197857639444e-08, "loss": 0.5781, "step": 2961 }, { "epoch": 0.97, "grad_norm": 0.2577527632133992, "learning_rate": 8.38384292653993e-08, "loss": 0.5848, "step": 2962 }, { "epoch": 0.97, "grad_norm": 0.32665260260729756, "learning_rate": 8.217156947590066e-08, "loss": 0.8441, "step": 2963 }, { "epoch": 0.97, "grad_norm": 0.2647686330540538, "learning_rate": 8.052140107292661e-08, "loss": 0.5776, "step": 2964 }, { "epoch": 0.97, "grad_norm": 0.25769073580283836, "learning_rate": 7.888792590283356e-08, "loss": 0.5959, "step": 2965 }, { "epoch": 0.97, "grad_norm": 0.32201554169655805, "learning_rate": 7.72711457932962e-08, "loss": 0.8188, "step": 2966 }, { "epoch": 0.97, "grad_norm": 0.25393183545186276, "learning_rate": 7.567106255331247e-08, "loss": 0.5694, "step": 2967 }, { "epoch": 0.97, "grad_norm": 0.25878967339860687, "learning_rate": 7.408767797319693e-08, "loss": 0.5902, "step": 2968 }, { "epoch": 0.97, "grad_norm": 0.3040649426616665, "learning_rate": 7.252099382458077e-08, "loss": 0.6186, "step": 2969 }, { "epoch": 0.97, "grad_norm": 0.32658673037518043, "learning_rate": 7.097101186040677e-08, "loss": 0.8317, "step": 2970 }, { "epoch": 0.97, "grad_norm": 0.2565823447843271, "learning_rate": 6.943773381493601e-08, "loss": 0.5764, "step": 2971 }, { "epoch": 0.97, "grad_norm": 0.34568353968099075, "learning_rate": 6.792116140373117e-08, "loss": 0.5949, "step": 2972 }, { "epoch": 0.97, "grad_norm": 0.32624328566325766, "learning_rate": 6.64212963236699e-08, "loss": 0.8425, "step": 2973 }, { "epoch": 0.97, "grad_norm": 0.25383592020522483, "learning_rate": 6.493814025293476e-08, "loss": 0.5822, "step": 2974 }, { "epoch": 0.97, "grad_norm": 0.30225789889671756, "learning_rate": 6.347169485100834e-08, "loss": 0.615, "step": 2975 }, { "epoch": 0.97, "grad_norm": 0.25438120578317863, "learning_rate": 6.202196175868479e-08, "loss": 0.5722, "step": 2976 }, { "epoch": 0.97, "grad_norm": 0.2587504713533989, "learning_rate": 6.058894259804992e-08, "loss": 0.5737, "step": 2977 }, { "epoch": 0.97, "grad_norm": 0.3238456427914919, "learning_rate": 5.917263897249281e-08, "loss": 0.8312, "step": 2978 }, { "epoch": 0.97, "grad_norm": 0.22348749649569893, "learning_rate": 5.7773052466700834e-08, "loss": 0.5502, "step": 2979 }, { "epoch": 0.97, "grad_norm": 0.3008927573702927, "learning_rate": 5.639018464665469e-08, "loss": 0.5766, "step": 2980 }, { "epoch": 0.97, "grad_norm": 0.3234697083206876, "learning_rate": 5.5024037059629993e-08, "loss": 0.8502, "step": 2981 }, { "epoch": 0.97, "grad_norm": 0.26484859044376086, "learning_rate": 5.367461123419071e-08, "loss": 0.5934, "step": 2982 }, { "epoch": 0.97, "grad_norm": 0.26107675592295354, "learning_rate": 5.234190868019573e-08, "loss": 0.5876, "step": 2983 }, { "epoch": 0.97, "grad_norm": 0.32731239820537233, "learning_rate": 5.1025930888790594e-08, "loss": 0.8703, "step": 2984 }, { "epoch": 0.97, "grad_norm": 0.25483376675824226, "learning_rate": 4.972667933240915e-08, "loss": 0.5562, "step": 2985 }, { "epoch": 0.98, "grad_norm": 0.31041103932011566, "learning_rate": 4.844415546476688e-08, "loss": 0.6265, "step": 2986 }, { "epoch": 0.98, "grad_norm": 0.6988992325755764, "learning_rate": 4.71783607208659e-08, "loss": 0.5533, "step": 2987 }, { "epoch": 0.98, "grad_norm": 0.2598621439634361, "learning_rate": 4.592929651698663e-08, "loss": 0.5632, "step": 2988 }, { "epoch": 0.98, "grad_norm": 0.31476062562289525, "learning_rate": 4.4696964250697805e-08, "loss": 0.8125, "step": 2989 }, { "epoch": 0.98, "grad_norm": 0.2587645655886903, "learning_rate": 4.348136530083813e-08, "loss": 0.5782, "step": 2990 }, { "epoch": 0.98, "grad_norm": 0.2627819747103884, "learning_rate": 4.228250102752962e-08, "loss": 0.5739, "step": 2991 }, { "epoch": 0.98, "grad_norm": 0.32316125117666966, "learning_rate": 4.110037277216427e-08, "loss": 0.8571, "step": 2992 }, { "epoch": 0.98, "grad_norm": 0.263178915800441, "learning_rate": 3.993498185741573e-08, "loss": 0.5763, "step": 2993 }, { "epoch": 0.98, "grad_norm": 0.25758644660259267, "learning_rate": 3.8786329587222614e-08, "loss": 0.5645, "step": 2994 }, { "epoch": 0.98, "grad_norm": 0.32344432499703984, "learning_rate": 3.7654417246801854e-08, "loss": 0.8409, "step": 2995 }, { "epoch": 0.98, "grad_norm": 0.2542229176994388, "learning_rate": 3.653924610263703e-08, "loss": 0.5844, "step": 2996 }, { "epoch": 0.98, "grad_norm": 0.31013663082918264, "learning_rate": 3.5440817402476735e-08, "loss": 0.604, "step": 2997 }, { "epoch": 0.98, "grad_norm": 0.26071750110753467, "learning_rate": 3.435913237534449e-08, "loss": 0.5902, "step": 2998 }, { "epoch": 0.98, "grad_norm": 0.3189353357211232, "learning_rate": 3.329419223152386e-08, "loss": 0.8017, "step": 2999 }, { "epoch": 0.98, "grad_norm": 1.3046828090738452, "learning_rate": 3.224599816256335e-08, "loss": 0.6077, "step": 3000 }, { "epoch": 0.98, "grad_norm": 0.25372503178980255, "learning_rate": 3.121455134127482e-08, "loss": 0.5731, "step": 3001 }, { "epoch": 0.98, "grad_norm": 0.26063455520934015, "learning_rate": 3.0199852921735104e-08, "loss": 0.5584, "step": 3002 }, { "epoch": 0.98, "grad_norm": 0.31986227774653675, "learning_rate": 2.920190403927436e-08, "loss": 0.8271, "step": 3003 }, { "epoch": 0.98, "grad_norm": 0.258119615530122, "learning_rate": 2.8220705810489412e-08, "loss": 0.5782, "step": 3004 }, { "epoch": 0.98, "grad_norm": 0.25362540766291997, "learning_rate": 2.7256259333230394e-08, "loss": 0.5823, "step": 3005 }, { "epoch": 0.98, "grad_norm": 0.3214189466882664, "learning_rate": 2.6308565686604114e-08, "loss": 0.8064, "step": 3006 }, { "epoch": 0.98, "grad_norm": 0.25456465123932526, "learning_rate": 2.5377625930977367e-08, "loss": 0.5606, "step": 3007 }, { "epoch": 0.98, "grad_norm": 0.30978192629158585, "learning_rate": 2.4463441107965278e-08, "loss": 0.6476, "step": 3008 }, { "epoch": 0.98, "grad_norm": 0.25829172096151937, "learning_rate": 2.3566012240439637e-08, "loss": 0.5572, "step": 3009 }, { "epoch": 0.98, "grad_norm": 0.3280426024126837, "learning_rate": 2.2685340332522232e-08, "loss": 0.8423, "step": 3010 }, { "epoch": 0.98, "grad_norm": 0.262933851834556, "learning_rate": 2.182142636958817e-08, "loss": 0.5879, "step": 3011 }, { "epoch": 0.98, "grad_norm": 0.25836853361119794, "learning_rate": 2.0974271318260907e-08, "loss": 0.5738, "step": 3012 }, { "epoch": 0.98, "grad_norm": 0.33278356703452416, "learning_rate": 2.0143876126412218e-08, "loss": 0.8083, "step": 3013 }, { "epoch": 0.98, "grad_norm": 0.2651356823927076, "learning_rate": 1.933024172316056e-08, "loss": 0.5964, "step": 3014 }, { "epoch": 0.98, "grad_norm": 0.2615960286421812, "learning_rate": 1.8533369018872704e-08, "loss": 0.5715, "step": 3015 }, { "epoch": 0.98, "grad_norm": 0.329008237694297, "learning_rate": 1.775325890516044e-08, "loss": 0.841, "step": 3016 }, { "epoch": 0.99, "grad_norm": 0.2602834621023953, "learning_rate": 1.698991225488056e-08, "loss": 0.564, "step": 3017 }, { "epoch": 0.99, "grad_norm": 0.25302192814544366, "learning_rate": 1.624332992213151e-08, "loss": 0.5742, "step": 3018 }, { "epoch": 0.99, "grad_norm": 0.31690815307222875, "learning_rate": 1.551351274225843e-08, "loss": 0.625, "step": 3019 }, { "epoch": 0.99, "grad_norm": 0.2593471043882813, "learning_rate": 1.4800461531841448e-08, "loss": 0.5892, "step": 3020 }, { "epoch": 0.99, "grad_norm": 0.3325597572816204, "learning_rate": 1.4104177088709036e-08, "loss": 0.8486, "step": 3021 }, { "epoch": 0.99, "grad_norm": 0.26034541233963443, "learning_rate": 1.342466019192301e-08, "loss": 0.567, "step": 3022 }, { "epoch": 0.99, "grad_norm": 0.2649158523991357, "learning_rate": 1.2761911601788524e-08, "loss": 0.6033, "step": 3023 }, { "epoch": 0.99, "grad_norm": 0.32030158516988405, "learning_rate": 1.2115932059847401e-08, "loss": 0.8593, "step": 3024 }, { "epoch": 0.99, "grad_norm": 0.5655216372000299, "learning_rate": 1.148672228887815e-08, "loss": 0.5899, "step": 3025 }, { "epoch": 0.99, "grad_norm": 0.2594487310611052, "learning_rate": 1.0874282992895946e-08, "loss": 0.5759, "step": 3026 }, { "epoch": 0.99, "grad_norm": 0.32746546391294806, "learning_rate": 1.0278614857154311e-08, "loss": 0.8419, "step": 3027 }, { "epoch": 0.99, "grad_norm": 0.2557523817520005, "learning_rate": 9.699718548136782e-09, "loss": 0.5749, "step": 3028 }, { "epoch": 0.99, "grad_norm": 0.27801307389693986, "learning_rate": 9.137594713563569e-09, "loss": 0.575, "step": 3029 }, { "epoch": 0.99, "grad_norm": 0.30068921710869256, "learning_rate": 8.592243982391557e-09, "loss": 0.6369, "step": 3030 }, { "epoch": 0.99, "grad_norm": 0.2571264883347178, "learning_rate": 8.063666964804317e-09, "loss": 0.5755, "step": 3031 }, { "epoch": 0.99, "grad_norm": 0.326247850447829, "learning_rate": 7.551864252223762e-09, "loss": 0.8447, "step": 3032 }, { "epoch": 0.99, "grad_norm": 0.2561155412281709, "learning_rate": 7.056836417296819e-09, "loss": 0.563, "step": 3033 }, { "epoch": 0.99, "grad_norm": 0.3321611256584422, "learning_rate": 6.578584013905431e-09, "loss": 0.5812, "step": 3034 }, { "epoch": 0.99, "grad_norm": 0.33362589120989933, "learning_rate": 6.117107577161551e-09, "loss": 0.8286, "step": 3035 }, { "epoch": 0.99, "grad_norm": 0.2578945470823204, "learning_rate": 5.672407623405484e-09, "loss": 0.5659, "step": 3036 }, { "epoch": 0.99, "grad_norm": 0.2854808469575818, "learning_rate": 5.244484650207548e-09, "loss": 0.5852, "step": 3037 }, { "epoch": 0.99, "grad_norm": 0.3255212649949795, "learning_rate": 4.83333913636308e-09, "loss": 0.8081, "step": 3038 }, { "epoch": 0.99, "grad_norm": 0.26159110083030285, "learning_rate": 4.4389715419024255e-09, "loss": 0.5872, "step": 3039 }, { "epoch": 0.99, "grad_norm": 0.2641786839518628, "learning_rate": 4.0613823080742905e-09, "loss": 0.581, "step": 3040 }, { "epoch": 0.99, "grad_norm": 0.31918250651497204, "learning_rate": 3.7005718573623892e-09, "loss": 0.6341, "step": 3041 }, { "epoch": 0.99, "grad_norm": 0.32405059260508784, "learning_rate": 3.3565405934721237e-09, "loss": 0.8267, "step": 3042 }, { "epoch": 0.99, "grad_norm": 0.2557068854580608, "learning_rate": 3.029288901337246e-09, "loss": 0.5779, "step": 3043 }, { "epoch": 0.99, "grad_norm": 0.25604376986730143, "learning_rate": 2.718817147113195e-09, "loss": 0.5614, "step": 3044 }, { "epoch": 0.99, "grad_norm": 0.25456024066003097, "learning_rate": 2.4251256781854247e-09, "loss": 0.5732, "step": 3045 }, { "epoch": 0.99, "grad_norm": 0.34250077982641713, "learning_rate": 2.148214823162742e-09, "loss": 0.8566, "step": 3046 }, { "epoch": 1.0, "grad_norm": 0.25390149283853686, "learning_rate": 1.8880848918739756e-09, "loss": 0.5504, "step": 3047 }, { "epoch": 1.0, "grad_norm": 0.26061427644168744, "learning_rate": 1.6447361753796352e-09, "loss": 0.5721, "step": 3048 }, { "epoch": 1.0, "grad_norm": 0.32629988695281265, "learning_rate": 1.4181689459569214e-09, "loss": 0.8501, "step": 3049 }, { "epoch": 1.0, "grad_norm": 0.26218442352716864, "learning_rate": 1.2083834571097185e-09, "loss": 0.5945, "step": 3050 }, { "epoch": 1.0, "grad_norm": 0.260229330406716, "learning_rate": 1.0153799435669298e-09, "loss": 0.5665, "step": 3051 }, { "epoch": 1.0, "grad_norm": 0.30881762469143403, "learning_rate": 8.391586212741498e-10, "loss": 0.6343, "step": 3052 }, { "epoch": 1.0, "grad_norm": 0.4231221784521575, "learning_rate": 6.797196874069878e-10, "loss": 0.8283, "step": 3053 }, { "epoch": 1.0, "grad_norm": 0.2578628422578347, "learning_rate": 5.370633203577447e-10, "loss": 0.5614, "step": 3054 }, { "epoch": 1.0, "grad_norm": 0.25897601122947544, "learning_rate": 4.111896797437398e-10, "loss": 0.5858, "step": 3055 }, { "epoch": 1.0, "grad_norm": 0.3280348115950767, "learning_rate": 3.0209890640231495e-10, "loss": 0.8377, "step": 3056 }, { "epoch": 1.0, "grad_norm": 0.24475625993804237, "learning_rate": 2.0979112239583042e-10, "loss": 0.389, "step": 3057 }, { "epoch": 1.0, "grad_norm": 0.3249026620501339, "learning_rate": 1.3426643100500347e-10, "loss": 0.8489, "step": 3058 }, { "epoch": 1.0, "grad_norm": 0.258428955188248, "learning_rate": 7.552491673390449e-11, "loss": 0.5652, "step": 3059 }, { "epoch": 1.0, "grad_norm": 0.25975550382546997, "learning_rate": 3.356664530829168e-11, "loss": 0.5875, "step": 3060 }, { "epoch": 1.0, "grad_norm": 0.322426859288248, "learning_rate": 8.391663673945616e-12, "loss": 0.8463, "step": 3061 }, { "epoch": 1.0, "grad_norm": 0.26632340692354883, "learning_rate": 0.0, "loss": 0.7333, "step": 3062 }, { "epoch": 1.0, "step": 3062, "total_flos": 0.0, "train_loss": 0.12032276010451015, "train_runtime": 5983.2516, "train_samples_per_second": 524.183, "train_steps_per_second": 0.512 } ], "logging_steps": 1.0, "max_steps": 3062, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": false, "should_training_stop": false }, "attributes": {} } }, "total_flos": 0.0, "train_batch_size": 8, "trial_name": null, "trial_params": null }