{ "best_metric": null, "best_model_checkpoint": null, "epoch": 2.959761549925484, "eval_steps": 168, "global_step": 2013, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0, "learning_rate": 1e-05, "loss": 1.3745, "step": 1 }, { "epoch": 0.0, "eval_loss": 1.6296857595443726, "eval_runtime": 2.6662, "eval_samples_per_second": 409.572, "eval_steps_per_second": 25.88, "step": 1 }, { "epoch": 0.0, "learning_rate": 2e-05, "loss": 1.42, "step": 2 }, { "epoch": 0.0, "learning_rate": 3e-05, "loss": 1.3057, "step": 3 }, { "epoch": 0.01, "learning_rate": 4e-05, "loss": 1.2307, "step": 4 }, { "epoch": 0.01, "learning_rate": 5e-05, "loss": 1.289, "step": 5 }, { "epoch": 0.01, "learning_rate": 6e-05, "loss": 1.4111, "step": 6 }, { "epoch": 0.01, "learning_rate": 7e-05, "loss": 1.3089, "step": 7 }, { "epoch": 0.01, "learning_rate": 8e-05, "loss": 1.3204, "step": 8 }, { "epoch": 0.01, "learning_rate": 9e-05, "loss": 1.3575, "step": 9 }, { "epoch": 0.01, "learning_rate": 0.0001, "loss": 1.3279, "step": 10 }, { "epoch": 0.02, "learning_rate": 0.00011000000000000002, "loss": 1.3149, "step": 11 }, { "epoch": 0.02, "learning_rate": 0.00012, "loss": 1.2578, "step": 12 }, { "epoch": 0.02, "learning_rate": 0.00013000000000000002, "loss": 1.2849, "step": 13 }, { "epoch": 0.02, "learning_rate": 0.00014, "loss": 1.2971, "step": 14 }, { "epoch": 0.02, "learning_rate": 0.00015000000000000001, "loss": 1.1473, "step": 15 }, { "epoch": 0.02, "learning_rate": 0.00016, "loss": 1.1943, "step": 16 }, { "epoch": 0.03, "learning_rate": 0.00017, "loss": 1.1877, "step": 17 }, { "epoch": 0.03, "learning_rate": 0.00018, "loss": 1.1984, "step": 18 }, { "epoch": 0.03, "learning_rate": 0.00019, "loss": 1.2647, "step": 19 }, { "epoch": 0.03, "learning_rate": 0.0002, "loss": 1.217, "step": 20 }, { "epoch": 0.03, "learning_rate": 0.00019999993046535236, "loss": 1.0274, "step": 21 }, { "epoch": 0.03, "learning_rate": 0.00019999972186150606, "loss": 1.2122, "step": 22 }, { "epoch": 0.03, "learning_rate": 0.00019999937418875124, "loss": 1.1868, "step": 23 }, { "epoch": 0.04, "learning_rate": 0.00019999888744757143, "loss": 1.2345, "step": 24 }, { "epoch": 0.04, "learning_rate": 0.00019999826163864348, "loss": 1.2127, "step": 25 }, { "epoch": 0.04, "learning_rate": 0.00019999749676283775, "loss": 1.2114, "step": 26 }, { "epoch": 0.04, "learning_rate": 0.00019999659282121792, "loss": 1.2224, "step": 27 }, { "epoch": 0.04, "learning_rate": 0.0001999955498150411, "loss": 1.1517, "step": 28 }, { "epoch": 0.04, "learning_rate": 0.0001999943677457578, "loss": 1.1631, "step": 29 }, { "epoch": 0.04, "learning_rate": 0.0001999930466150119, "loss": 1.0465, "step": 30 }, { "epoch": 0.05, "learning_rate": 0.0001999915864246407, "loss": 1.1847, "step": 31 }, { "epoch": 0.05, "learning_rate": 0.0001999899871766749, "loss": 1.1238, "step": 32 }, { "epoch": 0.05, "learning_rate": 0.0001999882488733385, "loss": 1.1491, "step": 33 }, { "epoch": 0.05, "learning_rate": 0.000199986371517049, "loss": 1.276, "step": 34 }, { "epoch": 0.05, "learning_rate": 0.0001999843551104172, "loss": 1.0911, "step": 35 }, { "epoch": 0.05, "learning_rate": 0.00019998219965624734, "loss": 1.1276, "step": 36 }, { "epoch": 0.06, "learning_rate": 0.00019997990515753693, "loss": 1.0981, "step": 37 }, { "epoch": 0.06, "learning_rate": 0.00019997747161747695, "loss": 1.0901, "step": 38 }, { "epoch": 0.06, "learning_rate": 0.0001999748990394517, "loss": 1.096, "step": 39 }, { "epoch": 0.06, "learning_rate": 0.00019997218742703887, "loss": 1.122, "step": 40 }, { "epoch": 0.06, "learning_rate": 0.00019996933678400946, "loss": 1.1132, "step": 41 }, { "epoch": 0.06, "learning_rate": 0.00019996634711432786, "loss": 1.1498, "step": 42 }, { "epoch": 0.06, "learning_rate": 0.00019996321842215173, "loss": 1.0708, "step": 43 }, { "epoch": 0.07, "learning_rate": 0.0001999599507118322, "loss": 1.1154, "step": 44 }, { "epoch": 0.07, "learning_rate": 0.00019995654398791355, "loss": 1.2118, "step": 45 }, { "epoch": 0.07, "learning_rate": 0.00019995299825513357, "loss": 1.0919, "step": 46 }, { "epoch": 0.07, "learning_rate": 0.00019994931351842327, "loss": 1.1364, "step": 47 }, { "epoch": 0.07, "learning_rate": 0.00019994548978290695, "loss": 1.1442, "step": 48 }, { "epoch": 0.07, "learning_rate": 0.0001999415270539023, "loss": 1.1248, "step": 49 }, { "epoch": 0.07, "learning_rate": 0.00019993742533692022, "loss": 1.1366, "step": 50 }, { "epoch": 0.08, "learning_rate": 0.00019993318463766495, "loss": 1.1437, "step": 51 }, { "epoch": 0.08, "learning_rate": 0.000199928804962034, "loss": 1.1191, "step": 52 }, { "epoch": 0.08, "learning_rate": 0.0001999242863161182, "loss": 1.0786, "step": 53 }, { "epoch": 0.08, "learning_rate": 0.00019991962870620153, "loss": 1.1951, "step": 54 }, { "epoch": 0.08, "learning_rate": 0.00019991483213876134, "loss": 1.1321, "step": 55 }, { "epoch": 0.08, "learning_rate": 0.00019990989662046818, "loss": 1.0876, "step": 56 }, { "epoch": 0.08, "learning_rate": 0.0001999048221581858, "loss": 1.1794, "step": 57 }, { "epoch": 0.09, "learning_rate": 0.00019989960875897126, "loss": 1.1796, "step": 58 }, { "epoch": 0.09, "learning_rate": 0.00019989425643007476, "loss": 1.1165, "step": 59 }, { "epoch": 0.09, "learning_rate": 0.0001998887651789398, "loss": 1.1978, "step": 60 }, { "epoch": 0.09, "learning_rate": 0.00019988313501320297, "loss": 1.1693, "step": 61 }, { "epoch": 0.09, "learning_rate": 0.00019987736594069414, "loss": 1.1553, "step": 62 }, { "epoch": 0.09, "learning_rate": 0.0001998714579694363, "loss": 1.1959, "step": 63 }, { "epoch": 0.1, "learning_rate": 0.00019986541110764565, "loss": 1.1945, "step": 64 }, { "epoch": 0.1, "learning_rate": 0.00019985922536373146, "loss": 1.121, "step": 65 }, { "epoch": 0.1, "learning_rate": 0.00019985290074629627, "loss": 1.122, "step": 66 }, { "epoch": 0.1, "learning_rate": 0.00019984643726413565, "loss": 1.1435, "step": 67 }, { "epoch": 0.1, "learning_rate": 0.00019983983492623833, "loss": 1.0413, "step": 68 }, { "epoch": 0.1, "learning_rate": 0.0001998330937417861, "loss": 1.078, "step": 69 }, { "epoch": 0.1, "learning_rate": 0.0001998262137201539, "loss": 1.0811, "step": 70 }, { "epoch": 0.11, "learning_rate": 0.00019981919487090972, "loss": 1.1639, "step": 71 }, { "epoch": 0.11, "learning_rate": 0.00019981203720381463, "loss": 1.164, "step": 72 }, { "epoch": 0.11, "learning_rate": 0.00019980474072882277, "loss": 1.1006, "step": 73 }, { "epoch": 0.11, "learning_rate": 0.00019979730545608126, "loss": 1.1926, "step": 74 }, { "epoch": 0.11, "learning_rate": 0.0001997897313959303, "loss": 1.1129, "step": 75 }, { "epoch": 0.11, "learning_rate": 0.00019978201855890308, "loss": 1.1367, "step": 76 }, { "epoch": 0.11, "learning_rate": 0.00019977416695572578, "loss": 1.1495, "step": 77 }, { "epoch": 0.12, "learning_rate": 0.0001997661765973176, "loss": 1.1567, "step": 78 }, { "epoch": 0.12, "learning_rate": 0.00019975804749479062, "loss": 1.2102, "step": 79 }, { "epoch": 0.12, "learning_rate": 0.00019974977965945, "loss": 1.1175, "step": 80 }, { "epoch": 0.12, "learning_rate": 0.0001997413731027937, "loss": 1.1243, "step": 81 }, { "epoch": 0.12, "learning_rate": 0.00019973282783651263, "loss": 1.1406, "step": 82 }, { "epoch": 0.12, "learning_rate": 0.00019972414387249072, "loss": 1.09, "step": 83 }, { "epoch": 0.13, "learning_rate": 0.00019971532122280464, "loss": 1.0115, "step": 84 }, { "epoch": 0.13, "learning_rate": 0.00019970635989972402, "loss": 1.0328, "step": 85 }, { "epoch": 0.13, "learning_rate": 0.00019969725991571128, "loss": 1.1226, "step": 86 }, { "epoch": 0.13, "learning_rate": 0.00019968802128342172, "loss": 1.0747, "step": 87 }, { "epoch": 0.13, "learning_rate": 0.00019967864401570343, "loss": 1.119, "step": 88 }, { "epoch": 0.13, "learning_rate": 0.00019966912812559732, "loss": 1.1125, "step": 89 }, { "epoch": 0.13, "learning_rate": 0.00019965947362633708, "loss": 1.0734, "step": 90 }, { "epoch": 0.14, "learning_rate": 0.0001996496805313491, "loss": 1.1798, "step": 91 }, { "epoch": 0.14, "learning_rate": 0.00019963974885425266, "loss": 1.1461, "step": 92 }, { "epoch": 0.14, "learning_rate": 0.0001996296786088596, "loss": 1.0397, "step": 93 }, { "epoch": 0.14, "learning_rate": 0.00019961946980917456, "loss": 1.17, "step": 94 }, { "epoch": 0.14, "learning_rate": 0.00019960912246939485, "loss": 1.0679, "step": 95 }, { "epoch": 0.14, "learning_rate": 0.00019959863660391045, "loss": 1.0839, "step": 96 }, { "epoch": 0.14, "learning_rate": 0.00019958801222730394, "loss": 1.0937, "step": 97 }, { "epoch": 0.15, "learning_rate": 0.00019957724935435063, "loss": 1.1668, "step": 98 }, { "epoch": 0.15, "learning_rate": 0.00019956634800001832, "loss": 1.0858, "step": 99 }, { "epoch": 0.15, "learning_rate": 0.00019955530817946748, "loss": 1.0935, "step": 100 }, { "epoch": 0.15, "learning_rate": 0.00019954412990805107, "loss": 1.1046, "step": 101 }, { "epoch": 0.15, "learning_rate": 0.00019953281320131468, "loss": 1.1319, "step": 102 }, { "epoch": 0.15, "learning_rate": 0.00019952135807499633, "loss": 1.1108, "step": 103 }, { "epoch": 0.15, "learning_rate": 0.0001995097645450266, "loss": 1.0485, "step": 104 }, { "epoch": 0.16, "learning_rate": 0.00019949803262752855, "loss": 1.0862, "step": 105 }, { "epoch": 0.16, "learning_rate": 0.00019948616233881768, "loss": 1.268, "step": 106 }, { "epoch": 0.16, "learning_rate": 0.00019947415369540189, "loss": 1.0926, "step": 107 }, { "epoch": 0.16, "learning_rate": 0.0001994620067139815, "loss": 1.1427, "step": 108 }, { "epoch": 0.16, "learning_rate": 0.00019944972141144928, "loss": 1.0754, "step": 109 }, { "epoch": 0.16, "learning_rate": 0.00019943729780489027, "loss": 1.0044, "step": 110 }, { "epoch": 0.17, "learning_rate": 0.0001994247359115819, "loss": 1.1304, "step": 111 }, { "epoch": 0.17, "learning_rate": 0.00019941203574899393, "loss": 1.1683, "step": 112 }, { "epoch": 0.17, "learning_rate": 0.00019939919733478838, "loss": 1.1559, "step": 113 }, { "epoch": 0.17, "learning_rate": 0.00019938622068681953, "loss": 1.1879, "step": 114 }, { "epoch": 0.17, "learning_rate": 0.00019937310582313392, "loss": 1.0613, "step": 115 }, { "epoch": 0.17, "learning_rate": 0.0001993598527619703, "loss": 1.1196, "step": 116 }, { "epoch": 0.17, "learning_rate": 0.0001993464615217596, "loss": 1.0762, "step": 117 }, { "epoch": 0.18, "learning_rate": 0.00019933293212112495, "loss": 1.1059, "step": 118 }, { "epoch": 0.18, "learning_rate": 0.00019931926457888156, "loss": 1.0831, "step": 119 }, { "epoch": 0.18, "learning_rate": 0.00019930545891403678, "loss": 1.0552, "step": 120 }, { "epoch": 0.18, "learning_rate": 0.00019929151514579008, "loss": 1.15, "step": 121 }, { "epoch": 0.18, "learning_rate": 0.00019927743329353295, "loss": 1.1038, "step": 122 }, { "epoch": 0.18, "learning_rate": 0.0001992632133768489, "loss": 1.067, "step": 123 }, { "epoch": 0.18, "learning_rate": 0.0001992488554155135, "loss": 1.1311, "step": 124 }, { "epoch": 0.19, "learning_rate": 0.00019923435942949426, "loss": 1.1402, "step": 125 }, { "epoch": 0.19, "learning_rate": 0.00019921972543895066, "loss": 1.0453, "step": 126 }, { "epoch": 0.19, "learning_rate": 0.00019920495346423402, "loss": 1.1567, "step": 127 }, { "epoch": 0.19, "learning_rate": 0.00019919004352588767, "loss": 1.137, "step": 128 }, { "epoch": 0.19, "learning_rate": 0.0001991749956446468, "loss": 0.9986, "step": 129 }, { "epoch": 0.19, "learning_rate": 0.00019915980984143832, "loss": 1.083, "step": 130 }, { "epoch": 0.2, "learning_rate": 0.00019914448613738106, "loss": 1.0619, "step": 131 }, { "epoch": 0.2, "learning_rate": 0.00019912902455378556, "loss": 1.1294, "step": 132 }, { "epoch": 0.2, "learning_rate": 0.00019911342511215414, "loss": 1.0965, "step": 133 }, { "epoch": 0.2, "learning_rate": 0.00019909768783418086, "loss": 1.0216, "step": 134 }, { "epoch": 0.2, "learning_rate": 0.00019908181274175138, "loss": 1.0081, "step": 135 }, { "epoch": 0.2, "learning_rate": 0.0001990657998569432, "loss": 1.0246, "step": 136 }, { "epoch": 0.2, "learning_rate": 0.0001990496492020252, "loss": 1.1249, "step": 137 }, { "epoch": 0.21, "learning_rate": 0.00019903336079945804, "loss": 1.0518, "step": 138 }, { "epoch": 0.21, "learning_rate": 0.00019901693467189386, "loss": 1.189, "step": 139 }, { "epoch": 0.21, "learning_rate": 0.00019900037084217637, "loss": 1.1475, "step": 140 }, { "epoch": 0.21, "learning_rate": 0.0001989836693333408, "loss": 1.2259, "step": 141 }, { "epoch": 0.21, "learning_rate": 0.0001989668301686138, "loss": 1.0399, "step": 142 }, { "epoch": 0.21, "learning_rate": 0.0001989498533714135, "loss": 1.128, "step": 143 }, { "epoch": 0.21, "learning_rate": 0.00019893273896534936, "loss": 1.014, "step": 144 }, { "epoch": 0.22, "learning_rate": 0.0001989154869742223, "loss": 1.1552, "step": 145 }, { "epoch": 0.22, "learning_rate": 0.00019889809742202455, "loss": 1.1159, "step": 146 }, { "epoch": 0.22, "learning_rate": 0.0001988805703329396, "loss": 1.0218, "step": 147 }, { "epoch": 0.22, "learning_rate": 0.00019886290573134228, "loss": 1.1723, "step": 148 }, { "epoch": 0.22, "learning_rate": 0.0001988451036417986, "loss": 1.2132, "step": 149 }, { "epoch": 0.22, "learning_rate": 0.00019882716408906585, "loss": 1.112, "step": 150 }, { "epoch": 0.23, "learning_rate": 0.0001988090870980924, "loss": 1.0856, "step": 151 }, { "epoch": 0.23, "learning_rate": 0.0001987908726940178, "loss": 1.0951, "step": 152 }, { "epoch": 0.23, "learning_rate": 0.00019877252090217271, "loss": 1.0218, "step": 153 }, { "epoch": 0.23, "learning_rate": 0.00019875403174807882, "loss": 1.0552, "step": 154 }, { "epoch": 0.23, "learning_rate": 0.00019873540525744887, "loss": 1.1481, "step": 155 }, { "epoch": 0.23, "learning_rate": 0.00019871664145618657, "loss": 1.169, "step": 156 }, { "epoch": 0.23, "learning_rate": 0.00019869774037038665, "loss": 1.0802, "step": 157 }, { "epoch": 0.24, "learning_rate": 0.0001986787020263347, "loss": 1.0871, "step": 158 }, { "epoch": 0.24, "learning_rate": 0.0001986595264505072, "loss": 1.1022, "step": 159 }, { "epoch": 0.24, "learning_rate": 0.00019864021366957147, "loss": 1.0257, "step": 160 }, { "epoch": 0.24, "learning_rate": 0.0001986207637103857, "loss": 1.0986, "step": 161 }, { "epoch": 0.24, "learning_rate": 0.00019860117659999878, "loss": 1.0837, "step": 162 }, { "epoch": 0.24, "learning_rate": 0.00019858145236565037, "loss": 1.1895, "step": 163 }, { "epoch": 0.24, "learning_rate": 0.00019856159103477086, "loss": 1.052, "step": 164 }, { "epoch": 0.25, "learning_rate": 0.00019854159263498123, "loss": 1.1184, "step": 165 }, { "epoch": 0.25, "learning_rate": 0.0001985214571940931, "loss": 1.0895, "step": 166 }, { "epoch": 0.25, "learning_rate": 0.00019850118474010872, "loss": 1.0764, "step": 167 }, { "epoch": 0.25, "learning_rate": 0.00019848077530122083, "loss": 1.1387, "step": 168 }, { "epoch": 0.25, "eval_loss": 1.084919810295105, "eval_runtime": 2.6029, "eval_samples_per_second": 419.538, "eval_steps_per_second": 26.509, "step": 168 }, { "epoch": 0.25, "learning_rate": 0.00019846022890581267, "loss": 1.0826, "step": 169 }, { "epoch": 0.25, "learning_rate": 0.000198439545582458, "loss": 1.1366, "step": 170 }, { "epoch": 0.25, "learning_rate": 0.000198418725359921, "loss": 1.1349, "step": 171 }, { "epoch": 0.26, "learning_rate": 0.00019839776826715614, "loss": 1.0636, "step": 172 }, { "epoch": 0.26, "learning_rate": 0.00019837667433330838, "loss": 1.1216, "step": 173 }, { "epoch": 0.26, "learning_rate": 0.0001983554435877128, "loss": 1.1051, "step": 174 }, { "epoch": 0.26, "learning_rate": 0.00019833407605989494, "loss": 1.1558, "step": 175 }, { "epoch": 0.26, "learning_rate": 0.00019831257177957044, "loss": 1.0364, "step": 176 }, { "epoch": 0.26, "learning_rate": 0.00019829093077664513, "loss": 1.0665, "step": 177 }, { "epoch": 0.27, "learning_rate": 0.00019826915308121504, "loss": 1.1994, "step": 178 }, { "epoch": 0.27, "learning_rate": 0.0001982472387235662, "loss": 1.1434, "step": 179 }, { "epoch": 0.27, "learning_rate": 0.0001982251877341748, "loss": 1.081, "step": 180 }, { "epoch": 0.27, "learning_rate": 0.000198203000143707, "loss": 1.0653, "step": 181 }, { "epoch": 0.27, "learning_rate": 0.0001981806759830189, "loss": 1.0269, "step": 182 }, { "epoch": 0.27, "learning_rate": 0.0001981582152831566, "loss": 1.1167, "step": 183 }, { "epoch": 0.27, "learning_rate": 0.00019813561807535598, "loss": 1.0608, "step": 184 }, { "epoch": 0.28, "learning_rate": 0.0001981128843910428, "loss": 1.0989, "step": 185 }, { "epoch": 0.28, "learning_rate": 0.0001980900142618327, "loss": 1.1405, "step": 186 }, { "epoch": 0.28, "learning_rate": 0.00019806700771953097, "loss": 1.0359, "step": 187 }, { "epoch": 0.28, "learning_rate": 0.0001980438647961327, "loss": 1.1073, "step": 188 }, { "epoch": 0.28, "learning_rate": 0.0001980205855238225, "loss": 1.0338, "step": 189 }, { "epoch": 0.28, "learning_rate": 0.00019799716993497475, "loss": 1.1285, "step": 190 }, { "epoch": 0.28, "learning_rate": 0.00019797361806215332, "loss": 1.1277, "step": 191 }, { "epoch": 0.29, "learning_rate": 0.00019794992993811165, "loss": 1.119, "step": 192 }, { "epoch": 0.29, "learning_rate": 0.00019792610559579265, "loss": 1.1224, "step": 193 }, { "epoch": 0.29, "learning_rate": 0.00019790214506832868, "loss": 1.1438, "step": 194 }, { "epoch": 0.29, "learning_rate": 0.0001978780483890414, "loss": 1.1462, "step": 195 }, { "epoch": 0.29, "learning_rate": 0.00019785381559144196, "loss": 1.042, "step": 196 }, { "epoch": 0.29, "learning_rate": 0.00019782944670923076, "loss": 1.1022, "step": 197 }, { "epoch": 0.3, "learning_rate": 0.00019780494177629735, "loss": 1.0564, "step": 198 }, { "epoch": 0.3, "learning_rate": 0.00019778030082672068, "loss": 1.0471, "step": 199 }, { "epoch": 0.3, "learning_rate": 0.00019775552389476864, "loss": 1.0636, "step": 200 }, { "epoch": 0.3, "learning_rate": 0.0001977306110148984, "loss": 1.0917, "step": 201 }, { "epoch": 0.3, "learning_rate": 0.00019770556222175608, "loss": 1.1965, "step": 202 }, { "epoch": 0.3, "learning_rate": 0.00019768037755017685, "loss": 1.073, "step": 203 }, { "epoch": 0.3, "learning_rate": 0.00019765505703518496, "loss": 1.0636, "step": 204 }, { "epoch": 0.31, "learning_rate": 0.00019762960071199333, "loss": 1.087, "step": 205 }, { "epoch": 0.31, "learning_rate": 0.000197604008616004, "loss": 1.0569, "step": 206 }, { "epoch": 0.31, "learning_rate": 0.00019757828078280766, "loss": 1.08, "step": 207 }, { "epoch": 0.31, "learning_rate": 0.00019755241724818387, "loss": 1.1536, "step": 208 }, { "epoch": 0.31, "learning_rate": 0.00019752641804810084, "loss": 1.1514, "step": 209 }, { "epoch": 0.31, "learning_rate": 0.00019750028321871546, "loss": 1.0691, "step": 210 }, { "epoch": 0.31, "learning_rate": 0.00019747401279637325, "loss": 1.1289, "step": 211 }, { "epoch": 0.32, "learning_rate": 0.00019744760681760832, "loss": 1.0834, "step": 212 }, { "epoch": 0.32, "learning_rate": 0.00019742106531914328, "loss": 1.0762, "step": 213 }, { "epoch": 0.32, "learning_rate": 0.0001973943883378892, "loss": 1.0913, "step": 214 }, { "epoch": 0.32, "learning_rate": 0.00019736757591094558, "loss": 1.132, "step": 215 }, { "epoch": 0.32, "learning_rate": 0.00019734062807560027, "loss": 1.0894, "step": 216 }, { "epoch": 0.32, "learning_rate": 0.00019731354486932944, "loss": 1.0327, "step": 217 }, { "epoch": 0.32, "learning_rate": 0.00019728632632979746, "loss": 1.112, "step": 218 }, { "epoch": 0.33, "learning_rate": 0.00019725897249485704, "loss": 1.0718, "step": 219 }, { "epoch": 0.33, "learning_rate": 0.00019723148340254892, "loss": 1.077, "step": 220 }, { "epoch": 0.33, "learning_rate": 0.00019720385909110198, "loss": 1.0335, "step": 221 }, { "epoch": 0.33, "learning_rate": 0.00019717609959893318, "loss": 1.0483, "step": 222 }, { "epoch": 0.33, "learning_rate": 0.00019714820496464746, "loss": 1.0901, "step": 223 }, { "epoch": 0.33, "learning_rate": 0.00019712017522703764, "loss": 0.9921, "step": 224 }, { "epoch": 0.34, "learning_rate": 0.00019709201042508455, "loss": 1.0829, "step": 225 }, { "epoch": 0.34, "learning_rate": 0.0001970637105979567, "loss": 1.0705, "step": 226 }, { "epoch": 0.34, "learning_rate": 0.0001970352757850105, "loss": 1.0481, "step": 227 }, { "epoch": 0.34, "learning_rate": 0.00019700670602579008, "loss": 0.9846, "step": 228 }, { "epoch": 0.34, "learning_rate": 0.0001969780013600272, "loss": 1.1492, "step": 229 }, { "epoch": 0.34, "learning_rate": 0.00019694916182764113, "loss": 1.1745, "step": 230 }, { "epoch": 0.34, "learning_rate": 0.00019692018746873892, "loss": 1.0451, "step": 231 }, { "epoch": 0.35, "learning_rate": 0.00019689107832361496, "loss": 1.1217, "step": 232 }, { "epoch": 0.35, "learning_rate": 0.00019686183443275116, "loss": 1.0788, "step": 233 }, { "epoch": 0.35, "learning_rate": 0.00019683245583681675, "loss": 1.0703, "step": 234 }, { "epoch": 0.35, "learning_rate": 0.00019680294257666837, "loss": 1.1521, "step": 235 }, { "epoch": 0.35, "learning_rate": 0.0001967732946933499, "loss": 1.0659, "step": 236 }, { "epoch": 0.35, "learning_rate": 0.00019674351222809242, "loss": 1.0321, "step": 237 }, { "epoch": 0.35, "learning_rate": 0.0001967135952223142, "loss": 1.0555, "step": 238 }, { "epoch": 0.36, "learning_rate": 0.00019668354371762066, "loss": 1.0648, "step": 239 }, { "epoch": 0.36, "learning_rate": 0.00019665335775580415, "loss": 1.0723, "step": 240 }, { "epoch": 0.36, "learning_rate": 0.0001966230373788441, "loss": 1.0264, "step": 241 }, { "epoch": 0.36, "learning_rate": 0.00019659258262890683, "loss": 1.0331, "step": 242 }, { "epoch": 0.36, "learning_rate": 0.00019656199354834558, "loss": 1.1514, "step": 243 }, { "epoch": 0.36, "learning_rate": 0.00019653127017970034, "loss": 1.069, "step": 244 }, { "epoch": 0.37, "learning_rate": 0.00019650041256569792, "loss": 0.9623, "step": 245 }, { "epoch": 0.37, "learning_rate": 0.00019646942074925172, "loss": 1.0021, "step": 246 }, { "epoch": 0.37, "learning_rate": 0.00019643829477346188, "loss": 1.1131, "step": 247 }, { "epoch": 0.37, "learning_rate": 0.0001964070346816151, "loss": 1.1426, "step": 248 }, { "epoch": 0.37, "learning_rate": 0.0001963756405171845, "loss": 1.0761, "step": 249 }, { "epoch": 0.37, "learning_rate": 0.00019634411232382978, "loss": 1.1112, "step": 250 }, { "epoch": 0.37, "learning_rate": 0.00019631245014539698, "loss": 1.081, "step": 251 }, { "epoch": 0.38, "learning_rate": 0.00019628065402591845, "loss": 1.1446, "step": 252 }, { "epoch": 0.38, "learning_rate": 0.00019624872400961284, "loss": 1.045, "step": 253 }, { "epoch": 0.38, "learning_rate": 0.00019621666014088494, "loss": 1.0337, "step": 254 }, { "epoch": 0.38, "learning_rate": 0.00019618446246432583, "loss": 1.1764, "step": 255 }, { "epoch": 0.38, "learning_rate": 0.00019615213102471257, "loss": 1.0323, "step": 256 }, { "epoch": 0.38, "learning_rate": 0.00019611966586700823, "loss": 1.0073, "step": 257 }, { "epoch": 0.38, "learning_rate": 0.00019608706703636188, "loss": 1.1615, "step": 258 }, { "epoch": 0.39, "learning_rate": 0.00019605433457810855, "loss": 1.1209, "step": 259 }, { "epoch": 0.39, "learning_rate": 0.00019602146853776894, "loss": 1.0721, "step": 260 }, { "epoch": 0.39, "learning_rate": 0.0001959884689610497, "loss": 1.0967, "step": 261 }, { "epoch": 0.39, "learning_rate": 0.00019595533589384308, "loss": 1.0284, "step": 262 }, { "epoch": 0.39, "learning_rate": 0.00019592206938222703, "loss": 1.0148, "step": 263 }, { "epoch": 0.39, "learning_rate": 0.00019588866947246498, "loss": 1.1434, "step": 264 }, { "epoch": 0.39, "learning_rate": 0.00019585513621100603, "loss": 1.1125, "step": 265 }, { "epoch": 0.4, "learning_rate": 0.0001958214696444846, "loss": 1.0812, "step": 266 }, { "epoch": 0.4, "learning_rate": 0.00019578766981972058, "loss": 1.0611, "step": 267 }, { "epoch": 0.4, "learning_rate": 0.00019575373678371909, "loss": 1.1029, "step": 268 }, { "epoch": 0.4, "learning_rate": 0.00019571967058367064, "loss": 1.0692, "step": 269 }, { "epoch": 0.4, "learning_rate": 0.00019568547126695083, "loss": 1.0581, "step": 270 }, { "epoch": 0.4, "learning_rate": 0.00019565113888112036, "loss": 0.9841, "step": 271 }, { "epoch": 0.41, "learning_rate": 0.00019561667347392508, "loss": 1.0173, "step": 272 }, { "epoch": 0.41, "learning_rate": 0.00019558207509329584, "loss": 1.0805, "step": 273 }, { "epoch": 0.41, "learning_rate": 0.00019554734378734824, "loss": 1.088, "step": 274 }, { "epoch": 0.41, "learning_rate": 0.00019551247960438296, "loss": 1.0481, "step": 275 }, { "epoch": 0.41, "learning_rate": 0.00019547748259288536, "loss": 1.1747, "step": 276 }, { "epoch": 0.41, "learning_rate": 0.0001954423528015255, "loss": 1.0407, "step": 277 }, { "epoch": 0.41, "learning_rate": 0.00019540709027915818, "loss": 1.1412, "step": 278 }, { "epoch": 0.42, "learning_rate": 0.0001953716950748227, "loss": 1.075, "step": 279 }, { "epoch": 0.42, "learning_rate": 0.00019533616723774294, "loss": 0.9863, "step": 280 }, { "epoch": 0.42, "learning_rate": 0.0001953005068173272, "loss": 1.1426, "step": 281 }, { "epoch": 0.42, "learning_rate": 0.0001952647138631682, "loss": 1.0621, "step": 282 }, { "epoch": 0.42, "learning_rate": 0.00019522878842504295, "loss": 1.1007, "step": 283 }, { "epoch": 0.42, "learning_rate": 0.00019519273055291266, "loss": 1.0632, "step": 284 }, { "epoch": 0.42, "learning_rate": 0.00019515654029692278, "loss": 1.126, "step": 285 }, { "epoch": 0.43, "learning_rate": 0.00019512021770740288, "loss": 1.0946, "step": 286 }, { "epoch": 0.43, "learning_rate": 0.0001950837628348665, "loss": 1.0639, "step": 287 }, { "epoch": 0.43, "learning_rate": 0.00019504717573001117, "loss": 1.1432, "step": 288 }, { "epoch": 0.43, "learning_rate": 0.00019501045644371832, "loss": 1.0619, "step": 289 }, { "epoch": 0.43, "learning_rate": 0.0001949736050270532, "loss": 1.0597, "step": 290 }, { "epoch": 0.43, "learning_rate": 0.00019493662153126481, "loss": 1.0743, "step": 291 }, { "epoch": 0.44, "learning_rate": 0.0001948995060077859, "loss": 1.1114, "step": 292 }, { "epoch": 0.44, "learning_rate": 0.00019486225850823266, "loss": 1.1435, "step": 293 }, { "epoch": 0.44, "learning_rate": 0.000194824879084405, "loss": 1.1396, "step": 294 }, { "epoch": 0.44, "learning_rate": 0.00019478736778828624, "loss": 1.1597, "step": 295 }, { "epoch": 0.44, "learning_rate": 0.00019474972467204297, "loss": 1.0976, "step": 296 }, { "epoch": 0.44, "learning_rate": 0.00019471194978802533, "loss": 1.0829, "step": 297 }, { "epoch": 0.44, "learning_rate": 0.0001946740431887665, "loss": 1.0437, "step": 298 }, { "epoch": 0.45, "learning_rate": 0.00019463600492698296, "loss": 1.0835, "step": 299 }, { "epoch": 0.45, "learning_rate": 0.00019459783505557424, "loss": 1.0558, "step": 300 }, { "epoch": 0.45, "learning_rate": 0.0001945595336276229, "loss": 1.0656, "step": 301 }, { "epoch": 0.45, "learning_rate": 0.00019452110069639452, "loss": 1.1487, "step": 302 }, { "epoch": 0.45, "learning_rate": 0.00019448253631533744, "loss": 1.1383, "step": 303 }, { "epoch": 0.45, "learning_rate": 0.00019444384053808288, "loss": 1.1582, "step": 304 }, { "epoch": 0.45, "learning_rate": 0.00019440501341844483, "loss": 0.9999, "step": 305 }, { "epoch": 0.46, "learning_rate": 0.00019436605501041987, "loss": 1.1317, "step": 306 }, { "epoch": 0.46, "learning_rate": 0.00019432696536818717, "loss": 1.0944, "step": 307 }, { "epoch": 0.46, "learning_rate": 0.00019428774454610843, "loss": 1.1624, "step": 308 }, { "epoch": 0.46, "learning_rate": 0.00019424839259872778, "loss": 1.1644, "step": 309 }, { "epoch": 0.46, "learning_rate": 0.00019420890958077167, "loss": 1.0486, "step": 310 }, { "epoch": 0.46, "learning_rate": 0.00019416929554714888, "loss": 1.0705, "step": 311 }, { "epoch": 0.46, "learning_rate": 0.00019412955055295034, "loss": 1.023, "step": 312 }, { "epoch": 0.47, "learning_rate": 0.00019408967465344917, "loss": 1.1144, "step": 313 }, { "epoch": 0.47, "learning_rate": 0.00019404966790410047, "loss": 1.0378, "step": 314 }, { "epoch": 0.47, "learning_rate": 0.00019400953036054138, "loss": 1.036, "step": 315 }, { "epoch": 0.47, "learning_rate": 0.00019396926207859084, "loss": 1.0735, "step": 316 }, { "epoch": 0.47, "learning_rate": 0.00019392886311424973, "loss": 1.0259, "step": 317 }, { "epoch": 0.47, "learning_rate": 0.0001938883335237006, "loss": 1.1603, "step": 318 }, { "epoch": 0.48, "learning_rate": 0.0001938476733633076, "loss": 1.1282, "step": 319 }, { "epoch": 0.48, "learning_rate": 0.0001938068826896166, "loss": 1.063, "step": 320 }, { "epoch": 0.48, "learning_rate": 0.00019376596155935486, "loss": 1.1176, "step": 321 }, { "epoch": 0.48, "learning_rate": 0.00019372491002943112, "loss": 1.1307, "step": 322 }, { "epoch": 0.48, "learning_rate": 0.00019368372815693549, "loss": 1.0412, "step": 323 }, { "epoch": 0.48, "learning_rate": 0.00019364241599913924, "loss": 1.1353, "step": 324 }, { "epoch": 0.48, "learning_rate": 0.00019360097361349494, "loss": 1.1293, "step": 325 }, { "epoch": 0.49, "learning_rate": 0.0001935594010576362, "loss": 1.0885, "step": 326 }, { "epoch": 0.49, "learning_rate": 0.00019351769838937775, "loss": 1.0944, "step": 327 }, { "epoch": 0.49, "learning_rate": 0.00019347586566671512, "loss": 1.1435, "step": 328 }, { "epoch": 0.49, "learning_rate": 0.0001934339029478248, "loss": 1.1217, "step": 329 }, { "epoch": 0.49, "learning_rate": 0.00019339181029106404, "loss": 1.1801, "step": 330 }, { "epoch": 0.49, "learning_rate": 0.00019334958775497083, "loss": 1.1846, "step": 331 }, { "epoch": 0.49, "learning_rate": 0.00019330723539826375, "loss": 1.0897, "step": 332 }, { "epoch": 0.5, "learning_rate": 0.00019326475327984192, "loss": 1.0643, "step": 333 }, { "epoch": 0.5, "learning_rate": 0.00019322214145878487, "loss": 1.0246, "step": 334 }, { "epoch": 0.5, "learning_rate": 0.0001931793999943526, "loss": 1.1108, "step": 335 }, { "epoch": 0.5, "learning_rate": 0.00019313652894598543, "loss": 1.0619, "step": 336 }, { "epoch": 0.5, "eval_loss": 1.048388123512268, "eval_runtime": 2.6045, "eval_samples_per_second": 419.273, "eval_steps_per_second": 26.493, "step": 336 }, { "epoch": 0.5, "learning_rate": 0.00019309352837330372, "loss": 1.0014, "step": 337 }, { "epoch": 0.5, "learning_rate": 0.0001930503983361081, "loss": 1.0786, "step": 338 }, { "epoch": 0.51, "learning_rate": 0.00019300713889437926, "loss": 1.014, "step": 339 }, { "epoch": 0.51, "learning_rate": 0.00019296375010827773, "loss": 1.1233, "step": 340 }, { "epoch": 0.51, "learning_rate": 0.000192920232038144, "loss": 1.1052, "step": 341 }, { "epoch": 0.51, "learning_rate": 0.0001928765847444984, "loss": 1.0138, "step": 342 }, { "epoch": 0.51, "learning_rate": 0.00019283280828804081, "loss": 1.1536, "step": 343 }, { "epoch": 0.51, "learning_rate": 0.00019278890272965096, "loss": 0.992, "step": 344 }, { "epoch": 0.51, "learning_rate": 0.0001927448681303879, "loss": 1.1165, "step": 345 }, { "epoch": 0.52, "learning_rate": 0.0001927007045514903, "loss": 1.0565, "step": 346 }, { "epoch": 0.52, "learning_rate": 0.00019265641205437611, "loss": 1.0664, "step": 347 }, { "epoch": 0.52, "learning_rate": 0.0001926119907006426, "loss": 1.0625, "step": 348 }, { "epoch": 0.52, "learning_rate": 0.00019256744055206622, "loss": 1.0393, "step": 349 }, { "epoch": 0.52, "learning_rate": 0.0001925227616706026, "loss": 1.125, "step": 350 }, { "epoch": 0.52, "learning_rate": 0.00019247795411838627, "loss": 1.0375, "step": 351 }, { "epoch": 0.52, "learning_rate": 0.00019243301795773086, "loss": 1.0648, "step": 352 }, { "epoch": 0.53, "learning_rate": 0.0001923879532511287, "loss": 1.0903, "step": 353 }, { "epoch": 0.53, "learning_rate": 0.000192342760061251, "loss": 1.1219, "step": 354 }, { "epoch": 0.53, "learning_rate": 0.00019229743845094755, "loss": 1.054, "step": 355 }, { "epoch": 0.53, "learning_rate": 0.0001922519884832469, "loss": 1.1206, "step": 356 }, { "epoch": 0.53, "learning_rate": 0.00019220641022135588, "loss": 1.1125, "step": 357 }, { "epoch": 0.53, "learning_rate": 0.00019216070372865996, "loss": 1.064, "step": 358 }, { "epoch": 0.54, "learning_rate": 0.0001921148690687228, "loss": 1.0843, "step": 359 }, { "epoch": 0.54, "learning_rate": 0.00019206890630528634, "loss": 1.1378, "step": 360 }, { "epoch": 0.54, "learning_rate": 0.00019202281550227064, "loss": 1.0399, "step": 361 }, { "epoch": 0.54, "learning_rate": 0.0001919765967237739, "loss": 1.1762, "step": 362 }, { "epoch": 0.54, "learning_rate": 0.0001919302500340722, "loss": 1.0538, "step": 363 }, { "epoch": 0.54, "learning_rate": 0.00019188377549761963, "loss": 1.0343, "step": 364 }, { "epoch": 0.54, "learning_rate": 0.0001918371731790479, "loss": 1.1027, "step": 365 }, { "epoch": 0.55, "learning_rate": 0.00019179044314316664, "loss": 1.036, "step": 366 }, { "epoch": 0.55, "learning_rate": 0.00019174358545496288, "loss": 1.041, "step": 367 }, { "epoch": 0.55, "learning_rate": 0.00019169660017960137, "loss": 1.0762, "step": 368 }, { "epoch": 0.55, "learning_rate": 0.00019164948738242409, "loss": 1.0807, "step": 369 }, { "epoch": 0.55, "learning_rate": 0.00019160224712895055, "loss": 1.037, "step": 370 }, { "epoch": 0.55, "learning_rate": 0.00019155487948487748, "loss": 1.0625, "step": 371 }, { "epoch": 0.55, "learning_rate": 0.0001915073845160786, "loss": 1.062, "step": 372 }, { "epoch": 0.56, "learning_rate": 0.00019145976228860496, "loss": 1.1882, "step": 373 }, { "epoch": 0.56, "learning_rate": 0.00019141201286868435, "loss": 1.1338, "step": 374 }, { "epoch": 0.56, "learning_rate": 0.00019136413632272163, "loss": 1.0174, "step": 375 }, { "epoch": 0.56, "learning_rate": 0.00019131613271729833, "loss": 1.0585, "step": 376 }, { "epoch": 0.56, "learning_rate": 0.00019126800211917276, "loss": 1.0495, "step": 377 }, { "epoch": 0.56, "learning_rate": 0.0001912197445952798, "loss": 1.123, "step": 378 }, { "epoch": 0.56, "learning_rate": 0.00019117136021273075, "loss": 1.0517, "step": 379 }, { "epoch": 0.57, "learning_rate": 0.0001911228490388136, "loss": 1.0545, "step": 380 }, { "epoch": 0.57, "learning_rate": 0.00019107421114099237, "loss": 1.0302, "step": 381 }, { "epoch": 0.57, "learning_rate": 0.00019102544658690748, "loss": 1.0908, "step": 382 }, { "epoch": 0.57, "learning_rate": 0.00019097655544437545, "loss": 1.1425, "step": 383 }, { "epoch": 0.57, "learning_rate": 0.00019092753778138886, "loss": 1.0686, "step": 384 }, { "epoch": 0.57, "learning_rate": 0.0001908783936661162, "loss": 1.06, "step": 385 }, { "epoch": 0.58, "learning_rate": 0.0001908291231669019, "loss": 1.1296, "step": 386 }, { "epoch": 0.58, "learning_rate": 0.00019077972635226604, "loss": 1.1029, "step": 387 }, { "epoch": 0.58, "learning_rate": 0.00019073020329090444, "loss": 1.0469, "step": 388 }, { "epoch": 0.58, "learning_rate": 0.0001906805540516885, "loss": 1.0427, "step": 389 }, { "epoch": 0.58, "learning_rate": 0.000190630778703665, "loss": 1.0075, "step": 390 }, { "epoch": 0.58, "learning_rate": 0.00019058087731605624, "loss": 1.1146, "step": 391 }, { "epoch": 0.58, "learning_rate": 0.0001905308499582597, "loss": 1.1161, "step": 392 }, { "epoch": 0.59, "learning_rate": 0.00019048069669984802, "loss": 1.1419, "step": 393 }, { "epoch": 0.59, "learning_rate": 0.00019043041761056907, "loss": 1.1586, "step": 394 }, { "epoch": 0.59, "learning_rate": 0.00019038001276034557, "loss": 1.0765, "step": 395 }, { "epoch": 0.59, "learning_rate": 0.00019032948221927524, "loss": 1.1225, "step": 396 }, { "epoch": 0.59, "learning_rate": 0.0001902788260576305, "loss": 1.0247, "step": 397 }, { "epoch": 0.59, "learning_rate": 0.00019022804434585852, "loss": 1.135, "step": 398 }, { "epoch": 0.59, "learning_rate": 0.0001901771371545811, "loss": 1.1122, "step": 399 }, { "epoch": 0.6, "learning_rate": 0.00019012610455459446, "loss": 1.075, "step": 400 }, { "epoch": 0.6, "learning_rate": 0.00019007494661686935, "loss": 1.1121, "step": 401 }, { "epoch": 0.6, "learning_rate": 0.0001900236634125507, "loss": 1.0531, "step": 402 }, { "epoch": 0.6, "learning_rate": 0.00018997225501295772, "loss": 1.0561, "step": 403 }, { "epoch": 0.6, "learning_rate": 0.00018992072148958368, "loss": 1.0803, "step": 404 }, { "epoch": 0.6, "learning_rate": 0.00018986906291409595, "loss": 1.0579, "step": 405 }, { "epoch": 0.61, "learning_rate": 0.00018981727935833567, "loss": 1.0614, "step": 406 }, { "epoch": 0.61, "learning_rate": 0.0001897653708943179, "loss": 0.9982, "step": 407 }, { "epoch": 0.61, "learning_rate": 0.00018971333759423142, "loss": 1.1498, "step": 408 }, { "epoch": 0.61, "learning_rate": 0.00018966117953043852, "loss": 1.1165, "step": 409 }, { "epoch": 0.61, "learning_rate": 0.00018960889677547505, "loss": 1.1155, "step": 410 }, { "epoch": 0.61, "learning_rate": 0.00018955648940205028, "loss": 1.0017, "step": 411 }, { "epoch": 0.61, "learning_rate": 0.00018950395748304678, "loss": 1.0556, "step": 412 }, { "epoch": 0.62, "learning_rate": 0.00018945130109152033, "loss": 1.0248, "step": 413 }, { "epoch": 0.62, "learning_rate": 0.00018939852030069981, "loss": 1.0155, "step": 414 }, { "epoch": 0.62, "learning_rate": 0.00018934561518398706, "loss": 1.0248, "step": 415 }, { "epoch": 0.62, "learning_rate": 0.00018929258581495685, "loss": 0.9835, "step": 416 }, { "epoch": 0.62, "learning_rate": 0.0001892394322673568, "loss": 1.1602, "step": 417 }, { "epoch": 0.62, "learning_rate": 0.0001891861546151071, "loss": 1.021, "step": 418 }, { "epoch": 0.62, "learning_rate": 0.00018913275293230069, "loss": 1.0526, "step": 419 }, { "epoch": 0.63, "learning_rate": 0.00018907922729320285, "loss": 1.0585, "step": 420 }, { "epoch": 0.63, "learning_rate": 0.00018902557777225135, "loss": 1.0327, "step": 421 }, { "epoch": 0.63, "learning_rate": 0.00018897180444405614, "loss": 1.0448, "step": 422 }, { "epoch": 0.63, "learning_rate": 0.0001889179073833995, "loss": 1.0776, "step": 423 }, { "epoch": 0.63, "learning_rate": 0.0001888638866652356, "loss": 1.0748, "step": 424 }, { "epoch": 0.63, "learning_rate": 0.0001888097423646907, "loss": 1.0482, "step": 425 }, { "epoch": 0.63, "learning_rate": 0.00018875547455706295, "loss": 1.0394, "step": 426 }, { "epoch": 0.64, "learning_rate": 0.00018870108331782217, "loss": 1.0646, "step": 427 }, { "epoch": 0.64, "learning_rate": 0.00018864656872260985, "loss": 1.0338, "step": 428 }, { "epoch": 0.64, "learning_rate": 0.00018859193084723913, "loss": 0.9848, "step": 429 }, { "epoch": 0.64, "learning_rate": 0.0001885371697676944, "loss": 1.0587, "step": 430 }, { "epoch": 0.64, "learning_rate": 0.0001884822855601316, "loss": 1.0711, "step": 431 }, { "epoch": 0.64, "learning_rate": 0.00018842727830087778, "loss": 1.0964, "step": 432 }, { "epoch": 0.65, "learning_rate": 0.00018837214806643115, "loss": 1.0254, "step": 433 }, { "epoch": 0.65, "learning_rate": 0.00018831689493346095, "loss": 1.0748, "step": 434 }, { "epoch": 0.65, "learning_rate": 0.00018826151897880728, "loss": 1.0797, "step": 435 }, { "epoch": 0.65, "learning_rate": 0.00018820602027948114, "loss": 1.1068, "step": 436 }, { "epoch": 0.65, "learning_rate": 0.00018815039891266418, "loss": 1.081, "step": 437 }, { "epoch": 0.65, "learning_rate": 0.0001880946549557086, "loss": 1.0685, "step": 438 }, { "epoch": 0.65, "learning_rate": 0.00018803878848613716, "loss": 1.0916, "step": 439 }, { "epoch": 0.66, "learning_rate": 0.00018798279958164295, "loss": 1.115, "step": 440 }, { "epoch": 0.66, "learning_rate": 0.00018792668832008936, "loss": 1.0048, "step": 441 }, { "epoch": 0.66, "learning_rate": 0.0001878704547795099, "loss": 1.0386, "step": 442 }, { "epoch": 0.66, "learning_rate": 0.00018781409903810821, "loss": 1.0283, "step": 443 }, { "epoch": 0.66, "learning_rate": 0.00018775762117425777, "loss": 1.085, "step": 444 }, { "epoch": 0.66, "learning_rate": 0.00018770102126650198, "loss": 1.0582, "step": 445 }, { "epoch": 0.66, "learning_rate": 0.00018764429939355392, "loss": 1.0705, "step": 446 }, { "epoch": 0.67, "learning_rate": 0.0001875874556342963, "loss": 1.1426, "step": 447 }, { "epoch": 0.67, "learning_rate": 0.00018753049006778132, "loss": 1.0337, "step": 448 }, { "epoch": 0.67, "learning_rate": 0.0001874734027732306, "loss": 1.0993, "step": 449 }, { "epoch": 0.67, "learning_rate": 0.00018741619383003507, "loss": 1.0661, "step": 450 }, { "epoch": 0.67, "learning_rate": 0.00018735886331775476, "loss": 1.0564, "step": 451 }, { "epoch": 0.67, "learning_rate": 0.00018730141131611882, "loss": 1.0989, "step": 452 }, { "epoch": 0.68, "learning_rate": 0.0001872438379050254, "loss": 1.0984, "step": 453 }, { "epoch": 0.68, "learning_rate": 0.00018718614316454133, "loss": 1.1173, "step": 454 }, { "epoch": 0.68, "learning_rate": 0.00018712832717490235, "loss": 1.1005, "step": 455 }, { "epoch": 0.68, "learning_rate": 0.00018707039001651277, "loss": 1.0008, "step": 456 }, { "epoch": 0.68, "learning_rate": 0.00018701233176994533, "loss": 1.0701, "step": 457 }, { "epoch": 0.68, "learning_rate": 0.00018695415251594123, "loss": 1.0831, "step": 458 }, { "epoch": 0.68, "learning_rate": 0.00018689585233541003, "loss": 1.1165, "step": 459 }, { "epoch": 0.69, "learning_rate": 0.00018683743130942928, "loss": 1.0884, "step": 460 }, { "epoch": 0.69, "learning_rate": 0.00018677888951924474, "loss": 0.9882, "step": 461 }, { "epoch": 0.69, "learning_rate": 0.00018672022704627002, "loss": 1.086, "step": 462 }, { "epoch": 0.69, "learning_rate": 0.00018666144397208668, "loss": 1.0545, "step": 463 }, { "epoch": 0.69, "learning_rate": 0.00018660254037844388, "loss": 1.0274, "step": 464 }, { "epoch": 0.69, "learning_rate": 0.0001865435163472584, "loss": 1.0795, "step": 465 }, { "epoch": 0.69, "learning_rate": 0.00018648437196061462, "loss": 1.022, "step": 466 }, { "epoch": 0.7, "learning_rate": 0.0001864251073007642, "loss": 1.0717, "step": 467 }, { "epoch": 0.7, "learning_rate": 0.00018636572245012606, "loss": 1.1501, "step": 468 }, { "epoch": 0.7, "learning_rate": 0.0001863062174912863, "loss": 1.1034, "step": 469 }, { "epoch": 0.7, "learning_rate": 0.00018624659250699805, "loss": 1.0784, "step": 470 }, { "epoch": 0.7, "learning_rate": 0.00018618684758018136, "loss": 1.1274, "step": 471 }, { "epoch": 0.7, "learning_rate": 0.0001861269827939231, "loss": 1.0643, "step": 472 }, { "epoch": 0.7, "learning_rate": 0.00018606699823147676, "loss": 1.1394, "step": 473 }, { "epoch": 0.71, "learning_rate": 0.00018600689397626246, "loss": 0.9665, "step": 474 }, { "epoch": 0.71, "learning_rate": 0.00018594667011186678, "loss": 1.058, "step": 475 }, { "epoch": 0.71, "learning_rate": 0.00018588632672204264, "loss": 1.0706, "step": 476 }, { "epoch": 0.71, "learning_rate": 0.0001858258638907091, "loss": 1.0414, "step": 477 }, { "epoch": 0.71, "learning_rate": 0.00018576528170195146, "loss": 1.1, "step": 478 }, { "epoch": 0.71, "learning_rate": 0.00018570458024002093, "loss": 1.1114, "step": 479 }, { "epoch": 0.72, "learning_rate": 0.00018564375958933459, "loss": 1.0596, "step": 480 }, { "epoch": 0.72, "learning_rate": 0.0001855828198344753, "loss": 1.0897, "step": 481 }, { "epoch": 0.72, "learning_rate": 0.00018552176106019155, "loss": 1.0316, "step": 482 }, { "epoch": 0.72, "learning_rate": 0.00018546058335139733, "loss": 1.0516, "step": 483 }, { "epoch": 0.72, "learning_rate": 0.0001853992867931721, "loss": 1.0477, "step": 484 }, { "epoch": 0.72, "learning_rate": 0.00018533787147076048, "loss": 1.0432, "step": 485 }, { "epoch": 0.72, "learning_rate": 0.00018527633746957234, "loss": 1.0568, "step": 486 }, { "epoch": 0.73, "learning_rate": 0.00018521468487518264, "loss": 1.114, "step": 487 }, { "epoch": 0.73, "learning_rate": 0.00018515291377333112, "loss": 1.0664, "step": 488 }, { "epoch": 0.73, "learning_rate": 0.0001850910242499225, "loss": 1.0162, "step": 489 }, { "epoch": 0.73, "learning_rate": 0.0001850290163910261, "loss": 1.0829, "step": 490 }, { "epoch": 0.73, "learning_rate": 0.00018496689028287572, "loss": 1.1078, "step": 491 }, { "epoch": 0.73, "learning_rate": 0.0001849046460118698, "loss": 1.0533, "step": 492 }, { "epoch": 0.73, "learning_rate": 0.00018484228366457095, "loss": 1.0923, "step": 493 }, { "epoch": 0.74, "learning_rate": 0.00018477980332770607, "loss": 1.0516, "step": 494 }, { "epoch": 0.74, "learning_rate": 0.00018471720508816614, "loss": 0.9826, "step": 495 }, { "epoch": 0.74, "learning_rate": 0.00018465448903300606, "loss": 1.1581, "step": 496 }, { "epoch": 0.74, "learning_rate": 0.0001845916552494446, "loss": 1.1268, "step": 497 }, { "epoch": 0.74, "learning_rate": 0.00018452870382486432, "loss": 1.0483, "step": 498 }, { "epoch": 0.74, "learning_rate": 0.00018446563484681127, "loss": 1.1792, "step": 499 }, { "epoch": 0.75, "learning_rate": 0.00018440244840299506, "loss": 1.0918, "step": 500 }, { "epoch": 0.75, "learning_rate": 0.0001843391445812886, "loss": 0.9691, "step": 501 }, { "epoch": 0.75, "learning_rate": 0.00018427572346972805, "loss": 1.0581, "step": 502 }, { "epoch": 0.75, "learning_rate": 0.0001842121851565128, "loss": 1.0072, "step": 503 }, { "epoch": 0.75, "learning_rate": 0.00018414852973000503, "loss": 0.9686, "step": 504 }, { "epoch": 0.75, "eval_loss": 1.0276715755462646, "eval_runtime": 2.6054, "eval_samples_per_second": 419.124, "eval_steps_per_second": 26.483, "step": 504 }, { "epoch": 0.75, "learning_rate": 0.00018408475727872995, "loss": 1.1221, "step": 505 }, { "epoch": 0.75, "learning_rate": 0.00018402086789137546, "loss": 1.087, "step": 506 }, { "epoch": 0.76, "learning_rate": 0.00018395686165679202, "loss": 1.0599, "step": 507 }, { "epoch": 0.76, "learning_rate": 0.00018389273866399275, "loss": 1.1844, "step": 508 }, { "epoch": 0.76, "learning_rate": 0.00018382849900215294, "loss": 1.046, "step": 509 }, { "epoch": 0.76, "learning_rate": 0.00018376414276061032, "loss": 0.9691, "step": 510 }, { "epoch": 0.76, "learning_rate": 0.00018369967002886464, "loss": 1.0996, "step": 511 }, { "epoch": 0.76, "learning_rate": 0.0001836350808965776, "loss": 1.083, "step": 512 }, { "epoch": 0.76, "learning_rate": 0.00018357037545357297, "loss": 1.0371, "step": 513 }, { "epoch": 0.77, "learning_rate": 0.00018350555378983608, "loss": 1.018, "step": 514 }, { "epoch": 0.77, "learning_rate": 0.00018344061599551398, "loss": 1.095, "step": 515 }, { "epoch": 0.77, "learning_rate": 0.00018337556216091517, "loss": 1.0871, "step": 516 }, { "epoch": 0.77, "learning_rate": 0.0001833103923765096, "loss": 1.0774, "step": 517 }, { "epoch": 0.77, "learning_rate": 0.00018324510673292842, "loss": 1.0337, "step": 518 }, { "epoch": 0.77, "learning_rate": 0.0001831797053209639, "loss": 1.0059, "step": 519 }, { "epoch": 0.77, "learning_rate": 0.00018311418823156936, "loss": 1.0744, "step": 520 }, { "epoch": 0.78, "learning_rate": 0.00018304855555585894, "loss": 0.9732, "step": 521 }, { "epoch": 0.78, "learning_rate": 0.00018298280738510752, "loss": 1.1176, "step": 522 }, { "epoch": 0.78, "learning_rate": 0.00018291694381075056, "loss": 1.1485, "step": 523 }, { "epoch": 0.78, "learning_rate": 0.00018285096492438424, "loss": 1.1044, "step": 524 }, { "epoch": 0.78, "learning_rate": 0.00018278487081776476, "loss": 0.9812, "step": 525 }, { "epoch": 0.78, "learning_rate": 0.00018271866158280884, "loss": 1.0966, "step": 526 }, { "epoch": 0.79, "learning_rate": 0.0001826523373115931, "loss": 1.2406, "step": 527 }, { "epoch": 0.79, "learning_rate": 0.0001825858980963543, "loss": 1.0727, "step": 528 }, { "epoch": 0.79, "learning_rate": 0.000182519344029489, "loss": 0.9966, "step": 529 }, { "epoch": 0.79, "learning_rate": 0.00018245267520355346, "loss": 1.081, "step": 530 }, { "epoch": 0.79, "learning_rate": 0.00018238589171126353, "loss": 1.1104, "step": 531 }, { "epoch": 0.79, "learning_rate": 0.00018231899364549455, "loss": 1.0535, "step": 532 }, { "epoch": 0.79, "learning_rate": 0.00018225198109928114, "loss": 1.0801, "step": 533 }, { "epoch": 0.8, "learning_rate": 0.00018218485416581726, "loss": 1.0726, "step": 534 }, { "epoch": 0.8, "learning_rate": 0.00018211761293845585, "loss": 1.0923, "step": 535 }, { "epoch": 0.8, "learning_rate": 0.00018205025751070875, "loss": 1.0551, "step": 536 }, { "epoch": 0.8, "learning_rate": 0.00018198278797624675, "loss": 1.0495, "step": 537 }, { "epoch": 0.8, "learning_rate": 0.0001819152044288992, "loss": 1.0589, "step": 538 }, { "epoch": 0.8, "learning_rate": 0.00018184750696265408, "loss": 1.0487, "step": 539 }, { "epoch": 0.8, "learning_rate": 0.0001817796956716578, "loss": 1.0491, "step": 540 }, { "epoch": 0.81, "learning_rate": 0.000181711770650215, "loss": 1.0981, "step": 541 }, { "epoch": 0.81, "learning_rate": 0.00018164373199278856, "loss": 1.1706, "step": 542 }, { "epoch": 0.81, "learning_rate": 0.0001815755797939994, "loss": 1.1024, "step": 543 }, { "epoch": 0.81, "learning_rate": 0.00018150731414862622, "loss": 1.0488, "step": 544 }, { "epoch": 0.81, "learning_rate": 0.00018143893515160564, "loss": 1.165, "step": 545 }, { "epoch": 0.81, "learning_rate": 0.00018137044289803181, "loss": 1.0346, "step": 546 }, { "epoch": 0.82, "learning_rate": 0.00018130183748315645, "loss": 1.1179, "step": 547 }, { "epoch": 0.82, "learning_rate": 0.0001812331190023886, "loss": 1.0027, "step": 548 }, { "epoch": 0.82, "learning_rate": 0.00018116428755129459, "loss": 1.1106, "step": 549 }, { "epoch": 0.82, "learning_rate": 0.00018109534322559783, "loss": 1.0479, "step": 550 }, { "epoch": 0.82, "learning_rate": 0.00018102628612117865, "loss": 1.0046, "step": 551 }, { "epoch": 0.82, "learning_rate": 0.0001809571163340744, "loss": 0.9883, "step": 552 }, { "epoch": 0.82, "learning_rate": 0.00018088783396047893, "loss": 1.1018, "step": 553 }, { "epoch": 0.83, "learning_rate": 0.00018081843909674276, "loss": 1.1389, "step": 554 }, { "epoch": 0.83, "learning_rate": 0.00018074893183937283, "loss": 1.0751, "step": 555 }, { "epoch": 0.83, "learning_rate": 0.00018067931228503246, "loss": 1.1475, "step": 556 }, { "epoch": 0.83, "learning_rate": 0.00018060958053054096, "loss": 1.0829, "step": 557 }, { "epoch": 0.83, "learning_rate": 0.00018053973667287387, "loss": 1.0272, "step": 558 }, { "epoch": 0.83, "learning_rate": 0.00018046978080916252, "loss": 1.0668, "step": 559 }, { "epoch": 0.83, "learning_rate": 0.00018039971303669407, "loss": 1.0988, "step": 560 }, { "epoch": 0.84, "learning_rate": 0.00018032953345291123, "loss": 1.0339, "step": 561 }, { "epoch": 0.84, "learning_rate": 0.0001802592421554123, "loss": 1.0654, "step": 562 }, { "epoch": 0.84, "learning_rate": 0.00018018883924195085, "loss": 1.0157, "step": 563 }, { "epoch": 0.84, "learning_rate": 0.00018011832481043576, "loss": 1.0738, "step": 564 }, { "epoch": 0.84, "learning_rate": 0.0001800476989589309, "loss": 1.0742, "step": 565 }, { "epoch": 0.84, "learning_rate": 0.0001799769617856552, "loss": 0.9861, "step": 566 }, { "epoch": 0.85, "learning_rate": 0.0001799061133889823, "loss": 1.0788, "step": 567 }, { "epoch": 0.85, "learning_rate": 0.00017983515386744061, "loss": 1.0539, "step": 568 }, { "epoch": 0.85, "learning_rate": 0.00017976408331971298, "loss": 1.0875, "step": 569 }, { "epoch": 0.85, "learning_rate": 0.0001796929018446368, "loss": 1.0765, "step": 570 }, { "epoch": 0.85, "learning_rate": 0.00017962160954120354, "loss": 1.1336, "step": 571 }, { "epoch": 0.85, "learning_rate": 0.000179550206508559, "loss": 0.9674, "step": 572 }, { "epoch": 0.85, "learning_rate": 0.00017947869284600282, "loss": 1.0607, "step": 573 }, { "epoch": 0.86, "learning_rate": 0.0001794070686529886, "loss": 0.9959, "step": 574 }, { "epoch": 0.86, "learning_rate": 0.00017933533402912354, "loss": 1.038, "step": 575 }, { "epoch": 0.86, "learning_rate": 0.0001792634890741685, "loss": 1.1342, "step": 576 }, { "epoch": 0.86, "learning_rate": 0.00017919153388803774, "loss": 1.0941, "step": 577 }, { "epoch": 0.86, "learning_rate": 0.00017911946857079888, "loss": 1.1286, "step": 578 }, { "epoch": 0.86, "learning_rate": 0.00017904729322267256, "loss": 1.0354, "step": 579 }, { "epoch": 0.86, "learning_rate": 0.0001789750079440326, "loss": 1.1314, "step": 580 }, { "epoch": 0.87, "learning_rate": 0.00017890261283540562, "loss": 1.0365, "step": 581 }, { "epoch": 0.87, "learning_rate": 0.00017883010799747099, "loss": 1.091, "step": 582 }, { "epoch": 0.87, "learning_rate": 0.00017875749353106062, "loss": 0.9995, "step": 583 }, { "epoch": 0.87, "learning_rate": 0.000178684769537159, "loss": 1.0435, "step": 584 }, { "epoch": 0.87, "learning_rate": 0.00017861193611690287, "loss": 1.0555, "step": 585 }, { "epoch": 0.87, "learning_rate": 0.00017853899337158112, "loss": 1.0637, "step": 586 }, { "epoch": 0.87, "learning_rate": 0.00017846594140263474, "loss": 1.064, "step": 587 }, { "epoch": 0.88, "learning_rate": 0.00017839278031165658, "loss": 0.9879, "step": 588 }, { "epoch": 0.88, "learning_rate": 0.00017831951020039126, "loss": 1.0846, "step": 589 }, { "epoch": 0.88, "learning_rate": 0.000178246131170735, "loss": 1.0373, "step": 590 }, { "epoch": 0.88, "learning_rate": 0.00017817264332473546, "loss": 1.0377, "step": 591 }, { "epoch": 0.88, "learning_rate": 0.00017809904676459177, "loss": 1.0932, "step": 592 }, { "epoch": 0.88, "learning_rate": 0.00017802534159265404, "loss": 1.085, "step": 593 }, { "epoch": 0.89, "learning_rate": 0.0001779515279114236, "loss": 1.0975, "step": 594 }, { "epoch": 0.89, "learning_rate": 0.0001778776058235526, "loss": 1.1283, "step": 595 }, { "epoch": 0.89, "learning_rate": 0.00017780357543184397, "loss": 1.0652, "step": 596 }, { "epoch": 0.89, "learning_rate": 0.00017772943683925122, "loss": 1.0336, "step": 597 }, { "epoch": 0.89, "learning_rate": 0.00017765519014887842, "loss": 0.9761, "step": 598 }, { "epoch": 0.89, "learning_rate": 0.0001775808354639799, "loss": 1.0688, "step": 599 }, { "epoch": 0.89, "learning_rate": 0.00017750637288796016, "loss": 1.1031, "step": 600 }, { "epoch": 0.9, "learning_rate": 0.00017743180252437383, "loss": 1.083, "step": 601 }, { "epoch": 0.9, "learning_rate": 0.00017735712447692538, "loss": 1.1612, "step": 602 }, { "epoch": 0.9, "learning_rate": 0.00017728233884946903, "loss": 1.1618, "step": 603 }, { "epoch": 0.9, "learning_rate": 0.00017720744574600863, "loss": 1.144, "step": 604 }, { "epoch": 0.9, "learning_rate": 0.0001771324452706975, "loss": 1.1174, "step": 605 }, { "epoch": 0.9, "learning_rate": 0.00017705733752783825, "loss": 0.9728, "step": 606 }, { "epoch": 0.9, "learning_rate": 0.0001769821226218827, "loss": 1.0599, "step": 607 }, { "epoch": 0.91, "learning_rate": 0.0001769068006574317, "loss": 1.0639, "step": 608 }, { "epoch": 0.91, "learning_rate": 0.00017683137173923495, "loss": 1.1278, "step": 609 }, { "epoch": 0.91, "learning_rate": 0.00017675583597219095, "loss": 0.9925, "step": 610 }, { "epoch": 0.91, "learning_rate": 0.0001766801934613467, "loss": 1.0457, "step": 611 }, { "epoch": 0.91, "learning_rate": 0.0001766044443118978, "loss": 1.0348, "step": 612 }, { "epoch": 0.91, "learning_rate": 0.000176528588629188, "loss": 1.022, "step": 613 }, { "epoch": 0.92, "learning_rate": 0.00017645262651870926, "loss": 1.0027, "step": 614 }, { "epoch": 0.92, "learning_rate": 0.00017637655808610156, "loss": 1.0491, "step": 615 }, { "epoch": 0.92, "learning_rate": 0.00017630038343715275, "loss": 1.0413, "step": 616 }, { "epoch": 0.92, "learning_rate": 0.00017622410267779834, "loss": 1.0358, "step": 617 }, { "epoch": 0.92, "learning_rate": 0.00017614771591412148, "loss": 1.1125, "step": 618 }, { "epoch": 0.92, "learning_rate": 0.00017607122325235267, "loss": 1.1185, "step": 619 }, { "epoch": 0.92, "learning_rate": 0.00017599462479886974, "loss": 1.0738, "step": 620 }, { "epoch": 0.93, "learning_rate": 0.00017591792066019765, "loss": 1.102, "step": 621 }, { "epoch": 0.93, "learning_rate": 0.00017584111094300827, "loss": 1.065, "step": 622 }, { "epoch": 0.93, "learning_rate": 0.0001757641957541203, "loss": 1.0514, "step": 623 }, { "epoch": 0.93, "learning_rate": 0.0001756871752004992, "loss": 1.0396, "step": 624 }, { "epoch": 0.93, "learning_rate": 0.00017561004938925688, "loss": 1.1027, "step": 625 }, { "epoch": 0.93, "learning_rate": 0.00017553281842765169, "loss": 1.0223, "step": 626 }, { "epoch": 0.93, "learning_rate": 0.00017545548242308816, "loss": 1.1793, "step": 627 }, { "epoch": 0.94, "learning_rate": 0.00017537804148311695, "loss": 1.0642, "step": 628 }, { "epoch": 0.94, "learning_rate": 0.00017530049571543464, "loss": 1.0682, "step": 629 }, { "epoch": 0.94, "learning_rate": 0.00017522284522788353, "loss": 1.0476, "step": 630 }, { "epoch": 0.94, "learning_rate": 0.00017514509012845164, "loss": 1.1064, "step": 631 }, { "epoch": 0.94, "learning_rate": 0.00017506723052527242, "loss": 1.0258, "step": 632 }, { "epoch": 0.94, "learning_rate": 0.00017498926652662476, "loss": 1.1954, "step": 633 }, { "epoch": 0.94, "learning_rate": 0.0001749111982409325, "loss": 1.0637, "step": 634 }, { "epoch": 0.95, "learning_rate": 0.00017483302577676475, "loss": 0.9685, "step": 635 }, { "epoch": 0.95, "learning_rate": 0.00017475474924283536, "loss": 1.0465, "step": 636 }, { "epoch": 0.95, "learning_rate": 0.000174676368748003, "loss": 1.0161, "step": 637 }, { "epoch": 0.95, "learning_rate": 0.00017459788440127083, "loss": 1.0479, "step": 638 }, { "epoch": 0.95, "learning_rate": 0.00017451929631178648, "loss": 1.1166, "step": 639 }, { "epoch": 0.95, "learning_rate": 0.0001744406045888419, "loss": 1.0634, "step": 640 }, { "epoch": 0.96, "learning_rate": 0.00017436180934187308, "loss": 1.0826, "step": 641 }, { "epoch": 0.96, "learning_rate": 0.00017428291068046, "loss": 1.07, "step": 642 }, { "epoch": 0.96, "learning_rate": 0.00017420390871432647, "loss": 1.1167, "step": 643 }, { "epoch": 0.96, "learning_rate": 0.00017412480355334005, "loss": 1.0347, "step": 644 }, { "epoch": 0.96, "learning_rate": 0.00017404559530751162, "loss": 1.0393, "step": 645 }, { "epoch": 0.96, "learning_rate": 0.00017396628408699555, "loss": 1.1108, "step": 646 }, { "epoch": 0.96, "learning_rate": 0.00017388687000208946, "loss": 1.006, "step": 647 }, { "epoch": 0.97, "learning_rate": 0.0001738073531632339, "loss": 1.0932, "step": 648 }, { "epoch": 0.97, "learning_rate": 0.0001737277336810124, "loss": 1.0123, "step": 649 }, { "epoch": 0.97, "learning_rate": 0.00017364801166615124, "loss": 1.1273, "step": 650 }, { "epoch": 0.97, "learning_rate": 0.0001735681872295192, "loss": 0.9893, "step": 651 }, { "epoch": 0.97, "learning_rate": 0.0001734882604821276, "loss": 1.0699, "step": 652 }, { "epoch": 0.97, "learning_rate": 0.00017340823153513002, "loss": 1.0901, "step": 653 }, { "epoch": 0.97, "learning_rate": 0.00017332810049982208, "loss": 1.0212, "step": 654 }, { "epoch": 0.98, "learning_rate": 0.00017324786748764155, "loss": 0.9898, "step": 655 }, { "epoch": 0.98, "learning_rate": 0.00017316753261016783, "loss": 1.0899, "step": 656 }, { "epoch": 0.98, "learning_rate": 0.00017308709597912213, "loss": 1.085, "step": 657 }, { "epoch": 0.98, "learning_rate": 0.00017300655770636708, "loss": 1.091, "step": 658 }, { "epoch": 0.98, "learning_rate": 0.00017292591790390665, "loss": 1.0502, "step": 659 }, { "epoch": 0.98, "learning_rate": 0.0001728451766838861, "loss": 1.2131, "step": 660 }, { "epoch": 0.99, "learning_rate": 0.00017276433415859167, "loss": 1.1256, "step": 661 }, { "epoch": 0.99, "learning_rate": 0.00017268339044045042, "loss": 1.0577, "step": 662 }, { "epoch": 0.99, "learning_rate": 0.00017260234564203032, "loss": 1.0012, "step": 663 }, { "epoch": 0.99, "learning_rate": 0.00017252119987603973, "loss": 1.0611, "step": 664 }, { "epoch": 0.99, "learning_rate": 0.00017243995325532755, "loss": 1.1251, "step": 665 }, { "epoch": 0.99, "learning_rate": 0.00017235860589288277, "loss": 1.0959, "step": 666 }, { "epoch": 0.99, "learning_rate": 0.0001722771579018347, "loss": 1.1413, "step": 667 }, { "epoch": 1.0, "learning_rate": 0.00017219560939545246, "loss": 1.0728, "step": 668 }, { "epoch": 1.0, "learning_rate": 0.00017211396048714498, "loss": 1.0461, "step": 669 }, { "epoch": 1.0, "learning_rate": 0.0001720322112904608, "loss": 1.1084, "step": 670 }, { "epoch": 1.0, "learning_rate": 0.00017195036191908797, "loss": 1.1316, "step": 671 }, { "epoch": 1.0, "learning_rate": 0.00017186841248685383, "loss": 1.0816, "step": 672 }, { "epoch": 1.0, "eval_loss": 1.0170178413391113, "eval_runtime": 2.6119, "eval_samples_per_second": 418.079, "eval_steps_per_second": 26.417, "step": 672 }, { "epoch": 1.0, "learning_rate": 0.0001717863631077249, "loss": 1.0711, "step": 673 }, { "epoch": 1.0, "learning_rate": 0.00017170421389580667, "loss": 1.1245, "step": 674 }, { "epoch": 1.01, "learning_rate": 0.00017162196496534342, "loss": 1.0519, "step": 675 }, { "epoch": 1.01, "learning_rate": 0.0001715396164307182, "loss": 1.104, "step": 676 }, { "epoch": 1.01, "learning_rate": 0.00017145716840645254, "loss": 1.1193, "step": 677 }, { "epoch": 1.01, "learning_rate": 0.00017137462100720631, "loss": 1.1238, "step": 678 }, { "epoch": 1.01, "learning_rate": 0.00017129197434777763, "loss": 1.004, "step": 679 }, { "epoch": 1.01, "learning_rate": 0.00017120922854310257, "loss": 1.0426, "step": 680 }, { "epoch": 1.01, "learning_rate": 0.00017112638370825515, "loss": 1.0308, "step": 681 }, { "epoch": 1.02, "learning_rate": 0.00017104343995844715, "loss": 1.0892, "step": 682 }, { "epoch": 1.02, "learning_rate": 0.00017096039740902784, "loss": 1.0115, "step": 683 }, { "epoch": 1.02, "learning_rate": 0.00017087725617548385, "loss": 1.1011, "step": 684 }, { "epoch": 1.0, "learning_rate": 0.00017079401637343914, "loss": 0.9829, "step": 685 }, { "epoch": 1.0, "learning_rate": 0.00017071067811865476, "loss": 0.9738, "step": 686 }, { "epoch": 1.0, "learning_rate": 0.0001706272415270286, "loss": 1.0563, "step": 687 }, { "epoch": 1.01, "learning_rate": 0.00017054370671459532, "loss": 1.0153, "step": 688 }, { "epoch": 1.01, "learning_rate": 0.0001704600737975262, "loss": 1.0638, "step": 689 }, { "epoch": 1.01, "learning_rate": 0.000170376342892129, "loss": 1.0053, "step": 690 }, { "epoch": 1.01, "learning_rate": 0.00017029251411484765, "loss": 1.0178, "step": 691 }, { "epoch": 1.01, "learning_rate": 0.00017020858758226229, "loss": 1.0755, "step": 692 }, { "epoch": 1.01, "learning_rate": 0.00017012456341108885, "loss": 0.9365, "step": 693 }, { "epoch": 1.01, "learning_rate": 0.00017004044171817925, "loss": 1.0666, "step": 694 }, { "epoch": 1.02, "learning_rate": 0.00016995622262052092, "loss": 1.041, "step": 695 }, { "epoch": 1.02, "learning_rate": 0.00016987190623523674, "loss": 1.0387, "step": 696 }, { "epoch": 1.02, "learning_rate": 0.00016978749267958495, "loss": 0.9332, "step": 697 }, { "epoch": 1.02, "learning_rate": 0.00016970298207095885, "loss": 1.0737, "step": 698 }, { "epoch": 1.02, "learning_rate": 0.00016961837452688676, "loss": 0.992, "step": 699 }, { "epoch": 1.02, "learning_rate": 0.00016953367016503182, "loss": 0.9997, "step": 700 }, { "epoch": 1.03, "learning_rate": 0.00016944886910319173, "loss": 1.1054, "step": 701 }, { "epoch": 1.03, "learning_rate": 0.00016936397145929878, "loss": 0.9876, "step": 702 }, { "epoch": 1.03, "learning_rate": 0.00016927897735141952, "loss": 1.0158, "step": 703 }, { "epoch": 1.03, "learning_rate": 0.00016919388689775464, "loss": 0.9771, "step": 704 }, { "epoch": 1.03, "learning_rate": 0.00016910870021663883, "loss": 0.942, "step": 705 }, { "epoch": 1.03, "learning_rate": 0.00016902341742654065, "loss": 1.0217, "step": 706 }, { "epoch": 1.03, "learning_rate": 0.00016893803864606222, "loss": 1.0346, "step": 707 }, { "epoch": 1.04, "learning_rate": 0.00016885256399393924, "loss": 0.9891, "step": 708 }, { "epoch": 1.04, "learning_rate": 0.00016876699358904068, "loss": 0.9697, "step": 709 }, { "epoch": 1.04, "learning_rate": 0.00016868132755036875, "loss": 1.0062, "step": 710 }, { "epoch": 1.04, "learning_rate": 0.00016859556599705856, "loss": 0.9822, "step": 711 }, { "epoch": 1.04, "learning_rate": 0.0001685097090483781, "loss": 1.0921, "step": 712 }, { "epoch": 1.04, "learning_rate": 0.00016842375682372805, "loss": 1.0126, "step": 713 }, { "epoch": 1.04, "learning_rate": 0.00016833770944264153, "loss": 1.0043, "step": 714 }, { "epoch": 1.05, "learning_rate": 0.00016825156702478407, "loss": 0.952, "step": 715 }, { "epoch": 1.05, "learning_rate": 0.00016816532968995328, "loss": 1.0423, "step": 716 }, { "epoch": 1.05, "learning_rate": 0.00016807899755807886, "loss": 1.0465, "step": 717 }, { "epoch": 1.05, "learning_rate": 0.00016799257074922224, "loss": 0.9827, "step": 718 }, { "epoch": 1.05, "learning_rate": 0.00016790604938357663, "loss": 0.9798, "step": 719 }, { "epoch": 1.05, "learning_rate": 0.00016781943358146664, "loss": 1.0268, "step": 720 }, { "epoch": 1.06, "learning_rate": 0.00016773272346334828, "loss": 1.0007, "step": 721 }, { "epoch": 1.06, "learning_rate": 0.0001676459191498087, "loss": 0.9989, "step": 722 }, { "epoch": 1.06, "learning_rate": 0.00016755902076156604, "loss": 0.9374, "step": 723 }, { "epoch": 1.06, "learning_rate": 0.00016747202841946928, "loss": 1.0031, "step": 724 }, { "epoch": 1.06, "learning_rate": 0.00016738494224449802, "loss": 0.9751, "step": 725 }, { "epoch": 1.06, "learning_rate": 0.00016729776235776246, "loss": 1.1055, "step": 726 }, { "epoch": 1.06, "learning_rate": 0.00016721048888050302, "loss": 1.0527, "step": 727 }, { "epoch": 1.07, "learning_rate": 0.0001671231219340903, "loss": 1.0048, "step": 728 }, { "epoch": 1.07, "learning_rate": 0.0001670356616400249, "loss": 0.957, "step": 729 }, { "epoch": 1.07, "learning_rate": 0.00016694810811993723, "loss": 1.0598, "step": 730 }, { "epoch": 1.07, "learning_rate": 0.00016686046149558736, "loss": 1.02, "step": 731 }, { "epoch": 1.07, "learning_rate": 0.00016677272188886483, "loss": 0.9973, "step": 732 }, { "epoch": 1.07, "learning_rate": 0.00016668488942178856, "loss": 1.0685, "step": 733 }, { "epoch": 1.07, "learning_rate": 0.00016659696421650645, "loss": 0.9783, "step": 734 }, { "epoch": 1.08, "learning_rate": 0.00016650894639529544, "loss": 0.9767, "step": 735 }, { "epoch": 1.08, "learning_rate": 0.00016642083608056141, "loss": 1.0192, "step": 736 }, { "epoch": 1.08, "learning_rate": 0.00016633263339483866, "loss": 1.0121, "step": 737 }, { "epoch": 1.08, "learning_rate": 0.00016624433846079012, "loss": 0.9817, "step": 738 }, { "epoch": 1.08, "learning_rate": 0.00016615595140120686, "loss": 1.1145, "step": 739 }, { "epoch": 1.08, "learning_rate": 0.00016606747233900815, "loss": 0.9862, "step": 740 }, { "epoch": 1.08, "learning_rate": 0.00016597890139724125, "loss": 1.0606, "step": 741 }, { "epoch": 1.09, "learning_rate": 0.0001658902386990811, "loss": 1.0416, "step": 742 }, { "epoch": 1.09, "learning_rate": 0.0001658014843678303, "loss": 0.9971, "step": 743 }, { "epoch": 1.09, "learning_rate": 0.00016571263852691888, "loss": 1.0318, "step": 744 }, { "epoch": 1.09, "learning_rate": 0.0001656237012999041, "loss": 1.0633, "step": 745 }, { "epoch": 1.09, "learning_rate": 0.0001655346728104704, "loss": 1.0418, "step": 746 }, { "epoch": 1.09, "learning_rate": 0.00016544555318242897, "loss": 0.9308, "step": 747 }, { "epoch": 1.1, "learning_rate": 0.00016535634253971794, "loss": 1.1049, "step": 748 }, { "epoch": 1.1, "learning_rate": 0.0001652670410064019, "loss": 0.9377, "step": 749 }, { "epoch": 1.1, "learning_rate": 0.00016517764870667182, "loss": 0.9934, "step": 750 }, { "epoch": 1.1, "learning_rate": 0.000165088165764845, "loss": 1.0467, "step": 751 }, { "epoch": 1.1, "learning_rate": 0.00016499859230536466, "loss": 1.0172, "step": 752 }, { "epoch": 1.1, "learning_rate": 0.0001649089284528001, "loss": 0.9922, "step": 753 }, { "epoch": 1.1, "learning_rate": 0.00016481917433184607, "loss": 1.0373, "step": 754 }, { "epoch": 1.11, "learning_rate": 0.0001647293300673231, "loss": 1.0377, "step": 755 }, { "epoch": 1.11, "learning_rate": 0.00016463939578417692, "loss": 0.9991, "step": 756 }, { "epoch": 1.11, "learning_rate": 0.00016454937160747854, "loss": 1.0657, "step": 757 }, { "epoch": 1.11, "learning_rate": 0.00016445925766242391, "loss": 0.9954, "step": 758 }, { "epoch": 1.11, "learning_rate": 0.0001643690540743339, "loss": 1.018, "step": 759 }, { "epoch": 1.11, "learning_rate": 0.00016427876096865394, "loss": 1.01, "step": 760 }, { "epoch": 1.11, "learning_rate": 0.0001641883784709541, "loss": 0.9318, "step": 761 }, { "epoch": 1.12, "learning_rate": 0.0001640979067069286, "loss": 1.0174, "step": 762 }, { "epoch": 1.12, "learning_rate": 0.00016400734580239594, "loss": 1.0886, "step": 763 }, { "epoch": 1.12, "learning_rate": 0.0001639166958832985, "loss": 1.0316, "step": 764 }, { "epoch": 1.12, "learning_rate": 0.0001638259570757025, "loss": 1.0514, "step": 765 }, { "epoch": 1.12, "learning_rate": 0.0001637351295057978, "loss": 0.9914, "step": 766 }, { "epoch": 1.12, "learning_rate": 0.00016364421329989755, "loss": 1.0529, "step": 767 }, { "epoch": 1.13, "learning_rate": 0.00016355320858443842, "loss": 0.9689, "step": 768 }, { "epoch": 1.13, "learning_rate": 0.00016346211548597995, "loss": 1.0398, "step": 769 }, { "epoch": 1.13, "learning_rate": 0.0001633709341312046, "loss": 1.0127, "step": 770 }, { "epoch": 1.13, "learning_rate": 0.00016327966464691778, "loss": 1.1388, "step": 771 }, { "epoch": 1.13, "learning_rate": 0.00016318830716004722, "loss": 0.9659, "step": 772 }, { "epoch": 1.13, "learning_rate": 0.00016309686179764317, "loss": 0.9907, "step": 773 }, { "epoch": 1.13, "learning_rate": 0.00016300532868687806, "loss": 0.9168, "step": 774 }, { "epoch": 1.14, "learning_rate": 0.0001629137079550463, "loss": 1.06, "step": 775 }, { "epoch": 1.14, "learning_rate": 0.00016282199972956425, "loss": 0.9826, "step": 776 }, { "epoch": 1.14, "learning_rate": 0.00016273020413796983, "loss": 1.0496, "step": 777 }, { "epoch": 1.14, "learning_rate": 0.0001626383213079226, "loss": 1.0245, "step": 778 }, { "epoch": 1.14, "learning_rate": 0.00016254635136720328, "loss": 1.036, "step": 779 }, { "epoch": 1.14, "learning_rate": 0.0001624542944437139, "loss": 1.0283, "step": 780 }, { "epoch": 1.14, "learning_rate": 0.00016236215066547734, "loss": 1.0078, "step": 781 }, { "epoch": 1.15, "learning_rate": 0.00016226992016063723, "loss": 0.9819, "step": 782 }, { "epoch": 1.15, "learning_rate": 0.00016217760305745803, "loss": 1.0687, "step": 783 }, { "epoch": 1.15, "learning_rate": 0.0001620851994843244, "loss": 1.0523, "step": 784 }, { "epoch": 1.15, "learning_rate": 0.00016199270956974128, "loss": 1.0279, "step": 785 }, { "epoch": 1.15, "learning_rate": 0.00016190013344233388, "loss": 1.0559, "step": 786 }, { "epoch": 1.15, "learning_rate": 0.00016180747123084705, "loss": 1.0844, "step": 787 }, { "epoch": 1.15, "learning_rate": 0.00016171472306414554, "loss": 1.0724, "step": 788 }, { "epoch": 1.16, "learning_rate": 0.00016162188907121354, "loss": 0.9696, "step": 789 }, { "epoch": 1.16, "learning_rate": 0.00016152896938115464, "loss": 0.9551, "step": 790 }, { "epoch": 1.16, "learning_rate": 0.0001614359641231916, "loss": 1.0032, "step": 791 }, { "epoch": 1.16, "learning_rate": 0.0001613428734266662, "loss": 1.1404, "step": 792 }, { "epoch": 1.16, "learning_rate": 0.00016124969742103897, "loss": 1.0329, "step": 793 }, { "epoch": 1.16, "learning_rate": 0.00016115643623588915, "loss": 1.039, "step": 794 }, { "epoch": 1.17, "learning_rate": 0.0001610630900009144, "loss": 1.0231, "step": 795 }, { "epoch": 1.17, "learning_rate": 0.0001609696588459307, "loss": 1.0659, "step": 796 }, { "epoch": 1.17, "learning_rate": 0.00016087614290087208, "loss": 1.0029, "step": 797 }, { "epoch": 1.17, "learning_rate": 0.0001607825422957905, "loss": 0.985, "step": 798 }, { "epoch": 1.17, "learning_rate": 0.00016068885716085567, "loss": 0.9392, "step": 799 }, { "epoch": 1.17, "learning_rate": 0.00016059508762635482, "loss": 1.006, "step": 800 }, { "epoch": 1.17, "learning_rate": 0.00016050123382269264, "loss": 1.0748, "step": 801 }, { "epoch": 1.18, "learning_rate": 0.0001604072958803909, "loss": 1.1378, "step": 802 }, { "epoch": 1.18, "learning_rate": 0.00016031327393008845, "loss": 1.058, "step": 803 }, { "epoch": 1.18, "learning_rate": 0.00016021916810254097, "loss": 0.9827, "step": 804 }, { "epoch": 1.18, "learning_rate": 0.00016012497852862075, "loss": 0.9572, "step": 805 }, { "epoch": 1.18, "learning_rate": 0.00016003070533931657, "loss": 1.0042, "step": 806 }, { "epoch": 1.18, "learning_rate": 0.00015993634866573347, "loss": 0.9521, "step": 807 }, { "epoch": 1.18, "learning_rate": 0.0001598419086390927, "loss": 0.9395, "step": 808 }, { "epoch": 1.19, "learning_rate": 0.00015974738539073125, "loss": 1.0902, "step": 809 }, { "epoch": 1.19, "learning_rate": 0.00015965277905210195, "loss": 1.0408, "step": 810 }, { "epoch": 1.19, "learning_rate": 0.00015955808975477319, "loss": 1.0436, "step": 811 }, { "epoch": 1.19, "learning_rate": 0.00015946331763042867, "loss": 1.0845, "step": 812 }, { "epoch": 1.19, "learning_rate": 0.00015936846281086736, "loss": 1.0752, "step": 813 }, { "epoch": 1.19, "learning_rate": 0.00015927352542800317, "loss": 1.0832, "step": 814 }, { "epoch": 1.2, "learning_rate": 0.00015917850561386488, "loss": 0.9901, "step": 815 }, { "epoch": 1.2, "learning_rate": 0.00015908340350059583, "loss": 1.0311, "step": 816 }, { "epoch": 1.2, "learning_rate": 0.00015898821922045385, "loss": 0.9858, "step": 817 }, { "epoch": 1.2, "learning_rate": 0.0001588929529058111, "loss": 0.9541, "step": 818 }, { "epoch": 1.2, "learning_rate": 0.00015879760468915372, "loss": 0.9516, "step": 819 }, { "epoch": 1.2, "learning_rate": 0.00015870217470308188, "loss": 1.0791, "step": 820 }, { "epoch": 1.2, "learning_rate": 0.00015860666308030932, "loss": 0.9099, "step": 821 }, { "epoch": 1.21, "learning_rate": 0.00015851106995366337, "loss": 1.0983, "step": 822 }, { "epoch": 1.21, "learning_rate": 0.00015841539545608478, "loss": 0.9951, "step": 823 }, { "epoch": 1.21, "learning_rate": 0.00015831963972062733, "loss": 0.9661, "step": 824 }, { "epoch": 1.21, "learning_rate": 0.00015822380288045792, "loss": 1.0111, "step": 825 }, { "epoch": 1.21, "learning_rate": 0.0001581278850688561, "loss": 1.0436, "step": 826 }, { "epoch": 1.21, "learning_rate": 0.00015803188641921417, "loss": 1.0916, "step": 827 }, { "epoch": 1.21, "learning_rate": 0.0001579358070650367, "loss": 1.0347, "step": 828 }, { "epoch": 1.22, "learning_rate": 0.0001578396471399406, "loss": 0.9978, "step": 829 }, { "epoch": 1.22, "learning_rate": 0.0001577434067776548, "loss": 1.0036, "step": 830 }, { "epoch": 1.22, "learning_rate": 0.00015764708611202015, "loss": 1.0387, "step": 831 }, { "epoch": 1.22, "learning_rate": 0.00015755068527698902, "loss": 1.0172, "step": 832 }, { "epoch": 1.22, "learning_rate": 0.00015745420440662543, "loss": 0.9723, "step": 833 }, { "epoch": 1.22, "learning_rate": 0.0001573576436351046, "loss": 0.9662, "step": 834 }, { "epoch": 1.23, "learning_rate": 0.000157261003096713, "loss": 0.9849, "step": 835 }, { "epoch": 1.23, "learning_rate": 0.00015716428292584787, "loss": 1.0198, "step": 836 }, { "epoch": 1.23, "learning_rate": 0.00015706748325701732, "loss": 0.9015, "step": 837 }, { "epoch": 1.23, "learning_rate": 0.0001569706042248399, "loss": 1.001, "step": 838 }, { "epoch": 1.23, "learning_rate": 0.0001568736459640447, "loss": 0.9681, "step": 839 }, { "epoch": 1.23, "learning_rate": 0.00015677660860947078, "loss": 1.0513, "step": 840 }, { "epoch": 1.23, "eval_loss": 1.008791446685791, "eval_runtime": 2.6133, "eval_samples_per_second": 417.867, "eval_steps_per_second": 26.404, "step": 840 }, { "epoch": 1.23, "learning_rate": 0.0001566794922960674, "loss": 0.9829, "step": 841 }, { "epoch": 1.24, "learning_rate": 0.00015658229715889347, "loss": 1.0362, "step": 842 }, { "epoch": 1.24, "learning_rate": 0.00015648502333311757, "loss": 0.9736, "step": 843 }, { "epoch": 1.24, "learning_rate": 0.0001563876709540178, "loss": 1.0457, "step": 844 }, { "epoch": 1.24, "learning_rate": 0.00015629024015698136, "loss": 0.9786, "step": 845 }, { "epoch": 1.24, "learning_rate": 0.00015619273107750462, "loss": 1.0595, "step": 846 }, { "epoch": 1.24, "learning_rate": 0.0001560951438511927, "loss": 1.0307, "step": 847 }, { "epoch": 1.24, "learning_rate": 0.00015599747861375955, "loss": 1.0386, "step": 848 }, { "epoch": 1.25, "learning_rate": 0.00015589973550102747, "loss": 0.9916, "step": 849 }, { "epoch": 1.25, "learning_rate": 0.00015580191464892716, "loss": 0.9652, "step": 850 }, { "epoch": 1.25, "learning_rate": 0.00015570401619349736, "loss": 0.9691, "step": 851 }, { "epoch": 1.25, "learning_rate": 0.00015560604027088477, "loss": 1.0006, "step": 852 }, { "epoch": 1.25, "learning_rate": 0.00015550798701734385, "loss": 1.0271, "step": 853 }, { "epoch": 1.25, "learning_rate": 0.00015540985656923645, "loss": 1.0591, "step": 854 }, { "epoch": 1.25, "learning_rate": 0.00015531164906303207, "loss": 0.967, "step": 855 }, { "epoch": 1.26, "learning_rate": 0.00015521336463530705, "loss": 1.0466, "step": 856 }, { "epoch": 1.26, "learning_rate": 0.0001551150034227449, "loss": 0.9953, "step": 857 }, { "epoch": 1.26, "learning_rate": 0.0001550165655621359, "loss": 0.9899, "step": 858 }, { "epoch": 1.26, "learning_rate": 0.00015491805119037684, "loss": 0.9742, "step": 859 }, { "epoch": 1.26, "learning_rate": 0.00015481946044447099, "loss": 0.9865, "step": 860 }, { "epoch": 1.26, "learning_rate": 0.0001547207934615278, "loss": 0.9384, "step": 861 }, { "epoch": 1.27, "learning_rate": 0.00015462205037876275, "loss": 1.0216, "step": 862 }, { "epoch": 1.27, "learning_rate": 0.00015452323133349714, "loss": 0.9467, "step": 863 }, { "epoch": 1.27, "learning_rate": 0.0001544243364631579, "loss": 1.0038, "step": 864 }, { "epoch": 1.27, "learning_rate": 0.0001543253659052775, "loss": 0.978, "step": 865 }, { "epoch": 1.27, "learning_rate": 0.00015422631979749354, "loss": 1.0434, "step": 866 }, { "epoch": 1.27, "learning_rate": 0.00015412719827754873, "loss": 1.0091, "step": 867 }, { "epoch": 1.27, "learning_rate": 0.00015402800148329071, "loss": 0.9598, "step": 868 }, { "epoch": 1.28, "learning_rate": 0.00015392872955267175, "loss": 0.9876, "step": 869 }, { "epoch": 1.28, "learning_rate": 0.00015382938262374865, "loss": 0.9559, "step": 870 }, { "epoch": 1.28, "learning_rate": 0.0001537299608346824, "loss": 0.9984, "step": 871 }, { "epoch": 1.28, "learning_rate": 0.00015363046432373824, "loss": 1.0171, "step": 872 }, { "epoch": 1.28, "learning_rate": 0.0001535308932292853, "loss": 1.0188, "step": 873 }, { "epoch": 1.28, "learning_rate": 0.00015343124768979637, "loss": 0.9613, "step": 874 }, { "epoch": 1.28, "learning_rate": 0.00015333152784384777, "loss": 0.9572, "step": 875 }, { "epoch": 1.29, "learning_rate": 0.0001532317338301192, "loss": 1.0093, "step": 876 }, { "epoch": 1.29, "learning_rate": 0.00015313186578739353, "loss": 0.9935, "step": 877 }, { "epoch": 1.29, "learning_rate": 0.0001530319238545565, "loss": 1.0371, "step": 878 }, { "epoch": 1.29, "learning_rate": 0.00015293190817059667, "loss": 1.0022, "step": 879 }, { "epoch": 1.29, "learning_rate": 0.00015283181887460517, "loss": 1.0033, "step": 880 }, { "epoch": 1.29, "learning_rate": 0.00015273165610577542, "loss": 0.9986, "step": 881 }, { "epoch": 1.3, "learning_rate": 0.00015263142000340312, "loss": 1.0495, "step": 882 }, { "epoch": 1.3, "learning_rate": 0.0001525311107068859, "loss": 1.017, "step": 883 }, { "epoch": 1.3, "learning_rate": 0.00015243072835572318, "loss": 0.9757, "step": 884 }, { "epoch": 1.3, "learning_rate": 0.000152330273089516, "loss": 1.0342, "step": 885 }, { "epoch": 1.3, "learning_rate": 0.0001522297450479668, "loss": 1.0059, "step": 886 }, { "epoch": 1.3, "learning_rate": 0.00015212914437087922, "loss": 0.9845, "step": 887 }, { "epoch": 1.3, "learning_rate": 0.0001520284711981579, "loss": 1.0365, "step": 888 }, { "epoch": 1.31, "learning_rate": 0.0001519277256698083, "loss": 0.9521, "step": 889 }, { "epoch": 1.31, "learning_rate": 0.0001518269079259366, "loss": 1.0867, "step": 890 }, { "epoch": 1.31, "learning_rate": 0.00015172601810674915, "loss": 1.0444, "step": 891 }, { "epoch": 1.31, "learning_rate": 0.00015162505635255287, "loss": 1.077, "step": 892 }, { "epoch": 1.31, "learning_rate": 0.00015152402280375454, "loss": 0.9883, "step": 893 }, { "epoch": 1.31, "learning_rate": 0.0001514229176008607, "loss": 0.9819, "step": 894 }, { "epoch": 1.31, "learning_rate": 0.00015132174088447776, "loss": 0.9912, "step": 895 }, { "epoch": 1.32, "learning_rate": 0.00015122049279531143, "loss": 0.9575, "step": 896 }, { "epoch": 1.32, "learning_rate": 0.00015111917347416671, "loss": 1.0356, "step": 897 }, { "epoch": 1.32, "learning_rate": 0.00015101778306194765, "loss": 0.9963, "step": 898 }, { "epoch": 1.32, "learning_rate": 0.0001509163216996572, "loss": 0.9728, "step": 899 }, { "epoch": 1.32, "learning_rate": 0.00015081478952839693, "loss": 1.0402, "step": 900 }, { "epoch": 1.32, "learning_rate": 0.00015071318668936695, "loss": 1.0287, "step": 901 }, { "epoch": 1.32, "learning_rate": 0.00015061151332386566, "loss": 1.0505, "step": 902 }, { "epoch": 1.33, "learning_rate": 0.00015050976957328938, "loss": 0.9814, "step": 903 }, { "epoch": 1.33, "learning_rate": 0.00015040795557913245, "loss": 1.0083, "step": 904 }, { "epoch": 1.33, "learning_rate": 0.00015030607148298696, "loss": 1.0871, "step": 905 }, { "epoch": 1.33, "learning_rate": 0.00015020411742654237, "loss": 1.0943, "step": 906 }, { "epoch": 1.33, "learning_rate": 0.0001501020935515854, "loss": 1.0631, "step": 907 }, { "epoch": 1.33, "learning_rate": 0.00015000000000000001, "loss": 1.0615, "step": 908 }, { "epoch": 1.34, "learning_rate": 0.00014989783691376696, "loss": 0.8933, "step": 909 }, { "epoch": 1.34, "learning_rate": 0.0001497956044349637, "loss": 1.012, "step": 910 }, { "epoch": 1.34, "learning_rate": 0.00014969330270576427, "loss": 0.9215, "step": 911 }, { "epoch": 1.34, "learning_rate": 0.00014959093186843895, "loss": 0.9894, "step": 912 }, { "epoch": 1.34, "learning_rate": 0.00014948849206535412, "loss": 1.0053, "step": 913 }, { "epoch": 1.34, "learning_rate": 0.00014938598343897214, "loss": 1.0975, "step": 914 }, { "epoch": 1.34, "learning_rate": 0.00014928340613185097, "loss": 1.068, "step": 915 }, { "epoch": 1.35, "learning_rate": 0.0001491807602866442, "loss": 0.9838, "step": 916 }, { "epoch": 1.35, "learning_rate": 0.00014907804604610063, "loss": 1.1493, "step": 917 }, { "epoch": 1.35, "learning_rate": 0.00014897526355306428, "loss": 0.9491, "step": 918 }, { "epoch": 1.35, "learning_rate": 0.000148872412950474, "loss": 1.0252, "step": 919 }, { "epoch": 1.35, "learning_rate": 0.00014876949438136347, "loss": 0.9555, "step": 920 }, { "epoch": 1.35, "learning_rate": 0.00014866650798886074, "loss": 0.9831, "step": 921 }, { "epoch": 1.35, "learning_rate": 0.0001485634539161883, "loss": 1.0957, "step": 922 }, { "epoch": 1.36, "learning_rate": 0.0001484603323066627, "loss": 0.9606, "step": 923 }, { "epoch": 1.36, "learning_rate": 0.00014835714330369446, "loss": 1.0643, "step": 924 }, { "epoch": 1.36, "learning_rate": 0.00014825388705078777, "loss": 1.0219, "step": 925 }, { "epoch": 1.36, "learning_rate": 0.00014815056369154038, "loss": 1.1315, "step": 926 }, { "epoch": 1.36, "learning_rate": 0.0001480471733696434, "loss": 1.0406, "step": 927 }, { "epoch": 1.36, "learning_rate": 0.000147943716228881, "loss": 1.0284, "step": 928 }, { "epoch": 1.37, "learning_rate": 0.00014784019241313026, "loss": 1.035, "step": 929 }, { "epoch": 1.37, "learning_rate": 0.00014773660206636105, "loss": 1.0562, "step": 930 }, { "epoch": 1.37, "learning_rate": 0.0001476329453326357, "loss": 0.9813, "step": 931 }, { "epoch": 1.37, "learning_rate": 0.000147529222356109, "loss": 1.0865, "step": 932 }, { "epoch": 1.37, "learning_rate": 0.0001474254332810277, "loss": 1.0074, "step": 933 }, { "epoch": 1.37, "learning_rate": 0.00014732157825173044, "loss": 1.0855, "step": 934 }, { "epoch": 1.37, "learning_rate": 0.00014721765741264786, "loss": 0.9785, "step": 935 }, { "epoch": 1.38, "learning_rate": 0.0001471136709083018, "loss": 1.011, "step": 936 }, { "epoch": 1.38, "learning_rate": 0.00014700961888330563, "loss": 1.0484, "step": 937 }, { "epoch": 1.38, "learning_rate": 0.0001469055014823637, "loss": 1.0435, "step": 938 }, { "epoch": 1.38, "learning_rate": 0.00014680131885027141, "loss": 1.0176, "step": 939 }, { "epoch": 1.38, "learning_rate": 0.00014669707113191483, "loss": 0.9542, "step": 940 }, { "epoch": 1.38, "learning_rate": 0.00014659275847227042, "loss": 0.9526, "step": 941 }, { "epoch": 1.38, "learning_rate": 0.00014648838101640518, "loss": 0.9681, "step": 942 }, { "epoch": 1.39, "learning_rate": 0.00014638393890947603, "loss": 0.9072, "step": 943 }, { "epoch": 1.39, "learning_rate": 0.0001462794322967299, "loss": 0.9939, "step": 944 }, { "epoch": 1.39, "learning_rate": 0.00014617486132350343, "loss": 1.018, "step": 945 }, { "epoch": 1.39, "learning_rate": 0.0001460702261352226, "loss": 0.8993, "step": 946 }, { "epoch": 1.39, "learning_rate": 0.00014596552687740302, "loss": 1.0134, "step": 947 }, { "epoch": 1.39, "learning_rate": 0.00014586076369564908, "loss": 0.947, "step": 948 }, { "epoch": 1.39, "learning_rate": 0.00014575593673565426, "loss": 1.0697, "step": 949 }, { "epoch": 1.4, "learning_rate": 0.00014565104614320065, "loss": 1.006, "step": 950 }, { "epoch": 1.4, "learning_rate": 0.00014554609206415885, "loss": 1.0262, "step": 951 }, { "epoch": 1.4, "learning_rate": 0.00014544107464448775, "loss": 0.9809, "step": 952 }, { "epoch": 1.4, "learning_rate": 0.0001453359940302344, "loss": 0.981, "step": 953 }, { "epoch": 1.4, "learning_rate": 0.00014523085036753354, "loss": 0.9925, "step": 954 }, { "epoch": 1.4, "learning_rate": 0.00014512564380260787, "loss": 1.0199, "step": 955 }, { "epoch": 1.41, "learning_rate": 0.00014502037448176734, "loss": 0.9715, "step": 956 }, { "epoch": 1.41, "learning_rate": 0.00014491504255140927, "loss": 1.0072, "step": 957 }, { "epoch": 1.41, "learning_rate": 0.000144809648158018, "loss": 1.0659, "step": 958 }, { "epoch": 1.41, "learning_rate": 0.00014470419144816483, "loss": 1.0538, "step": 959 }, { "epoch": 1.41, "learning_rate": 0.0001445986725685076, "loss": 1.0571, "step": 960 }, { "epoch": 1.41, "learning_rate": 0.00014449309166579072, "loss": 0.9701, "step": 961 }, { "epoch": 1.41, "learning_rate": 0.00014438744888684482, "loss": 0.9618, "step": 962 }, { "epoch": 1.42, "learning_rate": 0.0001442817443785865, "loss": 0.9179, "step": 963 }, { "epoch": 1.42, "learning_rate": 0.00014417597828801832, "loss": 1.0613, "step": 964 }, { "epoch": 1.42, "learning_rate": 0.00014407015076222846, "loss": 0.9558, "step": 965 }, { "epoch": 1.42, "learning_rate": 0.00014396426194839042, "loss": 0.9823, "step": 966 }, { "epoch": 1.42, "learning_rate": 0.00014385831199376317, "loss": 0.9968, "step": 967 }, { "epoch": 1.42, "learning_rate": 0.00014375230104569044, "loss": 0.9829, "step": 968 }, { "epoch": 1.42, "learning_rate": 0.00014364622925160098, "loss": 1.0552, "step": 969 }, { "epoch": 1.43, "learning_rate": 0.00014354009675900803, "loss": 0.993, "step": 970 }, { "epoch": 1.43, "learning_rate": 0.00014343390371550935, "loss": 1.0927, "step": 971 }, { "epoch": 1.43, "learning_rate": 0.00014332765026878687, "loss": 1.0387, "step": 972 }, { "epoch": 1.43, "learning_rate": 0.00014322133656660647, "loss": 0.9255, "step": 973 }, { "epoch": 1.43, "learning_rate": 0.00014311496275681783, "loss": 1.0093, "step": 974 }, { "epoch": 1.43, "learning_rate": 0.00014300852898735435, "loss": 1.0078, "step": 975 }, { "epoch": 1.44, "learning_rate": 0.00014290203540623267, "loss": 0.9161, "step": 976 }, { "epoch": 1.44, "learning_rate": 0.00014279548216155266, "loss": 1.03, "step": 977 }, { "epoch": 1.44, "learning_rate": 0.00014268886940149714, "loss": 1.0364, "step": 978 }, { "epoch": 1.44, "learning_rate": 0.0001425821972743318, "loss": 0.9768, "step": 979 }, { "epoch": 1.44, "learning_rate": 0.0001424754659284048, "loss": 1.1229, "step": 980 }, { "epoch": 1.44, "learning_rate": 0.0001423686755121466, "loss": 1.0362, "step": 981 }, { "epoch": 1.44, "learning_rate": 0.00014226182617406996, "loss": 0.9522, "step": 982 }, { "epoch": 1.45, "learning_rate": 0.00014215491806276944, "loss": 1.0479, "step": 983 }, { "epoch": 1.45, "learning_rate": 0.00014204795132692144, "loss": 1.0671, "step": 984 }, { "epoch": 1.45, "learning_rate": 0.00014194092611528384, "loss": 0.8983, "step": 985 }, { "epoch": 1.45, "learning_rate": 0.00014183384257669581, "loss": 1.004, "step": 986 }, { "epoch": 1.45, "learning_rate": 0.00014172670086007774, "loss": 1.0972, "step": 987 }, { "epoch": 1.45, "learning_rate": 0.00014161950111443077, "loss": 1.0198, "step": 988 }, { "epoch": 1.45, "learning_rate": 0.00014151224348883692, "loss": 1.0257, "step": 989 }, { "epoch": 1.46, "learning_rate": 0.00014140492813245856, "loss": 0.9717, "step": 990 }, { "epoch": 1.46, "learning_rate": 0.0001412975551945384, "loss": 0.9455, "step": 991 }, { "epoch": 1.46, "learning_rate": 0.0001411901248243993, "loss": 1.0372, "step": 992 }, { "epoch": 1.46, "learning_rate": 0.0001410826371714438, "loss": 0.9961, "step": 993 }, { "epoch": 1.46, "learning_rate": 0.00014097509238515432, "loss": 1.0599, "step": 994 }, { "epoch": 1.46, "learning_rate": 0.00014086749061509258, "loss": 1.0166, "step": 995 }, { "epoch": 1.46, "learning_rate": 0.00014075983201089964, "loss": 1.0254, "step": 996 }, { "epoch": 1.47, "learning_rate": 0.00014065211672229555, "loss": 0.9979, "step": 997 }, { "epoch": 1.47, "learning_rate": 0.00014054434489907915, "loss": 1.0365, "step": 998 }, { "epoch": 1.47, "learning_rate": 0.00014043651669112808, "loss": 1.0075, "step": 999 }, { "epoch": 1.47, "learning_rate": 0.00014032863224839814, "loss": 0.9743, "step": 1000 }, { "epoch": 1.47, "learning_rate": 0.00014022069172092352, "loss": 1.0056, "step": 1001 }, { "epoch": 1.47, "learning_rate": 0.00014011269525881636, "loss": 0.9647, "step": 1002 }, { "epoch": 1.48, "learning_rate": 0.00014000464301226656, "loss": 1.0912, "step": 1003 }, { "epoch": 1.48, "learning_rate": 0.00013989653513154165, "loss": 0.8811, "step": 1004 }, { "epoch": 1.48, "learning_rate": 0.00013978837176698646, "loss": 1.0667, "step": 1005 }, { "epoch": 1.48, "learning_rate": 0.000139680153069023, "loss": 1.0096, "step": 1006 }, { "epoch": 1.48, "learning_rate": 0.00013957187918815032, "loss": 0.926, "step": 1007 }, { "epoch": 1.48, "learning_rate": 0.0001394635502749441, "loss": 1.0814, "step": 1008 }, { "epoch": 1.48, "eval_loss": 1.0040607452392578, "eval_runtime": 2.6168, "eval_samples_per_second": 417.304, "eval_steps_per_second": 26.368, "step": 1008 }, { "epoch": 1.48, "learning_rate": 0.0001393551664800566, "loss": 1.0941, "step": 1009 }, { "epoch": 1.49, "learning_rate": 0.00013924672795421637, "loss": 1.044, "step": 1010 }, { "epoch": 1.49, "learning_rate": 0.00013913823484822815, "loss": 1.049, "step": 1011 }, { "epoch": 1.49, "learning_rate": 0.00013902968731297255, "loss": 0.9891, "step": 1012 }, { "epoch": 1.49, "learning_rate": 0.00013892108549940583, "loss": 0.9663, "step": 1013 }, { "epoch": 1.49, "learning_rate": 0.00013881242955855974, "loss": 1.0298, "step": 1014 }, { "epoch": 1.49, "learning_rate": 0.0001387037196415414, "loss": 1.0083, "step": 1015 }, { "epoch": 1.49, "learning_rate": 0.0001385949558995329, "loss": 0.9182, "step": 1016 }, { "epoch": 1.5, "learning_rate": 0.00013848613848379114, "loss": 1.013, "step": 1017 }, { "epoch": 1.5, "learning_rate": 0.00013837726754564785, "loss": 1.0022, "step": 1018 }, { "epoch": 1.5, "learning_rate": 0.000138268343236509, "loss": 0.9423, "step": 1019 }, { "epoch": 1.5, "learning_rate": 0.00013815936570785487, "loss": 1.058, "step": 1020 }, { "epoch": 1.5, "learning_rate": 0.00013805033511123975, "loss": 0.931, "step": 1021 }, { "epoch": 1.5, "learning_rate": 0.00013794125159829172, "loss": 1.0137, "step": 1022 }, { "epoch": 1.51, "learning_rate": 0.00013783211532071246, "loss": 1.0517, "step": 1023 }, { "epoch": 1.51, "learning_rate": 0.000137722926430277, "loss": 1.0259, "step": 1024 }, { "epoch": 1.51, "learning_rate": 0.00013761368507883359, "loss": 1.0263, "step": 1025 }, { "epoch": 1.51, "learning_rate": 0.00013750439141830339, "loss": 1.0286, "step": 1026 }, { "epoch": 1.51, "learning_rate": 0.00013739504560068033, "loss": 0.9749, "step": 1027 }, { "epoch": 1.51, "learning_rate": 0.00013728564777803088, "loss": 0.9317, "step": 1028 }, { "epoch": 1.51, "learning_rate": 0.00013717619810249378, "loss": 1.0653, "step": 1029 }, { "epoch": 1.52, "learning_rate": 0.00013706669672627997, "loss": 0.9623, "step": 1030 }, { "epoch": 1.52, "learning_rate": 0.00013695714380167223, "loss": 0.9911, "step": 1031 }, { "epoch": 1.52, "learning_rate": 0.000136847539481025, "loss": 0.9843, "step": 1032 }, { "epoch": 1.52, "learning_rate": 0.0001367378839167643, "loss": 0.981, "step": 1033 }, { "epoch": 1.52, "learning_rate": 0.00013662817726138728, "loss": 1.0651, "step": 1034 }, { "epoch": 1.52, "learning_rate": 0.00013651841966746232, "loss": 1.0602, "step": 1035 }, { "epoch": 1.52, "learning_rate": 0.0001364086112876284, "loss": 0.9524, "step": 1036 }, { "epoch": 1.53, "learning_rate": 0.00013629875227459532, "loss": 1.0264, "step": 1037 }, { "epoch": 1.53, "learning_rate": 0.00013618884278114324, "loss": 1.0691, "step": 1038 }, { "epoch": 1.53, "learning_rate": 0.00013607888296012259, "loss": 1.0527, "step": 1039 }, { "epoch": 1.53, "learning_rate": 0.0001359688729644536, "loss": 0.9629, "step": 1040 }, { "epoch": 1.53, "learning_rate": 0.0001358588129471264, "loss": 0.957, "step": 1041 }, { "epoch": 1.53, "learning_rate": 0.00013574870306120077, "loss": 1.0976, "step": 1042 }, { "epoch": 1.54, "learning_rate": 0.00013563854345980569, "loss": 0.9317, "step": 1043 }, { "epoch": 1.54, "learning_rate": 0.00013552833429613938, "loss": 1.0359, "step": 1044 }, { "epoch": 1.54, "learning_rate": 0.0001354180757234689, "loss": 1.0642, "step": 1045 }, { "epoch": 1.54, "learning_rate": 0.0001353077678951301, "loss": 1.0526, "step": 1046 }, { "epoch": 1.54, "learning_rate": 0.00013519741096452726, "loss": 1.0276, "step": 1047 }, { "epoch": 1.54, "learning_rate": 0.00013508700508513307, "loss": 1.0471, "step": 1048 }, { "epoch": 1.54, "learning_rate": 0.0001349765504104881, "loss": 1.0353, "step": 1049 }, { "epoch": 1.55, "learning_rate": 0.00013486604709420102, "loss": 1.0025, "step": 1050 }, { "epoch": 1.55, "learning_rate": 0.00013475549528994786, "loss": 0.9019, "step": 1051 }, { "epoch": 1.55, "learning_rate": 0.00013464489515147238, "loss": 1.0453, "step": 1052 }, { "epoch": 1.55, "learning_rate": 0.00013453424683258528, "loss": 1.0395, "step": 1053 }, { "epoch": 1.55, "learning_rate": 0.0001344235504871645, "loss": 0.8939, "step": 1054 }, { "epoch": 1.55, "learning_rate": 0.00013431280626915467, "loss": 0.9198, "step": 1055 }, { "epoch": 1.55, "learning_rate": 0.00013420201433256689, "loss": 1.0046, "step": 1056 }, { "epoch": 1.56, "learning_rate": 0.0001340911748314788, "loss": 0.9197, "step": 1057 }, { "epoch": 1.56, "learning_rate": 0.00013398028792003413, "loss": 0.9547, "step": 1058 }, { "epoch": 1.56, "learning_rate": 0.00013386935375244246, "loss": 0.968, "step": 1059 }, { "epoch": 1.56, "learning_rate": 0.00013375837248297926, "loss": 0.9611, "step": 1060 }, { "epoch": 1.56, "learning_rate": 0.00013364734426598527, "loss": 1.0125, "step": 1061 }, { "epoch": 1.56, "learning_rate": 0.00013353626925586672, "loss": 1.0179, "step": 1062 }, { "epoch": 1.56, "learning_rate": 0.00013342514760709485, "loss": 1.04, "step": 1063 }, { "epoch": 1.57, "learning_rate": 0.00013331397947420576, "loss": 0.9251, "step": 1064 }, { "epoch": 1.57, "learning_rate": 0.00013320276501180015, "loss": 1.0762, "step": 1065 }, { "epoch": 1.57, "learning_rate": 0.00013309150437454322, "loss": 1.0137, "step": 1066 }, { "epoch": 1.57, "learning_rate": 0.00013298019771716435, "loss": 0.9981, "step": 1067 }, { "epoch": 1.57, "learning_rate": 0.0001328688451944569, "loss": 1.003, "step": 1068 }, { "epoch": 1.57, "learning_rate": 0.00013275744696127805, "loss": 1.0307, "step": 1069 }, { "epoch": 1.58, "learning_rate": 0.00013264600317254853, "loss": 1.0257, "step": 1070 }, { "epoch": 1.58, "learning_rate": 0.00013253451398325249, "loss": 1.0426, "step": 1071 }, { "epoch": 1.58, "learning_rate": 0.00013242297954843711, "loss": 1.0167, "step": 1072 }, { "epoch": 1.58, "learning_rate": 0.00013231140002321253, "loss": 1.012, "step": 1073 }, { "epoch": 1.58, "learning_rate": 0.00013219977556275163, "loss": 1.0649, "step": 1074 }, { "epoch": 1.58, "learning_rate": 0.00013208810632228977, "loss": 0.9297, "step": 1075 }, { "epoch": 1.58, "learning_rate": 0.00013197639245712454, "loss": 0.9772, "step": 1076 }, { "epoch": 1.59, "learning_rate": 0.00013186463412261565, "loss": 1.0194, "step": 1077 }, { "epoch": 1.59, "learning_rate": 0.00013175283147418465, "loss": 1.0596, "step": 1078 }, { "epoch": 1.59, "learning_rate": 0.00013164098466731468, "loss": 0.9938, "step": 1079 }, { "epoch": 1.59, "learning_rate": 0.00013152909385755025, "loss": 0.9405, "step": 1080 }, { "epoch": 1.59, "learning_rate": 0.0001314171592004972, "loss": 1.0175, "step": 1081 }, { "epoch": 1.59, "learning_rate": 0.00013130518085182225, "loss": 0.9994, "step": 1082 }, { "epoch": 1.59, "learning_rate": 0.00013119315896725287, "loss": 0.9524, "step": 1083 }, { "epoch": 1.6, "learning_rate": 0.00013108109370257712, "loss": 0.9112, "step": 1084 }, { "epoch": 1.6, "learning_rate": 0.00013096898521364338, "loss": 0.9339, "step": 1085 }, { "epoch": 1.6, "learning_rate": 0.00013085683365636014, "loss": 0.9718, "step": 1086 }, { "epoch": 1.6, "learning_rate": 0.0001307446391866958, "loss": 0.969, "step": 1087 }, { "epoch": 1.6, "learning_rate": 0.00013063240196067836, "loss": 1.0255, "step": 1088 }, { "epoch": 1.6, "learning_rate": 0.00013052012213439536, "loss": 1.0119, "step": 1089 }, { "epoch": 1.61, "learning_rate": 0.00013040779986399362, "loss": 1.0396, "step": 1090 }, { "epoch": 1.61, "learning_rate": 0.00013029543530567884, "loss": 1.0202, "step": 1091 }, { "epoch": 1.61, "learning_rate": 0.0001301830286157157, "loss": 0.9024, "step": 1092 }, { "epoch": 1.61, "learning_rate": 0.00013007057995042732, "loss": 1.0079, "step": 1093 }, { "epoch": 1.61, "learning_rate": 0.0001299580894661953, "loss": 0.9771, "step": 1094 }, { "epoch": 1.61, "learning_rate": 0.0001298455573194594, "loss": 0.9942, "step": 1095 }, { "epoch": 1.61, "learning_rate": 0.00012973298366671725, "loss": 0.9879, "step": 1096 }, { "epoch": 1.62, "learning_rate": 0.00012962036866452422, "loss": 0.9365, "step": 1097 }, { "epoch": 1.62, "learning_rate": 0.0001295077124694932, "loss": 1.0128, "step": 1098 }, { "epoch": 1.62, "learning_rate": 0.00012939501523829444, "loss": 1.0707, "step": 1099 }, { "epoch": 1.62, "learning_rate": 0.00012928227712765504, "loss": 0.9769, "step": 1100 }, { "epoch": 1.62, "learning_rate": 0.00012916949829435922, "loss": 1.0208, "step": 1101 }, { "epoch": 1.62, "learning_rate": 0.0001290566788952477, "loss": 1.0376, "step": 1102 }, { "epoch": 1.62, "learning_rate": 0.00012894381908721756, "loss": 1.0588, "step": 1103 }, { "epoch": 1.63, "learning_rate": 0.0001288309190272222, "loss": 1.0217, "step": 1104 }, { "epoch": 1.63, "learning_rate": 0.00012871797887227087, "loss": 0.9684, "step": 1105 }, { "epoch": 1.63, "learning_rate": 0.00012860499877942875, "loss": 0.9753, "step": 1106 }, { "epoch": 1.63, "learning_rate": 0.00012849197890581638, "loss": 1.0094, "step": 1107 }, { "epoch": 1.63, "learning_rate": 0.00012837891940860972, "loss": 1.0346, "step": 1108 }, { "epoch": 1.63, "learning_rate": 0.00012826582044503978, "loss": 0.8741, "step": 1109 }, { "epoch": 1.63, "learning_rate": 0.00012815268217239252, "loss": 1.0223, "step": 1110 }, { "epoch": 1.64, "learning_rate": 0.00012803950474800862, "loss": 0.8748, "step": 1111 }, { "epoch": 1.64, "learning_rate": 0.00012792628832928302, "loss": 1.0296, "step": 1112 }, { "epoch": 1.64, "learning_rate": 0.000127813033073665, "loss": 0.9993, "step": 1113 }, { "epoch": 1.64, "learning_rate": 0.00012769973913865794, "loss": 1.0555, "step": 1114 }, { "epoch": 1.64, "learning_rate": 0.00012758640668181882, "loss": 1.0245, "step": 1115 }, { "epoch": 1.64, "learning_rate": 0.0001274730358607583, "loss": 0.9502, "step": 1116 }, { "epoch": 1.65, "learning_rate": 0.00012735962683314042, "loss": 1.0165, "step": 1117 }, { "epoch": 1.65, "learning_rate": 0.0001272461797566823, "loss": 1.0669, "step": 1118 }, { "epoch": 1.65, "learning_rate": 0.000127132694789154, "loss": 0.8676, "step": 1119 }, { "epoch": 1.65, "learning_rate": 0.0001270191720883782, "loss": 0.9432, "step": 1120 }, { "epoch": 1.65, "learning_rate": 0.00012690561181223024, "loss": 1.0614, "step": 1121 }, { "epoch": 1.65, "learning_rate": 0.0001267920141186375, "loss": 0.9924, "step": 1122 }, { "epoch": 1.65, "learning_rate": 0.00012667837916557954, "loss": 1.139, "step": 1123 }, { "epoch": 1.66, "learning_rate": 0.00012656470711108764, "loss": 1.0043, "step": 1124 }, { "epoch": 1.66, "learning_rate": 0.00012645099811324476, "loss": 1.0747, "step": 1125 }, { "epoch": 1.66, "learning_rate": 0.0001263372523301852, "loss": 0.9668, "step": 1126 }, { "epoch": 1.66, "learning_rate": 0.00012622346992009447, "loss": 0.9931, "step": 1127 }, { "epoch": 1.66, "learning_rate": 0.00012610965104120885, "loss": 0.9393, "step": 1128 }, { "epoch": 1.66, "learning_rate": 0.00012599579585181552, "loss": 0.9918, "step": 1129 }, { "epoch": 1.66, "learning_rate": 0.00012588190451025207, "loss": 1.0172, "step": 1130 }, { "epoch": 1.67, "learning_rate": 0.00012576797717490644, "loss": 1.0586, "step": 1131 }, { "epoch": 1.67, "learning_rate": 0.00012565401400421651, "loss": 1.0482, "step": 1132 }, { "epoch": 1.67, "learning_rate": 0.00012554001515667008, "loss": 1.0548, "step": 1133 }, { "epoch": 1.67, "learning_rate": 0.00012542598079080456, "loss": 1.0092, "step": 1134 }, { "epoch": 1.67, "learning_rate": 0.00012531191106520672, "loss": 1.0162, "step": 1135 }, { "epoch": 1.67, "learning_rate": 0.00012519780613851254, "loss": 1.0387, "step": 1136 }, { "epoch": 1.68, "learning_rate": 0.0001250836661694069, "loss": 0.9607, "step": 1137 }, { "epoch": 1.68, "learning_rate": 0.00012496949131662348, "loss": 1.0025, "step": 1138 }, { "epoch": 1.68, "learning_rate": 0.00012485528173894448, "loss": 1.0014, "step": 1139 }, { "epoch": 1.68, "learning_rate": 0.00012474103759520027, "loss": 0.9838, "step": 1140 }, { "epoch": 1.68, "learning_rate": 0.0001246267590442694, "loss": 1.0384, "step": 1141 }, { "epoch": 1.68, "learning_rate": 0.00012451244624507831, "loss": 0.9958, "step": 1142 }, { "epoch": 1.68, "learning_rate": 0.00012439809935660095, "loss": 0.9927, "step": 1143 }, { "epoch": 1.69, "learning_rate": 0.0001242837185378587, "loss": 1.0082, "step": 1144 }, { "epoch": 1.69, "learning_rate": 0.00012416930394792026, "loss": 0.9729, "step": 1145 }, { "epoch": 1.69, "learning_rate": 0.00012405485574590113, "loss": 1.0464, "step": 1146 }, { "epoch": 1.69, "learning_rate": 0.00012394037409096357, "loss": 0.987, "step": 1147 }, { "epoch": 1.69, "learning_rate": 0.0001238258591423165, "loss": 0.9402, "step": 1148 }, { "epoch": 1.69, "learning_rate": 0.00012371131105921504, "loss": 1.0293, "step": 1149 }, { "epoch": 1.69, "learning_rate": 0.00012359673000096033, "loss": 0.9418, "step": 1150 }, { "epoch": 1.7, "learning_rate": 0.0001234821161268995, "loss": 0.964, "step": 1151 }, { "epoch": 1.7, "learning_rate": 0.00012336746959642526, "loss": 0.9982, "step": 1152 }, { "epoch": 1.7, "learning_rate": 0.0001232527905689757, "loss": 0.9364, "step": 1153 }, { "epoch": 1.7, "learning_rate": 0.00012313807920403419, "loss": 0.9399, "step": 1154 }, { "epoch": 1.7, "learning_rate": 0.0001230233356611289, "loss": 1.015, "step": 1155 }, { "epoch": 1.7, "learning_rate": 0.000122908560099833, "loss": 1.0214, "step": 1156 }, { "epoch": 1.7, "learning_rate": 0.00012279375267976398, "loss": 1.0262, "step": 1157 }, { "epoch": 1.71, "learning_rate": 0.00012267891356058377, "loss": 1.0277, "step": 1158 }, { "epoch": 1.71, "learning_rate": 0.00012256404290199825, "loss": 1.0095, "step": 1159 }, { "epoch": 1.71, "learning_rate": 0.00012244914086375724, "loss": 1.0314, "step": 1160 }, { "epoch": 1.71, "learning_rate": 0.00012233420760565428, "loss": 0.8282, "step": 1161 }, { "epoch": 1.71, "learning_rate": 0.00012221924328752616, "loss": 0.9709, "step": 1162 }, { "epoch": 1.71, "learning_rate": 0.00012210424806925301, "loss": 0.941, "step": 1163 }, { "epoch": 1.72, "learning_rate": 0.00012198922211075778, "loss": 0.9716, "step": 1164 }, { "epoch": 1.72, "learning_rate": 0.00012187416557200633, "loss": 1.0125, "step": 1165 }, { "epoch": 1.72, "learning_rate": 0.00012175907861300697, "loss": 1.0159, "step": 1166 }, { "epoch": 1.72, "learning_rate": 0.00012164396139381029, "loss": 0.9306, "step": 1167 }, { "epoch": 1.72, "learning_rate": 0.00012152881407450905, "loss": 1.1056, "step": 1168 }, { "epoch": 1.72, "learning_rate": 0.00012141363681523776, "loss": 1.0113, "step": 1169 }, { "epoch": 1.72, "learning_rate": 0.00012129842977617265, "loss": 0.9983, "step": 1170 }, { "epoch": 1.73, "learning_rate": 0.00012118319311753137, "loss": 1.0076, "step": 1171 }, { "epoch": 1.73, "learning_rate": 0.00012106792699957263, "loss": 1.1181, "step": 1172 }, { "epoch": 1.73, "learning_rate": 0.00012095263158259631, "loss": 0.8759, "step": 1173 }, { "epoch": 1.73, "learning_rate": 0.00012083730702694291, "loss": 0.9855, "step": 1174 }, { "epoch": 1.73, "learning_rate": 0.00012072195349299345, "loss": 1.1361, "step": 1175 }, { "epoch": 1.73, "learning_rate": 0.00012060657114116926, "loss": 1.0275, "step": 1176 }, { "epoch": 1.73, "eval_loss": 0.9928944110870361, "eval_runtime": 2.6469, "eval_samples_per_second": 412.56, "eval_steps_per_second": 26.068, "step": 1176 }, { "epoch": 1.73, "learning_rate": 0.0001204911601319318, "loss": 1.0256, "step": 1177 }, { "epoch": 1.74, "learning_rate": 0.00012037572062578238, "loss": 0.9218, "step": 1178 }, { "epoch": 1.74, "learning_rate": 0.00012026025278326187, "loss": 1.0394, "step": 1179 }, { "epoch": 1.74, "learning_rate": 0.00012014475676495052, "loss": 1.0318, "step": 1180 }, { "epoch": 1.74, "learning_rate": 0.00012002923273146794, "loss": 1.0361, "step": 1181 }, { "epoch": 1.74, "learning_rate": 0.00011991368084347252, "loss": 1.0093, "step": 1182 }, { "epoch": 1.74, "learning_rate": 0.00011979810126166151, "loss": 0.9527, "step": 1183 }, { "epoch": 1.75, "learning_rate": 0.00011968249414677055, "loss": 1.0946, "step": 1184 }, { "epoch": 1.75, "learning_rate": 0.00011956685965957368, "loss": 1.0124, "step": 1185 }, { "epoch": 1.75, "learning_rate": 0.000119451197960883, "loss": 1.0074, "step": 1186 }, { "epoch": 1.75, "learning_rate": 0.00011933550921154834, "loss": 1.0315, "step": 1187 }, { "epoch": 1.75, "learning_rate": 0.0001192197935724573, "loss": 0.9915, "step": 1188 }, { "epoch": 1.75, "learning_rate": 0.00011910405120453476, "loss": 0.9823, "step": 1189 }, { "epoch": 1.75, "learning_rate": 0.00011898828226874284, "loss": 1.0294, "step": 1190 }, { "epoch": 1.76, "learning_rate": 0.00011887248692608057, "loss": 1.0176, "step": 1191 }, { "epoch": 1.76, "learning_rate": 0.00011875666533758372, "loss": 1.0486, "step": 1192 }, { "epoch": 1.76, "learning_rate": 0.00011864081766432456, "loss": 1.0237, "step": 1193 }, { "epoch": 1.76, "learning_rate": 0.00011852494406741165, "loss": 1.0469, "step": 1194 }, { "epoch": 1.76, "learning_rate": 0.00011840904470798955, "loss": 0.9545, "step": 1195 }, { "epoch": 1.76, "learning_rate": 0.00011829311974723867, "loss": 0.9812, "step": 1196 }, { "epoch": 1.76, "learning_rate": 0.00011817716934637509, "loss": 1.0503, "step": 1197 }, { "epoch": 1.77, "learning_rate": 0.0001180611936666502, "loss": 1.0693, "step": 1198 }, { "epoch": 1.77, "learning_rate": 0.00011794519286935055, "loss": 0.9627, "step": 1199 }, { "epoch": 1.77, "learning_rate": 0.00011782916711579759, "loss": 0.9728, "step": 1200 }, { "epoch": 1.77, "learning_rate": 0.0001177131165673476, "loss": 1.13, "step": 1201 }, { "epoch": 1.77, "learning_rate": 0.0001175970413853912, "loss": 0.9756, "step": 1202 }, { "epoch": 1.77, "learning_rate": 0.00011748094173135337, "loss": 1.0069, "step": 1203 }, { "epoch": 1.77, "learning_rate": 0.00011736481776669306, "loss": 1.033, "step": 1204 }, { "epoch": 1.78, "learning_rate": 0.00011724866965290302, "loss": 0.9906, "step": 1205 }, { "epoch": 1.78, "learning_rate": 0.00011713249755150965, "loss": 1.1008, "step": 1206 }, { "epoch": 1.78, "learning_rate": 0.00011701630162407266, "loss": 0.9987, "step": 1207 }, { "epoch": 1.78, "learning_rate": 0.00011690008203218493, "loss": 1.0122, "step": 1208 }, { "epoch": 1.78, "learning_rate": 0.0001167838389374722, "loss": 1.0495, "step": 1209 }, { "epoch": 1.78, "learning_rate": 0.0001166675725015929, "loss": 0.9875, "step": 1210 }, { "epoch": 1.79, "learning_rate": 0.00011655128288623802, "loss": 1.0231, "step": 1211 }, { "epoch": 1.79, "learning_rate": 0.00011643497025313061, "loss": 0.9342, "step": 1212 }, { "epoch": 1.79, "learning_rate": 0.00011631863476402594, "loss": 1.1006, "step": 1213 }, { "epoch": 1.79, "learning_rate": 0.00011620227658071087, "loss": 0.9264, "step": 1214 }, { "epoch": 1.79, "learning_rate": 0.00011608589586500391, "loss": 1.1099, "step": 1215 }, { "epoch": 1.79, "learning_rate": 0.00011596949277875495, "loss": 1.0326, "step": 1216 }, { "epoch": 1.79, "learning_rate": 0.0001158530674838449, "loss": 0.9235, "step": 1217 }, { "epoch": 1.8, "learning_rate": 0.00011573662014218564, "loss": 1.0227, "step": 1218 }, { "epoch": 1.8, "learning_rate": 0.00011562015091571963, "loss": 1.0028, "step": 1219 }, { "epoch": 1.8, "learning_rate": 0.00011550365996641979, "loss": 1.0744, "step": 1220 }, { "epoch": 1.8, "learning_rate": 0.00011538714745628931, "loss": 0.9521, "step": 1221 }, { "epoch": 1.8, "learning_rate": 0.00011527061354736129, "loss": 1.0171, "step": 1222 }, { "epoch": 1.8, "learning_rate": 0.00011515405840169861, "loss": 1.0481, "step": 1223 }, { "epoch": 1.8, "learning_rate": 0.00011503748218139369, "loss": 1.0034, "step": 1224 }, { "epoch": 1.81, "learning_rate": 0.00011492088504856826, "loss": 1.1384, "step": 1225 }, { "epoch": 1.81, "learning_rate": 0.00011480426716537315, "loss": 1.0268, "step": 1226 }, { "epoch": 1.81, "learning_rate": 0.00011468762869398802, "loss": 1.003, "step": 1227 }, { "epoch": 1.81, "learning_rate": 0.00011457096979662114, "loss": 1.1087, "step": 1228 }, { "epoch": 1.81, "learning_rate": 0.00011445429063550926, "loss": 1.0809, "step": 1229 }, { "epoch": 1.81, "learning_rate": 0.00011433759137291727, "loss": 1.0054, "step": 1230 }, { "epoch": 1.82, "learning_rate": 0.00011422087217113795, "loss": 0.9416, "step": 1231 }, { "epoch": 1.82, "learning_rate": 0.00011410413319249194, "loss": 1.0153, "step": 1232 }, { "epoch": 1.82, "learning_rate": 0.00011398737459932728, "loss": 1.0622, "step": 1233 }, { "epoch": 1.82, "learning_rate": 0.00011387059655401932, "loss": 1.0792, "step": 1234 }, { "epoch": 1.82, "learning_rate": 0.00011375379921897051, "loss": 0.9822, "step": 1235 }, { "epoch": 1.82, "learning_rate": 0.00011363698275661001, "loss": 0.9949, "step": 1236 }, { "epoch": 1.82, "learning_rate": 0.00011352014732939369, "loss": 0.9653, "step": 1237 }, { "epoch": 1.83, "learning_rate": 0.00011340329309980377, "loss": 1.0694, "step": 1238 }, { "epoch": 1.83, "learning_rate": 0.00011328642023034857, "loss": 0.9925, "step": 1239 }, { "epoch": 1.83, "learning_rate": 0.00011316952888356237, "loss": 0.9829, "step": 1240 }, { "epoch": 1.83, "learning_rate": 0.00011305261922200519, "loss": 0.9659, "step": 1241 }, { "epoch": 1.83, "learning_rate": 0.00011293569140826239, "loss": 1.109, "step": 1242 }, { "epoch": 1.83, "learning_rate": 0.00011281874560494472, "loss": 1.0614, "step": 1243 }, { "epoch": 1.83, "learning_rate": 0.00011270178197468789, "loss": 0.9013, "step": 1244 }, { "epoch": 1.84, "learning_rate": 0.00011258480068015235, "loss": 1.0049, "step": 1245 }, { "epoch": 1.84, "learning_rate": 0.00011246780188402322, "loss": 0.9746, "step": 1246 }, { "epoch": 1.84, "learning_rate": 0.00011235078574900984, "loss": 1.1433, "step": 1247 }, { "epoch": 1.84, "learning_rate": 0.00011223375243784573, "loss": 1.0196, "step": 1248 }, { "epoch": 1.84, "learning_rate": 0.00011211670211328833, "loss": 0.9859, "step": 1249 }, { "epoch": 1.84, "learning_rate": 0.0001119996349381187, "loss": 0.9037, "step": 1250 }, { "epoch": 1.85, "learning_rate": 0.0001118825510751413, "loss": 1.0481, "step": 1251 }, { "epoch": 1.85, "learning_rate": 0.00011176545068718385, "loss": 1.0324, "step": 1252 }, { "epoch": 1.85, "learning_rate": 0.00011164833393709706, "loss": 1.0155, "step": 1253 }, { "epoch": 1.85, "learning_rate": 0.00011153120098775434, "loss": 0.967, "step": 1254 }, { "epoch": 1.85, "learning_rate": 0.00011141405200205166, "loss": 0.9766, "step": 1255 }, { "epoch": 1.85, "learning_rate": 0.00011129688714290729, "loss": 1.0075, "step": 1256 }, { "epoch": 1.85, "learning_rate": 0.00011117970657326158, "loss": 0.9472, "step": 1257 }, { "epoch": 1.86, "learning_rate": 0.00011106251045607674, "loss": 0.9949, "step": 1258 }, { "epoch": 1.86, "learning_rate": 0.00011094529895433652, "loss": 1.0302, "step": 1259 }, { "epoch": 1.86, "learning_rate": 0.0001108280722310462, "loss": 1.0538, "step": 1260 }, { "epoch": 1.86, "learning_rate": 0.00011071083044923214, "loss": 1.0025, "step": 1261 }, { "epoch": 1.86, "learning_rate": 0.00011059357377194161, "loss": 1.024, "step": 1262 }, { "epoch": 1.86, "learning_rate": 0.00011047630236224271, "loss": 0.9452, "step": 1263 }, { "epoch": 1.86, "learning_rate": 0.00011035901638322392, "loss": 1.0055, "step": 1264 }, { "epoch": 1.87, "learning_rate": 0.00011024171599799409, "loss": 0.9875, "step": 1265 }, { "epoch": 1.87, "learning_rate": 0.00011012440136968196, "loss": 0.9582, "step": 1266 }, { "epoch": 1.87, "learning_rate": 0.00011000707266143617, "loss": 0.9986, "step": 1267 }, { "epoch": 1.87, "learning_rate": 0.00010988973003642499, "loss": 1.0328, "step": 1268 }, { "epoch": 1.87, "learning_rate": 0.0001097723736578359, "loss": 1.0108, "step": 1269 }, { "epoch": 1.87, "learning_rate": 0.00010965500368887567, "loss": 0.9941, "step": 1270 }, { "epoch": 1.87, "learning_rate": 0.00010953762029276982, "loss": 1.0842, "step": 1271 }, { "epoch": 1.88, "learning_rate": 0.00010942022363276264, "loss": 0.994, "step": 1272 }, { "epoch": 1.88, "learning_rate": 0.00010930281387211683, "loss": 1.0151, "step": 1273 }, { "epoch": 1.88, "learning_rate": 0.00010918539117411333, "loss": 1.0172, "step": 1274 }, { "epoch": 1.88, "learning_rate": 0.00010906795570205104, "loss": 1.0698, "step": 1275 }, { "epoch": 1.88, "learning_rate": 0.00010895050761924668, "loss": 0.9835, "step": 1276 }, { "epoch": 1.88, "learning_rate": 0.0001088330470890344, "loss": 0.9461, "step": 1277 }, { "epoch": 1.89, "learning_rate": 0.00010871557427476583, "loss": 1.0394, "step": 1278 }, { "epoch": 1.89, "learning_rate": 0.00010859808933980948, "loss": 0.9639, "step": 1279 }, { "epoch": 1.89, "learning_rate": 0.00010848059244755093, "loss": 0.9863, "step": 1280 }, { "epoch": 1.89, "learning_rate": 0.00010836308376139221, "loss": 1.0728, "step": 1281 }, { "epoch": 1.89, "learning_rate": 0.00010824556344475181, "loss": 0.9989, "step": 1282 }, { "epoch": 1.89, "learning_rate": 0.00010812803166106444, "loss": 0.962, "step": 1283 }, { "epoch": 1.89, "learning_rate": 0.00010801048857378071, "loss": 0.8658, "step": 1284 }, { "epoch": 1.9, "learning_rate": 0.00010789293434636698, "loss": 1.0488, "step": 1285 }, { "epoch": 1.9, "learning_rate": 0.00010777536914230508, "loss": 1.0183, "step": 1286 }, { "epoch": 1.9, "learning_rate": 0.00010765779312509208, "loss": 0.9535, "step": 1287 }, { "epoch": 1.9, "learning_rate": 0.00010754020645824017, "loss": 0.9978, "step": 1288 }, { "epoch": 1.9, "learning_rate": 0.00010742260930527625, "loss": 0.8929, "step": 1289 }, { "epoch": 1.9, "learning_rate": 0.0001073050018297419, "loss": 0.9762, "step": 1290 }, { "epoch": 1.9, "learning_rate": 0.00010718738419519297, "loss": 1.0559, "step": 1291 }, { "epoch": 1.91, "learning_rate": 0.00010706975656519946, "loss": 1.0327, "step": 1292 }, { "epoch": 1.91, "learning_rate": 0.00010695211910334537, "loss": 1.0322, "step": 1293 }, { "epoch": 1.91, "learning_rate": 0.00010683447197322817, "loss": 1.0542, "step": 1294 }, { "epoch": 1.91, "learning_rate": 0.00010671681533845899, "loss": 1.0521, "step": 1295 }, { "epoch": 1.91, "learning_rate": 0.00010659914936266206, "loss": 0.9967, "step": 1296 }, { "epoch": 1.91, "learning_rate": 0.00010648147420947461, "loss": 1.0491, "step": 1297 }, { "epoch": 1.92, "learning_rate": 0.00010636379004254664, "loss": 0.9035, "step": 1298 }, { "epoch": 1.92, "learning_rate": 0.00010624609702554069, "loss": 1.0704, "step": 1299 }, { "epoch": 1.92, "learning_rate": 0.00010612839532213164, "loss": 0.9533, "step": 1300 }, { "epoch": 1.92, "learning_rate": 0.00010601068509600642, "loss": 1.0396, "step": 1301 }, { "epoch": 1.92, "learning_rate": 0.00010589296651086376, "loss": 0.9543, "step": 1302 }, { "epoch": 1.92, "learning_rate": 0.0001057752397304141, "loss": 1.0591, "step": 1303 }, { "epoch": 1.92, "learning_rate": 0.00010565750491837925, "loss": 1.1191, "step": 1304 }, { "epoch": 1.93, "learning_rate": 0.00010553976223849218, "loss": 0.916, "step": 1305 }, { "epoch": 1.93, "learning_rate": 0.00010542201185449678, "loss": 0.9732, "step": 1306 }, { "epoch": 1.93, "learning_rate": 0.00010530425393014774, "loss": 1.01, "step": 1307 }, { "epoch": 1.93, "learning_rate": 0.00010518648862921012, "loss": 0.9849, "step": 1308 }, { "epoch": 1.93, "learning_rate": 0.0001050687161154593, "loss": 1.0519, "step": 1309 }, { "epoch": 1.93, "learning_rate": 0.00010495093655268071, "loss": 1.0539, "step": 1310 }, { "epoch": 1.93, "learning_rate": 0.00010483315010466952, "loss": 0.9922, "step": 1311 }, { "epoch": 1.94, "learning_rate": 0.00010471535693523057, "loss": 1.0048, "step": 1312 }, { "epoch": 1.94, "learning_rate": 0.00010459755720817797, "loss": 1.0576, "step": 1313 }, { "epoch": 1.94, "learning_rate": 0.00010447975108733492, "loss": 1.0268, "step": 1314 }, { "epoch": 1.94, "learning_rate": 0.00010436193873653361, "loss": 1.0566, "step": 1315 }, { "epoch": 1.94, "learning_rate": 0.00010424412031961484, "loss": 1.0294, "step": 1316 }, { "epoch": 1.94, "learning_rate": 0.00010412629600042785, "loss": 1.0808, "step": 1317 }, { "epoch": 1.94, "learning_rate": 0.00010400846594283012, "loss": 1.0487, "step": 1318 }, { "epoch": 1.95, "learning_rate": 0.00010389063031068698, "loss": 1.04, "step": 1319 }, { "epoch": 1.95, "learning_rate": 0.00010377278926787173, "loss": 1.033, "step": 1320 }, { "epoch": 1.95, "learning_rate": 0.000103654942978265, "loss": 0.9637, "step": 1321 }, { "epoch": 1.95, "learning_rate": 0.00010353709160575489, "loss": 0.9665, "step": 1322 }, { "epoch": 1.95, "learning_rate": 0.00010341923531423634, "loss": 1.0079, "step": 1323 }, { "epoch": 1.95, "learning_rate": 0.00010330137426761135, "loss": 0.9989, "step": 1324 }, { "epoch": 1.96, "learning_rate": 0.00010318350862978848, "loss": 1.0103, "step": 1325 }, { "epoch": 1.96, "learning_rate": 0.00010306563856468253, "loss": 0.9872, "step": 1326 }, { "epoch": 1.96, "learning_rate": 0.00010294776423621464, "loss": 0.9684, "step": 1327 }, { "epoch": 1.96, "learning_rate": 0.00010282988580831183, "loss": 0.9745, "step": 1328 }, { "epoch": 1.96, "learning_rate": 0.00010271200344490674, "loss": 1.0621, "step": 1329 }, { "epoch": 1.96, "learning_rate": 0.0001025941173099376, "loss": 1.0639, "step": 1330 }, { "epoch": 1.96, "learning_rate": 0.00010247622756734774, "loss": 0.914, "step": 1331 }, { "epoch": 1.97, "learning_rate": 0.00010235833438108571, "loss": 1.0135, "step": 1332 }, { "epoch": 1.97, "learning_rate": 0.00010224043791510465, "loss": 1.0132, "step": 1333 }, { "epoch": 1.97, "learning_rate": 0.00010212253833336237, "loss": 0.9912, "step": 1334 }, { "epoch": 1.97, "learning_rate": 0.00010200463579982098, "loss": 0.9869, "step": 1335 }, { "epoch": 1.97, "learning_rate": 0.0001018867304784467, "loss": 0.9784, "step": 1336 }, { "epoch": 1.97, "learning_rate": 0.00010176882253320967, "loss": 0.9837, "step": 1337 }, { "epoch": 1.97, "learning_rate": 0.0001016509121280836, "loss": 1.039, "step": 1338 }, { "epoch": 1.98, "learning_rate": 0.00010153299942704566, "loss": 0.9984, "step": 1339 }, { "epoch": 1.98, "learning_rate": 0.00010141508459407623, "loss": 1.0526, "step": 1340 }, { "epoch": 1.98, "learning_rate": 0.00010129716779315862, "loss": 1.0581, "step": 1341 }, { "epoch": 1.98, "learning_rate": 0.0001011792491882789, "loss": 1.0607, "step": 1342 }, { "epoch": 1.98, "learning_rate": 0.00010106132894342564, "loss": 0.9734, "step": 1343 }, { "epoch": 1.98, "learning_rate": 0.00010094340722258969, "loss": 0.8872, "step": 1344 }, { "epoch": 1.98, "eval_loss": 0.9883129000663757, "eval_runtime": 2.6191, "eval_samples_per_second": 416.937, "eval_steps_per_second": 26.345, "step": 1344 }, { "epoch": 1.99, "learning_rate": 0.00010082548418976399, "loss": 1.0538, "step": 1345 }, { "epoch": 1.99, "learning_rate": 0.00010070756000894321, "loss": 0.9638, "step": 1346 }, { "epoch": 1.99, "learning_rate": 0.00010058963484412372, "loss": 1.0556, "step": 1347 }, { "epoch": 1.99, "learning_rate": 0.00010047170885930324, "loss": 1.0462, "step": 1348 }, { "epoch": 1.99, "learning_rate": 0.00010035378221848053, "loss": 0.9999, "step": 1349 }, { "epoch": 1.99, "learning_rate": 0.00010023585508565538, "loss": 0.9644, "step": 1350 }, { "epoch": 1.99, "learning_rate": 0.00010011792762482826, "loss": 1.0601, "step": 1351 }, { "epoch": 2.0, "learning_rate": 0.0001, "loss": 0.9745, "step": 1352 }, { "epoch": 2.0, "learning_rate": 9.988207237517178e-05, "loss": 1.0911, "step": 1353 }, { "epoch": 2.0, "learning_rate": 9.976414491434463e-05, "loss": 0.9607, "step": 1354 }, { "epoch": 2.0, "learning_rate": 9.964621778151946e-05, "loss": 1.0151, "step": 1355 }, { "epoch": 2.0, "learning_rate": 9.95282911406968e-05, "loss": 1.036, "step": 1356 }, { "epoch": 2.0, "learning_rate": 9.94103651558763e-05, "loss": 1.0518, "step": 1357 }, { "epoch": 2.0, "learning_rate": 9.929243999105682e-05, "loss": 0.9251, "step": 1358 }, { "epoch": 2.01, "learning_rate": 9.917451581023607e-05, "loss": 0.9899, "step": 1359 }, { "epoch": 2.01, "learning_rate": 9.905659277741032e-05, "loss": 0.9737, "step": 1360 }, { "epoch": 2.01, "learning_rate": 9.89386710565744e-05, "loss": 1.0575, "step": 1361 }, { "epoch": 2.01, "learning_rate": 9.882075081172112e-05, "loss": 1.0074, "step": 1362 }, { "epoch": 2.01, "learning_rate": 9.870283220684142e-05, "loss": 0.9959, "step": 1363 }, { "epoch": 2.01, "learning_rate": 9.858491540592382e-05, "loss": 1.039, "step": 1364 }, { "epoch": 2.01, "learning_rate": 9.846700057295435e-05, "loss": 1.0228, "step": 1365 }, { "epoch": 2.02, "learning_rate": 9.834908787191642e-05, "loss": 1.014, "step": 1366 }, { "epoch": 2.02, "learning_rate": 9.823117746679034e-05, "loss": 0.9919, "step": 1367 }, { "epoch": 2.02, "learning_rate": 9.811326952155331e-05, "loss": 1.0975, "step": 1368 }, { "epoch": 2.02, "learning_rate": 9.799536420017906e-05, "loss": 0.934, "step": 1369 }, { "epoch": 2.0, "learning_rate": 9.787746166663764e-05, "loss": 0.9068, "step": 1370 }, { "epoch": 2.0, "learning_rate": 9.775956208489536e-05, "loss": 0.8948, "step": 1371 }, { "epoch": 2.0, "learning_rate": 9.764166561891432e-05, "loss": 0.9547, "step": 1372 }, { "epoch": 2.01, "learning_rate": 9.752377243265229e-05, "loss": 0.933, "step": 1373 }, { "epoch": 2.01, "learning_rate": 9.740588269006246e-05, "loss": 0.9899, "step": 1374 }, { "epoch": 2.01, "learning_rate": 9.728799655509327e-05, "loss": 0.9059, "step": 1375 }, { "epoch": 2.01, "learning_rate": 9.71701141916882e-05, "loss": 0.9062, "step": 1376 }, { "epoch": 2.01, "learning_rate": 9.705223576378539e-05, "loss": 0.9101, "step": 1377 }, { "epoch": 2.01, "learning_rate": 9.69343614353175e-05, "loss": 0.9635, "step": 1378 }, { "epoch": 2.01, "learning_rate": 9.681649137021158e-05, "loss": 0.8867, "step": 1379 }, { "epoch": 2.02, "learning_rate": 9.669862573238863e-05, "loss": 0.968, "step": 1380 }, { "epoch": 2.02, "learning_rate": 9.658076468576368e-05, "loss": 0.9432, "step": 1381 }, { "epoch": 2.02, "learning_rate": 9.646290839424515e-05, "loss": 0.9186, "step": 1382 }, { "epoch": 2.02, "learning_rate": 9.6345057021735e-05, "loss": 1.038, "step": 1383 }, { "epoch": 2.02, "learning_rate": 9.622721073212832e-05, "loss": 0.9284, "step": 1384 }, { "epoch": 2.02, "learning_rate": 9.610936968931303e-05, "loss": 0.9076, "step": 1385 }, { "epoch": 2.03, "learning_rate": 9.599153405716992e-05, "loss": 0.9196, "step": 1386 }, { "epoch": 2.03, "learning_rate": 9.587370399957217e-05, "loss": 0.9589, "step": 1387 }, { "epoch": 2.03, "learning_rate": 9.57558796803852e-05, "loss": 0.9658, "step": 1388 }, { "epoch": 2.03, "learning_rate": 9.563806126346642e-05, "loss": 0.8864, "step": 1389 }, { "epoch": 2.03, "learning_rate": 9.55202489126651e-05, "loss": 1.0115, "step": 1390 }, { "epoch": 2.03, "learning_rate": 9.540244279182205e-05, "loss": 0.9058, "step": 1391 }, { "epoch": 2.03, "learning_rate": 9.528464306476945e-05, "loss": 0.8686, "step": 1392 }, { "epoch": 2.04, "learning_rate": 9.516684989533051e-05, "loss": 0.9784, "step": 1393 }, { "epoch": 2.04, "learning_rate": 9.504906344731932e-05, "loss": 1.0002, "step": 1394 }, { "epoch": 2.04, "learning_rate": 9.49312838845407e-05, "loss": 0.9873, "step": 1395 }, { "epoch": 2.04, "learning_rate": 9.48135113707899e-05, "loss": 0.9026, "step": 1396 }, { "epoch": 2.04, "learning_rate": 9.46957460698523e-05, "loss": 0.9592, "step": 1397 }, { "epoch": 2.04, "learning_rate": 9.457798814550323e-05, "loss": 0.8996, "step": 1398 }, { "epoch": 2.04, "learning_rate": 9.446023776150787e-05, "loss": 1.0221, "step": 1399 }, { "epoch": 2.05, "learning_rate": 9.434249508162076e-05, "loss": 0.9343, "step": 1400 }, { "epoch": 2.05, "learning_rate": 9.422476026958593e-05, "loss": 0.9033, "step": 1401 }, { "epoch": 2.05, "learning_rate": 9.410703348913627e-05, "loss": 0.9904, "step": 1402 }, { "epoch": 2.05, "learning_rate": 9.398931490399363e-05, "loss": 0.8898, "step": 1403 }, { "epoch": 2.05, "learning_rate": 9.38716046778684e-05, "loss": 0.926, "step": 1404 }, { "epoch": 2.05, "learning_rate": 9.375390297445932e-05, "loss": 0.9515, "step": 1405 }, { "epoch": 2.06, "learning_rate": 9.363620995745337e-05, "loss": 0.9963, "step": 1406 }, { "epoch": 2.06, "learning_rate": 9.351852579052543e-05, "loss": 0.8657, "step": 1407 }, { "epoch": 2.06, "learning_rate": 9.340085063733797e-05, "loss": 0.9137, "step": 1408 }, { "epoch": 2.06, "learning_rate": 9.328318466154102e-05, "loss": 0.9354, "step": 1409 }, { "epoch": 2.06, "learning_rate": 9.316552802677184e-05, "loss": 0.8961, "step": 1410 }, { "epoch": 2.06, "learning_rate": 9.304788089665466e-05, "loss": 0.9796, "step": 1411 }, { "epoch": 2.06, "learning_rate": 9.293024343480055e-05, "loss": 0.8937, "step": 1412 }, { "epoch": 2.07, "learning_rate": 9.281261580480705e-05, "loss": 0.9589, "step": 1413 }, { "epoch": 2.07, "learning_rate": 9.269499817025814e-05, "loss": 1.025, "step": 1414 }, { "epoch": 2.07, "learning_rate": 9.257739069472374e-05, "loss": 0.9527, "step": 1415 }, { "epoch": 2.07, "learning_rate": 9.245979354175985e-05, "loss": 0.9236, "step": 1416 }, { "epoch": 2.07, "learning_rate": 9.234220687490794e-05, "loss": 0.8775, "step": 1417 }, { "epoch": 2.07, "learning_rate": 9.222463085769494e-05, "loss": 0.9171, "step": 1418 }, { "epoch": 2.07, "learning_rate": 9.210706565363305e-05, "loss": 0.8984, "step": 1419 }, { "epoch": 2.08, "learning_rate": 9.198951142621929e-05, "loss": 0.8724, "step": 1420 }, { "epoch": 2.08, "learning_rate": 9.187196833893558e-05, "loss": 0.918, "step": 1421 }, { "epoch": 2.08, "learning_rate": 9.175443655524821e-05, "loss": 0.9457, "step": 1422 }, { "epoch": 2.08, "learning_rate": 9.163691623860784e-05, "loss": 0.9508, "step": 1423 }, { "epoch": 2.08, "learning_rate": 9.151940755244912e-05, "loss": 0.8994, "step": 1424 }, { "epoch": 2.08, "learning_rate": 9.14019106601905e-05, "loss": 0.9118, "step": 1425 }, { "epoch": 2.08, "learning_rate": 9.128442572523417e-05, "loss": 0.8908, "step": 1426 }, { "epoch": 2.09, "learning_rate": 9.11669529109656e-05, "loss": 0.9755, "step": 1427 }, { "epoch": 2.09, "learning_rate": 9.104949238075336e-05, "loss": 0.9706, "step": 1428 }, { "epoch": 2.09, "learning_rate": 9.093204429794898e-05, "loss": 0.9156, "step": 1429 }, { "epoch": 2.09, "learning_rate": 9.081460882588667e-05, "loss": 0.9797, "step": 1430 }, { "epoch": 2.09, "learning_rate": 9.069718612788318e-05, "loss": 0.9402, "step": 1431 }, { "epoch": 2.09, "learning_rate": 9.057977636723739e-05, "loss": 0.9339, "step": 1432 }, { "epoch": 2.1, "learning_rate": 9.04623797072302e-05, "loss": 0.9026, "step": 1433 }, { "epoch": 2.1, "learning_rate": 9.034499631112437e-05, "loss": 0.9424, "step": 1434 }, { "epoch": 2.1, "learning_rate": 9.022762634216409e-05, "loss": 0.948, "step": 1435 }, { "epoch": 2.1, "learning_rate": 9.011026996357503e-05, "loss": 0.9323, "step": 1436 }, { "epoch": 2.1, "learning_rate": 8.999292733856385e-05, "loss": 1.0037, "step": 1437 }, { "epoch": 2.1, "learning_rate": 8.987559863031808e-05, "loss": 0.9549, "step": 1438 }, { "epoch": 2.1, "learning_rate": 8.975828400200592e-05, "loss": 0.915, "step": 1439 }, { "epoch": 2.11, "learning_rate": 8.964098361677607e-05, "loss": 0.9074, "step": 1440 }, { "epoch": 2.11, "learning_rate": 8.952369763775733e-05, "loss": 0.843, "step": 1441 }, { "epoch": 2.11, "learning_rate": 8.94064262280584e-05, "loss": 0.8654, "step": 1442 }, { "epoch": 2.11, "learning_rate": 8.928916955076792e-05, "loss": 0.9349, "step": 1443 }, { "epoch": 2.11, "learning_rate": 8.917192776895382e-05, "loss": 0.9353, "step": 1444 }, { "epoch": 2.11, "learning_rate": 8.90547010456635e-05, "loss": 0.9137, "step": 1445 }, { "epoch": 2.11, "learning_rate": 8.89374895439233e-05, "loss": 0.9528, "step": 1446 }, { "epoch": 2.12, "learning_rate": 8.882029342673844e-05, "loss": 0.8906, "step": 1447 }, { "epoch": 2.12, "learning_rate": 8.870311285709274e-05, "loss": 0.9351, "step": 1448 }, { "epoch": 2.12, "learning_rate": 8.858594799794835e-05, "loss": 1.0043, "step": 1449 }, { "epoch": 2.12, "learning_rate": 8.846879901224566e-05, "loss": 0.8462, "step": 1450 }, { "epoch": 2.12, "learning_rate": 8.835166606290295e-05, "loss": 1.01, "step": 1451 }, { "epoch": 2.12, "learning_rate": 8.823454931281616e-05, "loss": 0.8858, "step": 1452 }, { "epoch": 2.13, "learning_rate": 8.811744892485871e-05, "loss": 0.9954, "step": 1453 }, { "epoch": 2.13, "learning_rate": 8.800036506188129e-05, "loss": 0.9609, "step": 1454 }, { "epoch": 2.13, "learning_rate": 8.788329788671167e-05, "loss": 0.8891, "step": 1455 }, { "epoch": 2.13, "learning_rate": 8.776624756215429e-05, "loss": 0.9669, "step": 1456 }, { "epoch": 2.13, "learning_rate": 8.76492142509902e-05, "loss": 0.981, "step": 1457 }, { "epoch": 2.13, "learning_rate": 8.753219811597683e-05, "loss": 0.9607, "step": 1458 }, { "epoch": 2.13, "learning_rate": 8.741519931984766e-05, "loss": 0.8752, "step": 1459 }, { "epoch": 2.14, "learning_rate": 8.729821802531212e-05, "loss": 0.8649, "step": 1460 }, { "epoch": 2.14, "learning_rate": 8.71812543950553e-05, "loss": 0.9491, "step": 1461 }, { "epoch": 2.14, "learning_rate": 8.706430859173763e-05, "loss": 0.9335, "step": 1462 }, { "epoch": 2.14, "learning_rate": 8.694738077799488e-05, "loss": 0.9752, "step": 1463 }, { "epoch": 2.14, "learning_rate": 8.683047111643763e-05, "loss": 0.987, "step": 1464 }, { "epoch": 2.14, "learning_rate": 8.671357976965147e-05, "loss": 0.9728, "step": 1465 }, { "epoch": 2.14, "learning_rate": 8.659670690019625e-05, "loss": 0.8742, "step": 1466 }, { "epoch": 2.15, "learning_rate": 8.647985267060635e-05, "loss": 0.8658, "step": 1467 }, { "epoch": 2.15, "learning_rate": 8.636301724339004e-05, "loss": 1.0111, "step": 1468 }, { "epoch": 2.15, "learning_rate": 8.624620078102951e-05, "loss": 0.865, "step": 1469 }, { "epoch": 2.15, "learning_rate": 8.612940344598067e-05, "loss": 0.9406, "step": 1470 }, { "epoch": 2.15, "learning_rate": 8.601262540067274e-05, "loss": 0.9647, "step": 1471 }, { "epoch": 2.15, "learning_rate": 8.58958668075081e-05, "loss": 0.9346, "step": 1472 }, { "epoch": 2.15, "learning_rate": 8.577912782886206e-05, "loss": 0.9357, "step": 1473 }, { "epoch": 2.16, "learning_rate": 8.566240862708274e-05, "loss": 0.9226, "step": 1474 }, { "epoch": 2.16, "learning_rate": 8.554570936449074e-05, "loss": 0.8791, "step": 1475 }, { "epoch": 2.16, "learning_rate": 8.542903020337887e-05, "loss": 0.9111, "step": 1476 }, { "epoch": 2.16, "learning_rate": 8.531237130601199e-05, "loss": 0.928, "step": 1477 }, { "epoch": 2.16, "learning_rate": 8.519573283462687e-05, "loss": 0.9075, "step": 1478 }, { "epoch": 2.16, "learning_rate": 8.507911495143173e-05, "loss": 0.9911, "step": 1479 }, { "epoch": 2.17, "learning_rate": 8.496251781860633e-05, "loss": 0.9687, "step": 1480 }, { "epoch": 2.17, "learning_rate": 8.484594159830141e-05, "loss": 0.8915, "step": 1481 }, { "epoch": 2.17, "learning_rate": 8.472938645263875e-05, "loss": 0.9551, "step": 1482 }, { "epoch": 2.17, "learning_rate": 8.461285254371073e-05, "loss": 0.902, "step": 1483 }, { "epoch": 2.17, "learning_rate": 8.449634003358022e-05, "loss": 0.963, "step": 1484 }, { "epoch": 2.17, "learning_rate": 8.437984908428041e-05, "loss": 0.8784, "step": 1485 }, { "epoch": 2.17, "learning_rate": 8.426337985781438e-05, "loss": 0.9793, "step": 1486 }, { "epoch": 2.18, "learning_rate": 8.414693251615512e-05, "loss": 0.9435, "step": 1487 }, { "epoch": 2.18, "learning_rate": 8.403050722124509e-05, "loss": 0.9944, "step": 1488 }, { "epoch": 2.18, "learning_rate": 8.39141041349961e-05, "loss": 0.9912, "step": 1489 }, { "epoch": 2.18, "learning_rate": 8.379772341928915e-05, "loss": 0.8934, "step": 1490 }, { "epoch": 2.18, "learning_rate": 8.36813652359741e-05, "loss": 0.9868, "step": 1491 }, { "epoch": 2.18, "learning_rate": 8.356502974686941e-05, "loss": 0.8958, "step": 1492 }, { "epoch": 2.18, "learning_rate": 8.344871711376203e-05, "loss": 0.9075, "step": 1493 }, { "epoch": 2.19, "learning_rate": 8.33324274984071e-05, "loss": 0.954, "step": 1494 }, { "epoch": 2.19, "learning_rate": 8.321616106252783e-05, "loss": 0.9316, "step": 1495 }, { "epoch": 2.19, "learning_rate": 8.309991796781511e-05, "loss": 0.9949, "step": 1496 }, { "epoch": 2.19, "learning_rate": 8.298369837592735e-05, "loss": 1.0344, "step": 1497 }, { "epoch": 2.19, "learning_rate": 8.286750244849038e-05, "loss": 0.9034, "step": 1498 }, { "epoch": 2.19, "learning_rate": 8.275133034709699e-05, "loss": 0.9102, "step": 1499 }, { "epoch": 2.2, "learning_rate": 8.263518223330697e-05, "loss": 0.9013, "step": 1500 }, { "epoch": 2.2, "learning_rate": 8.251905826864665e-05, "loss": 1.0105, "step": 1501 }, { "epoch": 2.2, "learning_rate": 8.240295861460881e-05, "loss": 0.9215, "step": 1502 }, { "epoch": 2.2, "learning_rate": 8.228688343265242e-05, "loss": 0.8733, "step": 1503 }, { "epoch": 2.2, "learning_rate": 8.217083288420241e-05, "loss": 0.9214, "step": 1504 }, { "epoch": 2.2, "learning_rate": 8.205480713064946e-05, "loss": 0.9535, "step": 1505 }, { "epoch": 2.2, "learning_rate": 8.193880633334982e-05, "loss": 0.9155, "step": 1506 }, { "epoch": 2.21, "learning_rate": 8.182283065362493e-05, "loss": 0.9045, "step": 1507 }, { "epoch": 2.21, "learning_rate": 8.170688025276134e-05, "loss": 1.0051, "step": 1508 }, { "epoch": 2.21, "learning_rate": 8.159095529201049e-05, "loss": 0.9856, "step": 1509 }, { "epoch": 2.21, "learning_rate": 8.147505593258837e-05, "loss": 0.9189, "step": 1510 }, { "epoch": 2.21, "learning_rate": 8.135918233567545e-05, "loss": 0.9455, "step": 1511 }, { "epoch": 2.21, "learning_rate": 8.124333466241632e-05, "loss": 0.9351, "step": 1512 }, { "epoch": 2.21, "eval_loss": 0.9985308051109314, "eval_runtime": 2.6183, "eval_samples_per_second": 417.062, "eval_steps_per_second": 26.353, "step": 1512 }, { "epoch": 2.21, "learning_rate": 8.112751307391946e-05, "loss": 0.9408, "step": 1513 }, { "epoch": 2.22, "learning_rate": 8.101171773125716e-05, "loss": 0.8725, "step": 1514 }, { "epoch": 2.22, "learning_rate": 8.089594879546525e-05, "loss": 0.9437, "step": 1515 }, { "epoch": 2.22, "learning_rate": 8.078020642754274e-05, "loss": 0.904, "step": 1516 }, { "epoch": 2.22, "learning_rate": 8.066449078845168e-05, "loss": 0.93, "step": 1517 }, { "epoch": 2.22, "learning_rate": 8.054880203911705e-05, "loss": 1.0044, "step": 1518 }, { "epoch": 2.22, "learning_rate": 8.043314034042631e-05, "loss": 0.957, "step": 1519 }, { "epoch": 2.23, "learning_rate": 8.031750585322947e-05, "loss": 0.9136, "step": 1520 }, { "epoch": 2.23, "learning_rate": 8.020189873833852e-05, "loss": 0.9619, "step": 1521 }, { "epoch": 2.23, "learning_rate": 8.00863191565275e-05, "loss": 0.961, "step": 1522 }, { "epoch": 2.23, "learning_rate": 7.99707672685321e-05, "loss": 0.8426, "step": 1523 }, { "epoch": 2.23, "learning_rate": 7.985524323504948e-05, "loss": 0.9201, "step": 1524 }, { "epoch": 2.23, "learning_rate": 7.973974721673815e-05, "loss": 0.8758, "step": 1525 }, { "epoch": 2.23, "learning_rate": 7.962427937421763e-05, "loss": 0.8926, "step": 1526 }, { "epoch": 2.24, "learning_rate": 7.950883986806821e-05, "loss": 0.9891, "step": 1527 }, { "epoch": 2.24, "learning_rate": 7.939342885883076e-05, "loss": 0.9544, "step": 1528 }, { "epoch": 2.24, "learning_rate": 7.927804650700659e-05, "loss": 0.9669, "step": 1529 }, { "epoch": 2.24, "learning_rate": 7.916269297305712e-05, "loss": 0.9712, "step": 1530 }, { "epoch": 2.24, "learning_rate": 7.90473684174037e-05, "loss": 0.952, "step": 1531 }, { "epoch": 2.24, "learning_rate": 7.89320730004274e-05, "loss": 0.9196, "step": 1532 }, { "epoch": 2.24, "learning_rate": 7.881680688246869e-05, "loss": 1.0468, "step": 1533 }, { "epoch": 2.25, "learning_rate": 7.870157022382735e-05, "loss": 0.9278, "step": 1534 }, { "epoch": 2.25, "learning_rate": 7.858636318476226e-05, "loss": 0.9395, "step": 1535 }, { "epoch": 2.25, "learning_rate": 7.847118592549099e-05, "loss": 0.9261, "step": 1536 }, { "epoch": 2.25, "learning_rate": 7.835603860618972e-05, "loss": 0.8694, "step": 1537 }, { "epoch": 2.25, "learning_rate": 7.824092138699307e-05, "loss": 1.0352, "step": 1538 }, { "epoch": 2.25, "learning_rate": 7.812583442799368e-05, "loss": 0.8773, "step": 1539 }, { "epoch": 2.25, "learning_rate": 7.801077788924224e-05, "loss": 0.8136, "step": 1540 }, { "epoch": 2.26, "learning_rate": 7.789575193074704e-05, "loss": 0.9875, "step": 1541 }, { "epoch": 2.26, "learning_rate": 7.778075671247385e-05, "loss": 0.9006, "step": 1542 }, { "epoch": 2.26, "learning_rate": 7.766579239434575e-05, "loss": 0.9503, "step": 1543 }, { "epoch": 2.26, "learning_rate": 7.755085913624274e-05, "loss": 0.9282, "step": 1544 }, { "epoch": 2.26, "learning_rate": 7.743595709800176e-05, "loss": 0.9057, "step": 1545 }, { "epoch": 2.26, "learning_rate": 7.732108643941625e-05, "loss": 0.8577, "step": 1546 }, { "epoch": 2.27, "learning_rate": 7.720624732023603e-05, "loss": 1.0027, "step": 1547 }, { "epoch": 2.27, "learning_rate": 7.709143990016702e-05, "loss": 0.8771, "step": 1548 }, { "epoch": 2.27, "learning_rate": 7.697666433887108e-05, "loss": 0.9449, "step": 1549 }, { "epoch": 2.27, "learning_rate": 7.686192079596586e-05, "loss": 0.9362, "step": 1550 }, { "epoch": 2.27, "learning_rate": 7.674720943102432e-05, "loss": 0.9937, "step": 1551 }, { "epoch": 2.27, "learning_rate": 7.663253040357476e-05, "loss": 0.9218, "step": 1552 }, { "epoch": 2.27, "learning_rate": 7.651788387310052e-05, "loss": 0.9557, "step": 1553 }, { "epoch": 2.28, "learning_rate": 7.640326999903967e-05, "loss": 0.9602, "step": 1554 }, { "epoch": 2.28, "learning_rate": 7.628868894078501e-05, "loss": 0.9414, "step": 1555 }, { "epoch": 2.28, "learning_rate": 7.617414085768351e-05, "loss": 0.8552, "step": 1556 }, { "epoch": 2.28, "learning_rate": 7.605962590903643e-05, "loss": 0.9802, "step": 1557 }, { "epoch": 2.28, "learning_rate": 7.594514425409893e-05, "loss": 0.9204, "step": 1558 }, { "epoch": 2.28, "learning_rate": 7.583069605207975e-05, "loss": 0.9143, "step": 1559 }, { "epoch": 2.28, "learning_rate": 7.571628146214129e-05, "loss": 0.9383, "step": 1560 }, { "epoch": 2.29, "learning_rate": 7.560190064339908e-05, "loss": 0.9052, "step": 1561 }, { "epoch": 2.29, "learning_rate": 7.548755375492172e-05, "loss": 0.9063, "step": 1562 }, { "epoch": 2.29, "learning_rate": 7.537324095573064e-05, "loss": 0.9549, "step": 1563 }, { "epoch": 2.29, "learning_rate": 7.525896240479976e-05, "loss": 0.9499, "step": 1564 }, { "epoch": 2.29, "learning_rate": 7.514471826105556e-05, "loss": 0.8888, "step": 1565 }, { "epoch": 2.29, "learning_rate": 7.503050868337653e-05, "loss": 0.9178, "step": 1566 }, { "epoch": 2.3, "learning_rate": 7.491633383059313e-05, "loss": 0.9875, "step": 1567 }, { "epoch": 2.3, "learning_rate": 7.48021938614875e-05, "loss": 0.8715, "step": 1568 }, { "epoch": 2.3, "learning_rate": 7.468808893479327e-05, "loss": 0.9246, "step": 1569 }, { "epoch": 2.3, "learning_rate": 7.457401920919544e-05, "loss": 0.8635, "step": 1570 }, { "epoch": 2.3, "learning_rate": 7.445998484332993e-05, "loss": 0.95, "step": 1571 }, { "epoch": 2.3, "learning_rate": 7.434598599578351e-05, "loss": 0.9175, "step": 1572 }, { "epoch": 2.3, "learning_rate": 7.42320228250936e-05, "loss": 0.9439, "step": 1573 }, { "epoch": 2.31, "learning_rate": 7.411809548974792e-05, "loss": 0.9417, "step": 1574 }, { "epoch": 2.31, "learning_rate": 7.400420414818451e-05, "loss": 0.9001, "step": 1575 }, { "epoch": 2.31, "learning_rate": 7.389034895879118e-05, "loss": 0.8844, "step": 1576 }, { "epoch": 2.31, "learning_rate": 7.37765300799056e-05, "loss": 0.9377, "step": 1577 }, { "epoch": 2.31, "learning_rate": 7.366274766981483e-05, "loss": 0.9851, "step": 1578 }, { "epoch": 2.31, "learning_rate": 7.354900188675525e-05, "loss": 0.965, "step": 1579 }, { "epoch": 2.31, "learning_rate": 7.343529288891239e-05, "loss": 0.9737, "step": 1580 }, { "epoch": 2.32, "learning_rate": 7.332162083442049e-05, "loss": 0.9686, "step": 1581 }, { "epoch": 2.32, "learning_rate": 7.320798588136253e-05, "loss": 0.968, "step": 1582 }, { "epoch": 2.32, "learning_rate": 7.309438818776981e-05, "loss": 0.8647, "step": 1583 }, { "epoch": 2.32, "learning_rate": 7.29808279116218e-05, "loss": 0.8572, "step": 1584 }, { "epoch": 2.32, "learning_rate": 7.286730521084602e-05, "loss": 0.9264, "step": 1585 }, { "epoch": 2.32, "learning_rate": 7.275382024331772e-05, "loss": 0.9601, "step": 1586 }, { "epoch": 2.32, "learning_rate": 7.264037316685962e-05, "loss": 0.9996, "step": 1587 }, { "epoch": 2.33, "learning_rate": 7.252696413924174e-05, "loss": 0.9598, "step": 1588 }, { "epoch": 2.33, "learning_rate": 7.24135933181812e-05, "loss": 0.8675, "step": 1589 }, { "epoch": 2.33, "learning_rate": 7.23002608613421e-05, "loss": 0.9298, "step": 1590 }, { "epoch": 2.33, "learning_rate": 7.218696692633501e-05, "loss": 1.0448, "step": 1591 }, { "epoch": 2.33, "learning_rate": 7.2073711670717e-05, "loss": 0.9525, "step": 1592 }, { "epoch": 2.33, "learning_rate": 7.196049525199142e-05, "loss": 0.9621, "step": 1593 }, { "epoch": 2.34, "learning_rate": 7.184731782760746e-05, "loss": 0.933, "step": 1594 }, { "epoch": 2.34, "learning_rate": 7.173417955496024e-05, "loss": 1.0498, "step": 1595 }, { "epoch": 2.34, "learning_rate": 7.162108059139032e-05, "loss": 0.9508, "step": 1596 }, { "epoch": 2.34, "learning_rate": 7.150802109418366e-05, "loss": 0.9854, "step": 1597 }, { "epoch": 2.34, "learning_rate": 7.13950012205713e-05, "loss": 0.971, "step": 1598 }, { "epoch": 2.34, "learning_rate": 7.128202112772912e-05, "loss": 0.9346, "step": 1599 }, { "epoch": 2.34, "learning_rate": 7.116908097277781e-05, "loss": 1.0434, "step": 1600 }, { "epoch": 2.35, "learning_rate": 7.105618091278245e-05, "loss": 0.9354, "step": 1601 }, { "epoch": 2.35, "learning_rate": 7.094332110475234e-05, "loss": 0.9009, "step": 1602 }, { "epoch": 2.35, "learning_rate": 7.083050170564077e-05, "loss": 0.9379, "step": 1603 }, { "epoch": 2.35, "learning_rate": 7.071772287234497e-05, "loss": 0.9488, "step": 1604 }, { "epoch": 2.35, "learning_rate": 7.06049847617056e-05, "loss": 0.8778, "step": 1605 }, { "epoch": 2.35, "learning_rate": 7.049228753050681e-05, "loss": 0.8452, "step": 1606 }, { "epoch": 2.35, "learning_rate": 7.037963133547583e-05, "loss": 0.8394, "step": 1607 }, { "epoch": 2.36, "learning_rate": 7.026701633328276e-05, "loss": 0.8895, "step": 1608 }, { "epoch": 2.36, "learning_rate": 7.015444268054059e-05, "loss": 0.9667, "step": 1609 }, { "epoch": 2.36, "learning_rate": 7.004191053380469e-05, "loss": 0.9573, "step": 1610 }, { "epoch": 2.36, "learning_rate": 6.992942004957271e-05, "loss": 1.0102, "step": 1611 }, { "epoch": 2.36, "learning_rate": 6.981697138428434e-05, "loss": 0.9507, "step": 1612 }, { "epoch": 2.36, "learning_rate": 6.970456469432117e-05, "loss": 0.9597, "step": 1613 }, { "epoch": 2.37, "learning_rate": 6.959220013600641e-05, "loss": 0.8432, "step": 1614 }, { "epoch": 2.37, "learning_rate": 6.947987786560466e-05, "loss": 0.9855, "step": 1615 }, { "epoch": 2.37, "learning_rate": 6.936759803932167e-05, "loss": 0.922, "step": 1616 }, { "epoch": 2.37, "learning_rate": 6.925536081330424e-05, "loss": 0.9261, "step": 1617 }, { "epoch": 2.37, "learning_rate": 6.914316634363984e-05, "loss": 0.9046, "step": 1618 }, { "epoch": 2.37, "learning_rate": 6.903101478635662e-05, "loss": 0.93, "step": 1619 }, { "epoch": 2.37, "learning_rate": 6.891890629742288e-05, "loss": 0.8668, "step": 1620 }, { "epoch": 2.38, "learning_rate": 6.880684103274715e-05, "loss": 0.9852, "step": 1621 }, { "epoch": 2.38, "learning_rate": 6.869481914817779e-05, "loss": 0.976, "step": 1622 }, { "epoch": 2.38, "learning_rate": 6.85828407995028e-05, "loss": 0.9287, "step": 1623 }, { "epoch": 2.38, "learning_rate": 6.847090614244977e-05, "loss": 0.9192, "step": 1624 }, { "epoch": 2.38, "learning_rate": 6.835901533268536e-05, "loss": 0.9999, "step": 1625 }, { "epoch": 2.38, "learning_rate": 6.824716852581539e-05, "loss": 0.8986, "step": 1626 }, { "epoch": 2.38, "learning_rate": 6.813536587738436e-05, "loss": 0.9818, "step": 1627 }, { "epoch": 2.39, "learning_rate": 6.802360754287547e-05, "loss": 0.9658, "step": 1628 }, { "epoch": 2.39, "learning_rate": 6.791189367771025e-05, "loss": 0.8793, "step": 1629 }, { "epoch": 2.39, "learning_rate": 6.780022443724839e-05, "loss": 0.9363, "step": 1630 }, { "epoch": 2.39, "learning_rate": 6.768859997678751e-05, "loss": 0.9108, "step": 1631 }, { "epoch": 2.39, "learning_rate": 6.757702045156292e-05, "loss": 0.9561, "step": 1632 }, { "epoch": 2.39, "learning_rate": 6.74654860167475e-05, "loss": 0.8761, "step": 1633 }, { "epoch": 2.39, "learning_rate": 6.735399682745145e-05, "loss": 0.9307, "step": 1634 }, { "epoch": 2.4, "learning_rate": 6.724255303872197e-05, "loss": 0.8416, "step": 1635 }, { "epoch": 2.4, "learning_rate": 6.713115480554313e-05, "loss": 0.908, "step": 1636 }, { "epoch": 2.4, "learning_rate": 6.701980228283568e-05, "loss": 0.9147, "step": 1637 }, { "epoch": 2.4, "learning_rate": 6.690849562545678e-05, "loss": 0.9877, "step": 1638 }, { "epoch": 2.4, "learning_rate": 6.679723498819986e-05, "loss": 0.9294, "step": 1639 }, { "epoch": 2.4, "learning_rate": 6.668602052579424e-05, "loss": 0.929, "step": 1640 }, { "epoch": 2.41, "learning_rate": 6.657485239290515e-05, "loss": 0.8996, "step": 1641 }, { "epoch": 2.41, "learning_rate": 6.64637307441333e-05, "loss": 0.8994, "step": 1642 }, { "epoch": 2.41, "learning_rate": 6.635265573401474e-05, "loss": 0.9253, "step": 1643 }, { "epoch": 2.41, "learning_rate": 6.624162751702076e-05, "loss": 0.8867, "step": 1644 }, { "epoch": 2.41, "learning_rate": 6.613064624755753e-05, "loss": 0.987, "step": 1645 }, { "epoch": 2.41, "learning_rate": 6.601971207996591e-05, "loss": 0.9079, "step": 1646 }, { "epoch": 2.41, "learning_rate": 6.590882516852122e-05, "loss": 0.9696, "step": 1647 }, { "epoch": 2.42, "learning_rate": 6.579798566743314e-05, "loss": 0.964, "step": 1648 }, { "epoch": 2.42, "learning_rate": 6.568719373084538e-05, "loss": 0.9736, "step": 1649 }, { "epoch": 2.42, "learning_rate": 6.557644951283551e-05, "loss": 0.9806, "step": 1650 }, { "epoch": 2.42, "learning_rate": 6.546575316741474e-05, "loss": 1.0062, "step": 1651 }, { "epoch": 2.42, "learning_rate": 6.535510484852767e-05, "loss": 0.8794, "step": 1652 }, { "epoch": 2.42, "learning_rate": 6.524450471005213e-05, "loss": 0.8359, "step": 1653 }, { "epoch": 2.42, "learning_rate": 6.513395290579901e-05, "loss": 0.9442, "step": 1654 }, { "epoch": 2.43, "learning_rate": 6.50234495895119e-05, "loss": 0.8903, "step": 1655 }, { "epoch": 2.43, "learning_rate": 6.491299491486695e-05, "loss": 0.9625, "step": 1656 }, { "epoch": 2.43, "learning_rate": 6.480258903547276e-05, "loss": 0.929, "step": 1657 }, { "epoch": 2.43, "learning_rate": 6.469223210486992e-05, "loss": 0.9751, "step": 1658 }, { "epoch": 2.43, "learning_rate": 6.458192427653112e-05, "loss": 0.982, "step": 1659 }, { "epoch": 2.43, "learning_rate": 6.447166570386063e-05, "loss": 0.8681, "step": 1660 }, { "epoch": 2.44, "learning_rate": 6.436145654019432e-05, "loss": 0.9149, "step": 1661 }, { "epoch": 2.44, "learning_rate": 6.425129693879925e-05, "loss": 0.9411, "step": 1662 }, { "epoch": 2.44, "learning_rate": 6.414118705287359e-05, "loss": 0.9072, "step": 1663 }, { "epoch": 2.44, "learning_rate": 6.403112703554643e-05, "loss": 0.9784, "step": 1664 }, { "epoch": 2.44, "learning_rate": 6.392111703987744e-05, "loss": 0.8754, "step": 1665 }, { "epoch": 2.44, "learning_rate": 6.381115721885675e-05, "loss": 1.0216, "step": 1666 }, { "epoch": 2.44, "learning_rate": 6.370124772540469e-05, "loss": 0.8493, "step": 1667 }, { "epoch": 2.45, "learning_rate": 6.35913887123716e-05, "loss": 0.9215, "step": 1668 }, { "epoch": 2.45, "learning_rate": 6.348158033253773e-05, "loss": 0.919, "step": 1669 }, { "epoch": 2.45, "learning_rate": 6.337182273861273e-05, "loss": 0.9128, "step": 1670 }, { "epoch": 2.45, "learning_rate": 6.326211608323573e-05, "loss": 0.9469, "step": 1671 }, { "epoch": 2.45, "learning_rate": 6.315246051897503e-05, "loss": 0.8582, "step": 1672 }, { "epoch": 2.45, "learning_rate": 6.30428561983278e-05, "loss": 0.8487, "step": 1673 }, { "epoch": 2.45, "learning_rate": 6.293330327372005e-05, "loss": 0.8687, "step": 1674 }, { "epoch": 2.46, "learning_rate": 6.282380189750625e-05, "loss": 0.9028, "step": 1675 }, { "epoch": 2.46, "learning_rate": 6.271435222196916e-05, "loss": 0.9105, "step": 1676 }, { "epoch": 2.46, "learning_rate": 6.26049543993197e-05, "loss": 0.9384, "step": 1677 }, { "epoch": 2.46, "learning_rate": 6.249560858169661e-05, "loss": 0.9658, "step": 1678 }, { "epoch": 2.46, "learning_rate": 6.238631492116644e-05, "loss": 0.9193, "step": 1679 }, { "epoch": 2.46, "learning_rate": 6.227707356972301e-05, "loss": 0.9077, "step": 1680 }, { "epoch": 2.46, "eval_loss": 0.9968231916427612, "eval_runtime": 2.6101, "eval_samples_per_second": 418.377, "eval_steps_per_second": 26.436, "step": 1680 }, { "epoch": 2.46, "learning_rate": 6.216788467928758e-05, "loss": 0.9083, "step": 1681 }, { "epoch": 2.47, "learning_rate": 6.205874840170833e-05, "loss": 0.9397, "step": 1682 }, { "epoch": 2.47, "learning_rate": 6.194966488876027e-05, "loss": 0.9631, "step": 1683 }, { "epoch": 2.47, "learning_rate": 6.184063429214515e-05, "loss": 0.9214, "step": 1684 }, { "epoch": 2.47, "learning_rate": 6.173165676349103e-05, "loss": 0.9544, "step": 1685 }, { "epoch": 2.47, "learning_rate": 6.162273245435219e-05, "loss": 0.9303, "step": 1686 }, { "epoch": 2.47, "learning_rate": 6.151386151620887e-05, "loss": 0.8605, "step": 1687 }, { "epoch": 2.48, "learning_rate": 6.140504410046712e-05, "loss": 0.9125, "step": 1688 }, { "epoch": 2.48, "learning_rate": 6.129628035845861e-05, "loss": 0.9202, "step": 1689 }, { "epoch": 2.48, "learning_rate": 6.118757044144025e-05, "loss": 0.937, "step": 1690 }, { "epoch": 2.48, "learning_rate": 6.107891450059419e-05, "loss": 0.8941, "step": 1691 }, { "epoch": 2.48, "learning_rate": 6.097031268702746e-05, "loss": 0.9205, "step": 1692 }, { "epoch": 2.48, "learning_rate": 6.086176515177182e-05, "loss": 0.9164, "step": 1693 }, { "epoch": 2.48, "learning_rate": 6.0753272045783625e-05, "loss": 0.9261, "step": 1694 }, { "epoch": 2.49, "learning_rate": 6.0644833519943425e-05, "loss": 0.8642, "step": 1695 }, { "epoch": 2.49, "learning_rate": 6.053644972505593e-05, "loss": 0.9648, "step": 1696 }, { "epoch": 2.49, "learning_rate": 6.042812081184972e-05, "loss": 0.9213, "step": 1697 }, { "epoch": 2.49, "learning_rate": 6.0319846930977e-05, "loss": 0.9314, "step": 1698 }, { "epoch": 2.49, "learning_rate": 6.021162823301358e-05, "loss": 0.9055, "step": 1699 }, { "epoch": 2.49, "learning_rate": 6.010346486845837e-05, "loss": 0.9335, "step": 1700 }, { "epoch": 2.49, "learning_rate": 5.9995356987733466e-05, "loss": 0.9365, "step": 1701 }, { "epoch": 2.5, "learning_rate": 5.988730474118367e-05, "loss": 0.9817, "step": 1702 }, { "epoch": 2.5, "learning_rate": 5.977930827907649e-05, "loss": 0.9145, "step": 1703 }, { "epoch": 2.5, "learning_rate": 5.967136775160187e-05, "loss": 0.9974, "step": 1704 }, { "epoch": 2.5, "learning_rate": 5.956348330887196e-05, "loss": 0.9577, "step": 1705 }, { "epoch": 2.5, "learning_rate": 5.945565510092086e-05, "loss": 0.8716, "step": 1706 }, { "epoch": 2.5, "learning_rate": 5.93478832777045e-05, "loss": 0.9541, "step": 1707 }, { "epoch": 2.51, "learning_rate": 5.924016798910037e-05, "loss": 0.9464, "step": 1708 }, { "epoch": 2.51, "learning_rate": 5.913250938490744e-05, "loss": 0.9774, "step": 1709 }, { "epoch": 2.51, "learning_rate": 5.9024907614845716e-05, "loss": 0.8654, "step": 1710 }, { "epoch": 2.51, "learning_rate": 5.891736282855622e-05, "loss": 0.8559, "step": 1711 }, { "epoch": 2.51, "learning_rate": 5.880987517560075e-05, "loss": 1.0032, "step": 1712 }, { "epoch": 2.51, "learning_rate": 5.870244480546159e-05, "loss": 0.9373, "step": 1713 }, { "epoch": 2.51, "learning_rate": 5.859507186754146e-05, "loss": 0.8892, "step": 1714 }, { "epoch": 2.52, "learning_rate": 5.848775651116309e-05, "loss": 0.8809, "step": 1715 }, { "epoch": 2.52, "learning_rate": 5.838049888556925e-05, "loss": 0.9596, "step": 1716 }, { "epoch": 2.52, "learning_rate": 5.827329913992232e-05, "loss": 0.95, "step": 1717 }, { "epoch": 2.52, "learning_rate": 5.81661574233042e-05, "loss": 0.9694, "step": 1718 }, { "epoch": 2.52, "learning_rate": 5.80590738847162e-05, "loss": 0.9333, "step": 1719 }, { "epoch": 2.52, "learning_rate": 5.79520486730786e-05, "loss": 0.943, "step": 1720 }, { "epoch": 2.52, "learning_rate": 5.784508193723057e-05, "loss": 0.906, "step": 1721 }, { "epoch": 2.53, "learning_rate": 5.773817382593008e-05, "loss": 0.9575, "step": 1722 }, { "epoch": 2.53, "learning_rate": 5.763132448785339e-05, "loss": 0.8638, "step": 1723 }, { "epoch": 2.53, "learning_rate": 5.752453407159522e-05, "loss": 0.9667, "step": 1724 }, { "epoch": 2.53, "learning_rate": 5.741780272566821e-05, "loss": 1.0119, "step": 1725 }, { "epoch": 2.53, "learning_rate": 5.7311130598502885e-05, "loss": 0.8563, "step": 1726 }, { "epoch": 2.53, "learning_rate": 5.7204517838447405e-05, "loss": 0.9852, "step": 1727 }, { "epoch": 2.54, "learning_rate": 5.7097964593767375e-05, "loss": 0.8659, "step": 1728 }, { "epoch": 2.54, "learning_rate": 5.699147101264566e-05, "loss": 0.9579, "step": 1729 }, { "epoch": 2.54, "learning_rate": 5.688503724318217e-05, "loss": 0.9226, "step": 1730 }, { "epoch": 2.54, "learning_rate": 5.6778663433393574e-05, "loss": 0.9771, "step": 1731 }, { "epoch": 2.54, "learning_rate": 5.667234973121317e-05, "loss": 0.8307, "step": 1732 }, { "epoch": 2.54, "learning_rate": 5.6566096284490635e-05, "loss": 0.9012, "step": 1733 }, { "epoch": 2.54, "learning_rate": 5.645990324099197e-05, "loss": 0.8566, "step": 1734 }, { "epoch": 2.55, "learning_rate": 5.635377074839907e-05, "loss": 0.941, "step": 1735 }, { "epoch": 2.55, "learning_rate": 5.624769895430961e-05, "loss": 0.8925, "step": 1736 }, { "epoch": 2.55, "learning_rate": 5.614168800623687e-05, "loss": 0.9387, "step": 1737 }, { "epoch": 2.55, "learning_rate": 5.6035738051609555e-05, "loss": 0.9025, "step": 1738 }, { "epoch": 2.55, "learning_rate": 5.5929849237771556e-05, "loss": 0.9394, "step": 1739 }, { "epoch": 2.55, "learning_rate": 5.5824021711981686e-05, "loss": 0.9743, "step": 1740 }, { "epoch": 2.55, "learning_rate": 5.5718255621413526e-05, "loss": 0.9604, "step": 1741 }, { "epoch": 2.56, "learning_rate": 5.561255111315524e-05, "loss": 0.9223, "step": 1742 }, { "epoch": 2.56, "learning_rate": 5.550690833420928e-05, "loss": 0.8748, "step": 1743 }, { "epoch": 2.56, "learning_rate": 5.540132743149242e-05, "loss": 1.0009, "step": 1744 }, { "epoch": 2.56, "learning_rate": 5.5295808551835184e-05, "loss": 0.9533, "step": 1745 }, { "epoch": 2.56, "learning_rate": 5.5190351841982014e-05, "loss": 0.9026, "step": 1746 }, { "epoch": 2.56, "learning_rate": 5.508495744859077e-05, "loss": 0.9514, "step": 1747 }, { "epoch": 2.56, "learning_rate": 5.497962551823266e-05, "loss": 0.9328, "step": 1748 }, { "epoch": 2.57, "learning_rate": 5.487435619739214e-05, "loss": 0.9765, "step": 1749 }, { "epoch": 2.57, "learning_rate": 5.476914963246647e-05, "loss": 0.8417, "step": 1750 }, { "epoch": 2.57, "learning_rate": 5.4664005969765674e-05, "loss": 0.93, "step": 1751 }, { "epoch": 2.57, "learning_rate": 5.4558925355512256e-05, "loss": 0.8757, "step": 1752 }, { "epoch": 2.57, "learning_rate": 5.445390793584115e-05, "loss": 0.9529, "step": 1753 }, { "epoch": 2.57, "learning_rate": 5.434895385679937e-05, "loss": 0.8961, "step": 1754 }, { "epoch": 2.58, "learning_rate": 5.4244063264345745e-05, "loss": 0.844, "step": 1755 }, { "epoch": 2.58, "learning_rate": 5.4139236304350935e-05, "loss": 0.9071, "step": 1756 }, { "epoch": 2.58, "learning_rate": 5.403447312259702e-05, "loss": 0.8767, "step": 1757 }, { "epoch": 2.58, "learning_rate": 5.392977386477738e-05, "loss": 0.9043, "step": 1758 }, { "epoch": 2.58, "learning_rate": 5.382513867649663e-05, "loss": 1.0033, "step": 1759 }, { "epoch": 2.58, "learning_rate": 5.372056770327013e-05, "loss": 0.9496, "step": 1760 }, { "epoch": 2.58, "learning_rate": 5.361606109052397e-05, "loss": 0.9717, "step": 1761 }, { "epoch": 2.59, "learning_rate": 5.3511618983594845e-05, "loss": 0.9332, "step": 1762 }, { "epoch": 2.59, "learning_rate": 5.340724152772956e-05, "loss": 0.8972, "step": 1763 }, { "epoch": 2.59, "learning_rate": 5.33029288680852e-05, "loss": 1.0025, "step": 1764 }, { "epoch": 2.59, "learning_rate": 5.31986811497286e-05, "loss": 0.9639, "step": 1765 }, { "epoch": 2.59, "learning_rate": 5.309449851763633e-05, "loss": 0.9181, "step": 1766 }, { "epoch": 2.59, "learning_rate": 5.299038111669444e-05, "loss": 0.9608, "step": 1767 }, { "epoch": 2.59, "learning_rate": 5.288632909169823e-05, "loss": 0.9961, "step": 1768 }, { "epoch": 2.6, "learning_rate": 5.2782342587352154e-05, "loss": 0.9709, "step": 1769 }, { "epoch": 2.6, "learning_rate": 5.267842174826955e-05, "loss": 1.0208, "step": 1770 }, { "epoch": 2.6, "learning_rate": 5.2574566718972364e-05, "loss": 0.8965, "step": 1771 }, { "epoch": 2.6, "learning_rate": 5.247077764389099e-05, "loss": 0.9577, "step": 1772 }, { "epoch": 2.6, "learning_rate": 5.236705466736428e-05, "loss": 0.9513, "step": 1773 }, { "epoch": 2.6, "learning_rate": 5.226339793363898e-05, "loss": 0.995, "step": 1774 }, { "epoch": 2.61, "learning_rate": 5.215980758686978e-05, "loss": 0.8956, "step": 1775 }, { "epoch": 2.61, "learning_rate": 5.205628377111902e-05, "loss": 0.9213, "step": 1776 }, { "epoch": 2.61, "learning_rate": 5.195282663035661e-05, "loss": 0.9413, "step": 1777 }, { "epoch": 2.61, "learning_rate": 5.18494363084596e-05, "loss": 0.9791, "step": 1778 }, { "epoch": 2.61, "learning_rate": 5.174611294921224e-05, "loss": 0.9651, "step": 1779 }, { "epoch": 2.61, "learning_rate": 5.1642856696305575e-05, "loss": 0.9441, "step": 1780 }, { "epoch": 2.61, "learning_rate": 5.1539667693337335e-05, "loss": 0.9206, "step": 1781 }, { "epoch": 2.62, "learning_rate": 5.143654608381172e-05, "loss": 0.8989, "step": 1782 }, { "epoch": 2.62, "learning_rate": 5.133349201113929e-05, "loss": 0.9114, "step": 1783 }, { "epoch": 2.62, "learning_rate": 5.123050561863657e-05, "loss": 0.8995, "step": 1784 }, { "epoch": 2.62, "learning_rate": 5.112758704952598e-05, "loss": 0.952, "step": 1785 }, { "epoch": 2.62, "learning_rate": 5.1024736446935754e-05, "loss": 0.8444, "step": 1786 }, { "epoch": 2.62, "learning_rate": 5.092195395389937e-05, "loss": 0.9613, "step": 1787 }, { "epoch": 2.62, "learning_rate": 5.081923971335582e-05, "loss": 0.9385, "step": 1788 }, { "epoch": 2.63, "learning_rate": 5.071659386814907e-05, "loss": 0.8304, "step": 1789 }, { "epoch": 2.63, "learning_rate": 5.061401656102791e-05, "loss": 0.9589, "step": 1790 }, { "epoch": 2.63, "learning_rate": 5.051150793464592e-05, "loss": 0.9628, "step": 1791 }, { "epoch": 2.63, "learning_rate": 5.0409068131561067e-05, "loss": 0.9054, "step": 1792 }, { "epoch": 2.63, "learning_rate": 5.0306697294235714e-05, "loss": 0.8631, "step": 1793 }, { "epoch": 2.63, "learning_rate": 5.020439556503629e-05, "loss": 0.8949, "step": 1794 }, { "epoch": 2.63, "learning_rate": 5.0102163086233065e-05, "loss": 1.0095, "step": 1795 }, { "epoch": 2.64, "learning_rate": 5.000000000000002e-05, "loss": 1.0535, "step": 1796 }, { "epoch": 2.64, "learning_rate": 4.98979064484146e-05, "loss": 0.9593, "step": 1797 }, { "epoch": 2.64, "learning_rate": 4.979588257345766e-05, "loss": 0.9763, "step": 1798 }, { "epoch": 2.64, "learning_rate": 4.969392851701305e-05, "loss": 0.9788, "step": 1799 }, { "epoch": 2.64, "learning_rate": 4.959204442086753e-05, "loss": 0.9516, "step": 1800 }, { "epoch": 2.64, "learning_rate": 4.949023042671066e-05, "loss": 0.9403, "step": 1801 }, { "epoch": 2.65, "learning_rate": 4.938848667613436e-05, "loss": 0.9843, "step": 1802 }, { "epoch": 2.65, "learning_rate": 4.928681331063304e-05, "loss": 1.0064, "step": 1803 }, { "epoch": 2.65, "learning_rate": 4.918521047160308e-05, "loss": 0.9603, "step": 1804 }, { "epoch": 2.65, "learning_rate": 4.908367830034284e-05, "loss": 0.8546, "step": 1805 }, { "epoch": 2.65, "learning_rate": 4.8982216938052394e-05, "loss": 0.9444, "step": 1806 }, { "epoch": 2.65, "learning_rate": 4.888082652583331e-05, "loss": 0.9918, "step": 1807 }, { "epoch": 2.65, "learning_rate": 4.877950720468859e-05, "loss": 0.8489, "step": 1808 }, { "epoch": 2.66, "learning_rate": 4.8678259115522215e-05, "loss": 1.0195, "step": 1809 }, { "epoch": 2.66, "learning_rate": 4.8577082399139296e-05, "loss": 0.9962, "step": 1810 }, { "epoch": 2.66, "learning_rate": 4.8475977196245504e-05, "loss": 0.9836, "step": 1811 }, { "epoch": 2.66, "learning_rate": 4.837494364744711e-05, "loss": 0.8597, "step": 1812 }, { "epoch": 2.66, "learning_rate": 4.827398189325085e-05, "loss": 1.002, "step": 1813 }, { "epoch": 2.66, "learning_rate": 4.817309207406346e-05, "loss": 0.9228, "step": 1814 }, { "epoch": 2.66, "learning_rate": 4.8072274330191725e-05, "loss": 0.9954, "step": 1815 }, { "epoch": 2.67, "learning_rate": 4.7971528801842116e-05, "loss": 0.9124, "step": 1816 }, { "epoch": 2.67, "learning_rate": 4.787085562912076e-05, "loss": 0.925, "step": 1817 }, { "epoch": 2.67, "learning_rate": 4.777025495203319e-05, "loss": 0.9245, "step": 1818 }, { "epoch": 2.67, "learning_rate": 4.7669726910484e-05, "loss": 0.8668, "step": 1819 }, { "epoch": 2.67, "learning_rate": 4.756927164427685e-05, "loss": 1.0186, "step": 1820 }, { "epoch": 2.67, "learning_rate": 4.746888929311415e-05, "loss": 0.95, "step": 1821 }, { "epoch": 2.68, "learning_rate": 4.7368579996596904e-05, "loss": 0.9373, "step": 1822 }, { "epoch": 2.68, "learning_rate": 4.726834389422461e-05, "loss": 0.9235, "step": 1823 }, { "epoch": 2.68, "learning_rate": 4.716818112539485e-05, "loss": 0.8682, "step": 1824 }, { "epoch": 2.68, "learning_rate": 4.706809182940334e-05, "loss": 0.9166, "step": 1825 }, { "epoch": 2.68, "learning_rate": 4.6968076145443515e-05, "loss": 0.9526, "step": 1826 }, { "epoch": 2.68, "learning_rate": 4.686813421260646e-05, "loss": 0.8961, "step": 1827 }, { "epoch": 2.68, "learning_rate": 4.6768266169880804e-05, "loss": 0.9258, "step": 1828 }, { "epoch": 2.69, "learning_rate": 4.666847215615226e-05, "loss": 0.9526, "step": 1829 }, { "epoch": 2.69, "learning_rate": 4.656875231020368e-05, "loss": 0.9606, "step": 1830 }, { "epoch": 2.69, "learning_rate": 4.6469106770714745e-05, "loss": 0.9352, "step": 1831 }, { "epoch": 2.69, "learning_rate": 4.636953567626177e-05, "loss": 0.9786, "step": 1832 }, { "epoch": 2.69, "learning_rate": 4.6270039165317605e-05, "loss": 0.8797, "step": 1833 }, { "epoch": 2.69, "learning_rate": 4.617061737625139e-05, "loss": 0.9609, "step": 1834 }, { "epoch": 2.69, "learning_rate": 4.6071270447328276e-05, "loss": 0.8972, "step": 1835 }, { "epoch": 2.7, "learning_rate": 4.597199851670932e-05, "loss": 0.9233, "step": 1836 }, { "epoch": 2.7, "learning_rate": 4.587280172245129e-05, "loss": 0.9455, "step": 1837 }, { "epoch": 2.7, "learning_rate": 4.57736802025065e-05, "loss": 0.8729, "step": 1838 }, { "epoch": 2.7, "learning_rate": 4.567463409472255e-05, "loss": 0.8894, "step": 1839 }, { "epoch": 2.7, "learning_rate": 4.557566353684209e-05, "loss": 0.9784, "step": 1840 }, { "epoch": 2.7, "learning_rate": 4.547676866650289e-05, "loss": 0.9586, "step": 1841 }, { "epoch": 2.7, "learning_rate": 4.537794962123726e-05, "loss": 0.978, "step": 1842 }, { "epoch": 2.71, "learning_rate": 4.527920653847221e-05, "loss": 0.9431, "step": 1843 }, { "epoch": 2.71, "learning_rate": 4.518053955552903e-05, "loss": 0.8657, "step": 1844 }, { "epoch": 2.71, "learning_rate": 4.50819488096232e-05, "loss": 0.7781, "step": 1845 }, { "epoch": 2.71, "learning_rate": 4.498343443786416e-05, "loss": 0.9086, "step": 1846 }, { "epoch": 2.71, "learning_rate": 4.488499657725511e-05, "loss": 0.897, "step": 1847 }, { "epoch": 2.71, "learning_rate": 4.478663536469296e-05, "loss": 0.9494, "step": 1848 }, { "epoch": 2.71, "eval_loss": 0.9907075762748718, "eval_runtime": 2.6095, "eval_samples_per_second": 418.473, "eval_steps_per_second": 26.442, "step": 1848 }, { "epoch": 2.72, "learning_rate": 4.468835093696796e-05, "loss": 0.9201, "step": 1849 }, { "epoch": 2.72, "learning_rate": 4.4590143430763555e-05, "loss": 0.989, "step": 1850 }, { "epoch": 2.72, "learning_rate": 4.449201298265622e-05, "loss": 0.8615, "step": 1851 }, { "epoch": 2.72, "learning_rate": 4.4393959729115244e-05, "loss": 0.9388, "step": 1852 }, { "epoch": 2.72, "learning_rate": 4.429598380650266e-05, "loss": 0.9179, "step": 1853 }, { "epoch": 2.72, "learning_rate": 4.419808535107287e-05, "loss": 0.9471, "step": 1854 }, { "epoch": 2.72, "learning_rate": 4.4100264498972564e-05, "loss": 0.9584, "step": 1855 }, { "epoch": 2.73, "learning_rate": 4.4002521386240466e-05, "loss": 0.9077, "step": 1856 }, { "epoch": 2.73, "learning_rate": 4.3904856148807284e-05, "loss": 0.876, "step": 1857 }, { "epoch": 2.73, "learning_rate": 4.3807268922495406e-05, "loss": 0.9785, "step": 1858 }, { "epoch": 2.73, "learning_rate": 4.370975984301866e-05, "loss": 0.8762, "step": 1859 }, { "epoch": 2.73, "learning_rate": 4.3612329045982236e-05, "loss": 0.8797, "step": 1860 }, { "epoch": 2.73, "learning_rate": 4.351497666688246e-05, "loss": 1.05, "step": 1861 }, { "epoch": 2.73, "learning_rate": 4.341770284110656e-05, "loss": 0.9844, "step": 1862 }, { "epoch": 2.74, "learning_rate": 4.332050770393263e-05, "loss": 0.9814, "step": 1863 }, { "epoch": 2.74, "learning_rate": 4.322339139052921e-05, "loss": 0.7994, "step": 1864 }, { "epoch": 2.74, "learning_rate": 4.312635403595532e-05, "loss": 0.8213, "step": 1865 }, { "epoch": 2.74, "learning_rate": 4.3029395775160106e-05, "loss": 0.9036, "step": 1866 }, { "epoch": 2.74, "learning_rate": 4.293251674298269e-05, "loss": 0.9382, "step": 1867 }, { "epoch": 2.74, "learning_rate": 4.283571707415214e-05, "loss": 0.9933, "step": 1868 }, { "epoch": 2.75, "learning_rate": 4.273899690328702e-05, "loss": 0.897, "step": 1869 }, { "epoch": 2.75, "learning_rate": 4.264235636489542e-05, "loss": 0.9293, "step": 1870 }, { "epoch": 2.75, "learning_rate": 4.2545795593374594e-05, "loss": 0.9417, "step": 1871 }, { "epoch": 2.75, "learning_rate": 4.244931472301098e-05, "loss": 0.8857, "step": 1872 }, { "epoch": 2.75, "learning_rate": 4.235291388797986e-05, "loss": 0.8865, "step": 1873 }, { "epoch": 2.75, "learning_rate": 4.2256593222345185e-05, "loss": 0.8899, "step": 1874 }, { "epoch": 2.75, "learning_rate": 4.216035286005942e-05, "loss": 0.9032, "step": 1875 }, { "epoch": 2.76, "learning_rate": 4.206419293496333e-05, "loss": 0.9283, "step": 1876 }, { "epoch": 2.76, "learning_rate": 4.196811358078585e-05, "loss": 0.852, "step": 1877 }, { "epoch": 2.76, "learning_rate": 4.18721149311439e-05, "loss": 0.9111, "step": 1878 }, { "epoch": 2.76, "learning_rate": 4.177619711954211e-05, "loss": 0.9561, "step": 1879 }, { "epoch": 2.76, "learning_rate": 4.168036027937267e-05, "loss": 0.8661, "step": 1880 }, { "epoch": 2.76, "learning_rate": 4.1584604543915254e-05, "loss": 0.9675, "step": 1881 }, { "epoch": 2.76, "learning_rate": 4.148893004633663e-05, "loss": 0.9231, "step": 1882 }, { "epoch": 2.77, "learning_rate": 4.139333691969071e-05, "loss": 0.8708, "step": 1883 }, { "epoch": 2.77, "learning_rate": 4.129782529691815e-05, "loss": 0.9466, "step": 1884 }, { "epoch": 2.77, "learning_rate": 4.1202395310846296e-05, "loss": 0.8492, "step": 1885 }, { "epoch": 2.77, "learning_rate": 4.1107047094188946e-05, "loss": 0.8113, "step": 1886 }, { "epoch": 2.77, "learning_rate": 4.101178077954617e-05, "loss": 0.8231, "step": 1887 }, { "epoch": 2.77, "learning_rate": 4.091659649940419e-05, "loss": 1.025, "step": 1888 }, { "epoch": 2.77, "learning_rate": 4.082149438613514e-05, "loss": 0.9734, "step": 1889 }, { "epoch": 2.78, "learning_rate": 4.072647457199684e-05, "loss": 0.9829, "step": 1890 }, { "epoch": 2.78, "learning_rate": 4.063153718913267e-05, "loss": 0.9899, "step": 1891 }, { "epoch": 2.78, "learning_rate": 4.053668236957134e-05, "loss": 0.9136, "step": 1892 }, { "epoch": 2.78, "learning_rate": 4.044191024522686e-05, "loss": 0.955, "step": 1893 }, { "epoch": 2.78, "learning_rate": 4.034722094789809e-05, "loss": 0.9592, "step": 1894 }, { "epoch": 2.78, "learning_rate": 4.0252614609268766e-05, "loss": 0.9513, "step": 1895 }, { "epoch": 2.79, "learning_rate": 4.015809136090732e-05, "loss": 0.8921, "step": 1896 }, { "epoch": 2.79, "learning_rate": 4.0063651334266496e-05, "loss": 0.9401, "step": 1897 }, { "epoch": 2.79, "learning_rate": 3.996929466068344e-05, "loss": 0.9658, "step": 1898 }, { "epoch": 2.79, "learning_rate": 3.987502147137928e-05, "loss": 1.0133, "step": 1899 }, { "epoch": 2.79, "learning_rate": 3.978083189745907e-05, "loss": 0.8799, "step": 1900 }, { "epoch": 2.79, "learning_rate": 3.96867260699116e-05, "loss": 0.9336, "step": 1901 }, { "epoch": 2.79, "learning_rate": 3.9592704119609125e-05, "loss": 0.9089, "step": 1902 }, { "epoch": 2.8, "learning_rate": 3.9498766177307403e-05, "loss": 0.9544, "step": 1903 }, { "epoch": 2.8, "learning_rate": 3.9404912373645185e-05, "loss": 0.877, "step": 1904 }, { "epoch": 2.8, "learning_rate": 3.9311142839144365e-05, "loss": 0.9766, "step": 1905 }, { "epoch": 2.8, "learning_rate": 3.9217457704209536e-05, "loss": 0.9153, "step": 1906 }, { "epoch": 2.8, "learning_rate": 3.9123857099127936e-05, "loss": 0.8915, "step": 1907 }, { "epoch": 2.8, "learning_rate": 3.903034115406931e-05, "loss": 0.8892, "step": 1908 }, { "epoch": 2.8, "learning_rate": 3.893690999908562e-05, "loss": 0.9365, "step": 1909 }, { "epoch": 2.81, "learning_rate": 3.884356376411089e-05, "loss": 0.9644, "step": 1910 }, { "epoch": 2.81, "learning_rate": 3.875030257896105e-05, "loss": 0.8724, "step": 1911 }, { "epoch": 2.81, "learning_rate": 3.8657126573333804e-05, "loss": 0.9454, "step": 1912 }, { "epoch": 2.81, "learning_rate": 3.85640358768084e-05, "loss": 0.9356, "step": 1913 }, { "epoch": 2.81, "learning_rate": 3.8471030618845375e-05, "loss": 0.9073, "step": 1914 }, { "epoch": 2.81, "learning_rate": 3.837811092878649e-05, "loss": 0.9627, "step": 1915 }, { "epoch": 2.82, "learning_rate": 3.828527693585451e-05, "loss": 0.9601, "step": 1916 }, { "epoch": 2.82, "learning_rate": 3.819252876915297e-05, "loss": 0.9001, "step": 1917 }, { "epoch": 2.82, "learning_rate": 3.809986655766616e-05, "loss": 0.9256, "step": 1918 }, { "epoch": 2.82, "learning_rate": 3.800729043025871e-05, "loss": 0.9052, "step": 1919 }, { "epoch": 2.82, "learning_rate": 3.791480051567564e-05, "loss": 1.0201, "step": 1920 }, { "epoch": 2.82, "learning_rate": 3.7822396942542005e-05, "loss": 0.9417, "step": 1921 }, { "epoch": 2.82, "learning_rate": 3.7730079839362755e-05, "loss": 0.8914, "step": 1922 }, { "epoch": 2.83, "learning_rate": 3.76378493345227e-05, "loss": 0.9132, "step": 1923 }, { "epoch": 2.83, "learning_rate": 3.7545705556286126e-05, "loss": 0.8873, "step": 1924 }, { "epoch": 2.83, "learning_rate": 3.745364863279675e-05, "loss": 0.9416, "step": 1925 }, { "epoch": 2.83, "learning_rate": 3.7361678692077416e-05, "loss": 0.9625, "step": 1926 }, { "epoch": 2.83, "learning_rate": 3.726979586203019e-05, "loss": 0.9371, "step": 1927 }, { "epoch": 2.83, "learning_rate": 3.717800027043576e-05, "loss": 0.9022, "step": 1928 }, { "epoch": 2.83, "learning_rate": 3.708629204495371e-05, "loss": 0.9038, "step": 1929 }, { "epoch": 2.84, "learning_rate": 3.699467131312197e-05, "loss": 0.9053, "step": 1930 }, { "epoch": 2.84, "learning_rate": 3.6903138202356855e-05, "loss": 1.0199, "step": 1931 }, { "epoch": 2.84, "learning_rate": 3.681169283995279e-05, "loss": 0.9524, "step": 1932 }, { "epoch": 2.84, "learning_rate": 3.6720335353082246e-05, "loss": 0.8613, "step": 1933 }, { "epoch": 2.84, "learning_rate": 3.662906586879542e-05, "loss": 0.9004, "step": 1934 }, { "epoch": 2.84, "learning_rate": 3.653788451402009e-05, "loss": 0.9132, "step": 1935 }, { "epoch": 2.85, "learning_rate": 3.6446791415561574e-05, "loss": 0.866, "step": 1936 }, { "epoch": 2.85, "learning_rate": 3.635578670010242e-05, "loss": 0.975, "step": 1937 }, { "epoch": 2.85, "learning_rate": 3.626487049420223e-05, "loss": 0.9129, "step": 1938 }, { "epoch": 2.85, "learning_rate": 3.61740429242975e-05, "loss": 0.9127, "step": 1939 }, { "epoch": 2.85, "learning_rate": 3.6083304116701535e-05, "loss": 0.9235, "step": 1940 }, { "epoch": 2.85, "learning_rate": 3.599265419760408e-05, "loss": 0.9644, "step": 1941 }, { "epoch": 2.85, "learning_rate": 3.5902093293071425e-05, "loss": 0.9893, "step": 1942 }, { "epoch": 2.86, "learning_rate": 3.581162152904592e-05, "loss": 0.967, "step": 1943 }, { "epoch": 2.86, "learning_rate": 3.5721239031346066e-05, "loss": 0.8631, "step": 1944 }, { "epoch": 2.86, "learning_rate": 3.5630945925666134e-05, "loss": 0.9066, "step": 1945 }, { "epoch": 2.86, "learning_rate": 3.554074233757608e-05, "loss": 1.0079, "step": 1946 }, { "epoch": 2.86, "learning_rate": 3.545062839252147e-05, "loss": 0.8998, "step": 1947 }, { "epoch": 2.86, "learning_rate": 3.536060421582309e-05, "loss": 0.9351, "step": 1948 }, { "epoch": 2.86, "learning_rate": 3.5270669932676926e-05, "loss": 1.0081, "step": 1949 }, { "epoch": 2.87, "learning_rate": 3.518082566815396e-05, "loss": 0.9089, "step": 1950 }, { "epoch": 2.87, "learning_rate": 3.509107154719994e-05, "loss": 0.9823, "step": 1951 }, { "epoch": 2.87, "learning_rate": 3.500140769463533e-05, "loss": 0.9625, "step": 1952 }, { "epoch": 2.87, "learning_rate": 3.491183423515503e-05, "loss": 0.8762, "step": 1953 }, { "epoch": 2.87, "learning_rate": 3.48223512933282e-05, "loss": 0.9355, "step": 1954 }, { "epoch": 2.87, "learning_rate": 3.4732958993598154e-05, "loss": 0.9384, "step": 1955 }, { "epoch": 2.87, "learning_rate": 3.464365746028208e-05, "loss": 0.9222, "step": 1956 }, { "epoch": 2.88, "learning_rate": 3.455444681757105e-05, "loss": 0.9472, "step": 1957 }, { "epoch": 2.88, "learning_rate": 3.4465327189529664e-05, "loss": 0.9375, "step": 1958 }, { "epoch": 2.88, "learning_rate": 3.437629870009591e-05, "loss": 0.8795, "step": 1959 }, { "epoch": 2.88, "learning_rate": 3.428736147308115e-05, "loss": 0.8276, "step": 1960 }, { "epoch": 2.88, "learning_rate": 3.41985156321697e-05, "loss": 1.0247, "step": 1961 }, { "epoch": 2.88, "learning_rate": 3.410976130091892e-05, "loss": 0.9808, "step": 1962 }, { "epoch": 2.89, "learning_rate": 3.402109860275877e-05, "loss": 0.9254, "step": 1963 }, { "epoch": 2.89, "learning_rate": 3.393252766099187e-05, "loss": 0.8803, "step": 1964 }, { "epoch": 2.89, "learning_rate": 3.38440485987932e-05, "loss": 0.9022, "step": 1965 }, { "epoch": 2.89, "learning_rate": 3.375566153920992e-05, "loss": 0.8599, "step": 1966 }, { "epoch": 2.89, "learning_rate": 3.366736660516132e-05, "loss": 0.9316, "step": 1967 }, { "epoch": 2.89, "learning_rate": 3.3579163919438595e-05, "loss": 0.993, "step": 1968 }, { "epoch": 2.89, "learning_rate": 3.349105360470456e-05, "loss": 0.984, "step": 1969 }, { "epoch": 2.9, "learning_rate": 3.3403035783493605e-05, "loss": 0.9208, "step": 1970 }, { "epoch": 2.9, "learning_rate": 3.331511057821146e-05, "loss": 0.9139, "step": 1971 }, { "epoch": 2.9, "learning_rate": 3.322727811113516e-05, "loss": 0.9367, "step": 1972 }, { "epoch": 2.9, "learning_rate": 3.313953850441266e-05, "loss": 0.9278, "step": 1973 }, { "epoch": 2.9, "learning_rate": 3.305189188006281e-05, "loss": 0.9714, "step": 1974 }, { "epoch": 2.9, "learning_rate": 3.2964338359975134e-05, "loss": 0.8818, "step": 1975 }, { "epoch": 2.9, "learning_rate": 3.287687806590971e-05, "loss": 1.0362, "step": 1976 }, { "epoch": 2.91, "learning_rate": 3.2789511119496994e-05, "loss": 0.9065, "step": 1977 }, { "epoch": 2.91, "learning_rate": 3.270223764223755e-05, "loss": 0.9571, "step": 1978 }, { "epoch": 2.91, "learning_rate": 3.2615057755502e-05, "loss": 0.9128, "step": 1979 }, { "epoch": 2.91, "learning_rate": 3.252797158053077e-05, "loss": 0.8249, "step": 1980 }, { "epoch": 2.91, "learning_rate": 3.244097923843398e-05, "loss": 1.012, "step": 1981 }, { "epoch": 2.91, "learning_rate": 3.2354080850191324e-05, "loss": 0.854, "step": 1982 }, { "epoch": 2.92, "learning_rate": 3.226727653665171e-05, "loss": 0.8708, "step": 1983 }, { "epoch": 2.92, "learning_rate": 3.218056641853337e-05, "loss": 0.8986, "step": 1984 }, { "epoch": 2.92, "learning_rate": 3.2093950616423394e-05, "loss": 0.9326, "step": 1985 }, { "epoch": 2.92, "learning_rate": 3.200742925077775e-05, "loss": 0.941, "step": 1986 }, { "epoch": 2.92, "learning_rate": 3.192100244192116e-05, "loss": 1.0446, "step": 1987 }, { "epoch": 2.92, "learning_rate": 3.1834670310046734e-05, "loss": 0.971, "step": 1988 }, { "epoch": 2.92, "learning_rate": 3.174843297521596e-05, "loss": 0.9548, "step": 1989 }, { "epoch": 2.93, "learning_rate": 3.166229055735848e-05, "loss": 0.8474, "step": 1990 }, { "epoch": 2.93, "learning_rate": 3.157624317627195e-05, "loss": 0.8806, "step": 1991 }, { "epoch": 2.93, "learning_rate": 3.1490290951621904e-05, "loss": 0.9161, "step": 1992 }, { "epoch": 2.93, "learning_rate": 3.140443400294146e-05, "loss": 0.9025, "step": 1993 }, { "epoch": 2.93, "learning_rate": 3.1318672449631284e-05, "loss": 0.9566, "step": 1994 }, { "epoch": 2.93, "learning_rate": 3.123300641095935e-05, "loss": 0.8958, "step": 1995 }, { "epoch": 2.93, "learning_rate": 3.114743600606078e-05, "loss": 0.9001, "step": 1996 }, { "epoch": 2.94, "learning_rate": 3.106196135393782e-05, "loss": 0.993, "step": 1997 }, { "epoch": 2.94, "learning_rate": 3.09765825734594e-05, "loss": 0.932, "step": 1998 }, { "epoch": 2.94, "learning_rate": 3.089129978336118e-05, "loss": 0.9425, "step": 1999 }, { "epoch": 2.94, "learning_rate": 3.080611310224539e-05, "loss": 0.8835, "step": 2000 }, { "epoch": 2.94, "learning_rate": 3.0721022648580486e-05, "loss": 0.9171, "step": 2001 }, { "epoch": 2.94, "learning_rate": 3.063602854070123e-05, "loss": 0.8977, "step": 2002 }, { "epoch": 2.94, "learning_rate": 3.055113089680829e-05, "loss": 1.0178, "step": 2003 }, { "epoch": 2.95, "learning_rate": 3.0466329834968233e-05, "loss": 0.906, "step": 2004 }, { "epoch": 2.95, "learning_rate": 3.0381625473113284e-05, "loss": 0.9512, "step": 2005 }, { "epoch": 2.95, "learning_rate": 3.029701792904117e-05, "loss": 0.9653, "step": 2006 }, { "epoch": 2.95, "learning_rate": 3.0212507320415052e-05, "loss": 0.9445, "step": 2007 }, { "epoch": 2.95, "learning_rate": 3.0128093764763254e-05, "loss": 0.9406, "step": 2008 }, { "epoch": 2.95, "learning_rate": 3.0043777379479098e-05, "loss": 0.9888, "step": 2009 }, { "epoch": 2.96, "learning_rate": 2.9959558281820766e-05, "loss": 0.8496, "step": 2010 }, { "epoch": 2.96, "learning_rate": 2.9875436588911153e-05, "loss": 0.9583, "step": 2011 }, { "epoch": 2.96, "learning_rate": 2.979141241773775e-05, "loss": 0.9488, "step": 2012 }, { "epoch": 2.96, "learning_rate": 2.9707485885152363e-05, "loss": 0.9339, "step": 2013 } ], "logging_steps": 1, "max_steps": 2684, "num_input_tokens_seen": 0, "num_train_epochs": 4, "save_steps": 671, "total_flos": 6.253562983968932e+17, "train_batch_size": 2, "trial_name": null, "trial_params": null }