|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 14.981905910735826, |
|
"eval_steps": 500, |
|
"global_step": 6210, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0024125452352231603, |
|
"grad_norm": 6.09375, |
|
"learning_rate": 3.22061191626409e-07, |
|
"loss": 3.0105, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.012062726176115802, |
|
"grad_norm": 2.21875, |
|
"learning_rate": 1.610305958132045e-06, |
|
"loss": 3.0058, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.024125452352231604, |
|
"grad_norm": 2.125, |
|
"learning_rate": 3.22061191626409e-06, |
|
"loss": 3.059, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.03618817852834741, |
|
"grad_norm": 2.015625, |
|
"learning_rate": 4.830917874396135e-06, |
|
"loss": 3.011, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.04825090470446321, |
|
"grad_norm": 2.03125, |
|
"learning_rate": 6.44122383252818e-06, |
|
"loss": 3.0197, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.06031363088057901, |
|
"grad_norm": 2.953125, |
|
"learning_rate": 8.051529790660225e-06, |
|
"loss": 3.004, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.07237635705669482, |
|
"grad_norm": 2.625, |
|
"learning_rate": 9.66183574879227e-06, |
|
"loss": 2.9712, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.08443908323281062, |
|
"grad_norm": 3.984375, |
|
"learning_rate": 1.1272141706924317e-05, |
|
"loss": 2.9778, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.09650180940892641, |
|
"grad_norm": 2.34375, |
|
"learning_rate": 1.288244766505636e-05, |
|
"loss": 2.881, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.10856453558504221, |
|
"grad_norm": 2.0625, |
|
"learning_rate": 1.4492753623188407e-05, |
|
"loss": 2.8379, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.12062726176115803, |
|
"grad_norm": 2.96875, |
|
"learning_rate": 1.610305958132045e-05, |
|
"loss": 2.738, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.13268998793727382, |
|
"grad_norm": 1.8359375, |
|
"learning_rate": 1.77133655394525e-05, |
|
"loss": 2.6525, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.14475271411338964, |
|
"grad_norm": 2.609375, |
|
"learning_rate": 1.932367149758454e-05, |
|
"loss": 2.5732, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.15681544028950542, |
|
"grad_norm": 1.4609375, |
|
"learning_rate": 2.0933977455716587e-05, |
|
"loss": 2.4813, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.16887816646562123, |
|
"grad_norm": 1.0546875, |
|
"learning_rate": 2.2544283413848633e-05, |
|
"loss": 2.4334, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.18094089264173704, |
|
"grad_norm": 2.625, |
|
"learning_rate": 2.4154589371980676e-05, |
|
"loss": 2.3332, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.19300361881785283, |
|
"grad_norm": 1.1015625, |
|
"learning_rate": 2.576489533011272e-05, |
|
"loss": 2.2571, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.20506634499396864, |
|
"grad_norm": 5.34375, |
|
"learning_rate": 2.7375201288244768e-05, |
|
"loss": 2.1755, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.21712907117008443, |
|
"grad_norm": 1.0390625, |
|
"learning_rate": 2.8985507246376814e-05, |
|
"loss": 2.089, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.22919179734620024, |
|
"grad_norm": 3.328125, |
|
"learning_rate": 3.059581320450886e-05, |
|
"loss": 2.0223, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.24125452352231605, |
|
"grad_norm": 1.0390625, |
|
"learning_rate": 3.22061191626409e-05, |
|
"loss": 1.9332, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.25331724969843183, |
|
"grad_norm": 1.046875, |
|
"learning_rate": 3.381642512077295e-05, |
|
"loss": 1.8597, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.26537997587454765, |
|
"grad_norm": 1.0859375, |
|
"learning_rate": 3.5426731078905e-05, |
|
"loss": 1.8142, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.27744270205066346, |
|
"grad_norm": 0.625, |
|
"learning_rate": 3.7037037037037037e-05, |
|
"loss": 1.7501, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.28950542822677927, |
|
"grad_norm": 0.59375, |
|
"learning_rate": 3.864734299516908e-05, |
|
"loss": 1.6938, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.30156815440289503, |
|
"grad_norm": 0.478515625, |
|
"learning_rate": 4.025764895330113e-05, |
|
"loss": 1.6289, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.31363088057901084, |
|
"grad_norm": 0.671875, |
|
"learning_rate": 4.1867954911433174e-05, |
|
"loss": 1.5807, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.32569360675512665, |
|
"grad_norm": 0.423828125, |
|
"learning_rate": 4.347826086956522e-05, |
|
"loss": 1.5408, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.33775633293124246, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 4.5088566827697266e-05, |
|
"loss": 1.4961, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.3498190591073583, |
|
"grad_norm": 0.40234375, |
|
"learning_rate": 4.669887278582931e-05, |
|
"loss": 1.4528, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.3618817852834741, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 4.830917874396135e-05, |
|
"loss": 1.4274, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.37394451145958985, |
|
"grad_norm": 0.345703125, |
|
"learning_rate": 4.99194847020934e-05, |
|
"loss": 1.3942, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.38600723763570566, |
|
"grad_norm": 0.455078125, |
|
"learning_rate": 5.152979066022544e-05, |
|
"loss": 1.387, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.39806996381182147, |
|
"grad_norm": 0.404296875, |
|
"learning_rate": 5.3140096618357496e-05, |
|
"loss": 1.3418, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.4101326899879373, |
|
"grad_norm": 0.3046875, |
|
"learning_rate": 5.4750402576489535e-05, |
|
"loss": 1.3263, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.4221954161640531, |
|
"grad_norm": 0.279296875, |
|
"learning_rate": 5.6360708534621574e-05, |
|
"loss": 1.3058, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.43425814234016885, |
|
"grad_norm": 0.53125, |
|
"learning_rate": 5.797101449275363e-05, |
|
"loss": 1.2854, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.44632086851628466, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 5.9581320450885666e-05, |
|
"loss": 1.2769, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.4583835946924005, |
|
"grad_norm": 0.359375, |
|
"learning_rate": 6.119162640901772e-05, |
|
"loss": 1.2572, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.4704463208685163, |
|
"grad_norm": 0.2236328125, |
|
"learning_rate": 6.280193236714976e-05, |
|
"loss": 1.2506, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.4825090470446321, |
|
"grad_norm": 0.2734375, |
|
"learning_rate": 6.44122383252818e-05, |
|
"loss": 1.2466, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.4945717732207479, |
|
"grad_norm": 0.380859375, |
|
"learning_rate": 6.602254428341386e-05, |
|
"loss": 1.2347, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.5066344993968637, |
|
"grad_norm": 0.2119140625, |
|
"learning_rate": 6.76328502415459e-05, |
|
"loss": 1.2245, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.5186972255729795, |
|
"grad_norm": 0.2294921875, |
|
"learning_rate": 6.924315619967794e-05, |
|
"loss": 1.2147, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.5307599517490953, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 7.085346215781e-05, |
|
"loss": 1.2062, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.5428226779252111, |
|
"grad_norm": 0.416015625, |
|
"learning_rate": 7.246376811594203e-05, |
|
"loss": 1.2057, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.5548854041013269, |
|
"grad_norm": 0.357421875, |
|
"learning_rate": 7.407407407407407e-05, |
|
"loss": 1.1828, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.5669481302774427, |
|
"grad_norm": 0.279296875, |
|
"learning_rate": 7.568438003220612e-05, |
|
"loss": 1.1799, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.5790108564535585, |
|
"grad_norm": 0.390625, |
|
"learning_rate": 7.729468599033817e-05, |
|
"loss": 1.1689, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.5910735826296744, |
|
"grad_norm": 0.69921875, |
|
"learning_rate": 7.890499194847021e-05, |
|
"loss": 1.1764, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.6031363088057901, |
|
"grad_norm": 0.46484375, |
|
"learning_rate": 8.051529790660226e-05, |
|
"loss": 1.1552, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.6151990349819059, |
|
"grad_norm": 0.2421875, |
|
"learning_rate": 8.21256038647343e-05, |
|
"loss": 1.1596, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.6272617611580217, |
|
"grad_norm": 0.58984375, |
|
"learning_rate": 8.373590982286635e-05, |
|
"loss": 1.1637, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.6393244873341375, |
|
"grad_norm": 0.578125, |
|
"learning_rate": 8.53462157809984e-05, |
|
"loss": 1.15, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.6513872135102533, |
|
"grad_norm": 0.41015625, |
|
"learning_rate": 8.695652173913044e-05, |
|
"loss": 1.1566, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.6634499396863691, |
|
"grad_norm": 0.80859375, |
|
"learning_rate": 8.856682769726249e-05, |
|
"loss": 1.1402, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.6755126658624849, |
|
"grad_norm": 0.8984375, |
|
"learning_rate": 9.017713365539453e-05, |
|
"loss": 1.1368, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.6875753920386007, |
|
"grad_norm": 0.3125, |
|
"learning_rate": 9.178743961352657e-05, |
|
"loss": 1.1269, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.6996381182147166, |
|
"grad_norm": 0.341796875, |
|
"learning_rate": 9.339774557165862e-05, |
|
"loss": 1.1273, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.7117008443908324, |
|
"grad_norm": 0.26953125, |
|
"learning_rate": 9.500805152979067e-05, |
|
"loss": 1.1166, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.7237635705669482, |
|
"grad_norm": 0.2578125, |
|
"learning_rate": 9.66183574879227e-05, |
|
"loss": 1.1147, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.7358262967430639, |
|
"grad_norm": 0.310546875, |
|
"learning_rate": 9.822866344605476e-05, |
|
"loss": 1.1155, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.7478890229191797, |
|
"grad_norm": 0.41015625, |
|
"learning_rate": 9.98389694041868e-05, |
|
"loss": 1.1081, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.7599517490952955, |
|
"grad_norm": 0.3359375, |
|
"learning_rate": 0.00010144927536231885, |
|
"loss": 1.1083, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.7720144752714113, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 0.00010305958132045089, |
|
"loss": 1.0981, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.7840772014475271, |
|
"grad_norm": 0.478515625, |
|
"learning_rate": 0.00010466988727858293, |
|
"loss": 1.1001, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.7961399276236429, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 0.00010628019323671499, |
|
"loss": 1.1048, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.8082026537997588, |
|
"grad_norm": 0.71875, |
|
"learning_rate": 0.00010789049919484701, |
|
"loss": 1.0983, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.8202653799758746, |
|
"grad_norm": 0.388671875, |
|
"learning_rate": 0.00010950080515297907, |
|
"loss": 1.0982, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.8323281061519904, |
|
"grad_norm": 0.69921875, |
|
"learning_rate": 0.00011111111111111112, |
|
"loss": 1.0982, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.8443908323281062, |
|
"grad_norm": 0.53515625, |
|
"learning_rate": 0.00011272141706924315, |
|
"loss": 1.0968, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.856453558504222, |
|
"grad_norm": 0.609375, |
|
"learning_rate": 0.00011433172302737521, |
|
"loss": 1.084, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.8685162846803377, |
|
"grad_norm": 0.6015625, |
|
"learning_rate": 0.00011594202898550725, |
|
"loss": 1.0815, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.8805790108564535, |
|
"grad_norm": 0.373046875, |
|
"learning_rate": 0.00011755233494363929, |
|
"loss": 1.083, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.8926417370325693, |
|
"grad_norm": 0.68359375, |
|
"learning_rate": 0.00011916264090177133, |
|
"loss": 1.0735, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.9047044632086851, |
|
"grad_norm": 1.09375, |
|
"learning_rate": 0.00012077294685990339, |
|
"loss": 1.0852, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.916767189384801, |
|
"grad_norm": 0.4765625, |
|
"learning_rate": 0.00012238325281803544, |
|
"loss": 1.0852, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.9288299155609168, |
|
"grad_norm": 0.30859375, |
|
"learning_rate": 0.00012399355877616748, |
|
"loss": 1.0689, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.9408926417370326, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 0.00012560386473429953, |
|
"loss": 1.0771, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.9529553679131484, |
|
"grad_norm": 0.58203125, |
|
"learning_rate": 0.00012721417069243158, |
|
"loss": 1.0701, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.9650180940892642, |
|
"grad_norm": 0.333984375, |
|
"learning_rate": 0.0001288244766505636, |
|
"loss": 1.07, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.97708082026538, |
|
"grad_norm": 0.423828125, |
|
"learning_rate": 0.00013043478260869567, |
|
"loss": 1.0665, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.9891435464414958, |
|
"grad_norm": 0.3359375, |
|
"learning_rate": 0.00013204508856682771, |
|
"loss": 1.0614, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.9987937273823885, |
|
"eval_loss": 2.4759693145751953, |
|
"eval_runtime": 0.6381, |
|
"eval_samples_per_second": 15.671, |
|
"eval_steps_per_second": 1.567, |
|
"step": 414 |
|
}, |
|
{ |
|
"epoch": 1.0012062726176116, |
|
"grad_norm": 0.30859375, |
|
"learning_rate": 0.00013365539452495973, |
|
"loss": 1.0719, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 1.0132689987937273, |
|
"grad_norm": 0.32421875, |
|
"learning_rate": 0.0001352657004830918, |
|
"loss": 1.0513, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 1.0253317249698433, |
|
"grad_norm": 0.63671875, |
|
"learning_rate": 0.00013687600644122385, |
|
"loss": 1.0561, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 1.037394451145959, |
|
"grad_norm": 0.41796875, |
|
"learning_rate": 0.00013848631239935587, |
|
"loss": 1.0549, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 1.0494571773220749, |
|
"grad_norm": 0.337890625, |
|
"learning_rate": 0.00014009661835748792, |
|
"loss": 1.052, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 1.0615199034981906, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 0.00014170692431562, |
|
"loss": 1.0525, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 1.0735826296743065, |
|
"grad_norm": 0.7421875, |
|
"learning_rate": 0.000143317230273752, |
|
"loss": 1.0521, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 1.0856453558504222, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 0.00014492753623188405, |
|
"loss": 1.0538, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 1.097708082026538, |
|
"grad_norm": 0.6484375, |
|
"learning_rate": 0.00014653784219001613, |
|
"loss": 1.0593, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 1.1097708082026538, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 0.00014814814814814815, |
|
"loss": 1.0443, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 1.1218335343787695, |
|
"grad_norm": 0.373046875, |
|
"learning_rate": 0.0001497584541062802, |
|
"loss": 1.044, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 1.1338962605548855, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 0.00015136876006441224, |
|
"loss": 1.0496, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 1.1459589867310012, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 0.00015297906602254428, |
|
"loss": 1.0491, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 1.158021712907117, |
|
"grad_norm": 0.625, |
|
"learning_rate": 0.00015458937198067633, |
|
"loss": 1.0455, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 1.1700844390832328, |
|
"grad_norm": 0.546875, |
|
"learning_rate": 0.00015619967793880838, |
|
"loss": 1.0366, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 1.1821471652593487, |
|
"grad_norm": 0.59765625, |
|
"learning_rate": 0.00015780998389694042, |
|
"loss": 1.0428, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 1.1942098914354644, |
|
"grad_norm": 1.4140625, |
|
"learning_rate": 0.00015942028985507247, |
|
"loss": 1.0449, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 1.2062726176115803, |
|
"grad_norm": 0.6953125, |
|
"learning_rate": 0.00016103059581320451, |
|
"loss": 1.0363, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 1.218335343787696, |
|
"grad_norm": 0.7890625, |
|
"learning_rate": 0.00016264090177133656, |
|
"loss": 1.0377, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 1.2303980699638117, |
|
"grad_norm": 0.62890625, |
|
"learning_rate": 0.0001642512077294686, |
|
"loss": 1.0404, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 1.2424607961399277, |
|
"grad_norm": 0.69921875, |
|
"learning_rate": 0.00016586151368760065, |
|
"loss": 1.0448, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 1.2545235223160434, |
|
"grad_norm": 1.28125, |
|
"learning_rate": 0.0001674718196457327, |
|
"loss": 1.0441, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 1.2665862484921593, |
|
"grad_norm": 0.68359375, |
|
"learning_rate": 0.00016908212560386474, |
|
"loss": 1.0349, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 1.278648974668275, |
|
"grad_norm": 1.203125, |
|
"learning_rate": 0.0001706924315619968, |
|
"loss": 1.0504, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 1.290711700844391, |
|
"grad_norm": 0.8046875, |
|
"learning_rate": 0.00017230273752012884, |
|
"loss": 1.0362, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 1.3027744270205066, |
|
"grad_norm": 0.80859375, |
|
"learning_rate": 0.00017391304347826088, |
|
"loss": 1.0454, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 1.3148371531966223, |
|
"grad_norm": 1.0078125, |
|
"learning_rate": 0.00017552334943639293, |
|
"loss": 1.0466, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 1.3268998793727382, |
|
"grad_norm": 0.6171875, |
|
"learning_rate": 0.00017713365539452497, |
|
"loss": 1.0319, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 1.3389626055488542, |
|
"grad_norm": 0.3125, |
|
"learning_rate": 0.00017874396135265702, |
|
"loss": 1.038, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 1.3510253317249699, |
|
"grad_norm": 0.349609375, |
|
"learning_rate": 0.00018035426731078907, |
|
"loss": 1.0181, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 1.3630880579010856, |
|
"grad_norm": 0.34375, |
|
"learning_rate": 0.0001819645732689211, |
|
"loss": 1.0236, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 1.3751507840772015, |
|
"grad_norm": 0.36328125, |
|
"learning_rate": 0.00018357487922705313, |
|
"loss": 1.0274, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 1.3872135102533172, |
|
"grad_norm": 0.40234375, |
|
"learning_rate": 0.0001851851851851852, |
|
"loss": 1.0281, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 1.399276236429433, |
|
"grad_norm": 0.4765625, |
|
"learning_rate": 0.00018679549114331725, |
|
"loss": 1.0221, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 1.4113389626055488, |
|
"grad_norm": 0.474609375, |
|
"learning_rate": 0.00018840579710144927, |
|
"loss": 1.0354, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 1.4234016887816647, |
|
"grad_norm": 0.3671875, |
|
"learning_rate": 0.00019001610305958134, |
|
"loss": 1.0257, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 1.4354644149577804, |
|
"grad_norm": 0.3515625, |
|
"learning_rate": 0.0001916264090177134, |
|
"loss": 1.0246, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 1.4475271411338961, |
|
"grad_norm": 0.439453125, |
|
"learning_rate": 0.0001932367149758454, |
|
"loss": 1.0234, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 1.459589867310012, |
|
"grad_norm": 0.39453125, |
|
"learning_rate": 0.00019484702093397745, |
|
"loss": 1.0161, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 1.471652593486128, |
|
"grad_norm": 0.57421875, |
|
"learning_rate": 0.00019645732689210952, |
|
"loss": 1.017, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 1.4837153196622437, |
|
"grad_norm": 0.72265625, |
|
"learning_rate": 0.00019806763285024154, |
|
"loss": 1.0218, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 1.4957780458383594, |
|
"grad_norm": 0.609375, |
|
"learning_rate": 0.0001996779388083736, |
|
"loss": 1.0194, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 1.5078407720144753, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 0.00019999974723244348, |
|
"loss": 1.0206, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 1.5199034981905912, |
|
"grad_norm": 0.3828125, |
|
"learning_rate": 0.00019999872036643513, |
|
"loss": 1.025, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 1.531966224366707, |
|
"grad_norm": 0.66015625, |
|
"learning_rate": 0.0001999969036121076, |
|
"loss": 1.0246, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 1.5440289505428226, |
|
"grad_norm": 0.93359375, |
|
"learning_rate": 0.00019999429698381143, |
|
"loss": 1.0232, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 1.5560916767189386, |
|
"grad_norm": 0.70703125, |
|
"learning_rate": 0.00019999090050213636, |
|
"loss": 1.0167, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 1.5681544028950543, |
|
"grad_norm": 0.62109375, |
|
"learning_rate": 0.00019998671419391108, |
|
"loss": 1.0087, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 1.58021712907117, |
|
"grad_norm": 0.357421875, |
|
"learning_rate": 0.00019998173809220325, |
|
"loss": 1.0215, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 1.5922798552472859, |
|
"grad_norm": 0.34765625, |
|
"learning_rate": 0.00019997597223631895, |
|
"loss": 1.0208, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 1.6043425814234018, |
|
"grad_norm": 0.390625, |
|
"learning_rate": 0.0001999694166718026, |
|
"loss": 1.01, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 1.6164053075995175, |
|
"grad_norm": 0.37109375, |
|
"learning_rate": 0.00019996207145043653, |
|
"loss": 1.0138, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 1.6284680337756332, |
|
"grad_norm": 0.375, |
|
"learning_rate": 0.00019995393663024054, |
|
"loss": 1.0069, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 1.6405307599517491, |
|
"grad_norm": 0.76171875, |
|
"learning_rate": 0.00019994501227547137, |
|
"loss": 1.0115, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 1.652593486127865, |
|
"grad_norm": 0.34375, |
|
"learning_rate": 0.00019993529845662244, |
|
"loss": 1.0071, |
|
"step": 685 |
|
}, |
|
{ |
|
"epoch": 1.6646562123039808, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 0.00019992479525042303, |
|
"loss": 1.0058, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 1.6767189384800965, |
|
"grad_norm": 0.6171875, |
|
"learning_rate": 0.00019991350273983776, |
|
"loss": 1.006, |
|
"step": 695 |
|
}, |
|
{ |
|
"epoch": 1.6887816646562124, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 0.00019990142101406603, |
|
"loss": 1.0051, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 1.700844390832328, |
|
"grad_norm": 0.466796875, |
|
"learning_rate": 0.0001998885501685412, |
|
"loss": 1.0037, |
|
"step": 705 |
|
}, |
|
{ |
|
"epoch": 1.7129071170084438, |
|
"grad_norm": 0.5546875, |
|
"learning_rate": 0.0001998748903049299, |
|
"loss": 1.0072, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 1.7249698431845597, |
|
"grad_norm": 0.326171875, |
|
"learning_rate": 0.0001998604415311311, |
|
"loss": 1.0173, |
|
"step": 715 |
|
}, |
|
{ |
|
"epoch": 1.7370325693606756, |
|
"grad_norm": 0.7109375, |
|
"learning_rate": 0.00019984520396127553, |
|
"loss": 1.0094, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 1.7490952955367913, |
|
"grad_norm": 0.54296875, |
|
"learning_rate": 0.00019982917771572443, |
|
"loss": 1.0048, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 1.761158021712907, |
|
"grad_norm": 0.34375, |
|
"learning_rate": 0.00019981236292106896, |
|
"loss": 1.0063, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 1.773220747889023, |
|
"grad_norm": 0.3515625, |
|
"learning_rate": 0.0001997947597101288, |
|
"loss": 1.0042, |
|
"step": 735 |
|
}, |
|
{ |
|
"epoch": 1.7852834740651389, |
|
"grad_norm": 0.296875, |
|
"learning_rate": 0.00019977636822195153, |
|
"loss": 0.9947, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 1.7973462002412546, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 0.0001997571886018112, |
|
"loss": 1.0042, |
|
"step": 745 |
|
}, |
|
{ |
|
"epoch": 1.8094089264173703, |
|
"grad_norm": 0.578125, |
|
"learning_rate": 0.0001997372210012073, |
|
"loss": 1.0002, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 1.8214716525934862, |
|
"grad_norm": 0.5625, |
|
"learning_rate": 0.00019971646557786363, |
|
"loss": 1.0061, |
|
"step": 755 |
|
}, |
|
{ |
|
"epoch": 1.833534378769602, |
|
"grad_norm": 0.408203125, |
|
"learning_rate": 0.00019969492249572693, |
|
"loss": 1.0024, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 1.8455971049457176, |
|
"grad_norm": 0.40625, |
|
"learning_rate": 0.0001996725919249657, |
|
"loss": 1.001, |
|
"step": 765 |
|
}, |
|
{ |
|
"epoch": 1.8576598311218335, |
|
"grad_norm": 0.396484375, |
|
"learning_rate": 0.0001996494740419687, |
|
"loss": 1.0007, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 1.8697225572979495, |
|
"grad_norm": 0.34375, |
|
"learning_rate": 0.00019962556902934379, |
|
"loss": 0.998, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 1.8817852834740652, |
|
"grad_norm": 0.64453125, |
|
"learning_rate": 0.00019960087707591626, |
|
"loss": 0.9913, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 1.8938480096501809, |
|
"grad_norm": 1.140625, |
|
"learning_rate": 0.00019957539837672745, |
|
"loss": 1.0082, |
|
"step": 785 |
|
}, |
|
{ |
|
"epoch": 1.9059107358262968, |
|
"grad_norm": 0.62890625, |
|
"learning_rate": 0.00019954913313303323, |
|
"loss": 1.0001, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 1.9179734620024127, |
|
"grad_norm": 0.68359375, |
|
"learning_rate": 0.00019952208155230234, |
|
"loss": 1.0091, |
|
"step": 795 |
|
}, |
|
{ |
|
"epoch": 1.9300361881785284, |
|
"grad_norm": 0.5859375, |
|
"learning_rate": 0.0001994942438482148, |
|
"loss": 0.9919, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 1.942098914354644, |
|
"grad_norm": 0.419921875, |
|
"learning_rate": 0.00019946562024066014, |
|
"loss": 0.982, |
|
"step": 805 |
|
}, |
|
{ |
|
"epoch": 1.95416164053076, |
|
"grad_norm": 0.46484375, |
|
"learning_rate": 0.00019943621095573586, |
|
"loss": 1.0035, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 1.9662243667068757, |
|
"grad_norm": 0.373046875, |
|
"learning_rate": 0.00019940601622574542, |
|
"loss": 0.9854, |
|
"step": 815 |
|
}, |
|
{ |
|
"epoch": 1.9782870928829914, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 0.00019937503628919648, |
|
"loss": 0.9947, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 1.9903498190591074, |
|
"grad_norm": 0.546875, |
|
"learning_rate": 0.00019934327139079915, |
|
"loss": 1.0004, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_loss": 2.448092460632324, |
|
"eval_runtime": 0.534, |
|
"eval_samples_per_second": 18.726, |
|
"eval_steps_per_second": 1.873, |
|
"step": 829 |
|
}, |
|
{ |
|
"epoch": 2.0024125452352233, |
|
"grad_norm": 0.57421875, |
|
"learning_rate": 0.00019931072178146374, |
|
"loss": 0.9899, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 2.0144752714113388, |
|
"grad_norm": 0.423828125, |
|
"learning_rate": 0.0001992773877182992, |
|
"loss": 0.9899, |
|
"step": 835 |
|
}, |
|
{ |
|
"epoch": 2.0265379975874547, |
|
"grad_norm": 0.30859375, |
|
"learning_rate": 0.00019924326946461074, |
|
"loss": 0.9803, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 2.0386007237635706, |
|
"grad_norm": 0.291015625, |
|
"learning_rate": 0.00019920836728989794, |
|
"loss": 0.9873, |
|
"step": 845 |
|
}, |
|
{ |
|
"epoch": 2.0506634499396865, |
|
"grad_norm": 0.3203125, |
|
"learning_rate": 0.00019917268146985245, |
|
"loss": 0.9799, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 2.062726176115802, |
|
"grad_norm": 0.3515625, |
|
"learning_rate": 0.0001991362122863561, |
|
"loss": 0.9773, |
|
"step": 855 |
|
}, |
|
{ |
|
"epoch": 2.074788902291918, |
|
"grad_norm": 0.412109375, |
|
"learning_rate": 0.00019909896002747831, |
|
"loss": 0.9818, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 2.086851628468034, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 0.00019906092498747416, |
|
"loss": 0.9779, |
|
"step": 865 |
|
}, |
|
{ |
|
"epoch": 2.0989143546441498, |
|
"grad_norm": 0.478515625, |
|
"learning_rate": 0.0001990221074667818, |
|
"loss": 0.9807, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 2.1109770808202653, |
|
"grad_norm": 0.408203125, |
|
"learning_rate": 0.00019898250777202018, |
|
"loss": 0.9841, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 2.123039806996381, |
|
"grad_norm": 0.470703125, |
|
"learning_rate": 0.0001989421262159867, |
|
"loss": 0.9812, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 2.135102533172497, |
|
"grad_norm": 0.65625, |
|
"learning_rate": 0.00019890096311765465, |
|
"loss": 0.9709, |
|
"step": 885 |
|
}, |
|
{ |
|
"epoch": 2.147165259348613, |
|
"grad_norm": 0.349609375, |
|
"learning_rate": 0.0001988590188021707, |
|
"loss": 0.981, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 2.1592279855247285, |
|
"grad_norm": 0.47265625, |
|
"learning_rate": 0.0001988162936008523, |
|
"loss": 0.9715, |
|
"step": 895 |
|
}, |
|
{ |
|
"epoch": 2.1712907117008444, |
|
"grad_norm": 0.34765625, |
|
"learning_rate": 0.00019877278785118517, |
|
"loss": 0.9842, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 2.1833534378769603, |
|
"grad_norm": 0.333984375, |
|
"learning_rate": 0.00019872850189682053, |
|
"loss": 0.9776, |
|
"step": 905 |
|
}, |
|
{ |
|
"epoch": 2.195416164053076, |
|
"grad_norm": 0.341796875, |
|
"learning_rate": 0.00019868343608757245, |
|
"loss": 0.9921, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 2.2074788902291917, |
|
"grad_norm": 0.341796875, |
|
"learning_rate": 0.00019863759077941504, |
|
"loss": 0.9835, |
|
"step": 915 |
|
}, |
|
{ |
|
"epoch": 2.2195416164053077, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 0.00019859096633447965, |
|
"loss": 0.9828, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 2.2316043425814236, |
|
"grad_norm": 0.4765625, |
|
"learning_rate": 0.00019854356312105207, |
|
"loss": 0.9767, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 2.243667068757539, |
|
"grad_norm": 0.423828125, |
|
"learning_rate": 0.00019849538151356955, |
|
"loss": 0.982, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 2.255729794933655, |
|
"grad_norm": 0.384765625, |
|
"learning_rate": 0.00019844642189261778, |
|
"loss": 0.9848, |
|
"step": 935 |
|
}, |
|
{ |
|
"epoch": 2.267792521109771, |
|
"grad_norm": 0.32421875, |
|
"learning_rate": 0.00019839668464492811, |
|
"loss": 0.9735, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 2.2798552472858864, |
|
"grad_norm": 0.365234375, |
|
"learning_rate": 0.0001983461701633742, |
|
"loss": 0.977, |
|
"step": 945 |
|
}, |
|
{ |
|
"epoch": 2.2919179734620023, |
|
"grad_norm": 0.46875, |
|
"learning_rate": 0.00019829487884696925, |
|
"loss": 0.9785, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 2.3039806996381182, |
|
"grad_norm": 0.341796875, |
|
"learning_rate": 0.00019824281110086248, |
|
"loss": 0.9816, |
|
"step": 955 |
|
}, |
|
{ |
|
"epoch": 2.316043425814234, |
|
"grad_norm": 0.416015625, |
|
"learning_rate": 0.00019818996733633618, |
|
"loss": 0.988, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 2.3281061519903496, |
|
"grad_norm": 0.341796875, |
|
"learning_rate": 0.0001981363479708025, |
|
"loss": 0.9873, |
|
"step": 965 |
|
}, |
|
{ |
|
"epoch": 2.3401688781664656, |
|
"grad_norm": 0.349609375, |
|
"learning_rate": 0.00019808195342779992, |
|
"loss": 0.9812, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 2.3522316043425815, |
|
"grad_norm": 0.396484375, |
|
"learning_rate": 0.00019802678413699006, |
|
"loss": 0.9688, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 2.3642943305186974, |
|
"grad_norm": 0.333984375, |
|
"learning_rate": 0.0001979708405341543, |
|
"loss": 0.9744, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 2.376357056694813, |
|
"grad_norm": 0.36328125, |
|
"learning_rate": 0.00019791412306119027, |
|
"loss": 0.9682, |
|
"step": 985 |
|
}, |
|
{ |
|
"epoch": 2.388419782870929, |
|
"grad_norm": 0.3515625, |
|
"learning_rate": 0.00019785663216610844, |
|
"loss": 0.9854, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 2.4004825090470447, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 0.00019779836830302843, |
|
"loss": 0.9793, |
|
"step": 995 |
|
}, |
|
{ |
|
"epoch": 2.4125452352231607, |
|
"grad_norm": 0.359375, |
|
"learning_rate": 0.0001977393319321756, |
|
"loss": 0.9852, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 2.424607961399276, |
|
"grad_norm": 0.58203125, |
|
"learning_rate": 0.0001976795235198773, |
|
"loss": 0.9759, |
|
"step": 1005 |
|
}, |
|
{ |
|
"epoch": 2.436670687575392, |
|
"grad_norm": 0.42578125, |
|
"learning_rate": 0.00019761894353855926, |
|
"loss": 0.9689, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 2.448733413751508, |
|
"grad_norm": 0.55859375, |
|
"learning_rate": 0.0001975575924667418, |
|
"loss": 0.9717, |
|
"step": 1015 |
|
}, |
|
{ |
|
"epoch": 2.4607961399276235, |
|
"grad_norm": 0.37890625, |
|
"learning_rate": 0.00019749547078903604, |
|
"loss": 0.9769, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 2.4728588661037394, |
|
"grad_norm": 0.451171875, |
|
"learning_rate": 0.00019743257899614017, |
|
"loss": 0.9699, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 2.4849215922798553, |
|
"grad_norm": 0.330078125, |
|
"learning_rate": 0.00019736891758483544, |
|
"loss": 0.9818, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 2.4969843184559712, |
|
"grad_norm": 0.37109375, |
|
"learning_rate": 0.00019730448705798239, |
|
"loss": 0.9716, |
|
"step": 1035 |
|
}, |
|
{ |
|
"epoch": 2.5090470446320867, |
|
"grad_norm": 0.34375, |
|
"learning_rate": 0.00019723928792451668, |
|
"loss": 0.963, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 2.5211097708082026, |
|
"grad_norm": 0.333984375, |
|
"learning_rate": 0.00019717332069944528, |
|
"loss": 0.9629, |
|
"step": 1045 |
|
}, |
|
{ |
|
"epoch": 2.5331724969843186, |
|
"grad_norm": 0.353515625, |
|
"learning_rate": 0.00019710658590384227, |
|
"loss": 0.9741, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 2.545235223160434, |
|
"grad_norm": 0.59375, |
|
"learning_rate": 0.00019703908406484467, |
|
"loss": 0.9674, |
|
"step": 1055 |
|
}, |
|
{ |
|
"epoch": 2.55729794933655, |
|
"grad_norm": 0.392578125, |
|
"learning_rate": 0.00019697081571564854, |
|
"loss": 0.9714, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 2.569360675512666, |
|
"grad_norm": 0.39453125, |
|
"learning_rate": 0.00019690178139550443, |
|
"loss": 0.9656, |
|
"step": 1065 |
|
}, |
|
{ |
|
"epoch": 2.581423401688782, |
|
"grad_norm": 0.35546875, |
|
"learning_rate": 0.0001968319816497134, |
|
"loss": 0.968, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 2.5934861278648977, |
|
"grad_norm": 0.53515625, |
|
"learning_rate": 0.00019676141702962253, |
|
"loss": 0.9676, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 2.605548854041013, |
|
"grad_norm": 0.359375, |
|
"learning_rate": 0.00019669008809262062, |
|
"loss": 0.9733, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 2.617611580217129, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 0.00019661799540213382, |
|
"loss": 0.967, |
|
"step": 1085 |
|
}, |
|
{ |
|
"epoch": 2.6296743063932446, |
|
"grad_norm": 0.4375, |
|
"learning_rate": 0.0001965451395276212, |
|
"loss": 0.9751, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 2.6417370325693605, |
|
"grad_norm": 0.62109375, |
|
"learning_rate": 0.00019647152104457013, |
|
"loss": 0.9745, |
|
"step": 1095 |
|
}, |
|
{ |
|
"epoch": 2.6537997587454765, |
|
"grad_norm": 0.39453125, |
|
"learning_rate": 0.0001963971405344919, |
|
"loss": 0.9751, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 2.6658624849215924, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 0.0001963219985849169, |
|
"loss": 0.9724, |
|
"step": 1105 |
|
}, |
|
{ |
|
"epoch": 2.6779252110977083, |
|
"grad_norm": 0.376953125, |
|
"learning_rate": 0.00019624609578939027, |
|
"loss": 0.973, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 2.689987937273824, |
|
"grad_norm": 0.37890625, |
|
"learning_rate": 0.000196169432747467, |
|
"loss": 0.9771, |
|
"step": 1115 |
|
}, |
|
{ |
|
"epoch": 2.7020506634499397, |
|
"grad_norm": 0.46484375, |
|
"learning_rate": 0.00019609201006470724, |
|
"loss": 0.9671, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 2.7141133896260556, |
|
"grad_norm": 0.443359375, |
|
"learning_rate": 0.0001960138283526715, |
|
"loss": 0.9698, |
|
"step": 1125 |
|
}, |
|
{ |
|
"epoch": 2.726176115802171, |
|
"grad_norm": 0.3359375, |
|
"learning_rate": 0.000195934888228916, |
|
"loss": 0.9645, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 2.738238841978287, |
|
"grad_norm": 0.435546875, |
|
"learning_rate": 0.0001958551903169874, |
|
"loss": 0.9665, |
|
"step": 1135 |
|
}, |
|
{ |
|
"epoch": 2.750301568154403, |
|
"grad_norm": 0.365234375, |
|
"learning_rate": 0.0001957747352464184, |
|
"loss": 0.9719, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 2.762364294330519, |
|
"grad_norm": 0.4140625, |
|
"learning_rate": 0.0001956935236527222, |
|
"loss": 0.9624, |
|
"step": 1145 |
|
}, |
|
{ |
|
"epoch": 2.7744270205066344, |
|
"grad_norm": 0.357421875, |
|
"learning_rate": 0.00019561155617738797, |
|
"loss": 0.9686, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 2.7864897466827503, |
|
"grad_norm": 0.421875, |
|
"learning_rate": 0.00019552883346787552, |
|
"loss": 0.962, |
|
"step": 1155 |
|
}, |
|
{ |
|
"epoch": 2.798552472858866, |
|
"grad_norm": 0.423828125, |
|
"learning_rate": 0.0001954453561776102, |
|
"loss": 0.9684, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 2.8106151990349817, |
|
"grad_norm": 0.33984375, |
|
"learning_rate": 0.00019536112496597782, |
|
"loss": 0.9759, |
|
"step": 1165 |
|
}, |
|
{ |
|
"epoch": 2.8226779252110976, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 0.0001952761404983194, |
|
"loss": 0.9744, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 2.8347406513872135, |
|
"grad_norm": 0.37890625, |
|
"learning_rate": 0.00019519040344592593, |
|
"loss": 0.9714, |
|
"step": 1175 |
|
}, |
|
{ |
|
"epoch": 2.8468033775633295, |
|
"grad_norm": 0.388671875, |
|
"learning_rate": 0.000195103914486033, |
|
"loss": 0.965, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 2.8588661037394454, |
|
"grad_norm": 0.55078125, |
|
"learning_rate": 0.0001950166743018156, |
|
"loss": 0.9572, |
|
"step": 1185 |
|
}, |
|
{ |
|
"epoch": 2.870928829915561, |
|
"grad_norm": 0.44140625, |
|
"learning_rate": 0.00019492868358238249, |
|
"loss": 0.9563, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 2.882991556091677, |
|
"grad_norm": 0.326171875, |
|
"learning_rate": 0.00019483994302277102, |
|
"loss": 0.9597, |
|
"step": 1195 |
|
}, |
|
{ |
|
"epoch": 2.8950542822677923, |
|
"grad_norm": 0.42578125, |
|
"learning_rate": 0.00019475045332394153, |
|
"loss": 0.965, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 2.907117008443908, |
|
"grad_norm": 0.443359375, |
|
"learning_rate": 0.00019466021519277165, |
|
"loss": 0.9562, |
|
"step": 1205 |
|
}, |
|
{ |
|
"epoch": 2.919179734620024, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 0.000194569229342051, |
|
"loss": 0.9573, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 2.93124246079614, |
|
"grad_norm": 0.474609375, |
|
"learning_rate": 0.00019447749649047542, |
|
"loss": 0.9629, |
|
"step": 1215 |
|
}, |
|
{ |
|
"epoch": 2.943305186972256, |
|
"grad_norm": 0.6015625, |
|
"learning_rate": 0.0001943850173626412, |
|
"loss": 0.978, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 2.9553679131483714, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 0.00019429179268903957, |
|
"loss": 0.9696, |
|
"step": 1225 |
|
}, |
|
{ |
|
"epoch": 2.9674306393244874, |
|
"grad_norm": 0.3515625, |
|
"learning_rate": 0.0001941978232060507, |
|
"loss": 0.9682, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 2.9794933655006033, |
|
"grad_norm": 0.3828125, |
|
"learning_rate": 0.00019410310965593813, |
|
"loss": 0.9498, |
|
"step": 1235 |
|
}, |
|
{ |
|
"epoch": 2.9915560916767188, |
|
"grad_norm": 0.341796875, |
|
"learning_rate": 0.00019400765278684262, |
|
"loss": 0.9586, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 2.9987937273823886, |
|
"eval_loss": 2.4426419734954834, |
|
"eval_runtime": 0.66, |
|
"eval_samples_per_second": 15.151, |
|
"eval_steps_per_second": 1.515, |
|
"step": 1243 |
|
}, |
|
{ |
|
"epoch": 3.0036188178528347, |
|
"grad_norm": 0.41015625, |
|
"learning_rate": 0.00019391145335277655, |
|
"loss": 0.9537, |
|
"step": 1245 |
|
}, |
|
{ |
|
"epoch": 3.0156815440289506, |
|
"grad_norm": 0.423828125, |
|
"learning_rate": 0.00019381451211361773, |
|
"loss": 0.9384, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 3.0277442702050665, |
|
"grad_norm": 0.34765625, |
|
"learning_rate": 0.00019371682983510344, |
|
"loss": 0.9493, |
|
"step": 1255 |
|
}, |
|
{ |
|
"epoch": 3.039806996381182, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 0.00019361840728882447, |
|
"loss": 0.9441, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 3.051869722557298, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 0.000193519245252219, |
|
"loss": 0.9457, |
|
"step": 1265 |
|
}, |
|
{ |
|
"epoch": 3.063932448733414, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 0.00019341934450856633, |
|
"loss": 0.9498, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 3.0759951749095293, |
|
"grad_norm": 0.30078125, |
|
"learning_rate": 0.00019331870584698093, |
|
"loss": 0.9539, |
|
"step": 1275 |
|
}, |
|
{ |
|
"epoch": 3.0880579010856453, |
|
"grad_norm": 0.39453125, |
|
"learning_rate": 0.000193217330062406, |
|
"loss": 0.9564, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 3.100120627261761, |
|
"grad_norm": 0.296875, |
|
"learning_rate": 0.00019311521795560723, |
|
"loss": 0.9492, |
|
"step": 1285 |
|
}, |
|
{ |
|
"epoch": 3.112183353437877, |
|
"grad_norm": 0.359375, |
|
"learning_rate": 0.00019301237033316659, |
|
"loss": 0.9549, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 3.1242460796139926, |
|
"grad_norm": 0.38671875, |
|
"learning_rate": 0.00019290878800747578, |
|
"loss": 0.9511, |
|
"step": 1295 |
|
}, |
|
{ |
|
"epoch": 3.1363088057901085, |
|
"grad_norm": 0.3359375, |
|
"learning_rate": 0.00019280447179673, |
|
"loss": 0.9395, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 3.1483715319662244, |
|
"grad_norm": 0.388671875, |
|
"learning_rate": 0.00019269942252492133, |
|
"loss": 0.951, |
|
"step": 1305 |
|
}, |
|
{ |
|
"epoch": 3.1604342581423404, |
|
"grad_norm": 0.3828125, |
|
"learning_rate": 0.00019259364102183234, |
|
"loss": 0.9471, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 3.172496984318456, |
|
"grad_norm": 0.3515625, |
|
"learning_rate": 0.00019248712812302947, |
|
"loss": 0.9587, |
|
"step": 1315 |
|
}, |
|
{ |
|
"epoch": 3.1845597104945718, |
|
"grad_norm": 0.39453125, |
|
"learning_rate": 0.0001923798846698564, |
|
"loss": 0.9591, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 3.1966224366706877, |
|
"grad_norm": 0.44140625, |
|
"learning_rate": 0.0001922719115094275, |
|
"loss": 0.9437, |
|
"step": 1325 |
|
}, |
|
{ |
|
"epoch": 3.2086851628468036, |
|
"grad_norm": 0.6171875, |
|
"learning_rate": 0.00019216320949462103, |
|
"loss": 0.9494, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 3.220747889022919, |
|
"grad_norm": 0.8125, |
|
"learning_rate": 0.00019205377948407258, |
|
"loss": 0.9532, |
|
"step": 1335 |
|
}, |
|
{ |
|
"epoch": 3.232810615199035, |
|
"grad_norm": 0.361328125, |
|
"learning_rate": 0.00019194362234216795, |
|
"loss": 0.9568, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 3.244873341375151, |
|
"grad_norm": 0.94921875, |
|
"learning_rate": 0.0001918327389390368, |
|
"loss": 0.9531, |
|
"step": 1345 |
|
}, |
|
{ |
|
"epoch": 3.2569360675512664, |
|
"grad_norm": 0.384765625, |
|
"learning_rate": 0.00019172113015054532, |
|
"loss": 0.9561, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 3.2689987937273823, |
|
"grad_norm": 0.5703125, |
|
"learning_rate": 0.0001916087968582896, |
|
"loss": 0.9554, |
|
"step": 1355 |
|
}, |
|
{ |
|
"epoch": 3.2810615199034983, |
|
"grad_norm": 0.419921875, |
|
"learning_rate": 0.0001914957399495885, |
|
"loss": 0.95, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 3.293124246079614, |
|
"grad_norm": 0.64453125, |
|
"learning_rate": 0.00019138196031747681, |
|
"loss": 0.9599, |
|
"step": 1365 |
|
}, |
|
{ |
|
"epoch": 3.3051869722557297, |
|
"grad_norm": 0.5859375, |
|
"learning_rate": 0.00019126745886069807, |
|
"loss": 0.9546, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 3.3172496984318456, |
|
"grad_norm": 0.59375, |
|
"learning_rate": 0.0001911522364836974, |
|
"loss": 0.9521, |
|
"step": 1375 |
|
}, |
|
{ |
|
"epoch": 3.3293124246079615, |
|
"grad_norm": 0.365234375, |
|
"learning_rate": 0.0001910362940966147, |
|
"loss": 0.9571, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 3.341375150784077, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 0.0001909196326152769, |
|
"loss": 0.9514, |
|
"step": 1385 |
|
}, |
|
{ |
|
"epoch": 3.353437876960193, |
|
"grad_norm": 0.66796875, |
|
"learning_rate": 0.00019080225296119125, |
|
"loss": 0.9482, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 3.365500603136309, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 0.00019068415606153787, |
|
"loss": 0.9489, |
|
"step": 1395 |
|
}, |
|
{ |
|
"epoch": 3.3775633293124248, |
|
"grad_norm": 0.462890625, |
|
"learning_rate": 0.00019056534284916224, |
|
"loss": 0.944, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 3.3896260554885402, |
|
"grad_norm": 0.40625, |
|
"learning_rate": 0.00019044581426256804, |
|
"loss": 0.9536, |
|
"step": 1405 |
|
}, |
|
{ |
|
"epoch": 3.401688781664656, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 0.00019032557124590974, |
|
"loss": 0.9492, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 3.413751507840772, |
|
"grad_norm": 0.3359375, |
|
"learning_rate": 0.00019020461474898503, |
|
"loss": 0.945, |
|
"step": 1415 |
|
}, |
|
{ |
|
"epoch": 3.425814234016888, |
|
"grad_norm": 0.330078125, |
|
"learning_rate": 0.00019008294572722738, |
|
"loss": 0.955, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 3.4378769601930035, |
|
"grad_norm": 0.337890625, |
|
"learning_rate": 0.00018996056514169844, |
|
"loss": 0.9466, |
|
"step": 1425 |
|
}, |
|
{ |
|
"epoch": 3.4499396863691194, |
|
"grad_norm": 0.330078125, |
|
"learning_rate": 0.00018983747395908058, |
|
"loss": 0.9436, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 3.4620024125452353, |
|
"grad_norm": 0.34765625, |
|
"learning_rate": 0.0001897136731516691, |
|
"loss": 0.949, |
|
"step": 1435 |
|
}, |
|
{ |
|
"epoch": 3.4740651387213513, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 0.0001895891636973646, |
|
"loss": 0.9629, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 3.4861278648974667, |
|
"grad_norm": 0.359375, |
|
"learning_rate": 0.00018946394657966536, |
|
"loss": 0.9439, |
|
"step": 1445 |
|
}, |
|
{ |
|
"epoch": 3.4981905910735827, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 0.00018933802278765934, |
|
"loss": 0.9455, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 3.5102533172496986, |
|
"grad_norm": 0.365234375, |
|
"learning_rate": 0.00018921139331601667, |
|
"loss": 0.9462, |
|
"step": 1455 |
|
}, |
|
{ |
|
"epoch": 3.522316043425814, |
|
"grad_norm": 0.328125, |
|
"learning_rate": 0.00018908405916498154, |
|
"loss": 0.9499, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 3.53437876960193, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 0.00018895602134036434, |
|
"loss": 0.9433, |
|
"step": 1465 |
|
}, |
|
{ |
|
"epoch": 3.546441495778046, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 0.00018882728085353392, |
|
"loss": 0.9431, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 3.558504221954162, |
|
"grad_norm": 0.45703125, |
|
"learning_rate": 0.0001886978387214093, |
|
"loss": 0.9494, |
|
"step": 1475 |
|
}, |
|
{ |
|
"epoch": 3.5705669481302773, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 0.0001885676959664519, |
|
"loss": 0.9515, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 3.5826296743063932, |
|
"grad_norm": 0.5546875, |
|
"learning_rate": 0.00018843685361665723, |
|
"loss": 0.9413, |
|
"step": 1485 |
|
}, |
|
{ |
|
"epoch": 3.594692400482509, |
|
"grad_norm": 0.6171875, |
|
"learning_rate": 0.000188305312705547, |
|
"loss": 0.9398, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 3.6067551266586246, |
|
"grad_norm": 0.380859375, |
|
"learning_rate": 0.0001881730742721608, |
|
"loss": 0.9314, |
|
"step": 1495 |
|
}, |
|
{ |
|
"epoch": 3.6188178528347406, |
|
"grad_norm": 0.53515625, |
|
"learning_rate": 0.00018804013936104792, |
|
"loss": 0.9364, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 3.6308805790108565, |
|
"grad_norm": 0.3203125, |
|
"learning_rate": 0.0001879065090222591, |
|
"loss": 0.9502, |
|
"step": 1505 |
|
}, |
|
{ |
|
"epoch": 3.6429433051869724, |
|
"grad_norm": 0.37890625, |
|
"learning_rate": 0.00018777218431133833, |
|
"loss": 0.9477, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 3.6550060313630883, |
|
"grad_norm": 0.34765625, |
|
"learning_rate": 0.00018763716628931437, |
|
"loss": 0.9462, |
|
"step": 1515 |
|
}, |
|
{ |
|
"epoch": 3.667068757539204, |
|
"grad_norm": 0.431640625, |
|
"learning_rate": 0.00018750145602269247, |
|
"loss": 0.9442, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 3.6791314837153197, |
|
"grad_norm": 0.306640625, |
|
"learning_rate": 0.00018736505458344586, |
|
"loss": 0.9378, |
|
"step": 1525 |
|
}, |
|
{ |
|
"epoch": 3.691194209891435, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 0.0001872279630490074, |
|
"loss": 0.9423, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 3.703256936067551, |
|
"grad_norm": 0.298828125, |
|
"learning_rate": 0.00018709018250226094, |
|
"loss": 0.9389, |
|
"step": 1535 |
|
}, |
|
{ |
|
"epoch": 3.715319662243667, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 0.00018695171403153285, |
|
"loss": 0.9356, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 3.727382388419783, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 0.00018681255873058338, |
|
"loss": 0.9447, |
|
"step": 1545 |
|
}, |
|
{ |
|
"epoch": 3.739445114595899, |
|
"grad_norm": 0.3359375, |
|
"learning_rate": 0.00018667271769859812, |
|
"loss": 0.9409, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 3.7515078407720144, |
|
"grad_norm": 0.337890625, |
|
"learning_rate": 0.00018653219204017916, |
|
"loss": 0.9457, |
|
"step": 1555 |
|
}, |
|
{ |
|
"epoch": 3.7635705669481303, |
|
"grad_norm": 0.33984375, |
|
"learning_rate": 0.00018639098286533644, |
|
"loss": 0.954, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 3.7756332931242462, |
|
"grad_norm": 0.375, |
|
"learning_rate": 0.00018624909128947903, |
|
"loss": 0.9467, |
|
"step": 1565 |
|
}, |
|
{ |
|
"epoch": 3.7876960193003617, |
|
"grad_norm": 0.55078125, |
|
"learning_rate": 0.00018610651843340627, |
|
"loss": 0.9436, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 3.7997587454764776, |
|
"grad_norm": 0.392578125, |
|
"learning_rate": 0.00018596326542329888, |
|
"loss": 0.9381, |
|
"step": 1575 |
|
}, |
|
{ |
|
"epoch": 3.8118214716525936, |
|
"grad_norm": 0.41015625, |
|
"learning_rate": 0.0001858193333907101, |
|
"loss": 0.9408, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 3.8238841978287095, |
|
"grad_norm": 0.326171875, |
|
"learning_rate": 0.0001856747234725569, |
|
"loss": 0.9456, |
|
"step": 1585 |
|
}, |
|
{ |
|
"epoch": 3.835946924004825, |
|
"grad_norm": 0.357421875, |
|
"learning_rate": 0.00018552943681111067, |
|
"loss": 0.9405, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 3.848009650180941, |
|
"grad_norm": 0.384765625, |
|
"learning_rate": 0.00018538347455398846, |
|
"loss": 0.9551, |
|
"step": 1595 |
|
}, |
|
{ |
|
"epoch": 3.860072376357057, |
|
"grad_norm": 0.375, |
|
"learning_rate": 0.00018523683785414387, |
|
"loss": 0.9401, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 3.8721351025331723, |
|
"grad_norm": 0.328125, |
|
"learning_rate": 0.0001850895278698579, |
|
"loss": 0.9374, |
|
"step": 1605 |
|
}, |
|
{ |
|
"epoch": 3.884197828709288, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 0.00018494154576472976, |
|
"loss": 0.937, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 3.896260554885404, |
|
"grad_norm": 0.357421875, |
|
"learning_rate": 0.00018479289270766777, |
|
"loss": 0.9398, |
|
"step": 1615 |
|
}, |
|
{ |
|
"epoch": 3.90832328106152, |
|
"grad_norm": 0.298828125, |
|
"learning_rate": 0.00018464356987288013, |
|
"loss": 0.9402, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 3.920386007237636, |
|
"grad_norm": 0.310546875, |
|
"learning_rate": 0.0001844935784398655, |
|
"loss": 0.9443, |
|
"step": 1625 |
|
}, |
|
{ |
|
"epoch": 3.9324487334137515, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 0.00018434291959340394, |
|
"loss": 0.9441, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 3.9445114595898674, |
|
"grad_norm": 0.298828125, |
|
"learning_rate": 0.0001841915945235472, |
|
"loss": 0.939, |
|
"step": 1635 |
|
}, |
|
{ |
|
"epoch": 3.956574185765983, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 0.00018403960442560973, |
|
"loss": 0.94, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 3.9686369119420988, |
|
"grad_norm": 0.3515625, |
|
"learning_rate": 0.00018388695050015885, |
|
"loss": 0.939, |
|
"step": 1645 |
|
}, |
|
{ |
|
"epoch": 3.9806996381182147, |
|
"grad_norm": 0.408203125, |
|
"learning_rate": 0.00018373363395300554, |
|
"loss": 0.9399, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 3.9927623642943306, |
|
"grad_norm": 0.455078125, |
|
"learning_rate": 0.0001835796559951948, |
|
"loss": 0.9412, |
|
"step": 1655 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_loss": 2.449599504470825, |
|
"eval_runtime": 0.535, |
|
"eval_samples_per_second": 18.691, |
|
"eval_steps_per_second": 1.869, |
|
"step": 1658 |
|
}, |
|
{ |
|
"epoch": 4.0048250904704465, |
|
"grad_norm": 0.353515625, |
|
"learning_rate": 0.0001834250178429961, |
|
"loss": 0.9353, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 4.0168878166465625, |
|
"grad_norm": 0.390625, |
|
"learning_rate": 0.0001832697207178938, |
|
"loss": 0.9234, |
|
"step": 1665 |
|
}, |
|
{ |
|
"epoch": 4.0289505428226775, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 0.00018311376584657736, |
|
"loss": 0.9264, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 4.041013268998793, |
|
"grad_norm": 0.4375, |
|
"learning_rate": 0.00018295715446093192, |
|
"loss": 0.9233, |
|
"step": 1675 |
|
}, |
|
{ |
|
"epoch": 4.053075995174909, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 0.00018279988779802833, |
|
"loss": 0.9299, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 4.065138721351025, |
|
"grad_norm": 0.416015625, |
|
"learning_rate": 0.00018264196710011352, |
|
"loss": 0.9359, |
|
"step": 1685 |
|
}, |
|
{ |
|
"epoch": 4.077201447527141, |
|
"grad_norm": 0.37109375, |
|
"learning_rate": 0.00018248339361460057, |
|
"loss": 0.9266, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 4.089264173703257, |
|
"grad_norm": 0.36328125, |
|
"learning_rate": 0.00018232416859405895, |
|
"loss": 0.935, |
|
"step": 1695 |
|
}, |
|
{ |
|
"epoch": 4.101326899879373, |
|
"grad_norm": 0.341796875, |
|
"learning_rate": 0.00018216429329620456, |
|
"loss": 0.9294, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 4.113389626055489, |
|
"grad_norm": 0.359375, |
|
"learning_rate": 0.00018200376898388996, |
|
"loss": 0.9288, |
|
"step": 1705 |
|
}, |
|
{ |
|
"epoch": 4.125452352231604, |
|
"grad_norm": 0.330078125, |
|
"learning_rate": 0.00018184259692509406, |
|
"loss": 0.9264, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 4.13751507840772, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 0.00018168077839291247, |
|
"loss": 0.9421, |
|
"step": 1715 |
|
}, |
|
{ |
|
"epoch": 4.149577804583836, |
|
"grad_norm": 0.67578125, |
|
"learning_rate": 0.00018151831466554726, |
|
"loss": 0.9355, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 4.161640530759952, |
|
"grad_norm": 0.365234375, |
|
"learning_rate": 0.00018135520702629675, |
|
"loss": 0.9349, |
|
"step": 1725 |
|
}, |
|
{ |
|
"epoch": 4.173703256936068, |
|
"grad_norm": 0.458984375, |
|
"learning_rate": 0.00018119145676354575, |
|
"loss": 0.9345, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 4.185765983112184, |
|
"grad_norm": 0.5625, |
|
"learning_rate": 0.00018102706517075491, |
|
"loss": 0.9319, |
|
"step": 1735 |
|
}, |
|
{ |
|
"epoch": 4.1978287092882995, |
|
"grad_norm": 0.404296875, |
|
"learning_rate": 0.00018086203354645089, |
|
"loss": 0.9309, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 4.209891435464415, |
|
"grad_norm": 0.369140625, |
|
"learning_rate": 0.00018069636319421588, |
|
"loss": 0.9378, |
|
"step": 1745 |
|
}, |
|
{ |
|
"epoch": 4.2219541616405305, |
|
"grad_norm": 0.380859375, |
|
"learning_rate": 0.00018053005542267736, |
|
"loss": 0.9285, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 4.234016887816646, |
|
"grad_norm": 0.326171875, |
|
"learning_rate": 0.00018036311154549784, |
|
"loss": 0.9314, |
|
"step": 1755 |
|
}, |
|
{ |
|
"epoch": 4.246079613992762, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 0.00018019553288136435, |
|
"loss": 0.9234, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 4.258142340168878, |
|
"grad_norm": 0.41015625, |
|
"learning_rate": 0.00018002732075397812, |
|
"loss": 0.9286, |
|
"step": 1765 |
|
}, |
|
{ |
|
"epoch": 4.270205066344994, |
|
"grad_norm": 0.390625, |
|
"learning_rate": 0.00017985847649204417, |
|
"loss": 0.9267, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 4.28226779252111, |
|
"grad_norm": 0.333984375, |
|
"learning_rate": 0.00017968900142926057, |
|
"loss": 0.9275, |
|
"step": 1775 |
|
}, |
|
{ |
|
"epoch": 4.294330518697226, |
|
"grad_norm": 0.37890625, |
|
"learning_rate": 0.00017951889690430824, |
|
"loss": 0.9312, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 4.306393244873341, |
|
"grad_norm": 0.3984375, |
|
"learning_rate": 0.00017934816426084008, |
|
"loss": 0.9242, |
|
"step": 1785 |
|
}, |
|
{ |
|
"epoch": 4.318455971049457, |
|
"grad_norm": 0.333984375, |
|
"learning_rate": 0.00017917680484747065, |
|
"loss": 0.9297, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 4.330518697225573, |
|
"grad_norm": 0.3046875, |
|
"learning_rate": 0.00017900482001776517, |
|
"loss": 0.9206, |
|
"step": 1795 |
|
}, |
|
{ |
|
"epoch": 4.342581423401689, |
|
"grad_norm": 0.333984375, |
|
"learning_rate": 0.00017883221113022916, |
|
"loss": 0.927, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 4.354644149577805, |
|
"grad_norm": 0.337890625, |
|
"learning_rate": 0.00017865897954829748, |
|
"loss": 0.9324, |
|
"step": 1805 |
|
}, |
|
{ |
|
"epoch": 4.366706875753921, |
|
"grad_norm": 0.337890625, |
|
"learning_rate": 0.0001784851266403237, |
|
"loss": 0.9422, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 4.378769601930037, |
|
"grad_norm": 0.353515625, |
|
"learning_rate": 0.0001783106537795692, |
|
"loss": 0.9326, |
|
"step": 1815 |
|
}, |
|
{ |
|
"epoch": 4.390832328106152, |
|
"grad_norm": 0.470703125, |
|
"learning_rate": 0.00017813556234419234, |
|
"loss": 0.9366, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 4.402895054282268, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 0.00017795985371723763, |
|
"loss": 0.9328, |
|
"step": 1825 |
|
}, |
|
{ |
|
"epoch": 4.4149577804583835, |
|
"grad_norm": 0.326171875, |
|
"learning_rate": 0.00017778352928662474, |
|
"loss": 0.9288, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 4.427020506634499, |
|
"grad_norm": 0.357421875, |
|
"learning_rate": 0.00017760659044513757, |
|
"loss": 0.9311, |
|
"step": 1835 |
|
}, |
|
{ |
|
"epoch": 4.439083232810615, |
|
"grad_norm": 0.36328125, |
|
"learning_rate": 0.00017742903859041325, |
|
"loss": 0.9263, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 4.451145958986731, |
|
"grad_norm": 0.390625, |
|
"learning_rate": 0.0001772508751249311, |
|
"loss": 0.9248, |
|
"step": 1845 |
|
}, |
|
{ |
|
"epoch": 4.463208685162847, |
|
"grad_norm": 0.3203125, |
|
"learning_rate": 0.0001770721014560015, |
|
"loss": 0.9322, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 4.475271411338962, |
|
"grad_norm": 0.306640625, |
|
"learning_rate": 0.00017689271899575491, |
|
"loss": 0.9284, |
|
"step": 1855 |
|
}, |
|
{ |
|
"epoch": 4.487334137515078, |
|
"grad_norm": 0.302734375, |
|
"learning_rate": 0.00017671272916113052, |
|
"loss": 0.9323, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 4.499396863691194, |
|
"grad_norm": 0.30859375, |
|
"learning_rate": 0.00017653213337386517, |
|
"loss": 0.9302, |
|
"step": 1865 |
|
}, |
|
{ |
|
"epoch": 4.51145958986731, |
|
"grad_norm": 0.33984375, |
|
"learning_rate": 0.0001763509330604822, |
|
"loss": 0.9298, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 4.523522316043426, |
|
"grad_norm": 0.34375, |
|
"learning_rate": 0.00017616912965228001, |
|
"loss": 0.9278, |
|
"step": 1875 |
|
}, |
|
{ |
|
"epoch": 4.535585042219542, |
|
"grad_norm": 0.341796875, |
|
"learning_rate": 0.00017598672458532088, |
|
"loss": 0.9329, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 4.547647768395658, |
|
"grad_norm": 0.400390625, |
|
"learning_rate": 0.00017580371930041953, |
|
"loss": 0.9251, |
|
"step": 1885 |
|
}, |
|
{ |
|
"epoch": 4.559710494571773, |
|
"grad_norm": 0.34375, |
|
"learning_rate": 0.00017562011524313185, |
|
"loss": 0.9278, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 4.571773220747889, |
|
"grad_norm": 0.392578125, |
|
"learning_rate": 0.0001754359138637434, |
|
"loss": 0.9164, |
|
"step": 1895 |
|
}, |
|
{ |
|
"epoch": 4.583835946924005, |
|
"grad_norm": 0.349609375, |
|
"learning_rate": 0.00017525111661725797, |
|
"loss": 0.9184, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 4.595898673100121, |
|
"grad_norm": 0.328125, |
|
"learning_rate": 0.0001750657249633861, |
|
"loss": 0.9345, |
|
"step": 1905 |
|
}, |
|
{ |
|
"epoch": 4.6079613992762365, |
|
"grad_norm": 0.33984375, |
|
"learning_rate": 0.00017487974036653361, |
|
"loss": 0.9285, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 4.620024125452352, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 0.00017469316429578977, |
|
"loss": 0.9219, |
|
"step": 1915 |
|
}, |
|
{ |
|
"epoch": 4.632086851628468, |
|
"grad_norm": 0.365234375, |
|
"learning_rate": 0.00017450599822491615, |
|
"loss": 0.9192, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 4.644149577804583, |
|
"grad_norm": 0.447265625, |
|
"learning_rate": 0.00017431824363233457, |
|
"loss": 0.9301, |
|
"step": 1925 |
|
}, |
|
{ |
|
"epoch": 4.656212303980699, |
|
"grad_norm": 0.427734375, |
|
"learning_rate": 0.00017412990200111556, |
|
"loss": 0.9293, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 4.668275030156815, |
|
"grad_norm": 0.34375, |
|
"learning_rate": 0.00017394097481896676, |
|
"loss": 0.9215, |
|
"step": 1935 |
|
}, |
|
{ |
|
"epoch": 4.680337756332931, |
|
"grad_norm": 0.30078125, |
|
"learning_rate": 0.00017375146357822096, |
|
"loss": 0.927, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 4.692400482509047, |
|
"grad_norm": 0.33984375, |
|
"learning_rate": 0.00017356136977582454, |
|
"loss": 0.9251, |
|
"step": 1945 |
|
}, |
|
{ |
|
"epoch": 4.704463208685163, |
|
"grad_norm": 0.412109375, |
|
"learning_rate": 0.00017337069491332537, |
|
"loss": 0.9291, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 4.716525934861279, |
|
"grad_norm": 0.3984375, |
|
"learning_rate": 0.00017317944049686124, |
|
"loss": 0.9288, |
|
"step": 1955 |
|
}, |
|
{ |
|
"epoch": 4.728588661037395, |
|
"grad_norm": 0.408203125, |
|
"learning_rate": 0.00017298760803714775, |
|
"loss": 0.9165, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 4.74065138721351, |
|
"grad_norm": 0.55078125, |
|
"learning_rate": 0.00017279519904946647, |
|
"loss": 0.9309, |
|
"step": 1965 |
|
}, |
|
{ |
|
"epoch": 4.752714113389626, |
|
"grad_norm": 0.447265625, |
|
"learning_rate": 0.00017260221505365303, |
|
"loss": 0.919, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 4.764776839565742, |
|
"grad_norm": 0.37109375, |
|
"learning_rate": 0.00017240865757408495, |
|
"loss": 0.9239, |
|
"step": 1975 |
|
}, |
|
{ |
|
"epoch": 4.776839565741858, |
|
"grad_norm": 0.30859375, |
|
"learning_rate": 0.0001722145281396697, |
|
"loss": 0.9296, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 4.788902291917974, |
|
"grad_norm": 0.373046875, |
|
"learning_rate": 0.00017201982828383264, |
|
"loss": 0.9306, |
|
"step": 1985 |
|
}, |
|
{ |
|
"epoch": 4.8009650180940895, |
|
"grad_norm": 0.3359375, |
|
"learning_rate": 0.00017182455954450486, |
|
"loss": 0.9283, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 4.813027744270205, |
|
"grad_norm": 0.4140625, |
|
"learning_rate": 0.00017162872346411102, |
|
"loss": 0.9237, |
|
"step": 1995 |
|
}, |
|
{ |
|
"epoch": 4.825090470446321, |
|
"grad_norm": 0.333984375, |
|
"learning_rate": 0.00017143232158955732, |
|
"loss": 0.9293, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 4.837153196622436, |
|
"grad_norm": 0.40234375, |
|
"learning_rate": 0.00017123535547221907, |
|
"loss": 0.9295, |
|
"step": 2005 |
|
}, |
|
{ |
|
"epoch": 4.849215922798552, |
|
"grad_norm": 0.345703125, |
|
"learning_rate": 0.00017103782666792844, |
|
"loss": 0.9154, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 4.861278648974668, |
|
"grad_norm": 0.30859375, |
|
"learning_rate": 0.00017083973673696247, |
|
"loss": 0.9207, |
|
"step": 2015 |
|
}, |
|
{ |
|
"epoch": 4.873341375150784, |
|
"grad_norm": 0.359375, |
|
"learning_rate": 0.00017064108724403033, |
|
"loss": 0.9272, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 4.8854041013269, |
|
"grad_norm": 0.35546875, |
|
"learning_rate": 0.00017044187975826124, |
|
"loss": 0.9224, |
|
"step": 2025 |
|
}, |
|
{ |
|
"epoch": 4.897466827503016, |
|
"grad_norm": 0.408203125, |
|
"learning_rate": 0.00017024211585319203, |
|
"loss": 0.9241, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 4.909529553679132, |
|
"grad_norm": 0.357421875, |
|
"learning_rate": 0.00017004179710675463, |
|
"loss": 0.933, |
|
"step": 2035 |
|
}, |
|
{ |
|
"epoch": 4.921592279855247, |
|
"grad_norm": 0.294921875, |
|
"learning_rate": 0.00016984092510126367, |
|
"loss": 0.916, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 4.933655006031363, |
|
"grad_norm": 0.33203125, |
|
"learning_rate": 0.00016963950142340396, |
|
"loss": 0.9278, |
|
"step": 2045 |
|
}, |
|
{ |
|
"epoch": 4.945717732207479, |
|
"grad_norm": 0.33203125, |
|
"learning_rate": 0.00016943752766421794, |
|
"loss": 0.9284, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 4.957780458383595, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 0.0001692350054190932, |
|
"loss": 0.9358, |
|
"step": 2055 |
|
}, |
|
{ |
|
"epoch": 4.969843184559711, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 0.00016903193628774977, |
|
"loss": 0.925, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 4.981905910735827, |
|
"grad_norm": 0.359375, |
|
"learning_rate": 0.00016882832187422743, |
|
"loss": 0.9235, |
|
"step": 2065 |
|
}, |
|
{ |
|
"epoch": 4.9939686369119425, |
|
"grad_norm": 0.333984375, |
|
"learning_rate": 0.0001686241637868734, |
|
"loss": 0.9325, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 4.998793727382388, |
|
"eval_loss": 2.4599645137786865, |
|
"eval_runtime": 0.6669, |
|
"eval_samples_per_second": 14.995, |
|
"eval_steps_per_second": 1.499, |
|
"step": 2072 |
|
}, |
|
{ |
|
"epoch": 5.0060313630880575, |
|
"grad_norm": 0.33984375, |
|
"learning_rate": 0.000168419463638329, |
|
"loss": 0.9096, |
|
"step": 2075 |
|
}, |
|
{ |
|
"epoch": 5.018094089264173, |
|
"grad_norm": 0.3359375, |
|
"learning_rate": 0.00016821422304551766, |
|
"loss": 0.9202, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 5.030156815440289, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 0.00016800844362963147, |
|
"loss": 0.9134, |
|
"step": 2085 |
|
}, |
|
{ |
|
"epoch": 5.042219541616405, |
|
"grad_norm": 0.32421875, |
|
"learning_rate": 0.00016780212701611886, |
|
"loss": 0.909, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 5.054282267792521, |
|
"grad_norm": 0.3125, |
|
"learning_rate": 0.0001675952748346715, |
|
"loss": 0.9011, |
|
"step": 2095 |
|
}, |
|
{ |
|
"epoch": 5.066344993968637, |
|
"grad_norm": 0.36328125, |
|
"learning_rate": 0.00016738788871921152, |
|
"loss": 0.9104, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 5.078407720144753, |
|
"grad_norm": 0.35546875, |
|
"learning_rate": 0.0001671799703078786, |
|
"loss": 0.9222, |
|
"step": 2105 |
|
}, |
|
{ |
|
"epoch": 5.090470446320868, |
|
"grad_norm": 0.32421875, |
|
"learning_rate": 0.00016697152124301695, |
|
"loss": 0.9157, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 5.102533172496984, |
|
"grad_norm": 0.345703125, |
|
"learning_rate": 0.00016676254317116252, |
|
"loss": 0.922, |
|
"step": 2115 |
|
}, |
|
{ |
|
"epoch": 5.1145958986731, |
|
"grad_norm": 0.361328125, |
|
"learning_rate": 0.00016655303774302976, |
|
"loss": 0.9036, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 5.126658624849216, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 0.0001663430066134988, |
|
"loss": 0.9224, |
|
"step": 2125 |
|
}, |
|
{ |
|
"epoch": 5.138721351025332, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 0.0001661324514416022, |
|
"loss": 0.9104, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 5.150784077201448, |
|
"grad_norm": 0.4296875, |
|
"learning_rate": 0.0001659213738905119, |
|
"loss": 0.9145, |
|
"step": 2135 |
|
}, |
|
{ |
|
"epoch": 5.162846803377564, |
|
"grad_norm": 0.30859375, |
|
"learning_rate": 0.00016570977562752623, |
|
"loss": 0.9142, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 5.1749095295536796, |
|
"grad_norm": 0.36328125, |
|
"learning_rate": 0.00016549765832405653, |
|
"loss": 0.9082, |
|
"step": 2145 |
|
}, |
|
{ |
|
"epoch": 5.186972255729795, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 0.000165285023655614, |
|
"loss": 0.9093, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 5.1990349819059105, |
|
"grad_norm": 0.55859375, |
|
"learning_rate": 0.00016507187330179663, |
|
"loss": 0.9211, |
|
"step": 2155 |
|
}, |
|
{ |
|
"epoch": 5.211097708082026, |
|
"grad_norm": 0.42578125, |
|
"learning_rate": 0.0001648582089462756, |
|
"loss": 0.9133, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 5.223160434258142, |
|
"grad_norm": 0.365234375, |
|
"learning_rate": 0.00016464403227678238, |
|
"loss": 0.9211, |
|
"step": 2165 |
|
}, |
|
{ |
|
"epoch": 5.235223160434258, |
|
"grad_norm": 0.32421875, |
|
"learning_rate": 0.00016442934498509505, |
|
"loss": 0.9197, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 5.247285886610374, |
|
"grad_norm": 0.33203125, |
|
"learning_rate": 0.00016421414876702518, |
|
"loss": 0.909, |
|
"step": 2175 |
|
}, |
|
{ |
|
"epoch": 5.25934861278649, |
|
"grad_norm": 0.390625, |
|
"learning_rate": 0.00016399844532240433, |
|
"loss": 0.9158, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 5.271411338962605, |
|
"grad_norm": 0.328125, |
|
"learning_rate": 0.0001637822363550706, |
|
"loss": 0.914, |
|
"step": 2185 |
|
}, |
|
{ |
|
"epoch": 5.283474065138721, |
|
"grad_norm": 0.439453125, |
|
"learning_rate": 0.00016356552357285522, |
|
"loss": 0.9123, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 5.295536791314837, |
|
"grad_norm": 0.361328125, |
|
"learning_rate": 0.00016334830868756906, |
|
"loss": 0.9221, |
|
"step": 2195 |
|
}, |
|
{ |
|
"epoch": 5.307599517490953, |
|
"grad_norm": 0.41015625, |
|
"learning_rate": 0.00016313059341498901, |
|
"loss": 0.9146, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 5.319662243667069, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 0.0001629123794748447, |
|
"loss": 0.9072, |
|
"step": 2205 |
|
}, |
|
{ |
|
"epoch": 5.331724969843185, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 0.00016269366859080451, |
|
"loss": 0.9155, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 5.343787696019301, |
|
"grad_norm": 0.34765625, |
|
"learning_rate": 0.00016247446249046237, |
|
"loss": 0.9152, |
|
"step": 2215 |
|
}, |
|
{ |
|
"epoch": 5.355850422195417, |
|
"grad_norm": 0.45703125, |
|
"learning_rate": 0.00016225476290532374, |
|
"loss": 0.9229, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 5.367913148371532, |
|
"grad_norm": 0.40234375, |
|
"learning_rate": 0.0001620345715707922, |
|
"loss": 0.9192, |
|
"step": 2225 |
|
}, |
|
{ |
|
"epoch": 5.379975874547648, |
|
"grad_norm": 0.3828125, |
|
"learning_rate": 0.00016181389022615564, |
|
"loss": 0.9131, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 5.3920386007237635, |
|
"grad_norm": 0.337890625, |
|
"learning_rate": 0.00016159272061457255, |
|
"loss": 0.9173, |
|
"step": 2235 |
|
}, |
|
{ |
|
"epoch": 5.404101326899879, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 0.0001613710644830582, |
|
"loss": 0.9159, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 5.416164053075995, |
|
"grad_norm": 0.330078125, |
|
"learning_rate": 0.0001611489235824709, |
|
"loss": 0.9114, |
|
"step": 2245 |
|
}, |
|
{ |
|
"epoch": 5.428226779252111, |
|
"grad_norm": 0.34375, |
|
"learning_rate": 0.0001609262996674981, |
|
"loss": 0.926, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 5.440289505428227, |
|
"grad_norm": 0.404296875, |
|
"learning_rate": 0.00016070319449664264, |
|
"loss": 0.9121, |
|
"step": 2255 |
|
}, |
|
{ |
|
"epoch": 5.452352231604342, |
|
"grad_norm": 0.390625, |
|
"learning_rate": 0.00016047960983220873, |
|
"loss": 0.9099, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 5.464414957780458, |
|
"grad_norm": 0.5703125, |
|
"learning_rate": 0.0001602555474402881, |
|
"loss": 0.9105, |
|
"step": 2265 |
|
}, |
|
{ |
|
"epoch": 5.476477683956574, |
|
"grad_norm": 0.61328125, |
|
"learning_rate": 0.0001600310090907461, |
|
"loss": 0.913, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 5.48854041013269, |
|
"grad_norm": 0.330078125, |
|
"learning_rate": 0.00015980599655720758, |
|
"loss": 0.9153, |
|
"step": 2275 |
|
}, |
|
{ |
|
"epoch": 5.500603136308806, |
|
"grad_norm": 0.34765625, |
|
"learning_rate": 0.00015958051161704307, |
|
"loss": 0.9116, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 5.512665862484922, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 0.00015935455605135446, |
|
"loss": 0.9138, |
|
"step": 2285 |
|
}, |
|
{ |
|
"epoch": 5.524728588661038, |
|
"grad_norm": 0.41015625, |
|
"learning_rate": 0.00015912813164496131, |
|
"loss": 0.9166, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 5.536791314837153, |
|
"grad_norm": 0.380859375, |
|
"learning_rate": 0.00015890124018638638, |
|
"loss": 0.9174, |
|
"step": 2295 |
|
}, |
|
{ |
|
"epoch": 5.548854041013269, |
|
"grad_norm": 0.345703125, |
|
"learning_rate": 0.0001586738834678418, |
|
"loss": 0.9165, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 5.560916767189385, |
|
"grad_norm": 0.291015625, |
|
"learning_rate": 0.00015844606328521463, |
|
"loss": 0.912, |
|
"step": 2305 |
|
}, |
|
{ |
|
"epoch": 5.572979493365501, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 0.00015821778143805296, |
|
"loss": 0.9125, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 5.5850422195416165, |
|
"grad_norm": 0.310546875, |
|
"learning_rate": 0.0001579890397295515, |
|
"loss": 0.9264, |
|
"step": 2315 |
|
}, |
|
{ |
|
"epoch": 5.597104945717732, |
|
"grad_norm": 0.298828125, |
|
"learning_rate": 0.00015775983996653737, |
|
"loss": 0.9129, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 5.609167671893848, |
|
"grad_norm": 0.4375, |
|
"learning_rate": 0.00015753018395945598, |
|
"loss": 0.9139, |
|
"step": 2325 |
|
}, |
|
{ |
|
"epoch": 5.621230398069963, |
|
"grad_norm": 0.341796875, |
|
"learning_rate": 0.00015730007352235644, |
|
"loss": 0.9078, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 5.633293124246079, |
|
"grad_norm": 0.3515625, |
|
"learning_rate": 0.0001570695104728775, |
|
"loss": 0.905, |
|
"step": 2335 |
|
}, |
|
{ |
|
"epoch": 5.645355850422195, |
|
"grad_norm": 0.30859375, |
|
"learning_rate": 0.00015683849663223308, |
|
"loss": 0.906, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 5.657418576598311, |
|
"grad_norm": 0.337890625, |
|
"learning_rate": 0.0001566070338251978, |
|
"loss": 0.9144, |
|
"step": 2345 |
|
}, |
|
{ |
|
"epoch": 5.669481302774427, |
|
"grad_norm": 0.376953125, |
|
"learning_rate": 0.00015637512388009284, |
|
"loss": 0.9171, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 5.681544028950543, |
|
"grad_norm": 0.349609375, |
|
"learning_rate": 0.00015614276862877113, |
|
"loss": 0.9169, |
|
"step": 2355 |
|
}, |
|
{ |
|
"epoch": 5.693606755126659, |
|
"grad_norm": 0.353515625, |
|
"learning_rate": 0.00015590996990660317, |
|
"loss": 0.9131, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 5.705669481302774, |
|
"grad_norm": 0.37109375, |
|
"learning_rate": 0.0001556767295524624, |
|
"loss": 0.9248, |
|
"step": 2365 |
|
}, |
|
{ |
|
"epoch": 5.71773220747889, |
|
"grad_norm": 0.38671875, |
|
"learning_rate": 0.0001554430494087107, |
|
"loss": 0.9108, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 5.729794933655006, |
|
"grad_norm": 0.359375, |
|
"learning_rate": 0.00015520893132118385, |
|
"loss": 0.909, |
|
"step": 2375 |
|
}, |
|
{ |
|
"epoch": 5.741857659831122, |
|
"grad_norm": 0.578125, |
|
"learning_rate": 0.000154974377139177, |
|
"loss": 0.9094, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 5.753920386007238, |
|
"grad_norm": 0.61328125, |
|
"learning_rate": 0.00015473938871542986, |
|
"loss": 0.9155, |
|
"step": 2385 |
|
}, |
|
{ |
|
"epoch": 5.765983112183354, |
|
"grad_norm": 0.5546875, |
|
"learning_rate": 0.00015450396790611234, |
|
"loss": 0.9167, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 5.7780458383594695, |
|
"grad_norm": 0.3984375, |
|
"learning_rate": 0.00015426811657080967, |
|
"loss": 0.9265, |
|
"step": 2395 |
|
}, |
|
{ |
|
"epoch": 5.790108564535585, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 0.00015403183657250788, |
|
"loss": 0.9131, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 5.8021712907117005, |
|
"grad_norm": 0.30859375, |
|
"learning_rate": 0.00015379512977757896, |
|
"loss": 0.9088, |
|
"step": 2405 |
|
}, |
|
{ |
|
"epoch": 5.814234016887816, |
|
"grad_norm": 0.34375, |
|
"learning_rate": 0.00015355799805576612, |
|
"loss": 0.9101, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 5.826296743063932, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 0.00015332044328016914, |
|
"loss": 0.9118, |
|
"step": 2415 |
|
}, |
|
{ |
|
"epoch": 5.838359469240048, |
|
"grad_norm": 0.3515625, |
|
"learning_rate": 0.00015308246732722943, |
|
"loss": 0.9107, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 5.850422195416164, |
|
"grad_norm": 0.376953125, |
|
"learning_rate": 0.0001528440720767153, |
|
"loss": 0.9167, |
|
"step": 2425 |
|
}, |
|
{ |
|
"epoch": 5.86248492159228, |
|
"grad_norm": 0.306640625, |
|
"learning_rate": 0.00015260525941170712, |
|
"loss": 0.9059, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 5.874547647768396, |
|
"grad_norm": 0.3203125, |
|
"learning_rate": 0.00015236603121858232, |
|
"loss": 0.918, |
|
"step": 2435 |
|
}, |
|
{ |
|
"epoch": 5.886610373944512, |
|
"grad_norm": 0.35546875, |
|
"learning_rate": 0.00015212638938700068, |
|
"loss": 0.9084, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 5.898673100120627, |
|
"grad_norm": 0.328125, |
|
"learning_rate": 0.00015188633580988926, |
|
"loss": 0.9093, |
|
"step": 2445 |
|
}, |
|
{ |
|
"epoch": 5.910735826296743, |
|
"grad_norm": 0.328125, |
|
"learning_rate": 0.00015164587238342744, |
|
"loss": 0.9016, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 5.922798552472859, |
|
"grad_norm": 0.68359375, |
|
"learning_rate": 0.00015140500100703207, |
|
"loss": 0.9129, |
|
"step": 2455 |
|
}, |
|
{ |
|
"epoch": 5.934861278648975, |
|
"grad_norm": 0.37109375, |
|
"learning_rate": 0.00015116372358334233, |
|
"loss": 0.922, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 5.946924004825091, |
|
"grad_norm": 0.396484375, |
|
"learning_rate": 0.00015092204201820492, |
|
"loss": 0.915, |
|
"step": 2465 |
|
}, |
|
{ |
|
"epoch": 5.958986731001207, |
|
"grad_norm": 0.365234375, |
|
"learning_rate": 0.00015067995822065858, |
|
"loss": 0.9097, |
|
"step": 2470 |
|
}, |
|
{ |
|
"epoch": 5.9710494571773225, |
|
"grad_norm": 0.326171875, |
|
"learning_rate": 0.00015043747410291945, |
|
"loss": 0.9011, |
|
"step": 2475 |
|
}, |
|
{ |
|
"epoch": 5.9831121833534375, |
|
"grad_norm": 0.38671875, |
|
"learning_rate": 0.0001501945915803658, |
|
"loss": 0.919, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 5.9951749095295535, |
|
"grad_norm": 0.447265625, |
|
"learning_rate": 0.0001499513125715228, |
|
"loss": 0.9129, |
|
"step": 2485 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_loss": 2.4628818035125732, |
|
"eval_runtime": 0.5365, |
|
"eval_samples_per_second": 18.639, |
|
"eval_steps_per_second": 1.864, |
|
"step": 2487 |
|
}, |
|
{ |
|
"epoch": 6.007237635705669, |
|
"grad_norm": 0.376953125, |
|
"learning_rate": 0.00014970763899804763, |
|
"loss": 0.9123, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 6.019300361881785, |
|
"grad_norm": 0.37109375, |
|
"learning_rate": 0.00014946357278471388, |
|
"loss": 0.899, |
|
"step": 2495 |
|
}, |
|
{ |
|
"epoch": 6.031363088057901, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 0.0001492191158593968, |
|
"loss": 0.8955, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 6.043425814234017, |
|
"grad_norm": 0.478515625, |
|
"learning_rate": 0.0001489742701530578, |
|
"loss": 0.9005, |
|
"step": 2505 |
|
}, |
|
{ |
|
"epoch": 6.055488540410133, |
|
"grad_norm": 0.41796875, |
|
"learning_rate": 0.0001487290375997292, |
|
"loss": 0.8971, |
|
"step": 2510 |
|
}, |
|
{ |
|
"epoch": 6.067551266586248, |
|
"grad_norm": 0.40625, |
|
"learning_rate": 0.00014848342013649914, |
|
"loss": 0.9056, |
|
"step": 2515 |
|
}, |
|
{ |
|
"epoch": 6.079613992762364, |
|
"grad_norm": 0.326171875, |
|
"learning_rate": 0.00014823741970349606, |
|
"loss": 0.904, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 6.09167671893848, |
|
"grad_norm": 0.30859375, |
|
"learning_rate": 0.00014799103824387358, |
|
"loss": 0.8914, |
|
"step": 2525 |
|
}, |
|
{ |
|
"epoch": 6.103739445114596, |
|
"grad_norm": 0.361328125, |
|
"learning_rate": 0.0001477442777037949, |
|
"loss": 0.9086, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 6.115802171290712, |
|
"grad_norm": 0.369140625, |
|
"learning_rate": 0.0001474971400324177, |
|
"loss": 0.9118, |
|
"step": 2535 |
|
}, |
|
{ |
|
"epoch": 6.127864897466828, |
|
"grad_norm": 0.36328125, |
|
"learning_rate": 0.00014724962718187852, |
|
"loss": 0.9044, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 6.139927623642944, |
|
"grad_norm": 0.380859375, |
|
"learning_rate": 0.00014700174110727747, |
|
"loss": 0.9075, |
|
"step": 2545 |
|
}, |
|
{ |
|
"epoch": 6.151990349819059, |
|
"grad_norm": 0.33203125, |
|
"learning_rate": 0.00014675348376666278, |
|
"loss": 0.8888, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 6.164053075995175, |
|
"grad_norm": 0.34375, |
|
"learning_rate": 0.00014650485712101524, |
|
"loss": 0.8976, |
|
"step": 2555 |
|
}, |
|
{ |
|
"epoch": 6.1761158021712905, |
|
"grad_norm": 0.32421875, |
|
"learning_rate": 0.00014625586313423287, |
|
"loss": 0.9052, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 6.1881785283474064, |
|
"grad_norm": 0.3125, |
|
"learning_rate": 0.00014600650377311522, |
|
"loss": 0.8917, |
|
"step": 2565 |
|
}, |
|
{ |
|
"epoch": 6.200241254523522, |
|
"grad_norm": 0.33203125, |
|
"learning_rate": 0.00014575678100734796, |
|
"loss": 0.9049, |
|
"step": 2570 |
|
}, |
|
{ |
|
"epoch": 6.212303980699638, |
|
"grad_norm": 0.390625, |
|
"learning_rate": 0.0001455066968094873, |
|
"loss": 0.9041, |
|
"step": 2575 |
|
}, |
|
{ |
|
"epoch": 6.224366706875754, |
|
"grad_norm": 0.30859375, |
|
"learning_rate": 0.00014525625315494435, |
|
"loss": 0.9048, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 6.23642943305187, |
|
"grad_norm": 0.375, |
|
"learning_rate": 0.00014500545202196968, |
|
"loss": 0.9093, |
|
"step": 2585 |
|
}, |
|
{ |
|
"epoch": 6.248492159227985, |
|
"grad_norm": 0.4296875, |
|
"learning_rate": 0.00014475429539163742, |
|
"loss": 0.9003, |
|
"step": 2590 |
|
}, |
|
{ |
|
"epoch": 6.260554885404101, |
|
"grad_norm": 0.39453125, |
|
"learning_rate": 0.00014450278524782986, |
|
"loss": 0.9062, |
|
"step": 2595 |
|
}, |
|
{ |
|
"epoch": 6.272617611580217, |
|
"grad_norm": 0.4609375, |
|
"learning_rate": 0.00014425092357722168, |
|
"loss": 0.9032, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 6.284680337756333, |
|
"grad_norm": 0.3828125, |
|
"learning_rate": 0.00014399871236926422, |
|
"loss": 0.9009, |
|
"step": 2605 |
|
}, |
|
{ |
|
"epoch": 6.296743063932449, |
|
"grad_norm": 0.41796875, |
|
"learning_rate": 0.00014374615361616985, |
|
"loss": 0.8952, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 6.308805790108565, |
|
"grad_norm": 0.376953125, |
|
"learning_rate": 0.00014349324931289627, |
|
"loss": 0.8959, |
|
"step": 2615 |
|
}, |
|
{ |
|
"epoch": 6.320868516284681, |
|
"grad_norm": 0.302734375, |
|
"learning_rate": 0.0001432400014571305, |
|
"loss": 0.906, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 6.332931242460796, |
|
"grad_norm": 0.373046875, |
|
"learning_rate": 0.00014298641204927342, |
|
"loss": 0.899, |
|
"step": 2625 |
|
}, |
|
{ |
|
"epoch": 6.344993968636912, |
|
"grad_norm": 0.400390625, |
|
"learning_rate": 0.00014273248309242372, |
|
"loss": 0.895, |
|
"step": 2630 |
|
}, |
|
{ |
|
"epoch": 6.357056694813028, |
|
"grad_norm": 0.384765625, |
|
"learning_rate": 0.0001424782165923623, |
|
"loss": 0.8986, |
|
"step": 2635 |
|
}, |
|
{ |
|
"epoch": 6.3691194209891435, |
|
"grad_norm": 0.41796875, |
|
"learning_rate": 0.0001422236145575362, |
|
"loss": 0.9039, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 6.381182147165259, |
|
"grad_norm": 0.443359375, |
|
"learning_rate": 0.0001419686789990429, |
|
"loss": 0.9086, |
|
"step": 2645 |
|
}, |
|
{ |
|
"epoch": 6.393244873341375, |
|
"grad_norm": 0.369140625, |
|
"learning_rate": 0.0001417134119306144, |
|
"loss": 0.8945, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 6.405307599517491, |
|
"grad_norm": 0.33984375, |
|
"learning_rate": 0.00014145781536860122, |
|
"loss": 0.9083, |
|
"step": 2655 |
|
}, |
|
{ |
|
"epoch": 6.417370325693607, |
|
"grad_norm": 0.35546875, |
|
"learning_rate": 0.00014120189133195657, |
|
"loss": 0.9029, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 6.429433051869722, |
|
"grad_norm": 0.3359375, |
|
"learning_rate": 0.00014094564184222042, |
|
"loss": 0.9124, |
|
"step": 2665 |
|
}, |
|
{ |
|
"epoch": 6.441495778045838, |
|
"grad_norm": 0.361328125, |
|
"learning_rate": 0.00014068906892350343, |
|
"loss": 0.9106, |
|
"step": 2670 |
|
}, |
|
{ |
|
"epoch": 6.453558504221954, |
|
"grad_norm": 0.34375, |
|
"learning_rate": 0.00014043217460247109, |
|
"loss": 0.9021, |
|
"step": 2675 |
|
}, |
|
{ |
|
"epoch": 6.46562123039807, |
|
"grad_norm": 0.345703125, |
|
"learning_rate": 0.0001401749609083276, |
|
"loss": 0.907, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 6.477683956574186, |
|
"grad_norm": 0.3203125, |
|
"learning_rate": 0.0001399174298727998, |
|
"loss": 0.8993, |
|
"step": 2685 |
|
}, |
|
{ |
|
"epoch": 6.489746682750302, |
|
"grad_norm": 0.390625, |
|
"learning_rate": 0.0001396595835301214, |
|
"loss": 0.9022, |
|
"step": 2690 |
|
}, |
|
{ |
|
"epoch": 6.501809408926418, |
|
"grad_norm": 0.384765625, |
|
"learning_rate": 0.00013940142391701652, |
|
"loss": 0.9056, |
|
"step": 2695 |
|
}, |
|
{ |
|
"epoch": 6.513872135102533, |
|
"grad_norm": 0.328125, |
|
"learning_rate": 0.00013914295307268396, |
|
"loss": 0.9099, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 6.525934861278649, |
|
"grad_norm": 0.365234375, |
|
"learning_rate": 0.00013888417303878077, |
|
"loss": 0.9055, |
|
"step": 2705 |
|
}, |
|
{ |
|
"epoch": 6.537997587454765, |
|
"grad_norm": 0.36328125, |
|
"learning_rate": 0.00013862508585940644, |
|
"loss": 0.9037, |
|
"step": 2710 |
|
}, |
|
{ |
|
"epoch": 6.550060313630881, |
|
"grad_norm": 0.353515625, |
|
"learning_rate": 0.00013836569358108647, |
|
"loss": 0.8909, |
|
"step": 2715 |
|
}, |
|
{ |
|
"epoch": 6.5621230398069965, |
|
"grad_norm": 0.3125, |
|
"learning_rate": 0.00013810599825275638, |
|
"loss": 0.9075, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 6.574185765983112, |
|
"grad_norm": 0.33203125, |
|
"learning_rate": 0.00013784600192574557, |
|
"loss": 0.9112, |
|
"step": 2725 |
|
}, |
|
{ |
|
"epoch": 6.586248492159228, |
|
"grad_norm": 0.3828125, |
|
"learning_rate": 0.00013758570665376086, |
|
"loss": 0.9033, |
|
"step": 2730 |
|
}, |
|
{ |
|
"epoch": 6.598311218335343, |
|
"grad_norm": 0.306640625, |
|
"learning_rate": 0.00013732511449287055, |
|
"loss": 0.8993, |
|
"step": 2735 |
|
}, |
|
{ |
|
"epoch": 6.610373944511459, |
|
"grad_norm": 0.361328125, |
|
"learning_rate": 0.00013706422750148797, |
|
"loss": 0.8954, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 6.622436670687575, |
|
"grad_norm": 0.396484375, |
|
"learning_rate": 0.00013680304774035538, |
|
"loss": 0.9054, |
|
"step": 2745 |
|
}, |
|
{ |
|
"epoch": 6.634499396863691, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 0.00013654157727252766, |
|
"loss": 0.9018, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 6.646562123039807, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 0.00013627981816335583, |
|
"loss": 0.9036, |
|
"step": 2755 |
|
}, |
|
{ |
|
"epoch": 6.658624849215923, |
|
"grad_norm": 0.33203125, |
|
"learning_rate": 0.00013601777248047105, |
|
"loss": 0.9067, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 6.670687575392039, |
|
"grad_norm": 0.3125, |
|
"learning_rate": 0.000135755442293768, |
|
"loss": 0.8984, |
|
"step": 2765 |
|
}, |
|
{ |
|
"epoch": 6.682750301568154, |
|
"grad_norm": 0.30859375, |
|
"learning_rate": 0.00013549282967538872, |
|
"loss": 0.907, |
|
"step": 2770 |
|
}, |
|
{ |
|
"epoch": 6.69481302774427, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 0.0001352299366997062, |
|
"loss": 0.9024, |
|
"step": 2775 |
|
}, |
|
{ |
|
"epoch": 6.706875753920386, |
|
"grad_norm": 0.375, |
|
"learning_rate": 0.0001349667654433078, |
|
"loss": 0.9069, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 6.718938480096502, |
|
"grad_norm": 0.365234375, |
|
"learning_rate": 0.00013470331798497926, |
|
"loss": 0.9075, |
|
"step": 2785 |
|
}, |
|
{ |
|
"epoch": 6.731001206272618, |
|
"grad_norm": 0.330078125, |
|
"learning_rate": 0.0001344395964056878, |
|
"loss": 0.9034, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 6.743063932448734, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 0.0001341756027885661, |
|
"loss": 0.9037, |
|
"step": 2795 |
|
}, |
|
{ |
|
"epoch": 6.7551266586248495, |
|
"grad_norm": 0.3203125, |
|
"learning_rate": 0.00013391133921889558, |
|
"loss": 0.9075, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 6.7671893848009645, |
|
"grad_norm": 0.328125, |
|
"learning_rate": 0.00013364680778409, |
|
"loss": 0.9046, |
|
"step": 2805 |
|
}, |
|
{ |
|
"epoch": 6.7792521109770805, |
|
"grad_norm": 0.35546875, |
|
"learning_rate": 0.000133382010573679, |
|
"loss": 0.9002, |
|
"step": 2810 |
|
}, |
|
{ |
|
"epoch": 6.791314837153196, |
|
"grad_norm": 0.310546875, |
|
"learning_rate": 0.00013311694967929158, |
|
"loss": 0.9066, |
|
"step": 2815 |
|
}, |
|
{ |
|
"epoch": 6.803377563329312, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 0.00013285162719463961, |
|
"loss": 0.9057, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 6.815440289505428, |
|
"grad_norm": 0.345703125, |
|
"learning_rate": 0.0001325860452155012, |
|
"loss": 0.8974, |
|
"step": 2825 |
|
}, |
|
{ |
|
"epoch": 6.827503015681544, |
|
"grad_norm": 0.30859375, |
|
"learning_rate": 0.0001323202058397042, |
|
"loss": 0.9006, |
|
"step": 2830 |
|
}, |
|
{ |
|
"epoch": 6.83956574185766, |
|
"grad_norm": 0.32421875, |
|
"learning_rate": 0.00013205411116710972, |
|
"loss": 0.9023, |
|
"step": 2835 |
|
}, |
|
{ |
|
"epoch": 6.851628468033776, |
|
"grad_norm": 0.3203125, |
|
"learning_rate": 0.00013178776329959531, |
|
"loss": 0.903, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 6.863691194209892, |
|
"grad_norm": 0.380859375, |
|
"learning_rate": 0.00013152116434103867, |
|
"loss": 0.9027, |
|
"step": 2845 |
|
}, |
|
{ |
|
"epoch": 6.875753920386007, |
|
"grad_norm": 0.3828125, |
|
"learning_rate": 0.0001312543163973007, |
|
"loss": 0.9087, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 6.887816646562123, |
|
"grad_norm": 0.359375, |
|
"learning_rate": 0.00013098722157620917, |
|
"loss": 0.8917, |
|
"step": 2855 |
|
}, |
|
{ |
|
"epoch": 6.899879372738239, |
|
"grad_norm": 0.37109375, |
|
"learning_rate": 0.00013071988198754185, |
|
"loss": 0.902, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 6.911942098914355, |
|
"grad_norm": 0.357421875, |
|
"learning_rate": 0.00013045229974300993, |
|
"loss": 0.905, |
|
"step": 2865 |
|
}, |
|
{ |
|
"epoch": 6.924004825090471, |
|
"grad_norm": 0.3671875, |
|
"learning_rate": 0.0001301844769562414, |
|
"loss": 0.9072, |
|
"step": 2870 |
|
}, |
|
{ |
|
"epoch": 6.936067551266587, |
|
"grad_norm": 0.35546875, |
|
"learning_rate": 0.00012991641574276418, |
|
"loss": 0.9002, |
|
"step": 2875 |
|
}, |
|
{ |
|
"epoch": 6.9481302774427025, |
|
"grad_norm": 0.330078125, |
|
"learning_rate": 0.0001296481182199896, |
|
"loss": 0.8994, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 6.9601930036188175, |
|
"grad_norm": 0.3046875, |
|
"learning_rate": 0.0001293795865071956, |
|
"loss": 0.9076, |
|
"step": 2885 |
|
}, |
|
{ |
|
"epoch": 6.9722557297949335, |
|
"grad_norm": 0.328125, |
|
"learning_rate": 0.0001291108227255099, |
|
"loss": 0.9044, |
|
"step": 2890 |
|
}, |
|
{ |
|
"epoch": 6.984318455971049, |
|
"grad_norm": 0.333984375, |
|
"learning_rate": 0.00012884182899789343, |
|
"loss": 0.9003, |
|
"step": 2895 |
|
}, |
|
{ |
|
"epoch": 6.996381182147165, |
|
"grad_norm": 0.3203125, |
|
"learning_rate": 0.0001285726074491234, |
|
"loss": 0.8995, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 6.998793727382388, |
|
"eval_loss": 2.4702937602996826, |
|
"eval_runtime": 0.6137, |
|
"eval_samples_per_second": 16.294, |
|
"eval_steps_per_second": 1.629, |
|
"step": 2901 |
|
}, |
|
{ |
|
"epoch": 7.008443908323281, |
|
"grad_norm": 0.349609375, |
|
"learning_rate": 0.00012830316020577656, |
|
"loss": 0.8934, |
|
"step": 2905 |
|
}, |
|
{ |
|
"epoch": 7.020506634499397, |
|
"grad_norm": 0.34765625, |
|
"learning_rate": 0.00012803348939621252, |
|
"loss": 0.8901, |
|
"step": 2910 |
|
}, |
|
{ |
|
"epoch": 7.032569360675513, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 0.00012776359715055668, |
|
"loss": 0.8963, |
|
"step": 2915 |
|
}, |
|
{ |
|
"epoch": 7.044632086851628, |
|
"grad_norm": 0.32421875, |
|
"learning_rate": 0.0001274934856006837, |
|
"loss": 0.8909, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 7.056694813027744, |
|
"grad_norm": 0.353515625, |
|
"learning_rate": 0.00012722315688020047, |
|
"loss": 0.8836, |
|
"step": 2925 |
|
}, |
|
{ |
|
"epoch": 7.06875753920386, |
|
"grad_norm": 0.326171875, |
|
"learning_rate": 0.0001269526131244292, |
|
"loss": 0.8914, |
|
"step": 2930 |
|
}, |
|
{ |
|
"epoch": 7.080820265379976, |
|
"grad_norm": 0.32421875, |
|
"learning_rate": 0.0001266818564703909, |
|
"loss": 0.9006, |
|
"step": 2935 |
|
}, |
|
{ |
|
"epoch": 7.092882991556092, |
|
"grad_norm": 0.328125, |
|
"learning_rate": 0.00012641088905678802, |
|
"loss": 0.9071, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 7.104945717732208, |
|
"grad_norm": 0.32421875, |
|
"learning_rate": 0.00012613971302398794, |
|
"loss": 0.8939, |
|
"step": 2945 |
|
}, |
|
{ |
|
"epoch": 7.117008443908324, |
|
"grad_norm": 0.361328125, |
|
"learning_rate": 0.00012586833051400588, |
|
"loss": 0.8813, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 7.129071170084439, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 0.000125596743670488, |
|
"loss": 0.8852, |
|
"step": 2955 |
|
}, |
|
{ |
|
"epoch": 7.141133896260555, |
|
"grad_norm": 0.34765625, |
|
"learning_rate": 0.00012532495463869452, |
|
"loss": 0.8924, |
|
"step": 2960 |
|
}, |
|
{ |
|
"epoch": 7.1531966224366705, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 0.00012505296556548272, |
|
"loss": 0.891, |
|
"step": 2965 |
|
}, |
|
{ |
|
"epoch": 7.1652593486127865, |
|
"grad_norm": 0.3515625, |
|
"learning_rate": 0.00012478077859929, |
|
"loss": 0.8914, |
|
"step": 2970 |
|
}, |
|
{ |
|
"epoch": 7.177322074788902, |
|
"grad_norm": 0.328125, |
|
"learning_rate": 0.00012450839589011695, |
|
"loss": 0.8947, |
|
"step": 2975 |
|
}, |
|
{ |
|
"epoch": 7.189384800965018, |
|
"grad_norm": 0.3203125, |
|
"learning_rate": 0.00012423581958951026, |
|
"loss": 0.8955, |
|
"step": 2980 |
|
}, |
|
{ |
|
"epoch": 7.201447527141134, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 0.0001239630518505459, |
|
"loss": 0.898, |
|
"step": 2985 |
|
}, |
|
{ |
|
"epoch": 7.213510253317249, |
|
"grad_norm": 0.376953125, |
|
"learning_rate": 0.00012369009482781192, |
|
"loss": 0.8951, |
|
"step": 2990 |
|
}, |
|
{ |
|
"epoch": 7.225572979493365, |
|
"grad_norm": 0.53125, |
|
"learning_rate": 0.00012341695067739154, |
|
"loss": 0.8944, |
|
"step": 2995 |
|
}, |
|
{ |
|
"epoch": 7.237635705669481, |
|
"grad_norm": 0.421875, |
|
"learning_rate": 0.00012314362155684612, |
|
"loss": 0.8936, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 7.249698431845597, |
|
"grad_norm": 0.43359375, |
|
"learning_rate": 0.00012287010962519807, |
|
"loss": 0.8998, |
|
"step": 3005 |
|
}, |
|
{ |
|
"epoch": 7.261761158021713, |
|
"grad_norm": 0.3359375, |
|
"learning_rate": 0.00012259641704291384, |
|
"loss": 0.8959, |
|
"step": 3010 |
|
}, |
|
{ |
|
"epoch": 7.273823884197829, |
|
"grad_norm": 0.421875, |
|
"learning_rate": 0.00012232254597188688, |
|
"loss": 0.8942, |
|
"step": 3015 |
|
}, |
|
{ |
|
"epoch": 7.285886610373945, |
|
"grad_norm": 0.37890625, |
|
"learning_rate": 0.00012204849857542038, |
|
"loss": 0.8886, |
|
"step": 3020 |
|
}, |
|
{ |
|
"epoch": 7.297949336550061, |
|
"grad_norm": 0.333984375, |
|
"learning_rate": 0.00012177427701821051, |
|
"loss": 0.8859, |
|
"step": 3025 |
|
}, |
|
{ |
|
"epoch": 7.310012062726176, |
|
"grad_norm": 0.361328125, |
|
"learning_rate": 0.00012149988346632894, |
|
"loss": 0.8988, |
|
"step": 3030 |
|
}, |
|
{ |
|
"epoch": 7.322074788902292, |
|
"grad_norm": 0.328125, |
|
"learning_rate": 0.00012122532008720611, |
|
"loss": 0.9011, |
|
"step": 3035 |
|
}, |
|
{ |
|
"epoch": 7.334137515078408, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 0.00012095058904961379, |
|
"loss": 0.8903, |
|
"step": 3040 |
|
}, |
|
{ |
|
"epoch": 7.3462002412545235, |
|
"grad_norm": 0.32421875, |
|
"learning_rate": 0.00012067569252364809, |
|
"loss": 0.9006, |
|
"step": 3045 |
|
}, |
|
{ |
|
"epoch": 7.3582629674306395, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 0.00012040063268071243, |
|
"loss": 0.8968, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 7.370325693606755, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 0.00012012541169350011, |
|
"loss": 0.9044, |
|
"step": 3055 |
|
}, |
|
{ |
|
"epoch": 7.382388419782871, |
|
"grad_norm": 0.345703125, |
|
"learning_rate": 0.0001198500317359774, |
|
"loss": 0.8968, |
|
"step": 3060 |
|
}, |
|
{ |
|
"epoch": 7.394451145958986, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 0.00011957449498336626, |
|
"loss": 0.8956, |
|
"step": 3065 |
|
}, |
|
{ |
|
"epoch": 7.406513872135102, |
|
"grad_norm": 0.33984375, |
|
"learning_rate": 0.00011929880361212717, |
|
"loss": 0.8967, |
|
"step": 3070 |
|
}, |
|
{ |
|
"epoch": 7.418576598311218, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 0.00011902295979994192, |
|
"loss": 0.8995, |
|
"step": 3075 |
|
}, |
|
{ |
|
"epoch": 7.430639324487334, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 0.00011874696572569645, |
|
"loss": 0.8923, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 7.44270205066345, |
|
"grad_norm": 0.310546875, |
|
"learning_rate": 0.0001184708235694636, |
|
"loss": 0.8915, |
|
"step": 3085 |
|
}, |
|
{ |
|
"epoch": 7.454764776839566, |
|
"grad_norm": 0.3359375, |
|
"learning_rate": 0.00011819453551248592, |
|
"loss": 0.8951, |
|
"step": 3090 |
|
}, |
|
{ |
|
"epoch": 7.466827503015682, |
|
"grad_norm": 0.365234375, |
|
"learning_rate": 0.00011791810373715846, |
|
"loss": 0.8924, |
|
"step": 3095 |
|
}, |
|
{ |
|
"epoch": 7.478890229191798, |
|
"grad_norm": 0.443359375, |
|
"learning_rate": 0.0001176415304270114, |
|
"loss": 0.895, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 7.490952955367913, |
|
"grad_norm": 0.41796875, |
|
"learning_rate": 0.00011736481776669306, |
|
"loss": 0.8965, |
|
"step": 3105 |
|
}, |
|
{ |
|
"epoch": 7.503015681544029, |
|
"grad_norm": 0.36328125, |
|
"learning_rate": 0.00011708796794195227, |
|
"loss": 0.9014, |
|
"step": 3110 |
|
}, |
|
{ |
|
"epoch": 7.515078407720145, |
|
"grad_norm": 0.373046875, |
|
"learning_rate": 0.00011681098313962158, |
|
"loss": 0.8904, |
|
"step": 3115 |
|
}, |
|
{ |
|
"epoch": 7.527141133896261, |
|
"grad_norm": 0.34765625, |
|
"learning_rate": 0.00011653386554759946, |
|
"loss": 0.8932, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 7.5392038600723765, |
|
"grad_norm": 0.328125, |
|
"learning_rate": 0.00011625661735483348, |
|
"loss": 0.8999, |
|
"step": 3125 |
|
}, |
|
{ |
|
"epoch": 7.5512665862484925, |
|
"grad_norm": 0.37890625, |
|
"learning_rate": 0.00011597924075130273, |
|
"loss": 0.8913, |
|
"step": 3130 |
|
}, |
|
{ |
|
"epoch": 7.563329312424608, |
|
"grad_norm": 0.330078125, |
|
"learning_rate": 0.00011570173792800066, |
|
"loss": 0.8957, |
|
"step": 3135 |
|
}, |
|
{ |
|
"epoch": 7.575392038600723, |
|
"grad_norm": 0.34375, |
|
"learning_rate": 0.00011542411107691767, |
|
"loss": 0.8927, |
|
"step": 3140 |
|
}, |
|
{ |
|
"epoch": 7.587454764776839, |
|
"grad_norm": 0.376953125, |
|
"learning_rate": 0.00011514636239102393, |
|
"loss": 0.897, |
|
"step": 3145 |
|
}, |
|
{ |
|
"epoch": 7.599517490952955, |
|
"grad_norm": 0.328125, |
|
"learning_rate": 0.00011486849406425188, |
|
"loss": 0.8962, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 7.611580217129071, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 0.00011459050829147915, |
|
"loss": 0.8853, |
|
"step": 3155 |
|
}, |
|
{ |
|
"epoch": 7.623642943305187, |
|
"grad_norm": 0.3203125, |
|
"learning_rate": 0.00011431240726851091, |
|
"loss": 0.8978, |
|
"step": 3160 |
|
}, |
|
{ |
|
"epoch": 7.635705669481303, |
|
"grad_norm": 0.326171875, |
|
"learning_rate": 0.00011403419319206284, |
|
"loss": 0.894, |
|
"step": 3165 |
|
}, |
|
{ |
|
"epoch": 7.647768395657419, |
|
"grad_norm": 0.37890625, |
|
"learning_rate": 0.00011375586825974349, |
|
"loss": 0.8913, |
|
"step": 3170 |
|
}, |
|
{ |
|
"epoch": 7.659831121833534, |
|
"grad_norm": 0.32421875, |
|
"learning_rate": 0.0001134774346700371, |
|
"loss": 0.8915, |
|
"step": 3175 |
|
}, |
|
{ |
|
"epoch": 7.67189384800965, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 0.0001131988946222863, |
|
"loss": 0.8962, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 7.683956574185766, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 0.00011292025031667448, |
|
"loss": 0.8949, |
|
"step": 3185 |
|
}, |
|
{ |
|
"epoch": 7.696019300361882, |
|
"grad_norm": 0.33203125, |
|
"learning_rate": 0.00011264150395420865, |
|
"loss": 0.8974, |
|
"step": 3190 |
|
}, |
|
{ |
|
"epoch": 7.708082026537998, |
|
"grad_norm": 0.34765625, |
|
"learning_rate": 0.00011236265773670196, |
|
"loss": 0.8933, |
|
"step": 3195 |
|
}, |
|
{ |
|
"epoch": 7.720144752714114, |
|
"grad_norm": 0.330078125, |
|
"learning_rate": 0.00011208371386675624, |
|
"loss": 0.8996, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 7.7322074788902295, |
|
"grad_norm": 0.34765625, |
|
"learning_rate": 0.00011180467454774481, |
|
"loss": 0.892, |
|
"step": 3205 |
|
}, |
|
{ |
|
"epoch": 7.744270205066345, |
|
"grad_norm": 0.35546875, |
|
"learning_rate": 0.00011152554198379484, |
|
"loss": 0.8943, |
|
"step": 3210 |
|
}, |
|
{ |
|
"epoch": 7.7563329312424605, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 0.00011124631837976997, |
|
"loss": 0.8921, |
|
"step": 3215 |
|
}, |
|
{ |
|
"epoch": 7.768395657418576, |
|
"grad_norm": 0.373046875, |
|
"learning_rate": 0.00011096700594125318, |
|
"loss": 0.891, |
|
"step": 3220 |
|
}, |
|
{ |
|
"epoch": 7.780458383594692, |
|
"grad_norm": 0.328125, |
|
"learning_rate": 0.00011068760687452895, |
|
"loss": 0.8956, |
|
"step": 3225 |
|
}, |
|
{ |
|
"epoch": 7.792521109770808, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 0.00011040812338656615, |
|
"loss": 0.8986, |
|
"step": 3230 |
|
}, |
|
{ |
|
"epoch": 7.804583835946924, |
|
"grad_norm": 0.3203125, |
|
"learning_rate": 0.0001101285576850004, |
|
"loss": 0.8905, |
|
"step": 3235 |
|
}, |
|
{ |
|
"epoch": 7.81664656212304, |
|
"grad_norm": 0.306640625, |
|
"learning_rate": 0.00010984891197811687, |
|
"loss": 0.8861, |
|
"step": 3240 |
|
}, |
|
{ |
|
"epoch": 7.828709288299155, |
|
"grad_norm": 0.3359375, |
|
"learning_rate": 0.00010956918847483255, |
|
"loss": 0.8908, |
|
"step": 3245 |
|
}, |
|
{ |
|
"epoch": 7.840772014475271, |
|
"grad_norm": 0.328125, |
|
"learning_rate": 0.00010928938938467896, |
|
"loss": 0.8972, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 7.852834740651387, |
|
"grad_norm": 0.310546875, |
|
"learning_rate": 0.00010900951691778481, |
|
"loss": 0.8977, |
|
"step": 3255 |
|
}, |
|
{ |
|
"epoch": 7.864897466827503, |
|
"grad_norm": 0.3359375, |
|
"learning_rate": 0.00010872957328485819, |
|
"loss": 0.8974, |
|
"step": 3260 |
|
}, |
|
{ |
|
"epoch": 7.876960193003619, |
|
"grad_norm": 0.34765625, |
|
"learning_rate": 0.00010844956069716957, |
|
"loss": 0.8964, |
|
"step": 3265 |
|
}, |
|
{ |
|
"epoch": 7.889022919179735, |
|
"grad_norm": 0.30078125, |
|
"learning_rate": 0.00010816948136653386, |
|
"loss": 0.889, |
|
"step": 3270 |
|
}, |
|
{ |
|
"epoch": 7.901085645355851, |
|
"grad_norm": 0.34375, |
|
"learning_rate": 0.00010788933750529335, |
|
"loss": 0.8914, |
|
"step": 3275 |
|
}, |
|
{ |
|
"epoch": 7.913148371531967, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 0.00010760913132630002, |
|
"loss": 0.8923, |
|
"step": 3280 |
|
}, |
|
{ |
|
"epoch": 7.9252110977080825, |
|
"grad_norm": 0.328125, |
|
"learning_rate": 0.000107328865042898, |
|
"loss": 0.8954, |
|
"step": 3285 |
|
}, |
|
{ |
|
"epoch": 7.9372738238841976, |
|
"grad_norm": 0.34375, |
|
"learning_rate": 0.00010704854086890633, |
|
"loss": 0.8992, |
|
"step": 3290 |
|
}, |
|
{ |
|
"epoch": 7.9493365500603135, |
|
"grad_norm": 0.330078125, |
|
"learning_rate": 0.00010676816101860118, |
|
"loss": 0.8954, |
|
"step": 3295 |
|
}, |
|
{ |
|
"epoch": 7.961399276236429, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 0.00010648772770669861, |
|
"loss": 0.8952, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 7.973462002412545, |
|
"grad_norm": 0.353515625, |
|
"learning_rate": 0.00010620724314833696, |
|
"loss": 0.8903, |
|
"step": 3305 |
|
}, |
|
{ |
|
"epoch": 7.985524728588661, |
|
"grad_norm": 0.3359375, |
|
"learning_rate": 0.00010592670955905935, |
|
"loss": 0.8909, |
|
"step": 3310 |
|
}, |
|
{ |
|
"epoch": 7.997587454764777, |
|
"grad_norm": 0.357421875, |
|
"learning_rate": 0.00010564612915479612, |
|
"loss": 0.8999, |
|
"step": 3315 |
|
}, |
|
{ |
|
"epoch": 8.0, |
|
"eval_loss": 2.4829864501953125, |
|
"eval_runtime": 0.5356, |
|
"eval_samples_per_second": 18.67, |
|
"eval_steps_per_second": 1.867, |
|
"step": 3316 |
|
}, |
|
{ |
|
"epoch": 8.009650180940893, |
|
"grad_norm": 0.306640625, |
|
"learning_rate": 0.00010536550415184751, |
|
"loss": 0.889, |
|
"step": 3320 |
|
}, |
|
{ |
|
"epoch": 8.021712907117008, |
|
"grad_norm": 0.3359375, |
|
"learning_rate": 0.000105084836766866, |
|
"loss": 0.8866, |
|
"step": 3325 |
|
}, |
|
{ |
|
"epoch": 8.033775633293125, |
|
"grad_norm": 0.333984375, |
|
"learning_rate": 0.00010480412921683888, |
|
"loss": 0.8776, |
|
"step": 3330 |
|
}, |
|
{ |
|
"epoch": 8.04583835946924, |
|
"grad_norm": 0.361328125, |
|
"learning_rate": 0.00010452338371907064, |
|
"loss": 0.8824, |
|
"step": 3335 |
|
}, |
|
{ |
|
"epoch": 8.057901085645355, |
|
"grad_norm": 0.361328125, |
|
"learning_rate": 0.00010424260249116557, |
|
"loss": 0.8906, |
|
"step": 3340 |
|
}, |
|
{ |
|
"epoch": 8.069963811821472, |
|
"grad_norm": 0.328125, |
|
"learning_rate": 0.00010396178775101014, |
|
"loss": 0.8918, |
|
"step": 3345 |
|
}, |
|
{ |
|
"epoch": 8.082026537997587, |
|
"grad_norm": 0.359375, |
|
"learning_rate": 0.0001036809417167556, |
|
"loss": 0.8811, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 8.094089264173704, |
|
"grad_norm": 0.359375, |
|
"learning_rate": 0.00010340006660680032, |
|
"loss": 0.896, |
|
"step": 3355 |
|
}, |
|
{ |
|
"epoch": 8.106151990349819, |
|
"grad_norm": 0.34765625, |
|
"learning_rate": 0.00010311916463977242, |
|
"loss": 0.8858, |
|
"step": 3360 |
|
}, |
|
{ |
|
"epoch": 8.118214716525936, |
|
"grad_norm": 0.392578125, |
|
"learning_rate": 0.00010283823803451208, |
|
"loss": 0.8834, |
|
"step": 3365 |
|
}, |
|
{ |
|
"epoch": 8.13027744270205, |
|
"grad_norm": 0.32421875, |
|
"learning_rate": 0.00010255728901005417, |
|
"loss": 0.882, |
|
"step": 3370 |
|
}, |
|
{ |
|
"epoch": 8.142340168878167, |
|
"grad_norm": 0.33984375, |
|
"learning_rate": 0.00010227631978561056, |
|
"loss": 0.8819, |
|
"step": 3375 |
|
}, |
|
{ |
|
"epoch": 8.154402895054282, |
|
"grad_norm": 0.34765625, |
|
"learning_rate": 0.00010199533258055282, |
|
"loss": 0.8755, |
|
"step": 3380 |
|
}, |
|
{ |
|
"epoch": 8.166465621230397, |
|
"grad_norm": 0.333984375, |
|
"learning_rate": 0.00010171432961439438, |
|
"loss": 0.8847, |
|
"step": 3385 |
|
}, |
|
{ |
|
"epoch": 8.178528347406514, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 0.00010143331310677331, |
|
"loss": 0.8951, |
|
"step": 3390 |
|
}, |
|
{ |
|
"epoch": 8.19059107358263, |
|
"grad_norm": 0.373046875, |
|
"learning_rate": 0.00010115228527743452, |
|
"loss": 0.888, |
|
"step": 3395 |
|
}, |
|
{ |
|
"epoch": 8.202653799758746, |
|
"grad_norm": 0.345703125, |
|
"learning_rate": 0.00010087124834621248, |
|
"loss": 0.8875, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 8.214716525934861, |
|
"grad_norm": 0.345703125, |
|
"learning_rate": 0.00010059020453301345, |
|
"loss": 0.8902, |
|
"step": 3405 |
|
}, |
|
{ |
|
"epoch": 8.226779252110978, |
|
"grad_norm": 0.431640625, |
|
"learning_rate": 0.00010030915605779809, |
|
"loss": 0.8871, |
|
"step": 3410 |
|
}, |
|
{ |
|
"epoch": 8.238841978287093, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 0.00010002810514056391, |
|
"loss": 0.8884, |
|
"step": 3415 |
|
}, |
|
{ |
|
"epoch": 8.250904704463208, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 9.974705400132764e-05, |
|
"loss": 0.8859, |
|
"step": 3420 |
|
}, |
|
{ |
|
"epoch": 8.262967430639325, |
|
"grad_norm": 0.3125, |
|
"learning_rate": 9.946600486010785e-05, |
|
"loss": 0.8897, |
|
"step": 3425 |
|
}, |
|
{ |
|
"epoch": 8.27503015681544, |
|
"grad_norm": 0.326171875, |
|
"learning_rate": 9.918495993690724e-05, |
|
"loss": 0.8877, |
|
"step": 3430 |
|
}, |
|
{ |
|
"epoch": 8.287092882991557, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 9.890392145169531e-05, |
|
"loss": 0.8841, |
|
"step": 3435 |
|
}, |
|
{ |
|
"epoch": 8.299155609167672, |
|
"grad_norm": 0.447265625, |
|
"learning_rate": 9.862289162439059e-05, |
|
"loss": 0.8854, |
|
"step": 3440 |
|
}, |
|
{ |
|
"epoch": 8.311218335343789, |
|
"grad_norm": 0.341796875, |
|
"learning_rate": 9.834187267484326e-05, |
|
"loss": 0.8843, |
|
"step": 3445 |
|
}, |
|
{ |
|
"epoch": 8.323281061519904, |
|
"grad_norm": 0.361328125, |
|
"learning_rate": 9.806086682281758e-05, |
|
"loss": 0.8886, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 8.335343787696019, |
|
"grad_norm": 0.33984375, |
|
"learning_rate": 9.777987628797437e-05, |
|
"loss": 0.8878, |
|
"step": 3455 |
|
}, |
|
{ |
|
"epoch": 8.347406513872135, |
|
"grad_norm": 0.33203125, |
|
"learning_rate": 9.749890328985345e-05, |
|
"loss": 0.8811, |
|
"step": 3460 |
|
}, |
|
{ |
|
"epoch": 8.35946924004825, |
|
"grad_norm": 0.40234375, |
|
"learning_rate": 9.721795004785605e-05, |
|
"loss": 0.8902, |
|
"step": 3465 |
|
}, |
|
{ |
|
"epoch": 8.371531966224367, |
|
"grad_norm": 0.380859375, |
|
"learning_rate": 9.693701878122749e-05, |
|
"loss": 0.8854, |
|
"step": 3470 |
|
}, |
|
{ |
|
"epoch": 8.383594692400482, |
|
"grad_norm": 0.3671875, |
|
"learning_rate": 9.665611170903942e-05, |
|
"loss": 0.8934, |
|
"step": 3475 |
|
}, |
|
{ |
|
"epoch": 8.395657418576599, |
|
"grad_norm": 0.421875, |
|
"learning_rate": 9.637523105017229e-05, |
|
"loss": 0.8905, |
|
"step": 3480 |
|
}, |
|
{ |
|
"epoch": 8.407720144752714, |
|
"grad_norm": 0.4140625, |
|
"learning_rate": 9.609437902329814e-05, |
|
"loss": 0.8884, |
|
"step": 3485 |
|
}, |
|
{ |
|
"epoch": 8.41978287092883, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 9.581355784686266e-05, |
|
"loss": 0.8864, |
|
"step": 3490 |
|
}, |
|
{ |
|
"epoch": 8.431845597104946, |
|
"grad_norm": 0.35546875, |
|
"learning_rate": 9.553276973906786e-05, |
|
"loss": 0.8924, |
|
"step": 3495 |
|
}, |
|
{ |
|
"epoch": 8.443908323281061, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 9.525201691785462e-05, |
|
"loss": 0.8938, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 8.455971049457178, |
|
"grad_norm": 0.357421875, |
|
"learning_rate": 9.497130160088508e-05, |
|
"loss": 0.8937, |
|
"step": 3505 |
|
}, |
|
{ |
|
"epoch": 8.468033775633293, |
|
"grad_norm": 0.32421875, |
|
"learning_rate": 9.469062600552509e-05, |
|
"loss": 0.8847, |
|
"step": 3510 |
|
}, |
|
{ |
|
"epoch": 8.48009650180941, |
|
"grad_norm": 0.306640625, |
|
"learning_rate": 9.44099923488267e-05, |
|
"loss": 0.8882, |
|
"step": 3515 |
|
}, |
|
{ |
|
"epoch": 8.492159227985525, |
|
"grad_norm": 0.37890625, |
|
"learning_rate": 9.412940284751089e-05, |
|
"loss": 0.8885, |
|
"step": 3520 |
|
}, |
|
{ |
|
"epoch": 8.50422195416164, |
|
"grad_norm": 0.33984375, |
|
"learning_rate": 9.384885971794961e-05, |
|
"loss": 0.8938, |
|
"step": 3525 |
|
}, |
|
{ |
|
"epoch": 8.516284680337757, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 9.356836517614863e-05, |
|
"loss": 0.8938, |
|
"step": 3530 |
|
}, |
|
{ |
|
"epoch": 8.528347406513872, |
|
"grad_norm": 0.400390625, |
|
"learning_rate": 9.328792143773e-05, |
|
"loss": 0.8765, |
|
"step": 3535 |
|
}, |
|
{ |
|
"epoch": 8.540410132689988, |
|
"grad_norm": 0.38671875, |
|
"learning_rate": 9.300753071791434e-05, |
|
"loss": 0.8879, |
|
"step": 3540 |
|
}, |
|
{ |
|
"epoch": 8.552472858866103, |
|
"grad_norm": 0.361328125, |
|
"learning_rate": 9.272719523150355e-05, |
|
"loss": 0.8927, |
|
"step": 3545 |
|
}, |
|
{ |
|
"epoch": 8.56453558504222, |
|
"grad_norm": 0.35546875, |
|
"learning_rate": 9.24469171928632e-05, |
|
"loss": 0.8867, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 8.576598311218335, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 9.216669881590515e-05, |
|
"loss": 0.8951, |
|
"step": 3555 |
|
}, |
|
{ |
|
"epoch": 8.588661037394452, |
|
"grad_norm": 0.349609375, |
|
"learning_rate": 9.188654231406993e-05, |
|
"loss": 0.8918, |
|
"step": 3560 |
|
}, |
|
{ |
|
"epoch": 8.600723763570567, |
|
"grad_norm": 0.345703125, |
|
"learning_rate": 9.160644990030931e-05, |
|
"loss": 0.8894, |
|
"step": 3565 |
|
}, |
|
{ |
|
"epoch": 8.612786489746682, |
|
"grad_norm": 0.376953125, |
|
"learning_rate": 9.132642378706894e-05, |
|
"loss": 0.8868, |
|
"step": 3570 |
|
}, |
|
{ |
|
"epoch": 8.624849215922799, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 9.104646618627063e-05, |
|
"loss": 0.8848, |
|
"step": 3575 |
|
}, |
|
{ |
|
"epoch": 8.636911942098914, |
|
"grad_norm": 0.369140625, |
|
"learning_rate": 9.076657930929507e-05, |
|
"loss": 0.8997, |
|
"step": 3580 |
|
}, |
|
{ |
|
"epoch": 8.64897466827503, |
|
"grad_norm": 0.341796875, |
|
"learning_rate": 9.048676536696425e-05, |
|
"loss": 0.8922, |
|
"step": 3585 |
|
}, |
|
{ |
|
"epoch": 8.661037394451146, |
|
"grad_norm": 0.365234375, |
|
"learning_rate": 9.020702656952421e-05, |
|
"loss": 0.8889, |
|
"step": 3590 |
|
}, |
|
{ |
|
"epoch": 8.67310012062726, |
|
"grad_norm": 0.373046875, |
|
"learning_rate": 8.992736512662724e-05, |
|
"loss": 0.8801, |
|
"step": 3595 |
|
}, |
|
{ |
|
"epoch": 8.685162846803378, |
|
"grad_norm": 0.3125, |
|
"learning_rate": 8.964778324731467e-05, |
|
"loss": 0.8891, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 8.697225572979493, |
|
"grad_norm": 0.353515625, |
|
"learning_rate": 8.93682831399994e-05, |
|
"loss": 0.8902, |
|
"step": 3605 |
|
}, |
|
{ |
|
"epoch": 8.70928829915561, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 8.908886701244838e-05, |
|
"loss": 0.8901, |
|
"step": 3610 |
|
}, |
|
{ |
|
"epoch": 8.721351025331725, |
|
"grad_norm": 0.326171875, |
|
"learning_rate": 8.880953707176514e-05, |
|
"loss": 0.8915, |
|
"step": 3615 |
|
}, |
|
{ |
|
"epoch": 8.733413751507841, |
|
"grad_norm": 0.369140625, |
|
"learning_rate": 8.853029552437259e-05, |
|
"loss": 0.8906, |
|
"step": 3620 |
|
}, |
|
{ |
|
"epoch": 8.745476477683956, |
|
"grad_norm": 0.34765625, |
|
"learning_rate": 8.825114457599527e-05, |
|
"loss": 0.8869, |
|
"step": 3625 |
|
}, |
|
{ |
|
"epoch": 8.757539203860073, |
|
"grad_norm": 0.32421875, |
|
"learning_rate": 8.797208643164212e-05, |
|
"loss": 0.8998, |
|
"step": 3630 |
|
}, |
|
{ |
|
"epoch": 8.769601930036188, |
|
"grad_norm": 0.330078125, |
|
"learning_rate": 8.7693123295589e-05, |
|
"loss": 0.8776, |
|
"step": 3635 |
|
}, |
|
{ |
|
"epoch": 8.781664656212303, |
|
"grad_norm": 0.337890625, |
|
"learning_rate": 8.741425737136143e-05, |
|
"loss": 0.8951, |
|
"step": 3640 |
|
}, |
|
{ |
|
"epoch": 8.79372738238842, |
|
"grad_norm": 0.337890625, |
|
"learning_rate": 8.713549086171691e-05, |
|
"loss": 0.8865, |
|
"step": 3645 |
|
}, |
|
{ |
|
"epoch": 8.805790108564535, |
|
"grad_norm": 0.3203125, |
|
"learning_rate": 8.685682596862768e-05, |
|
"loss": 0.8861, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 8.817852834740652, |
|
"grad_norm": 0.35546875, |
|
"learning_rate": 8.65782648932634e-05, |
|
"loss": 0.8918, |
|
"step": 3655 |
|
}, |
|
{ |
|
"epoch": 8.829915560916767, |
|
"grad_norm": 0.330078125, |
|
"learning_rate": 8.629980983597358e-05, |
|
"loss": 0.893, |
|
"step": 3660 |
|
}, |
|
{ |
|
"epoch": 8.841978287092884, |
|
"grad_norm": 0.337890625, |
|
"learning_rate": 8.602146299627035e-05, |
|
"loss": 0.8871, |
|
"step": 3665 |
|
}, |
|
{ |
|
"epoch": 8.854041013268999, |
|
"grad_norm": 0.326171875, |
|
"learning_rate": 8.574322657281105e-05, |
|
"loss": 0.8899, |
|
"step": 3670 |
|
}, |
|
{ |
|
"epoch": 8.866103739445114, |
|
"grad_norm": 0.326171875, |
|
"learning_rate": 8.546510276338078e-05, |
|
"loss": 0.884, |
|
"step": 3675 |
|
}, |
|
{ |
|
"epoch": 8.87816646562123, |
|
"grad_norm": 0.44140625, |
|
"learning_rate": 8.518709376487515e-05, |
|
"loss": 0.8889, |
|
"step": 3680 |
|
}, |
|
{ |
|
"epoch": 8.890229191797346, |
|
"grad_norm": 0.333984375, |
|
"learning_rate": 8.490920177328282e-05, |
|
"loss": 0.8859, |
|
"step": 3685 |
|
}, |
|
{ |
|
"epoch": 8.902291917973463, |
|
"grad_norm": 0.3515625, |
|
"learning_rate": 8.463142898366834e-05, |
|
"loss": 0.8816, |
|
"step": 3690 |
|
}, |
|
{ |
|
"epoch": 8.914354644149578, |
|
"grad_norm": 0.330078125, |
|
"learning_rate": 8.435377759015455e-05, |
|
"loss": 0.8858, |
|
"step": 3695 |
|
}, |
|
{ |
|
"epoch": 8.926417370325694, |
|
"grad_norm": 0.326171875, |
|
"learning_rate": 8.407624978590543e-05, |
|
"loss": 0.8767, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 8.93848009650181, |
|
"grad_norm": 0.3984375, |
|
"learning_rate": 8.37988477631088e-05, |
|
"loss": 0.8946, |
|
"step": 3705 |
|
}, |
|
{ |
|
"epoch": 8.950542822677924, |
|
"grad_norm": 0.3125, |
|
"learning_rate": 8.352157371295884e-05, |
|
"loss": 0.883, |
|
"step": 3710 |
|
}, |
|
{ |
|
"epoch": 8.962605548854041, |
|
"grad_norm": 0.326171875, |
|
"learning_rate": 8.324442982563887e-05, |
|
"loss": 0.8866, |
|
"step": 3715 |
|
}, |
|
{ |
|
"epoch": 8.974668275030156, |
|
"grad_norm": 0.330078125, |
|
"learning_rate": 8.296741829030418e-05, |
|
"loss": 0.8844, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 8.986731001206273, |
|
"grad_norm": 0.34375, |
|
"learning_rate": 8.269054129506449e-05, |
|
"loss": 0.88, |
|
"step": 3725 |
|
}, |
|
{ |
|
"epoch": 8.998793727382388, |
|
"grad_norm": 0.341796875, |
|
"learning_rate": 8.241380102696681e-05, |
|
"loss": 0.8762, |
|
"step": 3730 |
|
}, |
|
{ |
|
"epoch": 8.998793727382388, |
|
"eval_loss": 2.493363618850708, |
|
"eval_runtime": 0.66, |
|
"eval_samples_per_second": 15.152, |
|
"eval_steps_per_second": 1.515, |
|
"step": 3730 |
|
}, |
|
{ |
|
"epoch": 9.010856453558505, |
|
"grad_norm": 0.36328125, |
|
"learning_rate": 8.213719967197817e-05, |
|
"loss": 0.8809, |
|
"step": 3735 |
|
}, |
|
{ |
|
"epoch": 9.02291917973462, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 8.186073941496835e-05, |
|
"loss": 0.8795, |
|
"step": 3740 |
|
}, |
|
{ |
|
"epoch": 9.034981905910735, |
|
"grad_norm": 0.33203125, |
|
"learning_rate": 8.158442243969256e-05, |
|
"loss": 0.8773, |
|
"step": 3745 |
|
}, |
|
{ |
|
"epoch": 9.047044632086852, |
|
"grad_norm": 0.3359375, |
|
"learning_rate": 8.130825092877418e-05, |
|
"loss": 0.8885, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 9.059107358262967, |
|
"grad_norm": 0.345703125, |
|
"learning_rate": 8.103222706368774e-05, |
|
"loss": 0.8857, |
|
"step": 3755 |
|
}, |
|
{ |
|
"epoch": 9.071170084439084, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 8.075635302474134e-05, |
|
"loss": 0.8747, |
|
"step": 3760 |
|
}, |
|
{ |
|
"epoch": 9.083232810615199, |
|
"grad_norm": 0.34765625, |
|
"learning_rate": 8.04806309910597e-05, |
|
"loss": 0.879, |
|
"step": 3765 |
|
}, |
|
{ |
|
"epoch": 9.095295536791316, |
|
"grad_norm": 0.35546875, |
|
"learning_rate": 8.020506314056677e-05, |
|
"loss": 0.878, |
|
"step": 3770 |
|
}, |
|
{ |
|
"epoch": 9.10735826296743, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 7.992965164996875e-05, |
|
"loss": 0.8809, |
|
"step": 3775 |
|
}, |
|
{ |
|
"epoch": 9.119420989143546, |
|
"grad_norm": 0.35546875, |
|
"learning_rate": 7.965439869473664e-05, |
|
"loss": 0.8815, |
|
"step": 3780 |
|
}, |
|
{ |
|
"epoch": 9.131483715319662, |
|
"grad_norm": 0.35546875, |
|
"learning_rate": 7.937930644908917e-05, |
|
"loss": 0.8826, |
|
"step": 3785 |
|
}, |
|
{ |
|
"epoch": 9.143546441495777, |
|
"grad_norm": 0.345703125, |
|
"learning_rate": 7.91043770859757e-05, |
|
"loss": 0.8863, |
|
"step": 3790 |
|
}, |
|
{ |
|
"epoch": 9.155609167671894, |
|
"grad_norm": 0.40234375, |
|
"learning_rate": 7.882961277705895e-05, |
|
"loss": 0.8792, |
|
"step": 3795 |
|
}, |
|
{ |
|
"epoch": 9.16767189384801, |
|
"grad_norm": 0.326171875, |
|
"learning_rate": 7.855501569269782e-05, |
|
"loss": 0.8743, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 9.179734620024126, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 7.82805880019304e-05, |
|
"loss": 0.881, |
|
"step": 3805 |
|
}, |
|
{ |
|
"epoch": 9.191797346200241, |
|
"grad_norm": 0.3203125, |
|
"learning_rate": 7.800633187245673e-05, |
|
"loss": 0.8873, |
|
"step": 3810 |
|
}, |
|
{ |
|
"epoch": 9.203860072376358, |
|
"grad_norm": 0.341796875, |
|
"learning_rate": 7.773224947062163e-05, |
|
"loss": 0.8766, |
|
"step": 3815 |
|
}, |
|
{ |
|
"epoch": 9.215922798552473, |
|
"grad_norm": 0.384765625, |
|
"learning_rate": 7.745834296139763e-05, |
|
"loss": 0.8818, |
|
"step": 3820 |
|
}, |
|
{ |
|
"epoch": 9.227985524728588, |
|
"grad_norm": 0.328125, |
|
"learning_rate": 7.718461450836804e-05, |
|
"loss": 0.8894, |
|
"step": 3825 |
|
}, |
|
{ |
|
"epoch": 9.240048250904705, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 7.691106627370955e-05, |
|
"loss": 0.8866, |
|
"step": 3830 |
|
}, |
|
{ |
|
"epoch": 9.25211097708082, |
|
"grad_norm": 0.341796875, |
|
"learning_rate": 7.663770041817534e-05, |
|
"loss": 0.8854, |
|
"step": 3835 |
|
}, |
|
{ |
|
"epoch": 9.264173703256937, |
|
"grad_norm": 0.328125, |
|
"learning_rate": 7.636451910107806e-05, |
|
"loss": 0.8832, |
|
"step": 3840 |
|
}, |
|
{ |
|
"epoch": 9.276236429433052, |
|
"grad_norm": 0.33984375, |
|
"learning_rate": 7.609152448027261e-05, |
|
"loss": 0.8853, |
|
"step": 3845 |
|
}, |
|
{ |
|
"epoch": 9.288299155609169, |
|
"grad_norm": 0.33984375, |
|
"learning_rate": 7.581871871213913e-05, |
|
"loss": 0.8826, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 9.300361881785284, |
|
"grad_norm": 0.384765625, |
|
"learning_rate": 7.554610395156624e-05, |
|
"loss": 0.8851, |
|
"step": 3855 |
|
}, |
|
{ |
|
"epoch": 9.312424607961399, |
|
"grad_norm": 0.32421875, |
|
"learning_rate": 7.527368235193355e-05, |
|
"loss": 0.8812, |
|
"step": 3860 |
|
}, |
|
{ |
|
"epoch": 9.324487334137515, |
|
"grad_norm": 0.412109375, |
|
"learning_rate": 7.500145606509501e-05, |
|
"loss": 0.8883, |
|
"step": 3865 |
|
}, |
|
{ |
|
"epoch": 9.33655006031363, |
|
"grad_norm": 0.337890625, |
|
"learning_rate": 7.472942724136174e-05, |
|
"loss": 0.885, |
|
"step": 3870 |
|
}, |
|
{ |
|
"epoch": 9.348612786489747, |
|
"grad_norm": 0.337890625, |
|
"learning_rate": 7.44575980294852e-05, |
|
"loss": 0.8857, |
|
"step": 3875 |
|
}, |
|
{ |
|
"epoch": 9.360675512665862, |
|
"grad_norm": 0.369140625, |
|
"learning_rate": 7.418597057664005e-05, |
|
"loss": 0.8872, |
|
"step": 3880 |
|
}, |
|
{ |
|
"epoch": 9.372738238841979, |
|
"grad_norm": 0.3125, |
|
"learning_rate": 7.391454702840722e-05, |
|
"loss": 0.8826, |
|
"step": 3885 |
|
}, |
|
{ |
|
"epoch": 9.384800965018094, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 7.36433295287571e-05, |
|
"loss": 0.8822, |
|
"step": 3890 |
|
}, |
|
{ |
|
"epoch": 9.39686369119421, |
|
"grad_norm": 0.34765625, |
|
"learning_rate": 7.337232022003244e-05, |
|
"loss": 0.8833, |
|
"step": 3895 |
|
}, |
|
{ |
|
"epoch": 9.408926417370326, |
|
"grad_norm": 0.359375, |
|
"learning_rate": 7.310152124293146e-05, |
|
"loss": 0.8782, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 9.420989143546441, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 7.28309347364911e-05, |
|
"loss": 0.8879, |
|
"step": 3905 |
|
}, |
|
{ |
|
"epoch": 9.433051869722558, |
|
"grad_norm": 0.326171875, |
|
"learning_rate": 7.256056283806986e-05, |
|
"loss": 0.8806, |
|
"step": 3910 |
|
}, |
|
{ |
|
"epoch": 9.445114595898673, |
|
"grad_norm": 0.33984375, |
|
"learning_rate": 7.229040768333115e-05, |
|
"loss": 0.8841, |
|
"step": 3915 |
|
}, |
|
{ |
|
"epoch": 9.45717732207479, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 7.202047140622621e-05, |
|
"loss": 0.8718, |
|
"step": 3920 |
|
}, |
|
{ |
|
"epoch": 9.469240048250905, |
|
"grad_norm": 0.341796875, |
|
"learning_rate": 7.175075613897756e-05, |
|
"loss": 0.8843, |
|
"step": 3925 |
|
}, |
|
{ |
|
"epoch": 9.48130277442702, |
|
"grad_norm": 0.34765625, |
|
"learning_rate": 7.14812640120618e-05, |
|
"loss": 0.8788, |
|
"step": 3930 |
|
}, |
|
{ |
|
"epoch": 9.493365500603137, |
|
"grad_norm": 0.33984375, |
|
"learning_rate": 7.121199715419295e-05, |
|
"loss": 0.8891, |
|
"step": 3935 |
|
}, |
|
{ |
|
"epoch": 9.505428226779252, |
|
"grad_norm": 0.3359375, |
|
"learning_rate": 7.094295769230577e-05, |
|
"loss": 0.8808, |
|
"step": 3940 |
|
}, |
|
{ |
|
"epoch": 9.517490952955368, |
|
"grad_norm": 0.306640625, |
|
"learning_rate": 7.067414775153871e-05, |
|
"loss": 0.8817, |
|
"step": 3945 |
|
}, |
|
{ |
|
"epoch": 9.529553679131483, |
|
"grad_norm": 0.359375, |
|
"learning_rate": 7.040556945521721e-05, |
|
"loss": 0.8837, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 9.5416164053076, |
|
"grad_norm": 0.30859375, |
|
"learning_rate": 7.013722492483708e-05, |
|
"loss": 0.8867, |
|
"step": 3955 |
|
}, |
|
{ |
|
"epoch": 9.553679131483715, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 6.986911628004753e-05, |
|
"loss": 0.8807, |
|
"step": 3960 |
|
}, |
|
{ |
|
"epoch": 9.56574185765983, |
|
"grad_norm": 0.306640625, |
|
"learning_rate": 6.96012456386345e-05, |
|
"loss": 0.873, |
|
"step": 3965 |
|
}, |
|
{ |
|
"epoch": 9.577804583835947, |
|
"grad_norm": 0.330078125, |
|
"learning_rate": 6.933361511650395e-05, |
|
"loss": 0.8852, |
|
"step": 3970 |
|
}, |
|
{ |
|
"epoch": 9.589867310012062, |
|
"grad_norm": 0.326171875, |
|
"learning_rate": 6.906622682766526e-05, |
|
"loss": 0.8914, |
|
"step": 3975 |
|
}, |
|
{ |
|
"epoch": 9.601930036188179, |
|
"grad_norm": 0.357421875, |
|
"learning_rate": 6.879908288421425e-05, |
|
"loss": 0.8691, |
|
"step": 3980 |
|
}, |
|
{ |
|
"epoch": 9.613992762364294, |
|
"grad_norm": 0.341796875, |
|
"learning_rate": 6.853218539631672e-05, |
|
"loss": 0.8839, |
|
"step": 3985 |
|
}, |
|
{ |
|
"epoch": 9.62605548854041, |
|
"grad_norm": 0.310546875, |
|
"learning_rate": 6.826553647219175e-05, |
|
"loss": 0.8879, |
|
"step": 3990 |
|
}, |
|
{ |
|
"epoch": 9.638118214716526, |
|
"grad_norm": 0.349609375, |
|
"learning_rate": 6.799913821809501e-05, |
|
"loss": 0.8807, |
|
"step": 3995 |
|
}, |
|
{ |
|
"epoch": 9.650180940892643, |
|
"grad_norm": 0.34375, |
|
"learning_rate": 6.773299273830207e-05, |
|
"loss": 0.8771, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 9.662243667068758, |
|
"grad_norm": 0.341796875, |
|
"learning_rate": 6.74671021350919e-05, |
|
"loss": 0.8814, |
|
"step": 4005 |
|
}, |
|
{ |
|
"epoch": 9.674306393244873, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 6.720146850873025e-05, |
|
"loss": 0.8844, |
|
"step": 4010 |
|
}, |
|
{ |
|
"epoch": 9.68636911942099, |
|
"grad_norm": 0.296875, |
|
"learning_rate": 6.69360939574529e-05, |
|
"loss": 0.8801, |
|
"step": 4015 |
|
}, |
|
{ |
|
"epoch": 9.698431845597105, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 6.667098057744927e-05, |
|
"loss": 0.8837, |
|
"step": 4020 |
|
}, |
|
{ |
|
"epoch": 9.710494571773221, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 6.640613046284581e-05, |
|
"loss": 0.8821, |
|
"step": 4025 |
|
}, |
|
{ |
|
"epoch": 9.722557297949336, |
|
"grad_norm": 0.310546875, |
|
"learning_rate": 6.614154570568934e-05, |
|
"loss": 0.89, |
|
"step": 4030 |
|
}, |
|
{ |
|
"epoch": 9.734620024125451, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 6.587722839593073e-05, |
|
"loss": 0.8775, |
|
"step": 4035 |
|
}, |
|
{ |
|
"epoch": 9.746682750301568, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 6.56131806214083e-05, |
|
"loss": 0.8802, |
|
"step": 4040 |
|
}, |
|
{ |
|
"epoch": 9.758745476477683, |
|
"grad_norm": 0.373046875, |
|
"learning_rate": 6.534940446783122e-05, |
|
"loss": 0.8763, |
|
"step": 4045 |
|
}, |
|
{ |
|
"epoch": 9.7708082026538, |
|
"grad_norm": 0.435546875, |
|
"learning_rate": 6.508590201876317e-05, |
|
"loss": 0.8889, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 9.782870928829915, |
|
"grad_norm": 0.33984375, |
|
"learning_rate": 6.482267535560583e-05, |
|
"loss": 0.8794, |
|
"step": 4055 |
|
}, |
|
{ |
|
"epoch": 9.794933655006032, |
|
"grad_norm": 0.3203125, |
|
"learning_rate": 6.455972655758253e-05, |
|
"loss": 0.8911, |
|
"step": 4060 |
|
}, |
|
{ |
|
"epoch": 9.806996381182147, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 6.429705770172168e-05, |
|
"loss": 0.8822, |
|
"step": 4065 |
|
}, |
|
{ |
|
"epoch": 9.819059107358264, |
|
"grad_norm": 0.32421875, |
|
"learning_rate": 6.403467086284041e-05, |
|
"loss": 0.881, |
|
"step": 4070 |
|
}, |
|
{ |
|
"epoch": 9.831121833534379, |
|
"grad_norm": 0.30078125, |
|
"learning_rate": 6.377256811352825e-05, |
|
"loss": 0.8814, |
|
"step": 4075 |
|
}, |
|
{ |
|
"epoch": 9.843184559710494, |
|
"grad_norm": 0.328125, |
|
"learning_rate": 6.351075152413068e-05, |
|
"loss": 0.8883, |
|
"step": 4080 |
|
}, |
|
{ |
|
"epoch": 9.85524728588661, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 6.324922316273282e-05, |
|
"loss": 0.8806, |
|
"step": 4085 |
|
}, |
|
{ |
|
"epoch": 9.867310012062726, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 6.298798509514312e-05, |
|
"loss": 0.8843, |
|
"step": 4090 |
|
}, |
|
{ |
|
"epoch": 9.879372738238843, |
|
"grad_norm": 0.353515625, |
|
"learning_rate": 6.272703938487694e-05, |
|
"loss": 0.8865, |
|
"step": 4095 |
|
}, |
|
{ |
|
"epoch": 9.891435464414958, |
|
"grad_norm": 0.3515625, |
|
"learning_rate": 6.246638809314036e-05, |
|
"loss": 0.882, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 9.903498190591074, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 6.220603327881375e-05, |
|
"loss": 0.8865, |
|
"step": 4105 |
|
}, |
|
{ |
|
"epoch": 9.91556091676719, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 6.194597699843581e-05, |
|
"loss": 0.8814, |
|
"step": 4110 |
|
}, |
|
{ |
|
"epoch": 9.927623642943304, |
|
"grad_norm": 0.3515625, |
|
"learning_rate": 6.168622130618694e-05, |
|
"loss": 0.8799, |
|
"step": 4115 |
|
}, |
|
{ |
|
"epoch": 9.939686369119421, |
|
"grad_norm": 0.3046875, |
|
"learning_rate": 6.142676825387328e-05, |
|
"loss": 0.8835, |
|
"step": 4120 |
|
}, |
|
{ |
|
"epoch": 9.951749095295536, |
|
"grad_norm": 0.32421875, |
|
"learning_rate": 6.116761989091042e-05, |
|
"loss": 0.882, |
|
"step": 4125 |
|
}, |
|
{ |
|
"epoch": 9.963811821471653, |
|
"grad_norm": 0.326171875, |
|
"learning_rate": 6.090877826430721e-05, |
|
"loss": 0.8727, |
|
"step": 4130 |
|
}, |
|
{ |
|
"epoch": 9.975874547647768, |
|
"grad_norm": 0.404296875, |
|
"learning_rate": 6.065024541864952e-05, |
|
"loss": 0.8867, |
|
"step": 4135 |
|
}, |
|
{ |
|
"epoch": 9.987937273823885, |
|
"grad_norm": 0.3515625, |
|
"learning_rate": 6.039202339608432e-05, |
|
"loss": 0.8825, |
|
"step": 4140 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"grad_norm": 0.333984375, |
|
"learning_rate": 6.0134114236303275e-05, |
|
"loss": 0.8821, |
|
"step": 4145 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"eval_loss": 2.4973604679107666, |
|
"eval_runtime": 0.5303, |
|
"eval_samples_per_second": 18.857, |
|
"eval_steps_per_second": 1.886, |
|
"step": 4145 |
|
}, |
|
{ |
|
"epoch": 10.012062726176115, |
|
"grad_norm": 0.337890625, |
|
"learning_rate": 5.987651997652677e-05, |
|
"loss": 0.8757, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 10.024125452352232, |
|
"grad_norm": 0.3125, |
|
"learning_rate": 5.961924265148777e-05, |
|
"loss": 0.8787, |
|
"step": 4155 |
|
}, |
|
{ |
|
"epoch": 10.036188178528347, |
|
"grad_norm": 0.326171875, |
|
"learning_rate": 5.93622842934159e-05, |
|
"loss": 0.8815, |
|
"step": 4160 |
|
}, |
|
{ |
|
"epoch": 10.048250904704464, |
|
"grad_norm": 0.3515625, |
|
"learning_rate": 5.9105646932021155e-05, |
|
"loss": 0.8806, |
|
"step": 4165 |
|
}, |
|
{ |
|
"epoch": 10.060313630880579, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 5.884933259447798e-05, |
|
"loss": 0.8848, |
|
"step": 4170 |
|
}, |
|
{ |
|
"epoch": 10.072376357056696, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 5.8593343305409355e-05, |
|
"loss": 0.8788, |
|
"step": 4175 |
|
}, |
|
{ |
|
"epoch": 10.08443908323281, |
|
"grad_norm": 0.310546875, |
|
"learning_rate": 5.833768108687063e-05, |
|
"loss": 0.8794, |
|
"step": 4180 |
|
}, |
|
{ |
|
"epoch": 10.096501809408926, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 5.8082347958333625e-05, |
|
"loss": 0.8835, |
|
"step": 4185 |
|
}, |
|
{ |
|
"epoch": 10.108564535585042, |
|
"grad_norm": 0.359375, |
|
"learning_rate": 5.7827345936670674e-05, |
|
"loss": 0.872, |
|
"step": 4190 |
|
}, |
|
{ |
|
"epoch": 10.120627261761157, |
|
"grad_norm": 0.333984375, |
|
"learning_rate": 5.7572677036138826e-05, |
|
"loss": 0.8782, |
|
"step": 4195 |
|
}, |
|
{ |
|
"epoch": 10.132689987937274, |
|
"grad_norm": 0.330078125, |
|
"learning_rate": 5.731834326836366e-05, |
|
"loss": 0.8814, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 10.14475271411339, |
|
"grad_norm": 0.369140625, |
|
"learning_rate": 5.7064346642323584e-05, |
|
"loss": 0.8783, |
|
"step": 4205 |
|
}, |
|
{ |
|
"epoch": 10.156815440289506, |
|
"grad_norm": 0.357421875, |
|
"learning_rate": 5.681068916433403e-05, |
|
"loss": 0.8805, |
|
"step": 4210 |
|
}, |
|
{ |
|
"epoch": 10.168878166465621, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 5.6557372838031384e-05, |
|
"loss": 0.8792, |
|
"step": 4215 |
|
}, |
|
{ |
|
"epoch": 10.180940892641736, |
|
"grad_norm": 0.34765625, |
|
"learning_rate": 5.6304399664357275e-05, |
|
"loss": 0.8759, |
|
"step": 4220 |
|
}, |
|
{ |
|
"epoch": 10.193003618817853, |
|
"grad_norm": 0.3125, |
|
"learning_rate": 5.605177164154293e-05, |
|
"loss": 0.8775, |
|
"step": 4225 |
|
}, |
|
{ |
|
"epoch": 10.205066344993968, |
|
"grad_norm": 0.3203125, |
|
"learning_rate": 5.579949076509305e-05, |
|
"loss": 0.8815, |
|
"step": 4230 |
|
}, |
|
{ |
|
"epoch": 10.217129071170085, |
|
"grad_norm": 0.310546875, |
|
"learning_rate": 5.5547559027770325e-05, |
|
"loss": 0.863, |
|
"step": 4235 |
|
}, |
|
{ |
|
"epoch": 10.2291917973462, |
|
"grad_norm": 0.32421875, |
|
"learning_rate": 5.529597841957959e-05, |
|
"loss": 0.8743, |
|
"step": 4240 |
|
}, |
|
{ |
|
"epoch": 10.241254523522317, |
|
"grad_norm": 0.3515625, |
|
"learning_rate": 5.5044750927752106e-05, |
|
"loss": 0.8785, |
|
"step": 4245 |
|
}, |
|
{ |
|
"epoch": 10.253317249698432, |
|
"grad_norm": 0.3203125, |
|
"learning_rate": 5.479387853672988e-05, |
|
"loss": 0.8758, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 10.265379975874549, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 5.4543363228149946e-05, |
|
"loss": 0.8833, |
|
"step": 4255 |
|
}, |
|
{ |
|
"epoch": 10.277442702050664, |
|
"grad_norm": 0.349609375, |
|
"learning_rate": 5.429320698082887e-05, |
|
"loss": 0.88, |
|
"step": 4260 |
|
}, |
|
{ |
|
"epoch": 10.289505428226779, |
|
"grad_norm": 0.326171875, |
|
"learning_rate": 5.404341177074686e-05, |
|
"loss": 0.876, |
|
"step": 4265 |
|
}, |
|
{ |
|
"epoch": 10.301568154402895, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 5.379397957103231e-05, |
|
"loss": 0.873, |
|
"step": 4270 |
|
}, |
|
{ |
|
"epoch": 10.31363088057901, |
|
"grad_norm": 0.337890625, |
|
"learning_rate": 5.354491235194635e-05, |
|
"loss": 0.8816, |
|
"step": 4275 |
|
}, |
|
{ |
|
"epoch": 10.325693606755127, |
|
"grad_norm": 0.3359375, |
|
"learning_rate": 5.3296212080866967e-05, |
|
"loss": 0.8828, |
|
"step": 4280 |
|
}, |
|
{ |
|
"epoch": 10.337756332931242, |
|
"grad_norm": 0.330078125, |
|
"learning_rate": 5.304788072227367e-05, |
|
"loss": 0.8825, |
|
"step": 4285 |
|
}, |
|
{ |
|
"epoch": 10.349819059107359, |
|
"grad_norm": 0.302734375, |
|
"learning_rate": 5.279992023773195e-05, |
|
"loss": 0.8735, |
|
"step": 4290 |
|
}, |
|
{ |
|
"epoch": 10.361881785283474, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 5.255233258587784e-05, |
|
"loss": 0.881, |
|
"step": 4295 |
|
}, |
|
{ |
|
"epoch": 10.37394451145959, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 5.2305119722402254e-05, |
|
"loss": 0.8755, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 10.386007237635706, |
|
"grad_norm": 0.328125, |
|
"learning_rate": 5.205828360003568e-05, |
|
"loss": 0.8776, |
|
"step": 4305 |
|
}, |
|
{ |
|
"epoch": 10.398069963811821, |
|
"grad_norm": 0.3125, |
|
"learning_rate": 5.181182616853286e-05, |
|
"loss": 0.8838, |
|
"step": 4310 |
|
}, |
|
{ |
|
"epoch": 10.410132689987938, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 5.15657493746571e-05, |
|
"loss": 0.8817, |
|
"step": 4315 |
|
}, |
|
{ |
|
"epoch": 10.422195416164053, |
|
"grad_norm": 0.34765625, |
|
"learning_rate": 5.1320055162165115e-05, |
|
"loss": 0.8752, |
|
"step": 4320 |
|
}, |
|
{ |
|
"epoch": 10.43425814234017, |
|
"grad_norm": 0.337890625, |
|
"learning_rate": 5.10747454717917e-05, |
|
"loss": 0.8791, |
|
"step": 4325 |
|
}, |
|
{ |
|
"epoch": 10.446320868516285, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 5.082982224123418e-05, |
|
"loss": 0.8745, |
|
"step": 4330 |
|
}, |
|
{ |
|
"epoch": 10.4583835946924, |
|
"grad_norm": 0.349609375, |
|
"learning_rate": 5.0585287405137305e-05, |
|
"loss": 0.8712, |
|
"step": 4335 |
|
}, |
|
{ |
|
"epoch": 10.470446320868517, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 5.034114289507792e-05, |
|
"loss": 0.871, |
|
"step": 4340 |
|
}, |
|
{ |
|
"epoch": 10.482509047044632, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 5.009739063954964e-05, |
|
"loss": 0.8759, |
|
"step": 4345 |
|
}, |
|
{ |
|
"epoch": 10.494571773220748, |
|
"grad_norm": 0.33203125, |
|
"learning_rate": 4.9854032563947714e-05, |
|
"loss": 0.8772, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 10.506634499396863, |
|
"grad_norm": 0.306640625, |
|
"learning_rate": 4.9611070590553724e-05, |
|
"loss": 0.8886, |
|
"step": 4355 |
|
}, |
|
{ |
|
"epoch": 10.51869722557298, |
|
"grad_norm": 0.341796875, |
|
"learning_rate": 4.936850663852053e-05, |
|
"loss": 0.8729, |
|
"step": 4360 |
|
}, |
|
{ |
|
"epoch": 10.530759951749095, |
|
"grad_norm": 0.337890625, |
|
"learning_rate": 4.912634262385695e-05, |
|
"loss": 0.8781, |
|
"step": 4365 |
|
}, |
|
{ |
|
"epoch": 10.54282267792521, |
|
"grad_norm": 0.3125, |
|
"learning_rate": 4.888458045941269e-05, |
|
"loss": 0.8788, |
|
"step": 4370 |
|
}, |
|
{ |
|
"epoch": 10.554885404101327, |
|
"grad_norm": 0.330078125, |
|
"learning_rate": 4.864322205486335e-05, |
|
"loss": 0.8897, |
|
"step": 4375 |
|
}, |
|
{ |
|
"epoch": 10.566948130277442, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 4.8402269316695134e-05, |
|
"loss": 0.878, |
|
"step": 4380 |
|
}, |
|
{ |
|
"epoch": 10.579010856453559, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 4.8161724148189934e-05, |
|
"loss": 0.8773, |
|
"step": 4385 |
|
}, |
|
{ |
|
"epoch": 10.591073582629674, |
|
"grad_norm": 0.330078125, |
|
"learning_rate": 4.792158844941023e-05, |
|
"loss": 0.8813, |
|
"step": 4390 |
|
}, |
|
{ |
|
"epoch": 10.60313630880579, |
|
"grad_norm": 0.328125, |
|
"learning_rate": 4.768186411718417e-05, |
|
"loss": 0.8865, |
|
"step": 4395 |
|
}, |
|
{ |
|
"epoch": 10.615199034981906, |
|
"grad_norm": 0.3203125, |
|
"learning_rate": 4.7442553045090474e-05, |
|
"loss": 0.8789, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 10.627261761158021, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 4.720365712344349e-05, |
|
"loss": 0.8774, |
|
"step": 4405 |
|
}, |
|
{ |
|
"epoch": 10.639324487334138, |
|
"grad_norm": 0.373046875, |
|
"learning_rate": 4.696517823927842e-05, |
|
"loss": 0.8797, |
|
"step": 4410 |
|
}, |
|
{ |
|
"epoch": 10.651387213510253, |
|
"grad_norm": 0.3359375, |
|
"learning_rate": 4.672711827633618e-05, |
|
"loss": 0.877, |
|
"step": 4415 |
|
}, |
|
{ |
|
"epoch": 10.66344993968637, |
|
"grad_norm": 0.353515625, |
|
"learning_rate": 4.6489479115048676e-05, |
|
"loss": 0.8766, |
|
"step": 4420 |
|
}, |
|
{ |
|
"epoch": 10.675512665862485, |
|
"grad_norm": 0.33984375, |
|
"learning_rate": 4.625226263252386e-05, |
|
"loss": 0.8804, |
|
"step": 4425 |
|
}, |
|
{ |
|
"epoch": 10.687575392038601, |
|
"grad_norm": 0.3203125, |
|
"learning_rate": 4.601547070253107e-05, |
|
"loss": 0.8889, |
|
"step": 4430 |
|
}, |
|
{ |
|
"epoch": 10.699638118214716, |
|
"grad_norm": 0.341796875, |
|
"learning_rate": 4.577910519548602e-05, |
|
"loss": 0.8747, |
|
"step": 4435 |
|
}, |
|
{ |
|
"epoch": 10.711700844390833, |
|
"grad_norm": 0.3515625, |
|
"learning_rate": 4.554316797843609e-05, |
|
"loss": 0.8816, |
|
"step": 4440 |
|
}, |
|
{ |
|
"epoch": 10.723763570566948, |
|
"grad_norm": 0.32421875, |
|
"learning_rate": 4.53076609150457e-05, |
|
"loss": 0.8796, |
|
"step": 4445 |
|
}, |
|
{ |
|
"epoch": 10.735826296743063, |
|
"grad_norm": 0.328125, |
|
"learning_rate": 4.5072585865581395e-05, |
|
"loss": 0.8836, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 10.74788902291918, |
|
"grad_norm": 0.3203125, |
|
"learning_rate": 4.483794468689728e-05, |
|
"loss": 0.8779, |
|
"step": 4455 |
|
}, |
|
{ |
|
"epoch": 10.759951749095295, |
|
"grad_norm": 0.326171875, |
|
"learning_rate": 4.460373923242041e-05, |
|
"loss": 0.8766, |
|
"step": 4460 |
|
}, |
|
{ |
|
"epoch": 10.772014475271412, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 4.436997135213592e-05, |
|
"loss": 0.8831, |
|
"step": 4465 |
|
}, |
|
{ |
|
"epoch": 10.784077201447527, |
|
"grad_norm": 0.3046875, |
|
"learning_rate": 4.413664289257265e-05, |
|
"loss": 0.8775, |
|
"step": 4470 |
|
}, |
|
{ |
|
"epoch": 10.796139927623642, |
|
"grad_norm": 0.3203125, |
|
"learning_rate": 4.3903755696788386e-05, |
|
"loss": 0.8744, |
|
"step": 4475 |
|
}, |
|
{ |
|
"epoch": 10.808202653799759, |
|
"grad_norm": 0.306640625, |
|
"learning_rate": 4.3671311604355494e-05, |
|
"loss": 0.8816, |
|
"step": 4480 |
|
}, |
|
{ |
|
"epoch": 10.820265379975874, |
|
"grad_norm": 0.3359375, |
|
"learning_rate": 4.343931245134616e-05, |
|
"loss": 0.8757, |
|
"step": 4485 |
|
}, |
|
{ |
|
"epoch": 10.83232810615199, |
|
"grad_norm": 0.34375, |
|
"learning_rate": 4.3207760070318026e-05, |
|
"loss": 0.8827, |
|
"step": 4490 |
|
}, |
|
{ |
|
"epoch": 10.844390832328106, |
|
"grad_norm": 0.32421875, |
|
"learning_rate": 4.2976656290299764e-05, |
|
"loss": 0.8823, |
|
"step": 4495 |
|
}, |
|
{ |
|
"epoch": 10.856453558504223, |
|
"grad_norm": 0.326171875, |
|
"learning_rate": 4.274600293677647e-05, |
|
"loss": 0.8856, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 10.868516284680338, |
|
"grad_norm": 0.326171875, |
|
"learning_rate": 4.2515801831675305e-05, |
|
"loss": 0.8786, |
|
"step": 4505 |
|
}, |
|
{ |
|
"epoch": 10.880579010856454, |
|
"grad_norm": 0.32421875, |
|
"learning_rate": 4.228605479335127e-05, |
|
"loss": 0.8808, |
|
"step": 4510 |
|
}, |
|
{ |
|
"epoch": 10.89264173703257, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 4.2056763636572574e-05, |
|
"loss": 0.8746, |
|
"step": 4515 |
|
}, |
|
{ |
|
"epoch": 10.904704463208684, |
|
"grad_norm": 0.330078125, |
|
"learning_rate": 4.182793017250643e-05, |
|
"loss": 0.8731, |
|
"step": 4520 |
|
}, |
|
{ |
|
"epoch": 10.916767189384801, |
|
"grad_norm": 0.32421875, |
|
"learning_rate": 4.1599556208704785e-05, |
|
"loss": 0.8791, |
|
"step": 4525 |
|
}, |
|
{ |
|
"epoch": 10.928829915560916, |
|
"grad_norm": 0.310546875, |
|
"learning_rate": 4.137164354908999e-05, |
|
"loss": 0.8785, |
|
"step": 4530 |
|
}, |
|
{ |
|
"epoch": 10.940892641737033, |
|
"grad_norm": 0.310546875, |
|
"learning_rate": 4.114419399394052e-05, |
|
"loss": 0.88, |
|
"step": 4535 |
|
}, |
|
{ |
|
"epoch": 10.952955367913148, |
|
"grad_norm": 0.32421875, |
|
"learning_rate": 4.0917209339876896e-05, |
|
"loss": 0.8801, |
|
"step": 4540 |
|
}, |
|
{ |
|
"epoch": 10.965018094089265, |
|
"grad_norm": 0.3359375, |
|
"learning_rate": 4.069069137984731e-05, |
|
"loss": 0.8809, |
|
"step": 4545 |
|
}, |
|
{ |
|
"epoch": 10.97708082026538, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 4.046464190311355e-05, |
|
"loss": 0.8911, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 10.989143546441495, |
|
"grad_norm": 0.333984375, |
|
"learning_rate": 4.023906269523686e-05, |
|
"loss": 0.8697, |
|
"step": 4555 |
|
}, |
|
{ |
|
"epoch": 10.998793727382388, |
|
"eval_loss": 2.5013487339019775, |
|
"eval_runtime": 0.6548, |
|
"eval_samples_per_second": 15.273, |
|
"eval_steps_per_second": 1.527, |
|
"step": 4559 |
|
}, |
|
{ |
|
"epoch": 11.001206272617612, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 4.001395553806391e-05, |
|
"loss": 0.8789, |
|
"step": 4560 |
|
}, |
|
{ |
|
"epoch": 11.013268998793727, |
|
"grad_norm": 0.33203125, |
|
"learning_rate": 3.9789322209712566e-05, |
|
"loss": 0.8779, |
|
"step": 4565 |
|
}, |
|
{ |
|
"epoch": 11.025331724969844, |
|
"grad_norm": 0.34375, |
|
"learning_rate": 3.956516448455794e-05, |
|
"loss": 0.8665, |
|
"step": 4570 |
|
}, |
|
{ |
|
"epoch": 11.037394451145959, |
|
"grad_norm": 0.33203125, |
|
"learning_rate": 3.9341484133218366e-05, |
|
"loss": 0.8736, |
|
"step": 4575 |
|
}, |
|
{ |
|
"epoch": 11.049457177322076, |
|
"grad_norm": 0.3359375, |
|
"learning_rate": 3.9118282922541474e-05, |
|
"loss": 0.8686, |
|
"step": 4580 |
|
}, |
|
{ |
|
"epoch": 11.06151990349819, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 3.8895562615590075e-05, |
|
"loss": 0.8758, |
|
"step": 4585 |
|
}, |
|
{ |
|
"epoch": 11.073582629674306, |
|
"grad_norm": 0.302734375, |
|
"learning_rate": 3.8673324971628357e-05, |
|
"loss": 0.8722, |
|
"step": 4590 |
|
}, |
|
{ |
|
"epoch": 11.085645355850422, |
|
"grad_norm": 0.3046875, |
|
"learning_rate": 3.8451571746108027e-05, |
|
"loss": 0.8769, |
|
"step": 4595 |
|
}, |
|
{ |
|
"epoch": 11.097708082026537, |
|
"grad_norm": 0.326171875, |
|
"learning_rate": 3.8230304690654304e-05, |
|
"loss": 0.8696, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 11.109770808202654, |
|
"grad_norm": 0.328125, |
|
"learning_rate": 3.800952555305216e-05, |
|
"loss": 0.8767, |
|
"step": 4605 |
|
}, |
|
{ |
|
"epoch": 11.12183353437877, |
|
"grad_norm": 0.337890625, |
|
"learning_rate": 3.778923607723252e-05, |
|
"loss": 0.8717, |
|
"step": 4610 |
|
}, |
|
{ |
|
"epoch": 11.133896260554886, |
|
"grad_norm": 0.341796875, |
|
"learning_rate": 3.756943800325852e-05, |
|
"loss": 0.8716, |
|
"step": 4615 |
|
}, |
|
{ |
|
"epoch": 11.145958986731001, |
|
"grad_norm": 0.345703125, |
|
"learning_rate": 3.7350133067311686e-05, |
|
"loss": 0.8779, |
|
"step": 4620 |
|
}, |
|
{ |
|
"epoch": 11.158021712907116, |
|
"grad_norm": 0.34375, |
|
"learning_rate": 3.7131323001678254e-05, |
|
"loss": 0.8674, |
|
"step": 4625 |
|
}, |
|
{ |
|
"epoch": 11.170084439083233, |
|
"grad_norm": 0.326171875, |
|
"learning_rate": 3.691300953473548e-05, |
|
"loss": 0.8784, |
|
"step": 4630 |
|
}, |
|
{ |
|
"epoch": 11.182147165259348, |
|
"grad_norm": 0.328125, |
|
"learning_rate": 3.669519439093801e-05, |
|
"loss": 0.8854, |
|
"step": 4635 |
|
}, |
|
{ |
|
"epoch": 11.194209891435465, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 3.6477879290804206e-05, |
|
"loss": 0.8706, |
|
"step": 4640 |
|
}, |
|
{ |
|
"epoch": 11.20627261761158, |
|
"grad_norm": 0.326171875, |
|
"learning_rate": 3.626106595090268e-05, |
|
"loss": 0.8776, |
|
"step": 4645 |
|
}, |
|
{ |
|
"epoch": 11.218335343787697, |
|
"grad_norm": 0.357421875, |
|
"learning_rate": 3.604475608383858e-05, |
|
"loss": 0.8725, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 11.230398069963812, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 3.582895139824014e-05, |
|
"loss": 0.8751, |
|
"step": 4655 |
|
}, |
|
{ |
|
"epoch": 11.242460796139927, |
|
"grad_norm": 0.33203125, |
|
"learning_rate": 3.561365359874513e-05, |
|
"loss": 0.8789, |
|
"step": 4660 |
|
}, |
|
{ |
|
"epoch": 11.254523522316044, |
|
"grad_norm": 0.310546875, |
|
"learning_rate": 3.539886438598756e-05, |
|
"loss": 0.8773, |
|
"step": 4665 |
|
}, |
|
{ |
|
"epoch": 11.266586248492159, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 3.518458545658401e-05, |
|
"loss": 0.877, |
|
"step": 4670 |
|
}, |
|
{ |
|
"epoch": 11.278648974668275, |
|
"grad_norm": 0.310546875, |
|
"learning_rate": 3.497081850312033e-05, |
|
"loss": 0.8754, |
|
"step": 4675 |
|
}, |
|
{ |
|
"epoch": 11.29071170084439, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 3.475756521413839e-05, |
|
"loss": 0.8763, |
|
"step": 4680 |
|
}, |
|
{ |
|
"epoch": 11.302774427020507, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 3.454482727412254e-05, |
|
"loss": 0.8778, |
|
"step": 4685 |
|
}, |
|
{ |
|
"epoch": 11.314837153196622, |
|
"grad_norm": 0.330078125, |
|
"learning_rate": 3.433260636348639e-05, |
|
"loss": 0.8784, |
|
"step": 4690 |
|
}, |
|
{ |
|
"epoch": 11.32689987937274, |
|
"grad_norm": 0.3125, |
|
"learning_rate": 3.412090415855963e-05, |
|
"loss": 0.8745, |
|
"step": 4695 |
|
}, |
|
{ |
|
"epoch": 11.338962605548854, |
|
"grad_norm": 0.32421875, |
|
"learning_rate": 3.3909722331574623e-05, |
|
"loss": 0.8756, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 11.35102533172497, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 3.369906255065328e-05, |
|
"loss": 0.8827, |
|
"step": 4705 |
|
}, |
|
{ |
|
"epoch": 11.363088057901086, |
|
"grad_norm": 0.328125, |
|
"learning_rate": 3.348892647979389e-05, |
|
"loss": 0.8763, |
|
"step": 4710 |
|
}, |
|
{ |
|
"epoch": 11.375150784077201, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 3.3279315778858036e-05, |
|
"loss": 0.8815, |
|
"step": 4715 |
|
}, |
|
{ |
|
"epoch": 11.387213510253318, |
|
"grad_norm": 0.306640625, |
|
"learning_rate": 3.307023210355729e-05, |
|
"loss": 0.8787, |
|
"step": 4720 |
|
}, |
|
{ |
|
"epoch": 11.399276236429433, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 3.2861677105440336e-05, |
|
"loss": 0.8794, |
|
"step": 4725 |
|
}, |
|
{ |
|
"epoch": 11.41133896260555, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 3.265365243187981e-05, |
|
"loss": 0.8753, |
|
"step": 4730 |
|
}, |
|
{ |
|
"epoch": 11.423401688781665, |
|
"grad_norm": 0.349609375, |
|
"learning_rate": 3.2446159726059345e-05, |
|
"loss": 0.8819, |
|
"step": 4735 |
|
}, |
|
{ |
|
"epoch": 11.43546441495778, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 3.223920062696052e-05, |
|
"loss": 0.8727, |
|
"step": 4740 |
|
}, |
|
{ |
|
"epoch": 11.447527141133897, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 3.203277676935007e-05, |
|
"loss": 0.8743, |
|
"step": 4745 |
|
}, |
|
{ |
|
"epoch": 11.459589867310012, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 3.182688978376678e-05, |
|
"loss": 0.8777, |
|
"step": 4750 |
|
}, |
|
{ |
|
"epoch": 11.471652593486128, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 3.1621541296508695e-05, |
|
"loss": 0.8759, |
|
"step": 4755 |
|
}, |
|
{ |
|
"epoch": 11.483715319662243, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 3.141673292962026e-05, |
|
"loss": 0.8693, |
|
"step": 4760 |
|
}, |
|
{ |
|
"epoch": 11.49577804583836, |
|
"grad_norm": 0.337890625, |
|
"learning_rate": 3.121246630087961e-05, |
|
"loss": 0.8747, |
|
"step": 4765 |
|
}, |
|
{ |
|
"epoch": 11.507840772014475, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 3.100874302378559e-05, |
|
"loss": 0.8789, |
|
"step": 4770 |
|
}, |
|
{ |
|
"epoch": 11.51990349819059, |
|
"grad_norm": 0.3203125, |
|
"learning_rate": 3.080556470754513e-05, |
|
"loss": 0.8858, |
|
"step": 4775 |
|
}, |
|
{ |
|
"epoch": 11.531966224366707, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 3.0602932957060635e-05, |
|
"loss": 0.8751, |
|
"step": 4780 |
|
}, |
|
{ |
|
"epoch": 11.544028950542822, |
|
"grad_norm": 0.337890625, |
|
"learning_rate": 3.0400849372917073e-05, |
|
"loss": 0.8828, |
|
"step": 4785 |
|
}, |
|
{ |
|
"epoch": 11.556091676718939, |
|
"grad_norm": 0.302734375, |
|
"learning_rate": 3.019931555136949e-05, |
|
"loss": 0.8751, |
|
"step": 4790 |
|
}, |
|
{ |
|
"epoch": 11.568154402895054, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 2.9998333084330333e-05, |
|
"loss": 0.8835, |
|
"step": 4795 |
|
}, |
|
{ |
|
"epoch": 11.58021712907117, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 2.979790355935703e-05, |
|
"loss": 0.8705, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 11.592279855247286, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 2.9598028559639212e-05, |
|
"loss": 0.8677, |
|
"step": 4805 |
|
}, |
|
{ |
|
"epoch": 11.604342581423401, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 2.9398709663986324e-05, |
|
"loss": 0.8772, |
|
"step": 4810 |
|
}, |
|
{ |
|
"epoch": 11.616405307599518, |
|
"grad_norm": 0.30859375, |
|
"learning_rate": 2.919994844681524e-05, |
|
"loss": 0.8788, |
|
"step": 4815 |
|
}, |
|
{ |
|
"epoch": 11.628468033775633, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 2.900174647813767e-05, |
|
"loss": 0.8767, |
|
"step": 4820 |
|
}, |
|
{ |
|
"epoch": 11.64053075995175, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 2.8804105323547837e-05, |
|
"loss": 0.8813, |
|
"step": 4825 |
|
}, |
|
{ |
|
"epoch": 11.652593486127865, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 2.8607026544210114e-05, |
|
"loss": 0.8797, |
|
"step": 4830 |
|
}, |
|
{ |
|
"epoch": 11.664656212303981, |
|
"grad_norm": 0.330078125, |
|
"learning_rate": 2.8410511696846676e-05, |
|
"loss": 0.8802, |
|
"step": 4835 |
|
}, |
|
{ |
|
"epoch": 11.676718938480096, |
|
"grad_norm": 0.328125, |
|
"learning_rate": 2.8214562333725227e-05, |
|
"loss": 0.8825, |
|
"step": 4840 |
|
}, |
|
{ |
|
"epoch": 11.688781664656211, |
|
"grad_norm": 0.328125, |
|
"learning_rate": 2.801918000264665e-05, |
|
"loss": 0.8741, |
|
"step": 4845 |
|
}, |
|
{ |
|
"epoch": 11.700844390832328, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 2.7824366246932985e-05, |
|
"loss": 0.8843, |
|
"step": 4850 |
|
}, |
|
{ |
|
"epoch": 11.712907117008443, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 2.7630122605414988e-05, |
|
"loss": 0.8783, |
|
"step": 4855 |
|
}, |
|
{ |
|
"epoch": 11.72496984318456, |
|
"grad_norm": 0.32421875, |
|
"learning_rate": 2.7436450612420095e-05, |
|
"loss": 0.8717, |
|
"step": 4860 |
|
}, |
|
{ |
|
"epoch": 11.737032569360675, |
|
"grad_norm": 0.33203125, |
|
"learning_rate": 2.7243351797760397e-05, |
|
"loss": 0.8807, |
|
"step": 4865 |
|
}, |
|
{ |
|
"epoch": 11.749095295536792, |
|
"grad_norm": 0.3125, |
|
"learning_rate": 2.705082768672036e-05, |
|
"loss": 0.8715, |
|
"step": 4870 |
|
}, |
|
{ |
|
"epoch": 11.761158021712907, |
|
"grad_norm": 0.349609375, |
|
"learning_rate": 2.6858879800044866e-05, |
|
"loss": 0.8765, |
|
"step": 4875 |
|
}, |
|
{ |
|
"epoch": 11.773220747889024, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 2.6667509653927334e-05, |
|
"loss": 0.8744, |
|
"step": 4880 |
|
}, |
|
{ |
|
"epoch": 11.785283474065139, |
|
"grad_norm": 0.3125, |
|
"learning_rate": 2.64767187599975e-05, |
|
"loss": 0.8793, |
|
"step": 4885 |
|
}, |
|
{ |
|
"epoch": 11.797346200241254, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 2.6286508625309624e-05, |
|
"loss": 0.8762, |
|
"step": 4890 |
|
}, |
|
{ |
|
"epoch": 11.80940892641737, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 2.6096880752330553e-05, |
|
"loss": 0.8839, |
|
"step": 4895 |
|
}, |
|
{ |
|
"epoch": 11.821471652593486, |
|
"grad_norm": 0.33984375, |
|
"learning_rate": 2.5907836638927928e-05, |
|
"loss": 0.8828, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 11.833534378769603, |
|
"grad_norm": 0.306640625, |
|
"learning_rate": 2.57193777783582e-05, |
|
"loss": 0.8782, |
|
"step": 4905 |
|
}, |
|
{ |
|
"epoch": 11.845597104945718, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 2.553150565925493e-05, |
|
"loss": 0.8708, |
|
"step": 4910 |
|
}, |
|
{ |
|
"epoch": 11.857659831121833, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 2.534422176561705e-05, |
|
"loss": 0.879, |
|
"step": 4915 |
|
}, |
|
{ |
|
"epoch": 11.86972255729795, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 2.515752757679707e-05, |
|
"loss": 0.8808, |
|
"step": 4920 |
|
}, |
|
{ |
|
"epoch": 11.881785283474064, |
|
"grad_norm": 0.353515625, |
|
"learning_rate": 2.4971424567489434e-05, |
|
"loss": 0.8777, |
|
"step": 4925 |
|
}, |
|
{ |
|
"epoch": 11.893848009650181, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 2.478591420771894e-05, |
|
"loss": 0.875, |
|
"step": 4930 |
|
}, |
|
{ |
|
"epoch": 11.905910735826296, |
|
"grad_norm": 0.30859375, |
|
"learning_rate": 2.4600997962828987e-05, |
|
"loss": 0.8767, |
|
"step": 4935 |
|
}, |
|
{ |
|
"epoch": 11.917973462002413, |
|
"grad_norm": 0.357421875, |
|
"learning_rate": 2.441667729347006e-05, |
|
"loss": 0.8761, |
|
"step": 4940 |
|
}, |
|
{ |
|
"epoch": 11.930036188178528, |
|
"grad_norm": 0.33984375, |
|
"learning_rate": 2.423295365558821e-05, |
|
"loss": 0.8821, |
|
"step": 4945 |
|
}, |
|
{ |
|
"epoch": 11.942098914354645, |
|
"grad_norm": 0.3125, |
|
"learning_rate": 2.404982850041363e-05, |
|
"loss": 0.8829, |
|
"step": 4950 |
|
}, |
|
{ |
|
"epoch": 11.95416164053076, |
|
"grad_norm": 0.3203125, |
|
"learning_rate": 2.3867303274448994e-05, |
|
"loss": 0.8794, |
|
"step": 4955 |
|
}, |
|
{ |
|
"epoch": 11.966224366706875, |
|
"grad_norm": 0.3203125, |
|
"learning_rate": 2.3685379419458164e-05, |
|
"loss": 0.882, |
|
"step": 4960 |
|
}, |
|
{ |
|
"epoch": 11.978287092882992, |
|
"grad_norm": 0.3046875, |
|
"learning_rate": 2.3504058372454884e-05, |
|
"loss": 0.8693, |
|
"step": 4965 |
|
}, |
|
{ |
|
"epoch": 11.990349819059107, |
|
"grad_norm": 0.337890625, |
|
"learning_rate": 2.3323341565691204e-05, |
|
"loss": 0.8729, |
|
"step": 4970 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"eval_loss": 2.5031228065490723, |
|
"eval_runtime": 0.5349, |
|
"eval_samples_per_second": 18.696, |
|
"eval_steps_per_second": 1.87, |
|
"step": 4974 |
|
}, |
|
{ |
|
"epoch": 12.002412545235224, |
|
"grad_norm": 0.310546875, |
|
"learning_rate": 2.3143230426646312e-05, |
|
"loss": 0.873, |
|
"step": 4975 |
|
}, |
|
{ |
|
"epoch": 12.014475271411339, |
|
"grad_norm": 0.30859375, |
|
"learning_rate": 2.2963726378015327e-05, |
|
"loss": 0.8753, |
|
"step": 4980 |
|
}, |
|
{ |
|
"epoch": 12.026537997587456, |
|
"grad_norm": 0.32421875, |
|
"learning_rate": 2.278483083769787e-05, |
|
"loss": 0.8806, |
|
"step": 4985 |
|
}, |
|
{ |
|
"epoch": 12.03860072376357, |
|
"grad_norm": 0.298828125, |
|
"learning_rate": 2.2606545218787012e-05, |
|
"loss": 0.8825, |
|
"step": 4990 |
|
}, |
|
{ |
|
"epoch": 12.050663449939686, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 2.242887092955801e-05, |
|
"loss": 0.8671, |
|
"step": 4995 |
|
}, |
|
{ |
|
"epoch": 12.062726176115802, |
|
"grad_norm": 0.298828125, |
|
"learning_rate": 2.2251809373457345e-05, |
|
"loss": 0.8707, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 12.074788902291917, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 2.207536194909142e-05, |
|
"loss": 0.8738, |
|
"step": 5005 |
|
}, |
|
{ |
|
"epoch": 12.086851628468034, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 2.189953005021569e-05, |
|
"loss": 0.8725, |
|
"step": 5010 |
|
}, |
|
{ |
|
"epoch": 12.09891435464415, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 2.1724315065723532e-05, |
|
"loss": 0.8718, |
|
"step": 5015 |
|
}, |
|
{ |
|
"epoch": 12.110977080820266, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 2.1549718379635377e-05, |
|
"loss": 0.8764, |
|
"step": 5020 |
|
}, |
|
{ |
|
"epoch": 12.123039806996381, |
|
"grad_norm": 0.32421875, |
|
"learning_rate": 2.1375741371087677e-05, |
|
"loss": 0.8717, |
|
"step": 5025 |
|
}, |
|
{ |
|
"epoch": 12.135102533172496, |
|
"grad_norm": 0.30859375, |
|
"learning_rate": 2.1202385414322078e-05, |
|
"loss": 0.8799, |
|
"step": 5030 |
|
}, |
|
{ |
|
"epoch": 12.147165259348613, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 2.1029651878674604e-05, |
|
"loss": 0.8683, |
|
"step": 5035 |
|
}, |
|
{ |
|
"epoch": 12.159227985524728, |
|
"grad_norm": 0.30859375, |
|
"learning_rate": 2.085754212856471e-05, |
|
"loss": 0.8839, |
|
"step": 5040 |
|
}, |
|
{ |
|
"epoch": 12.171290711700845, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 2.068605752348457e-05, |
|
"loss": 0.8786, |
|
"step": 5045 |
|
}, |
|
{ |
|
"epoch": 12.18335343787696, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 2.0515199417988452e-05, |
|
"loss": 0.8729, |
|
"step": 5050 |
|
}, |
|
{ |
|
"epoch": 12.195416164053077, |
|
"grad_norm": 0.3203125, |
|
"learning_rate": 2.0344969161681792e-05, |
|
"loss": 0.8784, |
|
"step": 5055 |
|
}, |
|
{ |
|
"epoch": 12.207478890229192, |
|
"grad_norm": 0.310546875, |
|
"learning_rate": 2.01753680992107e-05, |
|
"loss": 0.8831, |
|
"step": 5060 |
|
}, |
|
{ |
|
"epoch": 12.219541616405307, |
|
"grad_norm": 0.3203125, |
|
"learning_rate": 2.0006397570251356e-05, |
|
"loss": 0.872, |
|
"step": 5065 |
|
}, |
|
{ |
|
"epoch": 12.231604342581424, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 1.983805890949927e-05, |
|
"loss": 0.8756, |
|
"step": 5070 |
|
}, |
|
{ |
|
"epoch": 12.243667068757539, |
|
"grad_norm": 0.32421875, |
|
"learning_rate": 1.9670353446658873e-05, |
|
"loss": 0.8795, |
|
"step": 5075 |
|
}, |
|
{ |
|
"epoch": 12.255729794933655, |
|
"grad_norm": 0.30078125, |
|
"learning_rate": 1.9503282506432974e-05, |
|
"loss": 0.8744, |
|
"step": 5080 |
|
}, |
|
{ |
|
"epoch": 12.26779252110977, |
|
"grad_norm": 0.306640625, |
|
"learning_rate": 1.9336847408512328e-05, |
|
"loss": 0.8728, |
|
"step": 5085 |
|
}, |
|
{ |
|
"epoch": 12.279855247285887, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 1.917104946756515e-05, |
|
"loss": 0.867, |
|
"step": 5090 |
|
}, |
|
{ |
|
"epoch": 12.291917973462002, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 1.9005889993226735e-05, |
|
"loss": 0.8756, |
|
"step": 5095 |
|
}, |
|
{ |
|
"epoch": 12.303980699638117, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 1.884137029008921e-05, |
|
"loss": 0.8761, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 12.316043425814234, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 1.8677491657691116e-05, |
|
"loss": 0.8738, |
|
"step": 5105 |
|
}, |
|
{ |
|
"epoch": 12.32810615199035, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 1.8514255390507183e-05, |
|
"loss": 0.8793, |
|
"step": 5110 |
|
}, |
|
{ |
|
"epoch": 12.340168878166466, |
|
"grad_norm": 0.30859375, |
|
"learning_rate": 1.8351662777938127e-05, |
|
"loss": 0.8741, |
|
"step": 5115 |
|
}, |
|
{ |
|
"epoch": 12.352231604342581, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 1.818971510430042e-05, |
|
"loss": 0.876, |
|
"step": 5120 |
|
}, |
|
{ |
|
"epoch": 12.364294330518698, |
|
"grad_norm": 0.330078125, |
|
"learning_rate": 1.802841364881621e-05, |
|
"loss": 0.87, |
|
"step": 5125 |
|
}, |
|
{ |
|
"epoch": 12.376357056694813, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 1.7867759685603114e-05, |
|
"loss": 0.8818, |
|
"step": 5130 |
|
}, |
|
{ |
|
"epoch": 12.38841978287093, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 1.7707754483664308e-05, |
|
"loss": 0.8782, |
|
"step": 5135 |
|
}, |
|
{ |
|
"epoch": 12.400482509047045, |
|
"grad_norm": 0.3125, |
|
"learning_rate": 1.754839930687836e-05, |
|
"loss": 0.8743, |
|
"step": 5140 |
|
}, |
|
{ |
|
"epoch": 12.41254523522316, |
|
"grad_norm": 0.30859375, |
|
"learning_rate": 1.738969541398926e-05, |
|
"loss": 0.8783, |
|
"step": 5145 |
|
}, |
|
{ |
|
"epoch": 12.424607961399277, |
|
"grad_norm": 0.3203125, |
|
"learning_rate": 1.723164405859663e-05, |
|
"loss": 0.872, |
|
"step": 5150 |
|
}, |
|
{ |
|
"epoch": 12.436670687575392, |
|
"grad_norm": 0.341796875, |
|
"learning_rate": 1.7074246489145595e-05, |
|
"loss": 0.8739, |
|
"step": 5155 |
|
}, |
|
{ |
|
"epoch": 12.448733413751508, |
|
"grad_norm": 0.3125, |
|
"learning_rate": 1.691750394891707e-05, |
|
"loss": 0.8779, |
|
"step": 5160 |
|
}, |
|
{ |
|
"epoch": 12.460796139927623, |
|
"grad_norm": 0.310546875, |
|
"learning_rate": 1.6761417676017956e-05, |
|
"loss": 0.8758, |
|
"step": 5165 |
|
}, |
|
{ |
|
"epoch": 12.47285886610374, |
|
"grad_norm": 0.30859375, |
|
"learning_rate": 1.660598890337124e-05, |
|
"loss": 0.881, |
|
"step": 5170 |
|
}, |
|
{ |
|
"epoch": 12.484921592279855, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 1.6451218858706374e-05, |
|
"loss": 0.8806, |
|
"step": 5175 |
|
}, |
|
{ |
|
"epoch": 12.49698431845597, |
|
"grad_norm": 0.3125, |
|
"learning_rate": 1.629710876454945e-05, |
|
"loss": 0.8771, |
|
"step": 5180 |
|
}, |
|
{ |
|
"epoch": 12.509047044632087, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 1.6143659838213752e-05, |
|
"loss": 0.8784, |
|
"step": 5185 |
|
}, |
|
{ |
|
"epoch": 12.521109770808202, |
|
"grad_norm": 0.3046875, |
|
"learning_rate": 1.59908732917899e-05, |
|
"loss": 0.863, |
|
"step": 5190 |
|
}, |
|
{ |
|
"epoch": 12.533172496984319, |
|
"grad_norm": 0.310546875, |
|
"learning_rate": 1.583875033213642e-05, |
|
"loss": 0.87, |
|
"step": 5195 |
|
}, |
|
{ |
|
"epoch": 12.545235223160434, |
|
"grad_norm": 0.3203125, |
|
"learning_rate": 1.568729216087017e-05, |
|
"loss": 0.8741, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 12.55729794933655, |
|
"grad_norm": 0.33203125, |
|
"learning_rate": 1.5536499974356866e-05, |
|
"loss": 0.8759, |
|
"step": 5205 |
|
}, |
|
{ |
|
"epoch": 12.569360675512666, |
|
"grad_norm": 0.310546875, |
|
"learning_rate": 1.5386374963701633e-05, |
|
"loss": 0.8841, |
|
"step": 5210 |
|
}, |
|
{ |
|
"epoch": 12.581423401688781, |
|
"grad_norm": 0.3203125, |
|
"learning_rate": 1.5236918314739568e-05, |
|
"loss": 0.8843, |
|
"step": 5215 |
|
}, |
|
{ |
|
"epoch": 12.593486127864898, |
|
"grad_norm": 0.306640625, |
|
"learning_rate": 1.5088131208026367e-05, |
|
"loss": 0.8694, |
|
"step": 5220 |
|
}, |
|
{ |
|
"epoch": 12.605548854041013, |
|
"grad_norm": 0.3046875, |
|
"learning_rate": 1.4940014818829074e-05, |
|
"loss": 0.8738, |
|
"step": 5225 |
|
}, |
|
{ |
|
"epoch": 12.61761158021713, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 1.4792570317116671e-05, |
|
"loss": 0.8694, |
|
"step": 5230 |
|
}, |
|
{ |
|
"epoch": 12.629674306393245, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 1.4645798867551008e-05, |
|
"loss": 0.8774, |
|
"step": 5235 |
|
}, |
|
{ |
|
"epoch": 12.641737032569361, |
|
"grad_norm": 0.306640625, |
|
"learning_rate": 1.4499701629477446e-05, |
|
"loss": 0.8805, |
|
"step": 5240 |
|
}, |
|
{ |
|
"epoch": 12.653799758745476, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 1.4354279756915735e-05, |
|
"loss": 0.8815, |
|
"step": 5245 |
|
}, |
|
{ |
|
"epoch": 12.665862484921592, |
|
"grad_norm": 0.310546875, |
|
"learning_rate": 1.4209534398551016e-05, |
|
"loss": 0.875, |
|
"step": 5250 |
|
}, |
|
{ |
|
"epoch": 12.677925211097708, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 1.406546669772456e-05, |
|
"loss": 0.8757, |
|
"step": 5255 |
|
}, |
|
{ |
|
"epoch": 12.689987937273823, |
|
"grad_norm": 0.333984375, |
|
"learning_rate": 1.3922077792424881e-05, |
|
"loss": 0.8773, |
|
"step": 5260 |
|
}, |
|
{ |
|
"epoch": 12.70205066344994, |
|
"grad_norm": 0.32421875, |
|
"learning_rate": 1.3779368815278647e-05, |
|
"loss": 0.873, |
|
"step": 5265 |
|
}, |
|
{ |
|
"epoch": 12.714113389626055, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 1.3637340893541895e-05, |
|
"loss": 0.8811, |
|
"step": 5270 |
|
}, |
|
{ |
|
"epoch": 12.726176115802172, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 1.3495995149090911e-05, |
|
"loss": 0.8726, |
|
"step": 5275 |
|
}, |
|
{ |
|
"epoch": 12.738238841978287, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 1.335533269841347e-05, |
|
"loss": 0.8703, |
|
"step": 5280 |
|
}, |
|
{ |
|
"epoch": 12.750301568154402, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 1.321535465260012e-05, |
|
"loss": 0.8764, |
|
"step": 5285 |
|
}, |
|
{ |
|
"epoch": 12.762364294330519, |
|
"grad_norm": 0.33984375, |
|
"learning_rate": 1.307606211733522e-05, |
|
"loss": 0.8762, |
|
"step": 5290 |
|
}, |
|
{ |
|
"epoch": 12.774427020506634, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 1.2937456192888309e-05, |
|
"loss": 0.8707, |
|
"step": 5295 |
|
}, |
|
{ |
|
"epoch": 12.78648974668275, |
|
"grad_norm": 0.337890625, |
|
"learning_rate": 1.2799537974105402e-05, |
|
"loss": 0.8822, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 12.798552472858866, |
|
"grad_norm": 0.306640625, |
|
"learning_rate": 1.2662308550400336e-05, |
|
"loss": 0.8761, |
|
"step": 5305 |
|
}, |
|
{ |
|
"epoch": 12.810615199034983, |
|
"grad_norm": 0.365234375, |
|
"learning_rate": 1.252576900574618e-05, |
|
"loss": 0.8743, |
|
"step": 5310 |
|
}, |
|
{ |
|
"epoch": 12.822677925211098, |
|
"grad_norm": 0.3046875, |
|
"learning_rate": 1.2389920418666633e-05, |
|
"loss": 0.8689, |
|
"step": 5315 |
|
}, |
|
{ |
|
"epoch": 12.834740651387214, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 1.225476386222757e-05, |
|
"loss": 0.8698, |
|
"step": 5320 |
|
}, |
|
{ |
|
"epoch": 12.84680337756333, |
|
"grad_norm": 0.328125, |
|
"learning_rate": 1.2120300404028507e-05, |
|
"loss": 0.8802, |
|
"step": 5325 |
|
}, |
|
{ |
|
"epoch": 12.858866103739444, |
|
"grad_norm": 0.3046875, |
|
"learning_rate": 1.1986531106194143e-05, |
|
"loss": 0.8739, |
|
"step": 5330 |
|
}, |
|
{ |
|
"epoch": 12.870928829915561, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 1.1853457025366111e-05, |
|
"loss": 0.8814, |
|
"step": 5335 |
|
}, |
|
{ |
|
"epoch": 12.882991556091676, |
|
"grad_norm": 0.333984375, |
|
"learning_rate": 1.1721079212694452e-05, |
|
"loss": 0.8802, |
|
"step": 5340 |
|
}, |
|
{ |
|
"epoch": 12.895054282267793, |
|
"grad_norm": 0.33984375, |
|
"learning_rate": 1.1589398713829424e-05, |
|
"loss": 0.8746, |
|
"step": 5345 |
|
}, |
|
{ |
|
"epoch": 12.907117008443908, |
|
"grad_norm": 0.3046875, |
|
"learning_rate": 1.1458416568913232e-05, |
|
"loss": 0.8724, |
|
"step": 5350 |
|
}, |
|
{ |
|
"epoch": 12.919179734620023, |
|
"grad_norm": 0.3203125, |
|
"learning_rate": 1.1328133812571784e-05, |
|
"loss": 0.8836, |
|
"step": 5355 |
|
}, |
|
{ |
|
"epoch": 12.93124246079614, |
|
"grad_norm": 0.330078125, |
|
"learning_rate": 1.1198551473906493e-05, |
|
"loss": 0.8663, |
|
"step": 5360 |
|
}, |
|
{ |
|
"epoch": 12.943305186972255, |
|
"grad_norm": 0.3203125, |
|
"learning_rate": 1.1069670576486224e-05, |
|
"loss": 0.8722, |
|
"step": 5365 |
|
}, |
|
{ |
|
"epoch": 12.955367913148372, |
|
"grad_norm": 0.32421875, |
|
"learning_rate": 1.0941492138339183e-05, |
|
"loss": 0.8718, |
|
"step": 5370 |
|
}, |
|
{ |
|
"epoch": 12.967430639324487, |
|
"grad_norm": 0.3125, |
|
"learning_rate": 1.0814017171944812e-05, |
|
"loss": 0.875, |
|
"step": 5375 |
|
}, |
|
{ |
|
"epoch": 12.979493365500604, |
|
"grad_norm": 0.33984375, |
|
"learning_rate": 1.068724668422586e-05, |
|
"loss": 0.8768, |
|
"step": 5380 |
|
}, |
|
{ |
|
"epoch": 12.991556091676719, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 1.0561181676540444e-05, |
|
"loss": 0.8779, |
|
"step": 5385 |
|
}, |
|
{ |
|
"epoch": 12.998793727382388, |
|
"eval_loss": 2.5023179054260254, |
|
"eval_runtime": 0.6415, |
|
"eval_samples_per_second": 15.589, |
|
"eval_steps_per_second": 1.559, |
|
"step": 5388 |
|
}, |
|
{ |
|
"epoch": 13.003618817852836, |
|
"grad_norm": 0.310546875, |
|
"learning_rate": 1.0435823144674051e-05, |
|
"loss": 0.8808, |
|
"step": 5390 |
|
}, |
|
{ |
|
"epoch": 13.01568154402895, |
|
"grad_norm": 0.30859375, |
|
"learning_rate": 1.0311172078831766e-05, |
|
"loss": 0.882, |
|
"step": 5395 |
|
}, |
|
{ |
|
"epoch": 13.027744270205066, |
|
"grad_norm": 0.3125, |
|
"learning_rate": 1.01872294636304e-05, |
|
"loss": 0.8872, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 13.039806996381182, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 1.0063996278090704e-05, |
|
"loss": 0.8692, |
|
"step": 5405 |
|
}, |
|
{ |
|
"epoch": 13.051869722557297, |
|
"grad_norm": 0.333984375, |
|
"learning_rate": 9.941473495629683e-06, |
|
"loss": 0.8738, |
|
"step": 5410 |
|
}, |
|
{ |
|
"epoch": 13.063932448733414, |
|
"grad_norm": 0.30859375, |
|
"learning_rate": 9.81966208405285e-06, |
|
"loss": 0.8858, |
|
"step": 5415 |
|
}, |
|
{ |
|
"epoch": 13.07599517490953, |
|
"grad_norm": 0.310546875, |
|
"learning_rate": 9.698563005546668e-06, |
|
"loss": 0.878, |
|
"step": 5420 |
|
}, |
|
{ |
|
"epoch": 13.088057901085646, |
|
"grad_norm": 0.306640625, |
|
"learning_rate": 9.578177216670824e-06, |
|
"loss": 0.8625, |
|
"step": 5425 |
|
}, |
|
{ |
|
"epoch": 13.100120627261761, |
|
"grad_norm": 0.32421875, |
|
"learning_rate": 9.458505668350759e-06, |
|
"loss": 0.8794, |
|
"step": 5430 |
|
}, |
|
{ |
|
"epoch": 13.112183353437876, |
|
"grad_norm": 0.3125, |
|
"learning_rate": 9.339549305870187e-06, |
|
"loss": 0.8684, |
|
"step": 5435 |
|
}, |
|
{ |
|
"epoch": 13.124246079613993, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 9.221309068863527e-06, |
|
"loss": 0.874, |
|
"step": 5440 |
|
}, |
|
{ |
|
"epoch": 13.136308805790108, |
|
"grad_norm": 0.3203125, |
|
"learning_rate": 9.103785891308547e-06, |
|
"loss": 0.872, |
|
"step": 5445 |
|
}, |
|
{ |
|
"epoch": 13.148371531966225, |
|
"grad_norm": 0.326171875, |
|
"learning_rate": 8.986980701518966e-06, |
|
"loss": 0.8779, |
|
"step": 5450 |
|
}, |
|
{ |
|
"epoch": 13.16043425814234, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 8.870894422137177e-06, |
|
"loss": 0.8714, |
|
"step": 5455 |
|
}, |
|
{ |
|
"epoch": 13.172496984318457, |
|
"grad_norm": 0.3125, |
|
"learning_rate": 8.755527970126853e-06, |
|
"loss": 0.8753, |
|
"step": 5460 |
|
}, |
|
{ |
|
"epoch": 13.184559710494572, |
|
"grad_norm": 0.326171875, |
|
"learning_rate": 8.640882256765759e-06, |
|
"loss": 0.8712, |
|
"step": 5465 |
|
}, |
|
{ |
|
"epoch": 13.196622436670687, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 8.526958187638635e-06, |
|
"loss": 0.8703, |
|
"step": 5470 |
|
}, |
|
{ |
|
"epoch": 13.208685162846804, |
|
"grad_norm": 0.33203125, |
|
"learning_rate": 8.413756662629879e-06, |
|
"loss": 0.874, |
|
"step": 5475 |
|
}, |
|
{ |
|
"epoch": 13.220747889022919, |
|
"grad_norm": 0.33203125, |
|
"learning_rate": 8.301278575916538e-06, |
|
"loss": 0.8721, |
|
"step": 5480 |
|
}, |
|
{ |
|
"epoch": 13.232810615199035, |
|
"grad_norm": 0.310546875, |
|
"learning_rate": 8.189524815961291e-06, |
|
"loss": 0.8761, |
|
"step": 5485 |
|
}, |
|
{ |
|
"epoch": 13.24487334137515, |
|
"grad_norm": 0.32421875, |
|
"learning_rate": 8.07849626550531e-06, |
|
"loss": 0.8616, |
|
"step": 5490 |
|
}, |
|
{ |
|
"epoch": 13.256936067551267, |
|
"grad_norm": 0.3125, |
|
"learning_rate": 7.968193801561363e-06, |
|
"loss": 0.881, |
|
"step": 5495 |
|
}, |
|
{ |
|
"epoch": 13.268998793727382, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 7.858618295406895e-06, |
|
"loss": 0.8724, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 13.281061519903497, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 7.74977061257709e-06, |
|
"loss": 0.8826, |
|
"step": 5505 |
|
}, |
|
{ |
|
"epoch": 13.293124246079614, |
|
"grad_norm": 0.310546875, |
|
"learning_rate": 7.641651612858081e-06, |
|
"loss": 0.8864, |
|
"step": 5510 |
|
}, |
|
{ |
|
"epoch": 13.30518697225573, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 7.534262150280136e-06, |
|
"loss": 0.8781, |
|
"step": 5515 |
|
}, |
|
{ |
|
"epoch": 13.317249698431846, |
|
"grad_norm": 0.34375, |
|
"learning_rate": 7.427603073110967e-06, |
|
"loss": 0.8721, |
|
"step": 5520 |
|
}, |
|
{ |
|
"epoch": 13.329312424607961, |
|
"grad_norm": 0.310546875, |
|
"learning_rate": 7.321675223848923e-06, |
|
"loss": 0.8746, |
|
"step": 5525 |
|
}, |
|
{ |
|
"epoch": 13.341375150784078, |
|
"grad_norm": 0.3125, |
|
"learning_rate": 7.216479439216406e-06, |
|
"loss": 0.8657, |
|
"step": 5530 |
|
}, |
|
{ |
|
"epoch": 13.353437876960193, |
|
"grad_norm": 0.326171875, |
|
"learning_rate": 7.1120165501533e-06, |
|
"loss": 0.8668, |
|
"step": 5535 |
|
}, |
|
{ |
|
"epoch": 13.365500603136308, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 7.0082873818102964e-06, |
|
"loss": 0.8716, |
|
"step": 5540 |
|
}, |
|
{ |
|
"epoch": 13.377563329312425, |
|
"grad_norm": 0.326171875, |
|
"learning_rate": 6.90529275354247e-06, |
|
"loss": 0.8693, |
|
"step": 5545 |
|
}, |
|
{ |
|
"epoch": 13.38962605548854, |
|
"grad_norm": 0.3125, |
|
"learning_rate": 6.803033478902765e-06, |
|
"loss": 0.8786, |
|
"step": 5550 |
|
}, |
|
{ |
|
"epoch": 13.401688781664657, |
|
"grad_norm": 0.328125, |
|
"learning_rate": 6.7015103656355885e-06, |
|
"loss": 0.8846, |
|
"step": 5555 |
|
}, |
|
{ |
|
"epoch": 13.413751507840772, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 6.600724215670429e-06, |
|
"loss": 0.8735, |
|
"step": 5560 |
|
}, |
|
{ |
|
"epoch": 13.425814234016888, |
|
"grad_norm": 0.33203125, |
|
"learning_rate": 6.500675825115454e-06, |
|
"loss": 0.8701, |
|
"step": 5565 |
|
}, |
|
{ |
|
"epoch": 13.437876960193003, |
|
"grad_norm": 0.30859375, |
|
"learning_rate": 6.401365984251384e-06, |
|
"loss": 0.8764, |
|
"step": 5570 |
|
}, |
|
{ |
|
"epoch": 13.44993968636912, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 6.302795477525058e-06, |
|
"loss": 0.8824, |
|
"step": 5575 |
|
}, |
|
{ |
|
"epoch": 13.462002412545235, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 6.204965083543368e-06, |
|
"loss": 0.8773, |
|
"step": 5580 |
|
}, |
|
{ |
|
"epoch": 13.47406513872135, |
|
"grad_norm": 0.310546875, |
|
"learning_rate": 6.107875575067057e-06, |
|
"loss": 0.868, |
|
"step": 5585 |
|
}, |
|
{ |
|
"epoch": 13.486127864897467, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 6.011527719004617e-06, |
|
"loss": 0.8706, |
|
"step": 5590 |
|
}, |
|
{ |
|
"epoch": 13.498190591073582, |
|
"grad_norm": 0.330078125, |
|
"learning_rate": 5.915922276406249e-06, |
|
"loss": 0.878, |
|
"step": 5595 |
|
}, |
|
{ |
|
"epoch": 13.510253317249699, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 5.8210600024578234e-06, |
|
"loss": 0.8739, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 13.522316043425814, |
|
"grad_norm": 0.3125, |
|
"learning_rate": 5.726941646474992e-06, |
|
"loss": 0.8768, |
|
"step": 5605 |
|
}, |
|
{ |
|
"epoch": 13.534378769601929, |
|
"grad_norm": 0.30859375, |
|
"learning_rate": 5.633567951897145e-06, |
|
"loss": 0.8785, |
|
"step": 5610 |
|
}, |
|
{ |
|
"epoch": 13.546441495778046, |
|
"grad_norm": 0.326171875, |
|
"learning_rate": 5.5409396562816076e-06, |
|
"loss": 0.8751, |
|
"step": 5615 |
|
}, |
|
{ |
|
"epoch": 13.558504221954161, |
|
"grad_norm": 0.30859375, |
|
"learning_rate": 5.449057491297871e-06, |
|
"loss": 0.8861, |
|
"step": 5620 |
|
}, |
|
{ |
|
"epoch": 13.570566948130278, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 5.357922182721687e-06, |
|
"loss": 0.8782, |
|
"step": 5625 |
|
}, |
|
{ |
|
"epoch": 13.582629674306393, |
|
"grad_norm": 0.30859375, |
|
"learning_rate": 5.267534450429423e-06, |
|
"loss": 0.8705, |
|
"step": 5630 |
|
}, |
|
{ |
|
"epoch": 13.59469240048251, |
|
"grad_norm": 0.3125, |
|
"learning_rate": 5.177895008392353e-06, |
|
"loss": 0.8755, |
|
"step": 5635 |
|
}, |
|
{ |
|
"epoch": 13.606755126658625, |
|
"grad_norm": 0.302734375, |
|
"learning_rate": 5.08900456467103e-06, |
|
"loss": 0.8744, |
|
"step": 5640 |
|
}, |
|
{ |
|
"epoch": 13.618817852834741, |
|
"grad_norm": 0.3125, |
|
"learning_rate": 5.000863821409674e-06, |
|
"loss": 0.8811, |
|
"step": 5645 |
|
}, |
|
{ |
|
"epoch": 13.630880579010856, |
|
"grad_norm": 0.30859375, |
|
"learning_rate": 4.9134734748305986e-06, |
|
"loss": 0.8718, |
|
"step": 5650 |
|
}, |
|
{ |
|
"epoch": 13.642943305186972, |
|
"grad_norm": 0.310546875, |
|
"learning_rate": 4.826834215228826e-06, |
|
"loss": 0.8773, |
|
"step": 5655 |
|
}, |
|
{ |
|
"epoch": 13.655006031363088, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 4.740946726966466e-06, |
|
"loss": 0.8811, |
|
"step": 5660 |
|
}, |
|
{ |
|
"epoch": 13.667068757539203, |
|
"grad_norm": 0.310546875, |
|
"learning_rate": 4.655811688467448e-06, |
|
"loss": 0.8729, |
|
"step": 5665 |
|
}, |
|
{ |
|
"epoch": 13.67913148371532, |
|
"grad_norm": 0.337890625, |
|
"learning_rate": 4.5714297722121106e-06, |
|
"loss": 0.8811, |
|
"step": 5670 |
|
}, |
|
{ |
|
"epoch": 13.691194209891435, |
|
"grad_norm": 0.3046875, |
|
"learning_rate": 4.4878016447318704e-06, |
|
"loss": 0.8661, |
|
"step": 5675 |
|
}, |
|
{ |
|
"epoch": 13.703256936067552, |
|
"grad_norm": 0.345703125, |
|
"learning_rate": 4.404927966604011e-06, |
|
"loss": 0.8745, |
|
"step": 5680 |
|
}, |
|
{ |
|
"epoch": 13.715319662243667, |
|
"grad_norm": 0.310546875, |
|
"learning_rate": 4.322809392446392e-06, |
|
"loss": 0.8771, |
|
"step": 5685 |
|
}, |
|
{ |
|
"epoch": 13.727382388419782, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 4.241446570912344e-06, |
|
"loss": 0.874, |
|
"step": 5690 |
|
}, |
|
{ |
|
"epoch": 13.739445114595899, |
|
"grad_norm": 0.306640625, |
|
"learning_rate": 4.160840144685507e-06, |
|
"loss": 0.8768, |
|
"step": 5695 |
|
}, |
|
{ |
|
"epoch": 13.751507840772014, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 4.080990750474778e-06, |
|
"loss": 0.8706, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 13.76357056694813, |
|
"grad_norm": 0.306640625, |
|
"learning_rate": 4.001899019009281e-06, |
|
"loss": 0.8727, |
|
"step": 5705 |
|
}, |
|
{ |
|
"epoch": 13.775633293124246, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 3.92356557503335e-06, |
|
"loss": 0.8792, |
|
"step": 5710 |
|
}, |
|
{ |
|
"epoch": 13.787696019300363, |
|
"grad_norm": 0.3203125, |
|
"learning_rate": 3.84599103730161e-06, |
|
"loss": 0.864, |
|
"step": 5715 |
|
}, |
|
{ |
|
"epoch": 13.799758745476478, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 3.7691760185741387e-06, |
|
"loss": 0.8733, |
|
"step": 5720 |
|
}, |
|
{ |
|
"epoch": 13.811821471652593, |
|
"grad_norm": 0.337890625, |
|
"learning_rate": 3.6931211256115537e-06, |
|
"loss": 0.8753, |
|
"step": 5725 |
|
}, |
|
{ |
|
"epoch": 13.82388419782871, |
|
"grad_norm": 0.32421875, |
|
"learning_rate": 3.617826959170256e-06, |
|
"loss": 0.8714, |
|
"step": 5730 |
|
}, |
|
{ |
|
"epoch": 13.835946924004825, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 3.543294113997664e-06, |
|
"loss": 0.8741, |
|
"step": 5735 |
|
}, |
|
{ |
|
"epoch": 13.848009650180941, |
|
"grad_norm": 0.3046875, |
|
"learning_rate": 3.469523178827583e-06, |
|
"loss": 0.881, |
|
"step": 5740 |
|
}, |
|
{ |
|
"epoch": 13.860072376357056, |
|
"grad_norm": 0.306640625, |
|
"learning_rate": 3.3965147363754555e-06, |
|
"loss": 0.8778, |
|
"step": 5745 |
|
}, |
|
{ |
|
"epoch": 13.872135102533173, |
|
"grad_norm": 0.310546875, |
|
"learning_rate": 3.3242693633337983e-06, |
|
"loss": 0.8804, |
|
"step": 5750 |
|
}, |
|
{ |
|
"epoch": 13.884197828709288, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 3.252787630367715e-06, |
|
"loss": 0.882, |
|
"step": 5755 |
|
}, |
|
{ |
|
"epoch": 13.896260554885405, |
|
"grad_norm": 0.3046875, |
|
"learning_rate": 3.182070102110257e-06, |
|
"loss": 0.8672, |
|
"step": 5760 |
|
}, |
|
{ |
|
"epoch": 13.90832328106152, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 3.1121173371580825e-06, |
|
"loss": 0.8772, |
|
"step": 5765 |
|
}, |
|
{ |
|
"epoch": 13.920386007237635, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 3.042929888066992e-06, |
|
"loss": 0.8807, |
|
"step": 5770 |
|
}, |
|
{ |
|
"epoch": 13.932448733413752, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 2.974508301347534e-06, |
|
"loss": 0.8796, |
|
"step": 5775 |
|
}, |
|
{ |
|
"epoch": 13.944511459589867, |
|
"grad_norm": 0.306640625, |
|
"learning_rate": 2.9068531174607615e-06, |
|
"loss": 0.8712, |
|
"step": 5780 |
|
}, |
|
{ |
|
"epoch": 13.956574185765984, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 2.839964870813916e-06, |
|
"loss": 0.8744, |
|
"step": 5785 |
|
}, |
|
{ |
|
"epoch": 13.968636911942099, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 2.7738440897561723e-06, |
|
"loss": 0.8754, |
|
"step": 5790 |
|
}, |
|
{ |
|
"epoch": 13.980699638118214, |
|
"grad_norm": 0.3125, |
|
"learning_rate": 2.7084912965745778e-06, |
|
"loss": 0.8753, |
|
"step": 5795 |
|
}, |
|
{ |
|
"epoch": 13.99276236429433, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 2.6439070074897874e-06, |
|
"loss": 0.8743, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 14.0, |
|
"eval_loss": 2.5032882690429688, |
|
"eval_runtime": 0.5357, |
|
"eval_samples_per_second": 18.667, |
|
"eval_steps_per_second": 1.867, |
|
"step": 5803 |
|
}, |
|
{ |
|
"epoch": 14.004825090470446, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 2.580091732652101e-06, |
|
"loss": 0.8799, |
|
"step": 5805 |
|
}, |
|
{ |
|
"epoch": 14.016887816646562, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 2.5170459761373664e-06, |
|
"loss": 0.8781, |
|
"step": 5810 |
|
}, |
|
{ |
|
"epoch": 14.028950542822678, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 2.454770235943027e-06, |
|
"loss": 0.8751, |
|
"step": 5815 |
|
}, |
|
{ |
|
"epoch": 14.041013268998794, |
|
"grad_norm": 0.30859375, |
|
"learning_rate": 2.3932650039841687e-06, |
|
"loss": 0.887, |
|
"step": 5820 |
|
}, |
|
{ |
|
"epoch": 14.05307599517491, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 2.3325307660896577e-06, |
|
"loss": 0.8755, |
|
"step": 5825 |
|
}, |
|
{ |
|
"epoch": 14.065138721351026, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 2.2725680019982765e-06, |
|
"loss": 0.8792, |
|
"step": 5830 |
|
}, |
|
{ |
|
"epoch": 14.077201447527141, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 2.213377185354959e-06, |
|
"loss": 0.8719, |
|
"step": 5835 |
|
}, |
|
{ |
|
"epoch": 14.089264173703256, |
|
"grad_norm": 0.306640625, |
|
"learning_rate": 2.1549587837070394e-06, |
|
"loss": 0.8779, |
|
"step": 5840 |
|
}, |
|
{ |
|
"epoch": 14.101326899879373, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 2.0973132585005552e-06, |
|
"loss": 0.8774, |
|
"step": 5845 |
|
}, |
|
{ |
|
"epoch": 14.113389626055488, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 2.0404410650765817e-06, |
|
"loss": 0.8718, |
|
"step": 5850 |
|
}, |
|
{ |
|
"epoch": 14.125452352231605, |
|
"grad_norm": 0.330078125, |
|
"learning_rate": 1.9843426526677033e-06, |
|
"loss": 0.8853, |
|
"step": 5855 |
|
}, |
|
{ |
|
"epoch": 14.13751507840772, |
|
"grad_norm": 0.326171875, |
|
"learning_rate": 1.9290184643943944e-06, |
|
"loss": 0.8779, |
|
"step": 5860 |
|
}, |
|
{ |
|
"epoch": 14.149577804583837, |
|
"grad_norm": 0.306640625, |
|
"learning_rate": 1.874468937261531e-06, |
|
"loss": 0.8692, |
|
"step": 5865 |
|
}, |
|
{ |
|
"epoch": 14.161640530759952, |
|
"grad_norm": 0.310546875, |
|
"learning_rate": 1.820694502154996e-06, |
|
"loss": 0.8756, |
|
"step": 5870 |
|
}, |
|
{ |
|
"epoch": 14.173703256936067, |
|
"grad_norm": 0.296875, |
|
"learning_rate": 1.7676955838381804e-06, |
|
"loss": 0.8697, |
|
"step": 5875 |
|
}, |
|
{ |
|
"epoch": 14.185765983112184, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 1.71547260094872e-06, |
|
"loss": 0.8832, |
|
"step": 5880 |
|
}, |
|
{ |
|
"epoch": 14.197828709288299, |
|
"grad_norm": 0.302734375, |
|
"learning_rate": 1.6640259659951419e-06, |
|
"loss": 0.8761, |
|
"step": 5885 |
|
}, |
|
{ |
|
"epoch": 14.209891435464415, |
|
"grad_norm": 0.333984375, |
|
"learning_rate": 1.61335608535359e-06, |
|
"loss": 0.8785, |
|
"step": 5890 |
|
}, |
|
{ |
|
"epoch": 14.22195416164053, |
|
"grad_norm": 0.32421875, |
|
"learning_rate": 1.5634633592646609e-06, |
|
"loss": 0.8726, |
|
"step": 5895 |
|
}, |
|
{ |
|
"epoch": 14.234016887816647, |
|
"grad_norm": 0.3046875, |
|
"learning_rate": 1.5143481818302164e-06, |
|
"loss": 0.8694, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 14.246079613992762, |
|
"grad_norm": 0.333984375, |
|
"learning_rate": 1.4660109410102652e-06, |
|
"loss": 0.8852, |
|
"step": 5905 |
|
}, |
|
{ |
|
"epoch": 14.258142340168877, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 1.4184520186199202e-06, |
|
"loss": 0.8746, |
|
"step": 5910 |
|
}, |
|
{ |
|
"epoch": 14.270205066344994, |
|
"grad_norm": 0.3046875, |
|
"learning_rate": 1.3716717903263566e-06, |
|
"loss": 0.8767, |
|
"step": 5915 |
|
}, |
|
{ |
|
"epoch": 14.28226779252111, |
|
"grad_norm": 0.3203125, |
|
"learning_rate": 1.3256706256458585e-06, |
|
"loss": 0.8694, |
|
"step": 5920 |
|
}, |
|
{ |
|
"epoch": 14.294330518697226, |
|
"grad_norm": 0.310546875, |
|
"learning_rate": 1.2804488879408993e-06, |
|
"loss": 0.8709, |
|
"step": 5925 |
|
}, |
|
{ |
|
"epoch": 14.306393244873341, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 1.2360069344172887e-06, |
|
"loss": 0.8722, |
|
"step": 5930 |
|
}, |
|
{ |
|
"epoch": 14.318455971049458, |
|
"grad_norm": 0.30859375, |
|
"learning_rate": 1.1923451161213074e-06, |
|
"loss": 0.8754, |
|
"step": 5935 |
|
}, |
|
{ |
|
"epoch": 14.330518697225573, |
|
"grad_norm": 0.306640625, |
|
"learning_rate": 1.1494637779369766e-06, |
|
"loss": 0.8737, |
|
"step": 5940 |
|
}, |
|
{ |
|
"epoch": 14.342581423401688, |
|
"grad_norm": 0.306640625, |
|
"learning_rate": 1.107363258583305e-06, |
|
"loss": 0.8713, |
|
"step": 5945 |
|
}, |
|
{ |
|
"epoch": 14.354644149577805, |
|
"grad_norm": 0.302734375, |
|
"learning_rate": 1.0660438906116342e-06, |
|
"loss": 0.8744, |
|
"step": 5950 |
|
}, |
|
{ |
|
"epoch": 14.36670687575392, |
|
"grad_norm": 0.328125, |
|
"learning_rate": 1.0255060004030093e-06, |
|
"loss": 0.8805, |
|
"step": 5955 |
|
}, |
|
{ |
|
"epoch": 14.378769601930037, |
|
"grad_norm": 0.3046875, |
|
"learning_rate": 9.857499081655786e-07, |
|
"loss": 0.8786, |
|
"step": 5960 |
|
}, |
|
{ |
|
"epoch": 14.390832328106152, |
|
"grad_norm": 0.30078125, |
|
"learning_rate": 9.467759279320976e-07, |
|
"loss": 0.882, |
|
"step": 5965 |
|
}, |
|
{ |
|
"epoch": 14.402895054282268, |
|
"grad_norm": 0.3046875, |
|
"learning_rate": 9.085843675574079e-07, |
|
"loss": 0.8756, |
|
"step": 5970 |
|
}, |
|
{ |
|
"epoch": 14.414957780458383, |
|
"grad_norm": 0.3125, |
|
"learning_rate": 8.711755287160727e-07, |
|
"loss": 0.8743, |
|
"step": 5975 |
|
}, |
|
{ |
|
"epoch": 14.427020506634499, |
|
"grad_norm": 0.310546875, |
|
"learning_rate": 8.345497068998897e-07, |
|
"loss": 0.8792, |
|
"step": 5980 |
|
}, |
|
{ |
|
"epoch": 14.439083232810615, |
|
"grad_norm": 0.30859375, |
|
"learning_rate": 7.987071914156596e-07, |
|
"loss": 0.8664, |
|
"step": 5985 |
|
}, |
|
{ |
|
"epoch": 14.45114595898673, |
|
"grad_norm": 0.3046875, |
|
"learning_rate": 7.636482653828658e-07, |
|
"loss": 0.8744, |
|
"step": 5990 |
|
}, |
|
{ |
|
"epoch": 14.463208685162847, |
|
"grad_norm": 0.310546875, |
|
"learning_rate": 7.293732057313874e-07, |
|
"loss": 0.8674, |
|
"step": 5995 |
|
}, |
|
{ |
|
"epoch": 14.475271411338962, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 6.958822831994005e-07, |
|
"loss": 0.8744, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 14.487334137515079, |
|
"grad_norm": 0.30078125, |
|
"learning_rate": 6.631757623311586e-07, |
|
"loss": 0.872, |
|
"step": 6005 |
|
}, |
|
{ |
|
"epoch": 14.499396863691194, |
|
"grad_norm": 0.3046875, |
|
"learning_rate": 6.312539014749818e-07, |
|
"loss": 0.8838, |
|
"step": 6010 |
|
}, |
|
{ |
|
"epoch": 14.51145958986731, |
|
"grad_norm": 0.326171875, |
|
"learning_rate": 6.001169527811268e-07, |
|
"loss": 0.8684, |
|
"step": 6015 |
|
}, |
|
{ |
|
"epoch": 14.523522316043426, |
|
"grad_norm": 0.33203125, |
|
"learning_rate": 5.697651621998535e-07, |
|
"loss": 0.8776, |
|
"step": 6020 |
|
}, |
|
{ |
|
"epoch": 14.535585042219541, |
|
"grad_norm": 0.3125, |
|
"learning_rate": 5.401987694794941e-07, |
|
"loss": 0.8694, |
|
"step": 6025 |
|
}, |
|
{ |
|
"epoch": 14.547647768395658, |
|
"grad_norm": 0.3203125, |
|
"learning_rate": 5.114180081645214e-07, |
|
"loss": 0.8698, |
|
"step": 6030 |
|
}, |
|
{ |
|
"epoch": 14.559710494571773, |
|
"grad_norm": 0.310546875, |
|
"learning_rate": 4.834231055937055e-07, |
|
"loss": 0.8816, |
|
"step": 6035 |
|
}, |
|
{ |
|
"epoch": 14.57177322074789, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 4.5621428289834843e-07, |
|
"loss": 0.8839, |
|
"step": 6040 |
|
}, |
|
{ |
|
"epoch": 14.583835946924005, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 4.2979175500050817e-07, |
|
"loss": 0.8699, |
|
"step": 6045 |
|
}, |
|
{ |
|
"epoch": 14.595898673100121, |
|
"grad_norm": 0.306640625, |
|
"learning_rate": 4.0415573061133307e-07, |
|
"loss": 0.8657, |
|
"step": 6050 |
|
}, |
|
{ |
|
"epoch": 14.607961399276236, |
|
"grad_norm": 0.345703125, |
|
"learning_rate": 3.7930641222934103e-07, |
|
"loss": 0.8832, |
|
"step": 6055 |
|
}, |
|
{ |
|
"epoch": 14.620024125452352, |
|
"grad_norm": 0.30859375, |
|
"learning_rate": 3.552439961389431e-07, |
|
"loss": 0.8762, |
|
"step": 6060 |
|
}, |
|
{ |
|
"epoch": 14.632086851628468, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 3.3196867240876675e-07, |
|
"loss": 0.8634, |
|
"step": 6065 |
|
}, |
|
{ |
|
"epoch": 14.644149577804583, |
|
"grad_norm": 0.32421875, |
|
"learning_rate": 3.094806248902349e-07, |
|
"loss": 0.8749, |
|
"step": 6070 |
|
}, |
|
{ |
|
"epoch": 14.6562123039807, |
|
"grad_norm": 0.3046875, |
|
"learning_rate": 2.877800312160783e-07, |
|
"loss": 0.8708, |
|
"step": 6075 |
|
}, |
|
{ |
|
"epoch": 14.668275030156815, |
|
"grad_norm": 0.302734375, |
|
"learning_rate": 2.668670627989478e-07, |
|
"loss": 0.874, |
|
"step": 6080 |
|
}, |
|
{ |
|
"epoch": 14.680337756332932, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 2.4674188483003734e-07, |
|
"loss": 0.881, |
|
"step": 6085 |
|
}, |
|
{ |
|
"epoch": 14.692400482509047, |
|
"grad_norm": 0.310546875, |
|
"learning_rate": 2.274046562778409e-07, |
|
"loss": 0.864, |
|
"step": 6090 |
|
}, |
|
{ |
|
"epoch": 14.704463208685162, |
|
"grad_norm": 0.302734375, |
|
"learning_rate": 2.088555298867978e-07, |
|
"loss": 0.8669, |
|
"step": 6095 |
|
}, |
|
{ |
|
"epoch": 14.716525934861279, |
|
"grad_norm": 0.310546875, |
|
"learning_rate": 1.910946521761714e-07, |
|
"loss": 0.8725, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 14.728588661037394, |
|
"grad_norm": 0.30859375, |
|
"learning_rate": 1.7412216343885014e-07, |
|
"loss": 0.8664, |
|
"step": 6105 |
|
}, |
|
{ |
|
"epoch": 14.74065138721351, |
|
"grad_norm": 0.310546875, |
|
"learning_rate": 1.5793819774027053e-07, |
|
"loss": 0.875, |
|
"step": 6110 |
|
}, |
|
{ |
|
"epoch": 14.752714113389626, |
|
"grad_norm": 0.306640625, |
|
"learning_rate": 1.425428829173181e-07, |
|
"loss": 0.8681, |
|
"step": 6115 |
|
}, |
|
{ |
|
"epoch": 14.764776839565743, |
|
"grad_norm": 0.302734375, |
|
"learning_rate": 1.2793634057732818e-07, |
|
"loss": 0.8831, |
|
"step": 6120 |
|
}, |
|
{ |
|
"epoch": 14.776839565741858, |
|
"grad_norm": 0.310546875, |
|
"learning_rate": 1.1411868609717546e-07, |
|
"loss": 0.8769, |
|
"step": 6125 |
|
}, |
|
{ |
|
"epoch": 14.788902291917973, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 1.010900286222638e-07, |
|
"loss": 0.8717, |
|
"step": 6130 |
|
}, |
|
{ |
|
"epoch": 14.80096501809409, |
|
"grad_norm": 0.3203125, |
|
"learning_rate": 8.885047106578227e-08, |
|
"loss": 0.8743, |
|
"step": 6135 |
|
}, |
|
{ |
|
"epoch": 14.813027744270205, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 7.740011010778369e-08, |
|
"loss": 0.867, |
|
"step": 6140 |
|
}, |
|
{ |
|
"epoch": 14.825090470446321, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 6.673903619449639e-08, |
|
"loss": 0.875, |
|
"step": 6145 |
|
}, |
|
{ |
|
"epoch": 14.837153196622436, |
|
"grad_norm": 0.330078125, |
|
"learning_rate": 5.6867333537580226e-08, |
|
"loss": 0.8846, |
|
"step": 6150 |
|
}, |
|
{ |
|
"epoch": 14.849215922798553, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 4.7785080113449396e-08, |
|
"loss": 0.8832, |
|
"step": 6155 |
|
}, |
|
{ |
|
"epoch": 14.861278648974668, |
|
"grad_norm": 0.32421875, |
|
"learning_rate": 3.9492347662684057e-08, |
|
"loss": 0.8764, |
|
"step": 6160 |
|
}, |
|
{ |
|
"epoch": 14.873341375150783, |
|
"grad_norm": 0.3203125, |
|
"learning_rate": 3.1989201689452967e-08, |
|
"loss": 0.8753, |
|
"step": 6165 |
|
}, |
|
{ |
|
"epoch": 14.8854041013269, |
|
"grad_norm": 0.3046875, |
|
"learning_rate": 2.5275701460991674e-08, |
|
"loss": 0.8682, |
|
"step": 6170 |
|
}, |
|
{ |
|
"epoch": 14.897466827503015, |
|
"grad_norm": 0.306640625, |
|
"learning_rate": 1.9351900007114066e-08, |
|
"loss": 0.8719, |
|
"step": 6175 |
|
}, |
|
{ |
|
"epoch": 14.909529553679132, |
|
"grad_norm": 0.30078125, |
|
"learning_rate": 1.4217844119857048e-08, |
|
"loss": 0.8759, |
|
"step": 6180 |
|
}, |
|
{ |
|
"epoch": 14.921592279855247, |
|
"grad_norm": 0.30859375, |
|
"learning_rate": 9.873574353025384e-09, |
|
"loss": 0.8732, |
|
"step": 6185 |
|
}, |
|
{ |
|
"epoch": 14.933655006031364, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 6.3191250219474255e-09, |
|
"loss": 0.8793, |
|
"step": 6190 |
|
}, |
|
{ |
|
"epoch": 14.945717732207479, |
|
"grad_norm": 0.298828125, |
|
"learning_rate": 3.554524203175369e-09, |
|
"loss": 0.876, |
|
"step": 6195 |
|
}, |
|
{ |
|
"epoch": 14.957780458383596, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 1.5797937342298952e-09, |
|
"loss": 0.8788, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 14.96984318455971, |
|
"grad_norm": 0.30859375, |
|
"learning_rate": 3.949492134780464e-10, |
|
"loss": 0.8771, |
|
"step": 6205 |
|
}, |
|
{ |
|
"epoch": 14.981905910735826, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 0.0, |
|
"loss": 0.8746, |
|
"step": 6210 |
|
}, |
|
{ |
|
"epoch": 14.981905910735826, |
|
"eval_loss": 2.503845453262329, |
|
"eval_runtime": 0.5267, |
|
"eval_samples_per_second": 18.985, |
|
"eval_steps_per_second": 1.899, |
|
"step": 6210 |
|
}, |
|
{ |
|
"epoch": 14.981905910735826, |
|
"step": 6210, |
|
"total_flos": 4.863451355047526e+18, |
|
"train_loss": 0.9534509487582098, |
|
"train_runtime": 21285.3674, |
|
"train_samples_per_second": 18.683, |
|
"train_steps_per_second": 0.292 |
|
} |
|
], |
|
"logging_steps": 5, |
|
"max_steps": 6210, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 15, |
|
"save_steps": 100, |
|
"total_flos": 4.863451355047526e+18, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|