chansung's picture
Model save
f097832 verified
raw
history blame
210 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 14.981905910735826,
"eval_steps": 500,
"global_step": 6210,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0024125452352231603,
"grad_norm": 6.09375,
"learning_rate": 3.22061191626409e-07,
"loss": 3.0105,
"step": 1
},
{
"epoch": 0.012062726176115802,
"grad_norm": 2.21875,
"learning_rate": 1.610305958132045e-06,
"loss": 3.0058,
"step": 5
},
{
"epoch": 0.024125452352231604,
"grad_norm": 2.125,
"learning_rate": 3.22061191626409e-06,
"loss": 3.059,
"step": 10
},
{
"epoch": 0.03618817852834741,
"grad_norm": 2.015625,
"learning_rate": 4.830917874396135e-06,
"loss": 3.011,
"step": 15
},
{
"epoch": 0.04825090470446321,
"grad_norm": 2.03125,
"learning_rate": 6.44122383252818e-06,
"loss": 3.0197,
"step": 20
},
{
"epoch": 0.06031363088057901,
"grad_norm": 2.953125,
"learning_rate": 8.051529790660225e-06,
"loss": 3.004,
"step": 25
},
{
"epoch": 0.07237635705669482,
"grad_norm": 2.625,
"learning_rate": 9.66183574879227e-06,
"loss": 2.9712,
"step": 30
},
{
"epoch": 0.08443908323281062,
"grad_norm": 3.984375,
"learning_rate": 1.1272141706924317e-05,
"loss": 2.9778,
"step": 35
},
{
"epoch": 0.09650180940892641,
"grad_norm": 2.34375,
"learning_rate": 1.288244766505636e-05,
"loss": 2.881,
"step": 40
},
{
"epoch": 0.10856453558504221,
"grad_norm": 2.0625,
"learning_rate": 1.4492753623188407e-05,
"loss": 2.8379,
"step": 45
},
{
"epoch": 0.12062726176115803,
"grad_norm": 2.96875,
"learning_rate": 1.610305958132045e-05,
"loss": 2.738,
"step": 50
},
{
"epoch": 0.13268998793727382,
"grad_norm": 1.8359375,
"learning_rate": 1.77133655394525e-05,
"loss": 2.6525,
"step": 55
},
{
"epoch": 0.14475271411338964,
"grad_norm": 2.609375,
"learning_rate": 1.932367149758454e-05,
"loss": 2.5732,
"step": 60
},
{
"epoch": 0.15681544028950542,
"grad_norm": 1.4609375,
"learning_rate": 2.0933977455716587e-05,
"loss": 2.4813,
"step": 65
},
{
"epoch": 0.16887816646562123,
"grad_norm": 1.0546875,
"learning_rate": 2.2544283413848633e-05,
"loss": 2.4334,
"step": 70
},
{
"epoch": 0.18094089264173704,
"grad_norm": 2.625,
"learning_rate": 2.4154589371980676e-05,
"loss": 2.3332,
"step": 75
},
{
"epoch": 0.19300361881785283,
"grad_norm": 1.1015625,
"learning_rate": 2.576489533011272e-05,
"loss": 2.2571,
"step": 80
},
{
"epoch": 0.20506634499396864,
"grad_norm": 5.34375,
"learning_rate": 2.7375201288244768e-05,
"loss": 2.1755,
"step": 85
},
{
"epoch": 0.21712907117008443,
"grad_norm": 1.0390625,
"learning_rate": 2.8985507246376814e-05,
"loss": 2.089,
"step": 90
},
{
"epoch": 0.22919179734620024,
"grad_norm": 3.328125,
"learning_rate": 3.059581320450886e-05,
"loss": 2.0223,
"step": 95
},
{
"epoch": 0.24125452352231605,
"grad_norm": 1.0390625,
"learning_rate": 3.22061191626409e-05,
"loss": 1.9332,
"step": 100
},
{
"epoch": 0.25331724969843183,
"grad_norm": 1.046875,
"learning_rate": 3.381642512077295e-05,
"loss": 1.8597,
"step": 105
},
{
"epoch": 0.26537997587454765,
"grad_norm": 1.0859375,
"learning_rate": 3.5426731078905e-05,
"loss": 1.8142,
"step": 110
},
{
"epoch": 0.27744270205066346,
"grad_norm": 0.625,
"learning_rate": 3.7037037037037037e-05,
"loss": 1.7501,
"step": 115
},
{
"epoch": 0.28950542822677927,
"grad_norm": 0.59375,
"learning_rate": 3.864734299516908e-05,
"loss": 1.6938,
"step": 120
},
{
"epoch": 0.30156815440289503,
"grad_norm": 0.478515625,
"learning_rate": 4.025764895330113e-05,
"loss": 1.6289,
"step": 125
},
{
"epoch": 0.31363088057901084,
"grad_norm": 0.671875,
"learning_rate": 4.1867954911433174e-05,
"loss": 1.5807,
"step": 130
},
{
"epoch": 0.32569360675512665,
"grad_norm": 0.423828125,
"learning_rate": 4.347826086956522e-05,
"loss": 1.5408,
"step": 135
},
{
"epoch": 0.33775633293124246,
"grad_norm": 0.494140625,
"learning_rate": 4.5088566827697266e-05,
"loss": 1.4961,
"step": 140
},
{
"epoch": 0.3498190591073583,
"grad_norm": 0.40234375,
"learning_rate": 4.669887278582931e-05,
"loss": 1.4528,
"step": 145
},
{
"epoch": 0.3618817852834741,
"grad_norm": 0.322265625,
"learning_rate": 4.830917874396135e-05,
"loss": 1.4274,
"step": 150
},
{
"epoch": 0.37394451145958985,
"grad_norm": 0.345703125,
"learning_rate": 4.99194847020934e-05,
"loss": 1.3942,
"step": 155
},
{
"epoch": 0.38600723763570566,
"grad_norm": 0.455078125,
"learning_rate": 5.152979066022544e-05,
"loss": 1.387,
"step": 160
},
{
"epoch": 0.39806996381182147,
"grad_norm": 0.404296875,
"learning_rate": 5.3140096618357496e-05,
"loss": 1.3418,
"step": 165
},
{
"epoch": 0.4101326899879373,
"grad_norm": 0.3046875,
"learning_rate": 5.4750402576489535e-05,
"loss": 1.3263,
"step": 170
},
{
"epoch": 0.4221954161640531,
"grad_norm": 0.279296875,
"learning_rate": 5.6360708534621574e-05,
"loss": 1.3058,
"step": 175
},
{
"epoch": 0.43425814234016885,
"grad_norm": 0.53125,
"learning_rate": 5.797101449275363e-05,
"loss": 1.2854,
"step": 180
},
{
"epoch": 0.44632086851628466,
"grad_norm": 0.490234375,
"learning_rate": 5.9581320450885666e-05,
"loss": 1.2769,
"step": 185
},
{
"epoch": 0.4583835946924005,
"grad_norm": 0.359375,
"learning_rate": 6.119162640901772e-05,
"loss": 1.2572,
"step": 190
},
{
"epoch": 0.4704463208685163,
"grad_norm": 0.2236328125,
"learning_rate": 6.280193236714976e-05,
"loss": 1.2506,
"step": 195
},
{
"epoch": 0.4825090470446321,
"grad_norm": 0.2734375,
"learning_rate": 6.44122383252818e-05,
"loss": 1.2466,
"step": 200
},
{
"epoch": 0.4945717732207479,
"grad_norm": 0.380859375,
"learning_rate": 6.602254428341386e-05,
"loss": 1.2347,
"step": 205
},
{
"epoch": 0.5066344993968637,
"grad_norm": 0.2119140625,
"learning_rate": 6.76328502415459e-05,
"loss": 1.2245,
"step": 210
},
{
"epoch": 0.5186972255729795,
"grad_norm": 0.2294921875,
"learning_rate": 6.924315619967794e-05,
"loss": 1.2147,
"step": 215
},
{
"epoch": 0.5307599517490953,
"grad_norm": 0.322265625,
"learning_rate": 7.085346215781e-05,
"loss": 1.2062,
"step": 220
},
{
"epoch": 0.5428226779252111,
"grad_norm": 0.416015625,
"learning_rate": 7.246376811594203e-05,
"loss": 1.2057,
"step": 225
},
{
"epoch": 0.5548854041013269,
"grad_norm": 0.357421875,
"learning_rate": 7.407407407407407e-05,
"loss": 1.1828,
"step": 230
},
{
"epoch": 0.5669481302774427,
"grad_norm": 0.279296875,
"learning_rate": 7.568438003220612e-05,
"loss": 1.1799,
"step": 235
},
{
"epoch": 0.5790108564535585,
"grad_norm": 0.390625,
"learning_rate": 7.729468599033817e-05,
"loss": 1.1689,
"step": 240
},
{
"epoch": 0.5910735826296744,
"grad_norm": 0.69921875,
"learning_rate": 7.890499194847021e-05,
"loss": 1.1764,
"step": 245
},
{
"epoch": 0.6031363088057901,
"grad_norm": 0.46484375,
"learning_rate": 8.051529790660226e-05,
"loss": 1.1552,
"step": 250
},
{
"epoch": 0.6151990349819059,
"grad_norm": 0.2421875,
"learning_rate": 8.21256038647343e-05,
"loss": 1.1596,
"step": 255
},
{
"epoch": 0.6272617611580217,
"grad_norm": 0.58984375,
"learning_rate": 8.373590982286635e-05,
"loss": 1.1637,
"step": 260
},
{
"epoch": 0.6393244873341375,
"grad_norm": 0.578125,
"learning_rate": 8.53462157809984e-05,
"loss": 1.15,
"step": 265
},
{
"epoch": 0.6513872135102533,
"grad_norm": 0.41015625,
"learning_rate": 8.695652173913044e-05,
"loss": 1.1566,
"step": 270
},
{
"epoch": 0.6634499396863691,
"grad_norm": 0.80859375,
"learning_rate": 8.856682769726249e-05,
"loss": 1.1402,
"step": 275
},
{
"epoch": 0.6755126658624849,
"grad_norm": 0.8984375,
"learning_rate": 9.017713365539453e-05,
"loss": 1.1368,
"step": 280
},
{
"epoch": 0.6875753920386007,
"grad_norm": 0.3125,
"learning_rate": 9.178743961352657e-05,
"loss": 1.1269,
"step": 285
},
{
"epoch": 0.6996381182147166,
"grad_norm": 0.341796875,
"learning_rate": 9.339774557165862e-05,
"loss": 1.1273,
"step": 290
},
{
"epoch": 0.7117008443908324,
"grad_norm": 0.26953125,
"learning_rate": 9.500805152979067e-05,
"loss": 1.1166,
"step": 295
},
{
"epoch": 0.7237635705669482,
"grad_norm": 0.2578125,
"learning_rate": 9.66183574879227e-05,
"loss": 1.1147,
"step": 300
},
{
"epoch": 0.7358262967430639,
"grad_norm": 0.310546875,
"learning_rate": 9.822866344605476e-05,
"loss": 1.1155,
"step": 305
},
{
"epoch": 0.7478890229191797,
"grad_norm": 0.41015625,
"learning_rate": 9.98389694041868e-05,
"loss": 1.1081,
"step": 310
},
{
"epoch": 0.7599517490952955,
"grad_norm": 0.3359375,
"learning_rate": 0.00010144927536231885,
"loss": 1.1083,
"step": 315
},
{
"epoch": 0.7720144752714113,
"grad_norm": 0.314453125,
"learning_rate": 0.00010305958132045089,
"loss": 1.0981,
"step": 320
},
{
"epoch": 0.7840772014475271,
"grad_norm": 0.478515625,
"learning_rate": 0.00010466988727858293,
"loss": 1.1001,
"step": 325
},
{
"epoch": 0.7961399276236429,
"grad_norm": 0.5078125,
"learning_rate": 0.00010628019323671499,
"loss": 1.1048,
"step": 330
},
{
"epoch": 0.8082026537997588,
"grad_norm": 0.71875,
"learning_rate": 0.00010789049919484701,
"loss": 1.0983,
"step": 335
},
{
"epoch": 0.8202653799758746,
"grad_norm": 0.388671875,
"learning_rate": 0.00010950080515297907,
"loss": 1.0982,
"step": 340
},
{
"epoch": 0.8323281061519904,
"grad_norm": 0.69921875,
"learning_rate": 0.00011111111111111112,
"loss": 1.0982,
"step": 345
},
{
"epoch": 0.8443908323281062,
"grad_norm": 0.53515625,
"learning_rate": 0.00011272141706924315,
"loss": 1.0968,
"step": 350
},
{
"epoch": 0.856453558504222,
"grad_norm": 0.609375,
"learning_rate": 0.00011433172302737521,
"loss": 1.084,
"step": 355
},
{
"epoch": 0.8685162846803377,
"grad_norm": 0.6015625,
"learning_rate": 0.00011594202898550725,
"loss": 1.0815,
"step": 360
},
{
"epoch": 0.8805790108564535,
"grad_norm": 0.373046875,
"learning_rate": 0.00011755233494363929,
"loss": 1.083,
"step": 365
},
{
"epoch": 0.8926417370325693,
"grad_norm": 0.68359375,
"learning_rate": 0.00011916264090177133,
"loss": 1.0735,
"step": 370
},
{
"epoch": 0.9047044632086851,
"grad_norm": 1.09375,
"learning_rate": 0.00012077294685990339,
"loss": 1.0852,
"step": 375
},
{
"epoch": 0.916767189384801,
"grad_norm": 0.4765625,
"learning_rate": 0.00012238325281803544,
"loss": 1.0852,
"step": 380
},
{
"epoch": 0.9288299155609168,
"grad_norm": 0.30859375,
"learning_rate": 0.00012399355877616748,
"loss": 1.0689,
"step": 385
},
{
"epoch": 0.9408926417370326,
"grad_norm": 0.48828125,
"learning_rate": 0.00012560386473429953,
"loss": 1.0771,
"step": 390
},
{
"epoch": 0.9529553679131484,
"grad_norm": 0.58203125,
"learning_rate": 0.00012721417069243158,
"loss": 1.0701,
"step": 395
},
{
"epoch": 0.9650180940892642,
"grad_norm": 0.333984375,
"learning_rate": 0.0001288244766505636,
"loss": 1.07,
"step": 400
},
{
"epoch": 0.97708082026538,
"grad_norm": 0.423828125,
"learning_rate": 0.00013043478260869567,
"loss": 1.0665,
"step": 405
},
{
"epoch": 0.9891435464414958,
"grad_norm": 0.3359375,
"learning_rate": 0.00013204508856682771,
"loss": 1.0614,
"step": 410
},
{
"epoch": 0.9987937273823885,
"eval_loss": 2.4759693145751953,
"eval_runtime": 0.6381,
"eval_samples_per_second": 15.671,
"eval_steps_per_second": 1.567,
"step": 414
},
{
"epoch": 1.0012062726176116,
"grad_norm": 0.30859375,
"learning_rate": 0.00013365539452495973,
"loss": 1.0719,
"step": 415
},
{
"epoch": 1.0132689987937273,
"grad_norm": 0.32421875,
"learning_rate": 0.0001352657004830918,
"loss": 1.0513,
"step": 420
},
{
"epoch": 1.0253317249698433,
"grad_norm": 0.63671875,
"learning_rate": 0.00013687600644122385,
"loss": 1.0561,
"step": 425
},
{
"epoch": 1.037394451145959,
"grad_norm": 0.41796875,
"learning_rate": 0.00013848631239935587,
"loss": 1.0549,
"step": 430
},
{
"epoch": 1.0494571773220749,
"grad_norm": 0.337890625,
"learning_rate": 0.00014009661835748792,
"loss": 1.052,
"step": 435
},
{
"epoch": 1.0615199034981906,
"grad_norm": 0.482421875,
"learning_rate": 0.00014170692431562,
"loss": 1.0525,
"step": 440
},
{
"epoch": 1.0735826296743065,
"grad_norm": 0.7421875,
"learning_rate": 0.000143317230273752,
"loss": 1.0521,
"step": 445
},
{
"epoch": 1.0856453558504222,
"grad_norm": 0.50390625,
"learning_rate": 0.00014492753623188405,
"loss": 1.0538,
"step": 450
},
{
"epoch": 1.097708082026538,
"grad_norm": 0.6484375,
"learning_rate": 0.00014653784219001613,
"loss": 1.0593,
"step": 455
},
{
"epoch": 1.1097708082026538,
"grad_norm": 0.482421875,
"learning_rate": 0.00014814814814814815,
"loss": 1.0443,
"step": 460
},
{
"epoch": 1.1218335343787695,
"grad_norm": 0.373046875,
"learning_rate": 0.0001497584541062802,
"loss": 1.044,
"step": 465
},
{
"epoch": 1.1338962605548855,
"grad_norm": 0.48828125,
"learning_rate": 0.00015136876006441224,
"loss": 1.0496,
"step": 470
},
{
"epoch": 1.1459589867310012,
"grad_norm": 0.49609375,
"learning_rate": 0.00015297906602254428,
"loss": 1.0491,
"step": 475
},
{
"epoch": 1.158021712907117,
"grad_norm": 0.625,
"learning_rate": 0.00015458937198067633,
"loss": 1.0455,
"step": 480
},
{
"epoch": 1.1700844390832328,
"grad_norm": 0.546875,
"learning_rate": 0.00015619967793880838,
"loss": 1.0366,
"step": 485
},
{
"epoch": 1.1821471652593487,
"grad_norm": 0.59765625,
"learning_rate": 0.00015780998389694042,
"loss": 1.0428,
"step": 490
},
{
"epoch": 1.1942098914354644,
"grad_norm": 1.4140625,
"learning_rate": 0.00015942028985507247,
"loss": 1.0449,
"step": 495
},
{
"epoch": 1.2062726176115803,
"grad_norm": 0.6953125,
"learning_rate": 0.00016103059581320451,
"loss": 1.0363,
"step": 500
},
{
"epoch": 1.218335343787696,
"grad_norm": 0.7890625,
"learning_rate": 0.00016264090177133656,
"loss": 1.0377,
"step": 505
},
{
"epoch": 1.2303980699638117,
"grad_norm": 0.62890625,
"learning_rate": 0.0001642512077294686,
"loss": 1.0404,
"step": 510
},
{
"epoch": 1.2424607961399277,
"grad_norm": 0.69921875,
"learning_rate": 0.00016586151368760065,
"loss": 1.0448,
"step": 515
},
{
"epoch": 1.2545235223160434,
"grad_norm": 1.28125,
"learning_rate": 0.0001674718196457327,
"loss": 1.0441,
"step": 520
},
{
"epoch": 1.2665862484921593,
"grad_norm": 0.68359375,
"learning_rate": 0.00016908212560386474,
"loss": 1.0349,
"step": 525
},
{
"epoch": 1.278648974668275,
"grad_norm": 1.203125,
"learning_rate": 0.0001706924315619968,
"loss": 1.0504,
"step": 530
},
{
"epoch": 1.290711700844391,
"grad_norm": 0.8046875,
"learning_rate": 0.00017230273752012884,
"loss": 1.0362,
"step": 535
},
{
"epoch": 1.3027744270205066,
"grad_norm": 0.80859375,
"learning_rate": 0.00017391304347826088,
"loss": 1.0454,
"step": 540
},
{
"epoch": 1.3148371531966223,
"grad_norm": 1.0078125,
"learning_rate": 0.00017552334943639293,
"loss": 1.0466,
"step": 545
},
{
"epoch": 1.3268998793727382,
"grad_norm": 0.6171875,
"learning_rate": 0.00017713365539452497,
"loss": 1.0319,
"step": 550
},
{
"epoch": 1.3389626055488542,
"grad_norm": 0.3125,
"learning_rate": 0.00017874396135265702,
"loss": 1.038,
"step": 555
},
{
"epoch": 1.3510253317249699,
"grad_norm": 0.349609375,
"learning_rate": 0.00018035426731078907,
"loss": 1.0181,
"step": 560
},
{
"epoch": 1.3630880579010856,
"grad_norm": 0.34375,
"learning_rate": 0.0001819645732689211,
"loss": 1.0236,
"step": 565
},
{
"epoch": 1.3751507840772015,
"grad_norm": 0.36328125,
"learning_rate": 0.00018357487922705313,
"loss": 1.0274,
"step": 570
},
{
"epoch": 1.3872135102533172,
"grad_norm": 0.40234375,
"learning_rate": 0.0001851851851851852,
"loss": 1.0281,
"step": 575
},
{
"epoch": 1.399276236429433,
"grad_norm": 0.4765625,
"learning_rate": 0.00018679549114331725,
"loss": 1.0221,
"step": 580
},
{
"epoch": 1.4113389626055488,
"grad_norm": 0.474609375,
"learning_rate": 0.00018840579710144927,
"loss": 1.0354,
"step": 585
},
{
"epoch": 1.4234016887816647,
"grad_norm": 0.3671875,
"learning_rate": 0.00019001610305958134,
"loss": 1.0257,
"step": 590
},
{
"epoch": 1.4354644149577804,
"grad_norm": 0.3515625,
"learning_rate": 0.0001916264090177134,
"loss": 1.0246,
"step": 595
},
{
"epoch": 1.4475271411338961,
"grad_norm": 0.439453125,
"learning_rate": 0.0001932367149758454,
"loss": 1.0234,
"step": 600
},
{
"epoch": 1.459589867310012,
"grad_norm": 0.39453125,
"learning_rate": 0.00019484702093397745,
"loss": 1.0161,
"step": 605
},
{
"epoch": 1.471652593486128,
"grad_norm": 0.57421875,
"learning_rate": 0.00019645732689210952,
"loss": 1.017,
"step": 610
},
{
"epoch": 1.4837153196622437,
"grad_norm": 0.72265625,
"learning_rate": 0.00019806763285024154,
"loss": 1.0218,
"step": 615
},
{
"epoch": 1.4957780458383594,
"grad_norm": 0.609375,
"learning_rate": 0.0001996779388083736,
"loss": 1.0194,
"step": 620
},
{
"epoch": 1.5078407720144753,
"grad_norm": 0.515625,
"learning_rate": 0.00019999974723244348,
"loss": 1.0206,
"step": 625
},
{
"epoch": 1.5199034981905912,
"grad_norm": 0.3828125,
"learning_rate": 0.00019999872036643513,
"loss": 1.025,
"step": 630
},
{
"epoch": 1.531966224366707,
"grad_norm": 0.66015625,
"learning_rate": 0.0001999969036121076,
"loss": 1.0246,
"step": 635
},
{
"epoch": 1.5440289505428226,
"grad_norm": 0.93359375,
"learning_rate": 0.00019999429698381143,
"loss": 1.0232,
"step": 640
},
{
"epoch": 1.5560916767189386,
"grad_norm": 0.70703125,
"learning_rate": 0.00019999090050213636,
"loss": 1.0167,
"step": 645
},
{
"epoch": 1.5681544028950543,
"grad_norm": 0.62109375,
"learning_rate": 0.00019998671419391108,
"loss": 1.0087,
"step": 650
},
{
"epoch": 1.58021712907117,
"grad_norm": 0.357421875,
"learning_rate": 0.00019998173809220325,
"loss": 1.0215,
"step": 655
},
{
"epoch": 1.5922798552472859,
"grad_norm": 0.34765625,
"learning_rate": 0.00019997597223631895,
"loss": 1.0208,
"step": 660
},
{
"epoch": 1.6043425814234018,
"grad_norm": 0.390625,
"learning_rate": 0.0001999694166718026,
"loss": 1.01,
"step": 665
},
{
"epoch": 1.6164053075995175,
"grad_norm": 0.37109375,
"learning_rate": 0.00019996207145043653,
"loss": 1.0138,
"step": 670
},
{
"epoch": 1.6284680337756332,
"grad_norm": 0.375,
"learning_rate": 0.00019995393663024054,
"loss": 1.0069,
"step": 675
},
{
"epoch": 1.6405307599517491,
"grad_norm": 0.76171875,
"learning_rate": 0.00019994501227547137,
"loss": 1.0115,
"step": 680
},
{
"epoch": 1.652593486127865,
"grad_norm": 0.34375,
"learning_rate": 0.00019993529845662244,
"loss": 1.0071,
"step": 685
},
{
"epoch": 1.6646562123039808,
"grad_norm": 0.515625,
"learning_rate": 0.00019992479525042303,
"loss": 1.0058,
"step": 690
},
{
"epoch": 1.6767189384800965,
"grad_norm": 0.6171875,
"learning_rate": 0.00019991350273983776,
"loss": 1.006,
"step": 695
},
{
"epoch": 1.6887816646562124,
"grad_norm": 0.51171875,
"learning_rate": 0.00019990142101406603,
"loss": 1.0051,
"step": 700
},
{
"epoch": 1.700844390832328,
"grad_norm": 0.466796875,
"learning_rate": 0.0001998885501685412,
"loss": 1.0037,
"step": 705
},
{
"epoch": 1.7129071170084438,
"grad_norm": 0.5546875,
"learning_rate": 0.0001998748903049299,
"loss": 1.0072,
"step": 710
},
{
"epoch": 1.7249698431845597,
"grad_norm": 0.326171875,
"learning_rate": 0.0001998604415311311,
"loss": 1.0173,
"step": 715
},
{
"epoch": 1.7370325693606756,
"grad_norm": 0.7109375,
"learning_rate": 0.00019984520396127553,
"loss": 1.0094,
"step": 720
},
{
"epoch": 1.7490952955367913,
"grad_norm": 0.54296875,
"learning_rate": 0.00019982917771572443,
"loss": 1.0048,
"step": 725
},
{
"epoch": 1.761158021712907,
"grad_norm": 0.34375,
"learning_rate": 0.00019981236292106896,
"loss": 1.0063,
"step": 730
},
{
"epoch": 1.773220747889023,
"grad_norm": 0.3515625,
"learning_rate": 0.0001997947597101288,
"loss": 1.0042,
"step": 735
},
{
"epoch": 1.7852834740651389,
"grad_norm": 0.296875,
"learning_rate": 0.00019977636822195153,
"loss": 0.9947,
"step": 740
},
{
"epoch": 1.7973462002412546,
"grad_norm": 0.50390625,
"learning_rate": 0.0001997571886018112,
"loss": 1.0042,
"step": 745
},
{
"epoch": 1.8094089264173703,
"grad_norm": 0.578125,
"learning_rate": 0.0001997372210012073,
"loss": 1.0002,
"step": 750
},
{
"epoch": 1.8214716525934862,
"grad_norm": 0.5625,
"learning_rate": 0.00019971646557786363,
"loss": 1.0061,
"step": 755
},
{
"epoch": 1.833534378769602,
"grad_norm": 0.408203125,
"learning_rate": 0.00019969492249572693,
"loss": 1.0024,
"step": 760
},
{
"epoch": 1.8455971049457176,
"grad_norm": 0.40625,
"learning_rate": 0.0001996725919249657,
"loss": 1.001,
"step": 765
},
{
"epoch": 1.8576598311218335,
"grad_norm": 0.396484375,
"learning_rate": 0.0001996494740419687,
"loss": 1.0007,
"step": 770
},
{
"epoch": 1.8697225572979495,
"grad_norm": 0.34375,
"learning_rate": 0.00019962556902934379,
"loss": 0.998,
"step": 775
},
{
"epoch": 1.8817852834740652,
"grad_norm": 0.64453125,
"learning_rate": 0.00019960087707591626,
"loss": 0.9913,
"step": 780
},
{
"epoch": 1.8938480096501809,
"grad_norm": 1.140625,
"learning_rate": 0.00019957539837672745,
"loss": 1.0082,
"step": 785
},
{
"epoch": 1.9059107358262968,
"grad_norm": 0.62890625,
"learning_rate": 0.00019954913313303323,
"loss": 1.0001,
"step": 790
},
{
"epoch": 1.9179734620024127,
"grad_norm": 0.68359375,
"learning_rate": 0.00019952208155230234,
"loss": 1.0091,
"step": 795
},
{
"epoch": 1.9300361881785284,
"grad_norm": 0.5859375,
"learning_rate": 0.0001994942438482148,
"loss": 0.9919,
"step": 800
},
{
"epoch": 1.942098914354644,
"grad_norm": 0.419921875,
"learning_rate": 0.00019946562024066014,
"loss": 0.982,
"step": 805
},
{
"epoch": 1.95416164053076,
"grad_norm": 0.46484375,
"learning_rate": 0.00019943621095573586,
"loss": 1.0035,
"step": 810
},
{
"epoch": 1.9662243667068757,
"grad_norm": 0.373046875,
"learning_rate": 0.00019940601622574542,
"loss": 0.9854,
"step": 815
},
{
"epoch": 1.9782870928829914,
"grad_norm": 0.49609375,
"learning_rate": 0.00019937503628919648,
"loss": 0.9947,
"step": 820
},
{
"epoch": 1.9903498190591074,
"grad_norm": 0.546875,
"learning_rate": 0.00019934327139079915,
"loss": 1.0004,
"step": 825
},
{
"epoch": 2.0,
"eval_loss": 2.448092460632324,
"eval_runtime": 0.534,
"eval_samples_per_second": 18.726,
"eval_steps_per_second": 1.873,
"step": 829
},
{
"epoch": 2.0024125452352233,
"grad_norm": 0.57421875,
"learning_rate": 0.00019931072178146374,
"loss": 0.9899,
"step": 830
},
{
"epoch": 2.0144752714113388,
"grad_norm": 0.423828125,
"learning_rate": 0.0001992773877182992,
"loss": 0.9899,
"step": 835
},
{
"epoch": 2.0265379975874547,
"grad_norm": 0.30859375,
"learning_rate": 0.00019924326946461074,
"loss": 0.9803,
"step": 840
},
{
"epoch": 2.0386007237635706,
"grad_norm": 0.291015625,
"learning_rate": 0.00019920836728989794,
"loss": 0.9873,
"step": 845
},
{
"epoch": 2.0506634499396865,
"grad_norm": 0.3203125,
"learning_rate": 0.00019917268146985245,
"loss": 0.9799,
"step": 850
},
{
"epoch": 2.062726176115802,
"grad_norm": 0.3515625,
"learning_rate": 0.0001991362122863561,
"loss": 0.9773,
"step": 855
},
{
"epoch": 2.074788902291918,
"grad_norm": 0.412109375,
"learning_rate": 0.00019909896002747831,
"loss": 0.9818,
"step": 860
},
{
"epoch": 2.086851628468034,
"grad_norm": 0.318359375,
"learning_rate": 0.00019906092498747416,
"loss": 0.9779,
"step": 865
},
{
"epoch": 2.0989143546441498,
"grad_norm": 0.478515625,
"learning_rate": 0.0001990221074667818,
"loss": 0.9807,
"step": 870
},
{
"epoch": 2.1109770808202653,
"grad_norm": 0.408203125,
"learning_rate": 0.00019898250777202018,
"loss": 0.9841,
"step": 875
},
{
"epoch": 2.123039806996381,
"grad_norm": 0.470703125,
"learning_rate": 0.0001989421262159867,
"loss": 0.9812,
"step": 880
},
{
"epoch": 2.135102533172497,
"grad_norm": 0.65625,
"learning_rate": 0.00019890096311765465,
"loss": 0.9709,
"step": 885
},
{
"epoch": 2.147165259348613,
"grad_norm": 0.349609375,
"learning_rate": 0.0001988590188021707,
"loss": 0.981,
"step": 890
},
{
"epoch": 2.1592279855247285,
"grad_norm": 0.47265625,
"learning_rate": 0.0001988162936008523,
"loss": 0.9715,
"step": 895
},
{
"epoch": 2.1712907117008444,
"grad_norm": 0.34765625,
"learning_rate": 0.00019877278785118517,
"loss": 0.9842,
"step": 900
},
{
"epoch": 2.1833534378769603,
"grad_norm": 0.333984375,
"learning_rate": 0.00019872850189682053,
"loss": 0.9776,
"step": 905
},
{
"epoch": 2.195416164053076,
"grad_norm": 0.341796875,
"learning_rate": 0.00019868343608757245,
"loss": 0.9921,
"step": 910
},
{
"epoch": 2.2074788902291917,
"grad_norm": 0.341796875,
"learning_rate": 0.00019863759077941504,
"loss": 0.9835,
"step": 915
},
{
"epoch": 2.2195416164053077,
"grad_norm": 0.48828125,
"learning_rate": 0.00019859096633447965,
"loss": 0.9828,
"step": 920
},
{
"epoch": 2.2316043425814236,
"grad_norm": 0.4765625,
"learning_rate": 0.00019854356312105207,
"loss": 0.9767,
"step": 925
},
{
"epoch": 2.243667068757539,
"grad_norm": 0.423828125,
"learning_rate": 0.00019849538151356955,
"loss": 0.982,
"step": 930
},
{
"epoch": 2.255729794933655,
"grad_norm": 0.384765625,
"learning_rate": 0.00019844642189261778,
"loss": 0.9848,
"step": 935
},
{
"epoch": 2.267792521109771,
"grad_norm": 0.32421875,
"learning_rate": 0.00019839668464492811,
"loss": 0.9735,
"step": 940
},
{
"epoch": 2.2798552472858864,
"grad_norm": 0.365234375,
"learning_rate": 0.0001983461701633742,
"loss": 0.977,
"step": 945
},
{
"epoch": 2.2919179734620023,
"grad_norm": 0.46875,
"learning_rate": 0.00019829487884696925,
"loss": 0.9785,
"step": 950
},
{
"epoch": 2.3039806996381182,
"grad_norm": 0.341796875,
"learning_rate": 0.00019824281110086248,
"loss": 0.9816,
"step": 955
},
{
"epoch": 2.316043425814234,
"grad_norm": 0.416015625,
"learning_rate": 0.00019818996733633618,
"loss": 0.988,
"step": 960
},
{
"epoch": 2.3281061519903496,
"grad_norm": 0.341796875,
"learning_rate": 0.0001981363479708025,
"loss": 0.9873,
"step": 965
},
{
"epoch": 2.3401688781664656,
"grad_norm": 0.349609375,
"learning_rate": 0.00019808195342779992,
"loss": 0.9812,
"step": 970
},
{
"epoch": 2.3522316043425815,
"grad_norm": 0.396484375,
"learning_rate": 0.00019802678413699006,
"loss": 0.9688,
"step": 975
},
{
"epoch": 2.3642943305186974,
"grad_norm": 0.333984375,
"learning_rate": 0.0001979708405341543,
"loss": 0.9744,
"step": 980
},
{
"epoch": 2.376357056694813,
"grad_norm": 0.36328125,
"learning_rate": 0.00019791412306119027,
"loss": 0.9682,
"step": 985
},
{
"epoch": 2.388419782870929,
"grad_norm": 0.3515625,
"learning_rate": 0.00019785663216610844,
"loss": 0.9854,
"step": 990
},
{
"epoch": 2.4004825090470447,
"grad_norm": 0.318359375,
"learning_rate": 0.00019779836830302843,
"loss": 0.9793,
"step": 995
},
{
"epoch": 2.4125452352231607,
"grad_norm": 0.359375,
"learning_rate": 0.0001977393319321756,
"loss": 0.9852,
"step": 1000
},
{
"epoch": 2.424607961399276,
"grad_norm": 0.58203125,
"learning_rate": 0.0001976795235198773,
"loss": 0.9759,
"step": 1005
},
{
"epoch": 2.436670687575392,
"grad_norm": 0.42578125,
"learning_rate": 0.00019761894353855926,
"loss": 0.9689,
"step": 1010
},
{
"epoch": 2.448733413751508,
"grad_norm": 0.55859375,
"learning_rate": 0.0001975575924667418,
"loss": 0.9717,
"step": 1015
},
{
"epoch": 2.4607961399276235,
"grad_norm": 0.37890625,
"learning_rate": 0.00019749547078903604,
"loss": 0.9769,
"step": 1020
},
{
"epoch": 2.4728588661037394,
"grad_norm": 0.451171875,
"learning_rate": 0.00019743257899614017,
"loss": 0.9699,
"step": 1025
},
{
"epoch": 2.4849215922798553,
"grad_norm": 0.330078125,
"learning_rate": 0.00019736891758483544,
"loss": 0.9818,
"step": 1030
},
{
"epoch": 2.4969843184559712,
"grad_norm": 0.37109375,
"learning_rate": 0.00019730448705798239,
"loss": 0.9716,
"step": 1035
},
{
"epoch": 2.5090470446320867,
"grad_norm": 0.34375,
"learning_rate": 0.00019723928792451668,
"loss": 0.963,
"step": 1040
},
{
"epoch": 2.5211097708082026,
"grad_norm": 0.333984375,
"learning_rate": 0.00019717332069944528,
"loss": 0.9629,
"step": 1045
},
{
"epoch": 2.5331724969843186,
"grad_norm": 0.353515625,
"learning_rate": 0.00019710658590384227,
"loss": 0.9741,
"step": 1050
},
{
"epoch": 2.545235223160434,
"grad_norm": 0.59375,
"learning_rate": 0.00019703908406484467,
"loss": 0.9674,
"step": 1055
},
{
"epoch": 2.55729794933655,
"grad_norm": 0.392578125,
"learning_rate": 0.00019697081571564854,
"loss": 0.9714,
"step": 1060
},
{
"epoch": 2.569360675512666,
"grad_norm": 0.39453125,
"learning_rate": 0.00019690178139550443,
"loss": 0.9656,
"step": 1065
},
{
"epoch": 2.581423401688782,
"grad_norm": 0.35546875,
"learning_rate": 0.0001968319816497134,
"loss": 0.968,
"step": 1070
},
{
"epoch": 2.5934861278648977,
"grad_norm": 0.53515625,
"learning_rate": 0.00019676141702962253,
"loss": 0.9676,
"step": 1075
},
{
"epoch": 2.605548854041013,
"grad_norm": 0.359375,
"learning_rate": 0.00019669008809262062,
"loss": 0.9733,
"step": 1080
},
{
"epoch": 2.617611580217129,
"grad_norm": 0.486328125,
"learning_rate": 0.00019661799540213382,
"loss": 0.967,
"step": 1085
},
{
"epoch": 2.6296743063932446,
"grad_norm": 0.4375,
"learning_rate": 0.0001965451395276212,
"loss": 0.9751,
"step": 1090
},
{
"epoch": 2.6417370325693605,
"grad_norm": 0.62109375,
"learning_rate": 0.00019647152104457013,
"loss": 0.9745,
"step": 1095
},
{
"epoch": 2.6537997587454765,
"grad_norm": 0.39453125,
"learning_rate": 0.0001963971405344919,
"loss": 0.9751,
"step": 1100
},
{
"epoch": 2.6658624849215924,
"grad_norm": 0.318359375,
"learning_rate": 0.0001963219985849169,
"loss": 0.9724,
"step": 1105
},
{
"epoch": 2.6779252110977083,
"grad_norm": 0.376953125,
"learning_rate": 0.00019624609578939027,
"loss": 0.973,
"step": 1110
},
{
"epoch": 2.689987937273824,
"grad_norm": 0.37890625,
"learning_rate": 0.000196169432747467,
"loss": 0.9771,
"step": 1115
},
{
"epoch": 2.7020506634499397,
"grad_norm": 0.46484375,
"learning_rate": 0.00019609201006470724,
"loss": 0.9671,
"step": 1120
},
{
"epoch": 2.7141133896260556,
"grad_norm": 0.443359375,
"learning_rate": 0.0001960138283526715,
"loss": 0.9698,
"step": 1125
},
{
"epoch": 2.726176115802171,
"grad_norm": 0.3359375,
"learning_rate": 0.000195934888228916,
"loss": 0.9645,
"step": 1130
},
{
"epoch": 2.738238841978287,
"grad_norm": 0.435546875,
"learning_rate": 0.0001958551903169874,
"loss": 0.9665,
"step": 1135
},
{
"epoch": 2.750301568154403,
"grad_norm": 0.365234375,
"learning_rate": 0.0001957747352464184,
"loss": 0.9719,
"step": 1140
},
{
"epoch": 2.762364294330519,
"grad_norm": 0.4140625,
"learning_rate": 0.0001956935236527222,
"loss": 0.9624,
"step": 1145
},
{
"epoch": 2.7744270205066344,
"grad_norm": 0.357421875,
"learning_rate": 0.00019561155617738797,
"loss": 0.9686,
"step": 1150
},
{
"epoch": 2.7864897466827503,
"grad_norm": 0.421875,
"learning_rate": 0.00019552883346787552,
"loss": 0.962,
"step": 1155
},
{
"epoch": 2.798552472858866,
"grad_norm": 0.423828125,
"learning_rate": 0.0001954453561776102,
"loss": 0.9684,
"step": 1160
},
{
"epoch": 2.8106151990349817,
"grad_norm": 0.33984375,
"learning_rate": 0.00019536112496597782,
"loss": 0.9759,
"step": 1165
},
{
"epoch": 2.8226779252110976,
"grad_norm": 0.314453125,
"learning_rate": 0.0001952761404983194,
"loss": 0.9744,
"step": 1170
},
{
"epoch": 2.8347406513872135,
"grad_norm": 0.37890625,
"learning_rate": 0.00019519040344592593,
"loss": 0.9714,
"step": 1175
},
{
"epoch": 2.8468033775633295,
"grad_norm": 0.388671875,
"learning_rate": 0.000195103914486033,
"loss": 0.965,
"step": 1180
},
{
"epoch": 2.8588661037394454,
"grad_norm": 0.55078125,
"learning_rate": 0.0001950166743018156,
"loss": 0.9572,
"step": 1185
},
{
"epoch": 2.870928829915561,
"grad_norm": 0.44140625,
"learning_rate": 0.00019492868358238249,
"loss": 0.9563,
"step": 1190
},
{
"epoch": 2.882991556091677,
"grad_norm": 0.326171875,
"learning_rate": 0.00019483994302277102,
"loss": 0.9597,
"step": 1195
},
{
"epoch": 2.8950542822677923,
"grad_norm": 0.42578125,
"learning_rate": 0.00019475045332394153,
"loss": 0.965,
"step": 1200
},
{
"epoch": 2.907117008443908,
"grad_norm": 0.443359375,
"learning_rate": 0.00019466021519277165,
"loss": 0.9562,
"step": 1205
},
{
"epoch": 2.919179734620024,
"grad_norm": 0.322265625,
"learning_rate": 0.000194569229342051,
"loss": 0.9573,
"step": 1210
},
{
"epoch": 2.93124246079614,
"grad_norm": 0.474609375,
"learning_rate": 0.00019447749649047542,
"loss": 0.9629,
"step": 1215
},
{
"epoch": 2.943305186972256,
"grad_norm": 0.6015625,
"learning_rate": 0.0001943850173626412,
"loss": 0.978,
"step": 1220
},
{
"epoch": 2.9553679131483714,
"grad_norm": 0.50390625,
"learning_rate": 0.00019429179268903957,
"loss": 0.9696,
"step": 1225
},
{
"epoch": 2.9674306393244874,
"grad_norm": 0.3515625,
"learning_rate": 0.0001941978232060507,
"loss": 0.9682,
"step": 1230
},
{
"epoch": 2.9794933655006033,
"grad_norm": 0.3828125,
"learning_rate": 0.00019410310965593813,
"loss": 0.9498,
"step": 1235
},
{
"epoch": 2.9915560916767188,
"grad_norm": 0.341796875,
"learning_rate": 0.00019400765278684262,
"loss": 0.9586,
"step": 1240
},
{
"epoch": 2.9987937273823886,
"eval_loss": 2.4426419734954834,
"eval_runtime": 0.66,
"eval_samples_per_second": 15.151,
"eval_steps_per_second": 1.515,
"step": 1243
},
{
"epoch": 3.0036188178528347,
"grad_norm": 0.41015625,
"learning_rate": 0.00019391145335277655,
"loss": 0.9537,
"step": 1245
},
{
"epoch": 3.0156815440289506,
"grad_norm": 0.423828125,
"learning_rate": 0.00019381451211361773,
"loss": 0.9384,
"step": 1250
},
{
"epoch": 3.0277442702050665,
"grad_norm": 0.34765625,
"learning_rate": 0.00019371682983510344,
"loss": 0.9493,
"step": 1255
},
{
"epoch": 3.039806996381182,
"grad_norm": 0.314453125,
"learning_rate": 0.00019361840728882447,
"loss": 0.9441,
"step": 1260
},
{
"epoch": 3.051869722557298,
"grad_norm": 0.322265625,
"learning_rate": 0.000193519245252219,
"loss": 0.9457,
"step": 1265
},
{
"epoch": 3.063932448733414,
"grad_norm": 0.318359375,
"learning_rate": 0.00019341934450856633,
"loss": 0.9498,
"step": 1270
},
{
"epoch": 3.0759951749095293,
"grad_norm": 0.30078125,
"learning_rate": 0.00019331870584698093,
"loss": 0.9539,
"step": 1275
},
{
"epoch": 3.0880579010856453,
"grad_norm": 0.39453125,
"learning_rate": 0.000193217330062406,
"loss": 0.9564,
"step": 1280
},
{
"epoch": 3.100120627261761,
"grad_norm": 0.296875,
"learning_rate": 0.00019311521795560723,
"loss": 0.9492,
"step": 1285
},
{
"epoch": 3.112183353437877,
"grad_norm": 0.359375,
"learning_rate": 0.00019301237033316659,
"loss": 0.9549,
"step": 1290
},
{
"epoch": 3.1242460796139926,
"grad_norm": 0.38671875,
"learning_rate": 0.00019290878800747578,
"loss": 0.9511,
"step": 1295
},
{
"epoch": 3.1363088057901085,
"grad_norm": 0.3359375,
"learning_rate": 0.00019280447179673,
"loss": 0.9395,
"step": 1300
},
{
"epoch": 3.1483715319662244,
"grad_norm": 0.388671875,
"learning_rate": 0.00019269942252492133,
"loss": 0.951,
"step": 1305
},
{
"epoch": 3.1604342581423404,
"grad_norm": 0.3828125,
"learning_rate": 0.00019259364102183234,
"loss": 0.9471,
"step": 1310
},
{
"epoch": 3.172496984318456,
"grad_norm": 0.3515625,
"learning_rate": 0.00019248712812302947,
"loss": 0.9587,
"step": 1315
},
{
"epoch": 3.1845597104945718,
"grad_norm": 0.39453125,
"learning_rate": 0.0001923798846698564,
"loss": 0.9591,
"step": 1320
},
{
"epoch": 3.1966224366706877,
"grad_norm": 0.44140625,
"learning_rate": 0.0001922719115094275,
"loss": 0.9437,
"step": 1325
},
{
"epoch": 3.2086851628468036,
"grad_norm": 0.6171875,
"learning_rate": 0.00019216320949462103,
"loss": 0.9494,
"step": 1330
},
{
"epoch": 3.220747889022919,
"grad_norm": 0.8125,
"learning_rate": 0.00019205377948407258,
"loss": 0.9532,
"step": 1335
},
{
"epoch": 3.232810615199035,
"grad_norm": 0.361328125,
"learning_rate": 0.00019194362234216795,
"loss": 0.9568,
"step": 1340
},
{
"epoch": 3.244873341375151,
"grad_norm": 0.94921875,
"learning_rate": 0.0001918327389390368,
"loss": 0.9531,
"step": 1345
},
{
"epoch": 3.2569360675512664,
"grad_norm": 0.384765625,
"learning_rate": 0.00019172113015054532,
"loss": 0.9561,
"step": 1350
},
{
"epoch": 3.2689987937273823,
"grad_norm": 0.5703125,
"learning_rate": 0.0001916087968582896,
"loss": 0.9554,
"step": 1355
},
{
"epoch": 3.2810615199034983,
"grad_norm": 0.419921875,
"learning_rate": 0.0001914957399495885,
"loss": 0.95,
"step": 1360
},
{
"epoch": 3.293124246079614,
"grad_norm": 0.64453125,
"learning_rate": 0.00019138196031747681,
"loss": 0.9599,
"step": 1365
},
{
"epoch": 3.3051869722557297,
"grad_norm": 0.5859375,
"learning_rate": 0.00019126745886069807,
"loss": 0.9546,
"step": 1370
},
{
"epoch": 3.3172496984318456,
"grad_norm": 0.59375,
"learning_rate": 0.0001911522364836974,
"loss": 0.9521,
"step": 1375
},
{
"epoch": 3.3293124246079615,
"grad_norm": 0.365234375,
"learning_rate": 0.0001910362940966147,
"loss": 0.9571,
"step": 1380
},
{
"epoch": 3.341375150784077,
"grad_norm": 0.5078125,
"learning_rate": 0.0001909196326152769,
"loss": 0.9514,
"step": 1385
},
{
"epoch": 3.353437876960193,
"grad_norm": 0.66796875,
"learning_rate": 0.00019080225296119125,
"loss": 0.9482,
"step": 1390
},
{
"epoch": 3.365500603136309,
"grad_norm": 0.494140625,
"learning_rate": 0.00019068415606153787,
"loss": 0.9489,
"step": 1395
},
{
"epoch": 3.3775633293124248,
"grad_norm": 0.462890625,
"learning_rate": 0.00019056534284916224,
"loss": 0.944,
"step": 1400
},
{
"epoch": 3.3896260554885402,
"grad_norm": 0.40625,
"learning_rate": 0.00019044581426256804,
"loss": 0.9536,
"step": 1405
},
{
"epoch": 3.401688781664656,
"grad_norm": 0.318359375,
"learning_rate": 0.00019032557124590974,
"loss": 0.9492,
"step": 1410
},
{
"epoch": 3.413751507840772,
"grad_norm": 0.3359375,
"learning_rate": 0.00019020461474898503,
"loss": 0.945,
"step": 1415
},
{
"epoch": 3.425814234016888,
"grad_norm": 0.330078125,
"learning_rate": 0.00019008294572722738,
"loss": 0.955,
"step": 1420
},
{
"epoch": 3.4378769601930035,
"grad_norm": 0.337890625,
"learning_rate": 0.00018996056514169844,
"loss": 0.9466,
"step": 1425
},
{
"epoch": 3.4499396863691194,
"grad_norm": 0.330078125,
"learning_rate": 0.00018983747395908058,
"loss": 0.9436,
"step": 1430
},
{
"epoch": 3.4620024125452353,
"grad_norm": 0.34765625,
"learning_rate": 0.0001897136731516691,
"loss": 0.949,
"step": 1435
},
{
"epoch": 3.4740651387213513,
"grad_norm": 0.318359375,
"learning_rate": 0.0001895891636973646,
"loss": 0.9629,
"step": 1440
},
{
"epoch": 3.4861278648974667,
"grad_norm": 0.359375,
"learning_rate": 0.00018946394657966536,
"loss": 0.9439,
"step": 1445
},
{
"epoch": 3.4981905910735827,
"grad_norm": 0.318359375,
"learning_rate": 0.00018933802278765934,
"loss": 0.9455,
"step": 1450
},
{
"epoch": 3.5102533172496986,
"grad_norm": 0.365234375,
"learning_rate": 0.00018921139331601667,
"loss": 0.9462,
"step": 1455
},
{
"epoch": 3.522316043425814,
"grad_norm": 0.328125,
"learning_rate": 0.00018908405916498154,
"loss": 0.9499,
"step": 1460
},
{
"epoch": 3.53437876960193,
"grad_norm": 0.484375,
"learning_rate": 0.00018895602134036434,
"loss": 0.9433,
"step": 1465
},
{
"epoch": 3.546441495778046,
"grad_norm": 0.5078125,
"learning_rate": 0.00018882728085353392,
"loss": 0.9431,
"step": 1470
},
{
"epoch": 3.558504221954162,
"grad_norm": 0.45703125,
"learning_rate": 0.0001886978387214093,
"loss": 0.9494,
"step": 1475
},
{
"epoch": 3.5705669481302773,
"grad_norm": 0.484375,
"learning_rate": 0.0001885676959664519,
"loss": 0.9515,
"step": 1480
},
{
"epoch": 3.5826296743063932,
"grad_norm": 0.5546875,
"learning_rate": 0.00018843685361665723,
"loss": 0.9413,
"step": 1485
},
{
"epoch": 3.594692400482509,
"grad_norm": 0.6171875,
"learning_rate": 0.000188305312705547,
"loss": 0.9398,
"step": 1490
},
{
"epoch": 3.6067551266586246,
"grad_norm": 0.380859375,
"learning_rate": 0.0001881730742721608,
"loss": 0.9314,
"step": 1495
},
{
"epoch": 3.6188178528347406,
"grad_norm": 0.53515625,
"learning_rate": 0.00018804013936104792,
"loss": 0.9364,
"step": 1500
},
{
"epoch": 3.6308805790108565,
"grad_norm": 0.3203125,
"learning_rate": 0.0001879065090222591,
"loss": 0.9502,
"step": 1505
},
{
"epoch": 3.6429433051869724,
"grad_norm": 0.37890625,
"learning_rate": 0.00018777218431133833,
"loss": 0.9477,
"step": 1510
},
{
"epoch": 3.6550060313630883,
"grad_norm": 0.34765625,
"learning_rate": 0.00018763716628931437,
"loss": 0.9462,
"step": 1515
},
{
"epoch": 3.667068757539204,
"grad_norm": 0.431640625,
"learning_rate": 0.00018750145602269247,
"loss": 0.9442,
"step": 1520
},
{
"epoch": 3.6791314837153197,
"grad_norm": 0.306640625,
"learning_rate": 0.00018736505458344586,
"loss": 0.9378,
"step": 1525
},
{
"epoch": 3.691194209891435,
"grad_norm": 0.314453125,
"learning_rate": 0.0001872279630490074,
"loss": 0.9423,
"step": 1530
},
{
"epoch": 3.703256936067551,
"grad_norm": 0.298828125,
"learning_rate": 0.00018709018250226094,
"loss": 0.9389,
"step": 1535
},
{
"epoch": 3.715319662243667,
"grad_norm": 0.314453125,
"learning_rate": 0.00018695171403153285,
"loss": 0.9356,
"step": 1540
},
{
"epoch": 3.727382388419783,
"grad_norm": 0.52734375,
"learning_rate": 0.00018681255873058338,
"loss": 0.9447,
"step": 1545
},
{
"epoch": 3.739445114595899,
"grad_norm": 0.3359375,
"learning_rate": 0.00018667271769859812,
"loss": 0.9409,
"step": 1550
},
{
"epoch": 3.7515078407720144,
"grad_norm": 0.337890625,
"learning_rate": 0.00018653219204017916,
"loss": 0.9457,
"step": 1555
},
{
"epoch": 3.7635705669481303,
"grad_norm": 0.33984375,
"learning_rate": 0.00018639098286533644,
"loss": 0.954,
"step": 1560
},
{
"epoch": 3.7756332931242462,
"grad_norm": 0.375,
"learning_rate": 0.00018624909128947903,
"loss": 0.9467,
"step": 1565
},
{
"epoch": 3.7876960193003617,
"grad_norm": 0.55078125,
"learning_rate": 0.00018610651843340627,
"loss": 0.9436,
"step": 1570
},
{
"epoch": 3.7997587454764776,
"grad_norm": 0.392578125,
"learning_rate": 0.00018596326542329888,
"loss": 0.9381,
"step": 1575
},
{
"epoch": 3.8118214716525936,
"grad_norm": 0.41015625,
"learning_rate": 0.0001858193333907101,
"loss": 0.9408,
"step": 1580
},
{
"epoch": 3.8238841978287095,
"grad_norm": 0.326171875,
"learning_rate": 0.0001856747234725569,
"loss": 0.9456,
"step": 1585
},
{
"epoch": 3.835946924004825,
"grad_norm": 0.357421875,
"learning_rate": 0.00018552943681111067,
"loss": 0.9405,
"step": 1590
},
{
"epoch": 3.848009650180941,
"grad_norm": 0.384765625,
"learning_rate": 0.00018538347455398846,
"loss": 0.9551,
"step": 1595
},
{
"epoch": 3.860072376357057,
"grad_norm": 0.375,
"learning_rate": 0.00018523683785414387,
"loss": 0.9401,
"step": 1600
},
{
"epoch": 3.8721351025331723,
"grad_norm": 0.328125,
"learning_rate": 0.0001850895278698579,
"loss": 0.9374,
"step": 1605
},
{
"epoch": 3.884197828709288,
"grad_norm": 0.494140625,
"learning_rate": 0.00018494154576472976,
"loss": 0.937,
"step": 1610
},
{
"epoch": 3.896260554885404,
"grad_norm": 0.357421875,
"learning_rate": 0.00018479289270766777,
"loss": 0.9398,
"step": 1615
},
{
"epoch": 3.90832328106152,
"grad_norm": 0.298828125,
"learning_rate": 0.00018464356987288013,
"loss": 0.9402,
"step": 1620
},
{
"epoch": 3.920386007237636,
"grad_norm": 0.310546875,
"learning_rate": 0.0001844935784398655,
"loss": 0.9443,
"step": 1625
},
{
"epoch": 3.9324487334137515,
"grad_norm": 0.322265625,
"learning_rate": 0.00018434291959340394,
"loss": 0.9441,
"step": 1630
},
{
"epoch": 3.9445114595898674,
"grad_norm": 0.298828125,
"learning_rate": 0.0001841915945235472,
"loss": 0.939,
"step": 1635
},
{
"epoch": 3.956574185765983,
"grad_norm": 0.314453125,
"learning_rate": 0.00018403960442560973,
"loss": 0.94,
"step": 1640
},
{
"epoch": 3.9686369119420988,
"grad_norm": 0.3515625,
"learning_rate": 0.00018388695050015885,
"loss": 0.939,
"step": 1645
},
{
"epoch": 3.9806996381182147,
"grad_norm": 0.408203125,
"learning_rate": 0.00018373363395300554,
"loss": 0.9399,
"step": 1650
},
{
"epoch": 3.9927623642943306,
"grad_norm": 0.455078125,
"learning_rate": 0.0001835796559951948,
"loss": 0.9412,
"step": 1655
},
{
"epoch": 4.0,
"eval_loss": 2.449599504470825,
"eval_runtime": 0.535,
"eval_samples_per_second": 18.691,
"eval_steps_per_second": 1.869,
"step": 1658
},
{
"epoch": 4.0048250904704465,
"grad_norm": 0.353515625,
"learning_rate": 0.0001834250178429961,
"loss": 0.9353,
"step": 1660
},
{
"epoch": 4.0168878166465625,
"grad_norm": 0.390625,
"learning_rate": 0.0001832697207178938,
"loss": 0.9234,
"step": 1665
},
{
"epoch": 4.0289505428226775,
"grad_norm": 0.314453125,
"learning_rate": 0.00018311376584657736,
"loss": 0.9264,
"step": 1670
},
{
"epoch": 4.041013268998793,
"grad_norm": 0.4375,
"learning_rate": 0.00018295715446093192,
"loss": 0.9233,
"step": 1675
},
{
"epoch": 4.053075995174909,
"grad_norm": 0.318359375,
"learning_rate": 0.00018279988779802833,
"loss": 0.9299,
"step": 1680
},
{
"epoch": 4.065138721351025,
"grad_norm": 0.416015625,
"learning_rate": 0.00018264196710011352,
"loss": 0.9359,
"step": 1685
},
{
"epoch": 4.077201447527141,
"grad_norm": 0.37109375,
"learning_rate": 0.00018248339361460057,
"loss": 0.9266,
"step": 1690
},
{
"epoch": 4.089264173703257,
"grad_norm": 0.36328125,
"learning_rate": 0.00018232416859405895,
"loss": 0.935,
"step": 1695
},
{
"epoch": 4.101326899879373,
"grad_norm": 0.341796875,
"learning_rate": 0.00018216429329620456,
"loss": 0.9294,
"step": 1700
},
{
"epoch": 4.113389626055489,
"grad_norm": 0.359375,
"learning_rate": 0.00018200376898388996,
"loss": 0.9288,
"step": 1705
},
{
"epoch": 4.125452352231604,
"grad_norm": 0.330078125,
"learning_rate": 0.00018184259692509406,
"loss": 0.9264,
"step": 1710
},
{
"epoch": 4.13751507840772,
"grad_norm": 0.490234375,
"learning_rate": 0.00018168077839291247,
"loss": 0.9421,
"step": 1715
},
{
"epoch": 4.149577804583836,
"grad_norm": 0.67578125,
"learning_rate": 0.00018151831466554726,
"loss": 0.9355,
"step": 1720
},
{
"epoch": 4.161640530759952,
"grad_norm": 0.365234375,
"learning_rate": 0.00018135520702629675,
"loss": 0.9349,
"step": 1725
},
{
"epoch": 4.173703256936068,
"grad_norm": 0.458984375,
"learning_rate": 0.00018119145676354575,
"loss": 0.9345,
"step": 1730
},
{
"epoch": 4.185765983112184,
"grad_norm": 0.5625,
"learning_rate": 0.00018102706517075491,
"loss": 0.9319,
"step": 1735
},
{
"epoch": 4.1978287092882995,
"grad_norm": 0.404296875,
"learning_rate": 0.00018086203354645089,
"loss": 0.9309,
"step": 1740
},
{
"epoch": 4.209891435464415,
"grad_norm": 0.369140625,
"learning_rate": 0.00018069636319421588,
"loss": 0.9378,
"step": 1745
},
{
"epoch": 4.2219541616405305,
"grad_norm": 0.380859375,
"learning_rate": 0.00018053005542267736,
"loss": 0.9285,
"step": 1750
},
{
"epoch": 4.234016887816646,
"grad_norm": 0.326171875,
"learning_rate": 0.00018036311154549784,
"loss": 0.9314,
"step": 1755
},
{
"epoch": 4.246079613992762,
"grad_norm": 0.322265625,
"learning_rate": 0.00018019553288136435,
"loss": 0.9234,
"step": 1760
},
{
"epoch": 4.258142340168878,
"grad_norm": 0.41015625,
"learning_rate": 0.00018002732075397812,
"loss": 0.9286,
"step": 1765
},
{
"epoch": 4.270205066344994,
"grad_norm": 0.390625,
"learning_rate": 0.00017985847649204417,
"loss": 0.9267,
"step": 1770
},
{
"epoch": 4.28226779252111,
"grad_norm": 0.333984375,
"learning_rate": 0.00017968900142926057,
"loss": 0.9275,
"step": 1775
},
{
"epoch": 4.294330518697226,
"grad_norm": 0.37890625,
"learning_rate": 0.00017951889690430824,
"loss": 0.9312,
"step": 1780
},
{
"epoch": 4.306393244873341,
"grad_norm": 0.3984375,
"learning_rate": 0.00017934816426084008,
"loss": 0.9242,
"step": 1785
},
{
"epoch": 4.318455971049457,
"grad_norm": 0.333984375,
"learning_rate": 0.00017917680484747065,
"loss": 0.9297,
"step": 1790
},
{
"epoch": 4.330518697225573,
"grad_norm": 0.3046875,
"learning_rate": 0.00017900482001776517,
"loss": 0.9206,
"step": 1795
},
{
"epoch": 4.342581423401689,
"grad_norm": 0.333984375,
"learning_rate": 0.00017883221113022916,
"loss": 0.927,
"step": 1800
},
{
"epoch": 4.354644149577805,
"grad_norm": 0.337890625,
"learning_rate": 0.00017865897954829748,
"loss": 0.9324,
"step": 1805
},
{
"epoch": 4.366706875753921,
"grad_norm": 0.337890625,
"learning_rate": 0.0001784851266403237,
"loss": 0.9422,
"step": 1810
},
{
"epoch": 4.378769601930037,
"grad_norm": 0.353515625,
"learning_rate": 0.0001783106537795692,
"loss": 0.9326,
"step": 1815
},
{
"epoch": 4.390832328106152,
"grad_norm": 0.470703125,
"learning_rate": 0.00017813556234419234,
"loss": 0.9366,
"step": 1820
},
{
"epoch": 4.402895054282268,
"grad_norm": 0.31640625,
"learning_rate": 0.00017795985371723763,
"loss": 0.9328,
"step": 1825
},
{
"epoch": 4.4149577804583835,
"grad_norm": 0.326171875,
"learning_rate": 0.00017778352928662474,
"loss": 0.9288,
"step": 1830
},
{
"epoch": 4.427020506634499,
"grad_norm": 0.357421875,
"learning_rate": 0.00017760659044513757,
"loss": 0.9311,
"step": 1835
},
{
"epoch": 4.439083232810615,
"grad_norm": 0.36328125,
"learning_rate": 0.00017742903859041325,
"loss": 0.9263,
"step": 1840
},
{
"epoch": 4.451145958986731,
"grad_norm": 0.390625,
"learning_rate": 0.0001772508751249311,
"loss": 0.9248,
"step": 1845
},
{
"epoch": 4.463208685162847,
"grad_norm": 0.3203125,
"learning_rate": 0.0001770721014560015,
"loss": 0.9322,
"step": 1850
},
{
"epoch": 4.475271411338962,
"grad_norm": 0.306640625,
"learning_rate": 0.00017689271899575491,
"loss": 0.9284,
"step": 1855
},
{
"epoch": 4.487334137515078,
"grad_norm": 0.302734375,
"learning_rate": 0.00017671272916113052,
"loss": 0.9323,
"step": 1860
},
{
"epoch": 4.499396863691194,
"grad_norm": 0.30859375,
"learning_rate": 0.00017653213337386517,
"loss": 0.9302,
"step": 1865
},
{
"epoch": 4.51145958986731,
"grad_norm": 0.33984375,
"learning_rate": 0.0001763509330604822,
"loss": 0.9298,
"step": 1870
},
{
"epoch": 4.523522316043426,
"grad_norm": 0.34375,
"learning_rate": 0.00017616912965228001,
"loss": 0.9278,
"step": 1875
},
{
"epoch": 4.535585042219542,
"grad_norm": 0.341796875,
"learning_rate": 0.00017598672458532088,
"loss": 0.9329,
"step": 1880
},
{
"epoch": 4.547647768395658,
"grad_norm": 0.400390625,
"learning_rate": 0.00017580371930041953,
"loss": 0.9251,
"step": 1885
},
{
"epoch": 4.559710494571773,
"grad_norm": 0.34375,
"learning_rate": 0.00017562011524313185,
"loss": 0.9278,
"step": 1890
},
{
"epoch": 4.571773220747889,
"grad_norm": 0.392578125,
"learning_rate": 0.0001754359138637434,
"loss": 0.9164,
"step": 1895
},
{
"epoch": 4.583835946924005,
"grad_norm": 0.349609375,
"learning_rate": 0.00017525111661725797,
"loss": 0.9184,
"step": 1900
},
{
"epoch": 4.595898673100121,
"grad_norm": 0.328125,
"learning_rate": 0.0001750657249633861,
"loss": 0.9345,
"step": 1905
},
{
"epoch": 4.6079613992762365,
"grad_norm": 0.33984375,
"learning_rate": 0.00017487974036653361,
"loss": 0.9285,
"step": 1910
},
{
"epoch": 4.620024125452352,
"grad_norm": 0.52734375,
"learning_rate": 0.00017469316429578977,
"loss": 0.9219,
"step": 1915
},
{
"epoch": 4.632086851628468,
"grad_norm": 0.365234375,
"learning_rate": 0.00017450599822491615,
"loss": 0.9192,
"step": 1920
},
{
"epoch": 4.644149577804583,
"grad_norm": 0.447265625,
"learning_rate": 0.00017431824363233457,
"loss": 0.9301,
"step": 1925
},
{
"epoch": 4.656212303980699,
"grad_norm": 0.427734375,
"learning_rate": 0.00017412990200111556,
"loss": 0.9293,
"step": 1930
},
{
"epoch": 4.668275030156815,
"grad_norm": 0.34375,
"learning_rate": 0.00017394097481896676,
"loss": 0.9215,
"step": 1935
},
{
"epoch": 4.680337756332931,
"grad_norm": 0.30078125,
"learning_rate": 0.00017375146357822096,
"loss": 0.927,
"step": 1940
},
{
"epoch": 4.692400482509047,
"grad_norm": 0.33984375,
"learning_rate": 0.00017356136977582454,
"loss": 0.9251,
"step": 1945
},
{
"epoch": 4.704463208685163,
"grad_norm": 0.412109375,
"learning_rate": 0.00017337069491332537,
"loss": 0.9291,
"step": 1950
},
{
"epoch": 4.716525934861279,
"grad_norm": 0.3984375,
"learning_rate": 0.00017317944049686124,
"loss": 0.9288,
"step": 1955
},
{
"epoch": 4.728588661037395,
"grad_norm": 0.408203125,
"learning_rate": 0.00017298760803714775,
"loss": 0.9165,
"step": 1960
},
{
"epoch": 4.74065138721351,
"grad_norm": 0.55078125,
"learning_rate": 0.00017279519904946647,
"loss": 0.9309,
"step": 1965
},
{
"epoch": 4.752714113389626,
"grad_norm": 0.447265625,
"learning_rate": 0.00017260221505365303,
"loss": 0.919,
"step": 1970
},
{
"epoch": 4.764776839565742,
"grad_norm": 0.37109375,
"learning_rate": 0.00017240865757408495,
"loss": 0.9239,
"step": 1975
},
{
"epoch": 4.776839565741858,
"grad_norm": 0.30859375,
"learning_rate": 0.0001722145281396697,
"loss": 0.9296,
"step": 1980
},
{
"epoch": 4.788902291917974,
"grad_norm": 0.373046875,
"learning_rate": 0.00017201982828383264,
"loss": 0.9306,
"step": 1985
},
{
"epoch": 4.8009650180940895,
"grad_norm": 0.3359375,
"learning_rate": 0.00017182455954450486,
"loss": 0.9283,
"step": 1990
},
{
"epoch": 4.813027744270205,
"grad_norm": 0.4140625,
"learning_rate": 0.00017162872346411102,
"loss": 0.9237,
"step": 1995
},
{
"epoch": 4.825090470446321,
"grad_norm": 0.333984375,
"learning_rate": 0.00017143232158955732,
"loss": 0.9293,
"step": 2000
},
{
"epoch": 4.837153196622436,
"grad_norm": 0.40234375,
"learning_rate": 0.00017123535547221907,
"loss": 0.9295,
"step": 2005
},
{
"epoch": 4.849215922798552,
"grad_norm": 0.345703125,
"learning_rate": 0.00017103782666792844,
"loss": 0.9154,
"step": 2010
},
{
"epoch": 4.861278648974668,
"grad_norm": 0.30859375,
"learning_rate": 0.00017083973673696247,
"loss": 0.9207,
"step": 2015
},
{
"epoch": 4.873341375150784,
"grad_norm": 0.359375,
"learning_rate": 0.00017064108724403033,
"loss": 0.9272,
"step": 2020
},
{
"epoch": 4.8854041013269,
"grad_norm": 0.35546875,
"learning_rate": 0.00017044187975826124,
"loss": 0.9224,
"step": 2025
},
{
"epoch": 4.897466827503016,
"grad_norm": 0.408203125,
"learning_rate": 0.00017024211585319203,
"loss": 0.9241,
"step": 2030
},
{
"epoch": 4.909529553679132,
"grad_norm": 0.357421875,
"learning_rate": 0.00017004179710675463,
"loss": 0.933,
"step": 2035
},
{
"epoch": 4.921592279855247,
"grad_norm": 0.294921875,
"learning_rate": 0.00016984092510126367,
"loss": 0.916,
"step": 2040
},
{
"epoch": 4.933655006031363,
"grad_norm": 0.33203125,
"learning_rate": 0.00016963950142340396,
"loss": 0.9278,
"step": 2045
},
{
"epoch": 4.945717732207479,
"grad_norm": 0.33203125,
"learning_rate": 0.00016943752766421794,
"loss": 0.9284,
"step": 2050
},
{
"epoch": 4.957780458383595,
"grad_norm": 0.31640625,
"learning_rate": 0.0001692350054190932,
"loss": 0.9358,
"step": 2055
},
{
"epoch": 4.969843184559711,
"grad_norm": 0.515625,
"learning_rate": 0.00016903193628774977,
"loss": 0.925,
"step": 2060
},
{
"epoch": 4.981905910735827,
"grad_norm": 0.359375,
"learning_rate": 0.00016882832187422743,
"loss": 0.9235,
"step": 2065
},
{
"epoch": 4.9939686369119425,
"grad_norm": 0.333984375,
"learning_rate": 0.0001686241637868734,
"loss": 0.9325,
"step": 2070
},
{
"epoch": 4.998793727382388,
"eval_loss": 2.4599645137786865,
"eval_runtime": 0.6669,
"eval_samples_per_second": 14.995,
"eval_steps_per_second": 1.499,
"step": 2072
},
{
"epoch": 5.0060313630880575,
"grad_norm": 0.33984375,
"learning_rate": 0.000168419463638329,
"loss": 0.9096,
"step": 2075
},
{
"epoch": 5.018094089264173,
"grad_norm": 0.3359375,
"learning_rate": 0.00016821422304551766,
"loss": 0.9202,
"step": 2080
},
{
"epoch": 5.030156815440289,
"grad_norm": 0.318359375,
"learning_rate": 0.00016800844362963147,
"loss": 0.9134,
"step": 2085
},
{
"epoch": 5.042219541616405,
"grad_norm": 0.32421875,
"learning_rate": 0.00016780212701611886,
"loss": 0.909,
"step": 2090
},
{
"epoch": 5.054282267792521,
"grad_norm": 0.3125,
"learning_rate": 0.0001675952748346715,
"loss": 0.9011,
"step": 2095
},
{
"epoch": 5.066344993968637,
"grad_norm": 0.36328125,
"learning_rate": 0.00016738788871921152,
"loss": 0.9104,
"step": 2100
},
{
"epoch": 5.078407720144753,
"grad_norm": 0.35546875,
"learning_rate": 0.0001671799703078786,
"loss": 0.9222,
"step": 2105
},
{
"epoch": 5.090470446320868,
"grad_norm": 0.32421875,
"learning_rate": 0.00016697152124301695,
"loss": 0.9157,
"step": 2110
},
{
"epoch": 5.102533172496984,
"grad_norm": 0.345703125,
"learning_rate": 0.00016676254317116252,
"loss": 0.922,
"step": 2115
},
{
"epoch": 5.1145958986731,
"grad_norm": 0.361328125,
"learning_rate": 0.00016655303774302976,
"loss": 0.9036,
"step": 2120
},
{
"epoch": 5.126658624849216,
"grad_norm": 0.49609375,
"learning_rate": 0.0001663430066134988,
"loss": 0.9224,
"step": 2125
},
{
"epoch": 5.138721351025332,
"grad_norm": 0.50390625,
"learning_rate": 0.0001661324514416022,
"loss": 0.9104,
"step": 2130
},
{
"epoch": 5.150784077201448,
"grad_norm": 0.4296875,
"learning_rate": 0.0001659213738905119,
"loss": 0.9145,
"step": 2135
},
{
"epoch": 5.162846803377564,
"grad_norm": 0.30859375,
"learning_rate": 0.00016570977562752623,
"loss": 0.9142,
"step": 2140
},
{
"epoch": 5.1749095295536796,
"grad_norm": 0.36328125,
"learning_rate": 0.00016549765832405653,
"loss": 0.9082,
"step": 2145
},
{
"epoch": 5.186972255729795,
"grad_norm": 0.50390625,
"learning_rate": 0.000165285023655614,
"loss": 0.9093,
"step": 2150
},
{
"epoch": 5.1990349819059105,
"grad_norm": 0.55859375,
"learning_rate": 0.00016507187330179663,
"loss": 0.9211,
"step": 2155
},
{
"epoch": 5.211097708082026,
"grad_norm": 0.42578125,
"learning_rate": 0.0001648582089462756,
"loss": 0.9133,
"step": 2160
},
{
"epoch": 5.223160434258142,
"grad_norm": 0.365234375,
"learning_rate": 0.00016464403227678238,
"loss": 0.9211,
"step": 2165
},
{
"epoch": 5.235223160434258,
"grad_norm": 0.32421875,
"learning_rate": 0.00016442934498509505,
"loss": 0.9197,
"step": 2170
},
{
"epoch": 5.247285886610374,
"grad_norm": 0.33203125,
"learning_rate": 0.00016421414876702518,
"loss": 0.909,
"step": 2175
},
{
"epoch": 5.25934861278649,
"grad_norm": 0.390625,
"learning_rate": 0.00016399844532240433,
"loss": 0.9158,
"step": 2180
},
{
"epoch": 5.271411338962605,
"grad_norm": 0.328125,
"learning_rate": 0.0001637822363550706,
"loss": 0.914,
"step": 2185
},
{
"epoch": 5.283474065138721,
"grad_norm": 0.439453125,
"learning_rate": 0.00016356552357285522,
"loss": 0.9123,
"step": 2190
},
{
"epoch": 5.295536791314837,
"grad_norm": 0.361328125,
"learning_rate": 0.00016334830868756906,
"loss": 0.9221,
"step": 2195
},
{
"epoch": 5.307599517490953,
"grad_norm": 0.41015625,
"learning_rate": 0.00016313059341498901,
"loss": 0.9146,
"step": 2200
},
{
"epoch": 5.319662243667069,
"grad_norm": 0.318359375,
"learning_rate": 0.0001629123794748447,
"loss": 0.9072,
"step": 2205
},
{
"epoch": 5.331724969843185,
"grad_norm": 0.314453125,
"learning_rate": 0.00016269366859080451,
"loss": 0.9155,
"step": 2210
},
{
"epoch": 5.343787696019301,
"grad_norm": 0.34765625,
"learning_rate": 0.00016247446249046237,
"loss": 0.9152,
"step": 2215
},
{
"epoch": 5.355850422195417,
"grad_norm": 0.45703125,
"learning_rate": 0.00016225476290532374,
"loss": 0.9229,
"step": 2220
},
{
"epoch": 5.367913148371532,
"grad_norm": 0.40234375,
"learning_rate": 0.0001620345715707922,
"loss": 0.9192,
"step": 2225
},
{
"epoch": 5.379975874547648,
"grad_norm": 0.3828125,
"learning_rate": 0.00016181389022615564,
"loss": 0.9131,
"step": 2230
},
{
"epoch": 5.3920386007237635,
"grad_norm": 0.337890625,
"learning_rate": 0.00016159272061457255,
"loss": 0.9173,
"step": 2235
},
{
"epoch": 5.404101326899879,
"grad_norm": 0.314453125,
"learning_rate": 0.0001613710644830582,
"loss": 0.9159,
"step": 2240
},
{
"epoch": 5.416164053075995,
"grad_norm": 0.330078125,
"learning_rate": 0.0001611489235824709,
"loss": 0.9114,
"step": 2245
},
{
"epoch": 5.428226779252111,
"grad_norm": 0.34375,
"learning_rate": 0.0001609262996674981,
"loss": 0.926,
"step": 2250
},
{
"epoch": 5.440289505428227,
"grad_norm": 0.404296875,
"learning_rate": 0.00016070319449664264,
"loss": 0.9121,
"step": 2255
},
{
"epoch": 5.452352231604342,
"grad_norm": 0.390625,
"learning_rate": 0.00016047960983220873,
"loss": 0.9099,
"step": 2260
},
{
"epoch": 5.464414957780458,
"grad_norm": 0.5703125,
"learning_rate": 0.0001602555474402881,
"loss": 0.9105,
"step": 2265
},
{
"epoch": 5.476477683956574,
"grad_norm": 0.61328125,
"learning_rate": 0.0001600310090907461,
"loss": 0.913,
"step": 2270
},
{
"epoch": 5.48854041013269,
"grad_norm": 0.330078125,
"learning_rate": 0.00015980599655720758,
"loss": 0.9153,
"step": 2275
},
{
"epoch": 5.500603136308806,
"grad_norm": 0.34765625,
"learning_rate": 0.00015958051161704307,
"loss": 0.9116,
"step": 2280
},
{
"epoch": 5.512665862484922,
"grad_norm": 0.50390625,
"learning_rate": 0.00015935455605135446,
"loss": 0.9138,
"step": 2285
},
{
"epoch": 5.524728588661038,
"grad_norm": 0.41015625,
"learning_rate": 0.00015912813164496131,
"loss": 0.9166,
"step": 2290
},
{
"epoch": 5.536791314837153,
"grad_norm": 0.380859375,
"learning_rate": 0.00015890124018638638,
"loss": 0.9174,
"step": 2295
},
{
"epoch": 5.548854041013269,
"grad_norm": 0.345703125,
"learning_rate": 0.0001586738834678418,
"loss": 0.9165,
"step": 2300
},
{
"epoch": 5.560916767189385,
"grad_norm": 0.291015625,
"learning_rate": 0.00015844606328521463,
"loss": 0.912,
"step": 2305
},
{
"epoch": 5.572979493365501,
"grad_norm": 0.322265625,
"learning_rate": 0.00015821778143805296,
"loss": 0.9125,
"step": 2310
},
{
"epoch": 5.5850422195416165,
"grad_norm": 0.310546875,
"learning_rate": 0.0001579890397295515,
"loss": 0.9264,
"step": 2315
},
{
"epoch": 5.597104945717732,
"grad_norm": 0.298828125,
"learning_rate": 0.00015775983996653737,
"loss": 0.9129,
"step": 2320
},
{
"epoch": 5.609167671893848,
"grad_norm": 0.4375,
"learning_rate": 0.00015753018395945598,
"loss": 0.9139,
"step": 2325
},
{
"epoch": 5.621230398069963,
"grad_norm": 0.341796875,
"learning_rate": 0.00015730007352235644,
"loss": 0.9078,
"step": 2330
},
{
"epoch": 5.633293124246079,
"grad_norm": 0.3515625,
"learning_rate": 0.0001570695104728775,
"loss": 0.905,
"step": 2335
},
{
"epoch": 5.645355850422195,
"grad_norm": 0.30859375,
"learning_rate": 0.00015683849663223308,
"loss": 0.906,
"step": 2340
},
{
"epoch": 5.657418576598311,
"grad_norm": 0.337890625,
"learning_rate": 0.0001566070338251978,
"loss": 0.9144,
"step": 2345
},
{
"epoch": 5.669481302774427,
"grad_norm": 0.376953125,
"learning_rate": 0.00015637512388009284,
"loss": 0.9171,
"step": 2350
},
{
"epoch": 5.681544028950543,
"grad_norm": 0.349609375,
"learning_rate": 0.00015614276862877113,
"loss": 0.9169,
"step": 2355
},
{
"epoch": 5.693606755126659,
"grad_norm": 0.353515625,
"learning_rate": 0.00015590996990660317,
"loss": 0.9131,
"step": 2360
},
{
"epoch": 5.705669481302774,
"grad_norm": 0.37109375,
"learning_rate": 0.0001556767295524624,
"loss": 0.9248,
"step": 2365
},
{
"epoch": 5.71773220747889,
"grad_norm": 0.38671875,
"learning_rate": 0.0001554430494087107,
"loss": 0.9108,
"step": 2370
},
{
"epoch": 5.729794933655006,
"grad_norm": 0.359375,
"learning_rate": 0.00015520893132118385,
"loss": 0.909,
"step": 2375
},
{
"epoch": 5.741857659831122,
"grad_norm": 0.578125,
"learning_rate": 0.000154974377139177,
"loss": 0.9094,
"step": 2380
},
{
"epoch": 5.753920386007238,
"grad_norm": 0.61328125,
"learning_rate": 0.00015473938871542986,
"loss": 0.9155,
"step": 2385
},
{
"epoch": 5.765983112183354,
"grad_norm": 0.5546875,
"learning_rate": 0.00015450396790611234,
"loss": 0.9167,
"step": 2390
},
{
"epoch": 5.7780458383594695,
"grad_norm": 0.3984375,
"learning_rate": 0.00015426811657080967,
"loss": 0.9265,
"step": 2395
},
{
"epoch": 5.790108564535585,
"grad_norm": 0.31640625,
"learning_rate": 0.00015403183657250788,
"loss": 0.9131,
"step": 2400
},
{
"epoch": 5.8021712907117005,
"grad_norm": 0.30859375,
"learning_rate": 0.00015379512977757896,
"loss": 0.9088,
"step": 2405
},
{
"epoch": 5.814234016887816,
"grad_norm": 0.34375,
"learning_rate": 0.00015355799805576612,
"loss": 0.9101,
"step": 2410
},
{
"epoch": 5.826296743063932,
"grad_norm": 0.318359375,
"learning_rate": 0.00015332044328016914,
"loss": 0.9118,
"step": 2415
},
{
"epoch": 5.838359469240048,
"grad_norm": 0.3515625,
"learning_rate": 0.00015308246732722943,
"loss": 0.9107,
"step": 2420
},
{
"epoch": 5.850422195416164,
"grad_norm": 0.376953125,
"learning_rate": 0.0001528440720767153,
"loss": 0.9167,
"step": 2425
},
{
"epoch": 5.86248492159228,
"grad_norm": 0.306640625,
"learning_rate": 0.00015260525941170712,
"loss": 0.9059,
"step": 2430
},
{
"epoch": 5.874547647768396,
"grad_norm": 0.3203125,
"learning_rate": 0.00015236603121858232,
"loss": 0.918,
"step": 2435
},
{
"epoch": 5.886610373944512,
"grad_norm": 0.35546875,
"learning_rate": 0.00015212638938700068,
"loss": 0.9084,
"step": 2440
},
{
"epoch": 5.898673100120627,
"grad_norm": 0.328125,
"learning_rate": 0.00015188633580988926,
"loss": 0.9093,
"step": 2445
},
{
"epoch": 5.910735826296743,
"grad_norm": 0.328125,
"learning_rate": 0.00015164587238342744,
"loss": 0.9016,
"step": 2450
},
{
"epoch": 5.922798552472859,
"grad_norm": 0.68359375,
"learning_rate": 0.00015140500100703207,
"loss": 0.9129,
"step": 2455
},
{
"epoch": 5.934861278648975,
"grad_norm": 0.37109375,
"learning_rate": 0.00015116372358334233,
"loss": 0.922,
"step": 2460
},
{
"epoch": 5.946924004825091,
"grad_norm": 0.396484375,
"learning_rate": 0.00015092204201820492,
"loss": 0.915,
"step": 2465
},
{
"epoch": 5.958986731001207,
"grad_norm": 0.365234375,
"learning_rate": 0.00015067995822065858,
"loss": 0.9097,
"step": 2470
},
{
"epoch": 5.9710494571773225,
"grad_norm": 0.326171875,
"learning_rate": 0.00015043747410291945,
"loss": 0.9011,
"step": 2475
},
{
"epoch": 5.9831121833534375,
"grad_norm": 0.38671875,
"learning_rate": 0.0001501945915803658,
"loss": 0.919,
"step": 2480
},
{
"epoch": 5.9951749095295535,
"grad_norm": 0.447265625,
"learning_rate": 0.0001499513125715228,
"loss": 0.9129,
"step": 2485
},
{
"epoch": 6.0,
"eval_loss": 2.4628818035125732,
"eval_runtime": 0.5365,
"eval_samples_per_second": 18.639,
"eval_steps_per_second": 1.864,
"step": 2487
},
{
"epoch": 6.007237635705669,
"grad_norm": 0.376953125,
"learning_rate": 0.00014970763899804763,
"loss": 0.9123,
"step": 2490
},
{
"epoch": 6.019300361881785,
"grad_norm": 0.37109375,
"learning_rate": 0.00014946357278471388,
"loss": 0.899,
"step": 2495
},
{
"epoch": 6.031363088057901,
"grad_norm": 0.482421875,
"learning_rate": 0.0001492191158593968,
"loss": 0.8955,
"step": 2500
},
{
"epoch": 6.043425814234017,
"grad_norm": 0.478515625,
"learning_rate": 0.0001489742701530578,
"loss": 0.9005,
"step": 2505
},
{
"epoch": 6.055488540410133,
"grad_norm": 0.41796875,
"learning_rate": 0.0001487290375997292,
"loss": 0.8971,
"step": 2510
},
{
"epoch": 6.067551266586248,
"grad_norm": 0.40625,
"learning_rate": 0.00014848342013649914,
"loss": 0.9056,
"step": 2515
},
{
"epoch": 6.079613992762364,
"grad_norm": 0.326171875,
"learning_rate": 0.00014823741970349606,
"loss": 0.904,
"step": 2520
},
{
"epoch": 6.09167671893848,
"grad_norm": 0.30859375,
"learning_rate": 0.00014799103824387358,
"loss": 0.8914,
"step": 2525
},
{
"epoch": 6.103739445114596,
"grad_norm": 0.361328125,
"learning_rate": 0.0001477442777037949,
"loss": 0.9086,
"step": 2530
},
{
"epoch": 6.115802171290712,
"grad_norm": 0.369140625,
"learning_rate": 0.0001474971400324177,
"loss": 0.9118,
"step": 2535
},
{
"epoch": 6.127864897466828,
"grad_norm": 0.36328125,
"learning_rate": 0.00014724962718187852,
"loss": 0.9044,
"step": 2540
},
{
"epoch": 6.139927623642944,
"grad_norm": 0.380859375,
"learning_rate": 0.00014700174110727747,
"loss": 0.9075,
"step": 2545
},
{
"epoch": 6.151990349819059,
"grad_norm": 0.33203125,
"learning_rate": 0.00014675348376666278,
"loss": 0.8888,
"step": 2550
},
{
"epoch": 6.164053075995175,
"grad_norm": 0.34375,
"learning_rate": 0.00014650485712101524,
"loss": 0.8976,
"step": 2555
},
{
"epoch": 6.1761158021712905,
"grad_norm": 0.32421875,
"learning_rate": 0.00014625586313423287,
"loss": 0.9052,
"step": 2560
},
{
"epoch": 6.1881785283474064,
"grad_norm": 0.3125,
"learning_rate": 0.00014600650377311522,
"loss": 0.8917,
"step": 2565
},
{
"epoch": 6.200241254523522,
"grad_norm": 0.33203125,
"learning_rate": 0.00014575678100734796,
"loss": 0.9049,
"step": 2570
},
{
"epoch": 6.212303980699638,
"grad_norm": 0.390625,
"learning_rate": 0.0001455066968094873,
"loss": 0.9041,
"step": 2575
},
{
"epoch": 6.224366706875754,
"grad_norm": 0.30859375,
"learning_rate": 0.00014525625315494435,
"loss": 0.9048,
"step": 2580
},
{
"epoch": 6.23642943305187,
"grad_norm": 0.375,
"learning_rate": 0.00014500545202196968,
"loss": 0.9093,
"step": 2585
},
{
"epoch": 6.248492159227985,
"grad_norm": 0.4296875,
"learning_rate": 0.00014475429539163742,
"loss": 0.9003,
"step": 2590
},
{
"epoch": 6.260554885404101,
"grad_norm": 0.39453125,
"learning_rate": 0.00014450278524782986,
"loss": 0.9062,
"step": 2595
},
{
"epoch": 6.272617611580217,
"grad_norm": 0.4609375,
"learning_rate": 0.00014425092357722168,
"loss": 0.9032,
"step": 2600
},
{
"epoch": 6.284680337756333,
"grad_norm": 0.3828125,
"learning_rate": 0.00014399871236926422,
"loss": 0.9009,
"step": 2605
},
{
"epoch": 6.296743063932449,
"grad_norm": 0.41796875,
"learning_rate": 0.00014374615361616985,
"loss": 0.8952,
"step": 2610
},
{
"epoch": 6.308805790108565,
"grad_norm": 0.376953125,
"learning_rate": 0.00014349324931289627,
"loss": 0.8959,
"step": 2615
},
{
"epoch": 6.320868516284681,
"grad_norm": 0.302734375,
"learning_rate": 0.0001432400014571305,
"loss": 0.906,
"step": 2620
},
{
"epoch": 6.332931242460796,
"grad_norm": 0.373046875,
"learning_rate": 0.00014298641204927342,
"loss": 0.899,
"step": 2625
},
{
"epoch": 6.344993968636912,
"grad_norm": 0.400390625,
"learning_rate": 0.00014273248309242372,
"loss": 0.895,
"step": 2630
},
{
"epoch": 6.357056694813028,
"grad_norm": 0.384765625,
"learning_rate": 0.0001424782165923623,
"loss": 0.8986,
"step": 2635
},
{
"epoch": 6.3691194209891435,
"grad_norm": 0.41796875,
"learning_rate": 0.0001422236145575362,
"loss": 0.9039,
"step": 2640
},
{
"epoch": 6.381182147165259,
"grad_norm": 0.443359375,
"learning_rate": 0.0001419686789990429,
"loss": 0.9086,
"step": 2645
},
{
"epoch": 6.393244873341375,
"grad_norm": 0.369140625,
"learning_rate": 0.0001417134119306144,
"loss": 0.8945,
"step": 2650
},
{
"epoch": 6.405307599517491,
"grad_norm": 0.33984375,
"learning_rate": 0.00014145781536860122,
"loss": 0.9083,
"step": 2655
},
{
"epoch": 6.417370325693607,
"grad_norm": 0.35546875,
"learning_rate": 0.00014120189133195657,
"loss": 0.9029,
"step": 2660
},
{
"epoch": 6.429433051869722,
"grad_norm": 0.3359375,
"learning_rate": 0.00014094564184222042,
"loss": 0.9124,
"step": 2665
},
{
"epoch": 6.441495778045838,
"grad_norm": 0.361328125,
"learning_rate": 0.00014068906892350343,
"loss": 0.9106,
"step": 2670
},
{
"epoch": 6.453558504221954,
"grad_norm": 0.34375,
"learning_rate": 0.00014043217460247109,
"loss": 0.9021,
"step": 2675
},
{
"epoch": 6.46562123039807,
"grad_norm": 0.345703125,
"learning_rate": 0.0001401749609083276,
"loss": 0.907,
"step": 2680
},
{
"epoch": 6.477683956574186,
"grad_norm": 0.3203125,
"learning_rate": 0.0001399174298727998,
"loss": 0.8993,
"step": 2685
},
{
"epoch": 6.489746682750302,
"grad_norm": 0.390625,
"learning_rate": 0.0001396595835301214,
"loss": 0.9022,
"step": 2690
},
{
"epoch": 6.501809408926418,
"grad_norm": 0.384765625,
"learning_rate": 0.00013940142391701652,
"loss": 0.9056,
"step": 2695
},
{
"epoch": 6.513872135102533,
"grad_norm": 0.328125,
"learning_rate": 0.00013914295307268396,
"loss": 0.9099,
"step": 2700
},
{
"epoch": 6.525934861278649,
"grad_norm": 0.365234375,
"learning_rate": 0.00013888417303878077,
"loss": 0.9055,
"step": 2705
},
{
"epoch": 6.537997587454765,
"grad_norm": 0.36328125,
"learning_rate": 0.00013862508585940644,
"loss": 0.9037,
"step": 2710
},
{
"epoch": 6.550060313630881,
"grad_norm": 0.353515625,
"learning_rate": 0.00013836569358108647,
"loss": 0.8909,
"step": 2715
},
{
"epoch": 6.5621230398069965,
"grad_norm": 0.3125,
"learning_rate": 0.00013810599825275638,
"loss": 0.9075,
"step": 2720
},
{
"epoch": 6.574185765983112,
"grad_norm": 0.33203125,
"learning_rate": 0.00013784600192574557,
"loss": 0.9112,
"step": 2725
},
{
"epoch": 6.586248492159228,
"grad_norm": 0.3828125,
"learning_rate": 0.00013758570665376086,
"loss": 0.9033,
"step": 2730
},
{
"epoch": 6.598311218335343,
"grad_norm": 0.306640625,
"learning_rate": 0.00013732511449287055,
"loss": 0.8993,
"step": 2735
},
{
"epoch": 6.610373944511459,
"grad_norm": 0.361328125,
"learning_rate": 0.00013706422750148797,
"loss": 0.8954,
"step": 2740
},
{
"epoch": 6.622436670687575,
"grad_norm": 0.396484375,
"learning_rate": 0.00013680304774035538,
"loss": 0.9054,
"step": 2745
},
{
"epoch": 6.634499396863691,
"grad_norm": 0.314453125,
"learning_rate": 0.00013654157727252766,
"loss": 0.9018,
"step": 2750
},
{
"epoch": 6.646562123039807,
"grad_norm": 0.50390625,
"learning_rate": 0.00013627981816335583,
"loss": 0.9036,
"step": 2755
},
{
"epoch": 6.658624849215923,
"grad_norm": 0.33203125,
"learning_rate": 0.00013601777248047105,
"loss": 0.9067,
"step": 2760
},
{
"epoch": 6.670687575392039,
"grad_norm": 0.3125,
"learning_rate": 0.000135755442293768,
"loss": 0.8984,
"step": 2765
},
{
"epoch": 6.682750301568154,
"grad_norm": 0.30859375,
"learning_rate": 0.00013549282967538872,
"loss": 0.907,
"step": 2770
},
{
"epoch": 6.69481302774427,
"grad_norm": 0.314453125,
"learning_rate": 0.0001352299366997062,
"loss": 0.9024,
"step": 2775
},
{
"epoch": 6.706875753920386,
"grad_norm": 0.375,
"learning_rate": 0.0001349667654433078,
"loss": 0.9069,
"step": 2780
},
{
"epoch": 6.718938480096502,
"grad_norm": 0.365234375,
"learning_rate": 0.00013470331798497926,
"loss": 0.9075,
"step": 2785
},
{
"epoch": 6.731001206272618,
"grad_norm": 0.330078125,
"learning_rate": 0.0001344395964056878,
"loss": 0.9034,
"step": 2790
},
{
"epoch": 6.743063932448734,
"grad_norm": 0.318359375,
"learning_rate": 0.0001341756027885661,
"loss": 0.9037,
"step": 2795
},
{
"epoch": 6.7551266586248495,
"grad_norm": 0.3203125,
"learning_rate": 0.00013391133921889558,
"loss": 0.9075,
"step": 2800
},
{
"epoch": 6.7671893848009645,
"grad_norm": 0.328125,
"learning_rate": 0.00013364680778409,
"loss": 0.9046,
"step": 2805
},
{
"epoch": 6.7792521109770805,
"grad_norm": 0.35546875,
"learning_rate": 0.000133382010573679,
"loss": 0.9002,
"step": 2810
},
{
"epoch": 6.791314837153196,
"grad_norm": 0.310546875,
"learning_rate": 0.00013311694967929158,
"loss": 0.9066,
"step": 2815
},
{
"epoch": 6.803377563329312,
"grad_norm": 0.318359375,
"learning_rate": 0.00013285162719463961,
"loss": 0.9057,
"step": 2820
},
{
"epoch": 6.815440289505428,
"grad_norm": 0.345703125,
"learning_rate": 0.0001325860452155012,
"loss": 0.8974,
"step": 2825
},
{
"epoch": 6.827503015681544,
"grad_norm": 0.30859375,
"learning_rate": 0.0001323202058397042,
"loss": 0.9006,
"step": 2830
},
{
"epoch": 6.83956574185766,
"grad_norm": 0.32421875,
"learning_rate": 0.00013205411116710972,
"loss": 0.9023,
"step": 2835
},
{
"epoch": 6.851628468033776,
"grad_norm": 0.3203125,
"learning_rate": 0.00013178776329959531,
"loss": 0.903,
"step": 2840
},
{
"epoch": 6.863691194209892,
"grad_norm": 0.380859375,
"learning_rate": 0.00013152116434103867,
"loss": 0.9027,
"step": 2845
},
{
"epoch": 6.875753920386007,
"grad_norm": 0.3828125,
"learning_rate": 0.0001312543163973007,
"loss": 0.9087,
"step": 2850
},
{
"epoch": 6.887816646562123,
"grad_norm": 0.359375,
"learning_rate": 0.00013098722157620917,
"loss": 0.8917,
"step": 2855
},
{
"epoch": 6.899879372738239,
"grad_norm": 0.37109375,
"learning_rate": 0.00013071988198754185,
"loss": 0.902,
"step": 2860
},
{
"epoch": 6.911942098914355,
"grad_norm": 0.357421875,
"learning_rate": 0.00013045229974300993,
"loss": 0.905,
"step": 2865
},
{
"epoch": 6.924004825090471,
"grad_norm": 0.3671875,
"learning_rate": 0.0001301844769562414,
"loss": 0.9072,
"step": 2870
},
{
"epoch": 6.936067551266587,
"grad_norm": 0.35546875,
"learning_rate": 0.00012991641574276418,
"loss": 0.9002,
"step": 2875
},
{
"epoch": 6.9481302774427025,
"grad_norm": 0.330078125,
"learning_rate": 0.0001296481182199896,
"loss": 0.8994,
"step": 2880
},
{
"epoch": 6.9601930036188175,
"grad_norm": 0.3046875,
"learning_rate": 0.0001293795865071956,
"loss": 0.9076,
"step": 2885
},
{
"epoch": 6.9722557297949335,
"grad_norm": 0.328125,
"learning_rate": 0.0001291108227255099,
"loss": 0.9044,
"step": 2890
},
{
"epoch": 6.984318455971049,
"grad_norm": 0.333984375,
"learning_rate": 0.00012884182899789343,
"loss": 0.9003,
"step": 2895
},
{
"epoch": 6.996381182147165,
"grad_norm": 0.3203125,
"learning_rate": 0.0001285726074491234,
"loss": 0.8995,
"step": 2900
},
{
"epoch": 6.998793727382388,
"eval_loss": 2.4702937602996826,
"eval_runtime": 0.6137,
"eval_samples_per_second": 16.294,
"eval_steps_per_second": 1.629,
"step": 2901
},
{
"epoch": 7.008443908323281,
"grad_norm": 0.349609375,
"learning_rate": 0.00012830316020577656,
"loss": 0.8934,
"step": 2905
},
{
"epoch": 7.020506634499397,
"grad_norm": 0.34765625,
"learning_rate": 0.00012803348939621252,
"loss": 0.8901,
"step": 2910
},
{
"epoch": 7.032569360675513,
"grad_norm": 0.314453125,
"learning_rate": 0.00012776359715055668,
"loss": 0.8963,
"step": 2915
},
{
"epoch": 7.044632086851628,
"grad_norm": 0.32421875,
"learning_rate": 0.0001274934856006837,
"loss": 0.8909,
"step": 2920
},
{
"epoch": 7.056694813027744,
"grad_norm": 0.353515625,
"learning_rate": 0.00012722315688020047,
"loss": 0.8836,
"step": 2925
},
{
"epoch": 7.06875753920386,
"grad_norm": 0.326171875,
"learning_rate": 0.0001269526131244292,
"loss": 0.8914,
"step": 2930
},
{
"epoch": 7.080820265379976,
"grad_norm": 0.32421875,
"learning_rate": 0.0001266818564703909,
"loss": 0.9006,
"step": 2935
},
{
"epoch": 7.092882991556092,
"grad_norm": 0.328125,
"learning_rate": 0.00012641088905678802,
"loss": 0.9071,
"step": 2940
},
{
"epoch": 7.104945717732208,
"grad_norm": 0.32421875,
"learning_rate": 0.00012613971302398794,
"loss": 0.8939,
"step": 2945
},
{
"epoch": 7.117008443908324,
"grad_norm": 0.361328125,
"learning_rate": 0.00012586833051400588,
"loss": 0.8813,
"step": 2950
},
{
"epoch": 7.129071170084439,
"grad_norm": 0.314453125,
"learning_rate": 0.000125596743670488,
"loss": 0.8852,
"step": 2955
},
{
"epoch": 7.141133896260555,
"grad_norm": 0.34765625,
"learning_rate": 0.00012532495463869452,
"loss": 0.8924,
"step": 2960
},
{
"epoch": 7.1531966224366705,
"grad_norm": 0.322265625,
"learning_rate": 0.00012505296556548272,
"loss": 0.891,
"step": 2965
},
{
"epoch": 7.1652593486127865,
"grad_norm": 0.3515625,
"learning_rate": 0.00012478077859929,
"loss": 0.8914,
"step": 2970
},
{
"epoch": 7.177322074788902,
"grad_norm": 0.328125,
"learning_rate": 0.00012450839589011695,
"loss": 0.8947,
"step": 2975
},
{
"epoch": 7.189384800965018,
"grad_norm": 0.3203125,
"learning_rate": 0.00012423581958951026,
"loss": 0.8955,
"step": 2980
},
{
"epoch": 7.201447527141134,
"grad_norm": 0.318359375,
"learning_rate": 0.0001239630518505459,
"loss": 0.898,
"step": 2985
},
{
"epoch": 7.213510253317249,
"grad_norm": 0.376953125,
"learning_rate": 0.00012369009482781192,
"loss": 0.8951,
"step": 2990
},
{
"epoch": 7.225572979493365,
"grad_norm": 0.53125,
"learning_rate": 0.00012341695067739154,
"loss": 0.8944,
"step": 2995
},
{
"epoch": 7.237635705669481,
"grad_norm": 0.421875,
"learning_rate": 0.00012314362155684612,
"loss": 0.8936,
"step": 3000
},
{
"epoch": 7.249698431845597,
"grad_norm": 0.43359375,
"learning_rate": 0.00012287010962519807,
"loss": 0.8998,
"step": 3005
},
{
"epoch": 7.261761158021713,
"grad_norm": 0.3359375,
"learning_rate": 0.00012259641704291384,
"loss": 0.8959,
"step": 3010
},
{
"epoch": 7.273823884197829,
"grad_norm": 0.421875,
"learning_rate": 0.00012232254597188688,
"loss": 0.8942,
"step": 3015
},
{
"epoch": 7.285886610373945,
"grad_norm": 0.37890625,
"learning_rate": 0.00012204849857542038,
"loss": 0.8886,
"step": 3020
},
{
"epoch": 7.297949336550061,
"grad_norm": 0.333984375,
"learning_rate": 0.00012177427701821051,
"loss": 0.8859,
"step": 3025
},
{
"epoch": 7.310012062726176,
"grad_norm": 0.361328125,
"learning_rate": 0.00012149988346632894,
"loss": 0.8988,
"step": 3030
},
{
"epoch": 7.322074788902292,
"grad_norm": 0.328125,
"learning_rate": 0.00012122532008720611,
"loss": 0.9011,
"step": 3035
},
{
"epoch": 7.334137515078408,
"grad_norm": 0.322265625,
"learning_rate": 0.00012095058904961379,
"loss": 0.8903,
"step": 3040
},
{
"epoch": 7.3462002412545235,
"grad_norm": 0.32421875,
"learning_rate": 0.00012067569252364809,
"loss": 0.9006,
"step": 3045
},
{
"epoch": 7.3582629674306395,
"grad_norm": 0.322265625,
"learning_rate": 0.00012040063268071243,
"loss": 0.8968,
"step": 3050
},
{
"epoch": 7.370325693606755,
"grad_norm": 0.318359375,
"learning_rate": 0.00012012541169350011,
"loss": 0.9044,
"step": 3055
},
{
"epoch": 7.382388419782871,
"grad_norm": 0.345703125,
"learning_rate": 0.0001198500317359774,
"loss": 0.8968,
"step": 3060
},
{
"epoch": 7.394451145958986,
"grad_norm": 0.322265625,
"learning_rate": 0.00011957449498336626,
"loss": 0.8956,
"step": 3065
},
{
"epoch": 7.406513872135102,
"grad_norm": 0.33984375,
"learning_rate": 0.00011929880361212717,
"loss": 0.8967,
"step": 3070
},
{
"epoch": 7.418576598311218,
"grad_norm": 0.31640625,
"learning_rate": 0.00011902295979994192,
"loss": 0.8995,
"step": 3075
},
{
"epoch": 7.430639324487334,
"grad_norm": 0.314453125,
"learning_rate": 0.00011874696572569645,
"loss": 0.8923,
"step": 3080
},
{
"epoch": 7.44270205066345,
"grad_norm": 0.310546875,
"learning_rate": 0.0001184708235694636,
"loss": 0.8915,
"step": 3085
},
{
"epoch": 7.454764776839566,
"grad_norm": 0.3359375,
"learning_rate": 0.00011819453551248592,
"loss": 0.8951,
"step": 3090
},
{
"epoch": 7.466827503015682,
"grad_norm": 0.365234375,
"learning_rate": 0.00011791810373715846,
"loss": 0.8924,
"step": 3095
},
{
"epoch": 7.478890229191798,
"grad_norm": 0.443359375,
"learning_rate": 0.0001176415304270114,
"loss": 0.895,
"step": 3100
},
{
"epoch": 7.490952955367913,
"grad_norm": 0.41796875,
"learning_rate": 0.00011736481776669306,
"loss": 0.8965,
"step": 3105
},
{
"epoch": 7.503015681544029,
"grad_norm": 0.36328125,
"learning_rate": 0.00011708796794195227,
"loss": 0.9014,
"step": 3110
},
{
"epoch": 7.515078407720145,
"grad_norm": 0.373046875,
"learning_rate": 0.00011681098313962158,
"loss": 0.8904,
"step": 3115
},
{
"epoch": 7.527141133896261,
"grad_norm": 0.34765625,
"learning_rate": 0.00011653386554759946,
"loss": 0.8932,
"step": 3120
},
{
"epoch": 7.5392038600723765,
"grad_norm": 0.328125,
"learning_rate": 0.00011625661735483348,
"loss": 0.8999,
"step": 3125
},
{
"epoch": 7.5512665862484925,
"grad_norm": 0.37890625,
"learning_rate": 0.00011597924075130273,
"loss": 0.8913,
"step": 3130
},
{
"epoch": 7.563329312424608,
"grad_norm": 0.330078125,
"learning_rate": 0.00011570173792800066,
"loss": 0.8957,
"step": 3135
},
{
"epoch": 7.575392038600723,
"grad_norm": 0.34375,
"learning_rate": 0.00011542411107691767,
"loss": 0.8927,
"step": 3140
},
{
"epoch": 7.587454764776839,
"grad_norm": 0.376953125,
"learning_rate": 0.00011514636239102393,
"loss": 0.897,
"step": 3145
},
{
"epoch": 7.599517490952955,
"grad_norm": 0.328125,
"learning_rate": 0.00011486849406425188,
"loss": 0.8962,
"step": 3150
},
{
"epoch": 7.611580217129071,
"grad_norm": 0.318359375,
"learning_rate": 0.00011459050829147915,
"loss": 0.8853,
"step": 3155
},
{
"epoch": 7.623642943305187,
"grad_norm": 0.3203125,
"learning_rate": 0.00011431240726851091,
"loss": 0.8978,
"step": 3160
},
{
"epoch": 7.635705669481303,
"grad_norm": 0.326171875,
"learning_rate": 0.00011403419319206284,
"loss": 0.894,
"step": 3165
},
{
"epoch": 7.647768395657419,
"grad_norm": 0.37890625,
"learning_rate": 0.00011375586825974349,
"loss": 0.8913,
"step": 3170
},
{
"epoch": 7.659831121833534,
"grad_norm": 0.32421875,
"learning_rate": 0.0001134774346700371,
"loss": 0.8915,
"step": 3175
},
{
"epoch": 7.67189384800965,
"grad_norm": 0.322265625,
"learning_rate": 0.0001131988946222863,
"loss": 0.8962,
"step": 3180
},
{
"epoch": 7.683956574185766,
"grad_norm": 0.322265625,
"learning_rate": 0.00011292025031667448,
"loss": 0.8949,
"step": 3185
},
{
"epoch": 7.696019300361882,
"grad_norm": 0.33203125,
"learning_rate": 0.00011264150395420865,
"loss": 0.8974,
"step": 3190
},
{
"epoch": 7.708082026537998,
"grad_norm": 0.34765625,
"learning_rate": 0.00011236265773670196,
"loss": 0.8933,
"step": 3195
},
{
"epoch": 7.720144752714114,
"grad_norm": 0.330078125,
"learning_rate": 0.00011208371386675624,
"loss": 0.8996,
"step": 3200
},
{
"epoch": 7.7322074788902295,
"grad_norm": 0.34765625,
"learning_rate": 0.00011180467454774481,
"loss": 0.892,
"step": 3205
},
{
"epoch": 7.744270205066345,
"grad_norm": 0.35546875,
"learning_rate": 0.00011152554198379484,
"loss": 0.8943,
"step": 3210
},
{
"epoch": 7.7563329312424605,
"grad_norm": 0.31640625,
"learning_rate": 0.00011124631837976997,
"loss": 0.8921,
"step": 3215
},
{
"epoch": 7.768395657418576,
"grad_norm": 0.373046875,
"learning_rate": 0.00011096700594125318,
"loss": 0.891,
"step": 3220
},
{
"epoch": 7.780458383594692,
"grad_norm": 0.328125,
"learning_rate": 0.00011068760687452895,
"loss": 0.8956,
"step": 3225
},
{
"epoch": 7.792521109770808,
"grad_norm": 0.314453125,
"learning_rate": 0.00011040812338656615,
"loss": 0.8986,
"step": 3230
},
{
"epoch": 7.804583835946924,
"grad_norm": 0.3203125,
"learning_rate": 0.0001101285576850004,
"loss": 0.8905,
"step": 3235
},
{
"epoch": 7.81664656212304,
"grad_norm": 0.306640625,
"learning_rate": 0.00010984891197811687,
"loss": 0.8861,
"step": 3240
},
{
"epoch": 7.828709288299155,
"grad_norm": 0.3359375,
"learning_rate": 0.00010956918847483255,
"loss": 0.8908,
"step": 3245
},
{
"epoch": 7.840772014475271,
"grad_norm": 0.328125,
"learning_rate": 0.00010928938938467896,
"loss": 0.8972,
"step": 3250
},
{
"epoch": 7.852834740651387,
"grad_norm": 0.310546875,
"learning_rate": 0.00010900951691778481,
"loss": 0.8977,
"step": 3255
},
{
"epoch": 7.864897466827503,
"grad_norm": 0.3359375,
"learning_rate": 0.00010872957328485819,
"loss": 0.8974,
"step": 3260
},
{
"epoch": 7.876960193003619,
"grad_norm": 0.34765625,
"learning_rate": 0.00010844956069716957,
"loss": 0.8964,
"step": 3265
},
{
"epoch": 7.889022919179735,
"grad_norm": 0.30078125,
"learning_rate": 0.00010816948136653386,
"loss": 0.889,
"step": 3270
},
{
"epoch": 7.901085645355851,
"grad_norm": 0.34375,
"learning_rate": 0.00010788933750529335,
"loss": 0.8914,
"step": 3275
},
{
"epoch": 7.913148371531967,
"grad_norm": 0.314453125,
"learning_rate": 0.00010760913132630002,
"loss": 0.8923,
"step": 3280
},
{
"epoch": 7.9252110977080825,
"grad_norm": 0.328125,
"learning_rate": 0.000107328865042898,
"loss": 0.8954,
"step": 3285
},
{
"epoch": 7.9372738238841976,
"grad_norm": 0.34375,
"learning_rate": 0.00010704854086890633,
"loss": 0.8992,
"step": 3290
},
{
"epoch": 7.9493365500603135,
"grad_norm": 0.330078125,
"learning_rate": 0.00010676816101860118,
"loss": 0.8954,
"step": 3295
},
{
"epoch": 7.961399276236429,
"grad_norm": 0.31640625,
"learning_rate": 0.00010648772770669861,
"loss": 0.8952,
"step": 3300
},
{
"epoch": 7.973462002412545,
"grad_norm": 0.353515625,
"learning_rate": 0.00010620724314833696,
"loss": 0.8903,
"step": 3305
},
{
"epoch": 7.985524728588661,
"grad_norm": 0.3359375,
"learning_rate": 0.00010592670955905935,
"loss": 0.8909,
"step": 3310
},
{
"epoch": 7.997587454764777,
"grad_norm": 0.357421875,
"learning_rate": 0.00010564612915479612,
"loss": 0.8999,
"step": 3315
},
{
"epoch": 8.0,
"eval_loss": 2.4829864501953125,
"eval_runtime": 0.5356,
"eval_samples_per_second": 18.67,
"eval_steps_per_second": 1.867,
"step": 3316
},
{
"epoch": 8.009650180940893,
"grad_norm": 0.306640625,
"learning_rate": 0.00010536550415184751,
"loss": 0.889,
"step": 3320
},
{
"epoch": 8.021712907117008,
"grad_norm": 0.3359375,
"learning_rate": 0.000105084836766866,
"loss": 0.8866,
"step": 3325
},
{
"epoch": 8.033775633293125,
"grad_norm": 0.333984375,
"learning_rate": 0.00010480412921683888,
"loss": 0.8776,
"step": 3330
},
{
"epoch": 8.04583835946924,
"grad_norm": 0.361328125,
"learning_rate": 0.00010452338371907064,
"loss": 0.8824,
"step": 3335
},
{
"epoch": 8.057901085645355,
"grad_norm": 0.361328125,
"learning_rate": 0.00010424260249116557,
"loss": 0.8906,
"step": 3340
},
{
"epoch": 8.069963811821472,
"grad_norm": 0.328125,
"learning_rate": 0.00010396178775101014,
"loss": 0.8918,
"step": 3345
},
{
"epoch": 8.082026537997587,
"grad_norm": 0.359375,
"learning_rate": 0.0001036809417167556,
"loss": 0.8811,
"step": 3350
},
{
"epoch": 8.094089264173704,
"grad_norm": 0.359375,
"learning_rate": 0.00010340006660680032,
"loss": 0.896,
"step": 3355
},
{
"epoch": 8.106151990349819,
"grad_norm": 0.34765625,
"learning_rate": 0.00010311916463977242,
"loss": 0.8858,
"step": 3360
},
{
"epoch": 8.118214716525936,
"grad_norm": 0.392578125,
"learning_rate": 0.00010283823803451208,
"loss": 0.8834,
"step": 3365
},
{
"epoch": 8.13027744270205,
"grad_norm": 0.32421875,
"learning_rate": 0.00010255728901005417,
"loss": 0.882,
"step": 3370
},
{
"epoch": 8.142340168878167,
"grad_norm": 0.33984375,
"learning_rate": 0.00010227631978561056,
"loss": 0.8819,
"step": 3375
},
{
"epoch": 8.154402895054282,
"grad_norm": 0.34765625,
"learning_rate": 0.00010199533258055282,
"loss": 0.8755,
"step": 3380
},
{
"epoch": 8.166465621230397,
"grad_norm": 0.333984375,
"learning_rate": 0.00010171432961439438,
"loss": 0.8847,
"step": 3385
},
{
"epoch": 8.178528347406514,
"grad_norm": 0.318359375,
"learning_rate": 0.00010143331310677331,
"loss": 0.8951,
"step": 3390
},
{
"epoch": 8.19059107358263,
"grad_norm": 0.373046875,
"learning_rate": 0.00010115228527743452,
"loss": 0.888,
"step": 3395
},
{
"epoch": 8.202653799758746,
"grad_norm": 0.345703125,
"learning_rate": 0.00010087124834621248,
"loss": 0.8875,
"step": 3400
},
{
"epoch": 8.214716525934861,
"grad_norm": 0.345703125,
"learning_rate": 0.00010059020453301345,
"loss": 0.8902,
"step": 3405
},
{
"epoch": 8.226779252110978,
"grad_norm": 0.431640625,
"learning_rate": 0.00010030915605779809,
"loss": 0.8871,
"step": 3410
},
{
"epoch": 8.238841978287093,
"grad_norm": 0.318359375,
"learning_rate": 0.00010002810514056391,
"loss": 0.8884,
"step": 3415
},
{
"epoch": 8.250904704463208,
"grad_norm": 0.31640625,
"learning_rate": 9.974705400132764e-05,
"loss": 0.8859,
"step": 3420
},
{
"epoch": 8.262967430639325,
"grad_norm": 0.3125,
"learning_rate": 9.946600486010785e-05,
"loss": 0.8897,
"step": 3425
},
{
"epoch": 8.27503015681544,
"grad_norm": 0.326171875,
"learning_rate": 9.918495993690724e-05,
"loss": 0.8877,
"step": 3430
},
{
"epoch": 8.287092882991557,
"grad_norm": 0.31640625,
"learning_rate": 9.890392145169531e-05,
"loss": 0.8841,
"step": 3435
},
{
"epoch": 8.299155609167672,
"grad_norm": 0.447265625,
"learning_rate": 9.862289162439059e-05,
"loss": 0.8854,
"step": 3440
},
{
"epoch": 8.311218335343789,
"grad_norm": 0.341796875,
"learning_rate": 9.834187267484326e-05,
"loss": 0.8843,
"step": 3445
},
{
"epoch": 8.323281061519904,
"grad_norm": 0.361328125,
"learning_rate": 9.806086682281758e-05,
"loss": 0.8886,
"step": 3450
},
{
"epoch": 8.335343787696019,
"grad_norm": 0.33984375,
"learning_rate": 9.777987628797437e-05,
"loss": 0.8878,
"step": 3455
},
{
"epoch": 8.347406513872135,
"grad_norm": 0.33203125,
"learning_rate": 9.749890328985345e-05,
"loss": 0.8811,
"step": 3460
},
{
"epoch": 8.35946924004825,
"grad_norm": 0.40234375,
"learning_rate": 9.721795004785605e-05,
"loss": 0.8902,
"step": 3465
},
{
"epoch": 8.371531966224367,
"grad_norm": 0.380859375,
"learning_rate": 9.693701878122749e-05,
"loss": 0.8854,
"step": 3470
},
{
"epoch": 8.383594692400482,
"grad_norm": 0.3671875,
"learning_rate": 9.665611170903942e-05,
"loss": 0.8934,
"step": 3475
},
{
"epoch": 8.395657418576599,
"grad_norm": 0.421875,
"learning_rate": 9.637523105017229e-05,
"loss": 0.8905,
"step": 3480
},
{
"epoch": 8.407720144752714,
"grad_norm": 0.4140625,
"learning_rate": 9.609437902329814e-05,
"loss": 0.8884,
"step": 3485
},
{
"epoch": 8.41978287092883,
"grad_norm": 0.314453125,
"learning_rate": 9.581355784686266e-05,
"loss": 0.8864,
"step": 3490
},
{
"epoch": 8.431845597104946,
"grad_norm": 0.35546875,
"learning_rate": 9.553276973906786e-05,
"loss": 0.8924,
"step": 3495
},
{
"epoch": 8.443908323281061,
"grad_norm": 0.318359375,
"learning_rate": 9.525201691785462e-05,
"loss": 0.8938,
"step": 3500
},
{
"epoch": 8.455971049457178,
"grad_norm": 0.357421875,
"learning_rate": 9.497130160088508e-05,
"loss": 0.8937,
"step": 3505
},
{
"epoch": 8.468033775633293,
"grad_norm": 0.32421875,
"learning_rate": 9.469062600552509e-05,
"loss": 0.8847,
"step": 3510
},
{
"epoch": 8.48009650180941,
"grad_norm": 0.306640625,
"learning_rate": 9.44099923488267e-05,
"loss": 0.8882,
"step": 3515
},
{
"epoch": 8.492159227985525,
"grad_norm": 0.37890625,
"learning_rate": 9.412940284751089e-05,
"loss": 0.8885,
"step": 3520
},
{
"epoch": 8.50422195416164,
"grad_norm": 0.33984375,
"learning_rate": 9.384885971794961e-05,
"loss": 0.8938,
"step": 3525
},
{
"epoch": 8.516284680337757,
"grad_norm": 0.318359375,
"learning_rate": 9.356836517614863e-05,
"loss": 0.8938,
"step": 3530
},
{
"epoch": 8.528347406513872,
"grad_norm": 0.400390625,
"learning_rate": 9.328792143773e-05,
"loss": 0.8765,
"step": 3535
},
{
"epoch": 8.540410132689988,
"grad_norm": 0.38671875,
"learning_rate": 9.300753071791434e-05,
"loss": 0.8879,
"step": 3540
},
{
"epoch": 8.552472858866103,
"grad_norm": 0.361328125,
"learning_rate": 9.272719523150355e-05,
"loss": 0.8927,
"step": 3545
},
{
"epoch": 8.56453558504222,
"grad_norm": 0.35546875,
"learning_rate": 9.24469171928632e-05,
"loss": 0.8867,
"step": 3550
},
{
"epoch": 8.576598311218335,
"grad_norm": 0.494140625,
"learning_rate": 9.216669881590515e-05,
"loss": 0.8951,
"step": 3555
},
{
"epoch": 8.588661037394452,
"grad_norm": 0.349609375,
"learning_rate": 9.188654231406993e-05,
"loss": 0.8918,
"step": 3560
},
{
"epoch": 8.600723763570567,
"grad_norm": 0.345703125,
"learning_rate": 9.160644990030931e-05,
"loss": 0.8894,
"step": 3565
},
{
"epoch": 8.612786489746682,
"grad_norm": 0.376953125,
"learning_rate": 9.132642378706894e-05,
"loss": 0.8868,
"step": 3570
},
{
"epoch": 8.624849215922799,
"grad_norm": 0.322265625,
"learning_rate": 9.104646618627063e-05,
"loss": 0.8848,
"step": 3575
},
{
"epoch": 8.636911942098914,
"grad_norm": 0.369140625,
"learning_rate": 9.076657930929507e-05,
"loss": 0.8997,
"step": 3580
},
{
"epoch": 8.64897466827503,
"grad_norm": 0.341796875,
"learning_rate": 9.048676536696425e-05,
"loss": 0.8922,
"step": 3585
},
{
"epoch": 8.661037394451146,
"grad_norm": 0.365234375,
"learning_rate": 9.020702656952421e-05,
"loss": 0.8889,
"step": 3590
},
{
"epoch": 8.67310012062726,
"grad_norm": 0.373046875,
"learning_rate": 8.992736512662724e-05,
"loss": 0.8801,
"step": 3595
},
{
"epoch": 8.685162846803378,
"grad_norm": 0.3125,
"learning_rate": 8.964778324731467e-05,
"loss": 0.8891,
"step": 3600
},
{
"epoch": 8.697225572979493,
"grad_norm": 0.353515625,
"learning_rate": 8.93682831399994e-05,
"loss": 0.8902,
"step": 3605
},
{
"epoch": 8.70928829915561,
"grad_norm": 0.318359375,
"learning_rate": 8.908886701244838e-05,
"loss": 0.8901,
"step": 3610
},
{
"epoch": 8.721351025331725,
"grad_norm": 0.326171875,
"learning_rate": 8.880953707176514e-05,
"loss": 0.8915,
"step": 3615
},
{
"epoch": 8.733413751507841,
"grad_norm": 0.369140625,
"learning_rate": 8.853029552437259e-05,
"loss": 0.8906,
"step": 3620
},
{
"epoch": 8.745476477683956,
"grad_norm": 0.34765625,
"learning_rate": 8.825114457599527e-05,
"loss": 0.8869,
"step": 3625
},
{
"epoch": 8.757539203860073,
"grad_norm": 0.32421875,
"learning_rate": 8.797208643164212e-05,
"loss": 0.8998,
"step": 3630
},
{
"epoch": 8.769601930036188,
"grad_norm": 0.330078125,
"learning_rate": 8.7693123295589e-05,
"loss": 0.8776,
"step": 3635
},
{
"epoch": 8.781664656212303,
"grad_norm": 0.337890625,
"learning_rate": 8.741425737136143e-05,
"loss": 0.8951,
"step": 3640
},
{
"epoch": 8.79372738238842,
"grad_norm": 0.337890625,
"learning_rate": 8.713549086171691e-05,
"loss": 0.8865,
"step": 3645
},
{
"epoch": 8.805790108564535,
"grad_norm": 0.3203125,
"learning_rate": 8.685682596862768e-05,
"loss": 0.8861,
"step": 3650
},
{
"epoch": 8.817852834740652,
"grad_norm": 0.35546875,
"learning_rate": 8.65782648932634e-05,
"loss": 0.8918,
"step": 3655
},
{
"epoch": 8.829915560916767,
"grad_norm": 0.330078125,
"learning_rate": 8.629980983597358e-05,
"loss": 0.893,
"step": 3660
},
{
"epoch": 8.841978287092884,
"grad_norm": 0.337890625,
"learning_rate": 8.602146299627035e-05,
"loss": 0.8871,
"step": 3665
},
{
"epoch": 8.854041013268999,
"grad_norm": 0.326171875,
"learning_rate": 8.574322657281105e-05,
"loss": 0.8899,
"step": 3670
},
{
"epoch": 8.866103739445114,
"grad_norm": 0.326171875,
"learning_rate": 8.546510276338078e-05,
"loss": 0.884,
"step": 3675
},
{
"epoch": 8.87816646562123,
"grad_norm": 0.44140625,
"learning_rate": 8.518709376487515e-05,
"loss": 0.8889,
"step": 3680
},
{
"epoch": 8.890229191797346,
"grad_norm": 0.333984375,
"learning_rate": 8.490920177328282e-05,
"loss": 0.8859,
"step": 3685
},
{
"epoch": 8.902291917973463,
"grad_norm": 0.3515625,
"learning_rate": 8.463142898366834e-05,
"loss": 0.8816,
"step": 3690
},
{
"epoch": 8.914354644149578,
"grad_norm": 0.330078125,
"learning_rate": 8.435377759015455e-05,
"loss": 0.8858,
"step": 3695
},
{
"epoch": 8.926417370325694,
"grad_norm": 0.326171875,
"learning_rate": 8.407624978590543e-05,
"loss": 0.8767,
"step": 3700
},
{
"epoch": 8.93848009650181,
"grad_norm": 0.3984375,
"learning_rate": 8.37988477631088e-05,
"loss": 0.8946,
"step": 3705
},
{
"epoch": 8.950542822677924,
"grad_norm": 0.3125,
"learning_rate": 8.352157371295884e-05,
"loss": 0.883,
"step": 3710
},
{
"epoch": 8.962605548854041,
"grad_norm": 0.326171875,
"learning_rate": 8.324442982563887e-05,
"loss": 0.8866,
"step": 3715
},
{
"epoch": 8.974668275030156,
"grad_norm": 0.330078125,
"learning_rate": 8.296741829030418e-05,
"loss": 0.8844,
"step": 3720
},
{
"epoch": 8.986731001206273,
"grad_norm": 0.34375,
"learning_rate": 8.269054129506449e-05,
"loss": 0.88,
"step": 3725
},
{
"epoch": 8.998793727382388,
"grad_norm": 0.341796875,
"learning_rate": 8.241380102696681e-05,
"loss": 0.8762,
"step": 3730
},
{
"epoch": 8.998793727382388,
"eval_loss": 2.493363618850708,
"eval_runtime": 0.66,
"eval_samples_per_second": 15.152,
"eval_steps_per_second": 1.515,
"step": 3730
},
{
"epoch": 9.010856453558505,
"grad_norm": 0.36328125,
"learning_rate": 8.213719967197817e-05,
"loss": 0.8809,
"step": 3735
},
{
"epoch": 9.02291917973462,
"grad_norm": 0.322265625,
"learning_rate": 8.186073941496835e-05,
"loss": 0.8795,
"step": 3740
},
{
"epoch": 9.034981905910735,
"grad_norm": 0.33203125,
"learning_rate": 8.158442243969256e-05,
"loss": 0.8773,
"step": 3745
},
{
"epoch": 9.047044632086852,
"grad_norm": 0.3359375,
"learning_rate": 8.130825092877418e-05,
"loss": 0.8885,
"step": 3750
},
{
"epoch": 9.059107358262967,
"grad_norm": 0.345703125,
"learning_rate": 8.103222706368774e-05,
"loss": 0.8857,
"step": 3755
},
{
"epoch": 9.071170084439084,
"grad_norm": 0.322265625,
"learning_rate": 8.075635302474134e-05,
"loss": 0.8747,
"step": 3760
},
{
"epoch": 9.083232810615199,
"grad_norm": 0.34765625,
"learning_rate": 8.04806309910597e-05,
"loss": 0.879,
"step": 3765
},
{
"epoch": 9.095295536791316,
"grad_norm": 0.35546875,
"learning_rate": 8.020506314056677e-05,
"loss": 0.878,
"step": 3770
},
{
"epoch": 9.10735826296743,
"grad_norm": 0.322265625,
"learning_rate": 7.992965164996875e-05,
"loss": 0.8809,
"step": 3775
},
{
"epoch": 9.119420989143546,
"grad_norm": 0.35546875,
"learning_rate": 7.965439869473664e-05,
"loss": 0.8815,
"step": 3780
},
{
"epoch": 9.131483715319662,
"grad_norm": 0.35546875,
"learning_rate": 7.937930644908917e-05,
"loss": 0.8826,
"step": 3785
},
{
"epoch": 9.143546441495777,
"grad_norm": 0.345703125,
"learning_rate": 7.91043770859757e-05,
"loss": 0.8863,
"step": 3790
},
{
"epoch": 9.155609167671894,
"grad_norm": 0.40234375,
"learning_rate": 7.882961277705895e-05,
"loss": 0.8792,
"step": 3795
},
{
"epoch": 9.16767189384801,
"grad_norm": 0.326171875,
"learning_rate": 7.855501569269782e-05,
"loss": 0.8743,
"step": 3800
},
{
"epoch": 9.179734620024126,
"grad_norm": 0.314453125,
"learning_rate": 7.82805880019304e-05,
"loss": 0.881,
"step": 3805
},
{
"epoch": 9.191797346200241,
"grad_norm": 0.3203125,
"learning_rate": 7.800633187245673e-05,
"loss": 0.8873,
"step": 3810
},
{
"epoch": 9.203860072376358,
"grad_norm": 0.341796875,
"learning_rate": 7.773224947062163e-05,
"loss": 0.8766,
"step": 3815
},
{
"epoch": 9.215922798552473,
"grad_norm": 0.384765625,
"learning_rate": 7.745834296139763e-05,
"loss": 0.8818,
"step": 3820
},
{
"epoch": 9.227985524728588,
"grad_norm": 0.328125,
"learning_rate": 7.718461450836804e-05,
"loss": 0.8894,
"step": 3825
},
{
"epoch": 9.240048250904705,
"grad_norm": 0.314453125,
"learning_rate": 7.691106627370955e-05,
"loss": 0.8866,
"step": 3830
},
{
"epoch": 9.25211097708082,
"grad_norm": 0.341796875,
"learning_rate": 7.663770041817534e-05,
"loss": 0.8854,
"step": 3835
},
{
"epoch": 9.264173703256937,
"grad_norm": 0.328125,
"learning_rate": 7.636451910107806e-05,
"loss": 0.8832,
"step": 3840
},
{
"epoch": 9.276236429433052,
"grad_norm": 0.33984375,
"learning_rate": 7.609152448027261e-05,
"loss": 0.8853,
"step": 3845
},
{
"epoch": 9.288299155609169,
"grad_norm": 0.33984375,
"learning_rate": 7.581871871213913e-05,
"loss": 0.8826,
"step": 3850
},
{
"epoch": 9.300361881785284,
"grad_norm": 0.384765625,
"learning_rate": 7.554610395156624e-05,
"loss": 0.8851,
"step": 3855
},
{
"epoch": 9.312424607961399,
"grad_norm": 0.32421875,
"learning_rate": 7.527368235193355e-05,
"loss": 0.8812,
"step": 3860
},
{
"epoch": 9.324487334137515,
"grad_norm": 0.412109375,
"learning_rate": 7.500145606509501e-05,
"loss": 0.8883,
"step": 3865
},
{
"epoch": 9.33655006031363,
"grad_norm": 0.337890625,
"learning_rate": 7.472942724136174e-05,
"loss": 0.885,
"step": 3870
},
{
"epoch": 9.348612786489747,
"grad_norm": 0.337890625,
"learning_rate": 7.44575980294852e-05,
"loss": 0.8857,
"step": 3875
},
{
"epoch": 9.360675512665862,
"grad_norm": 0.369140625,
"learning_rate": 7.418597057664005e-05,
"loss": 0.8872,
"step": 3880
},
{
"epoch": 9.372738238841979,
"grad_norm": 0.3125,
"learning_rate": 7.391454702840722e-05,
"loss": 0.8826,
"step": 3885
},
{
"epoch": 9.384800965018094,
"grad_norm": 0.314453125,
"learning_rate": 7.36433295287571e-05,
"loss": 0.8822,
"step": 3890
},
{
"epoch": 9.39686369119421,
"grad_norm": 0.34765625,
"learning_rate": 7.337232022003244e-05,
"loss": 0.8833,
"step": 3895
},
{
"epoch": 9.408926417370326,
"grad_norm": 0.359375,
"learning_rate": 7.310152124293146e-05,
"loss": 0.8782,
"step": 3900
},
{
"epoch": 9.420989143546441,
"grad_norm": 0.314453125,
"learning_rate": 7.28309347364911e-05,
"loss": 0.8879,
"step": 3905
},
{
"epoch": 9.433051869722558,
"grad_norm": 0.326171875,
"learning_rate": 7.256056283806986e-05,
"loss": 0.8806,
"step": 3910
},
{
"epoch": 9.445114595898673,
"grad_norm": 0.33984375,
"learning_rate": 7.229040768333115e-05,
"loss": 0.8841,
"step": 3915
},
{
"epoch": 9.45717732207479,
"grad_norm": 0.314453125,
"learning_rate": 7.202047140622621e-05,
"loss": 0.8718,
"step": 3920
},
{
"epoch": 9.469240048250905,
"grad_norm": 0.341796875,
"learning_rate": 7.175075613897756e-05,
"loss": 0.8843,
"step": 3925
},
{
"epoch": 9.48130277442702,
"grad_norm": 0.34765625,
"learning_rate": 7.14812640120618e-05,
"loss": 0.8788,
"step": 3930
},
{
"epoch": 9.493365500603137,
"grad_norm": 0.33984375,
"learning_rate": 7.121199715419295e-05,
"loss": 0.8891,
"step": 3935
},
{
"epoch": 9.505428226779252,
"grad_norm": 0.3359375,
"learning_rate": 7.094295769230577e-05,
"loss": 0.8808,
"step": 3940
},
{
"epoch": 9.517490952955368,
"grad_norm": 0.306640625,
"learning_rate": 7.067414775153871e-05,
"loss": 0.8817,
"step": 3945
},
{
"epoch": 9.529553679131483,
"grad_norm": 0.359375,
"learning_rate": 7.040556945521721e-05,
"loss": 0.8837,
"step": 3950
},
{
"epoch": 9.5416164053076,
"grad_norm": 0.30859375,
"learning_rate": 7.013722492483708e-05,
"loss": 0.8867,
"step": 3955
},
{
"epoch": 9.553679131483715,
"grad_norm": 0.31640625,
"learning_rate": 6.986911628004753e-05,
"loss": 0.8807,
"step": 3960
},
{
"epoch": 9.56574185765983,
"grad_norm": 0.306640625,
"learning_rate": 6.96012456386345e-05,
"loss": 0.873,
"step": 3965
},
{
"epoch": 9.577804583835947,
"grad_norm": 0.330078125,
"learning_rate": 6.933361511650395e-05,
"loss": 0.8852,
"step": 3970
},
{
"epoch": 9.589867310012062,
"grad_norm": 0.326171875,
"learning_rate": 6.906622682766526e-05,
"loss": 0.8914,
"step": 3975
},
{
"epoch": 9.601930036188179,
"grad_norm": 0.357421875,
"learning_rate": 6.879908288421425e-05,
"loss": 0.8691,
"step": 3980
},
{
"epoch": 9.613992762364294,
"grad_norm": 0.341796875,
"learning_rate": 6.853218539631672e-05,
"loss": 0.8839,
"step": 3985
},
{
"epoch": 9.62605548854041,
"grad_norm": 0.310546875,
"learning_rate": 6.826553647219175e-05,
"loss": 0.8879,
"step": 3990
},
{
"epoch": 9.638118214716526,
"grad_norm": 0.349609375,
"learning_rate": 6.799913821809501e-05,
"loss": 0.8807,
"step": 3995
},
{
"epoch": 9.650180940892643,
"grad_norm": 0.34375,
"learning_rate": 6.773299273830207e-05,
"loss": 0.8771,
"step": 4000
},
{
"epoch": 9.662243667068758,
"grad_norm": 0.341796875,
"learning_rate": 6.74671021350919e-05,
"loss": 0.8814,
"step": 4005
},
{
"epoch": 9.674306393244873,
"grad_norm": 0.318359375,
"learning_rate": 6.720146850873025e-05,
"loss": 0.8844,
"step": 4010
},
{
"epoch": 9.68636911942099,
"grad_norm": 0.296875,
"learning_rate": 6.69360939574529e-05,
"loss": 0.8801,
"step": 4015
},
{
"epoch": 9.698431845597105,
"grad_norm": 0.31640625,
"learning_rate": 6.667098057744927e-05,
"loss": 0.8837,
"step": 4020
},
{
"epoch": 9.710494571773221,
"grad_norm": 0.314453125,
"learning_rate": 6.640613046284581e-05,
"loss": 0.8821,
"step": 4025
},
{
"epoch": 9.722557297949336,
"grad_norm": 0.310546875,
"learning_rate": 6.614154570568934e-05,
"loss": 0.89,
"step": 4030
},
{
"epoch": 9.734620024125451,
"grad_norm": 0.322265625,
"learning_rate": 6.587722839593073e-05,
"loss": 0.8775,
"step": 4035
},
{
"epoch": 9.746682750301568,
"grad_norm": 0.31640625,
"learning_rate": 6.56131806214083e-05,
"loss": 0.8802,
"step": 4040
},
{
"epoch": 9.758745476477683,
"grad_norm": 0.373046875,
"learning_rate": 6.534940446783122e-05,
"loss": 0.8763,
"step": 4045
},
{
"epoch": 9.7708082026538,
"grad_norm": 0.435546875,
"learning_rate": 6.508590201876317e-05,
"loss": 0.8889,
"step": 4050
},
{
"epoch": 9.782870928829915,
"grad_norm": 0.33984375,
"learning_rate": 6.482267535560583e-05,
"loss": 0.8794,
"step": 4055
},
{
"epoch": 9.794933655006032,
"grad_norm": 0.3203125,
"learning_rate": 6.455972655758253e-05,
"loss": 0.8911,
"step": 4060
},
{
"epoch": 9.806996381182147,
"grad_norm": 0.322265625,
"learning_rate": 6.429705770172168e-05,
"loss": 0.8822,
"step": 4065
},
{
"epoch": 9.819059107358264,
"grad_norm": 0.32421875,
"learning_rate": 6.403467086284041e-05,
"loss": 0.881,
"step": 4070
},
{
"epoch": 9.831121833534379,
"grad_norm": 0.30078125,
"learning_rate": 6.377256811352825e-05,
"loss": 0.8814,
"step": 4075
},
{
"epoch": 9.843184559710494,
"grad_norm": 0.328125,
"learning_rate": 6.351075152413068e-05,
"loss": 0.8883,
"step": 4080
},
{
"epoch": 9.85524728588661,
"grad_norm": 0.322265625,
"learning_rate": 6.324922316273282e-05,
"loss": 0.8806,
"step": 4085
},
{
"epoch": 9.867310012062726,
"grad_norm": 0.31640625,
"learning_rate": 6.298798509514312e-05,
"loss": 0.8843,
"step": 4090
},
{
"epoch": 9.879372738238843,
"grad_norm": 0.353515625,
"learning_rate": 6.272703938487694e-05,
"loss": 0.8865,
"step": 4095
},
{
"epoch": 9.891435464414958,
"grad_norm": 0.3515625,
"learning_rate": 6.246638809314036e-05,
"loss": 0.882,
"step": 4100
},
{
"epoch": 9.903498190591074,
"grad_norm": 0.322265625,
"learning_rate": 6.220603327881375e-05,
"loss": 0.8865,
"step": 4105
},
{
"epoch": 9.91556091676719,
"grad_norm": 0.318359375,
"learning_rate": 6.194597699843581e-05,
"loss": 0.8814,
"step": 4110
},
{
"epoch": 9.927623642943304,
"grad_norm": 0.3515625,
"learning_rate": 6.168622130618694e-05,
"loss": 0.8799,
"step": 4115
},
{
"epoch": 9.939686369119421,
"grad_norm": 0.3046875,
"learning_rate": 6.142676825387328e-05,
"loss": 0.8835,
"step": 4120
},
{
"epoch": 9.951749095295536,
"grad_norm": 0.32421875,
"learning_rate": 6.116761989091042e-05,
"loss": 0.882,
"step": 4125
},
{
"epoch": 9.963811821471653,
"grad_norm": 0.326171875,
"learning_rate": 6.090877826430721e-05,
"loss": 0.8727,
"step": 4130
},
{
"epoch": 9.975874547647768,
"grad_norm": 0.404296875,
"learning_rate": 6.065024541864952e-05,
"loss": 0.8867,
"step": 4135
},
{
"epoch": 9.987937273823885,
"grad_norm": 0.3515625,
"learning_rate": 6.039202339608432e-05,
"loss": 0.8825,
"step": 4140
},
{
"epoch": 10.0,
"grad_norm": 0.333984375,
"learning_rate": 6.0134114236303275e-05,
"loss": 0.8821,
"step": 4145
},
{
"epoch": 10.0,
"eval_loss": 2.4973604679107666,
"eval_runtime": 0.5303,
"eval_samples_per_second": 18.857,
"eval_steps_per_second": 1.886,
"step": 4145
},
{
"epoch": 10.012062726176115,
"grad_norm": 0.337890625,
"learning_rate": 5.987651997652677e-05,
"loss": 0.8757,
"step": 4150
},
{
"epoch": 10.024125452352232,
"grad_norm": 0.3125,
"learning_rate": 5.961924265148777e-05,
"loss": 0.8787,
"step": 4155
},
{
"epoch": 10.036188178528347,
"grad_norm": 0.326171875,
"learning_rate": 5.93622842934159e-05,
"loss": 0.8815,
"step": 4160
},
{
"epoch": 10.048250904704464,
"grad_norm": 0.3515625,
"learning_rate": 5.9105646932021155e-05,
"loss": 0.8806,
"step": 4165
},
{
"epoch": 10.060313630880579,
"grad_norm": 0.314453125,
"learning_rate": 5.884933259447798e-05,
"loss": 0.8848,
"step": 4170
},
{
"epoch": 10.072376357056696,
"grad_norm": 0.31640625,
"learning_rate": 5.8593343305409355e-05,
"loss": 0.8788,
"step": 4175
},
{
"epoch": 10.08443908323281,
"grad_norm": 0.310546875,
"learning_rate": 5.833768108687063e-05,
"loss": 0.8794,
"step": 4180
},
{
"epoch": 10.096501809408926,
"grad_norm": 0.31640625,
"learning_rate": 5.8082347958333625e-05,
"loss": 0.8835,
"step": 4185
},
{
"epoch": 10.108564535585042,
"grad_norm": 0.359375,
"learning_rate": 5.7827345936670674e-05,
"loss": 0.872,
"step": 4190
},
{
"epoch": 10.120627261761157,
"grad_norm": 0.333984375,
"learning_rate": 5.7572677036138826e-05,
"loss": 0.8782,
"step": 4195
},
{
"epoch": 10.132689987937274,
"grad_norm": 0.330078125,
"learning_rate": 5.731834326836366e-05,
"loss": 0.8814,
"step": 4200
},
{
"epoch": 10.14475271411339,
"grad_norm": 0.369140625,
"learning_rate": 5.7064346642323584e-05,
"loss": 0.8783,
"step": 4205
},
{
"epoch": 10.156815440289506,
"grad_norm": 0.357421875,
"learning_rate": 5.681068916433403e-05,
"loss": 0.8805,
"step": 4210
},
{
"epoch": 10.168878166465621,
"grad_norm": 0.314453125,
"learning_rate": 5.6557372838031384e-05,
"loss": 0.8792,
"step": 4215
},
{
"epoch": 10.180940892641736,
"grad_norm": 0.34765625,
"learning_rate": 5.6304399664357275e-05,
"loss": 0.8759,
"step": 4220
},
{
"epoch": 10.193003618817853,
"grad_norm": 0.3125,
"learning_rate": 5.605177164154293e-05,
"loss": 0.8775,
"step": 4225
},
{
"epoch": 10.205066344993968,
"grad_norm": 0.3203125,
"learning_rate": 5.579949076509305e-05,
"loss": 0.8815,
"step": 4230
},
{
"epoch": 10.217129071170085,
"grad_norm": 0.310546875,
"learning_rate": 5.5547559027770325e-05,
"loss": 0.863,
"step": 4235
},
{
"epoch": 10.2291917973462,
"grad_norm": 0.32421875,
"learning_rate": 5.529597841957959e-05,
"loss": 0.8743,
"step": 4240
},
{
"epoch": 10.241254523522317,
"grad_norm": 0.3515625,
"learning_rate": 5.5044750927752106e-05,
"loss": 0.8785,
"step": 4245
},
{
"epoch": 10.253317249698432,
"grad_norm": 0.3203125,
"learning_rate": 5.479387853672988e-05,
"loss": 0.8758,
"step": 4250
},
{
"epoch": 10.265379975874549,
"grad_norm": 0.31640625,
"learning_rate": 5.4543363228149946e-05,
"loss": 0.8833,
"step": 4255
},
{
"epoch": 10.277442702050664,
"grad_norm": 0.349609375,
"learning_rate": 5.429320698082887e-05,
"loss": 0.88,
"step": 4260
},
{
"epoch": 10.289505428226779,
"grad_norm": 0.326171875,
"learning_rate": 5.404341177074686e-05,
"loss": 0.876,
"step": 4265
},
{
"epoch": 10.301568154402895,
"grad_norm": 0.322265625,
"learning_rate": 5.379397957103231e-05,
"loss": 0.873,
"step": 4270
},
{
"epoch": 10.31363088057901,
"grad_norm": 0.337890625,
"learning_rate": 5.354491235194635e-05,
"loss": 0.8816,
"step": 4275
},
{
"epoch": 10.325693606755127,
"grad_norm": 0.3359375,
"learning_rate": 5.3296212080866967e-05,
"loss": 0.8828,
"step": 4280
},
{
"epoch": 10.337756332931242,
"grad_norm": 0.330078125,
"learning_rate": 5.304788072227367e-05,
"loss": 0.8825,
"step": 4285
},
{
"epoch": 10.349819059107359,
"grad_norm": 0.302734375,
"learning_rate": 5.279992023773195e-05,
"loss": 0.8735,
"step": 4290
},
{
"epoch": 10.361881785283474,
"grad_norm": 0.31640625,
"learning_rate": 5.255233258587784e-05,
"loss": 0.881,
"step": 4295
},
{
"epoch": 10.37394451145959,
"grad_norm": 0.31640625,
"learning_rate": 5.2305119722402254e-05,
"loss": 0.8755,
"step": 4300
},
{
"epoch": 10.386007237635706,
"grad_norm": 0.328125,
"learning_rate": 5.205828360003568e-05,
"loss": 0.8776,
"step": 4305
},
{
"epoch": 10.398069963811821,
"grad_norm": 0.3125,
"learning_rate": 5.181182616853286e-05,
"loss": 0.8838,
"step": 4310
},
{
"epoch": 10.410132689987938,
"grad_norm": 0.314453125,
"learning_rate": 5.15657493746571e-05,
"loss": 0.8817,
"step": 4315
},
{
"epoch": 10.422195416164053,
"grad_norm": 0.34765625,
"learning_rate": 5.1320055162165115e-05,
"loss": 0.8752,
"step": 4320
},
{
"epoch": 10.43425814234017,
"grad_norm": 0.337890625,
"learning_rate": 5.10747454717917e-05,
"loss": 0.8791,
"step": 4325
},
{
"epoch": 10.446320868516285,
"grad_norm": 0.318359375,
"learning_rate": 5.082982224123418e-05,
"loss": 0.8745,
"step": 4330
},
{
"epoch": 10.4583835946924,
"grad_norm": 0.349609375,
"learning_rate": 5.0585287405137305e-05,
"loss": 0.8712,
"step": 4335
},
{
"epoch": 10.470446320868517,
"grad_norm": 0.322265625,
"learning_rate": 5.034114289507792e-05,
"loss": 0.871,
"step": 4340
},
{
"epoch": 10.482509047044632,
"grad_norm": 0.31640625,
"learning_rate": 5.009739063954964e-05,
"loss": 0.8759,
"step": 4345
},
{
"epoch": 10.494571773220748,
"grad_norm": 0.33203125,
"learning_rate": 4.9854032563947714e-05,
"loss": 0.8772,
"step": 4350
},
{
"epoch": 10.506634499396863,
"grad_norm": 0.306640625,
"learning_rate": 4.9611070590553724e-05,
"loss": 0.8886,
"step": 4355
},
{
"epoch": 10.51869722557298,
"grad_norm": 0.341796875,
"learning_rate": 4.936850663852053e-05,
"loss": 0.8729,
"step": 4360
},
{
"epoch": 10.530759951749095,
"grad_norm": 0.337890625,
"learning_rate": 4.912634262385695e-05,
"loss": 0.8781,
"step": 4365
},
{
"epoch": 10.54282267792521,
"grad_norm": 0.3125,
"learning_rate": 4.888458045941269e-05,
"loss": 0.8788,
"step": 4370
},
{
"epoch": 10.554885404101327,
"grad_norm": 0.330078125,
"learning_rate": 4.864322205486335e-05,
"loss": 0.8897,
"step": 4375
},
{
"epoch": 10.566948130277442,
"grad_norm": 0.314453125,
"learning_rate": 4.8402269316695134e-05,
"loss": 0.878,
"step": 4380
},
{
"epoch": 10.579010856453559,
"grad_norm": 0.31640625,
"learning_rate": 4.8161724148189934e-05,
"loss": 0.8773,
"step": 4385
},
{
"epoch": 10.591073582629674,
"grad_norm": 0.330078125,
"learning_rate": 4.792158844941023e-05,
"loss": 0.8813,
"step": 4390
},
{
"epoch": 10.60313630880579,
"grad_norm": 0.328125,
"learning_rate": 4.768186411718417e-05,
"loss": 0.8865,
"step": 4395
},
{
"epoch": 10.615199034981906,
"grad_norm": 0.3203125,
"learning_rate": 4.7442553045090474e-05,
"loss": 0.8789,
"step": 4400
},
{
"epoch": 10.627261761158021,
"grad_norm": 0.318359375,
"learning_rate": 4.720365712344349e-05,
"loss": 0.8774,
"step": 4405
},
{
"epoch": 10.639324487334138,
"grad_norm": 0.373046875,
"learning_rate": 4.696517823927842e-05,
"loss": 0.8797,
"step": 4410
},
{
"epoch": 10.651387213510253,
"grad_norm": 0.3359375,
"learning_rate": 4.672711827633618e-05,
"loss": 0.877,
"step": 4415
},
{
"epoch": 10.66344993968637,
"grad_norm": 0.353515625,
"learning_rate": 4.6489479115048676e-05,
"loss": 0.8766,
"step": 4420
},
{
"epoch": 10.675512665862485,
"grad_norm": 0.33984375,
"learning_rate": 4.625226263252386e-05,
"loss": 0.8804,
"step": 4425
},
{
"epoch": 10.687575392038601,
"grad_norm": 0.3203125,
"learning_rate": 4.601547070253107e-05,
"loss": 0.8889,
"step": 4430
},
{
"epoch": 10.699638118214716,
"grad_norm": 0.341796875,
"learning_rate": 4.577910519548602e-05,
"loss": 0.8747,
"step": 4435
},
{
"epoch": 10.711700844390833,
"grad_norm": 0.3515625,
"learning_rate": 4.554316797843609e-05,
"loss": 0.8816,
"step": 4440
},
{
"epoch": 10.723763570566948,
"grad_norm": 0.32421875,
"learning_rate": 4.53076609150457e-05,
"loss": 0.8796,
"step": 4445
},
{
"epoch": 10.735826296743063,
"grad_norm": 0.328125,
"learning_rate": 4.5072585865581395e-05,
"loss": 0.8836,
"step": 4450
},
{
"epoch": 10.74788902291918,
"grad_norm": 0.3203125,
"learning_rate": 4.483794468689728e-05,
"loss": 0.8779,
"step": 4455
},
{
"epoch": 10.759951749095295,
"grad_norm": 0.326171875,
"learning_rate": 4.460373923242041e-05,
"loss": 0.8766,
"step": 4460
},
{
"epoch": 10.772014475271412,
"grad_norm": 0.322265625,
"learning_rate": 4.436997135213592e-05,
"loss": 0.8831,
"step": 4465
},
{
"epoch": 10.784077201447527,
"grad_norm": 0.3046875,
"learning_rate": 4.413664289257265e-05,
"loss": 0.8775,
"step": 4470
},
{
"epoch": 10.796139927623642,
"grad_norm": 0.3203125,
"learning_rate": 4.3903755696788386e-05,
"loss": 0.8744,
"step": 4475
},
{
"epoch": 10.808202653799759,
"grad_norm": 0.306640625,
"learning_rate": 4.3671311604355494e-05,
"loss": 0.8816,
"step": 4480
},
{
"epoch": 10.820265379975874,
"grad_norm": 0.3359375,
"learning_rate": 4.343931245134616e-05,
"loss": 0.8757,
"step": 4485
},
{
"epoch": 10.83232810615199,
"grad_norm": 0.34375,
"learning_rate": 4.3207760070318026e-05,
"loss": 0.8827,
"step": 4490
},
{
"epoch": 10.844390832328106,
"grad_norm": 0.32421875,
"learning_rate": 4.2976656290299764e-05,
"loss": 0.8823,
"step": 4495
},
{
"epoch": 10.856453558504223,
"grad_norm": 0.326171875,
"learning_rate": 4.274600293677647e-05,
"loss": 0.8856,
"step": 4500
},
{
"epoch": 10.868516284680338,
"grad_norm": 0.326171875,
"learning_rate": 4.2515801831675305e-05,
"loss": 0.8786,
"step": 4505
},
{
"epoch": 10.880579010856454,
"grad_norm": 0.32421875,
"learning_rate": 4.228605479335127e-05,
"loss": 0.8808,
"step": 4510
},
{
"epoch": 10.89264173703257,
"grad_norm": 0.318359375,
"learning_rate": 4.2056763636572574e-05,
"loss": 0.8746,
"step": 4515
},
{
"epoch": 10.904704463208684,
"grad_norm": 0.330078125,
"learning_rate": 4.182793017250643e-05,
"loss": 0.8731,
"step": 4520
},
{
"epoch": 10.916767189384801,
"grad_norm": 0.32421875,
"learning_rate": 4.1599556208704785e-05,
"loss": 0.8791,
"step": 4525
},
{
"epoch": 10.928829915560916,
"grad_norm": 0.310546875,
"learning_rate": 4.137164354908999e-05,
"loss": 0.8785,
"step": 4530
},
{
"epoch": 10.940892641737033,
"grad_norm": 0.310546875,
"learning_rate": 4.114419399394052e-05,
"loss": 0.88,
"step": 4535
},
{
"epoch": 10.952955367913148,
"grad_norm": 0.32421875,
"learning_rate": 4.0917209339876896e-05,
"loss": 0.8801,
"step": 4540
},
{
"epoch": 10.965018094089265,
"grad_norm": 0.3359375,
"learning_rate": 4.069069137984731e-05,
"loss": 0.8809,
"step": 4545
},
{
"epoch": 10.97708082026538,
"grad_norm": 0.318359375,
"learning_rate": 4.046464190311355e-05,
"loss": 0.8911,
"step": 4550
},
{
"epoch": 10.989143546441495,
"grad_norm": 0.333984375,
"learning_rate": 4.023906269523686e-05,
"loss": 0.8697,
"step": 4555
},
{
"epoch": 10.998793727382388,
"eval_loss": 2.5013487339019775,
"eval_runtime": 0.6548,
"eval_samples_per_second": 15.273,
"eval_steps_per_second": 1.527,
"step": 4559
},
{
"epoch": 11.001206272617612,
"grad_norm": 0.322265625,
"learning_rate": 4.001395553806391e-05,
"loss": 0.8789,
"step": 4560
},
{
"epoch": 11.013268998793727,
"grad_norm": 0.33203125,
"learning_rate": 3.9789322209712566e-05,
"loss": 0.8779,
"step": 4565
},
{
"epoch": 11.025331724969844,
"grad_norm": 0.34375,
"learning_rate": 3.956516448455794e-05,
"loss": 0.8665,
"step": 4570
},
{
"epoch": 11.037394451145959,
"grad_norm": 0.33203125,
"learning_rate": 3.9341484133218366e-05,
"loss": 0.8736,
"step": 4575
},
{
"epoch": 11.049457177322076,
"grad_norm": 0.3359375,
"learning_rate": 3.9118282922541474e-05,
"loss": 0.8686,
"step": 4580
},
{
"epoch": 11.06151990349819,
"grad_norm": 0.314453125,
"learning_rate": 3.8895562615590075e-05,
"loss": 0.8758,
"step": 4585
},
{
"epoch": 11.073582629674306,
"grad_norm": 0.302734375,
"learning_rate": 3.8673324971628357e-05,
"loss": 0.8722,
"step": 4590
},
{
"epoch": 11.085645355850422,
"grad_norm": 0.3046875,
"learning_rate": 3.8451571746108027e-05,
"loss": 0.8769,
"step": 4595
},
{
"epoch": 11.097708082026537,
"grad_norm": 0.326171875,
"learning_rate": 3.8230304690654304e-05,
"loss": 0.8696,
"step": 4600
},
{
"epoch": 11.109770808202654,
"grad_norm": 0.328125,
"learning_rate": 3.800952555305216e-05,
"loss": 0.8767,
"step": 4605
},
{
"epoch": 11.12183353437877,
"grad_norm": 0.337890625,
"learning_rate": 3.778923607723252e-05,
"loss": 0.8717,
"step": 4610
},
{
"epoch": 11.133896260554886,
"grad_norm": 0.341796875,
"learning_rate": 3.756943800325852e-05,
"loss": 0.8716,
"step": 4615
},
{
"epoch": 11.145958986731001,
"grad_norm": 0.345703125,
"learning_rate": 3.7350133067311686e-05,
"loss": 0.8779,
"step": 4620
},
{
"epoch": 11.158021712907116,
"grad_norm": 0.34375,
"learning_rate": 3.7131323001678254e-05,
"loss": 0.8674,
"step": 4625
},
{
"epoch": 11.170084439083233,
"grad_norm": 0.326171875,
"learning_rate": 3.691300953473548e-05,
"loss": 0.8784,
"step": 4630
},
{
"epoch": 11.182147165259348,
"grad_norm": 0.328125,
"learning_rate": 3.669519439093801e-05,
"loss": 0.8854,
"step": 4635
},
{
"epoch": 11.194209891435465,
"grad_norm": 0.314453125,
"learning_rate": 3.6477879290804206e-05,
"loss": 0.8706,
"step": 4640
},
{
"epoch": 11.20627261761158,
"grad_norm": 0.326171875,
"learning_rate": 3.626106595090268e-05,
"loss": 0.8776,
"step": 4645
},
{
"epoch": 11.218335343787697,
"grad_norm": 0.357421875,
"learning_rate": 3.604475608383858e-05,
"loss": 0.8725,
"step": 4650
},
{
"epoch": 11.230398069963812,
"grad_norm": 0.314453125,
"learning_rate": 3.582895139824014e-05,
"loss": 0.8751,
"step": 4655
},
{
"epoch": 11.242460796139927,
"grad_norm": 0.33203125,
"learning_rate": 3.561365359874513e-05,
"loss": 0.8789,
"step": 4660
},
{
"epoch": 11.254523522316044,
"grad_norm": 0.310546875,
"learning_rate": 3.539886438598756e-05,
"loss": 0.8773,
"step": 4665
},
{
"epoch": 11.266586248492159,
"grad_norm": 0.314453125,
"learning_rate": 3.518458545658401e-05,
"loss": 0.877,
"step": 4670
},
{
"epoch": 11.278648974668275,
"grad_norm": 0.310546875,
"learning_rate": 3.497081850312033e-05,
"loss": 0.8754,
"step": 4675
},
{
"epoch": 11.29071170084439,
"grad_norm": 0.322265625,
"learning_rate": 3.475756521413839e-05,
"loss": 0.8763,
"step": 4680
},
{
"epoch": 11.302774427020507,
"grad_norm": 0.31640625,
"learning_rate": 3.454482727412254e-05,
"loss": 0.8778,
"step": 4685
},
{
"epoch": 11.314837153196622,
"grad_norm": 0.330078125,
"learning_rate": 3.433260636348639e-05,
"loss": 0.8784,
"step": 4690
},
{
"epoch": 11.32689987937274,
"grad_norm": 0.3125,
"learning_rate": 3.412090415855963e-05,
"loss": 0.8745,
"step": 4695
},
{
"epoch": 11.338962605548854,
"grad_norm": 0.32421875,
"learning_rate": 3.3909722331574623e-05,
"loss": 0.8756,
"step": 4700
},
{
"epoch": 11.35102533172497,
"grad_norm": 0.318359375,
"learning_rate": 3.369906255065328e-05,
"loss": 0.8827,
"step": 4705
},
{
"epoch": 11.363088057901086,
"grad_norm": 0.328125,
"learning_rate": 3.348892647979389e-05,
"loss": 0.8763,
"step": 4710
},
{
"epoch": 11.375150784077201,
"grad_norm": 0.31640625,
"learning_rate": 3.3279315778858036e-05,
"loss": 0.8815,
"step": 4715
},
{
"epoch": 11.387213510253318,
"grad_norm": 0.306640625,
"learning_rate": 3.307023210355729e-05,
"loss": 0.8787,
"step": 4720
},
{
"epoch": 11.399276236429433,
"grad_norm": 0.318359375,
"learning_rate": 3.2861677105440336e-05,
"loss": 0.8794,
"step": 4725
},
{
"epoch": 11.41133896260555,
"grad_norm": 0.314453125,
"learning_rate": 3.265365243187981e-05,
"loss": 0.8753,
"step": 4730
},
{
"epoch": 11.423401688781665,
"grad_norm": 0.349609375,
"learning_rate": 3.2446159726059345e-05,
"loss": 0.8819,
"step": 4735
},
{
"epoch": 11.43546441495778,
"grad_norm": 0.318359375,
"learning_rate": 3.223920062696052e-05,
"loss": 0.8727,
"step": 4740
},
{
"epoch": 11.447527141133897,
"grad_norm": 0.322265625,
"learning_rate": 3.203277676935007e-05,
"loss": 0.8743,
"step": 4745
},
{
"epoch": 11.459589867310012,
"grad_norm": 0.314453125,
"learning_rate": 3.182688978376678e-05,
"loss": 0.8777,
"step": 4750
},
{
"epoch": 11.471652593486128,
"grad_norm": 0.31640625,
"learning_rate": 3.1621541296508695e-05,
"loss": 0.8759,
"step": 4755
},
{
"epoch": 11.483715319662243,
"grad_norm": 0.318359375,
"learning_rate": 3.141673292962026e-05,
"loss": 0.8693,
"step": 4760
},
{
"epoch": 11.49577804583836,
"grad_norm": 0.337890625,
"learning_rate": 3.121246630087961e-05,
"loss": 0.8747,
"step": 4765
},
{
"epoch": 11.507840772014475,
"grad_norm": 0.314453125,
"learning_rate": 3.100874302378559e-05,
"loss": 0.8789,
"step": 4770
},
{
"epoch": 11.51990349819059,
"grad_norm": 0.3203125,
"learning_rate": 3.080556470754513e-05,
"loss": 0.8858,
"step": 4775
},
{
"epoch": 11.531966224366707,
"grad_norm": 0.318359375,
"learning_rate": 3.0602932957060635e-05,
"loss": 0.8751,
"step": 4780
},
{
"epoch": 11.544028950542822,
"grad_norm": 0.337890625,
"learning_rate": 3.0400849372917073e-05,
"loss": 0.8828,
"step": 4785
},
{
"epoch": 11.556091676718939,
"grad_norm": 0.302734375,
"learning_rate": 3.019931555136949e-05,
"loss": 0.8751,
"step": 4790
},
{
"epoch": 11.568154402895054,
"grad_norm": 0.31640625,
"learning_rate": 2.9998333084330333e-05,
"loss": 0.8835,
"step": 4795
},
{
"epoch": 11.58021712907117,
"grad_norm": 0.322265625,
"learning_rate": 2.979790355935703e-05,
"loss": 0.8705,
"step": 4800
},
{
"epoch": 11.592279855247286,
"grad_norm": 0.322265625,
"learning_rate": 2.9598028559639212e-05,
"loss": 0.8677,
"step": 4805
},
{
"epoch": 11.604342581423401,
"grad_norm": 0.318359375,
"learning_rate": 2.9398709663986324e-05,
"loss": 0.8772,
"step": 4810
},
{
"epoch": 11.616405307599518,
"grad_norm": 0.30859375,
"learning_rate": 2.919994844681524e-05,
"loss": 0.8788,
"step": 4815
},
{
"epoch": 11.628468033775633,
"grad_norm": 0.322265625,
"learning_rate": 2.900174647813767e-05,
"loss": 0.8767,
"step": 4820
},
{
"epoch": 11.64053075995175,
"grad_norm": 0.314453125,
"learning_rate": 2.8804105323547837e-05,
"loss": 0.8813,
"step": 4825
},
{
"epoch": 11.652593486127865,
"grad_norm": 0.318359375,
"learning_rate": 2.8607026544210114e-05,
"loss": 0.8797,
"step": 4830
},
{
"epoch": 11.664656212303981,
"grad_norm": 0.330078125,
"learning_rate": 2.8410511696846676e-05,
"loss": 0.8802,
"step": 4835
},
{
"epoch": 11.676718938480096,
"grad_norm": 0.328125,
"learning_rate": 2.8214562333725227e-05,
"loss": 0.8825,
"step": 4840
},
{
"epoch": 11.688781664656211,
"grad_norm": 0.328125,
"learning_rate": 2.801918000264665e-05,
"loss": 0.8741,
"step": 4845
},
{
"epoch": 11.700844390832328,
"grad_norm": 0.314453125,
"learning_rate": 2.7824366246932985e-05,
"loss": 0.8843,
"step": 4850
},
{
"epoch": 11.712907117008443,
"grad_norm": 0.314453125,
"learning_rate": 2.7630122605414988e-05,
"loss": 0.8783,
"step": 4855
},
{
"epoch": 11.72496984318456,
"grad_norm": 0.32421875,
"learning_rate": 2.7436450612420095e-05,
"loss": 0.8717,
"step": 4860
},
{
"epoch": 11.737032569360675,
"grad_norm": 0.33203125,
"learning_rate": 2.7243351797760397e-05,
"loss": 0.8807,
"step": 4865
},
{
"epoch": 11.749095295536792,
"grad_norm": 0.3125,
"learning_rate": 2.705082768672036e-05,
"loss": 0.8715,
"step": 4870
},
{
"epoch": 11.761158021712907,
"grad_norm": 0.349609375,
"learning_rate": 2.6858879800044866e-05,
"loss": 0.8765,
"step": 4875
},
{
"epoch": 11.773220747889024,
"grad_norm": 0.31640625,
"learning_rate": 2.6667509653927334e-05,
"loss": 0.8744,
"step": 4880
},
{
"epoch": 11.785283474065139,
"grad_norm": 0.3125,
"learning_rate": 2.64767187599975e-05,
"loss": 0.8793,
"step": 4885
},
{
"epoch": 11.797346200241254,
"grad_norm": 0.31640625,
"learning_rate": 2.6286508625309624e-05,
"loss": 0.8762,
"step": 4890
},
{
"epoch": 11.80940892641737,
"grad_norm": 0.322265625,
"learning_rate": 2.6096880752330553e-05,
"loss": 0.8839,
"step": 4895
},
{
"epoch": 11.821471652593486,
"grad_norm": 0.33984375,
"learning_rate": 2.5907836638927928e-05,
"loss": 0.8828,
"step": 4900
},
{
"epoch": 11.833534378769603,
"grad_norm": 0.306640625,
"learning_rate": 2.57193777783582e-05,
"loss": 0.8782,
"step": 4905
},
{
"epoch": 11.845597104945718,
"grad_norm": 0.322265625,
"learning_rate": 2.553150565925493e-05,
"loss": 0.8708,
"step": 4910
},
{
"epoch": 11.857659831121833,
"grad_norm": 0.318359375,
"learning_rate": 2.534422176561705e-05,
"loss": 0.879,
"step": 4915
},
{
"epoch": 11.86972255729795,
"grad_norm": 0.314453125,
"learning_rate": 2.515752757679707e-05,
"loss": 0.8808,
"step": 4920
},
{
"epoch": 11.881785283474064,
"grad_norm": 0.353515625,
"learning_rate": 2.4971424567489434e-05,
"loss": 0.8777,
"step": 4925
},
{
"epoch": 11.893848009650181,
"grad_norm": 0.318359375,
"learning_rate": 2.478591420771894e-05,
"loss": 0.875,
"step": 4930
},
{
"epoch": 11.905910735826296,
"grad_norm": 0.30859375,
"learning_rate": 2.4600997962828987e-05,
"loss": 0.8767,
"step": 4935
},
{
"epoch": 11.917973462002413,
"grad_norm": 0.357421875,
"learning_rate": 2.441667729347006e-05,
"loss": 0.8761,
"step": 4940
},
{
"epoch": 11.930036188178528,
"grad_norm": 0.33984375,
"learning_rate": 2.423295365558821e-05,
"loss": 0.8821,
"step": 4945
},
{
"epoch": 11.942098914354645,
"grad_norm": 0.3125,
"learning_rate": 2.404982850041363e-05,
"loss": 0.8829,
"step": 4950
},
{
"epoch": 11.95416164053076,
"grad_norm": 0.3203125,
"learning_rate": 2.3867303274448994e-05,
"loss": 0.8794,
"step": 4955
},
{
"epoch": 11.966224366706875,
"grad_norm": 0.3203125,
"learning_rate": 2.3685379419458164e-05,
"loss": 0.882,
"step": 4960
},
{
"epoch": 11.978287092882992,
"grad_norm": 0.3046875,
"learning_rate": 2.3504058372454884e-05,
"loss": 0.8693,
"step": 4965
},
{
"epoch": 11.990349819059107,
"grad_norm": 0.337890625,
"learning_rate": 2.3323341565691204e-05,
"loss": 0.8729,
"step": 4970
},
{
"epoch": 12.0,
"eval_loss": 2.5031228065490723,
"eval_runtime": 0.5349,
"eval_samples_per_second": 18.696,
"eval_steps_per_second": 1.87,
"step": 4974
},
{
"epoch": 12.002412545235224,
"grad_norm": 0.310546875,
"learning_rate": 2.3143230426646312e-05,
"loss": 0.873,
"step": 4975
},
{
"epoch": 12.014475271411339,
"grad_norm": 0.30859375,
"learning_rate": 2.2963726378015327e-05,
"loss": 0.8753,
"step": 4980
},
{
"epoch": 12.026537997587456,
"grad_norm": 0.32421875,
"learning_rate": 2.278483083769787e-05,
"loss": 0.8806,
"step": 4985
},
{
"epoch": 12.03860072376357,
"grad_norm": 0.298828125,
"learning_rate": 2.2606545218787012e-05,
"loss": 0.8825,
"step": 4990
},
{
"epoch": 12.050663449939686,
"grad_norm": 0.31640625,
"learning_rate": 2.242887092955801e-05,
"loss": 0.8671,
"step": 4995
},
{
"epoch": 12.062726176115802,
"grad_norm": 0.298828125,
"learning_rate": 2.2251809373457345e-05,
"loss": 0.8707,
"step": 5000
},
{
"epoch": 12.074788902291917,
"grad_norm": 0.314453125,
"learning_rate": 2.207536194909142e-05,
"loss": 0.8738,
"step": 5005
},
{
"epoch": 12.086851628468034,
"grad_norm": 0.318359375,
"learning_rate": 2.189953005021569e-05,
"loss": 0.8725,
"step": 5010
},
{
"epoch": 12.09891435464415,
"grad_norm": 0.314453125,
"learning_rate": 2.1724315065723532e-05,
"loss": 0.8718,
"step": 5015
},
{
"epoch": 12.110977080820266,
"grad_norm": 0.314453125,
"learning_rate": 2.1549718379635377e-05,
"loss": 0.8764,
"step": 5020
},
{
"epoch": 12.123039806996381,
"grad_norm": 0.32421875,
"learning_rate": 2.1375741371087677e-05,
"loss": 0.8717,
"step": 5025
},
{
"epoch": 12.135102533172496,
"grad_norm": 0.30859375,
"learning_rate": 2.1202385414322078e-05,
"loss": 0.8799,
"step": 5030
},
{
"epoch": 12.147165259348613,
"grad_norm": 0.318359375,
"learning_rate": 2.1029651878674604e-05,
"loss": 0.8683,
"step": 5035
},
{
"epoch": 12.159227985524728,
"grad_norm": 0.30859375,
"learning_rate": 2.085754212856471e-05,
"loss": 0.8839,
"step": 5040
},
{
"epoch": 12.171290711700845,
"grad_norm": 0.314453125,
"learning_rate": 2.068605752348457e-05,
"loss": 0.8786,
"step": 5045
},
{
"epoch": 12.18335343787696,
"grad_norm": 0.318359375,
"learning_rate": 2.0515199417988452e-05,
"loss": 0.8729,
"step": 5050
},
{
"epoch": 12.195416164053077,
"grad_norm": 0.3203125,
"learning_rate": 2.0344969161681792e-05,
"loss": 0.8784,
"step": 5055
},
{
"epoch": 12.207478890229192,
"grad_norm": 0.310546875,
"learning_rate": 2.01753680992107e-05,
"loss": 0.8831,
"step": 5060
},
{
"epoch": 12.219541616405307,
"grad_norm": 0.3203125,
"learning_rate": 2.0006397570251356e-05,
"loss": 0.872,
"step": 5065
},
{
"epoch": 12.231604342581424,
"grad_norm": 0.318359375,
"learning_rate": 1.983805890949927e-05,
"loss": 0.8756,
"step": 5070
},
{
"epoch": 12.243667068757539,
"grad_norm": 0.32421875,
"learning_rate": 1.9670353446658873e-05,
"loss": 0.8795,
"step": 5075
},
{
"epoch": 12.255729794933655,
"grad_norm": 0.30078125,
"learning_rate": 1.9503282506432974e-05,
"loss": 0.8744,
"step": 5080
},
{
"epoch": 12.26779252110977,
"grad_norm": 0.306640625,
"learning_rate": 1.9336847408512328e-05,
"loss": 0.8728,
"step": 5085
},
{
"epoch": 12.279855247285887,
"grad_norm": 0.318359375,
"learning_rate": 1.917104946756515e-05,
"loss": 0.867,
"step": 5090
},
{
"epoch": 12.291917973462002,
"grad_norm": 0.318359375,
"learning_rate": 1.9005889993226735e-05,
"loss": 0.8756,
"step": 5095
},
{
"epoch": 12.303980699638117,
"grad_norm": 0.314453125,
"learning_rate": 1.884137029008921e-05,
"loss": 0.8761,
"step": 5100
},
{
"epoch": 12.316043425814234,
"grad_norm": 0.31640625,
"learning_rate": 1.8677491657691116e-05,
"loss": 0.8738,
"step": 5105
},
{
"epoch": 12.32810615199035,
"grad_norm": 0.322265625,
"learning_rate": 1.8514255390507183e-05,
"loss": 0.8793,
"step": 5110
},
{
"epoch": 12.340168878166466,
"grad_norm": 0.30859375,
"learning_rate": 1.8351662777938127e-05,
"loss": 0.8741,
"step": 5115
},
{
"epoch": 12.352231604342581,
"grad_norm": 0.322265625,
"learning_rate": 1.818971510430042e-05,
"loss": 0.876,
"step": 5120
},
{
"epoch": 12.364294330518698,
"grad_norm": 0.330078125,
"learning_rate": 1.802841364881621e-05,
"loss": 0.87,
"step": 5125
},
{
"epoch": 12.376357056694813,
"grad_norm": 0.318359375,
"learning_rate": 1.7867759685603114e-05,
"loss": 0.8818,
"step": 5130
},
{
"epoch": 12.38841978287093,
"grad_norm": 0.318359375,
"learning_rate": 1.7707754483664308e-05,
"loss": 0.8782,
"step": 5135
},
{
"epoch": 12.400482509047045,
"grad_norm": 0.3125,
"learning_rate": 1.754839930687836e-05,
"loss": 0.8743,
"step": 5140
},
{
"epoch": 12.41254523522316,
"grad_norm": 0.30859375,
"learning_rate": 1.738969541398926e-05,
"loss": 0.8783,
"step": 5145
},
{
"epoch": 12.424607961399277,
"grad_norm": 0.3203125,
"learning_rate": 1.723164405859663e-05,
"loss": 0.872,
"step": 5150
},
{
"epoch": 12.436670687575392,
"grad_norm": 0.341796875,
"learning_rate": 1.7074246489145595e-05,
"loss": 0.8739,
"step": 5155
},
{
"epoch": 12.448733413751508,
"grad_norm": 0.3125,
"learning_rate": 1.691750394891707e-05,
"loss": 0.8779,
"step": 5160
},
{
"epoch": 12.460796139927623,
"grad_norm": 0.310546875,
"learning_rate": 1.6761417676017956e-05,
"loss": 0.8758,
"step": 5165
},
{
"epoch": 12.47285886610374,
"grad_norm": 0.30859375,
"learning_rate": 1.660598890337124e-05,
"loss": 0.881,
"step": 5170
},
{
"epoch": 12.484921592279855,
"grad_norm": 0.318359375,
"learning_rate": 1.6451218858706374e-05,
"loss": 0.8806,
"step": 5175
},
{
"epoch": 12.49698431845597,
"grad_norm": 0.3125,
"learning_rate": 1.629710876454945e-05,
"loss": 0.8771,
"step": 5180
},
{
"epoch": 12.509047044632087,
"grad_norm": 0.322265625,
"learning_rate": 1.6143659838213752e-05,
"loss": 0.8784,
"step": 5185
},
{
"epoch": 12.521109770808202,
"grad_norm": 0.3046875,
"learning_rate": 1.59908732917899e-05,
"loss": 0.863,
"step": 5190
},
{
"epoch": 12.533172496984319,
"grad_norm": 0.310546875,
"learning_rate": 1.583875033213642e-05,
"loss": 0.87,
"step": 5195
},
{
"epoch": 12.545235223160434,
"grad_norm": 0.3203125,
"learning_rate": 1.568729216087017e-05,
"loss": 0.8741,
"step": 5200
},
{
"epoch": 12.55729794933655,
"grad_norm": 0.33203125,
"learning_rate": 1.5536499974356866e-05,
"loss": 0.8759,
"step": 5205
},
{
"epoch": 12.569360675512666,
"grad_norm": 0.310546875,
"learning_rate": 1.5386374963701633e-05,
"loss": 0.8841,
"step": 5210
},
{
"epoch": 12.581423401688781,
"grad_norm": 0.3203125,
"learning_rate": 1.5236918314739568e-05,
"loss": 0.8843,
"step": 5215
},
{
"epoch": 12.593486127864898,
"grad_norm": 0.306640625,
"learning_rate": 1.5088131208026367e-05,
"loss": 0.8694,
"step": 5220
},
{
"epoch": 12.605548854041013,
"grad_norm": 0.3046875,
"learning_rate": 1.4940014818829074e-05,
"loss": 0.8738,
"step": 5225
},
{
"epoch": 12.61761158021713,
"grad_norm": 0.322265625,
"learning_rate": 1.4792570317116671e-05,
"loss": 0.8694,
"step": 5230
},
{
"epoch": 12.629674306393245,
"grad_norm": 0.31640625,
"learning_rate": 1.4645798867551008e-05,
"loss": 0.8774,
"step": 5235
},
{
"epoch": 12.641737032569361,
"grad_norm": 0.306640625,
"learning_rate": 1.4499701629477446e-05,
"loss": 0.8805,
"step": 5240
},
{
"epoch": 12.653799758745476,
"grad_norm": 0.31640625,
"learning_rate": 1.4354279756915735e-05,
"loss": 0.8815,
"step": 5245
},
{
"epoch": 12.665862484921592,
"grad_norm": 0.310546875,
"learning_rate": 1.4209534398551016e-05,
"loss": 0.875,
"step": 5250
},
{
"epoch": 12.677925211097708,
"grad_norm": 0.31640625,
"learning_rate": 1.406546669772456e-05,
"loss": 0.8757,
"step": 5255
},
{
"epoch": 12.689987937273823,
"grad_norm": 0.333984375,
"learning_rate": 1.3922077792424881e-05,
"loss": 0.8773,
"step": 5260
},
{
"epoch": 12.70205066344994,
"grad_norm": 0.32421875,
"learning_rate": 1.3779368815278647e-05,
"loss": 0.873,
"step": 5265
},
{
"epoch": 12.714113389626055,
"grad_norm": 0.31640625,
"learning_rate": 1.3637340893541895e-05,
"loss": 0.8811,
"step": 5270
},
{
"epoch": 12.726176115802172,
"grad_norm": 0.314453125,
"learning_rate": 1.3495995149090911e-05,
"loss": 0.8726,
"step": 5275
},
{
"epoch": 12.738238841978287,
"grad_norm": 0.31640625,
"learning_rate": 1.335533269841347e-05,
"loss": 0.8703,
"step": 5280
},
{
"epoch": 12.750301568154402,
"grad_norm": 0.314453125,
"learning_rate": 1.321535465260012e-05,
"loss": 0.8764,
"step": 5285
},
{
"epoch": 12.762364294330519,
"grad_norm": 0.33984375,
"learning_rate": 1.307606211733522e-05,
"loss": 0.8762,
"step": 5290
},
{
"epoch": 12.774427020506634,
"grad_norm": 0.318359375,
"learning_rate": 1.2937456192888309e-05,
"loss": 0.8707,
"step": 5295
},
{
"epoch": 12.78648974668275,
"grad_norm": 0.337890625,
"learning_rate": 1.2799537974105402e-05,
"loss": 0.8822,
"step": 5300
},
{
"epoch": 12.798552472858866,
"grad_norm": 0.306640625,
"learning_rate": 1.2662308550400336e-05,
"loss": 0.8761,
"step": 5305
},
{
"epoch": 12.810615199034983,
"grad_norm": 0.365234375,
"learning_rate": 1.252576900574618e-05,
"loss": 0.8743,
"step": 5310
},
{
"epoch": 12.822677925211098,
"grad_norm": 0.3046875,
"learning_rate": 1.2389920418666633e-05,
"loss": 0.8689,
"step": 5315
},
{
"epoch": 12.834740651387214,
"grad_norm": 0.322265625,
"learning_rate": 1.225476386222757e-05,
"loss": 0.8698,
"step": 5320
},
{
"epoch": 12.84680337756333,
"grad_norm": 0.328125,
"learning_rate": 1.2120300404028507e-05,
"loss": 0.8802,
"step": 5325
},
{
"epoch": 12.858866103739444,
"grad_norm": 0.3046875,
"learning_rate": 1.1986531106194143e-05,
"loss": 0.8739,
"step": 5330
},
{
"epoch": 12.870928829915561,
"grad_norm": 0.322265625,
"learning_rate": 1.1853457025366111e-05,
"loss": 0.8814,
"step": 5335
},
{
"epoch": 12.882991556091676,
"grad_norm": 0.333984375,
"learning_rate": 1.1721079212694452e-05,
"loss": 0.8802,
"step": 5340
},
{
"epoch": 12.895054282267793,
"grad_norm": 0.33984375,
"learning_rate": 1.1589398713829424e-05,
"loss": 0.8746,
"step": 5345
},
{
"epoch": 12.907117008443908,
"grad_norm": 0.3046875,
"learning_rate": 1.1458416568913232e-05,
"loss": 0.8724,
"step": 5350
},
{
"epoch": 12.919179734620023,
"grad_norm": 0.3203125,
"learning_rate": 1.1328133812571784e-05,
"loss": 0.8836,
"step": 5355
},
{
"epoch": 12.93124246079614,
"grad_norm": 0.330078125,
"learning_rate": 1.1198551473906493e-05,
"loss": 0.8663,
"step": 5360
},
{
"epoch": 12.943305186972255,
"grad_norm": 0.3203125,
"learning_rate": 1.1069670576486224e-05,
"loss": 0.8722,
"step": 5365
},
{
"epoch": 12.955367913148372,
"grad_norm": 0.32421875,
"learning_rate": 1.0941492138339183e-05,
"loss": 0.8718,
"step": 5370
},
{
"epoch": 12.967430639324487,
"grad_norm": 0.3125,
"learning_rate": 1.0814017171944812e-05,
"loss": 0.875,
"step": 5375
},
{
"epoch": 12.979493365500604,
"grad_norm": 0.33984375,
"learning_rate": 1.068724668422586e-05,
"loss": 0.8768,
"step": 5380
},
{
"epoch": 12.991556091676719,
"grad_norm": 0.314453125,
"learning_rate": 1.0561181676540444e-05,
"loss": 0.8779,
"step": 5385
},
{
"epoch": 12.998793727382388,
"eval_loss": 2.5023179054260254,
"eval_runtime": 0.6415,
"eval_samples_per_second": 15.589,
"eval_steps_per_second": 1.559,
"step": 5388
},
{
"epoch": 13.003618817852836,
"grad_norm": 0.310546875,
"learning_rate": 1.0435823144674051e-05,
"loss": 0.8808,
"step": 5390
},
{
"epoch": 13.01568154402895,
"grad_norm": 0.30859375,
"learning_rate": 1.0311172078831766e-05,
"loss": 0.882,
"step": 5395
},
{
"epoch": 13.027744270205066,
"grad_norm": 0.3125,
"learning_rate": 1.01872294636304e-05,
"loss": 0.8872,
"step": 5400
},
{
"epoch": 13.039806996381182,
"grad_norm": 0.314453125,
"learning_rate": 1.0063996278090704e-05,
"loss": 0.8692,
"step": 5405
},
{
"epoch": 13.051869722557297,
"grad_norm": 0.333984375,
"learning_rate": 9.941473495629683e-06,
"loss": 0.8738,
"step": 5410
},
{
"epoch": 13.063932448733414,
"grad_norm": 0.30859375,
"learning_rate": 9.81966208405285e-06,
"loss": 0.8858,
"step": 5415
},
{
"epoch": 13.07599517490953,
"grad_norm": 0.310546875,
"learning_rate": 9.698563005546668e-06,
"loss": 0.878,
"step": 5420
},
{
"epoch": 13.088057901085646,
"grad_norm": 0.306640625,
"learning_rate": 9.578177216670824e-06,
"loss": 0.8625,
"step": 5425
},
{
"epoch": 13.100120627261761,
"grad_norm": 0.32421875,
"learning_rate": 9.458505668350759e-06,
"loss": 0.8794,
"step": 5430
},
{
"epoch": 13.112183353437876,
"grad_norm": 0.3125,
"learning_rate": 9.339549305870187e-06,
"loss": 0.8684,
"step": 5435
},
{
"epoch": 13.124246079613993,
"grad_norm": 0.314453125,
"learning_rate": 9.221309068863527e-06,
"loss": 0.874,
"step": 5440
},
{
"epoch": 13.136308805790108,
"grad_norm": 0.3203125,
"learning_rate": 9.103785891308547e-06,
"loss": 0.872,
"step": 5445
},
{
"epoch": 13.148371531966225,
"grad_norm": 0.326171875,
"learning_rate": 8.986980701518966e-06,
"loss": 0.8779,
"step": 5450
},
{
"epoch": 13.16043425814234,
"grad_norm": 0.31640625,
"learning_rate": 8.870894422137177e-06,
"loss": 0.8714,
"step": 5455
},
{
"epoch": 13.172496984318457,
"grad_norm": 0.3125,
"learning_rate": 8.755527970126853e-06,
"loss": 0.8753,
"step": 5460
},
{
"epoch": 13.184559710494572,
"grad_norm": 0.326171875,
"learning_rate": 8.640882256765759e-06,
"loss": 0.8712,
"step": 5465
},
{
"epoch": 13.196622436670687,
"grad_norm": 0.31640625,
"learning_rate": 8.526958187638635e-06,
"loss": 0.8703,
"step": 5470
},
{
"epoch": 13.208685162846804,
"grad_norm": 0.33203125,
"learning_rate": 8.413756662629879e-06,
"loss": 0.874,
"step": 5475
},
{
"epoch": 13.220747889022919,
"grad_norm": 0.33203125,
"learning_rate": 8.301278575916538e-06,
"loss": 0.8721,
"step": 5480
},
{
"epoch": 13.232810615199035,
"grad_norm": 0.310546875,
"learning_rate": 8.189524815961291e-06,
"loss": 0.8761,
"step": 5485
},
{
"epoch": 13.24487334137515,
"grad_norm": 0.32421875,
"learning_rate": 8.07849626550531e-06,
"loss": 0.8616,
"step": 5490
},
{
"epoch": 13.256936067551267,
"grad_norm": 0.3125,
"learning_rate": 7.968193801561363e-06,
"loss": 0.881,
"step": 5495
},
{
"epoch": 13.268998793727382,
"grad_norm": 0.31640625,
"learning_rate": 7.858618295406895e-06,
"loss": 0.8724,
"step": 5500
},
{
"epoch": 13.281061519903497,
"grad_norm": 0.314453125,
"learning_rate": 7.74977061257709e-06,
"loss": 0.8826,
"step": 5505
},
{
"epoch": 13.293124246079614,
"grad_norm": 0.310546875,
"learning_rate": 7.641651612858081e-06,
"loss": 0.8864,
"step": 5510
},
{
"epoch": 13.30518697225573,
"grad_norm": 0.31640625,
"learning_rate": 7.534262150280136e-06,
"loss": 0.8781,
"step": 5515
},
{
"epoch": 13.317249698431846,
"grad_norm": 0.34375,
"learning_rate": 7.427603073110967e-06,
"loss": 0.8721,
"step": 5520
},
{
"epoch": 13.329312424607961,
"grad_norm": 0.310546875,
"learning_rate": 7.321675223848923e-06,
"loss": 0.8746,
"step": 5525
},
{
"epoch": 13.341375150784078,
"grad_norm": 0.3125,
"learning_rate": 7.216479439216406e-06,
"loss": 0.8657,
"step": 5530
},
{
"epoch": 13.353437876960193,
"grad_norm": 0.326171875,
"learning_rate": 7.1120165501533e-06,
"loss": 0.8668,
"step": 5535
},
{
"epoch": 13.365500603136308,
"grad_norm": 0.318359375,
"learning_rate": 7.0082873818102964e-06,
"loss": 0.8716,
"step": 5540
},
{
"epoch": 13.377563329312425,
"grad_norm": 0.326171875,
"learning_rate": 6.90529275354247e-06,
"loss": 0.8693,
"step": 5545
},
{
"epoch": 13.38962605548854,
"grad_norm": 0.3125,
"learning_rate": 6.803033478902765e-06,
"loss": 0.8786,
"step": 5550
},
{
"epoch": 13.401688781664657,
"grad_norm": 0.328125,
"learning_rate": 6.7015103656355885e-06,
"loss": 0.8846,
"step": 5555
},
{
"epoch": 13.413751507840772,
"grad_norm": 0.314453125,
"learning_rate": 6.600724215670429e-06,
"loss": 0.8735,
"step": 5560
},
{
"epoch": 13.425814234016888,
"grad_norm": 0.33203125,
"learning_rate": 6.500675825115454e-06,
"loss": 0.8701,
"step": 5565
},
{
"epoch": 13.437876960193003,
"grad_norm": 0.30859375,
"learning_rate": 6.401365984251384e-06,
"loss": 0.8764,
"step": 5570
},
{
"epoch": 13.44993968636912,
"grad_norm": 0.322265625,
"learning_rate": 6.302795477525058e-06,
"loss": 0.8824,
"step": 5575
},
{
"epoch": 13.462002412545235,
"grad_norm": 0.322265625,
"learning_rate": 6.204965083543368e-06,
"loss": 0.8773,
"step": 5580
},
{
"epoch": 13.47406513872135,
"grad_norm": 0.310546875,
"learning_rate": 6.107875575067057e-06,
"loss": 0.868,
"step": 5585
},
{
"epoch": 13.486127864897467,
"grad_norm": 0.31640625,
"learning_rate": 6.011527719004617e-06,
"loss": 0.8706,
"step": 5590
},
{
"epoch": 13.498190591073582,
"grad_norm": 0.330078125,
"learning_rate": 5.915922276406249e-06,
"loss": 0.878,
"step": 5595
},
{
"epoch": 13.510253317249699,
"grad_norm": 0.314453125,
"learning_rate": 5.8210600024578234e-06,
"loss": 0.8739,
"step": 5600
},
{
"epoch": 13.522316043425814,
"grad_norm": 0.3125,
"learning_rate": 5.726941646474992e-06,
"loss": 0.8768,
"step": 5605
},
{
"epoch": 13.534378769601929,
"grad_norm": 0.30859375,
"learning_rate": 5.633567951897145e-06,
"loss": 0.8785,
"step": 5610
},
{
"epoch": 13.546441495778046,
"grad_norm": 0.326171875,
"learning_rate": 5.5409396562816076e-06,
"loss": 0.8751,
"step": 5615
},
{
"epoch": 13.558504221954161,
"grad_norm": 0.30859375,
"learning_rate": 5.449057491297871e-06,
"loss": 0.8861,
"step": 5620
},
{
"epoch": 13.570566948130278,
"grad_norm": 0.314453125,
"learning_rate": 5.357922182721687e-06,
"loss": 0.8782,
"step": 5625
},
{
"epoch": 13.582629674306393,
"grad_norm": 0.30859375,
"learning_rate": 5.267534450429423e-06,
"loss": 0.8705,
"step": 5630
},
{
"epoch": 13.59469240048251,
"grad_norm": 0.3125,
"learning_rate": 5.177895008392353e-06,
"loss": 0.8755,
"step": 5635
},
{
"epoch": 13.606755126658625,
"grad_norm": 0.302734375,
"learning_rate": 5.08900456467103e-06,
"loss": 0.8744,
"step": 5640
},
{
"epoch": 13.618817852834741,
"grad_norm": 0.3125,
"learning_rate": 5.000863821409674e-06,
"loss": 0.8811,
"step": 5645
},
{
"epoch": 13.630880579010856,
"grad_norm": 0.30859375,
"learning_rate": 4.9134734748305986e-06,
"loss": 0.8718,
"step": 5650
},
{
"epoch": 13.642943305186972,
"grad_norm": 0.310546875,
"learning_rate": 4.826834215228826e-06,
"loss": 0.8773,
"step": 5655
},
{
"epoch": 13.655006031363088,
"grad_norm": 0.314453125,
"learning_rate": 4.740946726966466e-06,
"loss": 0.8811,
"step": 5660
},
{
"epoch": 13.667068757539203,
"grad_norm": 0.310546875,
"learning_rate": 4.655811688467448e-06,
"loss": 0.8729,
"step": 5665
},
{
"epoch": 13.67913148371532,
"grad_norm": 0.337890625,
"learning_rate": 4.5714297722121106e-06,
"loss": 0.8811,
"step": 5670
},
{
"epoch": 13.691194209891435,
"grad_norm": 0.3046875,
"learning_rate": 4.4878016447318704e-06,
"loss": 0.8661,
"step": 5675
},
{
"epoch": 13.703256936067552,
"grad_norm": 0.345703125,
"learning_rate": 4.404927966604011e-06,
"loss": 0.8745,
"step": 5680
},
{
"epoch": 13.715319662243667,
"grad_norm": 0.310546875,
"learning_rate": 4.322809392446392e-06,
"loss": 0.8771,
"step": 5685
},
{
"epoch": 13.727382388419782,
"grad_norm": 0.31640625,
"learning_rate": 4.241446570912344e-06,
"loss": 0.874,
"step": 5690
},
{
"epoch": 13.739445114595899,
"grad_norm": 0.306640625,
"learning_rate": 4.160840144685507e-06,
"loss": 0.8768,
"step": 5695
},
{
"epoch": 13.751507840772014,
"grad_norm": 0.31640625,
"learning_rate": 4.080990750474778e-06,
"loss": 0.8706,
"step": 5700
},
{
"epoch": 13.76357056694813,
"grad_norm": 0.306640625,
"learning_rate": 4.001899019009281e-06,
"loss": 0.8727,
"step": 5705
},
{
"epoch": 13.775633293124246,
"grad_norm": 0.322265625,
"learning_rate": 3.92356557503335e-06,
"loss": 0.8792,
"step": 5710
},
{
"epoch": 13.787696019300363,
"grad_norm": 0.3203125,
"learning_rate": 3.84599103730161e-06,
"loss": 0.864,
"step": 5715
},
{
"epoch": 13.799758745476478,
"grad_norm": 0.31640625,
"learning_rate": 3.7691760185741387e-06,
"loss": 0.8733,
"step": 5720
},
{
"epoch": 13.811821471652593,
"grad_norm": 0.337890625,
"learning_rate": 3.6931211256115537e-06,
"loss": 0.8753,
"step": 5725
},
{
"epoch": 13.82388419782871,
"grad_norm": 0.32421875,
"learning_rate": 3.617826959170256e-06,
"loss": 0.8714,
"step": 5730
},
{
"epoch": 13.835946924004825,
"grad_norm": 0.318359375,
"learning_rate": 3.543294113997664e-06,
"loss": 0.8741,
"step": 5735
},
{
"epoch": 13.848009650180941,
"grad_norm": 0.3046875,
"learning_rate": 3.469523178827583e-06,
"loss": 0.881,
"step": 5740
},
{
"epoch": 13.860072376357056,
"grad_norm": 0.306640625,
"learning_rate": 3.3965147363754555e-06,
"loss": 0.8778,
"step": 5745
},
{
"epoch": 13.872135102533173,
"grad_norm": 0.310546875,
"learning_rate": 3.3242693633337983e-06,
"loss": 0.8804,
"step": 5750
},
{
"epoch": 13.884197828709288,
"grad_norm": 0.314453125,
"learning_rate": 3.252787630367715e-06,
"loss": 0.882,
"step": 5755
},
{
"epoch": 13.896260554885405,
"grad_norm": 0.3046875,
"learning_rate": 3.182070102110257e-06,
"loss": 0.8672,
"step": 5760
},
{
"epoch": 13.90832328106152,
"grad_norm": 0.314453125,
"learning_rate": 3.1121173371580825e-06,
"loss": 0.8772,
"step": 5765
},
{
"epoch": 13.920386007237635,
"grad_norm": 0.314453125,
"learning_rate": 3.042929888066992e-06,
"loss": 0.8807,
"step": 5770
},
{
"epoch": 13.932448733413752,
"grad_norm": 0.322265625,
"learning_rate": 2.974508301347534e-06,
"loss": 0.8796,
"step": 5775
},
{
"epoch": 13.944511459589867,
"grad_norm": 0.306640625,
"learning_rate": 2.9068531174607615e-06,
"loss": 0.8712,
"step": 5780
},
{
"epoch": 13.956574185765984,
"grad_norm": 0.318359375,
"learning_rate": 2.839964870813916e-06,
"loss": 0.8744,
"step": 5785
},
{
"epoch": 13.968636911942099,
"grad_norm": 0.314453125,
"learning_rate": 2.7738440897561723e-06,
"loss": 0.8754,
"step": 5790
},
{
"epoch": 13.980699638118214,
"grad_norm": 0.3125,
"learning_rate": 2.7084912965745778e-06,
"loss": 0.8753,
"step": 5795
},
{
"epoch": 13.99276236429433,
"grad_norm": 0.314453125,
"learning_rate": 2.6439070074897874e-06,
"loss": 0.8743,
"step": 5800
},
{
"epoch": 14.0,
"eval_loss": 2.5032882690429688,
"eval_runtime": 0.5357,
"eval_samples_per_second": 18.667,
"eval_steps_per_second": 1.867,
"step": 5803
},
{
"epoch": 14.004825090470446,
"grad_norm": 0.31640625,
"learning_rate": 2.580091732652101e-06,
"loss": 0.8799,
"step": 5805
},
{
"epoch": 14.016887816646562,
"grad_norm": 0.31640625,
"learning_rate": 2.5170459761373664e-06,
"loss": 0.8781,
"step": 5810
},
{
"epoch": 14.028950542822678,
"grad_norm": 0.31640625,
"learning_rate": 2.454770235943027e-06,
"loss": 0.8751,
"step": 5815
},
{
"epoch": 14.041013268998794,
"grad_norm": 0.30859375,
"learning_rate": 2.3932650039841687e-06,
"loss": 0.887,
"step": 5820
},
{
"epoch": 14.05307599517491,
"grad_norm": 0.31640625,
"learning_rate": 2.3325307660896577e-06,
"loss": 0.8755,
"step": 5825
},
{
"epoch": 14.065138721351026,
"grad_norm": 0.318359375,
"learning_rate": 2.2725680019982765e-06,
"loss": 0.8792,
"step": 5830
},
{
"epoch": 14.077201447527141,
"grad_norm": 0.322265625,
"learning_rate": 2.213377185354959e-06,
"loss": 0.8719,
"step": 5835
},
{
"epoch": 14.089264173703256,
"grad_norm": 0.306640625,
"learning_rate": 2.1549587837070394e-06,
"loss": 0.8779,
"step": 5840
},
{
"epoch": 14.101326899879373,
"grad_norm": 0.314453125,
"learning_rate": 2.0973132585005552e-06,
"loss": 0.8774,
"step": 5845
},
{
"epoch": 14.113389626055488,
"grad_norm": 0.318359375,
"learning_rate": 2.0404410650765817e-06,
"loss": 0.8718,
"step": 5850
},
{
"epoch": 14.125452352231605,
"grad_norm": 0.330078125,
"learning_rate": 1.9843426526677033e-06,
"loss": 0.8853,
"step": 5855
},
{
"epoch": 14.13751507840772,
"grad_norm": 0.326171875,
"learning_rate": 1.9290184643943944e-06,
"loss": 0.8779,
"step": 5860
},
{
"epoch": 14.149577804583837,
"grad_norm": 0.306640625,
"learning_rate": 1.874468937261531e-06,
"loss": 0.8692,
"step": 5865
},
{
"epoch": 14.161640530759952,
"grad_norm": 0.310546875,
"learning_rate": 1.820694502154996e-06,
"loss": 0.8756,
"step": 5870
},
{
"epoch": 14.173703256936067,
"grad_norm": 0.296875,
"learning_rate": 1.7676955838381804e-06,
"loss": 0.8697,
"step": 5875
},
{
"epoch": 14.185765983112184,
"grad_norm": 0.318359375,
"learning_rate": 1.71547260094872e-06,
"loss": 0.8832,
"step": 5880
},
{
"epoch": 14.197828709288299,
"grad_norm": 0.302734375,
"learning_rate": 1.6640259659951419e-06,
"loss": 0.8761,
"step": 5885
},
{
"epoch": 14.209891435464415,
"grad_norm": 0.333984375,
"learning_rate": 1.61335608535359e-06,
"loss": 0.8785,
"step": 5890
},
{
"epoch": 14.22195416164053,
"grad_norm": 0.32421875,
"learning_rate": 1.5634633592646609e-06,
"loss": 0.8726,
"step": 5895
},
{
"epoch": 14.234016887816647,
"grad_norm": 0.3046875,
"learning_rate": 1.5143481818302164e-06,
"loss": 0.8694,
"step": 5900
},
{
"epoch": 14.246079613992762,
"grad_norm": 0.333984375,
"learning_rate": 1.4660109410102652e-06,
"loss": 0.8852,
"step": 5905
},
{
"epoch": 14.258142340168877,
"grad_norm": 0.314453125,
"learning_rate": 1.4184520186199202e-06,
"loss": 0.8746,
"step": 5910
},
{
"epoch": 14.270205066344994,
"grad_norm": 0.3046875,
"learning_rate": 1.3716717903263566e-06,
"loss": 0.8767,
"step": 5915
},
{
"epoch": 14.28226779252111,
"grad_norm": 0.3203125,
"learning_rate": 1.3256706256458585e-06,
"loss": 0.8694,
"step": 5920
},
{
"epoch": 14.294330518697226,
"grad_norm": 0.310546875,
"learning_rate": 1.2804488879408993e-06,
"loss": 0.8709,
"step": 5925
},
{
"epoch": 14.306393244873341,
"grad_norm": 0.322265625,
"learning_rate": 1.2360069344172887e-06,
"loss": 0.8722,
"step": 5930
},
{
"epoch": 14.318455971049458,
"grad_norm": 0.30859375,
"learning_rate": 1.1923451161213074e-06,
"loss": 0.8754,
"step": 5935
},
{
"epoch": 14.330518697225573,
"grad_norm": 0.306640625,
"learning_rate": 1.1494637779369766e-06,
"loss": 0.8737,
"step": 5940
},
{
"epoch": 14.342581423401688,
"grad_norm": 0.306640625,
"learning_rate": 1.107363258583305e-06,
"loss": 0.8713,
"step": 5945
},
{
"epoch": 14.354644149577805,
"grad_norm": 0.302734375,
"learning_rate": 1.0660438906116342e-06,
"loss": 0.8744,
"step": 5950
},
{
"epoch": 14.36670687575392,
"grad_norm": 0.328125,
"learning_rate": 1.0255060004030093e-06,
"loss": 0.8805,
"step": 5955
},
{
"epoch": 14.378769601930037,
"grad_norm": 0.3046875,
"learning_rate": 9.857499081655786e-07,
"loss": 0.8786,
"step": 5960
},
{
"epoch": 14.390832328106152,
"grad_norm": 0.30078125,
"learning_rate": 9.467759279320976e-07,
"loss": 0.882,
"step": 5965
},
{
"epoch": 14.402895054282268,
"grad_norm": 0.3046875,
"learning_rate": 9.085843675574079e-07,
"loss": 0.8756,
"step": 5970
},
{
"epoch": 14.414957780458383,
"grad_norm": 0.3125,
"learning_rate": 8.711755287160727e-07,
"loss": 0.8743,
"step": 5975
},
{
"epoch": 14.427020506634499,
"grad_norm": 0.310546875,
"learning_rate": 8.345497068998897e-07,
"loss": 0.8792,
"step": 5980
},
{
"epoch": 14.439083232810615,
"grad_norm": 0.30859375,
"learning_rate": 7.987071914156596e-07,
"loss": 0.8664,
"step": 5985
},
{
"epoch": 14.45114595898673,
"grad_norm": 0.3046875,
"learning_rate": 7.636482653828658e-07,
"loss": 0.8744,
"step": 5990
},
{
"epoch": 14.463208685162847,
"grad_norm": 0.310546875,
"learning_rate": 7.293732057313874e-07,
"loss": 0.8674,
"step": 5995
},
{
"epoch": 14.475271411338962,
"grad_norm": 0.322265625,
"learning_rate": 6.958822831994005e-07,
"loss": 0.8744,
"step": 6000
},
{
"epoch": 14.487334137515079,
"grad_norm": 0.30078125,
"learning_rate": 6.631757623311586e-07,
"loss": 0.872,
"step": 6005
},
{
"epoch": 14.499396863691194,
"grad_norm": 0.3046875,
"learning_rate": 6.312539014749818e-07,
"loss": 0.8838,
"step": 6010
},
{
"epoch": 14.51145958986731,
"grad_norm": 0.326171875,
"learning_rate": 6.001169527811268e-07,
"loss": 0.8684,
"step": 6015
},
{
"epoch": 14.523522316043426,
"grad_norm": 0.33203125,
"learning_rate": 5.697651621998535e-07,
"loss": 0.8776,
"step": 6020
},
{
"epoch": 14.535585042219541,
"grad_norm": 0.3125,
"learning_rate": 5.401987694794941e-07,
"loss": 0.8694,
"step": 6025
},
{
"epoch": 14.547647768395658,
"grad_norm": 0.3203125,
"learning_rate": 5.114180081645214e-07,
"loss": 0.8698,
"step": 6030
},
{
"epoch": 14.559710494571773,
"grad_norm": 0.310546875,
"learning_rate": 4.834231055937055e-07,
"loss": 0.8816,
"step": 6035
},
{
"epoch": 14.57177322074789,
"grad_norm": 0.314453125,
"learning_rate": 4.5621428289834843e-07,
"loss": 0.8839,
"step": 6040
},
{
"epoch": 14.583835946924005,
"grad_norm": 0.318359375,
"learning_rate": 4.2979175500050817e-07,
"loss": 0.8699,
"step": 6045
},
{
"epoch": 14.595898673100121,
"grad_norm": 0.306640625,
"learning_rate": 4.0415573061133307e-07,
"loss": 0.8657,
"step": 6050
},
{
"epoch": 14.607961399276236,
"grad_norm": 0.345703125,
"learning_rate": 3.7930641222934103e-07,
"loss": 0.8832,
"step": 6055
},
{
"epoch": 14.620024125452352,
"grad_norm": 0.30859375,
"learning_rate": 3.552439961389431e-07,
"loss": 0.8762,
"step": 6060
},
{
"epoch": 14.632086851628468,
"grad_norm": 0.322265625,
"learning_rate": 3.3196867240876675e-07,
"loss": 0.8634,
"step": 6065
},
{
"epoch": 14.644149577804583,
"grad_norm": 0.32421875,
"learning_rate": 3.094806248902349e-07,
"loss": 0.8749,
"step": 6070
},
{
"epoch": 14.6562123039807,
"grad_norm": 0.3046875,
"learning_rate": 2.877800312160783e-07,
"loss": 0.8708,
"step": 6075
},
{
"epoch": 14.668275030156815,
"grad_norm": 0.302734375,
"learning_rate": 2.668670627989478e-07,
"loss": 0.874,
"step": 6080
},
{
"epoch": 14.680337756332932,
"grad_norm": 0.31640625,
"learning_rate": 2.4674188483003734e-07,
"loss": 0.881,
"step": 6085
},
{
"epoch": 14.692400482509047,
"grad_norm": 0.310546875,
"learning_rate": 2.274046562778409e-07,
"loss": 0.864,
"step": 6090
},
{
"epoch": 14.704463208685162,
"grad_norm": 0.302734375,
"learning_rate": 2.088555298867978e-07,
"loss": 0.8669,
"step": 6095
},
{
"epoch": 14.716525934861279,
"grad_norm": 0.310546875,
"learning_rate": 1.910946521761714e-07,
"loss": 0.8725,
"step": 6100
},
{
"epoch": 14.728588661037394,
"grad_norm": 0.30859375,
"learning_rate": 1.7412216343885014e-07,
"loss": 0.8664,
"step": 6105
},
{
"epoch": 14.74065138721351,
"grad_norm": 0.310546875,
"learning_rate": 1.5793819774027053e-07,
"loss": 0.875,
"step": 6110
},
{
"epoch": 14.752714113389626,
"grad_norm": 0.306640625,
"learning_rate": 1.425428829173181e-07,
"loss": 0.8681,
"step": 6115
},
{
"epoch": 14.764776839565743,
"grad_norm": 0.302734375,
"learning_rate": 1.2793634057732818e-07,
"loss": 0.8831,
"step": 6120
},
{
"epoch": 14.776839565741858,
"grad_norm": 0.310546875,
"learning_rate": 1.1411868609717546e-07,
"loss": 0.8769,
"step": 6125
},
{
"epoch": 14.788902291917973,
"grad_norm": 0.314453125,
"learning_rate": 1.010900286222638e-07,
"loss": 0.8717,
"step": 6130
},
{
"epoch": 14.80096501809409,
"grad_norm": 0.3203125,
"learning_rate": 8.885047106578227e-08,
"loss": 0.8743,
"step": 6135
},
{
"epoch": 14.813027744270205,
"grad_norm": 0.318359375,
"learning_rate": 7.740011010778369e-08,
"loss": 0.867,
"step": 6140
},
{
"epoch": 14.825090470446321,
"grad_norm": 0.31640625,
"learning_rate": 6.673903619449639e-08,
"loss": 0.875,
"step": 6145
},
{
"epoch": 14.837153196622436,
"grad_norm": 0.330078125,
"learning_rate": 5.6867333537580226e-08,
"loss": 0.8846,
"step": 6150
},
{
"epoch": 14.849215922798553,
"grad_norm": 0.322265625,
"learning_rate": 4.7785080113449396e-08,
"loss": 0.8832,
"step": 6155
},
{
"epoch": 14.861278648974668,
"grad_norm": 0.32421875,
"learning_rate": 3.9492347662684057e-08,
"loss": 0.8764,
"step": 6160
},
{
"epoch": 14.873341375150783,
"grad_norm": 0.3203125,
"learning_rate": 3.1989201689452967e-08,
"loss": 0.8753,
"step": 6165
},
{
"epoch": 14.8854041013269,
"grad_norm": 0.3046875,
"learning_rate": 2.5275701460991674e-08,
"loss": 0.8682,
"step": 6170
},
{
"epoch": 14.897466827503015,
"grad_norm": 0.306640625,
"learning_rate": 1.9351900007114066e-08,
"loss": 0.8719,
"step": 6175
},
{
"epoch": 14.909529553679132,
"grad_norm": 0.30078125,
"learning_rate": 1.4217844119857048e-08,
"loss": 0.8759,
"step": 6180
},
{
"epoch": 14.921592279855247,
"grad_norm": 0.30859375,
"learning_rate": 9.873574353025384e-09,
"loss": 0.8732,
"step": 6185
},
{
"epoch": 14.933655006031364,
"grad_norm": 0.31640625,
"learning_rate": 6.3191250219474255e-09,
"loss": 0.8793,
"step": 6190
},
{
"epoch": 14.945717732207479,
"grad_norm": 0.298828125,
"learning_rate": 3.554524203175369e-09,
"loss": 0.876,
"step": 6195
},
{
"epoch": 14.957780458383596,
"grad_norm": 0.318359375,
"learning_rate": 1.5797937342298952e-09,
"loss": 0.8788,
"step": 6200
},
{
"epoch": 14.96984318455971,
"grad_norm": 0.30859375,
"learning_rate": 3.949492134780464e-10,
"loss": 0.8771,
"step": 6205
},
{
"epoch": 14.981905910735826,
"grad_norm": 0.31640625,
"learning_rate": 0.0,
"loss": 0.8746,
"step": 6210
},
{
"epoch": 14.981905910735826,
"eval_loss": 2.503845453262329,
"eval_runtime": 0.5267,
"eval_samples_per_second": 18.985,
"eval_steps_per_second": 1.899,
"step": 6210
},
{
"epoch": 14.981905910735826,
"step": 6210,
"total_flos": 4.863451355047526e+18,
"train_loss": 0.9534509487582098,
"train_runtime": 21285.3674,
"train_samples_per_second": 18.683,
"train_steps_per_second": 0.292
}
],
"logging_steps": 5,
"max_steps": 6210,
"num_input_tokens_seen": 0,
"num_train_epochs": 15,
"save_steps": 100,
"total_flos": 4.863451355047526e+18,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}