gemma7b-summarize-gpt4o-80k / trainer_state.json
chansung's picture
Model save
085be5b verified
raw
history blame
58.5 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 15.0,
"eval_steps": 500,
"global_step": 1665,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.009009009009009009,
"grad_norm": 608.0,
"learning_rate": 1.1976047904191619e-06,
"loss": 58.5641,
"step": 1
},
{
"epoch": 0.04504504504504504,
"grad_norm": 536.0,
"learning_rate": 5.9880239520958085e-06,
"loss": 54.74,
"step": 5
},
{
"epoch": 0.09009009009009009,
"grad_norm": 488.0,
"learning_rate": 1.1976047904191617e-05,
"loss": 52.9637,
"step": 10
},
{
"epoch": 0.13513513513513514,
"grad_norm": 392.0,
"learning_rate": 1.7964071856287426e-05,
"loss": 40.7219,
"step": 15
},
{
"epoch": 0.18018018018018017,
"grad_norm": 46.5,
"learning_rate": 2.3952095808383234e-05,
"loss": 28.1416,
"step": 20
},
{
"epoch": 0.22522522522522523,
"grad_norm": 40.25,
"learning_rate": 2.994011976047904e-05,
"loss": 24.2461,
"step": 25
},
{
"epoch": 0.2702702702702703,
"grad_norm": 24.125,
"learning_rate": 3.592814371257485e-05,
"loss": 22.2451,
"step": 30
},
{
"epoch": 0.3153153153153153,
"grad_norm": 13.5625,
"learning_rate": 4.191616766467066e-05,
"loss": 20.2199,
"step": 35
},
{
"epoch": 0.36036036036036034,
"grad_norm": 7.5625,
"learning_rate": 4.790419161676647e-05,
"loss": 19.4029,
"step": 40
},
{
"epoch": 0.40540540540540543,
"grad_norm": 9.1875,
"learning_rate": 5.389221556886228e-05,
"loss": 18.5345,
"step": 45
},
{
"epoch": 0.45045045045045046,
"grad_norm": 15.75,
"learning_rate": 5.988023952095808e-05,
"loss": 17.2917,
"step": 50
},
{
"epoch": 0.4954954954954955,
"grad_norm": 37.25,
"learning_rate": 6.58682634730539e-05,
"loss": 15.8479,
"step": 55
},
{
"epoch": 0.5405405405405406,
"grad_norm": 59.75,
"learning_rate": 7.18562874251497e-05,
"loss": 10.5192,
"step": 60
},
{
"epoch": 0.5855855855855856,
"grad_norm": 13.375,
"learning_rate": 7.784431137724552e-05,
"loss": 3.399,
"step": 65
},
{
"epoch": 0.6306306306306306,
"grad_norm": 2.84375,
"learning_rate": 8.383233532934131e-05,
"loss": 2.1324,
"step": 70
},
{
"epoch": 0.6756756756756757,
"grad_norm": 2.125,
"learning_rate": 8.982035928143712e-05,
"loss": 1.8224,
"step": 75
},
{
"epoch": 0.7207207207207207,
"grad_norm": 3.703125,
"learning_rate": 9.580838323353294e-05,
"loss": 1.6034,
"step": 80
},
{
"epoch": 0.7657657657657657,
"grad_norm": 2.75,
"learning_rate": 0.00010179640718562875,
"loss": 1.4811,
"step": 85
},
{
"epoch": 0.8108108108108109,
"grad_norm": 7.53125,
"learning_rate": 0.00010778443113772456,
"loss": 1.4187,
"step": 90
},
{
"epoch": 0.8558558558558559,
"grad_norm": 2.234375,
"learning_rate": 0.00011377245508982037,
"loss": 1.3517,
"step": 95
},
{
"epoch": 0.9009009009009009,
"grad_norm": 8.375,
"learning_rate": 0.00011976047904191617,
"loss": 1.2567,
"step": 100
},
{
"epoch": 0.9459459459459459,
"grad_norm": 5.90625,
"learning_rate": 0.00012574850299401196,
"loss": 1.2634,
"step": 105
},
{
"epoch": 0.990990990990991,
"grad_norm": 4.1875,
"learning_rate": 0.0001317365269461078,
"loss": 1.2272,
"step": 110
},
{
"epoch": 1.0,
"eval_loss": 2.3900420665740967,
"eval_runtime": 1.0174,
"eval_samples_per_second": 4.914,
"eval_steps_per_second": 1.966,
"step": 111
},
{
"epoch": 1.0360360360360361,
"grad_norm": 3.234375,
"learning_rate": 0.00013772455089820359,
"loss": 1.1598,
"step": 115
},
{
"epoch": 1.0810810810810811,
"grad_norm": 2.3125,
"learning_rate": 0.0001437125748502994,
"loss": 1.1474,
"step": 120
},
{
"epoch": 1.1261261261261262,
"grad_norm": 6.375,
"learning_rate": 0.0001497005988023952,
"loss": 1.1051,
"step": 125
},
{
"epoch": 1.1711711711711712,
"grad_norm": 6.625,
"learning_rate": 0.00015568862275449103,
"loss": 1.0869,
"step": 130
},
{
"epoch": 1.2162162162162162,
"grad_norm": 1.0703125,
"learning_rate": 0.00016167664670658683,
"loss": 1.0864,
"step": 135
},
{
"epoch": 1.2612612612612613,
"grad_norm": 6.1875,
"learning_rate": 0.00016766467065868263,
"loss": 1.0419,
"step": 140
},
{
"epoch": 1.3063063063063063,
"grad_norm": 2.84375,
"learning_rate": 0.00017365269461077845,
"loss": 1.0647,
"step": 145
},
{
"epoch": 1.3513513513513513,
"grad_norm": 24.0,
"learning_rate": 0.00017964071856287425,
"loss": 1.0682,
"step": 150
},
{
"epoch": 1.3963963963963963,
"grad_norm": 5.3125,
"learning_rate": 0.00018562874251497007,
"loss": 1.0375,
"step": 155
},
{
"epoch": 1.4414414414414414,
"grad_norm": 2.5625,
"learning_rate": 0.00019161676646706587,
"loss": 1.0111,
"step": 160
},
{
"epoch": 1.4864864864864864,
"grad_norm": 4.25,
"learning_rate": 0.0001976047904191617,
"loss": 0.9605,
"step": 165
},
{
"epoch": 1.5315315315315314,
"grad_norm": 6.15625,
"learning_rate": 0.00019999802081131313,
"loss": 0.9843,
"step": 170
},
{
"epoch": 1.5765765765765765,
"grad_norm": 2.484375,
"learning_rate": 0.00019998592605305064,
"loss": 1.0225,
"step": 175
},
{
"epoch": 1.6216216216216215,
"grad_norm": 2.3125,
"learning_rate": 0.00019996283741405458,
"loss": 0.9908,
"step": 180
},
{
"epoch": 1.6666666666666665,
"grad_norm": 28.375,
"learning_rate": 0.00019992875743301967,
"loss": 1.0038,
"step": 185
},
{
"epoch": 1.7117117117117115,
"grad_norm": 2.390625,
"learning_rate": 0.00019988368985718576,
"loss": 0.9688,
"step": 190
},
{
"epoch": 1.7567567567567568,
"grad_norm": 1.75,
"learning_rate": 0.00019982763964192585,
"loss": 0.9638,
"step": 195
},
{
"epoch": 1.8018018018018018,
"grad_norm": 3.125,
"learning_rate": 0.00019976061295020106,
"loss": 0.9878,
"step": 200
},
{
"epoch": 1.8468468468468469,
"grad_norm": 1.921875,
"learning_rate": 0.00019968261715188324,
"loss": 0.9394,
"step": 205
},
{
"epoch": 1.8918918918918919,
"grad_norm": 1.546875,
"learning_rate": 0.00019959366082294438,
"loss": 0.9285,
"step": 210
},
{
"epoch": 1.936936936936937,
"grad_norm": 3.53125,
"learning_rate": 0.00019949375374451377,
"loss": 0.929,
"step": 215
},
{
"epoch": 1.981981981981982,
"grad_norm": 1.625,
"learning_rate": 0.00019938290690180247,
"loss": 0.9374,
"step": 220
},
{
"epoch": 2.0,
"eval_loss": 2.192767381668091,
"eval_runtime": 1.0169,
"eval_samples_per_second": 4.917,
"eval_steps_per_second": 1.967,
"step": 222
},
{
"epoch": 2.027027027027027,
"grad_norm": 2.25,
"learning_rate": 0.0001992611324828956,
"loss": 0.8849,
"step": 225
},
{
"epoch": 2.0720720720720722,
"grad_norm": 1.8125,
"learning_rate": 0.00019912844387741195,
"loss": 0.864,
"step": 230
},
{
"epoch": 2.1171171171171173,
"grad_norm": 1.09375,
"learning_rate": 0.00019898485567503187,
"loss": 0.8497,
"step": 235
},
{
"epoch": 2.1621621621621623,
"grad_norm": 1.9609375,
"learning_rate": 0.00019883038366389313,
"loss": 0.8844,
"step": 240
},
{
"epoch": 2.2072072072072073,
"grad_norm": 1.0234375,
"learning_rate": 0.0001986650448288548,
"loss": 0.8509,
"step": 245
},
{
"epoch": 2.2522522522522523,
"grad_norm": 3.078125,
"learning_rate": 0.00019848885734962982,
"loss": 0.8718,
"step": 250
},
{
"epoch": 2.2972972972972974,
"grad_norm": 3.390625,
"learning_rate": 0.00019830184059878606,
"loss": 0.881,
"step": 255
},
{
"epoch": 2.3423423423423424,
"grad_norm": 2.75,
"learning_rate": 0.0001981040151396161,
"loss": 0.854,
"step": 260
},
{
"epoch": 2.3873873873873874,
"grad_norm": 1.25,
"learning_rate": 0.0001978954027238763,
"loss": 0.8856,
"step": 265
},
{
"epoch": 2.4324324324324325,
"grad_norm": 1.640625,
"learning_rate": 0.00019767602628939507,
"loss": 0.8797,
"step": 270
},
{
"epoch": 2.4774774774774775,
"grad_norm": 2.28125,
"learning_rate": 0.00019744590995755088,
"loss": 0.8829,
"step": 275
},
{
"epoch": 2.5225225225225225,
"grad_norm": 1.265625,
"learning_rate": 0.00019720507903061974,
"loss": 0.8319,
"step": 280
},
{
"epoch": 2.5675675675675675,
"grad_norm": 1.6328125,
"learning_rate": 0.00019695355998899345,
"loss": 0.8489,
"step": 285
},
{
"epoch": 2.6126126126126126,
"grad_norm": 1.1328125,
"learning_rate": 0.00019669138048826766,
"loss": 0.8409,
"step": 290
},
{
"epoch": 2.6576576576576576,
"grad_norm": 1.1640625,
"learning_rate": 0.00019641856935620117,
"loss": 0.8454,
"step": 295
},
{
"epoch": 2.7027027027027026,
"grad_norm": 0.625,
"learning_rate": 0.00019613515658954624,
"loss": 0.8501,
"step": 300
},
{
"epoch": 2.7477477477477477,
"grad_norm": 1.0859375,
"learning_rate": 0.0001958411733507502,
"loss": 0.8223,
"step": 305
},
{
"epoch": 2.7927927927927927,
"grad_norm": 0.875,
"learning_rate": 0.00019553665196452903,
"loss": 0.8396,
"step": 310
},
{
"epoch": 2.8378378378378377,
"grad_norm": 1.328125,
"learning_rate": 0.0001952216259143132,
"loss": 0.836,
"step": 315
},
{
"epoch": 2.8828828828828827,
"grad_norm": 2.09375,
"learning_rate": 0.0001948961298385659,
"loss": 0.8453,
"step": 320
},
{
"epoch": 2.9279279279279278,
"grad_norm": 0.62109375,
"learning_rate": 0.00019456019952697462,
"loss": 0.824,
"step": 325
},
{
"epoch": 2.972972972972973,
"grad_norm": 1.7578125,
"learning_rate": 0.0001942138719165156,
"loss": 0.8471,
"step": 330
},
{
"epoch": 3.0,
"eval_loss": 2.168168544769287,
"eval_runtime": 1.0188,
"eval_samples_per_second": 4.908,
"eval_steps_per_second": 1.963,
"step": 333
},
{
"epoch": 3.018018018018018,
"grad_norm": 1.5078125,
"learning_rate": 0.00019385718508739262,
"loss": 0.8242,
"step": 335
},
{
"epoch": 3.063063063063063,
"grad_norm": 1.4375,
"learning_rate": 0.00019349017825885,
"loss": 0.7439,
"step": 340
},
{
"epoch": 3.108108108108108,
"grad_norm": 0.67578125,
"learning_rate": 0.00019311289178486018,
"loss": 0.7545,
"step": 345
},
{
"epoch": 3.153153153153153,
"grad_norm": 1.71875,
"learning_rate": 0.00019272536714968658,
"loss": 0.7531,
"step": 350
},
{
"epoch": 3.1981981981981984,
"grad_norm": 2.15625,
"learning_rate": 0.0001923276469633223,
"loss": 0.7731,
"step": 355
},
{
"epoch": 3.2432432432432434,
"grad_norm": 1.453125,
"learning_rate": 0.000191919774956805,
"loss": 0.7558,
"step": 360
},
{
"epoch": 3.2882882882882885,
"grad_norm": 1.71875,
"learning_rate": 0.0001915017959774084,
"loss": 0.782,
"step": 365
},
{
"epoch": 3.3333333333333335,
"grad_norm": 0.7265625,
"learning_rate": 0.00019107375598371112,
"loss": 0.7795,
"step": 370
},
{
"epoch": 3.3783783783783785,
"grad_norm": 1.71875,
"learning_rate": 0.0001906357020405435,
"loss": 0.7638,
"step": 375
},
{
"epoch": 3.4234234234234235,
"grad_norm": 1.7109375,
"learning_rate": 0.00019018768231381238,
"loss": 0.7511,
"step": 380
},
{
"epoch": 3.4684684684684686,
"grad_norm": 0.6015625,
"learning_rate": 0.0001897297460652053,
"loss": 0.7591,
"step": 385
},
{
"epoch": 3.5135135135135136,
"grad_norm": 7.625,
"learning_rate": 0.0001892619436467738,
"loss": 0.7433,
"step": 390
},
{
"epoch": 3.5585585585585586,
"grad_norm": 1.171875,
"learning_rate": 0.00018878432649539696,
"loss": 0.7673,
"step": 395
},
{
"epoch": 3.6036036036036037,
"grad_norm": 0.734375,
"learning_rate": 0.0001882969471271259,
"loss": 0.7633,
"step": 400
},
{
"epoch": 3.6486486486486487,
"grad_norm": 0.93359375,
"learning_rate": 0.00018779985913140924,
"loss": 0.7751,
"step": 405
},
{
"epoch": 3.6936936936936937,
"grad_norm": 0.9140625,
"learning_rate": 0.00018729311716520074,
"loss": 0.7628,
"step": 410
},
{
"epoch": 3.7387387387387387,
"grad_norm": 1.09375,
"learning_rate": 0.00018677677694694958,
"loss": 0.7553,
"step": 415
},
{
"epoch": 3.7837837837837838,
"grad_norm": 0.7578125,
"learning_rate": 0.00018625089525047385,
"loss": 0.7636,
"step": 420
},
{
"epoch": 3.828828828828829,
"grad_norm": 0.76953125,
"learning_rate": 0.00018571552989871806,
"loss": 0.7764,
"step": 425
},
{
"epoch": 3.873873873873874,
"grad_norm": 0.84765625,
"learning_rate": 0.00018517073975739514,
"loss": 0.7715,
"step": 430
},
{
"epoch": 3.918918918918919,
"grad_norm": 0.80859375,
"learning_rate": 0.0001846165847285141,
"loss": 0.77,
"step": 435
},
{
"epoch": 3.963963963963964,
"grad_norm": 3.0,
"learning_rate": 0.0001840531257437934,
"loss": 0.7873,
"step": 440
},
{
"epoch": 4.0,
"eval_loss": 2.2036337852478027,
"eval_runtime": 1.0177,
"eval_samples_per_second": 4.913,
"eval_steps_per_second": 1.965,
"step": 444
},
{
"epoch": 4.009009009009009,
"grad_norm": 2.0625,
"learning_rate": 0.00018348042475796122,
"loss": 0.7489,
"step": 445
},
{
"epoch": 4.054054054054054,
"grad_norm": 0.6484375,
"learning_rate": 0.00018289854474194347,
"loss": 0.6615,
"step": 450
},
{
"epoch": 4.099099099099099,
"grad_norm": 1.2421875,
"learning_rate": 0.00018230754967593963,
"loss": 0.6595,
"step": 455
},
{
"epoch": 4.1441441441441444,
"grad_norm": 0.7578125,
"learning_rate": 0.00018170750454238793,
"loss": 0.6718,
"step": 460
},
{
"epoch": 4.1891891891891895,
"grad_norm": 0.89453125,
"learning_rate": 0.00018109847531882047,
"loss": 0.6528,
"step": 465
},
{
"epoch": 4.2342342342342345,
"grad_norm": 1.15625,
"learning_rate": 0.0001804805289706083,
"loss": 0.6505,
"step": 470
},
{
"epoch": 4.2792792792792795,
"grad_norm": 1.25,
"learning_rate": 0.0001798537334435986,
"loss": 0.6708,
"step": 475
},
{
"epoch": 4.324324324324325,
"grad_norm": 0.89453125,
"learning_rate": 0.0001792181576566437,
"loss": 0.6702,
"step": 480
},
{
"epoch": 4.36936936936937,
"grad_norm": 0.734375,
"learning_rate": 0.0001785738714940231,
"loss": 0.6755,
"step": 485
},
{
"epoch": 4.414414414414415,
"grad_norm": 1.46875,
"learning_rate": 0.00017792094579775926,
"loss": 0.6825,
"step": 490
},
{
"epoch": 4.45945945945946,
"grad_norm": 1.0390625,
"learning_rate": 0.00017725945235982856,
"loss": 0.6864,
"step": 495
},
{
"epoch": 4.504504504504505,
"grad_norm": 0.62890625,
"learning_rate": 0.0001765894639142671,
"loss": 0.6758,
"step": 500
},
{
"epoch": 4.54954954954955,
"grad_norm": 0.71875,
"learning_rate": 0.0001759110541291736,
"loss": 0.6823,
"step": 505
},
{
"epoch": 4.594594594594595,
"grad_norm": 0.6875,
"learning_rate": 0.000175224297598609,
"loss": 0.6882,
"step": 510
},
{
"epoch": 4.63963963963964,
"grad_norm": 0.65234375,
"learning_rate": 0.00017452926983439475,
"loss": 0.6922,
"step": 515
},
{
"epoch": 4.684684684684685,
"grad_norm": 0.90234375,
"learning_rate": 0.0001738260472578097,
"loss": 0.6787,
"step": 520
},
{
"epoch": 4.72972972972973,
"grad_norm": 0.84765625,
"learning_rate": 0.00017311470719118765,
"loss": 0.7028,
"step": 525
},
{
"epoch": 4.774774774774775,
"grad_norm": 0.80859375,
"learning_rate": 0.00017239532784941495,
"loss": 0.6851,
"step": 530
},
{
"epoch": 4.81981981981982,
"grad_norm": 1.15625,
"learning_rate": 0.0001716679883313308,
"loss": 0.6756,
"step": 535
},
{
"epoch": 4.864864864864865,
"grad_norm": 1.34375,
"learning_rate": 0.00017093276861102988,
"loss": 0.6697,
"step": 540
},
{
"epoch": 4.90990990990991,
"grad_norm": 1.4296875,
"learning_rate": 0.00017018974952906884,
"loss": 0.689,
"step": 545
},
{
"epoch": 4.954954954954955,
"grad_norm": 0.68359375,
"learning_rate": 0.0001694390127835775,
"loss": 0.6894,
"step": 550
},
{
"epoch": 5.0,
"grad_norm": 1.390625,
"learning_rate": 0.00016868064092127587,
"loss": 0.685,
"step": 555
},
{
"epoch": 5.0,
"eval_loss": 2.297661304473877,
"eval_runtime": 1.0166,
"eval_samples_per_second": 4.918,
"eval_steps_per_second": 1.967,
"step": 555
},
{
"epoch": 5.045045045045045,
"grad_norm": 1.328125,
"learning_rate": 0.00016791471732839776,
"loss": 0.5743,
"step": 560
},
{
"epoch": 5.09009009009009,
"grad_norm": 0.7734375,
"learning_rate": 0.00016714132622152206,
"loss": 0.5415,
"step": 565
},
{
"epoch": 5.135135135135135,
"grad_norm": 0.78515625,
"learning_rate": 0.00016636055263831285,
"loss": 0.5752,
"step": 570
},
{
"epoch": 5.18018018018018,
"grad_norm": 0.91015625,
"learning_rate": 0.00016557248242816903,
"loss": 0.5778,
"step": 575
},
{
"epoch": 5.225225225225225,
"grad_norm": 1.015625,
"learning_rate": 0.00016477720224278492,
"loss": 0.5924,
"step": 580
},
{
"epoch": 5.27027027027027,
"grad_norm": 0.8046875,
"learning_rate": 0.00016397479952662248,
"loss": 0.5683,
"step": 585
},
{
"epoch": 5.315315315315315,
"grad_norm": 0.734375,
"learning_rate": 0.0001631653625072965,
"loss": 0.595,
"step": 590
},
{
"epoch": 5.36036036036036,
"grad_norm": 0.92578125,
"learning_rate": 0.00016234898018587337,
"loss": 0.5862,
"step": 595
},
{
"epoch": 5.405405405405405,
"grad_norm": 1.3671875,
"learning_rate": 0.00016152574232708534,
"loss": 0.5806,
"step": 600
},
{
"epoch": 5.45045045045045,
"grad_norm": 1.578125,
"learning_rate": 0.00016069573944946025,
"loss": 0.6002,
"step": 605
},
{
"epoch": 5.495495495495495,
"grad_norm": 1.1875,
"learning_rate": 0.00015985906281536875,
"loss": 0.6061,
"step": 610
},
{
"epoch": 5.54054054054054,
"grad_norm": 1.9765625,
"learning_rate": 0.00015901580442098968,
"loss": 0.6029,
"step": 615
},
{
"epoch": 5.585585585585585,
"grad_norm": 0.78515625,
"learning_rate": 0.00015816605698619452,
"loss": 0.5985,
"step": 620
},
{
"epoch": 5.63063063063063,
"grad_norm": 0.73046875,
"learning_rate": 0.0001573099139443525,
"loss": 0.5863,
"step": 625
},
{
"epoch": 5.675675675675675,
"grad_norm": 0.84375,
"learning_rate": 0.00015644746943205734,
"loss": 0.5927,
"step": 630
},
{
"epoch": 5.7207207207207205,
"grad_norm": 0.796875,
"learning_rate": 0.00015557881827877633,
"loss": 0.6072,
"step": 635
},
{
"epoch": 5.7657657657657655,
"grad_norm": 0.76171875,
"learning_rate": 0.0001547040559964236,
"loss": 0.6032,
"step": 640
},
{
"epoch": 5.8108108108108105,
"grad_norm": 0.7109375,
"learning_rate": 0.00015382327876885797,
"loss": 0.6135,
"step": 645
},
{
"epoch": 5.8558558558558556,
"grad_norm": 0.7265625,
"learning_rate": 0.00015293658344130734,
"loss": 0.6058,
"step": 650
},
{
"epoch": 5.900900900900901,
"grad_norm": 0.69140625,
"learning_rate": 0.00015204406750972,
"loss": 0.5948,
"step": 655
},
{
"epoch": 5.945945945945946,
"grad_norm": 0.65234375,
"learning_rate": 0.00015114582911004466,
"loss": 0.6108,
"step": 660
},
{
"epoch": 5.990990990990991,
"grad_norm": 0.79296875,
"learning_rate": 0.0001502419670074398,
"loss": 0.6223,
"step": 665
},
{
"epoch": 6.0,
"eval_loss": 2.4441299438476562,
"eval_runtime": 1.0183,
"eval_samples_per_second": 4.91,
"eval_steps_per_second": 1.964,
"step": 666
},
{
"epoch": 6.036036036036036,
"grad_norm": 1.109375,
"learning_rate": 0.00014933258058541415,
"loss": 0.5455,
"step": 670
},
{
"epoch": 6.081081081081081,
"grad_norm": 0.703125,
"learning_rate": 0.000148417769834899,
"loss": 0.5079,
"step": 675
},
{
"epoch": 6.126126126126126,
"grad_norm": 0.9375,
"learning_rate": 0.0001474976353432539,
"loss": 0.4906,
"step": 680
},
{
"epoch": 6.171171171171171,
"grad_norm": 0.96875,
"learning_rate": 0.00014657227828320635,
"loss": 0.5065,
"step": 685
},
{
"epoch": 6.216216216216216,
"grad_norm": 0.75,
"learning_rate": 0.0001456418004017278,
"loss": 0.4978,
"step": 690
},
{
"epoch": 6.261261261261261,
"grad_norm": 0.75390625,
"learning_rate": 0.00014470630400884575,
"loss": 0.4966,
"step": 695
},
{
"epoch": 6.306306306306306,
"grad_norm": 0.7578125,
"learning_rate": 0.00014376589196639467,
"loss": 0.4981,
"step": 700
},
{
"epoch": 6.351351351351352,
"grad_norm": 0.85546875,
"learning_rate": 0.00014282066767670572,
"loss": 0.5238,
"step": 705
},
{
"epoch": 6.396396396396397,
"grad_norm": 0.66015625,
"learning_rate": 0.0001418707350712372,
"loss": 0.5052,
"step": 710
},
{
"epoch": 6.441441441441442,
"grad_norm": 0.87890625,
"learning_rate": 0.00014091619859914694,
"loss": 0.5045,
"step": 715
},
{
"epoch": 6.486486486486487,
"grad_norm": 0.8515625,
"learning_rate": 0.0001399571632158076,
"loss": 0.5169,
"step": 720
},
{
"epoch": 6.531531531531532,
"grad_norm": 0.703125,
"learning_rate": 0.00013899373437126637,
"loss": 0.5208,
"step": 725
},
{
"epoch": 6.576576576576577,
"grad_norm": 0.66015625,
"learning_rate": 0.0001380260179986504,
"loss": 0.5192,
"step": 730
},
{
"epoch": 6.621621621621622,
"grad_norm": 0.77734375,
"learning_rate": 0.00013705412050251883,
"loss": 0.526,
"step": 735
},
{
"epoch": 6.666666666666667,
"grad_norm": 0.84375,
"learning_rate": 0.00013607814874716325,
"loss": 0.541,
"step": 740
},
{
"epoch": 6.711711711711712,
"grad_norm": 0.875,
"learning_rate": 0.0001350982100448575,
"loss": 0.5119,
"step": 745
},
{
"epoch": 6.756756756756757,
"grad_norm": 0.7578125,
"learning_rate": 0.00013411441214405829,
"loss": 0.5192,
"step": 750
},
{
"epoch": 6.801801801801802,
"grad_norm": 1.0703125,
"learning_rate": 0.00013312686321755761,
"loss": 0.5353,
"step": 755
},
{
"epoch": 6.846846846846847,
"grad_norm": 1.0078125,
"learning_rate": 0.00013213567185058888,
"loss": 0.5286,
"step": 760
},
{
"epoch": 6.891891891891892,
"grad_norm": 0.88671875,
"learning_rate": 0.0001311409470288874,
"loss": 0.5286,
"step": 765
},
{
"epoch": 6.936936936936937,
"grad_norm": 0.796875,
"learning_rate": 0.00013014279812670703,
"loss": 0.5353,
"step": 770
},
{
"epoch": 6.981981981981982,
"grad_norm": 0.6796875,
"learning_rate": 0.00012914133489479385,
"loss": 0.5378,
"step": 775
},
{
"epoch": 7.0,
"eval_loss": 2.6714916229248047,
"eval_runtime": 1.0156,
"eval_samples_per_second": 4.923,
"eval_steps_per_second": 1.969,
"step": 777
},
{
"epoch": 7.027027027027027,
"grad_norm": 0.74609375,
"learning_rate": 0.0001281366674483187,
"loss": 0.4628,
"step": 780
},
{
"epoch": 7.072072072072072,
"grad_norm": 0.75,
"learning_rate": 0.00012712890625476953,
"loss": 0.4205,
"step": 785
},
{
"epoch": 7.117117117117117,
"grad_norm": 0.84765625,
"learning_rate": 0.0001261181621218051,
"loss": 0.4286,
"step": 790
},
{
"epoch": 7.162162162162162,
"grad_norm": 0.8515625,
"learning_rate": 0.00012510454618507106,
"loss": 0.4393,
"step": 795
},
{
"epoch": 7.207207207207207,
"grad_norm": 0.765625,
"learning_rate": 0.0001240881698959802,
"loss": 0.4276,
"step": 800
},
{
"epoch": 7.252252252252252,
"grad_norm": 0.72265625,
"learning_rate": 0.00012306914500945788,
"loss": 0.4322,
"step": 805
},
{
"epoch": 7.297297297297297,
"grad_norm": 0.87890625,
"learning_rate": 0.00012204758357165409,
"loss": 0.442,
"step": 810
},
{
"epoch": 7.342342342342342,
"grad_norm": 0.7578125,
"learning_rate": 0.00012102359790762347,
"loss": 0.4401,
"step": 815
},
{
"epoch": 7.387387387387387,
"grad_norm": 0.73828125,
"learning_rate": 0.00011999730060897476,
"loss": 0.435,
"step": 820
},
{
"epoch": 7.4324324324324325,
"grad_norm": 0.8671875,
"learning_rate": 0.00011896880452149077,
"loss": 0.4448,
"step": 825
},
{
"epoch": 7.4774774774774775,
"grad_norm": 0.8203125,
"learning_rate": 0.00011793822273272066,
"loss": 0.4448,
"step": 830
},
{
"epoch": 7.5225225225225225,
"grad_norm": 0.78125,
"learning_rate": 0.00011690566855954523,
"loss": 0.4414,
"step": 835
},
{
"epoch": 7.5675675675675675,
"grad_norm": 0.8046875,
"learning_rate": 0.0001158712555357175,
"loss": 0.4461,
"step": 840
},
{
"epoch": 7.612612612612613,
"grad_norm": 0.78515625,
"learning_rate": 0.00011483509739937903,
"loss": 0.4455,
"step": 845
},
{
"epoch": 7.657657657657658,
"grad_norm": 0.84765625,
"learning_rate": 0.000113797308080554,
"loss": 0.4521,
"step": 850
},
{
"epoch": 7.702702702702703,
"grad_norm": 0.81640625,
"learning_rate": 0.00011275800168862214,
"loss": 0.4446,
"step": 855
},
{
"epoch": 7.747747747747748,
"grad_norm": 0.76171875,
"learning_rate": 0.00011171729249977169,
"loss": 0.4333,
"step": 860
},
{
"epoch": 7.792792792792793,
"grad_norm": 0.7734375,
"learning_rate": 0.00011067529494443463,
"loss": 0.4494,
"step": 865
},
{
"epoch": 7.837837837837838,
"grad_norm": 1.1796875,
"learning_rate": 0.00010963212359470421,
"loss": 0.4423,
"step": 870
},
{
"epoch": 7.882882882882883,
"grad_norm": 1.15625,
"learning_rate": 0.00010858789315173745,
"loss": 0.4424,
"step": 875
},
{
"epoch": 7.927927927927928,
"grad_norm": 0.76953125,
"learning_rate": 0.00010754271843314325,
"loss": 0.4491,
"step": 880
},
{
"epoch": 7.972972972972973,
"grad_norm": 0.9296875,
"learning_rate": 0.00010649671436035753,
"loss": 0.458,
"step": 885
},
{
"epoch": 8.0,
"eval_loss": 2.955517292022705,
"eval_runtime": 1.015,
"eval_samples_per_second": 4.926,
"eval_steps_per_second": 1.97,
"step": 888
},
{
"epoch": 8.018018018018019,
"grad_norm": 0.90625,
"learning_rate": 0.0001054499959460074,
"loss": 0.4181,
"step": 890
},
{
"epoch": 8.063063063063064,
"grad_norm": 0.87109375,
"learning_rate": 0.00010440267828126478,
"loss": 0.3664,
"step": 895
},
{
"epoch": 8.108108108108109,
"grad_norm": 1.0546875,
"learning_rate": 0.00010335487652319182,
"loss": 0.3723,
"step": 900
},
{
"epoch": 8.153153153153154,
"grad_norm": 0.734375,
"learning_rate": 0.00010230670588207873,
"loss": 0.3588,
"step": 905
},
{
"epoch": 8.198198198198199,
"grad_norm": 0.859375,
"learning_rate": 0.00010125828160877604,
"loss": 0.3619,
"step": 910
},
{
"epoch": 8.243243243243244,
"grad_norm": 0.765625,
"learning_rate": 0.00010020971898202218,
"loss": 0.3651,
"step": 915
},
{
"epoch": 8.288288288288289,
"grad_norm": 0.90625,
"learning_rate": 9.916113329576818e-05,
"loss": 0.3742,
"step": 920
},
{
"epoch": 8.333333333333334,
"grad_norm": 0.8359375,
"learning_rate": 9.811263984650056e-05,
"loss": 0.3752,
"step": 925
},
{
"epoch": 8.378378378378379,
"grad_norm": 1.015625,
"learning_rate": 9.706435392056394e-05,
"loss": 0.3746,
"step": 930
},
{
"epoch": 8.423423423423424,
"grad_norm": 0.8046875,
"learning_rate": 9.601639078148489e-05,
"loss": 0.3744,
"step": 935
},
{
"epoch": 8.468468468468469,
"grad_norm": 0.82421875,
"learning_rate": 9.496886565729823e-05,
"loss": 0.3765,
"step": 940
},
{
"epoch": 8.513513513513514,
"grad_norm": 0.73828125,
"learning_rate": 9.392189372787706e-05,
"loss": 0.371,
"step": 945
},
{
"epoch": 8.558558558558559,
"grad_norm": 0.796875,
"learning_rate": 9.287559011226843e-05,
"loss": 0.3747,
"step": 950
},
{
"epoch": 8.603603603603604,
"grad_norm": 0.80859375,
"learning_rate": 9.183006985603545e-05,
"loss": 0.372,
"step": 955
},
{
"epoch": 8.64864864864865,
"grad_norm": 0.7578125,
"learning_rate": 9.078544791860739e-05,
"loss": 0.3764,
"step": 960
},
{
"epoch": 8.693693693693694,
"grad_norm": 0.81640625,
"learning_rate": 8.974183916063968e-05,
"loss": 0.375,
"step": 965
},
{
"epoch": 8.73873873873874,
"grad_norm": 0.75,
"learning_rate": 8.869935833138414e-05,
"loss": 0.3836,
"step": 970
},
{
"epoch": 8.783783783783784,
"grad_norm": 0.75390625,
"learning_rate": 8.765812005607217e-05,
"loss": 0.3804,
"step": 975
},
{
"epoch": 8.82882882882883,
"grad_norm": 0.78515625,
"learning_rate": 8.6618238823311e-05,
"loss": 0.3857,
"step": 980
},
{
"epoch": 8.873873873873874,
"grad_norm": 0.81640625,
"learning_rate": 8.557982897249513e-05,
"loss": 0.383,
"step": 985
},
{
"epoch": 8.91891891891892,
"grad_norm": 0.8671875,
"learning_rate": 8.45430046812344e-05,
"loss": 0.3831,
"step": 990
},
{
"epoch": 8.963963963963964,
"grad_norm": 0.83203125,
"learning_rate": 8.35078799527996e-05,
"loss": 0.3843,
"step": 995
},
{
"epoch": 9.0,
"eval_loss": 3.43645977973938,
"eval_runtime": 1.0159,
"eval_samples_per_second": 4.922,
"eval_steps_per_second": 1.969,
"step": 999
},
{
"epoch": 9.00900900900901,
"grad_norm": 0.734375,
"learning_rate": 8.247456860358725e-05,
"loss": 0.3691,
"step": 1000
},
{
"epoch": 9.054054054054054,
"grad_norm": 1.0546875,
"learning_rate": 8.144318425060509e-05,
"loss": 0.32,
"step": 1005
},
{
"epoch": 9.0990990990991,
"grad_norm": 0.7890625,
"learning_rate": 8.04138402989794e-05,
"loss": 0.3192,
"step": 1010
},
{
"epoch": 9.144144144144144,
"grad_norm": 0.88671875,
"learning_rate": 7.938664992948549e-05,
"loss": 0.3166,
"step": 1015
},
{
"epoch": 9.18918918918919,
"grad_norm": 0.72265625,
"learning_rate": 7.83617260861032e-05,
"loss": 0.3123,
"step": 1020
},
{
"epoch": 9.234234234234235,
"grad_norm": 0.87890625,
"learning_rate": 7.733918146359815e-05,
"loss": 0.3192,
"step": 1025
},
{
"epoch": 9.27927927927928,
"grad_norm": 0.8046875,
"learning_rate": 7.631912849513025e-05,
"loss": 0.3127,
"step": 1030
},
{
"epoch": 9.324324324324325,
"grad_norm": 0.765625,
"learning_rate": 7.530167933989161e-05,
"loss": 0.3149,
"step": 1035
},
{
"epoch": 9.36936936936937,
"grad_norm": 0.7890625,
"learning_rate": 7.428694587077371e-05,
"loss": 0.318,
"step": 1040
},
{
"epoch": 9.414414414414415,
"grad_norm": 0.8125,
"learning_rate": 7.32750396620669e-05,
"loss": 0.3136,
"step": 1045
},
{
"epoch": 9.45945945945946,
"grad_norm": 0.859375,
"learning_rate": 7.226607197719202e-05,
"loss": 0.3193,
"step": 1050
},
{
"epoch": 9.504504504504505,
"grad_norm": 0.75390625,
"learning_rate": 7.126015375646666e-05,
"loss": 0.319,
"step": 1055
},
{
"epoch": 9.54954954954955,
"grad_norm": 0.75,
"learning_rate": 7.025739560490675e-05,
"loss": 0.3246,
"step": 1060
},
{
"epoch": 9.594594594594595,
"grad_norm": 0.79296875,
"learning_rate": 6.925790778006518e-05,
"loss": 0.3223,
"step": 1065
},
{
"epoch": 9.63963963963964,
"grad_norm": 0.77734375,
"learning_rate": 6.826180017990828e-05,
"loss": 0.3238,
"step": 1070
},
{
"epoch": 9.684684684684685,
"grad_norm": 0.80078125,
"learning_rate": 6.726918233073231e-05,
"loss": 0.321,
"step": 1075
},
{
"epoch": 9.72972972972973,
"grad_norm": 0.8203125,
"learning_rate": 6.62801633751204e-05,
"loss": 0.3161,
"step": 1080
},
{
"epoch": 9.774774774774775,
"grad_norm": 0.83203125,
"learning_rate": 6.52948520599419e-05,
"loss": 0.3221,
"step": 1085
},
{
"epoch": 9.81981981981982,
"grad_norm": 0.7734375,
"learning_rate": 6.431335672439531e-05,
"loss": 0.3251,
"step": 1090
},
{
"epoch": 9.864864864864865,
"grad_norm": 0.7890625,
"learning_rate": 6.333578528809574e-05,
"loss": 0.3246,
"step": 1095
},
{
"epoch": 9.90990990990991,
"grad_norm": 0.79296875,
"learning_rate": 6.236224523920879e-05,
"loss": 0.3238,
"step": 1100
},
{
"epoch": 9.954954954954955,
"grad_norm": 0.765625,
"learning_rate": 6.139284362263185e-05,
"loss": 0.3194,
"step": 1105
},
{
"epoch": 10.0,
"grad_norm": 0.7109375,
"learning_rate": 6.042768702822381e-05,
"loss": 0.3241,
"step": 1110
},
{
"epoch": 10.0,
"eval_loss": 3.8823204040527344,
"eval_runtime": 1.0143,
"eval_samples_per_second": 4.929,
"eval_steps_per_second": 1.972,
"step": 1110
},
{
"epoch": 10.045045045045045,
"grad_norm": 0.7890625,
"learning_rate": 5.9466881579085275e-05,
"loss": 0.2776,
"step": 1115
},
{
"epoch": 10.09009009009009,
"grad_norm": 0.7890625,
"learning_rate": 5.8510532919889804e-05,
"loss": 0.2767,
"step": 1120
},
{
"epoch": 10.135135135135135,
"grad_norm": 0.69140625,
"learning_rate": 5.7558746205267756e-05,
"loss": 0.2734,
"step": 1125
},
{
"epoch": 10.18018018018018,
"grad_norm": 0.8203125,
"learning_rate": 5.6611626088244194e-05,
"loss": 0.2792,
"step": 1130
},
{
"epoch": 10.225225225225225,
"grad_norm": 0.734375,
"learning_rate": 5.56692767087318e-05,
"loss": 0.2753,
"step": 1135
},
{
"epoch": 10.27027027027027,
"grad_norm": 0.70703125,
"learning_rate": 5.4731801682080206e-05,
"loss": 0.2763,
"step": 1140
},
{
"epoch": 10.315315315315315,
"grad_norm": 0.7265625,
"learning_rate": 5.3799304087683146e-05,
"loss": 0.2772,
"step": 1145
},
{
"epoch": 10.36036036036036,
"grad_norm": 0.79296875,
"learning_rate": 5.28718864576442e-05,
"loss": 0.2763,
"step": 1150
},
{
"epoch": 10.405405405405405,
"grad_norm": 0.75,
"learning_rate": 5.194965076550323e-05,
"loss": 0.2766,
"step": 1155
},
{
"epoch": 10.45045045045045,
"grad_norm": 0.84375,
"learning_rate": 5.1032698415023963e-05,
"loss": 0.2802,
"step": 1160
},
{
"epoch": 10.495495495495495,
"grad_norm": 0.8203125,
"learning_rate": 5.0121130229043786e-05,
"loss": 0.276,
"step": 1165
},
{
"epoch": 10.54054054054054,
"grad_norm": 0.79296875,
"learning_rate": 4.921504643838839e-05,
"loss": 0.2801,
"step": 1170
},
{
"epoch": 10.585585585585585,
"grad_norm": 0.7109375,
"learning_rate": 4.8314546670850594e-05,
"loss": 0.2775,
"step": 1175
},
{
"epoch": 10.63063063063063,
"grad_norm": 0.7421875,
"learning_rate": 4.741972994023601e-05,
"loss": 0.2806,
"step": 1180
},
{
"epoch": 10.675675675675675,
"grad_norm": 0.7421875,
"learning_rate": 4.6530694635476e-05,
"loss": 0.2775,
"step": 1185
},
{
"epoch": 10.72072072072072,
"grad_norm": 0.7421875,
"learning_rate": 4.5647538509809416e-05,
"loss": 0.2802,
"step": 1190
},
{
"epoch": 10.765765765765765,
"grad_norm": 0.73828125,
"learning_rate": 4.477035867003405e-05,
"loss": 0.2751,
"step": 1195
},
{
"epoch": 10.81081081081081,
"grad_norm": 0.7734375,
"learning_rate": 4.38992515658297e-05,
"loss": 0.2802,
"step": 1200
},
{
"epoch": 10.855855855855856,
"grad_norm": 0.79296875,
"learning_rate": 4.303431297915277e-05,
"loss": 0.2748,
"step": 1205
},
{
"epoch": 10.9009009009009,
"grad_norm": 0.77734375,
"learning_rate": 4.2175638013704655e-05,
"loss": 0.2808,
"step": 1210
},
{
"epoch": 10.945945945945946,
"grad_norm": 0.7421875,
"learning_rate": 4.1323321084474896e-05,
"loss": 0.2795,
"step": 1215
},
{
"epoch": 10.99099099099099,
"grad_norm": 0.76953125,
"learning_rate": 4.0477455907359715e-05,
"loss": 0.2825,
"step": 1220
},
{
"epoch": 11.0,
"eval_loss": 4.404350757598877,
"eval_runtime": 1.0163,
"eval_samples_per_second": 4.92,
"eval_steps_per_second": 1.968,
"step": 1221
},
{
"epoch": 11.036036036036036,
"grad_norm": 0.609375,
"learning_rate": 3.963813548885751e-05,
"loss": 0.2565,
"step": 1225
},
{
"epoch": 11.08108108108108,
"grad_norm": 0.7109375,
"learning_rate": 3.880545211584251e-05,
"loss": 0.2458,
"step": 1230
},
{
"epoch": 11.126126126126126,
"grad_norm": 0.6796875,
"learning_rate": 3.7979497345417234e-05,
"loss": 0.257,
"step": 1235
},
{
"epoch": 11.17117117117117,
"grad_norm": 0.6640625,
"learning_rate": 3.716036199484574e-05,
"loss": 0.2449,
"step": 1240
},
{
"epoch": 11.216216216216216,
"grad_norm": 0.83203125,
"learning_rate": 3.634813613156753e-05,
"loss": 0.2529,
"step": 1245
},
{
"epoch": 11.26126126126126,
"grad_norm": 0.7421875,
"learning_rate": 3.554290906329438e-05,
"loss": 0.2519,
"step": 1250
},
{
"epoch": 11.306306306306306,
"grad_norm": 0.734375,
"learning_rate": 3.474476932819062e-05,
"loss": 0.2551,
"step": 1255
},
{
"epoch": 11.35135135135135,
"grad_norm": 0.69140625,
"learning_rate": 3.395380468513805e-05,
"loss": 0.2492,
"step": 1260
},
{
"epoch": 11.396396396396396,
"grad_norm": 0.703125,
"learning_rate": 3.317010210408617e-05,
"loss": 0.2501,
"step": 1265
},
{
"epoch": 11.441441441441441,
"grad_norm": 0.734375,
"learning_rate": 3.2393747756489845e-05,
"loss": 0.2493,
"step": 1270
},
{
"epoch": 11.486486486486486,
"grad_norm": 0.70703125,
"learning_rate": 3.162482700583402e-05,
"loss": 0.2475,
"step": 1275
},
{
"epoch": 11.531531531531531,
"grad_norm": 0.7578125,
"learning_rate": 3.0863424398248064e-05,
"loss": 0.2524,
"step": 1280
},
{
"epoch": 11.576576576576576,
"grad_norm": 0.6953125,
"learning_rate": 3.0109623653209163e-05,
"loss": 0.2489,
"step": 1285
},
{
"epoch": 11.621621621621621,
"grad_norm": 0.72265625,
"learning_rate": 2.936350765433713e-05,
"loss": 0.2507,
"step": 1290
},
{
"epoch": 11.666666666666666,
"grad_norm": 0.68359375,
"learning_rate": 2.862515844028111e-05,
"loss": 0.2478,
"step": 1295
},
{
"epoch": 11.711711711711711,
"grad_norm": 0.7109375,
"learning_rate": 2.789465719569906e-05,
"loss": 0.2521,
"step": 1300
},
{
"epoch": 11.756756756756756,
"grad_norm": 0.6875,
"learning_rate": 2.717208424233091e-05,
"loss": 0.2525,
"step": 1305
},
{
"epoch": 11.801801801801801,
"grad_norm": 0.7421875,
"learning_rate": 2.6457519030167143e-05,
"loss": 0.2566,
"step": 1310
},
{
"epoch": 11.846846846846846,
"grad_norm": 0.7109375,
"learning_rate": 2.57510401287128e-05,
"loss": 0.2495,
"step": 1315
},
{
"epoch": 11.891891891891891,
"grad_norm": 0.6796875,
"learning_rate": 2.5052725218348394e-05,
"loss": 0.2526,
"step": 1320
},
{
"epoch": 11.936936936936936,
"grad_norm": 0.71484375,
"learning_rate": 2.4362651081788667e-05,
"loss": 0.249,
"step": 1325
},
{
"epoch": 11.981981981981981,
"grad_norm": 0.77734375,
"learning_rate": 2.3680893595640043e-05,
"loss": 0.2549,
"step": 1330
},
{
"epoch": 12.0,
"eval_loss": 4.838156700134277,
"eval_runtime": 1.0155,
"eval_samples_per_second": 4.923,
"eval_steps_per_second": 1.969,
"step": 1332
},
{
"epoch": 12.027027027027026,
"grad_norm": 0.625,
"learning_rate": 2.3007527722057488e-05,
"loss": 0.2452,
"step": 1335
},
{
"epoch": 12.072072072072071,
"grad_norm": 0.6484375,
"learning_rate": 2.234262750050241e-05,
"loss": 0.2436,
"step": 1340
},
{
"epoch": 12.117117117117116,
"grad_norm": 0.6171875,
"learning_rate": 2.168626603960138e-05,
"loss": 0.2396,
"step": 1345
},
{
"epoch": 12.162162162162161,
"grad_norm": 0.671875,
"learning_rate": 2.1038515509107736e-05,
"loss": 0.2344,
"step": 1350
},
{
"epoch": 12.207207207207206,
"grad_norm": 0.71875,
"learning_rate": 2.0399447131966132e-05,
"loss": 0.2359,
"step": 1355
},
{
"epoch": 12.252252252252251,
"grad_norm": 0.65234375,
"learning_rate": 1.976913117648128e-05,
"loss": 0.2374,
"step": 1360
},
{
"epoch": 12.297297297297296,
"grad_norm": 0.6015625,
"learning_rate": 1.9147636948591653e-05,
"loss": 0.2406,
"step": 1365
},
{
"epoch": 12.342342342342342,
"grad_norm": 0.65234375,
"learning_rate": 1.8535032784249028e-05,
"loss": 0.2392,
"step": 1370
},
{
"epoch": 12.387387387387387,
"grad_norm": 0.703125,
"learning_rate": 1.7931386041904506e-05,
"loss": 0.2347,
"step": 1375
},
{
"epoch": 12.432432432432432,
"grad_norm": 0.66796875,
"learning_rate": 1.733676309510245e-05,
"loss": 0.241,
"step": 1380
},
{
"epoch": 12.477477477477478,
"grad_norm": 0.71875,
"learning_rate": 1.6751229325182195e-05,
"loss": 0.2392,
"step": 1385
},
{
"epoch": 12.522522522522522,
"grad_norm": 0.65234375,
"learning_rate": 1.6174849114089064e-05,
"loss": 0.2389,
"step": 1390
},
{
"epoch": 12.567567567567568,
"grad_norm": 0.703125,
"learning_rate": 1.5607685837295516e-05,
"loss": 0.2397,
"step": 1395
},
{
"epoch": 12.612612612612612,
"grad_norm": 0.6640625,
"learning_rate": 1.5049801856832557e-05,
"loss": 0.2343,
"step": 1400
},
{
"epoch": 12.657657657657658,
"grad_norm": 0.67578125,
"learning_rate": 1.4501258514432836e-05,
"loss": 0.2364,
"step": 1405
},
{
"epoch": 12.702702702702704,
"grad_norm": 0.69140625,
"learning_rate": 1.3962116124785863e-05,
"loss": 0.2416,
"step": 1410
},
{
"epoch": 12.747747747747749,
"grad_norm": 0.6796875,
"learning_rate": 1.3432433968906044e-05,
"loss": 0.238,
"step": 1415
},
{
"epoch": 12.792792792792794,
"grad_norm": 0.625,
"learning_rate": 1.2912270287614736e-05,
"loss": 0.2355,
"step": 1420
},
{
"epoch": 12.837837837837839,
"grad_norm": 0.65625,
"learning_rate": 1.240168227513614e-05,
"loss": 0.2388,
"step": 1425
},
{
"epoch": 12.882882882882884,
"grad_norm": 0.6640625,
"learning_rate": 1.1900726072808666e-05,
"loss": 0.2351,
"step": 1430
},
{
"epoch": 12.927927927927929,
"grad_norm": 0.65234375,
"learning_rate": 1.1409456762911985e-05,
"loss": 0.2393,
"step": 1435
},
{
"epoch": 12.972972972972974,
"grad_norm": 0.65234375,
"learning_rate": 1.0927928362610462e-05,
"loss": 0.2408,
"step": 1440
},
{
"epoch": 13.0,
"eval_loss": 5.061110496520996,
"eval_runtime": 1.0149,
"eval_samples_per_second": 4.927,
"eval_steps_per_second": 1.971,
"step": 1443
},
{
"epoch": 13.018018018018019,
"grad_norm": 0.68359375,
"learning_rate": 1.0456193818013726e-05,
"loss": 0.2395,
"step": 1445
},
{
"epoch": 13.063063063063064,
"grad_norm": 0.66015625,
"learning_rate": 9.99430499835503e-06,
"loss": 0.2342,
"step": 1450
},
{
"epoch": 13.108108108108109,
"grad_norm": 0.75390625,
"learning_rate": 9.542312690288036e-06,
"loss": 0.2342,
"step": 1455
},
{
"epoch": 13.153153153153154,
"grad_norm": 0.6015625,
"learning_rate": 9.100266592302542e-06,
"loss": 0.2332,
"step": 1460
},
{
"epoch": 13.198198198198199,
"grad_norm": 0.671875,
"learning_rate": 8.668215309259997e-06,
"loss": 0.2327,
"step": 1465
},
{
"epoch": 13.243243243243244,
"grad_norm": 0.62890625,
"learning_rate": 8.246206347049079e-06,
"loss": 0.2353,
"step": 1470
},
{
"epoch": 13.288288288288289,
"grad_norm": 0.66796875,
"learning_rate": 7.834286107362343e-06,
"loss": 0.2331,
"step": 1475
},
{
"epoch": 13.333333333333334,
"grad_norm": 0.67578125,
"learning_rate": 7.432499882594091e-06,
"loss": 0.2323,
"step": 1480
},
{
"epoch": 13.378378378378379,
"grad_norm": 0.68359375,
"learning_rate": 7.040891850860287e-06,
"loss": 0.2354,
"step": 1485
},
{
"epoch": 13.423423423423424,
"grad_norm": 0.6328125,
"learning_rate": 6.659505071140959e-06,
"loss": 0.2345,
"step": 1490
},
{
"epoch": 13.468468468468469,
"grad_norm": 0.62109375,
"learning_rate": 6.2883814785457575e-06,
"loss": 0.2376,
"step": 1495
},
{
"epoch": 13.513513513513514,
"grad_norm": 0.64453125,
"learning_rate": 5.927561879702903e-06,
"loss": 0.2361,
"step": 1500
},
{
"epoch": 13.558558558558559,
"grad_norm": 0.72265625,
"learning_rate": 5.577085948272431e-06,
"loss": 0.2331,
"step": 1505
},
{
"epoch": 13.603603603603604,
"grad_norm": 0.609375,
"learning_rate": 5.236992220583814e-06,
"loss": 0.2312,
"step": 1510
},
{
"epoch": 13.64864864864865,
"grad_norm": 0.6953125,
"learning_rate": 4.90731809139876e-06,
"loss": 0.2347,
"step": 1515
},
{
"epoch": 13.693693693693694,
"grad_norm": 0.62890625,
"learning_rate": 4.588099809799551e-06,
"loss": 0.2359,
"step": 1520
},
{
"epoch": 13.73873873873874,
"grad_norm": 0.62890625,
"learning_rate": 4.279372475203181e-06,
"loss": 0.2321,
"step": 1525
},
{
"epoch": 13.783783783783784,
"grad_norm": 0.65625,
"learning_rate": 3.981170033502158e-06,
"loss": 0.2326,
"step": 1530
},
{
"epoch": 13.82882882882883,
"grad_norm": 0.6328125,
"learning_rate": 3.6935252733319213e-06,
"loss": 0.2341,
"step": 1535
},
{
"epoch": 13.873873873873874,
"grad_norm": 0.6171875,
"learning_rate": 3.4164698224656066e-06,
"loss": 0.2308,
"step": 1540
},
{
"epoch": 13.91891891891892,
"grad_norm": 0.66796875,
"learning_rate": 3.1500341443364556e-06,
"loss": 0.2338,
"step": 1545
},
{
"epoch": 13.963963963963964,
"grad_norm": 0.66796875,
"learning_rate": 2.8942475346882435e-06,
"loss": 0.2361,
"step": 1550
},
{
"epoch": 14.0,
"eval_loss": 5.106085777282715,
"eval_runtime": 1.0149,
"eval_samples_per_second": 4.927,
"eval_steps_per_second": 1.971,
"step": 1554
},
{
"epoch": 14.00900900900901,
"grad_norm": 0.61328125,
"learning_rate": 2.649138118354011e-06,
"loss": 0.2352,
"step": 1555
},
{
"epoch": 14.054054054054054,
"grad_norm": 0.6796875,
"learning_rate": 2.4147328461637144e-06,
"loss": 0.2335,
"step": 1560
},
{
"epoch": 14.0990990990991,
"grad_norm": 0.6484375,
"learning_rate": 2.1910574919808304e-06,
"loss": 0.2308,
"step": 1565
},
{
"epoch": 14.144144144144144,
"grad_norm": 0.62890625,
"learning_rate": 1.978136649868345e-06,
"loss": 0.2345,
"step": 1570
},
{
"epoch": 14.18918918918919,
"grad_norm": 0.66015625,
"learning_rate": 1.775993731384662e-06,
"loss": 0.235,
"step": 1575
},
{
"epoch": 14.234234234234235,
"grad_norm": 0.6796875,
"learning_rate": 1.584650963009271e-06,
"loss": 0.2337,
"step": 1580
},
{
"epoch": 14.27927927927928,
"grad_norm": 0.68359375,
"learning_rate": 1.4041293836989377e-06,
"loss": 0.2319,
"step": 1585
},
{
"epoch": 14.324324324324325,
"grad_norm": 0.64453125,
"learning_rate": 1.234448842574365e-06,
"loss": 0.2332,
"step": 1590
},
{
"epoch": 14.36936936936937,
"grad_norm": 0.6328125,
"learning_rate": 1.075627996737627e-06,
"loss": 0.23,
"step": 1595
},
{
"epoch": 14.414414414414415,
"grad_norm": 0.6875,
"learning_rate": 9.276843092208553e-07,
"loss": 0.2353,
"step": 1600
},
{
"epoch": 14.45945945945946,
"grad_norm": 0.640625,
"learning_rate": 7.90634047066019e-07,
"loss": 0.2339,
"step": 1605
},
{
"epoch": 14.504504504504505,
"grad_norm": 0.6796875,
"learning_rate": 6.644922795363218e-07,
"loss": 0.2381,
"step": 1610
},
{
"epoch": 14.54954954954955,
"grad_norm": 0.66015625,
"learning_rate": 5.492728764592725e-07,
"loss": 0.2335,
"step": 1615
},
{
"epoch": 14.594594594594595,
"grad_norm": 0.65234375,
"learning_rate": 4.4498850670164906e-07,
"loss": 0.2343,
"step": 1620
},
{
"epoch": 14.63963963963964,
"grad_norm": 0.6015625,
"learning_rate": 3.5165063677645714e-07,
"loss": 0.2369,
"step": 1625
},
{
"epoch": 14.684684684684685,
"grad_norm": 0.65234375,
"learning_rate": 2.6926952958221674e-07,
"loss": 0.2367,
"step": 1630
},
{
"epoch": 14.72972972972973,
"grad_norm": 0.6328125,
"learning_rate": 1.9785424327440906e-07,
"loss": 0.2302,
"step": 1635
},
{
"epoch": 14.774774774774775,
"grad_norm": 0.6640625,
"learning_rate": 1.374126302696066e-07,
"loss": 0.237,
"step": 1640
},
{
"epoch": 14.81981981981982,
"grad_norm": 0.65625,
"learning_rate": 8.795133638197506e-08,
"loss": 0.2298,
"step": 1645
},
{
"epoch": 14.864864864864865,
"grad_norm": 0.7109375,
"learning_rate": 4.94758000925799e-08,
"loss": 0.2372,
"step": 1650
},
{
"epoch": 14.90990990990991,
"grad_norm": 0.62890625,
"learning_rate": 2.1990251951398144e-08,
"loss": 0.2312,
"step": 1655
},
{
"epoch": 14.954954954954955,
"grad_norm": 0.62890625,
"learning_rate": 5.497714112157048e-09,
"loss": 0.2311,
"step": 1660
},
{
"epoch": 15.0,
"grad_norm": 0.62109375,
"learning_rate": 0.0,
"loss": 0.2319,
"step": 1665
},
{
"epoch": 15.0,
"eval_loss": 5.111132621765137,
"eval_runtime": 1.0137,
"eval_samples_per_second": 4.932,
"eval_steps_per_second": 1.973,
"step": 1665
},
{
"epoch": 15.0,
"step": 1665,
"total_flos": 2.545573832974926e+18,
"train_loss": 1.4823458194016694,
"train_runtime": 13248.7618,
"train_samples_per_second": 2.007,
"train_steps_per_second": 0.126
}
],
"logging_steps": 5,
"max_steps": 1665,
"num_input_tokens_seen": 0,
"num_train_epochs": 15,
"save_steps": 100,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 2.545573832974926e+18,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}