Coyoteranger's picture
Upload 28 files
4952d9a verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.03930611603165453,
"eval_steps": 500,
"global_step": 1500,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0002620407735443635,
"grad_norm": 1.2820378541946411,
"learning_rate": 4.999563265377426e-05,
"loss": 0.6609,
"step": 10
},
{
"epoch": 0.000524081547088727,
"grad_norm": 1.3106062412261963,
"learning_rate": 4.999126530754852e-05,
"loss": 0.8066,
"step": 20
},
{
"epoch": 0.0007861223206330905,
"grad_norm": 1.6985974311828613,
"learning_rate": 4.9986897961322784e-05,
"loss": 0.6863,
"step": 30
},
{
"epoch": 0.001048163094177454,
"grad_norm": 1.2552741765975952,
"learning_rate": 4.9982530615097044e-05,
"loss": 0.5674,
"step": 40
},
{
"epoch": 0.0013102038677218176,
"grad_norm": 1.3928155899047852,
"learning_rate": 4.9978163268871304e-05,
"loss": 0.5992,
"step": 50
},
{
"epoch": 0.001572244641266181,
"grad_norm": 1.5794411897659302,
"learning_rate": 4.9973795922645565e-05,
"loss": 0.7939,
"step": 60
},
{
"epoch": 0.0018342854148105446,
"grad_norm": 1.4024512767791748,
"learning_rate": 4.9969428576419825e-05,
"loss": 0.6178,
"step": 70
},
{
"epoch": 0.002096326188354908,
"grad_norm": 1.2232664823532104,
"learning_rate": 4.9965061230194085e-05,
"loss": 0.7544,
"step": 80
},
{
"epoch": 0.0023583669618992714,
"grad_norm": 1.5368870496749878,
"learning_rate": 4.9960693883968346e-05,
"loss": 0.6645,
"step": 90
},
{
"epoch": 0.002620407735443635,
"grad_norm": 1.6177372932434082,
"learning_rate": 4.9956326537742606e-05,
"loss": 0.6329,
"step": 100
},
{
"epoch": 0.0028824485089879986,
"grad_norm": 1.3803173303604126,
"learning_rate": 4.995195919151687e-05,
"loss": 0.7114,
"step": 110
},
{
"epoch": 0.003144489282532362,
"grad_norm": 1.5191670656204224,
"learning_rate": 4.9947591845291134e-05,
"loss": 0.8717,
"step": 120
},
{
"epoch": 0.0034065300560767254,
"grad_norm": 1.2967548370361328,
"learning_rate": 4.994322449906539e-05,
"loss": 0.7618,
"step": 130
},
{
"epoch": 0.003668570829621089,
"grad_norm": 1.8742738962173462,
"learning_rate": 4.9938857152839655e-05,
"loss": 0.5896,
"step": 140
},
{
"epoch": 0.003930611603165452,
"grad_norm": 1.697966456413269,
"learning_rate": 4.993448980661391e-05,
"loss": 0.7493,
"step": 150
},
{
"epoch": 0.004192652376709816,
"grad_norm": 1.5282775163650513,
"learning_rate": 4.9930122460388175e-05,
"loss": 0.6359,
"step": 160
},
{
"epoch": 0.00445469315025418,
"grad_norm": 1.2776225805282593,
"learning_rate": 4.992575511416243e-05,
"loss": 0.7677,
"step": 170
},
{
"epoch": 0.004716733923798543,
"grad_norm": 1.2036848068237305,
"learning_rate": 4.9921387767936696e-05,
"loss": 0.6234,
"step": 180
},
{
"epoch": 0.0049787746973429066,
"grad_norm": 1.1325912475585938,
"learning_rate": 4.991702042171095e-05,
"loss": 0.624,
"step": 190
},
{
"epoch": 0.00524081547088727,
"grad_norm": 1.3846409320831299,
"learning_rate": 4.991265307548522e-05,
"loss": 0.5979,
"step": 200
},
{
"epoch": 0.005502856244431633,
"grad_norm": 1.9792439937591553,
"learning_rate": 4.990828572925948e-05,
"loss": 0.7897,
"step": 210
},
{
"epoch": 0.005764897017975997,
"grad_norm": 1.8546253442764282,
"learning_rate": 4.990391838303374e-05,
"loss": 0.7035,
"step": 220
},
{
"epoch": 0.006026937791520361,
"grad_norm": 1.5434975624084473,
"learning_rate": 4.9899551036808e-05,
"loss": 0.7092,
"step": 230
},
{
"epoch": 0.006288978565064724,
"grad_norm": 1.0328209400177002,
"learning_rate": 4.989518369058226e-05,
"loss": 0.6239,
"step": 240
},
{
"epoch": 0.006551019338609088,
"grad_norm": 0.9389006495475769,
"learning_rate": 4.989081634435652e-05,
"loss": 0.6957,
"step": 250
},
{
"epoch": 0.006813060112153451,
"grad_norm": 1.0274962186813354,
"learning_rate": 4.988644899813078e-05,
"loss": 0.5302,
"step": 260
},
{
"epoch": 0.0070751008856978145,
"grad_norm": 1.2626285552978516,
"learning_rate": 4.988208165190504e-05,
"loss": 0.6541,
"step": 270
},
{
"epoch": 0.007337141659242178,
"grad_norm": 1.4558316469192505,
"learning_rate": 4.98777143056793e-05,
"loss": 0.7284,
"step": 280
},
{
"epoch": 0.007599182432786541,
"grad_norm": 1.3997328281402588,
"learning_rate": 4.987334695945356e-05,
"loss": 0.6888,
"step": 290
},
{
"epoch": 0.007861223206330904,
"grad_norm": 1.7742432355880737,
"learning_rate": 4.986897961322782e-05,
"loss": 0.6869,
"step": 300
},
{
"epoch": 0.008123263979875268,
"grad_norm": 1.4925923347473145,
"learning_rate": 4.986461226700208e-05,
"loss": 0.731,
"step": 310
},
{
"epoch": 0.008385304753419632,
"grad_norm": 1.5941082239151,
"learning_rate": 4.986024492077634e-05,
"loss": 0.7149,
"step": 320
},
{
"epoch": 0.008647345526963996,
"grad_norm": 1.507450819015503,
"learning_rate": 4.98558775745506e-05,
"loss": 0.6443,
"step": 330
},
{
"epoch": 0.00890938630050836,
"grad_norm": 0.9866214990615845,
"learning_rate": 4.985151022832487e-05,
"loss": 0.6003,
"step": 340
},
{
"epoch": 0.009171427074052723,
"grad_norm": 0.9682250022888184,
"learning_rate": 4.984714288209912e-05,
"loss": 0.6602,
"step": 350
},
{
"epoch": 0.009433467847597085,
"grad_norm": 1.7567181587219238,
"learning_rate": 4.984277553587338e-05,
"loss": 0.5647,
"step": 360
},
{
"epoch": 0.00969550862114145,
"grad_norm": 1.4297257661819458,
"learning_rate": 4.9838408189647643e-05,
"loss": 0.6628,
"step": 370
},
{
"epoch": 0.009957549394685813,
"grad_norm": 1.515763521194458,
"learning_rate": 4.9834040843421904e-05,
"loss": 0.568,
"step": 380
},
{
"epoch": 0.010219590168230177,
"grad_norm": 1.1806342601776123,
"learning_rate": 4.9829673497196164e-05,
"loss": 0.6729,
"step": 390
},
{
"epoch": 0.01048163094177454,
"grad_norm": 1.1328020095825195,
"learning_rate": 4.9825306150970425e-05,
"loss": 0.6963,
"step": 400
},
{
"epoch": 0.010743671715318903,
"grad_norm": 0.7740004658699036,
"learning_rate": 4.9820938804744685e-05,
"loss": 0.6654,
"step": 410
},
{
"epoch": 0.011005712488863267,
"grad_norm": 0.9519413113594055,
"learning_rate": 4.9816571458518945e-05,
"loss": 0.6487,
"step": 420
},
{
"epoch": 0.01126775326240763,
"grad_norm": 0.8964847922325134,
"learning_rate": 4.981220411229321e-05,
"loss": 0.5667,
"step": 430
},
{
"epoch": 0.011529794035951994,
"grad_norm": 1.428072452545166,
"learning_rate": 4.9807836766067466e-05,
"loss": 0.8164,
"step": 440
},
{
"epoch": 0.011791834809496358,
"grad_norm": 1.4375147819519043,
"learning_rate": 4.980346941984173e-05,
"loss": 0.5476,
"step": 450
},
{
"epoch": 0.012053875583040722,
"grad_norm": 1.1702146530151367,
"learning_rate": 4.979910207361599e-05,
"loss": 0.7342,
"step": 460
},
{
"epoch": 0.012315916356585084,
"grad_norm": 1.2703320980072021,
"learning_rate": 4.9794734727390254e-05,
"loss": 0.5767,
"step": 470
},
{
"epoch": 0.012577957130129448,
"grad_norm": 1.2520267963409424,
"learning_rate": 4.979036738116451e-05,
"loss": 0.5969,
"step": 480
},
{
"epoch": 0.012839997903673812,
"grad_norm": 1.413979172706604,
"learning_rate": 4.9786000034938775e-05,
"loss": 0.7011,
"step": 490
},
{
"epoch": 0.013102038677218176,
"grad_norm": 1.3918565511703491,
"learning_rate": 4.978163268871303e-05,
"loss": 0.6041,
"step": 500
},
{
"epoch": 0.01336407945076254,
"grad_norm": 0.9175894260406494,
"learning_rate": 4.9777265342487296e-05,
"loss": 0.5052,
"step": 510
},
{
"epoch": 0.013626120224306901,
"grad_norm": 1.296505331993103,
"learning_rate": 4.9772897996261556e-05,
"loss": 0.6076,
"step": 520
},
{
"epoch": 0.013888160997851265,
"grad_norm": 1.2490183115005493,
"learning_rate": 4.9768530650035816e-05,
"loss": 0.6287,
"step": 530
},
{
"epoch": 0.014150201771395629,
"grad_norm": 1.398285984992981,
"learning_rate": 4.976416330381008e-05,
"loss": 0.7905,
"step": 540
},
{
"epoch": 0.014412242544939993,
"grad_norm": 1.3094829320907593,
"learning_rate": 4.975979595758434e-05,
"loss": 0.5133,
"step": 550
},
{
"epoch": 0.014674283318484357,
"grad_norm": 1.7128199338912964,
"learning_rate": 4.97554286113586e-05,
"loss": 0.824,
"step": 560
},
{
"epoch": 0.014936324092028719,
"grad_norm": 1.1319103240966797,
"learning_rate": 4.975106126513286e-05,
"loss": 0.5893,
"step": 570
},
{
"epoch": 0.015198364865573083,
"grad_norm": 1.6176029443740845,
"learning_rate": 4.974669391890712e-05,
"loss": 0.581,
"step": 580
},
{
"epoch": 0.015460405639117446,
"grad_norm": 1.590836524963379,
"learning_rate": 4.974232657268138e-05,
"loss": 0.5814,
"step": 590
},
{
"epoch": 0.01572244641266181,
"grad_norm": 1.6922227144241333,
"learning_rate": 4.973795922645564e-05,
"loss": 0.5917,
"step": 600
},
{
"epoch": 0.015984487186206174,
"grad_norm": 1.3479337692260742,
"learning_rate": 4.97335918802299e-05,
"loss": 0.7356,
"step": 610
},
{
"epoch": 0.016246527959750536,
"grad_norm": 2.2019124031066895,
"learning_rate": 4.972922453400416e-05,
"loss": 0.6197,
"step": 620
},
{
"epoch": 0.016508568733294902,
"grad_norm": 1.6211423873901367,
"learning_rate": 4.972485718777842e-05,
"loss": 0.6299,
"step": 630
},
{
"epoch": 0.016770609506839264,
"grad_norm": 1.157416582107544,
"learning_rate": 4.972048984155268e-05,
"loss": 0.6343,
"step": 640
},
{
"epoch": 0.01703265028038363,
"grad_norm": 1.2576712369918823,
"learning_rate": 4.971612249532695e-05,
"loss": 0.6126,
"step": 650
},
{
"epoch": 0.01729469105392799,
"grad_norm": 1.3852715492248535,
"learning_rate": 4.97117551491012e-05,
"loss": 0.6352,
"step": 660
},
{
"epoch": 0.017556731827472354,
"grad_norm": 1.0178048610687256,
"learning_rate": 4.970738780287547e-05,
"loss": 0.5923,
"step": 670
},
{
"epoch": 0.01781877260101672,
"grad_norm": 0.8760583996772766,
"learning_rate": 4.970302045664972e-05,
"loss": 0.6158,
"step": 680
},
{
"epoch": 0.01808081337456108,
"grad_norm": 0.8956984281539917,
"learning_rate": 4.969865311042398e-05,
"loss": 0.5746,
"step": 690
},
{
"epoch": 0.018342854148105447,
"grad_norm": 1.1126501560211182,
"learning_rate": 4.969428576419824e-05,
"loss": 0.6254,
"step": 700
},
{
"epoch": 0.01860489492164981,
"grad_norm": 1.168455958366394,
"learning_rate": 4.96899184179725e-05,
"loss": 0.652,
"step": 710
},
{
"epoch": 0.01886693569519417,
"grad_norm": 1.3628567457199097,
"learning_rate": 4.9685551071746764e-05,
"loss": 0.5789,
"step": 720
},
{
"epoch": 0.019128976468738536,
"grad_norm": 1.1971865892410278,
"learning_rate": 4.9681183725521024e-05,
"loss": 0.6408,
"step": 730
},
{
"epoch": 0.0193910172422829,
"grad_norm": 1.1916868686676025,
"learning_rate": 4.9676816379295285e-05,
"loss": 0.6461,
"step": 740
},
{
"epoch": 0.019653058015827264,
"grad_norm": 1.1797837018966675,
"learning_rate": 4.9672449033069545e-05,
"loss": 0.5843,
"step": 750
},
{
"epoch": 0.019915098789371626,
"grad_norm": 0.7941935658454895,
"learning_rate": 4.966808168684381e-05,
"loss": 0.6165,
"step": 760
},
{
"epoch": 0.02017713956291599,
"grad_norm": 1.4876329898834229,
"learning_rate": 4.9663714340618066e-05,
"loss": 0.6347,
"step": 770
},
{
"epoch": 0.020439180336460354,
"grad_norm": 1.1482038497924805,
"learning_rate": 4.965934699439233e-05,
"loss": 0.5662,
"step": 780
},
{
"epoch": 0.020701221110004716,
"grad_norm": 1.3942419290542603,
"learning_rate": 4.9654979648166586e-05,
"loss": 0.6189,
"step": 790
},
{
"epoch": 0.02096326188354908,
"grad_norm": 0.8826277256011963,
"learning_rate": 4.9650612301940854e-05,
"loss": 0.6801,
"step": 800
},
{
"epoch": 0.021225302657093444,
"grad_norm": 1.3729712963104248,
"learning_rate": 4.964624495571511e-05,
"loss": 0.5789,
"step": 810
},
{
"epoch": 0.021487343430637806,
"grad_norm": 0.747199296951294,
"learning_rate": 4.9641877609489374e-05,
"loss": 0.651,
"step": 820
},
{
"epoch": 0.02174938420418217,
"grad_norm": 0.7911145091056824,
"learning_rate": 4.963751026326363e-05,
"loss": 0.6834,
"step": 830
},
{
"epoch": 0.022011424977726533,
"grad_norm": 1.1725844144821167,
"learning_rate": 4.9633142917037895e-05,
"loss": 0.6687,
"step": 840
},
{
"epoch": 0.0222734657512709,
"grad_norm": 1.2759829759597778,
"learning_rate": 4.9628775570812156e-05,
"loss": 0.6612,
"step": 850
},
{
"epoch": 0.02253550652481526,
"grad_norm": 1.497684359550476,
"learning_rate": 4.9624408224586416e-05,
"loss": 0.686,
"step": 860
},
{
"epoch": 0.022797547298359623,
"grad_norm": 1.4431102275848389,
"learning_rate": 4.9620040878360676e-05,
"loss": 0.5838,
"step": 870
},
{
"epoch": 0.02305958807190399,
"grad_norm": 0.8864196538925171,
"learning_rate": 4.961567353213494e-05,
"loss": 0.6076,
"step": 880
},
{
"epoch": 0.02332162884544835,
"grad_norm": 1.4421597719192505,
"learning_rate": 4.96113061859092e-05,
"loss": 0.6669,
"step": 890
},
{
"epoch": 0.023583669618992716,
"grad_norm": 1.541601300239563,
"learning_rate": 4.960693883968346e-05,
"loss": 0.477,
"step": 900
},
{
"epoch": 0.02384571039253708,
"grad_norm": 1.0725853443145752,
"learning_rate": 4.960257149345772e-05,
"loss": 0.5399,
"step": 910
},
{
"epoch": 0.024107751166081444,
"grad_norm": 1.4579834938049316,
"learning_rate": 4.959820414723198e-05,
"loss": 0.636,
"step": 920
},
{
"epoch": 0.024369791939625806,
"grad_norm": 1.1018449068069458,
"learning_rate": 4.959383680100624e-05,
"loss": 0.7731,
"step": 930
},
{
"epoch": 0.024631832713170168,
"grad_norm": 1.3531861305236816,
"learning_rate": 4.95894694547805e-05,
"loss": 0.7017,
"step": 940
},
{
"epoch": 0.024893873486714534,
"grad_norm": 1.1225773096084595,
"learning_rate": 4.958510210855476e-05,
"loss": 0.7228,
"step": 950
},
{
"epoch": 0.025155914260258896,
"grad_norm": 0.6442508697509766,
"learning_rate": 4.958073476232902e-05,
"loss": 0.4421,
"step": 960
},
{
"epoch": 0.02541795503380326,
"grad_norm": 1.1338638067245483,
"learning_rate": 4.957636741610328e-05,
"loss": 0.6533,
"step": 970
},
{
"epoch": 0.025679995807347623,
"grad_norm": 1.0796573162078857,
"learning_rate": 4.957200006987755e-05,
"loss": 0.6407,
"step": 980
},
{
"epoch": 0.025942036580891985,
"grad_norm": 1.001578450202942,
"learning_rate": 4.95676327236518e-05,
"loss": 0.652,
"step": 990
},
{
"epoch": 0.02620407735443635,
"grad_norm": 1.521545648574829,
"learning_rate": 4.956326537742607e-05,
"loss": 0.4812,
"step": 1000
},
{
"epoch": 0.026466118127980713,
"grad_norm": 1.1855580806732178,
"learning_rate": 4.955889803120032e-05,
"loss": 0.5927,
"step": 1010
},
{
"epoch": 0.02672815890152508,
"grad_norm": 1.3463046550750732,
"learning_rate": 4.955453068497459e-05,
"loss": 0.7304,
"step": 1020
},
{
"epoch": 0.02699019967506944,
"grad_norm": 1.0500322580337524,
"learning_rate": 4.955016333874884e-05,
"loss": 0.5409,
"step": 1030
},
{
"epoch": 0.027252240448613803,
"grad_norm": 1.4060779809951782,
"learning_rate": 4.95457959925231e-05,
"loss": 0.7348,
"step": 1040
},
{
"epoch": 0.02751428122215817,
"grad_norm": 1.433759331703186,
"learning_rate": 4.954142864629736e-05,
"loss": 0.6079,
"step": 1050
},
{
"epoch": 0.02777632199570253,
"grad_norm": 2.5895800590515137,
"learning_rate": 4.9537061300071624e-05,
"loss": 0.5789,
"step": 1060
},
{
"epoch": 0.028038362769246896,
"grad_norm": 1.1626940965652466,
"learning_rate": 4.953269395384589e-05,
"loss": 0.5478,
"step": 1070
},
{
"epoch": 0.028300403542791258,
"grad_norm": 1.6183115243911743,
"learning_rate": 4.9528326607620144e-05,
"loss": 0.6782,
"step": 1080
},
{
"epoch": 0.02856244431633562,
"grad_norm": 1.5321258306503296,
"learning_rate": 4.952395926139441e-05,
"loss": 0.5878,
"step": 1090
},
{
"epoch": 0.028824485089879986,
"grad_norm": 0.9456602334976196,
"learning_rate": 4.9519591915168665e-05,
"loss": 0.6277,
"step": 1100
},
{
"epoch": 0.029086525863424348,
"grad_norm": 1.0988554954528809,
"learning_rate": 4.951522456894293e-05,
"loss": 0.5019,
"step": 1110
},
{
"epoch": 0.029348566636968713,
"grad_norm": 1.053934931755066,
"learning_rate": 4.9510857222717186e-05,
"loss": 0.6578,
"step": 1120
},
{
"epoch": 0.029610607410513076,
"grad_norm": 1.292533040046692,
"learning_rate": 4.950648987649145e-05,
"loss": 0.6115,
"step": 1130
},
{
"epoch": 0.029872648184057438,
"grad_norm": 0.8503355383872986,
"learning_rate": 4.950212253026571e-05,
"loss": 0.56,
"step": 1140
},
{
"epoch": 0.030134688957601803,
"grad_norm": 1.5698297023773193,
"learning_rate": 4.9497755184039974e-05,
"loss": 0.6208,
"step": 1150
},
{
"epoch": 0.030396729731146165,
"grad_norm": 1.3464306592941284,
"learning_rate": 4.9493387837814234e-05,
"loss": 0.5812,
"step": 1160
},
{
"epoch": 0.03065877050469053,
"grad_norm": 0.9388158321380615,
"learning_rate": 4.9489020491588495e-05,
"loss": 0.5684,
"step": 1170
},
{
"epoch": 0.030920811278234893,
"grad_norm": 1.271061897277832,
"learning_rate": 4.9484653145362755e-05,
"loss": 0.5456,
"step": 1180
},
{
"epoch": 0.03118285205177926,
"grad_norm": 0.8666424751281738,
"learning_rate": 4.9480285799137015e-05,
"loss": 0.5418,
"step": 1190
},
{
"epoch": 0.03144489282532362,
"grad_norm": 1.2482844591140747,
"learning_rate": 4.9475918452911276e-05,
"loss": 0.7439,
"step": 1200
},
{
"epoch": 0.03170693359886798,
"grad_norm": 1.226784110069275,
"learning_rate": 4.9471551106685536e-05,
"loss": 0.5813,
"step": 1210
},
{
"epoch": 0.03196897437241235,
"grad_norm": 1.1704344749450684,
"learning_rate": 4.9467183760459797e-05,
"loss": 0.757,
"step": 1220
},
{
"epoch": 0.032231015145956714,
"grad_norm": 1.0429288148880005,
"learning_rate": 4.946281641423406e-05,
"loss": 0.5162,
"step": 1230
},
{
"epoch": 0.03249305591950107,
"grad_norm": 1.3331559896469116,
"learning_rate": 4.945844906800832e-05,
"loss": 0.6224,
"step": 1240
},
{
"epoch": 0.03275509669304544,
"grad_norm": 1.2090318202972412,
"learning_rate": 4.945408172178258e-05,
"loss": 0.6473,
"step": 1250
},
{
"epoch": 0.033017137466589803,
"grad_norm": 1.3969329595565796,
"learning_rate": 4.944971437555684e-05,
"loss": 0.5806,
"step": 1260
},
{
"epoch": 0.03327917824013416,
"grad_norm": 1.6032638549804688,
"learning_rate": 4.94453470293311e-05,
"loss": 0.5766,
"step": 1270
},
{
"epoch": 0.03354121901367853,
"grad_norm": 1.0155426263809204,
"learning_rate": 4.944097968310536e-05,
"loss": 0.5711,
"step": 1280
},
{
"epoch": 0.03380325978722289,
"grad_norm": 1.2512564659118652,
"learning_rate": 4.9436612336879626e-05,
"loss": 0.6783,
"step": 1290
},
{
"epoch": 0.03406530056076726,
"grad_norm": 1.0756443738937378,
"learning_rate": 4.943224499065388e-05,
"loss": 0.6255,
"step": 1300
},
{
"epoch": 0.03432734133431162,
"grad_norm": 1.2002949714660645,
"learning_rate": 4.942787764442815e-05,
"loss": 0.6057,
"step": 1310
},
{
"epoch": 0.03458938210785598,
"grad_norm": 0.6843838691711426,
"learning_rate": 4.94235102982024e-05,
"loss": 0.5411,
"step": 1320
},
{
"epoch": 0.03485142288140035,
"grad_norm": 0.8038992285728455,
"learning_rate": 4.941914295197667e-05,
"loss": 0.6097,
"step": 1330
},
{
"epoch": 0.03511346365494471,
"grad_norm": 1.2226810455322266,
"learning_rate": 4.941477560575092e-05,
"loss": 0.5731,
"step": 1340
},
{
"epoch": 0.03537550442848907,
"grad_norm": 1.236194372177124,
"learning_rate": 4.941040825952519e-05,
"loss": 0.5677,
"step": 1350
},
{
"epoch": 0.03563754520203344,
"grad_norm": 1.2129302024841309,
"learning_rate": 4.940604091329944e-05,
"loss": 0.6031,
"step": 1360
},
{
"epoch": 0.0358995859755778,
"grad_norm": 1.5371068716049194,
"learning_rate": 4.94016735670737e-05,
"loss": 0.5957,
"step": 1370
},
{
"epoch": 0.03616162674912216,
"grad_norm": 1.6025753021240234,
"learning_rate": 4.939730622084797e-05,
"loss": 0.6261,
"step": 1380
},
{
"epoch": 0.03642366752266653,
"grad_norm": 1.3252272605895996,
"learning_rate": 4.939293887462222e-05,
"loss": 0.626,
"step": 1390
},
{
"epoch": 0.036685708296210894,
"grad_norm": 1.0813733339309692,
"learning_rate": 4.938857152839649e-05,
"loss": 0.6659,
"step": 1400
},
{
"epoch": 0.03694774906975525,
"grad_norm": 1.280362844467163,
"learning_rate": 4.9384204182170744e-05,
"loss": 0.5828,
"step": 1410
},
{
"epoch": 0.03720978984329962,
"grad_norm": 0.7822287082672119,
"learning_rate": 4.937983683594501e-05,
"loss": 0.4939,
"step": 1420
},
{
"epoch": 0.03747183061684398,
"grad_norm": 1.0170814990997314,
"learning_rate": 4.9375469489719265e-05,
"loss": 0.5846,
"step": 1430
},
{
"epoch": 0.03773387139038834,
"grad_norm": 1.4380531311035156,
"learning_rate": 4.937110214349353e-05,
"loss": 0.5203,
"step": 1440
},
{
"epoch": 0.03799591216393271,
"grad_norm": 0.9485931396484375,
"learning_rate": 4.9366734797267785e-05,
"loss": 0.4703,
"step": 1450
},
{
"epoch": 0.03825795293747707,
"grad_norm": 1.4566253423690796,
"learning_rate": 4.936236745104205e-05,
"loss": 0.6879,
"step": 1460
},
{
"epoch": 0.03851999371102143,
"grad_norm": 1.3309845924377441,
"learning_rate": 4.935800010481631e-05,
"loss": 0.6359,
"step": 1470
},
{
"epoch": 0.0387820344845658,
"grad_norm": 0.8272371292114258,
"learning_rate": 4.9353632758590573e-05,
"loss": 0.5521,
"step": 1480
},
{
"epoch": 0.03904407525811016,
"grad_norm": 1.122835397720337,
"learning_rate": 4.9349265412364834e-05,
"loss": 0.5871,
"step": 1490
},
{
"epoch": 0.03930611603165453,
"grad_norm": 1.701686978340149,
"learning_rate": 4.9344898066139094e-05,
"loss": 0.5549,
"step": 1500
}
],
"logging_steps": 10,
"max_steps": 114486,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 5608556199936000.0,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}