Qwen2.5-0.5B-Open-R1-Distill / trainer_state.json
Qucy's picture
Model save
e5ea3de verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.9994447529150472,
"eval_steps": 100,
"global_step": 675,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.007403294466037387,
"grad_norm": 0.6776805171272973,
"learning_rate": 1.4705882352941177e-06,
"loss": 1.2512,
"step": 5
},
{
"epoch": 0.014806588932074774,
"grad_norm": 0.44944873495655524,
"learning_rate": 2.9411764705882355e-06,
"loss": 1.2594,
"step": 10
},
{
"epoch": 0.02220988339811216,
"grad_norm": 0.34624498626136563,
"learning_rate": 4.411764705882353e-06,
"loss": 1.2327,
"step": 15
},
{
"epoch": 0.029613177864149548,
"grad_norm": 0.3384934751143345,
"learning_rate": 5.882352941176471e-06,
"loss": 1.1823,
"step": 20
},
{
"epoch": 0.037016472330186935,
"grad_norm": 0.2415385954324663,
"learning_rate": 7.352941176470589e-06,
"loss": 1.1483,
"step": 25
},
{
"epoch": 0.04441976679622432,
"grad_norm": 0.21956265650508958,
"learning_rate": 8.823529411764707e-06,
"loss": 1.1051,
"step": 30
},
{
"epoch": 0.05182306126226171,
"grad_norm": 0.18350394805421755,
"learning_rate": 1.0294117647058823e-05,
"loss": 1.0659,
"step": 35
},
{
"epoch": 0.059226355728299096,
"grad_norm": 0.16210146685861945,
"learning_rate": 1.1764705882352942e-05,
"loss": 1.0381,
"step": 40
},
{
"epoch": 0.06662965019433648,
"grad_norm": 0.1607004827826561,
"learning_rate": 1.323529411764706e-05,
"loss": 1.0217,
"step": 45
},
{
"epoch": 0.07403294466037387,
"grad_norm": 0.16745984502215897,
"learning_rate": 1.4705882352941179e-05,
"loss": 1.038,
"step": 50
},
{
"epoch": 0.08143623912641125,
"grad_norm": 0.14843663256682382,
"learning_rate": 1.6176470588235296e-05,
"loss": 1.0276,
"step": 55
},
{
"epoch": 0.08883953359244864,
"grad_norm": 0.17168834832578023,
"learning_rate": 1.7647058823529414e-05,
"loss": 1.0032,
"step": 60
},
{
"epoch": 0.09624282805848602,
"grad_norm": 0.15126385928527755,
"learning_rate": 1.911764705882353e-05,
"loss": 0.9954,
"step": 65
},
{
"epoch": 0.10364612252452342,
"grad_norm": 0.1688834438402727,
"learning_rate": 1.9999464266898485e-05,
"loss": 0.9919,
"step": 70
},
{
"epoch": 0.1110494169905608,
"grad_norm": 0.20449596627344765,
"learning_rate": 1.9993437928712977e-05,
"loss": 0.9839,
"step": 75
},
{
"epoch": 0.11845271145659819,
"grad_norm": 0.17590208096396392,
"learning_rate": 1.998071963486563e-05,
"loss": 0.9714,
"step": 80
},
{
"epoch": 0.12585600592263557,
"grad_norm": 0.14935660715610657,
"learning_rate": 1.9961317901970953e-05,
"loss": 0.9577,
"step": 85
},
{
"epoch": 0.13325930038867295,
"grad_norm": 0.1656804473286765,
"learning_rate": 1.993524572210807e-05,
"loss": 0.9607,
"step": 90
},
{
"epoch": 0.14066259485471036,
"grad_norm": 0.20165204812335902,
"learning_rate": 1.990252055412077e-05,
"loss": 0.9538,
"step": 95
},
{
"epoch": 0.14806588932074774,
"grad_norm": 0.1672105719290636,
"learning_rate": 1.9863164311926433e-05,
"loss": 0.9842,
"step": 100
},
{
"epoch": 0.14806588932074774,
"eval_loss": 0.9855244755744934,
"eval_runtime": 7.1036,
"eval_samples_per_second": 18.019,
"eval_steps_per_second": 2.252,
"step": 100
},
{
"epoch": 0.15546918378678512,
"grad_norm": 0.1594521756746442,
"learning_rate": 1.981720334984174e-05,
"loss": 0.9583,
"step": 105
},
{
"epoch": 0.1628724782528225,
"grad_norm": 0.17121067681141633,
"learning_rate": 1.9764668444934853e-05,
"loss": 0.9474,
"step": 110
},
{
"epoch": 0.17027577271885988,
"grad_norm": 0.15907478498991337,
"learning_rate": 1.970559477641606e-05,
"loss": 0.9207,
"step": 115
},
{
"epoch": 0.1776790671848973,
"grad_norm": 0.1606853101811812,
"learning_rate": 1.9640021902080523e-05,
"loss": 0.9532,
"step": 120
},
{
"epoch": 0.18508236165093467,
"grad_norm": 0.14842172195462983,
"learning_rate": 1.9567993731818988e-05,
"loss": 0.9512,
"step": 125
},
{
"epoch": 0.19248565611697205,
"grad_norm": 0.16039228542123013,
"learning_rate": 1.9489558498214197e-05,
"loss": 0.9443,
"step": 130
},
{
"epoch": 0.19988895058300943,
"grad_norm": 0.15683712656065646,
"learning_rate": 1.9404768724242667e-05,
"loss": 0.9257,
"step": 135
},
{
"epoch": 0.20729224504904684,
"grad_norm": 0.17249812910241194,
"learning_rate": 1.931368118810346e-05,
"loss": 0.9545,
"step": 140
},
{
"epoch": 0.21469553951508422,
"grad_norm": 0.15206299782477828,
"learning_rate": 1.92163568851975e-05,
"loss": 0.938,
"step": 145
},
{
"epoch": 0.2220988339811216,
"grad_norm": 0.16815584502250794,
"learning_rate": 1.911286098728296e-05,
"loss": 0.9285,
"step": 150
},
{
"epoch": 0.22950212844715898,
"grad_norm": 0.16122796157665575,
"learning_rate": 1.900326279883392e-05,
"loss": 0.9624,
"step": 155
},
{
"epoch": 0.23690542291319638,
"grad_norm": 0.15224168391003395,
"learning_rate": 1.8887635710631716e-05,
"loss": 0.9662,
"step": 160
},
{
"epoch": 0.24430871737923376,
"grad_norm": 0.14815411902225328,
"learning_rate": 1.8766057150619865e-05,
"loss": 0.9346,
"step": 165
},
{
"epoch": 0.25171201184527114,
"grad_norm": 0.15183932254655913,
"learning_rate": 1.8638608532055635e-05,
"loss": 0.9533,
"step": 170
},
{
"epoch": 0.2591153063113085,
"grad_norm": 0.1446345977964626,
"learning_rate": 1.8505375198992856e-05,
"loss": 0.9395,
"step": 175
},
{
"epoch": 0.2665186007773459,
"grad_norm": 0.14315212915090392,
"learning_rate": 1.836644636913258e-05,
"loss": 0.9099,
"step": 180
},
{
"epoch": 0.2739218952433833,
"grad_norm": 0.16987297421517,
"learning_rate": 1.8221915074079764e-05,
"loss": 0.9342,
"step": 185
},
{
"epoch": 0.2813251897094207,
"grad_norm": 0.17161893073243528,
"learning_rate": 1.8071878097046064e-05,
"loss": 0.9115,
"step": 190
},
{
"epoch": 0.2887284841754581,
"grad_norm": 0.1544233807044902,
"learning_rate": 1.7916435908040413e-05,
"loss": 0.9309,
"step": 195
},
{
"epoch": 0.2961317786414955,
"grad_norm": 0.14748037798669633,
"learning_rate": 1.7755692596590778e-05,
"loss": 0.9333,
"step": 200
},
{
"epoch": 0.2961317786414955,
"eval_loss": 0.9494131803512573,
"eval_runtime": 7.1466,
"eval_samples_per_second": 17.911,
"eval_steps_per_second": 2.239,
"step": 200
},
{
"epoch": 0.30353507310753286,
"grad_norm": 0.1555651675700521,
"learning_rate": 1.7589755802042188e-05,
"loss": 0.9287,
"step": 205
},
{
"epoch": 0.31093836757357024,
"grad_norm": 0.16353203572144123,
"learning_rate": 1.7418736641477636e-05,
"loss": 0.9099,
"step": 210
},
{
"epoch": 0.3183416620396076,
"grad_norm": 0.15764904906951305,
"learning_rate": 1.7242749635310222e-05,
"loss": 0.913,
"step": 215
},
{
"epoch": 0.325744956505645,
"grad_norm": 0.1448977765793318,
"learning_rate": 1.7061912630596252e-05,
"loss": 0.9173,
"step": 220
},
{
"epoch": 0.3331482509716824,
"grad_norm": 0.15258137653598713,
"learning_rate": 1.6876346722120747e-05,
"loss": 0.9319,
"step": 225
},
{
"epoch": 0.34055154543771976,
"grad_norm": 0.1563094102159032,
"learning_rate": 1.6686176171308125e-05,
"loss": 0.955,
"step": 230
},
{
"epoch": 0.3479548399037572,
"grad_norm": 0.14423025166628695,
"learning_rate": 1.6491528323012412e-05,
"loss": 0.9132,
"step": 235
},
{
"epoch": 0.3553581343697946,
"grad_norm": 0.14263303346026474,
"learning_rate": 1.6292533520242663e-05,
"loss": 0.9157,
"step": 240
},
{
"epoch": 0.36276142883583196,
"grad_norm": 0.13731880620436607,
"learning_rate": 1.6089325016880737e-05,
"loss": 0.9058,
"step": 245
},
{
"epoch": 0.37016472330186934,
"grad_norm": 0.17337841623245412,
"learning_rate": 1.588203888844982e-05,
"loss": 0.9261,
"step": 250
},
{
"epoch": 0.3775680177679067,
"grad_norm": 0.1517393366410306,
"learning_rate": 1.5670813940993504e-05,
"loss": 0.8936,
"step": 255
},
{
"epoch": 0.3849713122339441,
"grad_norm": 0.14624943669912388,
"learning_rate": 1.5455791618126407e-05,
"loss": 0.8844,
"step": 260
},
{
"epoch": 0.3923746066999815,
"grad_norm": 0.14950642484342547,
"learning_rate": 1.5237115906318565e-05,
"loss": 0.9132,
"step": 265
},
{
"epoch": 0.39977790116601886,
"grad_norm": 0.15048654068633793,
"learning_rate": 1.5014933238477069e-05,
"loss": 0.8891,
"step": 270
},
{
"epoch": 0.4071811956320563,
"grad_norm": 0.1570568189193127,
"learning_rate": 1.4789392395889468e-05,
"loss": 0.9186,
"step": 275
},
{
"epoch": 0.41458449009809367,
"grad_norm": 0.15381562012177635,
"learning_rate": 1.4560644408594602e-05,
"loss": 0.8927,
"step": 280
},
{
"epoch": 0.42198778456413105,
"grad_norm": 0.1444417700490563,
"learning_rate": 1.432884245424761e-05,
"loss": 0.9081,
"step": 285
},
{
"epoch": 0.42939107903016843,
"grad_norm": 0.15807566475131207,
"learning_rate": 1.4094141755546816e-05,
"loss": 0.9389,
"step": 290
},
{
"epoch": 0.4367943734962058,
"grad_norm": 0.14968340220236578,
"learning_rate": 1.3856699476291176e-05,
"loss": 0.8926,
"step": 295
},
{
"epoch": 0.4441976679622432,
"grad_norm": 0.15781464037029244,
"learning_rate": 1.3616674616137902e-05,
"loss": 0.9161,
"step": 300
},
{
"epoch": 0.4441976679622432,
"eval_loss": 0.930830717086792,
"eval_runtime": 7.1062,
"eval_samples_per_second": 18.013,
"eval_steps_per_second": 2.252,
"step": 300
},
{
"epoch": 0.4516009624282806,
"grad_norm": 0.15525393114371358,
"learning_rate": 1.3374227904130724e-05,
"loss": 0.9071,
"step": 305
},
{
"epoch": 0.45900425689431795,
"grad_norm": 0.1457498963275538,
"learning_rate": 1.3129521691070108e-05,
"loss": 0.8833,
"step": 310
},
{
"epoch": 0.46640755136035533,
"grad_norm": 0.14987682585253773,
"learning_rate": 1.2882719840797473e-05,
"loss": 0.9022,
"step": 315
},
{
"epoch": 0.47381084582639277,
"grad_norm": 0.15509542057237247,
"learning_rate": 1.2633987620466229e-05,
"loss": 0.8885,
"step": 320
},
{
"epoch": 0.48121414029243015,
"grad_norm": 0.14584510044567442,
"learning_rate": 1.2383491589873122e-05,
"loss": 0.8899,
"step": 325
},
{
"epoch": 0.48861743475846753,
"grad_norm": 0.14309664992352822,
"learning_rate": 1.213139948992394e-05,
"loss": 0.8998,
"step": 330
},
{
"epoch": 0.4960207292245049,
"grad_norm": 0.15130038484765435,
"learning_rate": 1.187788013030837e-05,
"loss": 0.8988,
"step": 335
},
{
"epoch": 0.5034240236905423,
"grad_norm": 0.1459259790309168,
"learning_rate": 1.1623103276459086e-05,
"loss": 0.9007,
"step": 340
},
{
"epoch": 0.5108273181565797,
"grad_norm": 0.14777529707412382,
"learning_rate": 1.1367239535870913e-05,
"loss": 0.8925,
"step": 345
},
{
"epoch": 0.518230612622617,
"grad_norm": 0.14724048128355402,
"learning_rate": 1.1110460243856051e-05,
"loss": 0.8785,
"step": 350
},
{
"epoch": 0.5256339070886544,
"grad_norm": 0.14527394357147336,
"learning_rate": 1.085293734881197e-05,
"loss": 0.8982,
"step": 355
},
{
"epoch": 0.5330372015546918,
"grad_norm": 0.14050877787386462,
"learning_rate": 1.0594843297078736e-05,
"loss": 0.9181,
"step": 360
},
{
"epoch": 0.5404404960207292,
"grad_norm": 0.15865485844086924,
"learning_rate": 1.0336350917462925e-05,
"loss": 0.9062,
"step": 365
},
{
"epoch": 0.5478437904867666,
"grad_norm": 0.13521171802296975,
"learning_rate": 1.0077633305505402e-05,
"loss": 0.8927,
"step": 370
},
{
"epoch": 0.555247084952804,
"grad_norm": 0.13603561195542918,
"learning_rate": 9.818863707570476e-06,
"loss": 0.9154,
"step": 375
},
{
"epoch": 0.5626503794188414,
"grad_norm": 0.13965992628983345,
"learning_rate": 9.560215404834094e-06,
"loss": 0.9016,
"step": 380
},
{
"epoch": 0.5700536738848788,
"grad_norm": 0.13666000951700663,
"learning_rate": 9.30186159724869e-06,
"loss": 0.8612,
"step": 385
},
{
"epoch": 0.5774569683509162,
"grad_norm": 0.13625895333241347,
"learning_rate": 9.043975287562443e-06,
"loss": 0.9001,
"step": 390
},
{
"epoch": 0.5848602628169536,
"grad_norm": 0.14371829685671048,
"learning_rate": 8.786729165470584e-06,
"loss": 0.8738,
"step": 395
},
{
"epoch": 0.592263557282991,
"grad_norm": 0.13916699255941037,
"learning_rate": 8.530295491976338e-06,
"loss": 0.8804,
"step": 400
},
{
"epoch": 0.592263557282991,
"eval_loss": 0.9166682958602905,
"eval_runtime": 7.0621,
"eval_samples_per_second": 18.125,
"eval_steps_per_second": 2.266,
"step": 400
},
{
"epoch": 0.5996668517490283,
"grad_norm": 0.13004150614702822,
"learning_rate": 8.274845984038916e-06,
"loss": 0.8647,
"step": 405
},
{
"epoch": 0.6070701462150657,
"grad_norm": 0.14629481441132058,
"learning_rate": 8.020551699585843e-06,
"loss": 0.9007,
"step": 410
},
{
"epoch": 0.6144734406811031,
"grad_norm": 0.12877499400506406,
"learning_rate": 7.76758292296659e-06,
"loss": 0.8745,
"step": 415
},
{
"epoch": 0.6218767351471405,
"grad_norm": 0.13752238505053258,
"learning_rate": 7.5161090509242005e-06,
"loss": 0.8928,
"step": 420
},
{
"epoch": 0.6292800296131779,
"grad_norm": 0.14148995267833064,
"learning_rate": 7.2662984791613186e-06,
"loss": 0.8845,
"step": 425
},
{
"epoch": 0.6366833240792152,
"grad_norm": 0.14452849454657804,
"learning_rate": 7.01831848957653e-06,
"loss": 0.8991,
"step": 430
},
{
"epoch": 0.6440866185452526,
"grad_norm": 0.13297035448949285,
"learning_rate": 6.772335138246548e-06,
"loss": 0.8966,
"step": 435
},
{
"epoch": 0.65148991301129,
"grad_norm": 0.13999651657873824,
"learning_rate": 6.528513144229256e-06,
"loss": 0.8901,
"step": 440
},
{
"epoch": 0.6588932074773274,
"grad_norm": 0.13786802129264458,
"learning_rate": 6.287015779262064e-06,
"loss": 0.8988,
"step": 445
},
{
"epoch": 0.6662965019433648,
"grad_norm": 0.13362818972180024,
"learning_rate": 6.048004758429451e-06,
"loss": 0.8773,
"step": 450
},
{
"epoch": 0.6736997964094021,
"grad_norm": 0.14094460764095268,
"learning_rate": 5.811640131872867e-06,
"loss": 0.9022,
"step": 455
},
{
"epoch": 0.6811030908754395,
"grad_norm": 0.13177032254471754,
"learning_rate": 5.578080177615575e-06,
"loss": 0.8681,
"step": 460
},
{
"epoch": 0.688506385341477,
"grad_norm": 0.1447285259606947,
"learning_rate": 5.347481295574141e-06,
"loss": 0.8642,
"step": 465
},
{
"epoch": 0.6959096798075144,
"grad_norm": 0.16212399303569955,
"learning_rate": 5.119997902827584e-06,
"loss": 0.8768,
"step": 470
},
{
"epoch": 0.7033129742735518,
"grad_norm": 0.13407281956450473,
"learning_rate": 4.8957823302142916e-06,
"loss": 0.8861,
"step": 475
},
{
"epoch": 0.7107162687395892,
"grad_norm": 0.12431950401558224,
"learning_rate": 4.674984720325961e-06,
"loss": 0.8663,
"step": 480
},
{
"epoch": 0.7181195632056265,
"grad_norm": 0.13146748720425666,
"learning_rate": 4.457752926966888e-06,
"loss": 0.8618,
"step": 485
},
{
"epoch": 0.7255228576716639,
"grad_norm": 0.12584510121252945,
"learning_rate": 4.244232416145839e-06,
"loss": 0.8854,
"step": 490
},
{
"epoch": 0.7329261521377013,
"grad_norm": 0.12990134879461734,
"learning_rate": 4.0345661686669745e-06,
"loss": 0.8762,
"step": 495
},
{
"epoch": 0.7403294466037387,
"grad_norm": 0.13783047027951226,
"learning_rate": 3.828894584384867e-06,
"loss": 0.8859,
"step": 500
},
{
"epoch": 0.7403294466037387,
"eval_loss": 0.9078959226608276,
"eval_runtime": 7.1187,
"eval_samples_per_second": 17.981,
"eval_steps_per_second": 2.248,
"step": 500
},
{
"epoch": 0.747732741069776,
"grad_norm": 0.12586000345704224,
"learning_rate": 3.62735538818787e-06,
"loss": 0.8688,
"step": 505
},
{
"epoch": 0.7551360355358134,
"grad_norm": 0.14104016395887148,
"learning_rate": 3.4300835377726904e-06,
"loss": 0.8703,
"step": 510
},
{
"epoch": 0.7625393300018508,
"grad_norm": 0.13533570426883107,
"learning_rate": 3.2372111332720045e-06,
"loss": 0.9138,
"step": 515
},
{
"epoch": 0.7699426244678882,
"grad_norm": 0.14074880883541166,
"learning_rate": 3.048867328795588e-06,
"loss": 0.8618,
"step": 520
},
{
"epoch": 0.7773459189339256,
"grad_norm": 0.13230663319443067,
"learning_rate": 2.865178245944218e-06,
"loss": 0.8604,
"step": 525
},
{
"epoch": 0.784749213399963,
"grad_norm": 0.1247899466464526,
"learning_rate": 2.686266889354211e-06,
"loss": 0.8859,
"step": 530
},
{
"epoch": 0.7921525078660003,
"grad_norm": 0.13575024760681823,
"learning_rate": 2.5122530643292274e-06,
"loss": 0.8905,
"step": 535
},
{
"epoch": 0.7995558023320377,
"grad_norm": 0.1357493128741245,
"learning_rate": 2.3432532966144526e-06,
"loss": 0.8852,
"step": 540
},
{
"epoch": 0.8069590967980751,
"grad_norm": 0.13498792821706115,
"learning_rate": 2.1793807543668857e-06,
"loss": 0.8828,
"step": 545
},
{
"epoch": 0.8143623912641126,
"grad_norm": 0.12019174745605471,
"learning_rate": 2.0207451723739633e-06,
"loss": 0.8742,
"step": 550
},
{
"epoch": 0.82176568573015,
"grad_norm": 0.12150186530956911,
"learning_rate": 1.8674527785713247e-06,
"loss": 0.8802,
"step": 555
},
{
"epoch": 0.8291689801961873,
"grad_norm": 0.13657600204630863,
"learning_rate": 1.7196062229088606e-06,
"loss": 0.8458,
"step": 560
},
{
"epoch": 0.8365722746622247,
"grad_norm": 0.14316460964604913,
"learning_rate": 1.577304508612717e-06,
"loss": 0.8772,
"step": 565
},
{
"epoch": 0.8439755691282621,
"grad_norm": 0.11702464461690702,
"learning_rate": 1.4406429258892762e-06,
"loss": 0.8978,
"step": 570
},
{
"epoch": 0.8513788635942995,
"grad_norm": 0.13649799424857104,
"learning_rate": 1.3097129881154936e-06,
"loss": 0.8679,
"step": 575
},
{
"epoch": 0.8587821580603369,
"grad_norm": 0.12879197044894133,
"learning_rate": 1.1846023705583442e-06,
"loss": 0.8611,
"step": 580
},
{
"epoch": 0.8661854525263742,
"grad_norm": 0.12611530487166148,
"learning_rate": 1.065394851664394e-06,
"loss": 0.8847,
"step": 585
},
{
"epoch": 0.8735887469924116,
"grad_norm": 0.12199166350158228,
"learning_rate": 9.521702569588199e-07,
"loss": 0.9041,
"step": 590
},
{
"epoch": 0.880992041458449,
"grad_norm": 0.13250148039242704,
"learning_rate": 8.450044055914497e-07,
"loss": 0.8703,
"step": 595
},
{
"epoch": 0.8883953359244864,
"grad_norm": 0.12335176677515089,
"learning_rate": 7.439690595656013e-07,
"loss": 0.8931,
"step": 600
},
{
"epoch": 0.8883953359244864,
"eval_loss": 0.904110312461853,
"eval_runtime": 7.0453,
"eval_samples_per_second": 18.168,
"eval_steps_per_second": 2.271,
"step": 600
},
{
"epoch": 0.8957986303905238,
"grad_norm": 0.13803312908970108,
"learning_rate": 6.491318756837417e-07,
"loss": 0.8773,
"step": 605
},
{
"epoch": 0.9032019248565611,
"grad_norm": 0.11990945958525849,
"learning_rate": 5.605563602421149e-07,
"loss": 0.8506,
"step": 610
},
{
"epoch": 0.9106052193225985,
"grad_norm": 0.12252169940099521,
"learning_rate": 4.783018265047179e-07,
"loss": 0.9066,
"step": 615
},
{
"epoch": 0.9180085137886359,
"grad_norm": 0.12997936768176632,
"learning_rate": 4.024233549850509e-07,
"loss": 0.8924,
"step": 620
},
{
"epoch": 0.9254118082546733,
"grad_norm": 0.12018543528554308,
"learning_rate": 3.329717565622825e-07,
"loss": 0.8889,
"step": 625
},
{
"epoch": 0.9328151027207107,
"grad_norm": 0.12623553627779757,
"learning_rate": 2.6999353845651113e-07,
"loss": 0.8718,
"step": 630
},
{
"epoch": 0.9402183971867482,
"grad_norm": 0.12569602093691798,
"learning_rate": 2.1353087308590314e-07,
"loss": 0.8885,
"step": 635
},
{
"epoch": 0.9476216916527855,
"grad_norm": 0.12693697747520694,
"learning_rate": 1.6362156982656085e-07,
"loss": 0.8778,
"step": 640
},
{
"epoch": 0.9550249861188229,
"grad_norm": 0.11916388416111799,
"learning_rate": 1.2029904969404482e-07,
"loss": 0.8574,
"step": 645
},
{
"epoch": 0.9624282805848603,
"grad_norm": 0.13423693830897884,
"learning_rate": 8.359232296349163e-08,
"loss": 0.8633,
"step": 650
},
{
"epoch": 0.9698315750508977,
"grad_norm": 0.12941855976651295,
"learning_rate": 5.3525969743324356e-08,
"loss": 0.8779,
"step": 655
},
{
"epoch": 0.9772348695169351,
"grad_norm": 0.11997265846004904,
"learning_rate": 3.012012351554017e-08,
"loss": 0.8493,
"step": 660
},
{
"epoch": 0.9846381639829724,
"grad_norm": 0.13237634154671962,
"learning_rate": 1.3390457653639221e-08,
"loss": 0.8863,
"step": 665
},
{
"epoch": 0.9920414584490098,
"grad_norm": 0.13967439000954648,
"learning_rate": 3.3481749271768726e-09,
"loss": 0.8954,
"step": 670
},
{
"epoch": 0.9994447529150472,
"grad_norm": 0.12596999235093093,
"learning_rate": 0.0,
"loss": 0.861,
"step": 675
},
{
"epoch": 0.9994447529150472,
"step": 675,
"total_flos": 1.6624354892709888e+17,
"train_loss": 0.9221899901496039,
"train_runtime": 5284.4545,
"train_samples_per_second": 4.089,
"train_steps_per_second": 0.128
}
],
"logging_steps": 5,
"max_steps": 675,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": false,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 1.6624354892709888e+17,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}