Mistral-Codon-v1-1M / trainer_state.json
RaphaelMourad's picture
Upload 10 files
004dd08 verified
{
"best_metric": 3.7249155044555664,
"best_model_checkpoint": "./results/models/checkpoint-58650",
"epoch": 30.0,
"eval_steps": 500,
"global_step": 58650,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.2557544757033248,
"grad_norm": 0.047607421875,
"learning_rate": 0.003979539641943734,
"loss": 1.9283,
"step": 500
},
{
"epoch": 0.5115089514066496,
"grad_norm": 0.05810546875,
"learning_rate": 0.003959079283887468,
"loss": 1.89,
"step": 1000
},
{
"epoch": 0.7672634271099744,
"grad_norm": 0.048828125,
"learning_rate": 0.003938618925831202,
"loss": 1.8823,
"step": 1500
},
{
"epoch": 1.0,
"eval_loss": 3.7650158405303955,
"eval_runtime": 1.1856,
"eval_samples_per_second": 421.713,
"eval_steps_per_second": 0.843,
"step": 1955
},
{
"epoch": 1.0230179028132993,
"grad_norm": 0.048583984375,
"learning_rate": 0.003918158567774936,
"loss": 1.8791,
"step": 2000
},
{
"epoch": 1.278772378516624,
"grad_norm": 0.0458984375,
"learning_rate": 0.00389769820971867,
"loss": 1.8765,
"step": 2500
},
{
"epoch": 1.5345268542199488,
"grad_norm": 0.04443359375,
"learning_rate": 0.003877237851662404,
"loss": 1.8758,
"step": 3000
},
{
"epoch": 1.7902813299232738,
"grad_norm": 0.04638671875,
"learning_rate": 0.003856777493606138,
"loss": 1.874,
"step": 3500
},
{
"epoch": 2.0,
"eval_loss": 3.755614757537842,
"eval_runtime": 1.2689,
"eval_samples_per_second": 394.03,
"eval_steps_per_second": 0.788,
"step": 3910
},
{
"epoch": 2.0460358056265986,
"grad_norm": 0.0322265625,
"learning_rate": 0.0038363171355498722,
"loss": 1.8728,
"step": 4000
},
{
"epoch": 2.3017902813299234,
"grad_norm": 0.03857421875,
"learning_rate": 0.0038158567774936062,
"loss": 1.8721,
"step": 4500
},
{
"epoch": 2.557544757033248,
"grad_norm": 0.051025390625,
"learning_rate": 0.0037953964194373403,
"loss": 1.8714,
"step": 5000
},
{
"epoch": 2.813299232736573,
"grad_norm": 0.043701171875,
"learning_rate": 0.0037749360613810743,
"loss": 1.8704,
"step": 5500
},
{
"epoch": 3.0,
"eval_loss": 3.749542474746704,
"eval_runtime": 1.1996,
"eval_samples_per_second": 416.822,
"eval_steps_per_second": 0.834,
"step": 5865
},
{
"epoch": 3.0690537084398977,
"grad_norm": 0.045654296875,
"learning_rate": 0.0037544757033248083,
"loss": 1.8699,
"step": 6000
},
{
"epoch": 3.3248081841432224,
"grad_norm": 0.038818359375,
"learning_rate": 0.0037340153452685423,
"loss": 1.8695,
"step": 6500
},
{
"epoch": 3.580562659846547,
"grad_norm": 0.04541015625,
"learning_rate": 0.0037135549872122763,
"loss": 1.8688,
"step": 7000
},
{
"epoch": 3.836317135549872,
"grad_norm": 0.037109375,
"learning_rate": 0.0036930946291560103,
"loss": 1.8686,
"step": 7500
},
{
"epoch": 4.0,
"eval_loss": 3.7466166019439697,
"eval_runtime": 1.1726,
"eval_samples_per_second": 426.391,
"eval_steps_per_second": 0.853,
"step": 7820
},
{
"epoch": 4.092071611253197,
"grad_norm": 0.04833984375,
"learning_rate": 0.0036726342710997444,
"loss": 1.8681,
"step": 8000
},
{
"epoch": 4.3478260869565215,
"grad_norm": 0.047607421875,
"learning_rate": 0.003652173913043478,
"loss": 1.8678,
"step": 8500
},
{
"epoch": 4.603580562659847,
"grad_norm": 0.056640625,
"learning_rate": 0.0036317135549872124,
"loss": 1.8674,
"step": 9000
},
{
"epoch": 4.859335038363171,
"grad_norm": 0.0400390625,
"learning_rate": 0.0036112531969309464,
"loss": 1.867,
"step": 9500
},
{
"epoch": 5.0,
"eval_loss": 3.7444469928741455,
"eval_runtime": 1.1971,
"eval_samples_per_second": 417.668,
"eval_steps_per_second": 0.835,
"step": 9775
},
{
"epoch": 5.115089514066496,
"grad_norm": 0.04736328125,
"learning_rate": 0.0035907928388746804,
"loss": 1.8668,
"step": 10000
},
{
"epoch": 5.370843989769821,
"grad_norm": 0.03564453125,
"learning_rate": 0.0035703324808184144,
"loss": 1.8668,
"step": 10500
},
{
"epoch": 5.626598465473146,
"grad_norm": 0.047119140625,
"learning_rate": 0.003549872122762148,
"loss": 1.866,
"step": 11000
},
{
"epoch": 5.882352941176471,
"grad_norm": 0.04052734375,
"learning_rate": 0.0035294117647058825,
"loss": 1.8662,
"step": 11500
},
{
"epoch": 6.0,
"eval_loss": 3.742023229598999,
"eval_runtime": 1.1689,
"eval_samples_per_second": 427.744,
"eval_steps_per_second": 0.855,
"step": 11730
},
{
"epoch": 6.138107416879795,
"grad_norm": 0.047119140625,
"learning_rate": 0.0035089514066496165,
"loss": 1.8652,
"step": 12000
},
{
"epoch": 6.3938618925831205,
"grad_norm": 0.041259765625,
"learning_rate": 0.0034884910485933505,
"loss": 1.8656,
"step": 12500
},
{
"epoch": 6.649616368286445,
"grad_norm": 0.040283203125,
"learning_rate": 0.0034680306905370845,
"loss": 1.8655,
"step": 13000
},
{
"epoch": 6.90537084398977,
"grad_norm": 0.04736328125,
"learning_rate": 0.003447570332480818,
"loss": 1.8653,
"step": 13500
},
{
"epoch": 7.0,
"eval_loss": 3.7410786151885986,
"eval_runtime": 1.2028,
"eval_samples_per_second": 415.68,
"eval_steps_per_second": 0.831,
"step": 13685
},
{
"epoch": 7.161125319693094,
"grad_norm": 0.0439453125,
"learning_rate": 0.0034271099744245526,
"loss": 1.865,
"step": 14000
},
{
"epoch": 7.41687979539642,
"grad_norm": 0.04052734375,
"learning_rate": 0.0034066496163682866,
"loss": 1.8648,
"step": 14500
},
{
"epoch": 7.672634271099744,
"grad_norm": 0.060302734375,
"learning_rate": 0.0033861892583120206,
"loss": 1.865,
"step": 15000
},
{
"epoch": 7.928388746803069,
"grad_norm": 0.0400390625,
"learning_rate": 0.0033657289002557546,
"loss": 1.8641,
"step": 15500
},
{
"epoch": 8.0,
"eval_loss": 3.7378389835357666,
"eval_runtime": 1.1927,
"eval_samples_per_second": 419.217,
"eval_steps_per_second": 0.838,
"step": 15640
},
{
"epoch": 8.184143222506394,
"grad_norm": 0.04345703125,
"learning_rate": 0.003345268542199488,
"loss": 1.8644,
"step": 16000
},
{
"epoch": 8.43989769820972,
"grad_norm": 0.04736328125,
"learning_rate": 0.0033248081841432226,
"loss": 1.8638,
"step": 16500
},
{
"epoch": 8.695652173913043,
"grad_norm": 0.05126953125,
"learning_rate": 0.0033043478260869567,
"loss": 1.8639,
"step": 17000
},
{
"epoch": 8.951406649616368,
"grad_norm": 0.05224609375,
"learning_rate": 0.0032838874680306907,
"loss": 1.8641,
"step": 17500
},
{
"epoch": 9.0,
"eval_loss": 3.7366719245910645,
"eval_runtime": 1.2549,
"eval_samples_per_second": 398.446,
"eval_steps_per_second": 0.797,
"step": 17595
},
{
"epoch": 9.207161125319693,
"grad_norm": 0.06396484375,
"learning_rate": 0.0032634271099744247,
"loss": 1.8636,
"step": 18000
},
{
"epoch": 9.462915601023019,
"grad_norm": 0.03857421875,
"learning_rate": 0.0032429667519181583,
"loss": 1.8655,
"step": 18500
},
{
"epoch": 9.718670076726342,
"grad_norm": 0.038818359375,
"learning_rate": 0.0032225063938618927,
"loss": 1.8655,
"step": 19000
},
{
"epoch": 9.974424552429667,
"grad_norm": 0.04150390625,
"learning_rate": 0.0032020460358056268,
"loss": 1.8646,
"step": 19500
},
{
"epoch": 10.0,
"eval_loss": 3.7368271350860596,
"eval_runtime": 1.1654,
"eval_samples_per_second": 429.02,
"eval_steps_per_second": 0.858,
"step": 19550
},
{
"epoch": 10.230179028132993,
"grad_norm": 0.051513671875,
"learning_rate": 0.0031815856777493608,
"loss": 1.8637,
"step": 20000
},
{
"epoch": 10.485933503836318,
"grad_norm": 0.048095703125,
"learning_rate": 0.003161125319693095,
"loss": 1.8641,
"step": 20500
},
{
"epoch": 10.741687979539641,
"grad_norm": 0.037841796875,
"learning_rate": 0.0031406649616368284,
"loss": 1.8633,
"step": 21000
},
{
"epoch": 10.997442455242966,
"grad_norm": 0.04931640625,
"learning_rate": 0.003120204603580563,
"loss": 1.8633,
"step": 21500
},
{
"epoch": 11.0,
"eval_loss": 3.736281156539917,
"eval_runtime": 1.3206,
"eval_samples_per_second": 378.604,
"eval_steps_per_second": 0.757,
"step": 21505
},
{
"epoch": 11.253196930946292,
"grad_norm": 0.04833984375,
"learning_rate": 0.003099744245524297,
"loss": 1.8636,
"step": 22000
},
{
"epoch": 11.508951406649617,
"grad_norm": 0.0478515625,
"learning_rate": 0.003079283887468031,
"loss": 1.863,
"step": 22500
},
{
"epoch": 11.764705882352942,
"grad_norm": 0.048583984375,
"learning_rate": 0.0030588235294117644,
"loss": 1.8629,
"step": 23000
},
{
"epoch": 12.0,
"eval_loss": 3.734565496444702,
"eval_runtime": 1.1835,
"eval_samples_per_second": 422.479,
"eval_steps_per_second": 0.845,
"step": 23460
},
{
"epoch": 12.020460358056265,
"grad_norm": 0.044189453125,
"learning_rate": 0.0030383631713554985,
"loss": 1.8625,
"step": 23500
},
{
"epoch": 12.27621483375959,
"grad_norm": 0.048828125,
"learning_rate": 0.003017902813299233,
"loss": 1.8629,
"step": 24000
},
{
"epoch": 12.531969309462916,
"grad_norm": 0.04931640625,
"learning_rate": 0.002997442455242967,
"loss": 1.8624,
"step": 24500
},
{
"epoch": 12.787723785166241,
"grad_norm": 0.04345703125,
"learning_rate": 0.002976982097186701,
"loss": 1.8621,
"step": 25000
},
{
"epoch": 13.0,
"eval_loss": 3.7329983711242676,
"eval_runtime": 1.3627,
"eval_samples_per_second": 366.916,
"eval_steps_per_second": 0.734,
"step": 25415
},
{
"epoch": 13.043478260869565,
"grad_norm": 0.0546875,
"learning_rate": 0.0029565217391304345,
"loss": 1.8624,
"step": 25500
},
{
"epoch": 13.29923273657289,
"grad_norm": 0.048095703125,
"learning_rate": 0.002936061381074169,
"loss": 1.8624,
"step": 26000
},
{
"epoch": 13.554987212276215,
"grad_norm": 0.05517578125,
"learning_rate": 0.002915601023017903,
"loss": 1.8618,
"step": 26500
},
{
"epoch": 13.81074168797954,
"grad_norm": 0.04443359375,
"learning_rate": 0.002895140664961637,
"loss": 1.8619,
"step": 27000
},
{
"epoch": 14.0,
"eval_loss": 3.7325527667999268,
"eval_runtime": 1.1565,
"eval_samples_per_second": 432.347,
"eval_steps_per_second": 0.865,
"step": 27370
},
{
"epoch": 14.066496163682864,
"grad_norm": 0.043212890625,
"learning_rate": 0.002874680306905371,
"loss": 1.862,
"step": 27500
},
{
"epoch": 14.322250639386189,
"grad_norm": 0.06640625,
"learning_rate": 0.0028542199488491046,
"loss": 1.8616,
"step": 28000
},
{
"epoch": 14.578005115089514,
"grad_norm": 0.050048828125,
"learning_rate": 0.002833759590792839,
"loss": 1.8615,
"step": 28500
},
{
"epoch": 14.83375959079284,
"grad_norm": 0.046875,
"learning_rate": 0.002813299232736573,
"loss": 1.8623,
"step": 29000
},
{
"epoch": 15.0,
"eval_loss": 3.729966402053833,
"eval_runtime": 1.1923,
"eval_samples_per_second": 419.349,
"eval_steps_per_second": 0.839,
"step": 29325
},
{
"epoch": 15.089514066496164,
"grad_norm": 0.0458984375,
"learning_rate": 0.002792838874680307,
"loss": 1.8615,
"step": 29500
},
{
"epoch": 15.345268542199488,
"grad_norm": 0.042236328125,
"learning_rate": 0.002772378516624041,
"loss": 1.8616,
"step": 30000
},
{
"epoch": 15.601023017902813,
"grad_norm": 0.037353515625,
"learning_rate": 0.0027519181585677747,
"loss": 1.8618,
"step": 30500
},
{
"epoch": 15.856777493606138,
"grad_norm": 0.06591796875,
"learning_rate": 0.002731457800511509,
"loss": 1.8613,
"step": 31000
},
{
"epoch": 16.0,
"eval_loss": 3.7326087951660156,
"eval_runtime": 1.1839,
"eval_samples_per_second": 422.34,
"eval_steps_per_second": 0.845,
"step": 31280
},
{
"epoch": 16.11253196930946,
"grad_norm": 0.038818359375,
"learning_rate": 0.002710997442455243,
"loss": 1.8615,
"step": 31500
},
{
"epoch": 16.36828644501279,
"grad_norm": 0.0458984375,
"learning_rate": 0.002690537084398977,
"loss": 1.8619,
"step": 32000
},
{
"epoch": 16.624040920716112,
"grad_norm": 0.044677734375,
"learning_rate": 0.002670076726342711,
"loss": 1.8616,
"step": 32500
},
{
"epoch": 16.87979539641944,
"grad_norm": 0.04541015625,
"learning_rate": 0.0026496163682864448,
"loss": 1.8616,
"step": 33000
},
{
"epoch": 17.0,
"eval_loss": 3.729959011077881,
"eval_runtime": 1.2063,
"eval_samples_per_second": 414.478,
"eval_steps_per_second": 0.829,
"step": 33235
},
{
"epoch": 17.135549872122763,
"grad_norm": 0.04541015625,
"learning_rate": 0.0026291560102301792,
"loss": 1.8616,
"step": 33500
},
{
"epoch": 17.391304347826086,
"grad_norm": 0.048583984375,
"learning_rate": 0.0026086956521739132,
"loss": 1.8608,
"step": 34000
},
{
"epoch": 17.647058823529413,
"grad_norm": 0.047119140625,
"learning_rate": 0.0025882352941176473,
"loss": 1.8609,
"step": 34500
},
{
"epoch": 17.902813299232736,
"grad_norm": 0.041259765625,
"learning_rate": 0.002567774936061381,
"loss": 1.8606,
"step": 35000
},
{
"epoch": 18.0,
"eval_loss": 3.7307207584381104,
"eval_runtime": 1.1724,
"eval_samples_per_second": 426.468,
"eval_steps_per_second": 0.853,
"step": 35190
},
{
"epoch": 18.15856777493606,
"grad_norm": 0.04052734375,
"learning_rate": 0.002547314578005115,
"loss": 1.8613,
"step": 35500
},
{
"epoch": 18.414322250639387,
"grad_norm": 0.042236328125,
"learning_rate": 0.0025268542199488493,
"loss": 1.8608,
"step": 36000
},
{
"epoch": 18.67007672634271,
"grad_norm": 0.046142578125,
"learning_rate": 0.0025063938618925833,
"loss": 1.8607,
"step": 36500
},
{
"epoch": 18.925831202046037,
"grad_norm": 0.03955078125,
"learning_rate": 0.0024859335038363174,
"loss": 1.8606,
"step": 37000
},
{
"epoch": 19.0,
"eval_loss": 3.729092836380005,
"eval_runtime": 1.1564,
"eval_samples_per_second": 432.369,
"eval_steps_per_second": 0.865,
"step": 37145
},
{
"epoch": 19.18158567774936,
"grad_norm": 0.041259765625,
"learning_rate": 0.002465473145780051,
"loss": 1.8604,
"step": 37500
},
{
"epoch": 19.437340153452684,
"grad_norm": 0.047607421875,
"learning_rate": 0.002445012787723785,
"loss": 1.8607,
"step": 38000
},
{
"epoch": 19.69309462915601,
"grad_norm": 0.056396484375,
"learning_rate": 0.0024245524296675194,
"loss": 1.8605,
"step": 38500
},
{
"epoch": 19.948849104859335,
"grad_norm": 0.05517578125,
"learning_rate": 0.0024040920716112534,
"loss": 1.8606,
"step": 39000
},
{
"epoch": 20.0,
"eval_loss": 3.727680206298828,
"eval_runtime": 1.1607,
"eval_samples_per_second": 430.785,
"eval_steps_per_second": 0.862,
"step": 39100
},
{
"epoch": 20.20460358056266,
"grad_norm": 0.043212890625,
"learning_rate": 0.0023836317135549874,
"loss": 1.8605,
"step": 39500
},
{
"epoch": 20.460358056265985,
"grad_norm": 0.04345703125,
"learning_rate": 0.002363171355498721,
"loss": 1.8605,
"step": 40000
},
{
"epoch": 20.71611253196931,
"grad_norm": 0.037841796875,
"learning_rate": 0.002342710997442455,
"loss": 1.8606,
"step": 40500
},
{
"epoch": 20.971867007672635,
"grad_norm": 0.051513671875,
"learning_rate": 0.0023222506393861895,
"loss": 1.8601,
"step": 41000
},
{
"epoch": 21.0,
"eval_loss": 3.7270610332489014,
"eval_runtime": 1.1618,
"eval_samples_per_second": 430.348,
"eval_steps_per_second": 0.861,
"step": 41055
},
{
"epoch": 21.22762148337596,
"grad_norm": 0.05126953125,
"learning_rate": 0.0023017902813299235,
"loss": 1.8601,
"step": 41500
},
{
"epoch": 21.483375959079282,
"grad_norm": 0.056884765625,
"learning_rate": 0.0022813299232736575,
"loss": 1.8601,
"step": 42000
},
{
"epoch": 21.73913043478261,
"grad_norm": 0.045166015625,
"learning_rate": 0.002260869565217391,
"loss": 1.8602,
"step": 42500
},
{
"epoch": 21.994884910485933,
"grad_norm": 0.036376953125,
"learning_rate": 0.002240409207161125,
"loss": 1.86,
"step": 43000
},
{
"epoch": 22.0,
"eval_loss": 3.72686505317688,
"eval_runtime": 1.2281,
"eval_samples_per_second": 407.144,
"eval_steps_per_second": 0.814,
"step": 43010
},
{
"epoch": 22.25063938618926,
"grad_norm": 0.04931640625,
"learning_rate": 0.0022199488491048596,
"loss": 1.86,
"step": 43500
},
{
"epoch": 22.506393861892583,
"grad_norm": 0.0380859375,
"learning_rate": 0.0021994884910485936,
"loss": 1.8596,
"step": 44000
},
{
"epoch": 22.762148337595907,
"grad_norm": 0.039306640625,
"learning_rate": 0.0021790281329923276,
"loss": 1.8602,
"step": 44500
},
{
"epoch": 23.0,
"eval_loss": 3.7282092571258545,
"eval_runtime": 1.1696,
"eval_samples_per_second": 427.495,
"eval_steps_per_second": 0.855,
"step": 44965
},
{
"epoch": 23.017902813299234,
"grad_norm": 0.046142578125,
"learning_rate": 0.002158567774936061,
"loss": 1.86,
"step": 45000
},
{
"epoch": 23.273657289002557,
"grad_norm": 0.039306640625,
"learning_rate": 0.002138107416879795,
"loss": 1.8598,
"step": 45500
},
{
"epoch": 23.529411764705884,
"grad_norm": 0.04296875,
"learning_rate": 0.0021176470588235297,
"loss": 1.86,
"step": 46000
},
{
"epoch": 23.785166240409207,
"grad_norm": 0.054931640625,
"learning_rate": 0.0020971867007672637,
"loss": 1.8596,
"step": 46500
},
{
"epoch": 24.0,
"eval_loss": 3.7264657020568848,
"eval_runtime": 1.1709,
"eval_samples_per_second": 427.022,
"eval_steps_per_second": 0.854,
"step": 46920
},
{
"epoch": 24.04092071611253,
"grad_norm": 0.0419921875,
"learning_rate": 0.0020767263427109977,
"loss": 1.8599,
"step": 47000
},
{
"epoch": 24.296675191815858,
"grad_norm": 0.043701171875,
"learning_rate": 0.0020562659846547313,
"loss": 1.8603,
"step": 47500
},
{
"epoch": 24.55242966751918,
"grad_norm": 0.06884765625,
"learning_rate": 0.0020358056265984653,
"loss": 1.8592,
"step": 48000
},
{
"epoch": 24.808184143222505,
"grad_norm": 0.037353515625,
"learning_rate": 0.0020153452685421997,
"loss": 1.8596,
"step": 48500
},
{
"epoch": 25.0,
"eval_loss": 3.7268757820129395,
"eval_runtime": 1.2077,
"eval_samples_per_second": 414.025,
"eval_steps_per_second": 0.828,
"step": 48875
},
{
"epoch": 25.06393861892583,
"grad_norm": 0.0498046875,
"learning_rate": 0.0019948849104859333,
"loss": 1.8595,
"step": 49000
},
{
"epoch": 25.319693094629155,
"grad_norm": 0.0419921875,
"learning_rate": 0.0019744245524296678,
"loss": 1.8592,
"step": 49500
},
{
"epoch": 25.575447570332482,
"grad_norm": 0.0478515625,
"learning_rate": 0.001953964194373402,
"loss": 1.8595,
"step": 50000
},
{
"epoch": 25.831202046035806,
"grad_norm": 0.04052734375,
"learning_rate": 0.0019335038363171356,
"loss": 1.8598,
"step": 50500
},
{
"epoch": 26.0,
"eval_loss": 3.7266719341278076,
"eval_runtime": 1.1718,
"eval_samples_per_second": 426.711,
"eval_steps_per_second": 0.853,
"step": 50830
},
{
"epoch": 26.08695652173913,
"grad_norm": 0.046142578125,
"learning_rate": 0.0019130434782608696,
"loss": 1.8595,
"step": 51000
},
{
"epoch": 26.342710997442456,
"grad_norm": 0.043701171875,
"learning_rate": 0.0018925831202046036,
"loss": 1.8595,
"step": 51500
},
{
"epoch": 26.59846547314578,
"grad_norm": 0.03369140625,
"learning_rate": 0.0018721227621483376,
"loss": 1.8593,
"step": 52000
},
{
"epoch": 26.854219948849106,
"grad_norm": 0.03662109375,
"learning_rate": 0.0018516624040920714,
"loss": 1.8595,
"step": 52500
},
{
"epoch": 27.0,
"eval_loss": 3.72537899017334,
"eval_runtime": 1.2057,
"eval_samples_per_second": 414.687,
"eval_steps_per_second": 0.829,
"step": 52785
},
{
"epoch": 27.10997442455243,
"grad_norm": 0.044921875,
"learning_rate": 0.0018312020460358057,
"loss": 1.8596,
"step": 53000
},
{
"epoch": 27.365728900255753,
"grad_norm": 0.04248046875,
"learning_rate": 0.0018107416879795397,
"loss": 1.8592,
"step": 53500
},
{
"epoch": 27.62148337595908,
"grad_norm": 0.04833984375,
"learning_rate": 0.0017902813299232737,
"loss": 1.8598,
"step": 54000
},
{
"epoch": 27.877237851662404,
"grad_norm": 0.0400390625,
"learning_rate": 0.0017698209718670077,
"loss": 1.8592,
"step": 54500
},
{
"epoch": 28.0,
"eval_loss": 3.7256040573120117,
"eval_runtime": 1.1731,
"eval_samples_per_second": 426.237,
"eval_steps_per_second": 0.852,
"step": 54740
},
{
"epoch": 28.132992327365727,
"grad_norm": 0.045654296875,
"learning_rate": 0.0017493606138107415,
"loss": 1.8593,
"step": 55000
},
{
"epoch": 28.388746803069054,
"grad_norm": 0.0439453125,
"learning_rate": 0.0017289002557544758,
"loss": 1.8589,
"step": 55500
},
{
"epoch": 28.644501278772378,
"grad_norm": 0.047607421875,
"learning_rate": 0.0017084398976982098,
"loss": 1.8591,
"step": 56000
},
{
"epoch": 28.900255754475705,
"grad_norm": 0.05029296875,
"learning_rate": 0.0016879795396419438,
"loss": 1.8592,
"step": 56500
},
{
"epoch": 29.0,
"eval_loss": 3.7257955074310303,
"eval_runtime": 1.1794,
"eval_samples_per_second": 423.928,
"eval_steps_per_second": 0.848,
"step": 56695
},
{
"epoch": 29.156010230179028,
"grad_norm": 0.045166015625,
"learning_rate": 0.0016675191815856778,
"loss": 1.8589,
"step": 57000
},
{
"epoch": 29.41176470588235,
"grad_norm": 0.037109375,
"learning_rate": 0.0016470588235294116,
"loss": 1.8591,
"step": 57500
},
{
"epoch": 29.66751918158568,
"grad_norm": 0.0390625,
"learning_rate": 0.0016265984654731459,
"loss": 1.8591,
"step": 58000
},
{
"epoch": 29.923273657289002,
"grad_norm": 0.052734375,
"learning_rate": 0.0016061381074168799,
"loss": 1.8596,
"step": 58500
},
{
"epoch": 30.0,
"eval_loss": 3.7249155044555664,
"eval_runtime": 1.1467,
"eval_samples_per_second": 436.021,
"eval_steps_per_second": 0.872,
"step": 58650
}
],
"logging_steps": 500,
"max_steps": 97750,
"num_input_tokens_seen": 0,
"num_train_epochs": 50,
"save_steps": 500,
"stateful_callbacks": {
"EarlyStoppingCallback": {
"args": {
"early_stopping_patience": 3,
"early_stopping_threshold": 0.0
},
"attributes": {
"early_stopping_patience_counter": 0
}
},
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 8.390099826432e+16,
"train_batch_size": 512,
"trial_name": null,
"trial_params": null
}