DebertaV2-Base-10M_babylm-A__qqp / trainer_state.json
Ar4l's picture
Upload folder using huggingface_hub
d6f2afb verified
{
"best_metric": 0.8391068037456617,
"best_model_checkpoint": "/home/ubuntu/utah/babylm-24/src/evaluation/results/finetune/DebertaV2-Base-10M_babylm-A/qqp/checkpoint-227405",
"epoch": 5.0,
"eval_steps": 500,
"global_step": 227405,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.01099360172379675,
"grad_norm": 4.236701011657715,
"learning_rate": 2.993403838965722e-05,
"loss": 0.6086,
"step": 500
},
{
"epoch": 0.0219872034475935,
"grad_norm": 4.105135440826416,
"learning_rate": 2.986807677931444e-05,
"loss": 0.5618,
"step": 1000
},
{
"epoch": 0.03298080517139025,
"grad_norm": 5.613928318023682,
"learning_rate": 2.980211516897166e-05,
"loss": 0.5252,
"step": 1500
},
{
"epoch": 0.043974406895187,
"grad_norm": 12.754159927368164,
"learning_rate": 2.973615355862888e-05,
"loss": 0.5182,
"step": 2000
},
{
"epoch": 0.05496800861898375,
"grad_norm": 5.488389015197754,
"learning_rate": 2.9670191948286098e-05,
"loss": 0.505,
"step": 2500
},
{
"epoch": 0.0659616103427805,
"grad_norm": 5.46230936050415,
"learning_rate": 2.9604230337943316e-05,
"loss": 0.5044,
"step": 3000
},
{
"epoch": 0.07695521206657725,
"grad_norm": 13.249972343444824,
"learning_rate": 2.9538268727600537e-05,
"loss": 0.498,
"step": 3500
},
{
"epoch": 0.087948813790374,
"grad_norm": 13.043343544006348,
"learning_rate": 2.9472307117257757e-05,
"loss": 0.479,
"step": 4000
},
{
"epoch": 0.09894241551417075,
"grad_norm": 14.147295951843262,
"learning_rate": 2.9406345506914978e-05,
"loss": 0.4692,
"step": 4500
},
{
"epoch": 0.1099360172379675,
"grad_norm": 8.739255905151367,
"learning_rate": 2.9340383896572196e-05,
"loss": 0.4703,
"step": 5000
},
{
"epoch": 0.12092961896176425,
"grad_norm": 11.371148109436035,
"learning_rate": 2.9274422286229417e-05,
"loss": 0.4557,
"step": 5500
},
{
"epoch": 0.131923220685561,
"grad_norm": 8.44424057006836,
"learning_rate": 2.9208460675886637e-05,
"loss": 0.4635,
"step": 6000
},
{
"epoch": 0.14291682240935774,
"grad_norm": 3.9929113388061523,
"learning_rate": 2.914249906554385e-05,
"loss": 0.4651,
"step": 6500
},
{
"epoch": 0.1539104241331545,
"grad_norm": 3.4015634059906006,
"learning_rate": 2.9076537455201072e-05,
"loss": 0.4537,
"step": 7000
},
{
"epoch": 0.16490402585695124,
"grad_norm": 4.32379674911499,
"learning_rate": 2.9010575844858293e-05,
"loss": 0.4445,
"step": 7500
},
{
"epoch": 0.175897627580748,
"grad_norm": 5.210988521575928,
"learning_rate": 2.8944614234515514e-05,
"loss": 0.4326,
"step": 8000
},
{
"epoch": 0.18689122930454474,
"grad_norm": 3.908278226852417,
"learning_rate": 2.887865262417273e-05,
"loss": 0.4454,
"step": 8500
},
{
"epoch": 0.1978848310283415,
"grad_norm": 2.8256139755249023,
"learning_rate": 2.8812691013829952e-05,
"loss": 0.438,
"step": 9000
},
{
"epoch": 0.20887843275213824,
"grad_norm": 7.012328624725342,
"learning_rate": 2.8746729403487173e-05,
"loss": 0.4333,
"step": 9500
},
{
"epoch": 0.219872034475935,
"grad_norm": 5.030713081359863,
"learning_rate": 2.8680767793144394e-05,
"loss": 0.4588,
"step": 10000
},
{
"epoch": 0.23086563619973174,
"grad_norm": 10.047335624694824,
"learning_rate": 2.8614806182801608e-05,
"loss": 0.4418,
"step": 10500
},
{
"epoch": 0.2418592379235285,
"grad_norm": 5.562270164489746,
"learning_rate": 2.854884457245883e-05,
"loss": 0.4375,
"step": 11000
},
{
"epoch": 0.25285283964732524,
"grad_norm": 14.433082580566406,
"learning_rate": 2.848288296211605e-05,
"loss": 0.4278,
"step": 11500
},
{
"epoch": 0.263846441371122,
"grad_norm": 3.507910966873169,
"learning_rate": 2.8416921351773267e-05,
"loss": 0.4331,
"step": 12000
},
{
"epoch": 0.2748400430949188,
"grad_norm": 13.184822082519531,
"learning_rate": 2.8350959741430488e-05,
"loss": 0.4325,
"step": 12500
},
{
"epoch": 0.2858336448187155,
"grad_norm": 11.627528190612793,
"learning_rate": 2.828499813108771e-05,
"loss": 0.4283,
"step": 13000
},
{
"epoch": 0.29682724654251225,
"grad_norm": 4.054474830627441,
"learning_rate": 2.821903652074493e-05,
"loss": 0.4361,
"step": 13500
},
{
"epoch": 0.307820848266309,
"grad_norm": 7.193812847137451,
"learning_rate": 2.8153074910402147e-05,
"loss": 0.4251,
"step": 14000
},
{
"epoch": 0.3188144499901058,
"grad_norm": 4.036972999572754,
"learning_rate": 2.8087113300059365e-05,
"loss": 0.4233,
"step": 14500
},
{
"epoch": 0.3298080517139025,
"grad_norm": 6.451129913330078,
"learning_rate": 2.8021151689716586e-05,
"loss": 0.4407,
"step": 15000
},
{
"epoch": 0.34080165343769925,
"grad_norm": 6.5472612380981445,
"learning_rate": 2.7955190079373803e-05,
"loss": 0.4158,
"step": 15500
},
{
"epoch": 0.351795255161496,
"grad_norm": 16.589092254638672,
"learning_rate": 2.7889228469031024e-05,
"loss": 0.4261,
"step": 16000
},
{
"epoch": 0.3627888568852928,
"grad_norm": 6.696326732635498,
"learning_rate": 2.7823266858688245e-05,
"loss": 0.4111,
"step": 16500
},
{
"epoch": 0.3737824586090895,
"grad_norm": 4.396112442016602,
"learning_rate": 2.7757305248345466e-05,
"loss": 0.4236,
"step": 17000
},
{
"epoch": 0.38477606033288625,
"grad_norm": 7.1260986328125,
"learning_rate": 2.7691343638002683e-05,
"loss": 0.413,
"step": 17500
},
{
"epoch": 0.395769662056683,
"grad_norm": 8.553855895996094,
"learning_rate": 2.7625382027659904e-05,
"loss": 0.4105,
"step": 18000
},
{
"epoch": 0.4067632637804798,
"grad_norm": 1.411335825920105,
"learning_rate": 2.755942041731712e-05,
"loss": 0.3977,
"step": 18500
},
{
"epoch": 0.4177568655042765,
"grad_norm": 12.823638916015625,
"learning_rate": 2.7493458806974342e-05,
"loss": 0.4143,
"step": 19000
},
{
"epoch": 0.42875046722807325,
"grad_norm": 11.823991775512695,
"learning_rate": 2.742749719663156e-05,
"loss": 0.4002,
"step": 19500
},
{
"epoch": 0.43974406895187,
"grad_norm": 5.152065753936768,
"learning_rate": 2.736153558628878e-05,
"loss": 0.4001,
"step": 20000
},
{
"epoch": 0.4507376706756668,
"grad_norm": 2.0797653198242188,
"learning_rate": 2.7295573975946e-05,
"loss": 0.4137,
"step": 20500
},
{
"epoch": 0.4617312723994635,
"grad_norm": 5.874008655548096,
"learning_rate": 2.722961236560322e-05,
"loss": 0.4166,
"step": 21000
},
{
"epoch": 0.47272487412326025,
"grad_norm": 20.658824920654297,
"learning_rate": 2.716365075526044e-05,
"loss": 0.4072,
"step": 21500
},
{
"epoch": 0.483718475847057,
"grad_norm": 11.262660026550293,
"learning_rate": 2.709768914491766e-05,
"loss": 0.4048,
"step": 22000
},
{
"epoch": 0.4947120775708538,
"grad_norm": 8.16556167602539,
"learning_rate": 2.7031727534574878e-05,
"loss": 0.4059,
"step": 22500
},
{
"epoch": 0.5057056792946505,
"grad_norm": 12.176790237426758,
"learning_rate": 2.6965765924232095e-05,
"loss": 0.415,
"step": 23000
},
{
"epoch": 0.5166992810184473,
"grad_norm": 8.630789756774902,
"learning_rate": 2.6899804313889316e-05,
"loss": 0.4059,
"step": 23500
},
{
"epoch": 0.527692882742244,
"grad_norm": 11.014044761657715,
"learning_rate": 2.6833842703546537e-05,
"loss": 0.3804,
"step": 24000
},
{
"epoch": 0.5386864844660407,
"grad_norm": 12.287202835083008,
"learning_rate": 2.6767881093203758e-05,
"loss": 0.4046,
"step": 24500
},
{
"epoch": 0.5496800861898375,
"grad_norm": 6.118470668792725,
"learning_rate": 2.6701919482860975e-05,
"loss": 0.398,
"step": 25000
},
{
"epoch": 0.5606736879136343,
"grad_norm": 17.233190536499023,
"learning_rate": 2.6635957872518196e-05,
"loss": 0.4013,
"step": 25500
},
{
"epoch": 0.571667289637431,
"grad_norm": 15.902141571044922,
"learning_rate": 2.6569996262175417e-05,
"loss": 0.4158,
"step": 26000
},
{
"epoch": 0.5826608913612278,
"grad_norm": 6.975952625274658,
"learning_rate": 2.650403465183263e-05,
"loss": 0.3939,
"step": 26500
},
{
"epoch": 0.5936544930850245,
"grad_norm": 19.319835662841797,
"learning_rate": 2.6438073041489852e-05,
"loss": 0.4021,
"step": 27000
},
{
"epoch": 0.6046480948088213,
"grad_norm": 3.9395010471343994,
"learning_rate": 2.6372111431147073e-05,
"loss": 0.4015,
"step": 27500
},
{
"epoch": 0.615641696532618,
"grad_norm": 7.098001956939697,
"learning_rate": 2.6306149820804294e-05,
"loss": 0.3876,
"step": 28000
},
{
"epoch": 0.6266352982564147,
"grad_norm": 3.967722177505493,
"learning_rate": 2.624018821046151e-05,
"loss": 0.3935,
"step": 28500
},
{
"epoch": 0.6376288999802115,
"grad_norm": 5.257786273956299,
"learning_rate": 2.6174226600118732e-05,
"loss": 0.3959,
"step": 29000
},
{
"epoch": 0.6486225017040083,
"grad_norm": 17.10645294189453,
"learning_rate": 2.6108264989775953e-05,
"loss": 0.3948,
"step": 29500
},
{
"epoch": 0.659616103427805,
"grad_norm": 14.60950756072998,
"learning_rate": 2.6042303379433174e-05,
"loss": 0.4078,
"step": 30000
},
{
"epoch": 0.6706097051516018,
"grad_norm": 6.0776166915893555,
"learning_rate": 2.5976341769090388e-05,
"loss": 0.3831,
"step": 30500
},
{
"epoch": 0.6816033068753985,
"grad_norm": 1.830690622329712,
"learning_rate": 2.591038015874761e-05,
"loss": 0.3911,
"step": 31000
},
{
"epoch": 0.6925969085991953,
"grad_norm": 16.717496871948242,
"learning_rate": 2.584441854840483e-05,
"loss": 0.3926,
"step": 31500
},
{
"epoch": 0.703590510322992,
"grad_norm": 4.424517631530762,
"learning_rate": 2.5778456938062047e-05,
"loss": 0.3931,
"step": 32000
},
{
"epoch": 0.7145841120467887,
"grad_norm": 6.995429992675781,
"learning_rate": 2.5712495327719268e-05,
"loss": 0.3912,
"step": 32500
},
{
"epoch": 0.7255777137705856,
"grad_norm": 6.78953742980957,
"learning_rate": 2.564653371737649e-05,
"loss": 0.3858,
"step": 33000
},
{
"epoch": 0.7365713154943823,
"grad_norm": 12.592966079711914,
"learning_rate": 2.558057210703371e-05,
"loss": 0.3822,
"step": 33500
},
{
"epoch": 0.747564917218179,
"grad_norm": 1.8982641696929932,
"learning_rate": 2.5514610496690927e-05,
"loss": 0.3964,
"step": 34000
},
{
"epoch": 0.7585585189419758,
"grad_norm": 3.2166171073913574,
"learning_rate": 2.5448648886348144e-05,
"loss": 0.3902,
"step": 34500
},
{
"epoch": 0.7695521206657725,
"grad_norm": 10.258162498474121,
"learning_rate": 2.5382687276005365e-05,
"loss": 0.3691,
"step": 35000
},
{
"epoch": 0.7805457223895693,
"grad_norm": 2.9836630821228027,
"learning_rate": 2.5316725665662583e-05,
"loss": 0.3937,
"step": 35500
},
{
"epoch": 0.791539324113366,
"grad_norm": 10.085326194763184,
"learning_rate": 2.5250764055319804e-05,
"loss": 0.3897,
"step": 36000
},
{
"epoch": 0.8025329258371627,
"grad_norm": 8.519903182983398,
"learning_rate": 2.5184802444977024e-05,
"loss": 0.361,
"step": 36500
},
{
"epoch": 0.8135265275609596,
"grad_norm": 4.746450901031494,
"learning_rate": 2.5118840834634245e-05,
"loss": 0.3845,
"step": 37000
},
{
"epoch": 0.8245201292847563,
"grad_norm": 13.058253288269043,
"learning_rate": 2.5052879224291463e-05,
"loss": 0.3871,
"step": 37500
},
{
"epoch": 0.835513731008553,
"grad_norm": 2.8094441890716553,
"learning_rate": 2.4986917613948684e-05,
"loss": 0.4012,
"step": 38000
},
{
"epoch": 0.8465073327323498,
"grad_norm": 4.1435770988464355,
"learning_rate": 2.49209560036059e-05,
"loss": 0.3802,
"step": 38500
},
{
"epoch": 0.8575009344561465,
"grad_norm": 12.436211585998535,
"learning_rate": 2.4854994393263122e-05,
"loss": 0.3705,
"step": 39000
},
{
"epoch": 0.8684945361799433,
"grad_norm": 16.29452133178711,
"learning_rate": 2.478903278292034e-05,
"loss": 0.3921,
"step": 39500
},
{
"epoch": 0.87948813790374,
"grad_norm": 6.898037910461426,
"learning_rate": 2.472307117257756e-05,
"loss": 0.3925,
"step": 40000
},
{
"epoch": 0.8904817396275367,
"grad_norm": 6.402541160583496,
"learning_rate": 2.465710956223478e-05,
"loss": 0.3771,
"step": 40500
},
{
"epoch": 0.9014753413513336,
"grad_norm": 3.24283504486084,
"learning_rate": 2.4591147951892e-05,
"loss": 0.3698,
"step": 41000
},
{
"epoch": 0.9124689430751303,
"grad_norm": 7.773282527923584,
"learning_rate": 2.452518634154922e-05,
"loss": 0.3823,
"step": 41500
},
{
"epoch": 0.923462544798927,
"grad_norm": 4.645416736602783,
"learning_rate": 2.445922473120644e-05,
"loss": 0.3792,
"step": 42000
},
{
"epoch": 0.9344561465227238,
"grad_norm": 8.859955787658691,
"learning_rate": 2.4393263120863658e-05,
"loss": 0.3988,
"step": 42500
},
{
"epoch": 0.9454497482465205,
"grad_norm": 5.384950160980225,
"learning_rate": 2.4327301510520875e-05,
"loss": 0.367,
"step": 43000
},
{
"epoch": 0.9564433499703173,
"grad_norm": 21.994779586791992,
"learning_rate": 2.4261339900178096e-05,
"loss": 0.3768,
"step": 43500
},
{
"epoch": 0.967436951694114,
"grad_norm": 9.221137046813965,
"learning_rate": 2.4195378289835317e-05,
"loss": 0.3767,
"step": 44000
},
{
"epoch": 0.9784305534179107,
"grad_norm": 1.9626529216766357,
"learning_rate": 2.4129416679492538e-05,
"loss": 0.3626,
"step": 44500
},
{
"epoch": 0.9894241551417076,
"grad_norm": 19.04450798034668,
"learning_rate": 2.4063455069149755e-05,
"loss": 0.3754,
"step": 45000
},
{
"epoch": 1.0,
"eval_accuracy": 0.8456591367721558,
"eval_f1": 0.7902943944078505,
"eval_loss": 0.37383726239204407,
"eval_mcc": 0.6683280718139261,
"eval_runtime": 29.7825,
"eval_samples_per_second": 678.754,
"eval_steps_per_second": 84.848,
"step": 45481
},
{
"epoch": 1.0004177568655044,
"grad_norm": 14.85083293914795,
"learning_rate": 2.3997493458806976e-05,
"loss": 0.3783,
"step": 45500
},
{
"epoch": 1.011411358589301,
"grad_norm": 7.004974365234375,
"learning_rate": 2.3931531848464197e-05,
"loss": 0.3307,
"step": 46000
},
{
"epoch": 1.0224049603130978,
"grad_norm": 1.900647759437561,
"learning_rate": 2.386557023812141e-05,
"loss": 0.3454,
"step": 46500
},
{
"epoch": 1.0333985620368946,
"grad_norm": 13.517301559448242,
"learning_rate": 2.379960862777863e-05,
"loss": 0.3404,
"step": 47000
},
{
"epoch": 1.0443921637606912,
"grad_norm": 11.086533546447754,
"learning_rate": 2.3733647017435853e-05,
"loss": 0.344,
"step": 47500
},
{
"epoch": 1.055385765484488,
"grad_norm": 1.5423752069473267,
"learning_rate": 2.3667685407093073e-05,
"loss": 0.348,
"step": 48000
},
{
"epoch": 1.0663793672082849,
"grad_norm": 13.40974235534668,
"learning_rate": 2.360172379675029e-05,
"loss": 0.3353,
"step": 48500
},
{
"epoch": 1.0773729689320815,
"grad_norm": 8.961548805236816,
"learning_rate": 2.353576218640751e-05,
"loss": 0.3442,
"step": 49000
},
{
"epoch": 1.0883665706558783,
"grad_norm": 6.879663944244385,
"learning_rate": 2.3469800576064733e-05,
"loss": 0.331,
"step": 49500
},
{
"epoch": 1.099360172379675,
"grad_norm": 10.819347381591797,
"learning_rate": 2.3403838965721953e-05,
"loss": 0.3431,
"step": 50000
},
{
"epoch": 1.1103537741034717,
"grad_norm": 18.258974075317383,
"learning_rate": 2.3337877355379167e-05,
"loss": 0.3468,
"step": 50500
},
{
"epoch": 1.1213473758272685,
"grad_norm": 11.075167655944824,
"learning_rate": 2.3271915745036388e-05,
"loss": 0.3313,
"step": 51000
},
{
"epoch": 1.1323409775510653,
"grad_norm": 13.255118370056152,
"learning_rate": 2.320595413469361e-05,
"loss": 0.3379,
"step": 51500
},
{
"epoch": 1.1433345792748622,
"grad_norm": 9.165648460388184,
"learning_rate": 2.3139992524350827e-05,
"loss": 0.3402,
"step": 52000
},
{
"epoch": 1.1543281809986587,
"grad_norm": 20.563486099243164,
"learning_rate": 2.3074030914008047e-05,
"loss": 0.3429,
"step": 52500
},
{
"epoch": 1.1653217827224556,
"grad_norm": 23.879840850830078,
"learning_rate": 2.3008069303665268e-05,
"loss": 0.3437,
"step": 53000
},
{
"epoch": 1.1763153844462524,
"grad_norm": 16.95256996154785,
"learning_rate": 2.294210769332249e-05,
"loss": 0.3484,
"step": 53500
},
{
"epoch": 1.187308986170049,
"grad_norm": 23.673189163208008,
"learning_rate": 2.2876146082979707e-05,
"loss": 0.3175,
"step": 54000
},
{
"epoch": 1.1983025878938458,
"grad_norm": 12.443720817565918,
"learning_rate": 2.2810184472636927e-05,
"loss": 0.3471,
"step": 54500
},
{
"epoch": 1.2092961896176426,
"grad_norm": 6.558742046356201,
"learning_rate": 2.2744222862294145e-05,
"loss": 0.3463,
"step": 55000
},
{
"epoch": 1.2202897913414392,
"grad_norm": 3.0714826583862305,
"learning_rate": 2.2678261251951362e-05,
"loss": 0.3352,
"step": 55500
},
{
"epoch": 1.231283393065236,
"grad_norm": 6.919187068939209,
"learning_rate": 2.2612299641608583e-05,
"loss": 0.335,
"step": 56000
},
{
"epoch": 1.2422769947890329,
"grad_norm": 8.951086044311523,
"learning_rate": 2.2546338031265804e-05,
"loss": 0.3437,
"step": 56500
},
{
"epoch": 1.2532705965128295,
"grad_norm": 81.6339111328125,
"learning_rate": 2.2480376420923025e-05,
"loss": 0.318,
"step": 57000
},
{
"epoch": 1.2642641982366263,
"grad_norm": 10.197173118591309,
"learning_rate": 2.2414414810580242e-05,
"loss": 0.3478,
"step": 57500
},
{
"epoch": 1.275257799960423,
"grad_norm": 3.3102078437805176,
"learning_rate": 2.2348453200237463e-05,
"loss": 0.3316,
"step": 58000
},
{
"epoch": 1.2862514016842197,
"grad_norm": 7.871964454650879,
"learning_rate": 2.2282491589894684e-05,
"loss": 0.3329,
"step": 58500
},
{
"epoch": 1.2972450034080165,
"grad_norm": 13.741714477539062,
"learning_rate": 2.22165299795519e-05,
"loss": 0.3285,
"step": 59000
},
{
"epoch": 1.3082386051318133,
"grad_norm": 5.765045166015625,
"learning_rate": 2.215056836920912e-05,
"loss": 0.3363,
"step": 59500
},
{
"epoch": 1.31923220685561,
"grad_norm": 21.365049362182617,
"learning_rate": 2.208460675886634e-05,
"loss": 0.3475,
"step": 60000
},
{
"epoch": 1.3302258085794068,
"grad_norm": 16.869543075561523,
"learning_rate": 2.201864514852356e-05,
"loss": 0.3252,
"step": 60500
},
{
"epoch": 1.3412194103032036,
"grad_norm": 31.089399337768555,
"learning_rate": 2.1952683538180778e-05,
"loss": 0.3344,
"step": 61000
},
{
"epoch": 1.3522130120270002,
"grad_norm": 11.333529472351074,
"learning_rate": 2.1886721927838e-05,
"loss": 0.3641,
"step": 61500
},
{
"epoch": 1.363206613750797,
"grad_norm": 9.713915824890137,
"learning_rate": 2.182076031749522e-05,
"loss": 0.3415,
"step": 62000
},
{
"epoch": 1.3742002154745938,
"grad_norm": 8.068568229675293,
"learning_rate": 2.175479870715244e-05,
"loss": 0.3591,
"step": 62500
},
{
"epoch": 1.3851938171983904,
"grad_norm": 19.801572799682617,
"learning_rate": 2.1688837096809655e-05,
"loss": 0.335,
"step": 63000
},
{
"epoch": 1.3961874189221872,
"grad_norm": 28.160655975341797,
"learning_rate": 2.1622875486466876e-05,
"loss": 0.3374,
"step": 63500
},
{
"epoch": 1.407181020645984,
"grad_norm": 2.872919797897339,
"learning_rate": 2.1556913876124096e-05,
"loss": 0.3403,
"step": 64000
},
{
"epoch": 1.4181746223697809,
"grad_norm": 2.8728220462799072,
"learning_rate": 2.1490952265781317e-05,
"loss": 0.3384,
"step": 64500
},
{
"epoch": 1.4291682240935775,
"grad_norm": 5.093236923217773,
"learning_rate": 2.1424990655438535e-05,
"loss": 0.3305,
"step": 65000
},
{
"epoch": 1.4401618258173743,
"grad_norm": 17.081308364868164,
"learning_rate": 2.1359029045095756e-05,
"loss": 0.3589,
"step": 65500
},
{
"epoch": 1.451155427541171,
"grad_norm": 2.3248064517974854,
"learning_rate": 2.1293067434752976e-05,
"loss": 0.3313,
"step": 66000
},
{
"epoch": 1.462149029264968,
"grad_norm": 24.70163917541504,
"learning_rate": 2.1227105824410194e-05,
"loss": 0.3338,
"step": 66500
},
{
"epoch": 1.4731426309887645,
"grad_norm": 19.80680274963379,
"learning_rate": 2.116114421406741e-05,
"loss": 0.3487,
"step": 67000
},
{
"epoch": 1.4841362327125613,
"grad_norm": 2.365659713745117,
"learning_rate": 2.1095182603724632e-05,
"loss": 0.3479,
"step": 67500
},
{
"epoch": 1.4951298344363582,
"grad_norm": 2.9082655906677246,
"learning_rate": 2.1029220993381853e-05,
"loss": 0.3427,
"step": 68000
},
{
"epoch": 1.5061234361601548,
"grad_norm": 1.5247036218643188,
"learning_rate": 2.096325938303907e-05,
"loss": 0.321,
"step": 68500
},
{
"epoch": 1.5171170378839516,
"grad_norm": 13.082464218139648,
"learning_rate": 2.089729777269629e-05,
"loss": 0.347,
"step": 69000
},
{
"epoch": 1.5281106396077484,
"grad_norm": 32.83438491821289,
"learning_rate": 2.0831336162353512e-05,
"loss": 0.3382,
"step": 69500
},
{
"epoch": 1.539104241331545,
"grad_norm": 19.514705657958984,
"learning_rate": 2.0765374552010733e-05,
"loss": 0.334,
"step": 70000
},
{
"epoch": 1.5500978430553418,
"grad_norm": 11.129077911376953,
"learning_rate": 2.069941294166795e-05,
"loss": 0.3376,
"step": 70500
},
{
"epoch": 1.5610914447791386,
"grad_norm": 14.449658393859863,
"learning_rate": 2.0633451331325168e-05,
"loss": 0.3297,
"step": 71000
},
{
"epoch": 1.5720850465029352,
"grad_norm": 4.129580974578857,
"learning_rate": 2.056748972098239e-05,
"loss": 0.3405,
"step": 71500
},
{
"epoch": 1.583078648226732,
"grad_norm": 4.104194164276123,
"learning_rate": 2.0501528110639606e-05,
"loss": 0.3327,
"step": 72000
},
{
"epoch": 1.5940722499505289,
"grad_norm": 12.376803398132324,
"learning_rate": 2.0435566500296827e-05,
"loss": 0.3178,
"step": 72500
},
{
"epoch": 1.6050658516743255,
"grad_norm": 18.45488739013672,
"learning_rate": 2.0369604889954048e-05,
"loss": 0.3457,
"step": 73000
},
{
"epoch": 1.6160594533981223,
"grad_norm": 6.8812174797058105,
"learning_rate": 2.030364327961127e-05,
"loss": 0.3452,
"step": 73500
},
{
"epoch": 1.6270530551219191,
"grad_norm": 3.4659981727600098,
"learning_rate": 2.0237681669268486e-05,
"loss": 0.3407,
"step": 74000
},
{
"epoch": 1.6380466568457157,
"grad_norm": 21.697237014770508,
"learning_rate": 2.0171720058925707e-05,
"loss": 0.3493,
"step": 74500
},
{
"epoch": 1.6490402585695125,
"grad_norm": 20.997262954711914,
"learning_rate": 2.0105758448582925e-05,
"loss": 0.3453,
"step": 75000
},
{
"epoch": 1.6600338602933093,
"grad_norm": 8.582404136657715,
"learning_rate": 2.0039796838240142e-05,
"loss": 0.344,
"step": 75500
},
{
"epoch": 1.671027462017106,
"grad_norm": 15.028887748718262,
"learning_rate": 1.9973835227897363e-05,
"loss": 0.3428,
"step": 76000
},
{
"epoch": 1.6820210637409028,
"grad_norm": 15.200948715209961,
"learning_rate": 1.9907873617554584e-05,
"loss": 0.3531,
"step": 76500
},
{
"epoch": 1.6930146654646996,
"grad_norm": 12.243021011352539,
"learning_rate": 1.9841912007211805e-05,
"loss": 0.3288,
"step": 77000
},
{
"epoch": 1.7040082671884962,
"grad_norm": 0.7417749166488647,
"learning_rate": 1.9775950396869022e-05,
"loss": 0.3305,
"step": 77500
},
{
"epoch": 1.715001868912293,
"grad_norm": 12.099386215209961,
"learning_rate": 1.9709988786526243e-05,
"loss": 0.3521,
"step": 78000
},
{
"epoch": 1.7259954706360898,
"grad_norm": 10.566434860229492,
"learning_rate": 1.9644027176183464e-05,
"loss": 0.3466,
"step": 78500
},
{
"epoch": 1.7369890723598864,
"grad_norm": 1.6488581895828247,
"learning_rate": 1.957806556584068e-05,
"loss": 0.3453,
"step": 79000
},
{
"epoch": 1.7479826740836832,
"grad_norm": 15.446043968200684,
"learning_rate": 1.95121039554979e-05,
"loss": 0.3547,
"step": 79500
},
{
"epoch": 1.75897627580748,
"grad_norm": 3.6907153129577637,
"learning_rate": 1.944614234515512e-05,
"loss": 0.3346,
"step": 80000
},
{
"epoch": 1.7699698775312767,
"grad_norm": 13.95593547821045,
"learning_rate": 1.938018073481234e-05,
"loss": 0.3325,
"step": 80500
},
{
"epoch": 1.7809634792550737,
"grad_norm": 7.613198757171631,
"learning_rate": 1.9314219124469558e-05,
"loss": 0.3302,
"step": 81000
},
{
"epoch": 1.7919570809788703,
"grad_norm": 17.56180191040039,
"learning_rate": 1.924825751412678e-05,
"loss": 0.3303,
"step": 81500
},
{
"epoch": 1.802950682702667,
"grad_norm": 30.033525466918945,
"learning_rate": 1.9182295903784e-05,
"loss": 0.327,
"step": 82000
},
{
"epoch": 1.813944284426464,
"grad_norm": 2.658094644546509,
"learning_rate": 1.911633429344122e-05,
"loss": 0.3462,
"step": 82500
},
{
"epoch": 1.8249378861502605,
"grad_norm": 8.311567306518555,
"learning_rate": 1.9050372683098434e-05,
"loss": 0.3347,
"step": 83000
},
{
"epoch": 1.8359314878740571,
"grad_norm": 17.150461196899414,
"learning_rate": 1.8984411072755655e-05,
"loss": 0.3282,
"step": 83500
},
{
"epoch": 1.8469250895978542,
"grad_norm": 10.157220840454102,
"learning_rate": 1.8918449462412876e-05,
"loss": 0.3193,
"step": 84000
},
{
"epoch": 1.8579186913216508,
"grad_norm": 4.80257511138916,
"learning_rate": 1.8852487852070097e-05,
"loss": 0.3541,
"step": 84500
},
{
"epoch": 1.8689122930454476,
"grad_norm": 3.8798446655273438,
"learning_rate": 1.8786526241727314e-05,
"loss": 0.3483,
"step": 85000
},
{
"epoch": 1.8799058947692444,
"grad_norm": 8.881115913391113,
"learning_rate": 1.8720564631384535e-05,
"loss": 0.3302,
"step": 85500
},
{
"epoch": 1.890899496493041,
"grad_norm": 0.945717453956604,
"learning_rate": 1.8654603021041756e-05,
"loss": 0.336,
"step": 86000
},
{
"epoch": 1.9018930982168378,
"grad_norm": 5.144163608551025,
"learning_rate": 1.8588641410698974e-05,
"loss": 0.3148,
"step": 86500
},
{
"epoch": 1.9128866999406346,
"grad_norm": 1.5305918455123901,
"learning_rate": 1.852267980035619e-05,
"loss": 0.3395,
"step": 87000
},
{
"epoch": 1.9238803016644312,
"grad_norm": 15.06664752960205,
"learning_rate": 1.8456718190013412e-05,
"loss": 0.341,
"step": 87500
},
{
"epoch": 1.934873903388228,
"grad_norm": 33.20983123779297,
"learning_rate": 1.8390756579670633e-05,
"loss": 0.3619,
"step": 88000
},
{
"epoch": 1.9458675051120249,
"grad_norm": 11.427024841308594,
"learning_rate": 1.832479496932785e-05,
"loss": 0.3474,
"step": 88500
},
{
"epoch": 1.9568611068358215,
"grad_norm": 23.793506622314453,
"learning_rate": 1.825883335898507e-05,
"loss": 0.321,
"step": 89000
},
{
"epoch": 1.9678547085596183,
"grad_norm": 3.9075679779052734,
"learning_rate": 1.8192871748642292e-05,
"loss": 0.33,
"step": 89500
},
{
"epoch": 1.9788483102834151,
"grad_norm": 1.7106132507324219,
"learning_rate": 1.8126910138299513e-05,
"loss": 0.3562,
"step": 90000
},
{
"epoch": 1.9898419120072117,
"grad_norm": 1.6460707187652588,
"learning_rate": 1.806094852795673e-05,
"loss": 0.3378,
"step": 90500
},
{
"epoch": 2.0,
"eval_accuracy": 0.8638139963150024,
"eval_f1": 0.8192264758027448,
"eval_loss": 0.3989393711090088,
"eval_mcc": 0.7101352338901957,
"eval_runtime": 29.2332,
"eval_samples_per_second": 691.508,
"eval_steps_per_second": 86.443,
"step": 90962
},
{
"epoch": 2.0008355137310088,
"grad_norm": 0.9355267882347107,
"learning_rate": 1.7994986917613948e-05,
"loss": 0.3212,
"step": 91000
},
{
"epoch": 2.0118291154548054,
"grad_norm": 11.889479637145996,
"learning_rate": 1.792902530727117e-05,
"loss": 0.3003,
"step": 91500
},
{
"epoch": 2.022822717178602,
"grad_norm": 5.802761077880859,
"learning_rate": 1.7863063696928386e-05,
"loss": 0.2923,
"step": 92000
},
{
"epoch": 2.033816318902399,
"grad_norm": 7.432724475860596,
"learning_rate": 1.7797102086585607e-05,
"loss": 0.293,
"step": 92500
},
{
"epoch": 2.0448099206261956,
"grad_norm": 4.792222499847412,
"learning_rate": 1.7731140476242828e-05,
"loss": 0.3043,
"step": 93000
},
{
"epoch": 2.055803522349992,
"grad_norm": 8.417468070983887,
"learning_rate": 1.766517886590005e-05,
"loss": 0.2916,
"step": 93500
},
{
"epoch": 2.0667971240737892,
"grad_norm": 1.0828003883361816,
"learning_rate": 1.7599217255557266e-05,
"loss": 0.3112,
"step": 94000
},
{
"epoch": 2.077790725797586,
"grad_norm": 3.0800647735595703,
"learning_rate": 1.7533255645214487e-05,
"loss": 0.2804,
"step": 94500
},
{
"epoch": 2.0887843275213824,
"grad_norm": 3.254809617996216,
"learning_rate": 1.7467294034871704e-05,
"loss": 0.2909,
"step": 95000
},
{
"epoch": 2.0997779292451795,
"grad_norm": 2.6388871669769287,
"learning_rate": 1.7401332424528925e-05,
"loss": 0.305,
"step": 95500
},
{
"epoch": 2.110771530968976,
"grad_norm": 16.700735092163086,
"learning_rate": 1.7335370814186143e-05,
"loss": 0.2901,
"step": 96000
},
{
"epoch": 2.1217651326927727,
"grad_norm": 17.04293441772461,
"learning_rate": 1.7269409203843363e-05,
"loss": 0.3217,
"step": 96500
},
{
"epoch": 2.1327587344165697,
"grad_norm": 1.1329630613327026,
"learning_rate": 1.7203447593500584e-05,
"loss": 0.2898,
"step": 97000
},
{
"epoch": 2.1437523361403663,
"grad_norm": 36.23415756225586,
"learning_rate": 1.7137485983157802e-05,
"loss": 0.2864,
"step": 97500
},
{
"epoch": 2.154745937864163,
"grad_norm": 106.94963836669922,
"learning_rate": 1.7071524372815023e-05,
"loss": 0.3083,
"step": 98000
},
{
"epoch": 2.16573953958796,
"grad_norm": 0.37686920166015625,
"learning_rate": 1.7005562762472243e-05,
"loss": 0.3238,
"step": 98500
},
{
"epoch": 2.1767331413117565,
"grad_norm": 36.30667495727539,
"learning_rate": 1.693960115212946e-05,
"loss": 0.2917,
"step": 99000
},
{
"epoch": 2.187726743035553,
"grad_norm": 35.45988082885742,
"learning_rate": 1.687363954178668e-05,
"loss": 0.3009,
"step": 99500
},
{
"epoch": 2.19872034475935,
"grad_norm": 0.25279441475868225,
"learning_rate": 1.68076779314439e-05,
"loss": 0.3158,
"step": 100000
},
{
"epoch": 2.209713946483147,
"grad_norm": 10.676984786987305,
"learning_rate": 1.674171632110112e-05,
"loss": 0.3005,
"step": 100500
},
{
"epoch": 2.2207075482069434,
"grad_norm": 19.04880714416504,
"learning_rate": 1.6675754710758337e-05,
"loss": 0.3018,
"step": 101000
},
{
"epoch": 2.2317011499307404,
"grad_norm": 35.431583404541016,
"learning_rate": 1.660979310041556e-05,
"loss": 0.3135,
"step": 101500
},
{
"epoch": 2.242694751654537,
"grad_norm": 0.3378468155860901,
"learning_rate": 1.654383149007278e-05,
"loss": 0.2779,
"step": 102000
},
{
"epoch": 2.2536883533783336,
"grad_norm": 46.81476974487305,
"learning_rate": 1.647786987973e-05,
"loss": 0.279,
"step": 102500
},
{
"epoch": 2.2646819551021307,
"grad_norm": 36.49277877807617,
"learning_rate": 1.6411908269387214e-05,
"loss": 0.3184,
"step": 103000
},
{
"epoch": 2.2756755568259273,
"grad_norm": 12.877152442932129,
"learning_rate": 1.6345946659044435e-05,
"loss": 0.3024,
"step": 103500
},
{
"epoch": 2.2866691585497243,
"grad_norm": 4.798713684082031,
"learning_rate": 1.6279985048701656e-05,
"loss": 0.3014,
"step": 104000
},
{
"epoch": 2.297662760273521,
"grad_norm": 6.63606071472168,
"learning_rate": 1.6214023438358877e-05,
"loss": 0.3214,
"step": 104500
},
{
"epoch": 2.3086563619973175,
"grad_norm": 13.403897285461426,
"learning_rate": 1.6148061828016094e-05,
"loss": 0.2943,
"step": 105000
},
{
"epoch": 2.3196499637211145,
"grad_norm": 33.9350471496582,
"learning_rate": 1.6082100217673315e-05,
"loss": 0.302,
"step": 105500
},
{
"epoch": 2.330643565444911,
"grad_norm": 3.330829620361328,
"learning_rate": 1.6016138607330536e-05,
"loss": 0.3087,
"step": 106000
},
{
"epoch": 2.3416371671687077,
"grad_norm": 1.2686516046524048,
"learning_rate": 1.5950176996987753e-05,
"loss": 0.3007,
"step": 106500
},
{
"epoch": 2.3526307688925048,
"grad_norm": 20.976926803588867,
"learning_rate": 1.5884215386644974e-05,
"loss": 0.3094,
"step": 107000
},
{
"epoch": 2.3636243706163014,
"grad_norm": 0.7142143249511719,
"learning_rate": 1.581825377630219e-05,
"loss": 0.3169,
"step": 107500
},
{
"epoch": 2.374617972340098,
"grad_norm": 6.738494873046875,
"learning_rate": 1.5752292165959412e-05,
"loss": 0.3101,
"step": 108000
},
{
"epoch": 2.385611574063895,
"grad_norm": 0.8053629398345947,
"learning_rate": 1.568633055561663e-05,
"loss": 0.3208,
"step": 108500
},
{
"epoch": 2.3966051757876916,
"grad_norm": 0.35285481810569763,
"learning_rate": 1.562036894527385e-05,
"loss": 0.2903,
"step": 109000
},
{
"epoch": 2.407598777511488,
"grad_norm": 0.9598795771598816,
"learning_rate": 1.555440733493107e-05,
"loss": 0.3189,
"step": 109500
},
{
"epoch": 2.4185923792352853,
"grad_norm": 8.283425331115723,
"learning_rate": 1.5488445724588292e-05,
"loss": 0.2922,
"step": 110000
},
{
"epoch": 2.429585980959082,
"grad_norm": 2.2365481853485107,
"learning_rate": 1.542248411424551e-05,
"loss": 0.2865,
"step": 110500
},
{
"epoch": 2.4405795826828784,
"grad_norm": 22.584705352783203,
"learning_rate": 1.535652250390273e-05,
"loss": 0.2883,
"step": 111000
},
{
"epoch": 2.4515731844066755,
"grad_norm": 1.3138020038604736,
"learning_rate": 1.5290560893559948e-05,
"loss": 0.3233,
"step": 111500
},
{
"epoch": 2.462566786130472,
"grad_norm": 17.076557159423828,
"learning_rate": 1.5224599283217167e-05,
"loss": 0.3017,
"step": 112000
},
{
"epoch": 2.4735603878542687,
"grad_norm": 139.9231719970703,
"learning_rate": 1.5158637672874386e-05,
"loss": 0.3033,
"step": 112500
},
{
"epoch": 2.4845539895780657,
"grad_norm": 8.334077835083008,
"learning_rate": 1.5092676062531607e-05,
"loss": 0.2925,
"step": 113000
},
{
"epoch": 2.4955475913018623,
"grad_norm": 0.4488193094730377,
"learning_rate": 1.5026714452188828e-05,
"loss": 0.314,
"step": 113500
},
{
"epoch": 2.506541193025659,
"grad_norm": 18.986644744873047,
"learning_rate": 1.4960752841846047e-05,
"loss": 0.3011,
"step": 114000
},
{
"epoch": 2.517534794749456,
"grad_norm": 0.16863927245140076,
"learning_rate": 1.4894791231503265e-05,
"loss": 0.2845,
"step": 114500
},
{
"epoch": 2.5285283964732526,
"grad_norm": 19.12157440185547,
"learning_rate": 1.4828829621160486e-05,
"loss": 0.3092,
"step": 115000
},
{
"epoch": 2.539521998197049,
"grad_norm": 6.872998237609863,
"learning_rate": 1.4762868010817705e-05,
"loss": 0.3072,
"step": 115500
},
{
"epoch": 2.550515599920846,
"grad_norm": 0.4193851947784424,
"learning_rate": 1.4696906400474924e-05,
"loss": 0.3004,
"step": 116000
},
{
"epoch": 2.561509201644643,
"grad_norm": 0.6917738318443298,
"learning_rate": 1.4630944790132143e-05,
"loss": 0.2967,
"step": 116500
},
{
"epoch": 2.5725028033684394,
"grad_norm": 10.825478553771973,
"learning_rate": 1.4564983179789362e-05,
"loss": 0.3086,
"step": 117000
},
{
"epoch": 2.5834964050922364,
"grad_norm": 68.07927703857422,
"learning_rate": 1.4499021569446583e-05,
"loss": 0.2964,
"step": 117500
},
{
"epoch": 2.594490006816033,
"grad_norm": 5.563518047332764,
"learning_rate": 1.4433059959103802e-05,
"loss": 0.3067,
"step": 118000
},
{
"epoch": 2.6054836085398296,
"grad_norm": 4.1622633934021,
"learning_rate": 1.4367098348761021e-05,
"loss": 0.3019,
"step": 118500
},
{
"epoch": 2.6164772102636267,
"grad_norm": 20.468860626220703,
"learning_rate": 1.430113673841824e-05,
"loss": 0.2857,
"step": 119000
},
{
"epoch": 2.6274708119874233,
"grad_norm": 19.43634605407715,
"learning_rate": 1.4235175128075461e-05,
"loss": 0.3079,
"step": 119500
},
{
"epoch": 2.63846441371122,
"grad_norm": 6.021149158477783,
"learning_rate": 1.416921351773268e-05,
"loss": 0.2901,
"step": 120000
},
{
"epoch": 2.649458015435017,
"grad_norm": 8.589285850524902,
"learning_rate": 1.41032519073899e-05,
"loss": 0.3109,
"step": 120500
},
{
"epoch": 2.6604516171588135,
"grad_norm": 16.921823501586914,
"learning_rate": 1.4037290297047119e-05,
"loss": 0.3002,
"step": 121000
},
{
"epoch": 2.67144521888261,
"grad_norm": 16.486186981201172,
"learning_rate": 1.3971328686704338e-05,
"loss": 0.306,
"step": 121500
},
{
"epoch": 2.682438820606407,
"grad_norm": 8.290379524230957,
"learning_rate": 1.3905367076361559e-05,
"loss": 0.3005,
"step": 122000
},
{
"epoch": 2.6934324223302037,
"grad_norm": 0.8587543964385986,
"learning_rate": 1.3839405466018776e-05,
"loss": 0.29,
"step": 122500
},
{
"epoch": 2.7044260240540003,
"grad_norm": 45.68854904174805,
"learning_rate": 1.3773443855675997e-05,
"loss": 0.3037,
"step": 123000
},
{
"epoch": 2.7154196257777974,
"grad_norm": 13.316100120544434,
"learning_rate": 1.3707482245333216e-05,
"loss": 0.2717,
"step": 123500
},
{
"epoch": 2.726413227501594,
"grad_norm": 5.796350479125977,
"learning_rate": 1.3641520634990437e-05,
"loss": 0.3116,
"step": 124000
},
{
"epoch": 2.7374068292253906,
"grad_norm": 10.975761413574219,
"learning_rate": 1.3575559024647655e-05,
"loss": 0.3089,
"step": 124500
},
{
"epoch": 2.7484004309491876,
"grad_norm": 99.86316680908203,
"learning_rate": 1.3509597414304875e-05,
"loss": 0.3071,
"step": 125000
},
{
"epoch": 2.759394032672984,
"grad_norm": 92.33716583251953,
"learning_rate": 1.3443635803962095e-05,
"loss": 0.2881,
"step": 125500
},
{
"epoch": 2.770387634396781,
"grad_norm": 20.75370979309082,
"learning_rate": 1.3377674193619314e-05,
"loss": 0.2922,
"step": 126000
},
{
"epoch": 2.781381236120578,
"grad_norm": 63.51997756958008,
"learning_rate": 1.3311712583276533e-05,
"loss": 0.3016,
"step": 126500
},
{
"epoch": 2.7923748378443745,
"grad_norm": 12.819772720336914,
"learning_rate": 1.3245750972933752e-05,
"loss": 0.3005,
"step": 127000
},
{
"epoch": 2.803368439568171,
"grad_norm": 14.167094230651855,
"learning_rate": 1.3179789362590973e-05,
"loss": 0.3099,
"step": 127500
},
{
"epoch": 2.814362041291968,
"grad_norm": 14.828591346740723,
"learning_rate": 1.3113827752248192e-05,
"loss": 0.2759,
"step": 128000
},
{
"epoch": 2.8253556430157647,
"grad_norm": 9.91226577758789,
"learning_rate": 1.3047866141905411e-05,
"loss": 0.3025,
"step": 128500
},
{
"epoch": 2.8363492447395617,
"grad_norm": 38.544525146484375,
"learning_rate": 1.298190453156263e-05,
"loss": 0.3038,
"step": 129000
},
{
"epoch": 2.8473428464633583,
"grad_norm": 5.008056640625,
"learning_rate": 1.2915942921219851e-05,
"loss": 0.2947,
"step": 129500
},
{
"epoch": 2.858336448187155,
"grad_norm": 14.466870307922363,
"learning_rate": 1.284998131087707e-05,
"loss": 0.2989,
"step": 130000
},
{
"epoch": 2.869330049910952,
"grad_norm": 0.3647148907184601,
"learning_rate": 1.278401970053429e-05,
"loss": 0.2864,
"step": 130500
},
{
"epoch": 2.8803236516347486,
"grad_norm": 0.18057258427143097,
"learning_rate": 1.2718058090191509e-05,
"loss": 0.2894,
"step": 131000
},
{
"epoch": 2.891317253358545,
"grad_norm": 0.5057438015937805,
"learning_rate": 1.2652096479848728e-05,
"loss": 0.302,
"step": 131500
},
{
"epoch": 2.902310855082342,
"grad_norm": 10.934133529663086,
"learning_rate": 1.2586134869505949e-05,
"loss": 0.3047,
"step": 132000
},
{
"epoch": 2.913304456806139,
"grad_norm": 2.3341269493103027,
"learning_rate": 1.2520173259163166e-05,
"loss": 0.3011,
"step": 132500
},
{
"epoch": 2.924298058529936,
"grad_norm": 0.3195688724517822,
"learning_rate": 1.2454211648820387e-05,
"loss": 0.2939,
"step": 133000
},
{
"epoch": 2.9352916602537324,
"grad_norm": 8.257743835449219,
"learning_rate": 1.2388250038477606e-05,
"loss": 0.3057,
"step": 133500
},
{
"epoch": 2.946285261977529,
"grad_norm": 20.577478408813477,
"learning_rate": 1.2322288428134827e-05,
"loss": 0.2966,
"step": 134000
},
{
"epoch": 2.957278863701326,
"grad_norm": 0.777562141418457,
"learning_rate": 1.2256326817792044e-05,
"loss": 0.3013,
"step": 134500
},
{
"epoch": 2.9682724654251227,
"grad_norm": 58.6212158203125,
"learning_rate": 1.2190365207449265e-05,
"loss": 0.2874,
"step": 135000
},
{
"epoch": 2.9792660671489193,
"grad_norm": 69.42217254638672,
"learning_rate": 1.2124403597106484e-05,
"loss": 0.293,
"step": 135500
},
{
"epoch": 2.9902596688727163,
"grad_norm": 4.408263683319092,
"learning_rate": 1.2058441986763704e-05,
"loss": 0.3005,
"step": 136000
},
{
"epoch": 3.0,
"eval_accuracy": 0.8718278408050537,
"eval_f1": 0.8299757201916136,
"eval_loss": 0.42042940855026245,
"eval_mcc": 0.7272909835972381,
"eval_runtime": 28.09,
"eval_samples_per_second": 719.65,
"eval_steps_per_second": 89.961,
"step": 136443
},
{
"epoch": 3.001253270596513,
"grad_norm": 6.014369010925293,
"learning_rate": 1.1992480376420923e-05,
"loss": 0.281,
"step": 136500
},
{
"epoch": 3.0122468723203095,
"grad_norm": 10.639359474182129,
"learning_rate": 1.1926518766078142e-05,
"loss": 0.2556,
"step": 137000
},
{
"epoch": 3.023240474044106,
"grad_norm": 3.0724806785583496,
"learning_rate": 1.1860557155735363e-05,
"loss": 0.2414,
"step": 137500
},
{
"epoch": 3.034234075767903,
"grad_norm": 0.3316449522972107,
"learning_rate": 1.1794595545392582e-05,
"loss": 0.2539,
"step": 138000
},
{
"epoch": 3.0452276774916998,
"grad_norm": 23.327177047729492,
"learning_rate": 1.1728633935049801e-05,
"loss": 0.2374,
"step": 138500
},
{
"epoch": 3.056221279215497,
"grad_norm": 0.7128089666366577,
"learning_rate": 1.166267232470702e-05,
"loss": 0.2476,
"step": 139000
},
{
"epoch": 3.0672148809392934,
"grad_norm": 189.15638732910156,
"learning_rate": 1.1596710714364241e-05,
"loss": 0.2431,
"step": 139500
},
{
"epoch": 3.07820848266309,
"grad_norm": 17.80859375,
"learning_rate": 1.153074910402146e-05,
"loss": 0.225,
"step": 140000
},
{
"epoch": 3.089202084386887,
"grad_norm": 0.14886409044265747,
"learning_rate": 1.146478749367868e-05,
"loss": 0.2495,
"step": 140500
},
{
"epoch": 3.1001956861106836,
"grad_norm": 0.5925188064575195,
"learning_rate": 1.1398825883335899e-05,
"loss": 0.2541,
"step": 141000
},
{
"epoch": 3.1111892878344802,
"grad_norm": 9.857983589172363,
"learning_rate": 1.1332864272993118e-05,
"loss": 0.2565,
"step": 141500
},
{
"epoch": 3.1221828895582773,
"grad_norm": 0.7951391339302063,
"learning_rate": 1.1266902662650339e-05,
"loss": 0.2561,
"step": 142000
},
{
"epoch": 3.133176491282074,
"grad_norm": 61.58017349243164,
"learning_rate": 1.1200941052307556e-05,
"loss": 0.2361,
"step": 142500
},
{
"epoch": 3.1441700930058705,
"grad_norm": 9.199590682983398,
"learning_rate": 1.1134979441964777e-05,
"loss": 0.2559,
"step": 143000
},
{
"epoch": 3.1551636947296675,
"grad_norm": 1.7396503686904907,
"learning_rate": 1.1069017831621996e-05,
"loss": 0.2543,
"step": 143500
},
{
"epoch": 3.166157296453464,
"grad_norm": 185.94760131835938,
"learning_rate": 1.1003056221279217e-05,
"loss": 0.2309,
"step": 144000
},
{
"epoch": 3.1771508981772607,
"grad_norm": 42.58454132080078,
"learning_rate": 1.0937094610936434e-05,
"loss": 0.2634,
"step": 144500
},
{
"epoch": 3.1881444999010577,
"grad_norm": 0.19487299025058746,
"learning_rate": 1.0871133000593655e-05,
"loss": 0.2668,
"step": 145000
},
{
"epoch": 3.1991381016248543,
"grad_norm": 0.11774999648332596,
"learning_rate": 1.0805171390250874e-05,
"loss": 0.2531,
"step": 145500
},
{
"epoch": 3.210131703348651,
"grad_norm": 0.8709030747413635,
"learning_rate": 1.0739209779908093e-05,
"loss": 0.2579,
"step": 146000
},
{
"epoch": 3.221125305072448,
"grad_norm": 0.26620733737945557,
"learning_rate": 1.0673248169565313e-05,
"loss": 0.2456,
"step": 146500
},
{
"epoch": 3.2321189067962446,
"grad_norm": 30.161376953125,
"learning_rate": 1.0607286559222532e-05,
"loss": 0.2588,
"step": 147000
},
{
"epoch": 3.243112508520041,
"grad_norm": 14.72189998626709,
"learning_rate": 1.0541324948879753e-05,
"loss": 0.2538,
"step": 147500
},
{
"epoch": 3.2541061102438382,
"grad_norm": 22.82953453063965,
"learning_rate": 1.0475363338536972e-05,
"loss": 0.2398,
"step": 148000
},
{
"epoch": 3.265099711967635,
"grad_norm": 1.3340407609939575,
"learning_rate": 1.0409401728194191e-05,
"loss": 0.2619,
"step": 148500
},
{
"epoch": 3.2760933136914314,
"grad_norm": 4.700684070587158,
"learning_rate": 1.034344011785141e-05,
"loss": 0.2563,
"step": 149000
},
{
"epoch": 3.2870869154152285,
"grad_norm": 49.13290786743164,
"learning_rate": 1.0277478507508631e-05,
"loss": 0.2756,
"step": 149500
},
{
"epoch": 3.298080517139025,
"grad_norm": 0.5340966582298279,
"learning_rate": 1.021151689716585e-05,
"loss": 0.249,
"step": 150000
},
{
"epoch": 3.3090741188628217,
"grad_norm": 156.9650115966797,
"learning_rate": 1.014555528682307e-05,
"loss": 0.237,
"step": 150500
},
{
"epoch": 3.3200677205866187,
"grad_norm": 0.09667583554983139,
"learning_rate": 1.0079593676480288e-05,
"loss": 0.2621,
"step": 151000
},
{
"epoch": 3.3310613223104153,
"grad_norm": 9.222663879394531,
"learning_rate": 1.0013632066137508e-05,
"loss": 0.2457,
"step": 151500
},
{
"epoch": 3.342054924034212,
"grad_norm": 14.612710952758789,
"learning_rate": 9.947670455794728e-06,
"loss": 0.2555,
"step": 152000
},
{
"epoch": 3.353048525758009,
"grad_norm": 50.92832565307617,
"learning_rate": 9.881708845451946e-06,
"loss": 0.2394,
"step": 152500
},
{
"epoch": 3.3640421274818055,
"grad_norm": 1.5905165672302246,
"learning_rate": 9.815747235109167e-06,
"loss": 0.2408,
"step": 153000
},
{
"epoch": 3.375035729205602,
"grad_norm": 34.99452209472656,
"learning_rate": 9.749785624766386e-06,
"loss": 0.2351,
"step": 153500
},
{
"epoch": 3.386029330929399,
"grad_norm": 1.3218666315078735,
"learning_rate": 9.683824014423607e-06,
"loss": 0.2494,
"step": 154000
},
{
"epoch": 3.3970229326531958,
"grad_norm": 19.163127899169922,
"learning_rate": 9.617862404080824e-06,
"loss": 0.2716,
"step": 154500
},
{
"epoch": 3.4080165343769924,
"grad_norm": 0.3468831479549408,
"learning_rate": 9.551900793738045e-06,
"loss": 0.2617,
"step": 155000
},
{
"epoch": 3.4190101361007894,
"grad_norm": 19.626012802124023,
"learning_rate": 9.485939183395264e-06,
"loss": 0.2651,
"step": 155500
},
{
"epoch": 3.430003737824586,
"grad_norm": 17.755313873291016,
"learning_rate": 9.419977573052483e-06,
"loss": 0.2487,
"step": 156000
},
{
"epoch": 3.4409973395483826,
"grad_norm": 14.6954984664917,
"learning_rate": 9.354015962709702e-06,
"loss": 0.2579,
"step": 156500
},
{
"epoch": 3.4519909412721796,
"grad_norm": 24.834569931030273,
"learning_rate": 9.288054352366922e-06,
"loss": 0.2343,
"step": 157000
},
{
"epoch": 3.4629845429959762,
"grad_norm": 11.748075485229492,
"learning_rate": 9.222092742024142e-06,
"loss": 0.2493,
"step": 157500
},
{
"epoch": 3.473978144719773,
"grad_norm": 0.26250067353248596,
"learning_rate": 9.156131131681362e-06,
"loss": 0.2518,
"step": 158000
},
{
"epoch": 3.48497174644357,
"grad_norm": 0.31238773465156555,
"learning_rate": 9.09016952133858e-06,
"loss": 0.2563,
"step": 158500
},
{
"epoch": 3.4959653481673665,
"grad_norm": 0.38414067029953003,
"learning_rate": 9.0242079109958e-06,
"loss": 0.2605,
"step": 159000
},
{
"epoch": 3.506958949891163,
"grad_norm": 0.26335904002189636,
"learning_rate": 8.95824630065302e-06,
"loss": 0.25,
"step": 159500
},
{
"epoch": 3.51795255161496,
"grad_norm": 11.267284393310547,
"learning_rate": 8.89228469031024e-06,
"loss": 0.2446,
"step": 160000
},
{
"epoch": 3.5289461533387567,
"grad_norm": 13.036714553833008,
"learning_rate": 8.826323079967459e-06,
"loss": 0.2341,
"step": 160500
},
{
"epoch": 3.5399397550625533,
"grad_norm": 9.3615140914917,
"learning_rate": 8.760361469624678e-06,
"loss": 0.2632,
"step": 161000
},
{
"epoch": 3.5509333567863504,
"grad_norm": 0.3549996018409729,
"learning_rate": 8.694399859281897e-06,
"loss": 0.2507,
"step": 161500
},
{
"epoch": 3.561926958510147,
"grad_norm": 0.46619582176208496,
"learning_rate": 8.628438248939118e-06,
"loss": 0.2543,
"step": 162000
},
{
"epoch": 3.572920560233944,
"grad_norm": 0.9738485217094421,
"learning_rate": 8.562476638596337e-06,
"loss": 0.2672,
"step": 162500
},
{
"epoch": 3.5839141619577406,
"grad_norm": 12.682645797729492,
"learning_rate": 8.496515028253557e-06,
"loss": 0.2422,
"step": 163000
},
{
"epoch": 3.594907763681537,
"grad_norm": 0.33584246039390564,
"learning_rate": 8.430553417910776e-06,
"loss": 0.2597,
"step": 163500
},
{
"epoch": 3.6059013654053342,
"grad_norm": 8.06340217590332,
"learning_rate": 8.364591807567997e-06,
"loss": 0.271,
"step": 164000
},
{
"epoch": 3.616894967129131,
"grad_norm": 0.4343748688697815,
"learning_rate": 8.298630197225216e-06,
"loss": 0.2686,
"step": 164500
},
{
"epoch": 3.6278885688529274,
"grad_norm": 38.23839569091797,
"learning_rate": 8.232668586882435e-06,
"loss": 0.2605,
"step": 165000
},
{
"epoch": 3.6388821705767245,
"grad_norm": 20.085224151611328,
"learning_rate": 8.166706976539654e-06,
"loss": 0.2449,
"step": 165500
},
{
"epoch": 3.649875772300521,
"grad_norm": 0.3538534641265869,
"learning_rate": 8.100745366196875e-06,
"loss": 0.2436,
"step": 166000
},
{
"epoch": 3.660869374024318,
"grad_norm": 0.3242553174495697,
"learning_rate": 8.034783755854094e-06,
"loss": 0.2468,
"step": 166500
},
{
"epoch": 3.6718629757481147,
"grad_norm": 28.696617126464844,
"learning_rate": 7.968822145511311e-06,
"loss": 0.2621,
"step": 167000
},
{
"epoch": 3.6828565774719113,
"grad_norm": 0.26111406087875366,
"learning_rate": 7.902860535168532e-06,
"loss": 0.2519,
"step": 167500
},
{
"epoch": 3.6938501791957083,
"grad_norm": 0.24540553987026215,
"learning_rate": 7.836898924825751e-06,
"loss": 0.2515,
"step": 168000
},
{
"epoch": 3.704843780919505,
"grad_norm": 0.4676073491573334,
"learning_rate": 7.770937314482972e-06,
"loss": 0.256,
"step": 168500
},
{
"epoch": 3.7158373826433015,
"grad_norm": 87.25594329833984,
"learning_rate": 7.70497570414019e-06,
"loss": 0.2354,
"step": 169000
},
{
"epoch": 3.7268309843670986,
"grad_norm": 1.1010403633117676,
"learning_rate": 7.63901409379741e-06,
"loss": 0.2394,
"step": 169500
},
{
"epoch": 3.737824586090895,
"grad_norm": 0.1542312502861023,
"learning_rate": 7.57305248345463e-06,
"loss": 0.2491,
"step": 170000
},
{
"epoch": 3.748818187814692,
"grad_norm": 2.4090046882629395,
"learning_rate": 7.50709087311185e-06,
"loss": 0.2337,
"step": 170500
},
{
"epoch": 3.759811789538489,
"grad_norm": 6.501917362213135,
"learning_rate": 7.441129262769069e-06,
"loss": 0.2406,
"step": 171000
},
{
"epoch": 3.7708053912622854,
"grad_norm": 19.246479034423828,
"learning_rate": 7.375167652426288e-06,
"loss": 0.2547,
"step": 171500
},
{
"epoch": 3.781798992986082,
"grad_norm": 0.10231161117553711,
"learning_rate": 7.309206042083508e-06,
"loss": 0.2379,
"step": 172000
},
{
"epoch": 3.792792594709879,
"grad_norm": 10.832609176635742,
"learning_rate": 7.243244431740727e-06,
"loss": 0.2711,
"step": 172500
},
{
"epoch": 3.8037861964336757,
"grad_norm": 18.556346893310547,
"learning_rate": 7.177282821397947e-06,
"loss": 0.2265,
"step": 173000
},
{
"epoch": 3.8147797981574723,
"grad_norm": 1.0246055126190186,
"learning_rate": 7.111321211055166e-06,
"loss": 0.2334,
"step": 173500
},
{
"epoch": 3.8257733998812693,
"grad_norm": 1.1454087495803833,
"learning_rate": 7.0453596007123855e-06,
"loss": 0.2491,
"step": 174000
},
{
"epoch": 3.836767001605066,
"grad_norm": 11.247049331665039,
"learning_rate": 6.979397990369605e-06,
"loss": 0.2632,
"step": 174500
},
{
"epoch": 3.8477606033288625,
"grad_norm": 14.466601371765137,
"learning_rate": 6.913436380026825e-06,
"loss": 0.2639,
"step": 175000
},
{
"epoch": 3.8587542050526595,
"grad_norm": 7.91213321685791,
"learning_rate": 6.847474769684044e-06,
"loss": 0.2527,
"step": 175500
},
{
"epoch": 3.869747806776456,
"grad_norm": 26.528411865234375,
"learning_rate": 6.781513159341264e-06,
"loss": 0.2438,
"step": 176000
},
{
"epoch": 3.8807414085002527,
"grad_norm": 0.7833952903747559,
"learning_rate": 6.715551548998483e-06,
"loss": 0.2585,
"step": 176500
},
{
"epoch": 3.8917350102240498,
"grad_norm": 53.77830123901367,
"learning_rate": 6.649589938655703e-06,
"loss": 0.2615,
"step": 177000
},
{
"epoch": 3.9027286119478464,
"grad_norm": 0.6139953136444092,
"learning_rate": 6.583628328312922e-06,
"loss": 0.2644,
"step": 177500
},
{
"epoch": 3.913722213671643,
"grad_norm": 1.4486163854599,
"learning_rate": 6.517666717970142e-06,
"loss": 0.2302,
"step": 178000
},
{
"epoch": 3.92471581539544,
"grad_norm": 5.287415504455566,
"learning_rate": 6.451705107627361e-06,
"loss": 0.2727,
"step": 178500
},
{
"epoch": 3.9357094171192366,
"grad_norm": 26.611614227294922,
"learning_rate": 6.3857434972845804e-06,
"loss": 0.2514,
"step": 179000
},
{
"epoch": 3.946703018843033,
"grad_norm": 14.361977577209473,
"learning_rate": 6.3197818869418e-06,
"loss": 0.2476,
"step": 179500
},
{
"epoch": 3.9576966205668302,
"grad_norm": 1.3597434759140015,
"learning_rate": 6.25382027659902e-06,
"loss": 0.2406,
"step": 180000
},
{
"epoch": 3.968690222290627,
"grad_norm": 10.218100547790527,
"learning_rate": 6.187858666256239e-06,
"loss": 0.254,
"step": 180500
},
{
"epoch": 3.9796838240144234,
"grad_norm": 0.7666225433349609,
"learning_rate": 6.121897055913459e-06,
"loss": 0.2741,
"step": 181000
},
{
"epoch": 3.9906774257382205,
"grad_norm": 36.5604133605957,
"learning_rate": 6.055935445570678e-06,
"loss": 0.2374,
"step": 181500
},
{
"epoch": 4.0,
"eval_accuracy": 0.8765273094177246,
"eval_f1": 0.8365850464842216,
"eval_loss": 0.5333936214447021,
"eval_mcc": 0.7376033359055921,
"eval_runtime": 27.772,
"eval_samples_per_second": 727.891,
"eval_steps_per_second": 90.991,
"step": 181924
},
{
"epoch": 4.0016710274620175,
"grad_norm": 0.4658304750919342,
"learning_rate": 5.989973835227898e-06,
"loss": 0.2447,
"step": 182000
},
{
"epoch": 4.012664629185814,
"grad_norm": 0.2597205340862274,
"learning_rate": 5.924012224885117e-06,
"loss": 0.1824,
"step": 182500
},
{
"epoch": 4.023658230909611,
"grad_norm": 0.4755733907222748,
"learning_rate": 5.858050614542337e-06,
"loss": 0.2266,
"step": 183000
},
{
"epoch": 4.034651832633408,
"grad_norm": 0.3996201753616333,
"learning_rate": 5.792089004199556e-06,
"loss": 0.1901,
"step": 183500
},
{
"epoch": 4.045645434357204,
"grad_norm": 0.8636412620544434,
"learning_rate": 5.726127393856775e-06,
"loss": 0.1955,
"step": 184000
},
{
"epoch": 4.056639036081001,
"grad_norm": 0.3265284597873688,
"learning_rate": 5.6601657835139945e-06,
"loss": 0.1979,
"step": 184500
},
{
"epoch": 4.067632637804798,
"grad_norm": 0.1188616007566452,
"learning_rate": 5.5942041731712145e-06,
"loss": 0.2066,
"step": 185000
},
{
"epoch": 4.078626239528594,
"grad_norm": 0.7546807527542114,
"learning_rate": 5.528242562828434e-06,
"loss": 0.1959,
"step": 185500
},
{
"epoch": 4.089619841252391,
"grad_norm": 0.12139397114515305,
"learning_rate": 5.462280952485654e-06,
"loss": 0.1969,
"step": 186000
},
{
"epoch": 4.100613442976188,
"grad_norm": 0.13494807481765747,
"learning_rate": 5.396319342142873e-06,
"loss": 0.1814,
"step": 186500
},
{
"epoch": 4.111607044699984,
"grad_norm": 0.023194080218672752,
"learning_rate": 5.330357731800093e-06,
"loss": 0.1877,
"step": 187000
},
{
"epoch": 4.122600646423781,
"grad_norm": 0.07898598164319992,
"learning_rate": 5.264396121457312e-06,
"loss": 0.209,
"step": 187500
},
{
"epoch": 4.1335942481475785,
"grad_norm": 0.11429109424352646,
"learning_rate": 5.198434511114532e-06,
"loss": 0.1794,
"step": 188000
},
{
"epoch": 4.144587849871375,
"grad_norm": 0.1126711368560791,
"learning_rate": 5.132472900771751e-06,
"loss": 0.2143,
"step": 188500
},
{
"epoch": 4.155581451595172,
"grad_norm": 36.669212341308594,
"learning_rate": 5.06651129042897e-06,
"loss": 0.1995,
"step": 189000
},
{
"epoch": 4.166575053318969,
"grad_norm": 0.0557066835463047,
"learning_rate": 5.0005496800861894e-06,
"loss": 0.1977,
"step": 189500
},
{
"epoch": 4.177568655042765,
"grad_norm": 0.11854979395866394,
"learning_rate": 4.9345880697434094e-06,
"loss": 0.203,
"step": 190000
},
{
"epoch": 4.188562256766562,
"grad_norm": 47.56736755371094,
"learning_rate": 4.868626459400629e-06,
"loss": 0.1673,
"step": 190500
},
{
"epoch": 4.199555858490359,
"grad_norm": 1.5440220832824707,
"learning_rate": 4.802664849057849e-06,
"loss": 0.1998,
"step": 191000
},
{
"epoch": 4.210549460214155,
"grad_norm": 0.11512400209903717,
"learning_rate": 4.736703238715068e-06,
"loss": 0.2027,
"step": 191500
},
{
"epoch": 4.221543061937952,
"grad_norm": 13.000309944152832,
"learning_rate": 4.670741628372288e-06,
"loss": 0.2268,
"step": 192000
},
{
"epoch": 4.232536663661749,
"grad_norm": 0.053511910140514374,
"learning_rate": 4.604780018029507e-06,
"loss": 0.2119,
"step": 192500
},
{
"epoch": 4.243530265385545,
"grad_norm": 1.0577130317687988,
"learning_rate": 4.538818407686727e-06,
"loss": 0.207,
"step": 193000
},
{
"epoch": 4.254523867109342,
"grad_norm": 0.5129163861274719,
"learning_rate": 4.472856797343946e-06,
"loss": 0.1971,
"step": 193500
},
{
"epoch": 4.265517468833139,
"grad_norm": 38.92678451538086,
"learning_rate": 4.406895187001165e-06,
"loss": 0.1915,
"step": 194000
},
{
"epoch": 4.276511070556936,
"grad_norm": 0.2508489489555359,
"learning_rate": 4.340933576658384e-06,
"loss": 0.2076,
"step": 194500
},
{
"epoch": 4.287504672280733,
"grad_norm": 0.8289797306060791,
"learning_rate": 4.274971966315604e-06,
"loss": 0.1906,
"step": 195000
},
{
"epoch": 4.29849827400453,
"grad_norm": 0.1511843502521515,
"learning_rate": 4.2090103559728235e-06,
"loss": 0.1908,
"step": 195500
},
{
"epoch": 4.309491875728326,
"grad_norm": 36.47195053100586,
"learning_rate": 4.1430487456300435e-06,
"loss": 0.2198,
"step": 196000
},
{
"epoch": 4.320485477452123,
"grad_norm": 59.39978790283203,
"learning_rate": 4.077087135287263e-06,
"loss": 0.1958,
"step": 196500
},
{
"epoch": 4.33147907917592,
"grad_norm": 25.194355010986328,
"learning_rate": 4.011125524944483e-06,
"loss": 0.185,
"step": 197000
},
{
"epoch": 4.342472680899716,
"grad_norm": 20.661163330078125,
"learning_rate": 3.945163914601702e-06,
"loss": 0.2032,
"step": 197500
},
{
"epoch": 4.353466282623513,
"grad_norm": 0.04815911129117012,
"learning_rate": 3.879202304258922e-06,
"loss": 0.194,
"step": 198000
},
{
"epoch": 4.36445988434731,
"grad_norm": 0.18730510771274567,
"learning_rate": 3.8132406939161414e-06,
"loss": 0.236,
"step": 198500
},
{
"epoch": 4.375453486071106,
"grad_norm": 15.972749710083008,
"learning_rate": 3.7472790835733606e-06,
"loss": 0.196,
"step": 199000
},
{
"epoch": 4.386447087794903,
"grad_norm": 0.25309285521507263,
"learning_rate": 3.68131747323058e-06,
"loss": 0.2161,
"step": 199500
},
{
"epoch": 4.3974406895187,
"grad_norm": 0.17074181139469147,
"learning_rate": 3.6153558628877997e-06,
"loss": 0.1836,
"step": 200000
},
{
"epoch": 4.4084342912424965,
"grad_norm": 20.413162231445312,
"learning_rate": 3.549394252545019e-06,
"loss": 0.1911,
"step": 200500
},
{
"epoch": 4.419427892966294,
"grad_norm": 0.5931562781333923,
"learning_rate": 3.4834326422022384e-06,
"loss": 0.1847,
"step": 201000
},
{
"epoch": 4.430421494690091,
"grad_norm": 37.10576248168945,
"learning_rate": 3.417471031859458e-06,
"loss": 0.1657,
"step": 201500
},
{
"epoch": 4.441415096413887,
"grad_norm": 0.06108024716377258,
"learning_rate": 3.3515094215166776e-06,
"loss": 0.1949,
"step": 202000
},
{
"epoch": 4.452408698137684,
"grad_norm": 14.7476224899292,
"learning_rate": 3.285547811173897e-06,
"loss": 0.2052,
"step": 202500
},
{
"epoch": 4.463402299861481,
"grad_norm": 0.4280465841293335,
"learning_rate": 3.2195862008311163e-06,
"loss": 0.2127,
"step": 203000
},
{
"epoch": 4.474395901585277,
"grad_norm": 0.38156208395957947,
"learning_rate": 3.153624590488336e-06,
"loss": 0.1949,
"step": 203500
},
{
"epoch": 4.485389503309074,
"grad_norm": 325.33026123046875,
"learning_rate": 3.0876629801455555e-06,
"loss": 0.1976,
"step": 204000
},
{
"epoch": 4.496383105032871,
"grad_norm": 99.75337982177734,
"learning_rate": 3.021701369802775e-06,
"loss": 0.2031,
"step": 204500
},
{
"epoch": 4.507376706756667,
"grad_norm": 0.17061945796012878,
"learning_rate": 2.9557397594599946e-06,
"loss": 0.1964,
"step": 205000
},
{
"epoch": 4.518370308480464,
"grad_norm": 25.07261085510254,
"learning_rate": 2.8897781491172138e-06,
"loss": 0.1875,
"step": 205500
},
{
"epoch": 4.529363910204261,
"grad_norm": 0.2692670226097107,
"learning_rate": 2.8238165387744334e-06,
"loss": 0.1978,
"step": 206000
},
{
"epoch": 4.5403575119280575,
"grad_norm": 2.56193470954895,
"learning_rate": 2.757854928431653e-06,
"loss": 0.207,
"step": 206500
},
{
"epoch": 4.5513511136518545,
"grad_norm": 1.1347905397415161,
"learning_rate": 2.6918933180888725e-06,
"loss": 0.2049,
"step": 207000
},
{
"epoch": 4.562344715375652,
"grad_norm": 0.9405034184455872,
"learning_rate": 2.625931707746092e-06,
"loss": 0.182,
"step": 207500
},
{
"epoch": 4.573338317099449,
"grad_norm": 0.10386385023593903,
"learning_rate": 2.5599700974033112e-06,
"loss": 0.1816,
"step": 208000
},
{
"epoch": 4.584331918823245,
"grad_norm": 1.0305184125900269,
"learning_rate": 2.494008487060531e-06,
"loss": 0.208,
"step": 208500
},
{
"epoch": 4.595325520547042,
"grad_norm": 5.062295913696289,
"learning_rate": 2.4280468767177504e-06,
"loss": 0.1918,
"step": 209000
},
{
"epoch": 4.606319122270838,
"grad_norm": 13.542932510375977,
"learning_rate": 2.36208526637497e-06,
"loss": 0.2016,
"step": 209500
},
{
"epoch": 4.617312723994635,
"grad_norm": 28.13912582397461,
"learning_rate": 2.2961236560321896e-06,
"loss": 0.2065,
"step": 210000
},
{
"epoch": 4.628306325718432,
"grad_norm": 38.89891052246094,
"learning_rate": 2.2301620456894087e-06,
"loss": 0.1951,
"step": 210500
},
{
"epoch": 4.639299927442229,
"grad_norm": 220.25010681152344,
"learning_rate": 2.1642004353466283e-06,
"loss": 0.2243,
"step": 211000
},
{
"epoch": 4.650293529166025,
"grad_norm": 0.11063925921916962,
"learning_rate": 2.098238825003848e-06,
"loss": 0.1862,
"step": 211500
},
{
"epoch": 4.661287130889822,
"grad_norm": 0.9656747579574585,
"learning_rate": 2.0322772146610674e-06,
"loss": 0.1796,
"step": 212000
},
{
"epoch": 4.672280732613618,
"grad_norm": 0.03588191047310829,
"learning_rate": 1.966315604318287e-06,
"loss": 0.1936,
"step": 212500
},
{
"epoch": 4.6832743343374155,
"grad_norm": 25.791149139404297,
"learning_rate": 1.9003539939755062e-06,
"loss": 0.2102,
"step": 213000
},
{
"epoch": 4.6942679360612125,
"grad_norm": 1.5398284196853638,
"learning_rate": 1.8343923836327257e-06,
"loss": 0.1941,
"step": 213500
},
{
"epoch": 4.7052615377850096,
"grad_norm": 0.8514572978019714,
"learning_rate": 1.7684307732899453e-06,
"loss": 0.2124,
"step": 214000
},
{
"epoch": 4.716255139508806,
"grad_norm": 0.1109534353017807,
"learning_rate": 1.7024691629471647e-06,
"loss": 0.1896,
"step": 214500
},
{
"epoch": 4.727248741232603,
"grad_norm": 33.977500915527344,
"learning_rate": 1.6365075526043843e-06,
"loss": 0.1868,
"step": 215000
},
{
"epoch": 4.738242342956399,
"grad_norm": 0.09221459925174713,
"learning_rate": 1.5705459422616038e-06,
"loss": 0.2147,
"step": 215500
},
{
"epoch": 4.749235944680196,
"grad_norm": 0.13753363490104675,
"learning_rate": 1.5045843319188232e-06,
"loss": 0.1805,
"step": 216000
},
{
"epoch": 4.760229546403993,
"grad_norm": 0.03300468996167183,
"learning_rate": 1.4386227215760428e-06,
"loss": 0.195,
"step": 216500
},
{
"epoch": 4.77122314812779,
"grad_norm": 52.13509750366211,
"learning_rate": 1.3726611112332624e-06,
"loss": 0.1934,
"step": 217000
},
{
"epoch": 4.782216749851586,
"grad_norm": 0.05201047658920288,
"learning_rate": 1.3066995008904817e-06,
"loss": 0.1986,
"step": 217500
},
{
"epoch": 4.793210351575383,
"grad_norm": 0.15796062350273132,
"learning_rate": 1.2407378905477013e-06,
"loss": 0.189,
"step": 218000
},
{
"epoch": 4.80420395329918,
"grad_norm": 0.603727400302887,
"learning_rate": 1.1747762802049207e-06,
"loss": 0.194,
"step": 218500
},
{
"epoch": 4.815197555022976,
"grad_norm": 19.412994384765625,
"learning_rate": 1.1088146698621402e-06,
"loss": 0.2022,
"step": 219000
},
{
"epoch": 4.8261911567467735,
"grad_norm": 0.08504907041788101,
"learning_rate": 1.0428530595193598e-06,
"loss": 0.1858,
"step": 219500
},
{
"epoch": 4.8371847584705705,
"grad_norm": 0.07863516360521317,
"learning_rate": 9.768914491765792e-07,
"loss": 0.1942,
"step": 220000
},
{
"epoch": 4.848178360194367,
"grad_norm": 23.51129722595215,
"learning_rate": 9.109298388337987e-07,
"loss": 0.1912,
"step": 220500
},
{
"epoch": 4.859171961918164,
"grad_norm": 5.780854225158691,
"learning_rate": 8.449682284910183e-07,
"loss": 0.2185,
"step": 221000
},
{
"epoch": 4.870165563641961,
"grad_norm": 0.05857408419251442,
"learning_rate": 7.790066181482378e-07,
"loss": 0.1914,
"step": 221500
},
{
"epoch": 4.881159165365757,
"grad_norm": 0.18249481916427612,
"learning_rate": 7.130450078054573e-07,
"loss": 0.1738,
"step": 222000
},
{
"epoch": 4.892152767089554,
"grad_norm": 418.6382751464844,
"learning_rate": 6.470833974626767e-07,
"loss": 0.2041,
"step": 222500
},
{
"epoch": 4.903146368813351,
"grad_norm": 0.7230046987533569,
"learning_rate": 5.811217871198962e-07,
"loss": 0.1957,
"step": 223000
},
{
"epoch": 4.914139970537147,
"grad_norm": 0.05270848050713539,
"learning_rate": 5.151601767771158e-07,
"loss": 0.1835,
"step": 223500
},
{
"epoch": 4.925133572260944,
"grad_norm": 1709.6539306640625,
"learning_rate": 4.491985664343352e-07,
"loss": 0.1921,
"step": 224000
},
{
"epoch": 4.936127173984741,
"grad_norm": 602.2431640625,
"learning_rate": 3.8323695609155474e-07,
"loss": 0.1727,
"step": 224500
},
{
"epoch": 4.947120775708537,
"grad_norm": 0.297931432723999,
"learning_rate": 3.1727534574877426e-07,
"loss": 0.1915,
"step": 225000
},
{
"epoch": 4.958114377432334,
"grad_norm": 0.3502364456653595,
"learning_rate": 2.5131373540599373e-07,
"loss": 0.1776,
"step": 225500
},
{
"epoch": 4.9691079791561314,
"grad_norm": 0.023652415722608566,
"learning_rate": 1.8535212506321323e-07,
"loss": 0.1978,
"step": 226000
},
{
"epoch": 4.980101580879928,
"grad_norm": 0.24965056777000427,
"learning_rate": 1.1939051472043273e-07,
"loss": 0.1924,
"step": 226500
},
{
"epoch": 4.991095182603725,
"grad_norm": 0.07366069406270981,
"learning_rate": 5.342890437765221e-08,
"loss": 0.1959,
"step": 227000
},
{
"epoch": 5.0,
"eval_accuracy": 0.8784565925598145,
"eval_f1": 0.8391068037456617,
"eval_loss": 0.6183628439903259,
"eval_mcc": 0.7416813202825935,
"eval_runtime": 27.9497,
"eval_samples_per_second": 723.264,
"eval_steps_per_second": 90.412,
"step": 227405
},
{
"epoch": 5.0,
"step": 227405,
"total_flos": 1.0733045580407808e+17,
"train_loss": 0.30088049875882883,
"train_runtime": 13193.5971,
"train_samples_per_second": 137.887,
"train_steps_per_second": 17.236
}
],
"logging_steps": 500,
"max_steps": 227405,
"num_input_tokens_seen": 0,
"num_train_epochs": 5,
"save_steps": 500,
"stateful_callbacks": {
"EarlyStoppingCallback": {
"args": {
"early_stopping_patience": 3,
"early_stopping_threshold": 0.001
},
"attributes": {
"early_stopping_patience_counter": 0
}
},
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1.0733045580407808e+17,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}