LLM / trainer_state.json
anthonymeo's picture
Upload 15 files
9715f5a verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 2.5188916876574305,
"eval_steps": 100,
"global_step": 1000,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.012594458438287154,
"grad_norm": 1.2670824527740479,
"learning_rate": 4.999782569758238e-05,
"loss": 2.5898,
"num_input_tokens_seen": 17064,
"step": 5
},
{
"epoch": 0.02518891687657431,
"grad_norm": 1.8668310642242432,
"learning_rate": 4.9991303168536793e-05,
"loss": 2.69,
"num_input_tokens_seen": 31136,
"step": 10
},
{
"epoch": 0.037783375314861464,
"grad_norm": 1.4463753700256348,
"learning_rate": 4.9980433547419305e-05,
"loss": 2.2158,
"num_input_tokens_seen": 47520,
"step": 15
},
{
"epoch": 0.05037783375314862,
"grad_norm": 1.9465786218643188,
"learning_rate": 4.996521872493738e-05,
"loss": 1.8421,
"num_input_tokens_seen": 66432,
"step": 20
},
{
"epoch": 0.06297229219143577,
"grad_norm": 1.3088130950927734,
"learning_rate": 4.994566134762105e-05,
"loss": 2.0699,
"num_input_tokens_seen": 83544,
"step": 25
},
{
"epoch": 0.07556675062972293,
"grad_norm": 1.5839314460754395,
"learning_rate": 4.992176481736254e-05,
"loss": 1.4037,
"num_input_tokens_seen": 97680,
"step": 30
},
{
"epoch": 0.08816120906801007,
"grad_norm": 2.81331729888916,
"learning_rate": 4.989353329082452e-05,
"loss": 1.8734,
"num_input_tokens_seen": 115456,
"step": 35
},
{
"epoch": 0.10075566750629723,
"grad_norm": 1.999861240386963,
"learning_rate": 4.986097167871711e-05,
"loss": 1.584,
"num_input_tokens_seen": 132872,
"step": 40
},
{
"epoch": 0.11335012594458438,
"grad_norm": 1.38965904712677,
"learning_rate": 4.982408564494367e-05,
"loss": 1.0849,
"num_input_tokens_seen": 148296,
"step": 45
},
{
"epoch": 0.12594458438287154,
"grad_norm": 1.410891056060791,
"learning_rate": 4.978288160561558e-05,
"loss": 1.0605,
"num_input_tokens_seen": 167504,
"step": 50
},
{
"epoch": 0.1385390428211587,
"grad_norm": 1.8659024238586426,
"learning_rate": 4.9737366727936235e-05,
"loss": 1.596,
"num_input_tokens_seen": 182536,
"step": 55
},
{
"epoch": 0.15113350125944586,
"grad_norm": 0.825343906879425,
"learning_rate": 4.968754892895432e-05,
"loss": 1.1978,
"num_input_tokens_seen": 201568,
"step": 60
},
{
"epoch": 0.163727959697733,
"grad_norm": 1.3539998531341553,
"learning_rate": 4.963343687418669e-05,
"loss": 1.1129,
"num_input_tokens_seen": 217584,
"step": 65
},
{
"epoch": 0.17632241813602015,
"grad_norm": 1.6213319301605225,
"learning_rate": 4.9575039976111084e-05,
"loss": 1.3955,
"num_input_tokens_seen": 235280,
"step": 70
},
{
"epoch": 0.1889168765743073,
"grad_norm": 1.7147520780563354,
"learning_rate": 4.9512368392528806e-05,
"loss": 1.029,
"num_input_tokens_seen": 253544,
"step": 75
},
{
"epoch": 0.20151133501259447,
"grad_norm": 1.0490851402282715,
"learning_rate": 4.9445433024797936e-05,
"loss": 1.1447,
"num_input_tokens_seen": 270112,
"step": 80
},
{
"epoch": 0.2141057934508816,
"grad_norm": 1.7510355710983276,
"learning_rate": 4.937424551593702e-05,
"loss": 1.0042,
"num_input_tokens_seen": 287296,
"step": 85
},
{
"epoch": 0.22670025188916876,
"grad_norm": 1.5315502882003784,
"learning_rate": 4.929881824859985e-05,
"loss": 1.3522,
"num_input_tokens_seen": 302304,
"step": 90
},
{
"epoch": 0.23929471032745592,
"grad_norm": 1.2641905546188354,
"learning_rate": 4.9219164342921634e-05,
"loss": 1.2398,
"num_input_tokens_seen": 317624,
"step": 95
},
{
"epoch": 0.2518891687657431,
"grad_norm": 1.116264820098877,
"learning_rate": 4.9135297654236724e-05,
"loss": 0.8733,
"num_input_tokens_seen": 332920,
"step": 100
},
{
"epoch": 0.2518891687657431,
"eval_accuracy": 0.7302086027292422,
"eval_loss": 1.238808035850525,
"eval_runtime": 533.8541,
"eval_samples_per_second": 0.332,
"eval_steps_per_second": 0.332,
"num_input_tokens_seen": 332920,
"step": 100
},
{
"epoch": 0.26448362720403024,
"grad_norm": 1.540872573852539,
"learning_rate": 4.904723277066864e-05,
"loss": 1.1192,
"num_input_tokens_seen": 347680,
"step": 105
},
{
"epoch": 0.2770780856423174,
"grad_norm": 1.2526932954788208,
"learning_rate": 4.8954985010592534e-05,
"loss": 1.1241,
"num_input_tokens_seen": 364952,
"step": 110
},
{
"epoch": 0.28967254408060455,
"grad_norm": 1.2870938777923584,
"learning_rate": 4.8858570419970616e-05,
"loss": 1.1762,
"num_input_tokens_seen": 383104,
"step": 115
},
{
"epoch": 0.3022670025188917,
"grad_norm": 1.6976910829544067,
"learning_rate": 4.875800576956108e-05,
"loss": 1.0041,
"num_input_tokens_seen": 401448,
"step": 120
},
{
"epoch": 0.3148614609571788,
"grad_norm": 1.2784565687179565,
"learning_rate": 4.865330855200094e-05,
"loss": 0.9044,
"num_input_tokens_seen": 419184,
"step": 125
},
{
"epoch": 0.327455919395466,
"grad_norm": 2.2473819255828857,
"learning_rate": 4.854449697876325e-05,
"loss": 1.5228,
"num_input_tokens_seen": 436976,
"step": 130
},
{
"epoch": 0.34005037783375314,
"grad_norm": 1.252365231513977,
"learning_rate": 4.843158997698936e-05,
"loss": 1.2336,
"num_input_tokens_seen": 455432,
"step": 135
},
{
"epoch": 0.3526448362720403,
"grad_norm": 1.5875591039657593,
"learning_rate": 4.831460718619661e-05,
"loss": 1.0085,
"num_input_tokens_seen": 473896,
"step": 140
},
{
"epoch": 0.36523929471032746,
"grad_norm": 1.8101682662963867,
"learning_rate": 4.819356895486219e-05,
"loss": 1.3333,
"num_input_tokens_seen": 488936,
"step": 145
},
{
"epoch": 0.3778337531486146,
"grad_norm": 1.2537903785705566,
"learning_rate": 4.806849633688363e-05,
"loss": 1.1225,
"num_input_tokens_seen": 507800,
"step": 150
},
{
"epoch": 0.3904282115869018,
"grad_norm": 1.2231327295303345,
"learning_rate": 4.7939411087916566e-05,
"loss": 1.0405,
"num_input_tokens_seen": 524704,
"step": 155
},
{
"epoch": 0.40302267002518893,
"grad_norm": 0.9885507822036743,
"learning_rate": 4.7806335661590526e-05,
"loss": 1.0268,
"num_input_tokens_seen": 542320,
"step": 160
},
{
"epoch": 0.4156171284634761,
"grad_norm": 1.273047685623169,
"learning_rate": 4.7669293205603196e-05,
"loss": 1.1071,
"num_input_tokens_seen": 560488,
"step": 165
},
{
"epoch": 0.4282115869017632,
"grad_norm": 1.8347108364105225,
"learning_rate": 4.752830755769405e-05,
"loss": 1.169,
"num_input_tokens_seen": 577680,
"step": 170
},
{
"epoch": 0.44080604534005036,
"grad_norm": 1.9854867458343506,
"learning_rate": 4.73834032414979e-05,
"loss": 0.9519,
"num_input_tokens_seen": 596208,
"step": 175
},
{
"epoch": 0.4534005037783375,
"grad_norm": 1.2936229705810547,
"learning_rate": 4.723460546227914e-05,
"loss": 1.2277,
"num_input_tokens_seen": 613120,
"step": 180
},
{
"epoch": 0.4659949622166247,
"grad_norm": 0.9850680232048035,
"learning_rate": 4.7081940102547463e-05,
"loss": 0.9588,
"num_input_tokens_seen": 630336,
"step": 185
},
{
"epoch": 0.47858942065491183,
"grad_norm": 1.643853783607483,
"learning_rate": 4.692543371755572e-05,
"loss": 1.0816,
"num_input_tokens_seen": 644488,
"step": 190
},
{
"epoch": 0.491183879093199,
"grad_norm": 1.7307960987091064,
"learning_rate": 4.6765113530680825e-05,
"loss": 0.8216,
"num_input_tokens_seen": 660432,
"step": 195
},
{
"epoch": 0.5037783375314862,
"grad_norm": 1.731389045715332,
"learning_rate": 4.660100742868836e-05,
"loss": 0.8872,
"num_input_tokens_seen": 680080,
"step": 200
},
{
"epoch": 0.5037783375314862,
"eval_accuracy": 0.7530999930068348,
"eval_loss": 1.129120111465454,
"eval_runtime": 534.7922,
"eval_samples_per_second": 0.331,
"eval_steps_per_second": 0.331,
"num_input_tokens_seen": 680080,
"step": 200
},
{
"epoch": 0.5163727959697733,
"grad_norm": 2.0154528617858887,
"learning_rate": 4.643314395688188e-05,
"loss": 1.0193,
"num_input_tokens_seen": 695688,
"step": 205
},
{
"epoch": 0.5289672544080605,
"grad_norm": 1.5947645902633667,
"learning_rate": 4.626155231413758e-05,
"loss": 1.1722,
"num_input_tokens_seen": 718192,
"step": 210
},
{
"epoch": 0.5415617128463476,
"grad_norm": 1.233261227607727,
"learning_rate": 4.608626234782536e-05,
"loss": 0.9888,
"num_input_tokens_seen": 733136,
"step": 215
},
{
"epoch": 0.5541561712846348,
"grad_norm": 2.006932497024536,
"learning_rate": 4.5907304548617024e-05,
"loss": 1.018,
"num_input_tokens_seen": 748016,
"step": 220
},
{
"epoch": 0.5667506297229219,
"grad_norm": 1.6554147005081177,
"learning_rate": 4.572471004518261e-05,
"loss": 0.8822,
"num_input_tokens_seen": 763264,
"step": 225
},
{
"epoch": 0.5793450881612091,
"grad_norm": 2.6174421310424805,
"learning_rate": 4.553851059877573e-05,
"loss": 1.1984,
"num_input_tokens_seen": 778632,
"step": 230
},
{
"epoch": 0.5919395465994962,
"grad_norm": 1.639770269393921,
"learning_rate": 4.534873859770892e-05,
"loss": 1.0492,
"num_input_tokens_seen": 799096,
"step": 235
},
{
"epoch": 0.6045340050377834,
"grad_norm": 1.863054871559143,
"learning_rate": 4.515542705171981e-05,
"loss": 0.7339,
"num_input_tokens_seen": 815240,
"step": 240
},
{
"epoch": 0.6171284634760705,
"grad_norm": 2.6653268337249756,
"learning_rate": 4.495860958622937e-05,
"loss": 1.2535,
"num_input_tokens_seen": 831320,
"step": 245
},
{
"epoch": 0.6297229219143576,
"grad_norm": 3.9443094730377197,
"learning_rate": 4.475832043649287e-05,
"loss": 1.883,
"num_input_tokens_seen": 850376,
"step": 250
},
{
"epoch": 0.6423173803526449,
"grad_norm": 3.2649178504943848,
"learning_rate": 4.455459444164492e-05,
"loss": 0.9972,
"num_input_tokens_seen": 868192,
"step": 255
},
{
"epoch": 0.654911838790932,
"grad_norm": 2.0120456218719482,
"learning_rate": 4.4347467038639364e-05,
"loss": 1.0848,
"num_input_tokens_seen": 883560,
"step": 260
},
{
"epoch": 0.6675062972292192,
"grad_norm": 1.7830870151519775,
"learning_rate": 4.4136974256085236e-05,
"loss": 0.8549,
"num_input_tokens_seen": 902952,
"step": 265
},
{
"epoch": 0.6801007556675063,
"grad_norm": 1.284114956855774,
"learning_rate": 4.392315270797985e-05,
"loss": 0.5925,
"num_input_tokens_seen": 919584,
"step": 270
},
{
"epoch": 0.6926952141057935,
"grad_norm": 1.58772873878479,
"learning_rate": 4.3706039587339894e-05,
"loss": 1.2722,
"num_input_tokens_seen": 940200,
"step": 275
},
{
"epoch": 0.7052896725440806,
"grad_norm": 1.5675506591796875,
"learning_rate": 4.3485672659732034e-05,
"loss": 0.9741,
"num_input_tokens_seen": 961256,
"step": 280
},
{
"epoch": 0.7178841309823678,
"grad_norm": 1.5801304578781128,
"learning_rate": 4.3262090256703736e-05,
"loss": 0.9787,
"num_input_tokens_seen": 978000,
"step": 285
},
{
"epoch": 0.7304785894206549,
"grad_norm": 1.2644524574279785,
"learning_rate": 4.303533126911577e-05,
"loss": 1.1364,
"num_input_tokens_seen": 997512,
"step": 290
},
{
"epoch": 0.743073047858942,
"grad_norm": 1.307681918144226,
"learning_rate": 4.280543514037731e-05,
"loss": 1.1322,
"num_input_tokens_seen": 1016824,
"step": 295
},
{
"epoch": 0.7556675062972292,
"grad_norm": 1.8267008066177368,
"learning_rate": 4.257244185958505e-05,
"loss": 1.0074,
"num_input_tokens_seen": 1036168,
"step": 300
},
{
"epoch": 0.7556675062972292,
"eval_accuracy": 0.7609178510672793,
"eval_loss": 1.095440149307251,
"eval_runtime": 537.6469,
"eval_samples_per_second": 0.329,
"eval_steps_per_second": 0.329,
"num_input_tokens_seen": 1036168,
"step": 300
},
{
"epoch": 0.7682619647355163,
"grad_norm": 1.6457531452178955,
"learning_rate": 4.233639195456729e-05,
"loss": 0.939,
"num_input_tokens_seen": 1053264,
"step": 305
},
{
"epoch": 0.7808564231738035,
"grad_norm": 1.752995252609253,
"learning_rate": 4.2097326484834346e-05,
"loss": 1.0468,
"num_input_tokens_seen": 1068696,
"step": 310
},
{
"epoch": 0.7934508816120907,
"grad_norm": 1.001474142074585,
"learning_rate": 4.1855287034436555e-05,
"loss": 0.8325,
"num_input_tokens_seen": 1085264,
"step": 315
},
{
"epoch": 0.8060453400503779,
"grad_norm": 1.4869558811187744,
"learning_rate": 4.1610315704730854e-05,
"loss": 0.8035,
"num_input_tokens_seen": 1102368,
"step": 320
},
{
"epoch": 0.818639798488665,
"grad_norm": 1.8002692461013794,
"learning_rate": 4.136245510705762e-05,
"loss": 1.0207,
"num_input_tokens_seen": 1117768,
"step": 325
},
{
"epoch": 0.8312342569269522,
"grad_norm": 1.747013807296753,
"learning_rate": 4.111174835532857e-05,
"loss": 1.2914,
"num_input_tokens_seen": 1133440,
"step": 330
},
{
"epoch": 0.8438287153652393,
"grad_norm": 1.6267316341400146,
"learning_rate": 4.085823905852745e-05,
"loss": 1.2979,
"num_input_tokens_seen": 1146480,
"step": 335
},
{
"epoch": 0.8564231738035264,
"grad_norm": 1.2453759908676147,
"learning_rate": 4.06019713131244e-05,
"loss": 0.6644,
"num_input_tokens_seen": 1162248,
"step": 340
},
{
"epoch": 0.8690176322418136,
"grad_norm": 0.9214743971824646,
"learning_rate": 4.034298969540567e-05,
"loss": 1.1669,
"num_input_tokens_seen": 1179224,
"step": 345
},
{
"epoch": 0.8816120906801007,
"grad_norm": 1.0969208478927612,
"learning_rate": 4.008133925371988e-05,
"loss": 1.2072,
"num_input_tokens_seen": 1195240,
"step": 350
},
{
"epoch": 0.8942065491183879,
"grad_norm": 1.6422204971313477,
"learning_rate": 3.981706550064208e-05,
"loss": 0.9078,
"num_input_tokens_seen": 1213056,
"step": 355
},
{
"epoch": 0.906801007556675,
"grad_norm": 1.5316386222839355,
"learning_rate": 3.955021440505706e-05,
"loss": 0.4814,
"num_input_tokens_seen": 1230744,
"step": 360
},
{
"epoch": 0.9193954659949622,
"grad_norm": 1.2854864597320557,
"learning_rate": 3.928083238416342e-05,
"loss": 0.9272,
"num_input_tokens_seen": 1246904,
"step": 365
},
{
"epoch": 0.9319899244332494,
"grad_norm": 1.3557329177856445,
"learning_rate": 3.9008966295399494e-05,
"loss": 0.9995,
"num_input_tokens_seen": 1263880,
"step": 370
},
{
"epoch": 0.9445843828715366,
"grad_norm": 1.2092177867889404,
"learning_rate": 3.873466342829281e-05,
"loss": 0.8696,
"num_input_tokens_seen": 1283816,
"step": 375
},
{
"epoch": 0.9571788413098237,
"grad_norm": 2.7501771450042725,
"learning_rate": 3.845797149623434e-05,
"loss": 1.4119,
"num_input_tokens_seen": 1300192,
"step": 380
},
{
"epoch": 0.9697732997481109,
"grad_norm": 1.222266435623169,
"learning_rate": 3.817893862817902e-05,
"loss": 0.8804,
"num_input_tokens_seen": 1317224,
"step": 385
},
{
"epoch": 0.982367758186398,
"grad_norm": 0.6030636429786682,
"learning_rate": 3.789761336027403e-05,
"loss": 0.5944,
"num_input_tokens_seen": 1335296,
"step": 390
},
{
"epoch": 0.9949622166246851,
"grad_norm": 1.5092536211013794,
"learning_rate": 3.761404462741618e-05,
"loss": 1.0303,
"num_input_tokens_seen": 1351920,
"step": 395
},
{
"epoch": 1.0075566750629723,
"grad_norm": 2.0633888244628906,
"learning_rate": 3.7328281754739974e-05,
"loss": 0.9671,
"num_input_tokens_seen": 1370864,
"step": 400
},
{
"epoch": 1.0075566750629723,
"eval_accuracy": 0.7659399606125864,
"eval_loss": 1.0698517560958862,
"eval_runtime": 535.9639,
"eval_samples_per_second": 0.33,
"eval_steps_per_second": 0.33,
"num_input_tokens_seen": 1370864,
"step": 400
},
{
"epoch": 1.0201511335012594,
"grad_norm": 1.7792761325836182,
"learning_rate": 3.704037444903782e-05,
"loss": 1.0106,
"num_input_tokens_seen": 1390136,
"step": 405
},
{
"epoch": 1.0327455919395465,
"grad_norm": 1.8169097900390625,
"learning_rate": 3.6750372790113766e-05,
"loss": 0.7452,
"num_input_tokens_seen": 1411432,
"step": 410
},
{
"epoch": 1.0453400503778338,
"grad_norm": 1.3711882829666138,
"learning_rate": 3.645832722207248e-05,
"loss": 0.9704,
"num_input_tokens_seen": 1429024,
"step": 415
},
{
"epoch": 1.057934508816121,
"grad_norm": 0.9055928587913513,
"learning_rate": 3.6164288544544725e-05,
"loss": 0.5268,
"num_input_tokens_seen": 1445848,
"step": 420
},
{
"epoch": 1.070528967254408,
"grad_norm": 1.2504442930221558,
"learning_rate": 3.586830790385109e-05,
"loss": 0.6362,
"num_input_tokens_seen": 1463232,
"step": 425
},
{
"epoch": 1.0831234256926952,
"grad_norm": 2.108982563018799,
"learning_rate": 3.55704367841054e-05,
"loss": 0.5694,
"num_input_tokens_seen": 1478584,
"step": 430
},
{
"epoch": 1.0957178841309823,
"grad_norm": 2.6890852451324463,
"learning_rate": 3.52707269982593e-05,
"loss": 0.5836,
"num_input_tokens_seen": 1495112,
"step": 435
},
{
"epoch": 1.1083123425692696,
"grad_norm": 3.2210803031921387,
"learning_rate": 3.496923067908977e-05,
"loss": 1.0356,
"num_input_tokens_seen": 1513000,
"step": 440
},
{
"epoch": 1.1209068010075567,
"grad_norm": 1.5383672714233398,
"learning_rate": 3.466600027013084e-05,
"loss": 1.0125,
"num_input_tokens_seen": 1526896,
"step": 445
},
{
"epoch": 1.1335012594458438,
"grad_norm": 2.0438127517700195,
"learning_rate": 3.436108851655143e-05,
"loss": 1.0554,
"num_input_tokens_seen": 1542448,
"step": 450
},
{
"epoch": 1.146095717884131,
"grad_norm": 2.1045315265655518,
"learning_rate": 3.4054548455980565e-05,
"loss": 0.714,
"num_input_tokens_seen": 1557656,
"step": 455
},
{
"epoch": 1.1586901763224182,
"grad_norm": 2.4777379035949707,
"learning_rate": 3.3746433409281844e-05,
"loss": 0.8676,
"num_input_tokens_seen": 1575192,
"step": 460
},
{
"epoch": 1.1712846347607053,
"grad_norm": 2.529090642929077,
"learning_rate": 3.3436796971278526e-05,
"loss": 0.6624,
"num_input_tokens_seen": 1596112,
"step": 465
},
{
"epoch": 1.1838790931989924,
"grad_norm": 2.93548846244812,
"learning_rate": 3.312569300143108e-05,
"loss": 0.795,
"num_input_tokens_seen": 1610768,
"step": 470
},
{
"epoch": 1.1964735516372795,
"grad_norm": 2.8596348762512207,
"learning_rate": 3.2813175614468604e-05,
"loss": 1.3433,
"num_input_tokens_seen": 1627672,
"step": 475
},
{
"epoch": 1.2090680100755669,
"grad_norm": 3.336879253387451,
"learning_rate": 3.24992991709759e-05,
"loss": 1.0594,
"num_input_tokens_seen": 1641912,
"step": 480
},
{
"epoch": 1.221662468513854,
"grad_norm": 2.030346155166626,
"learning_rate": 3.218411826793777e-05,
"loss": 0.972,
"num_input_tokens_seen": 1659832,
"step": 485
},
{
"epoch": 1.234256926952141,
"grad_norm": 2.6392228603363037,
"learning_rate": 3.186768772924216e-05,
"loss": 1.032,
"num_input_tokens_seen": 1679064,
"step": 490
},
{
"epoch": 1.2468513853904282,
"grad_norm": 2.443204402923584,
"learning_rate": 3.1550062596143886e-05,
"loss": 1.3751,
"num_input_tokens_seen": 1692608,
"step": 495
},
{
"epoch": 1.2594458438287153,
"grad_norm": 1.4907230138778687,
"learning_rate": 3.1231298117690554e-05,
"loss": 0.7884,
"num_input_tokens_seen": 1709712,
"step": 500
},
{
"epoch": 1.2594458438287153,
"eval_accuracy": 0.7672396326295029,
"eval_loss": 1.0675625801086426,
"eval_runtime": 535.3969,
"eval_samples_per_second": 0.331,
"eval_steps_per_second": 0.331,
"num_input_tokens_seen": 1709712,
"step": 500
},
{
"epoch": 1.2720403022670026,
"grad_norm": 0.9061765074729919,
"learning_rate": 3.091144974111224e-05,
"loss": 0.781,
"num_input_tokens_seen": 1729424,
"step": 505
},
{
"epoch": 1.2846347607052897,
"grad_norm": 2.395761251449585,
"learning_rate": 3.059057310217683e-05,
"loss": 0.8264,
"num_input_tokens_seen": 1749136,
"step": 510
},
{
"epoch": 1.2972292191435768,
"grad_norm": 2.069194793701172,
"learning_rate": 3.0268724015512463e-05,
"loss": 0.6579,
"num_input_tokens_seen": 1765216,
"step": 515
},
{
"epoch": 1.309823677581864,
"grad_norm": 1.5534199476242065,
"learning_rate": 2.994595846489892e-05,
"loss": 0.7195,
"num_input_tokens_seen": 1781320,
"step": 520
},
{
"epoch": 1.322418136020151,
"grad_norm": 1.4709012508392334,
"learning_rate": 2.9622332593529563e-05,
"loss": 0.5226,
"num_input_tokens_seen": 1797760,
"step": 525
},
{
"epoch": 1.3350125944584383,
"grad_norm": 2.4601175785064697,
"learning_rate": 2.9297902694245542e-05,
"loss": 1.2005,
"num_input_tokens_seen": 1813224,
"step": 530
},
{
"epoch": 1.3476070528967254,
"grad_norm": 2.9571943283081055,
"learning_rate": 2.8972725199744033e-05,
"loss": 0.7554,
"num_input_tokens_seen": 1830728,
"step": 535
},
{
"epoch": 1.3602015113350125,
"grad_norm": 1.610404372215271,
"learning_rate": 2.864685667276201e-05,
"loss": 1.0766,
"num_input_tokens_seen": 1848816,
"step": 540
},
{
"epoch": 1.3727959697732999,
"grad_norm": 2.7103452682495117,
"learning_rate": 2.8320353796237553e-05,
"loss": 0.8778,
"num_input_tokens_seen": 1863224,
"step": 545
},
{
"epoch": 1.385390428211587,
"grad_norm": 3.026928424835205,
"learning_rate": 2.7993273363450184e-05,
"loss": 0.6799,
"num_input_tokens_seen": 1880288,
"step": 550
},
{
"epoch": 1.397984886649874,
"grad_norm": 1.5982580184936523,
"learning_rate": 2.7665672268141956e-05,
"loss": 0.4951,
"num_input_tokens_seen": 1896552,
"step": 555
},
{
"epoch": 1.4105793450881612,
"grad_norm": 2.5049729347229004,
"learning_rate": 2.7337607494621152e-05,
"loss": 0.9428,
"num_input_tokens_seen": 1915872,
"step": 560
},
{
"epoch": 1.4231738035264483,
"grad_norm": 2.13607120513916,
"learning_rate": 2.7009136107850185e-05,
"loss": 0.8936,
"num_input_tokens_seen": 1934704,
"step": 565
},
{
"epoch": 1.4357682619647356,
"grad_norm": 3.7449216842651367,
"learning_rate": 2.668031524351949e-05,
"loss": 0.8481,
"num_input_tokens_seen": 1951816,
"step": 570
},
{
"epoch": 1.4483627204030227,
"grad_norm": 2.881800651550293,
"learning_rate": 2.6351202098109083e-05,
"loss": 1.1778,
"num_input_tokens_seen": 1970640,
"step": 575
},
{
"epoch": 1.4609571788413098,
"grad_norm": 3.845482110977173,
"learning_rate": 2.6021853918939587e-05,
"loss": 0.7675,
"num_input_tokens_seen": 1986504,
"step": 580
},
{
"epoch": 1.473551637279597,
"grad_norm": 2.6726672649383545,
"learning_rate": 2.5692327994214383e-05,
"loss": 0.8112,
"num_input_tokens_seen": 2003440,
"step": 585
},
{
"epoch": 1.486146095717884,
"grad_norm": 3.1416702270507812,
"learning_rate": 2.536268164305465e-05,
"loss": 1.3447,
"num_input_tokens_seen": 2020568,
"step": 590
},
{
"epoch": 1.4987405541561714,
"grad_norm": 1.6394110918045044,
"learning_rate": 2.5032972205529044e-05,
"loss": 0.9512,
"num_input_tokens_seen": 2037096,
"step": 595
},
{
"epoch": 1.5113350125944585,
"grad_norm": 4.283825874328613,
"learning_rate": 2.4703257032679744e-05,
"loss": 1.0526,
"num_input_tokens_seen": 2053296,
"step": 600
},
{
"epoch": 1.5113350125944585,
"eval_accuracy": 0.7689241678712095,
"eval_loss": 1.0595225095748901,
"eval_runtime": 536.6614,
"eval_samples_per_second": 0.33,
"eval_steps_per_second": 0.33,
"num_input_tokens_seen": 2053296,
"step": 600
},
{
"epoch": 1.5239294710327456,
"grad_norm": 2.488013505935669,
"learning_rate": 2.437359347654655e-05,
"loss": 0.7499,
"num_input_tokens_seen": 2070840,
"step": 605
},
{
"epoch": 1.536523929471033,
"grad_norm": 2.6227219104766846,
"learning_rate": 2.4044038880190824e-05,
"loss": 0.6816,
"num_input_tokens_seen": 2087952,
"step": 610
},
{
"epoch": 1.5491183879093198,
"grad_norm": 1.874306082725525,
"learning_rate": 2.3714650567721016e-05,
"loss": 0.7719,
"num_input_tokens_seen": 2103824,
"step": 615
},
{
"epoch": 1.561712846347607,
"grad_norm": 2.8610432147979736,
"learning_rate": 2.338548583432144e-05,
"loss": 1.1229,
"num_input_tokens_seen": 2120240,
"step": 620
},
{
"epoch": 1.5743073047858942,
"grad_norm": 6.437571048736572,
"learning_rate": 2.305660193628618e-05,
"loss": 1.1712,
"num_input_tokens_seen": 2135416,
"step": 625
},
{
"epoch": 1.5869017632241813,
"grad_norm": 2.076413154602051,
"learning_rate": 2.272805608105958e-05,
"loss": 0.6688,
"num_input_tokens_seen": 2151904,
"step": 630
},
{
"epoch": 1.5994962216624686,
"grad_norm": 2.047494888305664,
"learning_rate": 2.2399905417285434e-05,
"loss": 0.8043,
"num_input_tokens_seen": 2168952,
"step": 635
},
{
"epoch": 1.6120906801007555,
"grad_norm": 2.4243392944335938,
"learning_rate": 2.2072207024866266e-05,
"loss": 0.5582,
"num_input_tokens_seen": 2185192,
"step": 640
},
{
"epoch": 1.6246851385390428,
"grad_norm": 2.66733717918396,
"learning_rate": 2.1745017905034625e-05,
"loss": 1.1033,
"num_input_tokens_seen": 2200856,
"step": 645
},
{
"epoch": 1.63727959697733,
"grad_norm": 2.6124303340911865,
"learning_rate": 2.141839497043806e-05,
"loss": 0.8529,
"num_input_tokens_seen": 2215080,
"step": 650
},
{
"epoch": 1.649874055415617,
"grad_norm": 2.1553502082824707,
"learning_rate": 2.1092395035239472e-05,
"loss": 0.7331,
"num_input_tokens_seen": 2229808,
"step": 655
},
{
"epoch": 1.6624685138539044,
"grad_norm": 2.305765390396118,
"learning_rate": 2.076707480523464e-05,
"loss": 0.9966,
"num_input_tokens_seen": 2247584,
"step": 660
},
{
"epoch": 1.6750629722921915,
"grad_norm": 1.9919112920761108,
"learning_rate": 2.0442490867988582e-05,
"loss": 1.0719,
"num_input_tokens_seen": 2264280,
"step": 665
},
{
"epoch": 1.6876574307304786,
"grad_norm": 2.5872271060943604,
"learning_rate": 2.011869968299245e-05,
"loss": 0.8667,
"num_input_tokens_seen": 2281600,
"step": 670
},
{
"epoch": 1.700251889168766,
"grad_norm": 2.303976535797119,
"learning_rate": 1.9795757571842744e-05,
"loss": 0.559,
"num_input_tokens_seen": 2298688,
"step": 675
},
{
"epoch": 1.7128463476070528,
"grad_norm": 2.5704333782196045,
"learning_rate": 1.947372070844452e-05,
"loss": 0.899,
"num_input_tokens_seen": 2318312,
"step": 680
},
{
"epoch": 1.7254408060453401,
"grad_norm": 2.638066530227661,
"learning_rate": 1.915264510924022e-05,
"loss": 0.7052,
"num_input_tokens_seen": 2334496,
"step": 685
},
{
"epoch": 1.7380352644836272,
"grad_norm": 2.046243906021118,
"learning_rate": 1.883258662346596e-05,
"loss": 0.9922,
"num_input_tokens_seen": 2355640,
"step": 690
},
{
"epoch": 1.7506297229219143,
"grad_norm": 2.731766700744629,
"learning_rate": 1.8513600923436923e-05,
"loss": 0.9633,
"num_input_tokens_seen": 2375368,
"step": 695
},
{
"epoch": 1.7632241813602016,
"grad_norm": 4.190962791442871,
"learning_rate": 1.8195743494863387e-05,
"loss": 1.2255,
"num_input_tokens_seen": 2392384,
"step": 700
},
{
"epoch": 1.7632241813602016,
"eval_accuracy": 0.7701849009613134,
"eval_loss": 1.0550936460494995,
"eval_runtime": 535.3581,
"eval_samples_per_second": 0.331,
"eval_steps_per_second": 0.331,
"num_input_tokens_seen": 2392384,
"step": 700
},
{
"epoch": 1.7758186397984885,
"grad_norm": 6.717029094696045,
"learning_rate": 1.787906962719939e-05,
"loss": 0.7964,
"num_input_tokens_seen": 2409672,
"step": 705
},
{
"epoch": 1.7884130982367759,
"grad_norm": 2.1568119525909424,
"learning_rate": 1.7563634404025414e-05,
"loss": 0.7568,
"num_input_tokens_seen": 2425144,
"step": 710
},
{
"epoch": 1.801007556675063,
"grad_norm": 2.602142810821533,
"learning_rate": 1.7249492693466934e-05,
"loss": 0.739,
"num_input_tokens_seen": 2445872,
"step": 715
},
{
"epoch": 1.81360201511335,
"grad_norm": 2.0826470851898193,
"learning_rate": 1.6936699138650397e-05,
"loss": 0.7168,
"num_input_tokens_seen": 2463232,
"step": 720
},
{
"epoch": 1.8261964735516374,
"grad_norm": 3.3417489528656006,
"learning_rate": 1.6625308148198413e-05,
"loss": 0.7757,
"num_input_tokens_seen": 2480816,
"step": 725
},
{
"epoch": 1.8387909319899243,
"grad_norm": 2.6503310203552246,
"learning_rate": 1.6315373886765646e-05,
"loss": 0.6779,
"num_input_tokens_seen": 2498488,
"step": 730
},
{
"epoch": 1.8513853904282116,
"grad_norm": 3.9675190448760986,
"learning_rate": 1.600695026561721e-05,
"loss": 0.9367,
"num_input_tokens_seen": 2516792,
"step": 735
},
{
"epoch": 1.8639798488664987,
"grad_norm": 2.195193290710449,
"learning_rate": 1.5700090933251115e-05,
"loss": 0.447,
"num_input_tokens_seen": 2533232,
"step": 740
},
{
"epoch": 1.8765743073047858,
"grad_norm": 2.9985008239746094,
"learning_rate": 1.5394849266066416e-05,
"loss": 0.6294,
"num_input_tokens_seen": 2552824,
"step": 745
},
{
"epoch": 1.8891687657430731,
"grad_norm": 3.8788790702819824,
"learning_rate": 1.509127835907872e-05,
"loss": 1.3992,
"num_input_tokens_seen": 2571512,
"step": 750
},
{
"epoch": 1.9017632241813602,
"grad_norm": 1.619795799255371,
"learning_rate": 1.4789431016684558e-05,
"loss": 0.8268,
"num_input_tokens_seen": 2588120,
"step": 755
},
{
"epoch": 1.9143576826196473,
"grad_norm": 3.272700786590576,
"learning_rate": 1.4489359743476461e-05,
"loss": 0.6765,
"num_input_tokens_seen": 2605248,
"step": 760
},
{
"epoch": 1.9269521410579347,
"grad_norm": 2.847982168197632,
"learning_rate": 1.4191116735110007e-05,
"loss": 1.1278,
"num_input_tokens_seen": 2621656,
"step": 765
},
{
"epoch": 1.9395465994962215,
"grad_norm": 3.8507871627807617,
"learning_rate": 1.3894753869224725e-05,
"loss": 0.6863,
"num_input_tokens_seen": 2639440,
"step": 770
},
{
"epoch": 1.9521410579345089,
"grad_norm": 2.1526479721069336,
"learning_rate": 1.3600322696420275e-05,
"loss": 0.6884,
"num_input_tokens_seen": 2657952,
"step": 775
},
{
"epoch": 1.964735516372796,
"grad_norm": 2.447094202041626,
"learning_rate": 1.330787443128953e-05,
"loss": 0.6405,
"num_input_tokens_seen": 2673752,
"step": 780
},
{
"epoch": 1.977329974811083,
"grad_norm": 1.717159628868103,
"learning_rate": 1.3017459943510084e-05,
"loss": 0.6037,
"num_input_tokens_seen": 2689440,
"step": 785
},
{
"epoch": 1.9899244332493704,
"grad_norm": 3.210221529006958,
"learning_rate": 1.2729129748995749e-05,
"loss": 0.9172,
"num_input_tokens_seen": 2706624,
"step": 790
},
{
"epoch": 2.0025188916876573,
"grad_norm": 1.6862154006958008,
"learning_rate": 1.2442934001109671e-05,
"loss": 0.7949,
"num_input_tokens_seen": 2723496,
"step": 795
},
{
"epoch": 2.0151133501259446,
"grad_norm": 2.2537808418273926,
"learning_rate": 1.2158922481940361e-05,
"loss": 0.5226,
"num_input_tokens_seen": 2738304,
"step": 800
},
{
"epoch": 2.0151133501259446,
"eval_accuracy": 0.7704949500614456,
"eval_loss": 1.0528764724731445,
"eval_runtime": 535.8681,
"eval_samples_per_second": 0.33,
"eval_steps_per_second": 0.33,
"num_input_tokens_seen": 2738304,
"step": 800
},
{
"epoch": 2.027707808564232,
"grad_norm": 5.461361408233643,
"learning_rate": 1.1877144593642439e-05,
"loss": 0.9266,
"num_input_tokens_seen": 2755120,
"step": 805
},
{
"epoch": 2.040302267002519,
"grad_norm": 2.930683135986328,
"learning_rate": 1.1597649349843413e-05,
"loss": 0.9805,
"num_input_tokens_seen": 2771080,
"step": 810
},
{
"epoch": 2.052896725440806,
"grad_norm": 3.590545654296875,
"learning_rate": 1.1320485367118017e-05,
"loss": 1.1796,
"num_input_tokens_seen": 2786752,
"step": 815
},
{
"epoch": 2.065491183879093,
"grad_norm": 2.6750876903533936,
"learning_rate": 1.1045700856531668e-05,
"loss": 0.4839,
"num_input_tokens_seen": 2804392,
"step": 820
},
{
"epoch": 2.0780856423173804,
"grad_norm": 1.7390782833099365,
"learning_rate": 1.0773343615254446e-05,
"loss": 0.359,
"num_input_tokens_seen": 2818640,
"step": 825
},
{
"epoch": 2.0906801007556677,
"grad_norm": 3.9132091999053955,
"learning_rate": 1.0503461018246977e-05,
"loss": 1.0472,
"num_input_tokens_seen": 2836256,
"step": 830
},
{
"epoch": 2.1032745591939546,
"grad_norm": 3.056783437728882,
"learning_rate": 1.0236100010019919e-05,
"loss": 0.8781,
"num_input_tokens_seen": 2855496,
"step": 835
},
{
"epoch": 2.115869017632242,
"grad_norm": 3.0513293743133545,
"learning_rate": 9.971307096468203e-06,
"loss": 0.9041,
"num_input_tokens_seen": 2872200,
"step": 840
},
{
"epoch": 2.1284634760705288,
"grad_norm": 2.966620445251465,
"learning_rate": 9.709128336781592e-06,
"loss": 0.7187,
"num_input_tokens_seen": 2888256,
"step": 845
},
{
"epoch": 2.141057934508816,
"grad_norm": 2.2629482746124268,
"learning_rate": 9.449609335432972e-06,
"loss": 0.8696,
"num_input_tokens_seen": 2905920,
"step": 850
},
{
"epoch": 2.1536523929471034,
"grad_norm": 5.053290367126465,
"learning_rate": 9.192795234245697e-06,
"loss": 0.6862,
"num_input_tokens_seen": 2921056,
"step": 855
},
{
"epoch": 2.1662468513853903,
"grad_norm": 2.8935582637786865,
"learning_rate": 8.938730704541473e-06,
"loss": 0.702,
"num_input_tokens_seen": 2942096,
"step": 860
},
{
"epoch": 2.1788413098236776,
"grad_norm": 4.863999843597412,
"learning_rate": 8.687459939369983e-06,
"loss": 0.9868,
"num_input_tokens_seen": 2959320,
"step": 865
},
{
"epoch": 2.1914357682619645,
"grad_norm": 2.9665706157684326,
"learning_rate": 8.439026645821802e-06,
"loss": 0.5647,
"num_input_tokens_seen": 2976800,
"step": 870
},
{
"epoch": 2.204030226700252,
"grad_norm": 4.912269115447998,
"learning_rate": 8.193474037425794e-06,
"loss": 0.7983,
"num_input_tokens_seen": 2994536,
"step": 875
},
{
"epoch": 2.216624685138539,
"grad_norm": 3.118041515350342,
"learning_rate": 7.950844826632373e-06,
"loss": 0.9227,
"num_input_tokens_seen": 3010504,
"step": 880
},
{
"epoch": 2.229219143576826,
"grad_norm": 1.5938191413879395,
"learning_rate": 7.711181217383896e-06,
"loss": 0.9203,
"num_input_tokens_seen": 3027240,
"step": 885
},
{
"epoch": 2.2418136020151134,
"grad_norm": 2.1337177753448486,
"learning_rate": 7.474524897773555e-06,
"loss": 0.5222,
"num_input_tokens_seen": 3044024,
"step": 890
},
{
"epoch": 2.2544080604534007,
"grad_norm": 3.2635252475738525,
"learning_rate": 7.240917032794003e-06,
"loss": 0.5499,
"num_input_tokens_seen": 3060632,
"step": 895
},
{
"epoch": 2.2670025188916876,
"grad_norm": 2.3565027713775635,
"learning_rate": 7.010398257176878e-06,
"loss": 0.7812,
"num_input_tokens_seen": 3075440,
"step": 900
},
{
"epoch": 2.2670025188916876,
"eval_accuracy": 0.7686542323215242,
"eval_loss": 1.077448844909668,
"eval_runtime": 536.5671,
"eval_samples_per_second": 0.33,
"eval_steps_per_second": 0.33,
"num_input_tokens_seen": 3075440,
"step": 900
},
{
"epoch": 2.279596977329975,
"grad_norm": 2.752980947494507,
"learning_rate": 6.78300866832467e-06,
"loss": 0.9713,
"num_input_tokens_seen": 3094664,
"step": 905
},
{
"epoch": 2.292191435768262,
"grad_norm": 3.7448384761810303,
"learning_rate": 6.558787819336002e-06,
"loss": 0.6824,
"num_input_tokens_seen": 3111856,
"step": 910
},
{
"epoch": 2.304785894206549,
"grad_norm": 4.047982692718506,
"learning_rate": 6.337774712125597e-06,
"loss": 0.7068,
"num_input_tokens_seen": 3128312,
"step": 915
},
{
"epoch": 2.3173803526448364,
"grad_norm": 1.6455196142196655,
"learning_rate": 6.120007790640123e-06,
"loss": 0.7046,
"num_input_tokens_seen": 3146240,
"step": 920
},
{
"epoch": 2.3299748110831233,
"grad_norm": 3.6144297122955322,
"learning_rate": 5.905524934171086e-06,
"loss": 0.493,
"num_input_tokens_seen": 3164896,
"step": 925
},
{
"epoch": 2.3425692695214106,
"grad_norm": 2.6038029193878174,
"learning_rate": 5.694363450765958e-06,
"loss": 0.5957,
"num_input_tokens_seen": 3180744,
"step": 930
},
{
"epoch": 2.355163727959698,
"grad_norm": 2.931267023086548,
"learning_rate": 5.486560070738647e-06,
"loss": 0.5896,
"num_input_tokens_seen": 3196144,
"step": 935
},
{
"epoch": 2.367758186397985,
"grad_norm": 3.1074607372283936,
"learning_rate": 5.282150940280481e-06,
"loss": 0.7852,
"num_input_tokens_seen": 3214888,
"step": 940
},
{
"epoch": 2.380352644836272,
"grad_norm": 2.5608012676239014,
"learning_rate": 5.081171615172781e-06,
"loss": 0.6306,
"num_input_tokens_seen": 3231664,
"step": 945
},
{
"epoch": 2.392947103274559,
"grad_norm": 3.663986921310425,
"learning_rate": 4.883657054602148e-06,
"loss": 0.5392,
"num_input_tokens_seen": 3249000,
"step": 950
},
{
"epoch": 2.4055415617128464,
"grad_norm": 3.1661274433135986,
"learning_rate": 4.689641615079499e-06,
"loss": 0.6259,
"num_input_tokens_seen": 3267528,
"step": 955
},
{
"epoch": 2.4181360201511337,
"grad_norm": 3.51973819732666,
"learning_rate": 4.499159044463983e-06,
"loss": 1.0237,
"num_input_tokens_seen": 3284096,
"step": 960
},
{
"epoch": 2.4307304785894206,
"grad_norm": 3.3107500076293945,
"learning_rate": 4.312242476092698e-06,
"loss": 0.4255,
"num_input_tokens_seen": 3300344,
"step": 965
},
{
"epoch": 2.443324937027708,
"grad_norm": 2.484478712081909,
"learning_rate": 4.1289244230173715e-06,
"loss": 0.4904,
"num_input_tokens_seen": 3317032,
"step": 970
},
{
"epoch": 2.455919395465995,
"grad_norm": 2.773125410079956,
"learning_rate": 3.9492367723488685e-06,
"loss": 0.7385,
"num_input_tokens_seen": 3335752,
"step": 975
},
{
"epoch": 2.468513853904282,
"grad_norm": 4.129367351531982,
"learning_rate": 3.773210779710662e-06,
"loss": 0.5983,
"num_input_tokens_seen": 3352968,
"step": 980
},
{
"epoch": 2.4811083123425695,
"grad_norm": 2.1071560382843018,
"learning_rate": 3.600877063802055e-06,
"loss": 0.6822,
"num_input_tokens_seen": 3366088,
"step": 985
},
{
"epoch": 2.4937027707808563,
"grad_norm": 1.8715922832489014,
"learning_rate": 3.4322656010722542e-06,
"loss": 0.5104,
"num_input_tokens_seen": 3382936,
"step": 990
},
{
"epoch": 2.5062972292191437,
"grad_norm": 3.2055535316467285,
"learning_rate": 3.267405720506156e-06,
"loss": 0.6008,
"num_input_tokens_seen": 3397648,
"step": 995
},
{
"epoch": 2.5188916876574305,
"grad_norm": 3.4338555335998535,
"learning_rate": 3.106326098522705e-06,
"loss": 0.6864,
"num_input_tokens_seen": 3416024,
"step": 1000
},
{
"epoch": 2.5188916876574305,
"eval_accuracy": 0.7695559062966267,
"eval_loss": 1.0843136310577393,
"eval_runtime": 538.1124,
"eval_samples_per_second": 0.329,
"eval_steps_per_second": 0.329,
"num_input_tokens_seen": 3416024,
"step": 1000
}
],
"logging_steps": 5,
"max_steps": 1191,
"num_input_tokens_seen": 3416024,
"num_train_epochs": 3,
"save_steps": 100,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 1.5121076737651507e+17,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}