{ "best_metric": null, "best_model_checkpoint": null, "epoch": 2.5188916876574305, "eval_steps": 100, "global_step": 1000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.012594458438287154, "grad_norm": 1.2670824527740479, "learning_rate": 4.999782569758238e-05, "loss": 2.5898, "num_input_tokens_seen": 17064, "step": 5 }, { "epoch": 0.02518891687657431, "grad_norm": 1.8668310642242432, "learning_rate": 4.9991303168536793e-05, "loss": 2.69, "num_input_tokens_seen": 31136, "step": 10 }, { "epoch": 0.037783375314861464, "grad_norm": 1.4463753700256348, "learning_rate": 4.9980433547419305e-05, "loss": 2.2158, "num_input_tokens_seen": 47520, "step": 15 }, { "epoch": 0.05037783375314862, "grad_norm": 1.9465786218643188, "learning_rate": 4.996521872493738e-05, "loss": 1.8421, "num_input_tokens_seen": 66432, "step": 20 }, { "epoch": 0.06297229219143577, "grad_norm": 1.3088130950927734, "learning_rate": 4.994566134762105e-05, "loss": 2.0699, "num_input_tokens_seen": 83544, "step": 25 }, { "epoch": 0.07556675062972293, "grad_norm": 1.5839314460754395, "learning_rate": 4.992176481736254e-05, "loss": 1.4037, "num_input_tokens_seen": 97680, "step": 30 }, { "epoch": 0.08816120906801007, "grad_norm": 2.81331729888916, "learning_rate": 4.989353329082452e-05, "loss": 1.8734, "num_input_tokens_seen": 115456, "step": 35 }, { "epoch": 0.10075566750629723, "grad_norm": 1.999861240386963, "learning_rate": 4.986097167871711e-05, "loss": 1.584, "num_input_tokens_seen": 132872, "step": 40 }, { "epoch": 0.11335012594458438, "grad_norm": 1.38965904712677, "learning_rate": 4.982408564494367e-05, "loss": 1.0849, "num_input_tokens_seen": 148296, "step": 45 }, { "epoch": 0.12594458438287154, "grad_norm": 1.410891056060791, "learning_rate": 4.978288160561558e-05, "loss": 1.0605, "num_input_tokens_seen": 167504, "step": 50 }, { "epoch": 0.1385390428211587, "grad_norm": 1.8659024238586426, "learning_rate": 4.9737366727936235e-05, "loss": 1.596, "num_input_tokens_seen": 182536, "step": 55 }, { "epoch": 0.15113350125944586, "grad_norm": 0.825343906879425, "learning_rate": 4.968754892895432e-05, "loss": 1.1978, "num_input_tokens_seen": 201568, "step": 60 }, { "epoch": 0.163727959697733, "grad_norm": 1.3539998531341553, "learning_rate": 4.963343687418669e-05, "loss": 1.1129, "num_input_tokens_seen": 217584, "step": 65 }, { "epoch": 0.17632241813602015, "grad_norm": 1.6213319301605225, "learning_rate": 4.9575039976111084e-05, "loss": 1.3955, "num_input_tokens_seen": 235280, "step": 70 }, { "epoch": 0.1889168765743073, "grad_norm": 1.7147520780563354, "learning_rate": 4.9512368392528806e-05, "loss": 1.029, "num_input_tokens_seen": 253544, "step": 75 }, { "epoch": 0.20151133501259447, "grad_norm": 1.0490851402282715, "learning_rate": 4.9445433024797936e-05, "loss": 1.1447, "num_input_tokens_seen": 270112, "step": 80 }, { "epoch": 0.2141057934508816, "grad_norm": 1.7510355710983276, "learning_rate": 4.937424551593702e-05, "loss": 1.0042, "num_input_tokens_seen": 287296, "step": 85 }, { "epoch": 0.22670025188916876, "grad_norm": 1.5315502882003784, "learning_rate": 4.929881824859985e-05, "loss": 1.3522, "num_input_tokens_seen": 302304, "step": 90 }, { "epoch": 0.23929471032745592, "grad_norm": 1.2641905546188354, "learning_rate": 4.9219164342921634e-05, "loss": 1.2398, "num_input_tokens_seen": 317624, "step": 95 }, { "epoch": 0.2518891687657431, "grad_norm": 1.116264820098877, "learning_rate": 4.9135297654236724e-05, "loss": 0.8733, "num_input_tokens_seen": 332920, "step": 100 }, { "epoch": 0.2518891687657431, "eval_accuracy": 0.7302086027292422, "eval_loss": 1.238808035850525, "eval_runtime": 533.8541, "eval_samples_per_second": 0.332, "eval_steps_per_second": 0.332, "num_input_tokens_seen": 332920, "step": 100 }, { "epoch": 0.26448362720403024, "grad_norm": 1.540872573852539, "learning_rate": 4.904723277066864e-05, "loss": 1.1192, "num_input_tokens_seen": 347680, "step": 105 }, { "epoch": 0.2770780856423174, "grad_norm": 1.2526932954788208, "learning_rate": 4.8954985010592534e-05, "loss": 1.1241, "num_input_tokens_seen": 364952, "step": 110 }, { "epoch": 0.28967254408060455, "grad_norm": 1.2870938777923584, "learning_rate": 4.8858570419970616e-05, "loss": 1.1762, "num_input_tokens_seen": 383104, "step": 115 }, { "epoch": 0.3022670025188917, "grad_norm": 1.6976910829544067, "learning_rate": 4.875800576956108e-05, "loss": 1.0041, "num_input_tokens_seen": 401448, "step": 120 }, { "epoch": 0.3148614609571788, "grad_norm": 1.2784565687179565, "learning_rate": 4.865330855200094e-05, "loss": 0.9044, "num_input_tokens_seen": 419184, "step": 125 }, { "epoch": 0.327455919395466, "grad_norm": 2.2473819255828857, "learning_rate": 4.854449697876325e-05, "loss": 1.5228, "num_input_tokens_seen": 436976, "step": 130 }, { "epoch": 0.34005037783375314, "grad_norm": 1.252365231513977, "learning_rate": 4.843158997698936e-05, "loss": 1.2336, "num_input_tokens_seen": 455432, "step": 135 }, { "epoch": 0.3526448362720403, "grad_norm": 1.5875591039657593, "learning_rate": 4.831460718619661e-05, "loss": 1.0085, "num_input_tokens_seen": 473896, "step": 140 }, { "epoch": 0.36523929471032746, "grad_norm": 1.8101682662963867, "learning_rate": 4.819356895486219e-05, "loss": 1.3333, "num_input_tokens_seen": 488936, "step": 145 }, { "epoch": 0.3778337531486146, "grad_norm": 1.2537903785705566, "learning_rate": 4.806849633688363e-05, "loss": 1.1225, "num_input_tokens_seen": 507800, "step": 150 }, { "epoch": 0.3904282115869018, "grad_norm": 1.2231327295303345, "learning_rate": 4.7939411087916566e-05, "loss": 1.0405, "num_input_tokens_seen": 524704, "step": 155 }, { "epoch": 0.40302267002518893, "grad_norm": 0.9885507822036743, "learning_rate": 4.7806335661590526e-05, "loss": 1.0268, "num_input_tokens_seen": 542320, "step": 160 }, { "epoch": 0.4156171284634761, "grad_norm": 1.273047685623169, "learning_rate": 4.7669293205603196e-05, "loss": 1.1071, "num_input_tokens_seen": 560488, "step": 165 }, { "epoch": 0.4282115869017632, "grad_norm": 1.8347108364105225, "learning_rate": 4.752830755769405e-05, "loss": 1.169, "num_input_tokens_seen": 577680, "step": 170 }, { "epoch": 0.44080604534005036, "grad_norm": 1.9854867458343506, "learning_rate": 4.73834032414979e-05, "loss": 0.9519, "num_input_tokens_seen": 596208, "step": 175 }, { "epoch": 0.4534005037783375, "grad_norm": 1.2936229705810547, "learning_rate": 4.723460546227914e-05, "loss": 1.2277, "num_input_tokens_seen": 613120, "step": 180 }, { "epoch": 0.4659949622166247, "grad_norm": 0.9850680232048035, "learning_rate": 4.7081940102547463e-05, "loss": 0.9588, "num_input_tokens_seen": 630336, "step": 185 }, { "epoch": 0.47858942065491183, "grad_norm": 1.643853783607483, "learning_rate": 4.692543371755572e-05, "loss": 1.0816, "num_input_tokens_seen": 644488, "step": 190 }, { "epoch": 0.491183879093199, "grad_norm": 1.7307960987091064, "learning_rate": 4.6765113530680825e-05, "loss": 0.8216, "num_input_tokens_seen": 660432, "step": 195 }, { "epoch": 0.5037783375314862, "grad_norm": 1.731389045715332, "learning_rate": 4.660100742868836e-05, "loss": 0.8872, "num_input_tokens_seen": 680080, "step": 200 }, { "epoch": 0.5037783375314862, "eval_accuracy": 0.7530999930068348, "eval_loss": 1.129120111465454, "eval_runtime": 534.7922, "eval_samples_per_second": 0.331, "eval_steps_per_second": 0.331, "num_input_tokens_seen": 680080, "step": 200 }, { "epoch": 0.5163727959697733, "grad_norm": 2.0154528617858887, "learning_rate": 4.643314395688188e-05, "loss": 1.0193, "num_input_tokens_seen": 695688, "step": 205 }, { "epoch": 0.5289672544080605, "grad_norm": 1.5947645902633667, "learning_rate": 4.626155231413758e-05, "loss": 1.1722, "num_input_tokens_seen": 718192, "step": 210 }, { "epoch": 0.5415617128463476, "grad_norm": 1.233261227607727, "learning_rate": 4.608626234782536e-05, "loss": 0.9888, "num_input_tokens_seen": 733136, "step": 215 }, { "epoch": 0.5541561712846348, "grad_norm": 2.006932497024536, "learning_rate": 4.5907304548617024e-05, "loss": 1.018, "num_input_tokens_seen": 748016, "step": 220 }, { "epoch": 0.5667506297229219, "grad_norm": 1.6554147005081177, "learning_rate": 4.572471004518261e-05, "loss": 0.8822, "num_input_tokens_seen": 763264, "step": 225 }, { "epoch": 0.5793450881612091, "grad_norm": 2.6174421310424805, "learning_rate": 4.553851059877573e-05, "loss": 1.1984, "num_input_tokens_seen": 778632, "step": 230 }, { "epoch": 0.5919395465994962, "grad_norm": 1.639770269393921, "learning_rate": 4.534873859770892e-05, "loss": 1.0492, "num_input_tokens_seen": 799096, "step": 235 }, { "epoch": 0.6045340050377834, "grad_norm": 1.863054871559143, "learning_rate": 4.515542705171981e-05, "loss": 0.7339, "num_input_tokens_seen": 815240, "step": 240 }, { "epoch": 0.6171284634760705, "grad_norm": 2.6653268337249756, "learning_rate": 4.495860958622937e-05, "loss": 1.2535, "num_input_tokens_seen": 831320, "step": 245 }, { "epoch": 0.6297229219143576, "grad_norm": 3.9443094730377197, "learning_rate": 4.475832043649287e-05, "loss": 1.883, "num_input_tokens_seen": 850376, "step": 250 }, { "epoch": 0.6423173803526449, "grad_norm": 3.2649178504943848, "learning_rate": 4.455459444164492e-05, "loss": 0.9972, "num_input_tokens_seen": 868192, "step": 255 }, { "epoch": 0.654911838790932, "grad_norm": 2.0120456218719482, "learning_rate": 4.4347467038639364e-05, "loss": 1.0848, "num_input_tokens_seen": 883560, "step": 260 }, { "epoch": 0.6675062972292192, "grad_norm": 1.7830870151519775, "learning_rate": 4.4136974256085236e-05, "loss": 0.8549, "num_input_tokens_seen": 902952, "step": 265 }, { "epoch": 0.6801007556675063, "grad_norm": 1.284114956855774, "learning_rate": 4.392315270797985e-05, "loss": 0.5925, "num_input_tokens_seen": 919584, "step": 270 }, { "epoch": 0.6926952141057935, "grad_norm": 1.58772873878479, "learning_rate": 4.3706039587339894e-05, "loss": 1.2722, "num_input_tokens_seen": 940200, "step": 275 }, { "epoch": 0.7052896725440806, "grad_norm": 1.5675506591796875, "learning_rate": 4.3485672659732034e-05, "loss": 0.9741, "num_input_tokens_seen": 961256, "step": 280 }, { "epoch": 0.7178841309823678, "grad_norm": 1.5801304578781128, "learning_rate": 4.3262090256703736e-05, "loss": 0.9787, "num_input_tokens_seen": 978000, "step": 285 }, { "epoch": 0.7304785894206549, "grad_norm": 1.2644524574279785, "learning_rate": 4.303533126911577e-05, "loss": 1.1364, "num_input_tokens_seen": 997512, "step": 290 }, { "epoch": 0.743073047858942, "grad_norm": 1.307681918144226, "learning_rate": 4.280543514037731e-05, "loss": 1.1322, "num_input_tokens_seen": 1016824, "step": 295 }, { "epoch": 0.7556675062972292, "grad_norm": 1.8267008066177368, "learning_rate": 4.257244185958505e-05, "loss": 1.0074, "num_input_tokens_seen": 1036168, "step": 300 }, { "epoch": 0.7556675062972292, "eval_accuracy": 0.7609178510672793, "eval_loss": 1.095440149307251, "eval_runtime": 537.6469, "eval_samples_per_second": 0.329, "eval_steps_per_second": 0.329, "num_input_tokens_seen": 1036168, "step": 300 }, { "epoch": 0.7682619647355163, "grad_norm": 1.6457531452178955, "learning_rate": 4.233639195456729e-05, "loss": 0.939, "num_input_tokens_seen": 1053264, "step": 305 }, { "epoch": 0.7808564231738035, "grad_norm": 1.752995252609253, "learning_rate": 4.2097326484834346e-05, "loss": 1.0468, "num_input_tokens_seen": 1068696, "step": 310 }, { "epoch": 0.7934508816120907, "grad_norm": 1.001474142074585, "learning_rate": 4.1855287034436555e-05, "loss": 0.8325, "num_input_tokens_seen": 1085264, "step": 315 }, { "epoch": 0.8060453400503779, "grad_norm": 1.4869558811187744, "learning_rate": 4.1610315704730854e-05, "loss": 0.8035, "num_input_tokens_seen": 1102368, "step": 320 }, { "epoch": 0.818639798488665, "grad_norm": 1.8002692461013794, "learning_rate": 4.136245510705762e-05, "loss": 1.0207, "num_input_tokens_seen": 1117768, "step": 325 }, { "epoch": 0.8312342569269522, "grad_norm": 1.747013807296753, "learning_rate": 4.111174835532857e-05, "loss": 1.2914, "num_input_tokens_seen": 1133440, "step": 330 }, { "epoch": 0.8438287153652393, "grad_norm": 1.6267316341400146, "learning_rate": 4.085823905852745e-05, "loss": 1.2979, "num_input_tokens_seen": 1146480, "step": 335 }, { "epoch": 0.8564231738035264, "grad_norm": 1.2453759908676147, "learning_rate": 4.06019713131244e-05, "loss": 0.6644, "num_input_tokens_seen": 1162248, "step": 340 }, { "epoch": 0.8690176322418136, "grad_norm": 0.9214743971824646, "learning_rate": 4.034298969540567e-05, "loss": 1.1669, "num_input_tokens_seen": 1179224, "step": 345 }, { "epoch": 0.8816120906801007, "grad_norm": 1.0969208478927612, "learning_rate": 4.008133925371988e-05, "loss": 1.2072, "num_input_tokens_seen": 1195240, "step": 350 }, { "epoch": 0.8942065491183879, "grad_norm": 1.6422204971313477, "learning_rate": 3.981706550064208e-05, "loss": 0.9078, "num_input_tokens_seen": 1213056, "step": 355 }, { "epoch": 0.906801007556675, "grad_norm": 1.5316386222839355, "learning_rate": 3.955021440505706e-05, "loss": 0.4814, "num_input_tokens_seen": 1230744, "step": 360 }, { "epoch": 0.9193954659949622, "grad_norm": 1.2854864597320557, "learning_rate": 3.928083238416342e-05, "loss": 0.9272, "num_input_tokens_seen": 1246904, "step": 365 }, { "epoch": 0.9319899244332494, "grad_norm": 1.3557329177856445, "learning_rate": 3.9008966295399494e-05, "loss": 0.9995, "num_input_tokens_seen": 1263880, "step": 370 }, { "epoch": 0.9445843828715366, "grad_norm": 1.2092177867889404, "learning_rate": 3.873466342829281e-05, "loss": 0.8696, "num_input_tokens_seen": 1283816, "step": 375 }, { "epoch": 0.9571788413098237, "grad_norm": 2.7501771450042725, "learning_rate": 3.845797149623434e-05, "loss": 1.4119, "num_input_tokens_seen": 1300192, "step": 380 }, { "epoch": 0.9697732997481109, "grad_norm": 1.222266435623169, "learning_rate": 3.817893862817902e-05, "loss": 0.8804, "num_input_tokens_seen": 1317224, "step": 385 }, { "epoch": 0.982367758186398, "grad_norm": 0.6030636429786682, "learning_rate": 3.789761336027403e-05, "loss": 0.5944, "num_input_tokens_seen": 1335296, "step": 390 }, { "epoch": 0.9949622166246851, "grad_norm": 1.5092536211013794, "learning_rate": 3.761404462741618e-05, "loss": 1.0303, "num_input_tokens_seen": 1351920, "step": 395 }, { "epoch": 1.0075566750629723, "grad_norm": 2.0633888244628906, "learning_rate": 3.7328281754739974e-05, "loss": 0.9671, "num_input_tokens_seen": 1370864, "step": 400 }, { "epoch": 1.0075566750629723, "eval_accuracy": 0.7659399606125864, "eval_loss": 1.0698517560958862, "eval_runtime": 535.9639, "eval_samples_per_second": 0.33, "eval_steps_per_second": 0.33, "num_input_tokens_seen": 1370864, "step": 400 }, { "epoch": 1.0201511335012594, "grad_norm": 1.7792761325836182, "learning_rate": 3.704037444903782e-05, "loss": 1.0106, "num_input_tokens_seen": 1390136, "step": 405 }, { "epoch": 1.0327455919395465, "grad_norm": 1.8169097900390625, "learning_rate": 3.6750372790113766e-05, "loss": 0.7452, "num_input_tokens_seen": 1411432, "step": 410 }, { "epoch": 1.0453400503778338, "grad_norm": 1.3711882829666138, "learning_rate": 3.645832722207248e-05, "loss": 0.9704, "num_input_tokens_seen": 1429024, "step": 415 }, { "epoch": 1.057934508816121, "grad_norm": 0.9055928587913513, "learning_rate": 3.6164288544544725e-05, "loss": 0.5268, "num_input_tokens_seen": 1445848, "step": 420 }, { "epoch": 1.070528967254408, "grad_norm": 1.2504442930221558, "learning_rate": 3.586830790385109e-05, "loss": 0.6362, "num_input_tokens_seen": 1463232, "step": 425 }, { "epoch": 1.0831234256926952, "grad_norm": 2.108982563018799, "learning_rate": 3.55704367841054e-05, "loss": 0.5694, "num_input_tokens_seen": 1478584, "step": 430 }, { "epoch": 1.0957178841309823, "grad_norm": 2.6890852451324463, "learning_rate": 3.52707269982593e-05, "loss": 0.5836, "num_input_tokens_seen": 1495112, "step": 435 }, { "epoch": 1.1083123425692696, "grad_norm": 3.2210803031921387, "learning_rate": 3.496923067908977e-05, "loss": 1.0356, "num_input_tokens_seen": 1513000, "step": 440 }, { "epoch": 1.1209068010075567, "grad_norm": 1.5383672714233398, "learning_rate": 3.466600027013084e-05, "loss": 1.0125, "num_input_tokens_seen": 1526896, "step": 445 }, { "epoch": 1.1335012594458438, "grad_norm": 2.0438127517700195, "learning_rate": 3.436108851655143e-05, "loss": 1.0554, "num_input_tokens_seen": 1542448, "step": 450 }, { "epoch": 1.146095717884131, "grad_norm": 2.1045315265655518, "learning_rate": 3.4054548455980565e-05, "loss": 0.714, "num_input_tokens_seen": 1557656, "step": 455 }, { "epoch": 1.1586901763224182, "grad_norm": 2.4777379035949707, "learning_rate": 3.3746433409281844e-05, "loss": 0.8676, "num_input_tokens_seen": 1575192, "step": 460 }, { "epoch": 1.1712846347607053, "grad_norm": 2.529090642929077, "learning_rate": 3.3436796971278526e-05, "loss": 0.6624, "num_input_tokens_seen": 1596112, "step": 465 }, { "epoch": 1.1838790931989924, "grad_norm": 2.93548846244812, "learning_rate": 3.312569300143108e-05, "loss": 0.795, "num_input_tokens_seen": 1610768, "step": 470 }, { "epoch": 1.1964735516372795, "grad_norm": 2.8596348762512207, "learning_rate": 3.2813175614468604e-05, "loss": 1.3433, "num_input_tokens_seen": 1627672, "step": 475 }, { "epoch": 1.2090680100755669, "grad_norm": 3.336879253387451, "learning_rate": 3.24992991709759e-05, "loss": 1.0594, "num_input_tokens_seen": 1641912, "step": 480 }, { "epoch": 1.221662468513854, "grad_norm": 2.030346155166626, "learning_rate": 3.218411826793777e-05, "loss": 0.972, "num_input_tokens_seen": 1659832, "step": 485 }, { "epoch": 1.234256926952141, "grad_norm": 2.6392228603363037, "learning_rate": 3.186768772924216e-05, "loss": 1.032, "num_input_tokens_seen": 1679064, "step": 490 }, { "epoch": 1.2468513853904282, "grad_norm": 2.443204402923584, "learning_rate": 3.1550062596143886e-05, "loss": 1.3751, "num_input_tokens_seen": 1692608, "step": 495 }, { "epoch": 1.2594458438287153, "grad_norm": 1.4907230138778687, "learning_rate": 3.1231298117690554e-05, "loss": 0.7884, "num_input_tokens_seen": 1709712, "step": 500 }, { "epoch": 1.2594458438287153, "eval_accuracy": 0.7672396326295029, "eval_loss": 1.0675625801086426, "eval_runtime": 535.3969, "eval_samples_per_second": 0.331, "eval_steps_per_second": 0.331, "num_input_tokens_seen": 1709712, "step": 500 }, { "epoch": 1.2720403022670026, "grad_norm": 0.9061765074729919, "learning_rate": 3.091144974111224e-05, "loss": 0.781, "num_input_tokens_seen": 1729424, "step": 505 }, { "epoch": 1.2846347607052897, "grad_norm": 2.395761251449585, "learning_rate": 3.059057310217683e-05, "loss": 0.8264, "num_input_tokens_seen": 1749136, "step": 510 }, { "epoch": 1.2972292191435768, "grad_norm": 2.069194793701172, "learning_rate": 3.0268724015512463e-05, "loss": 0.6579, "num_input_tokens_seen": 1765216, "step": 515 }, { "epoch": 1.309823677581864, "grad_norm": 1.5534199476242065, "learning_rate": 2.994595846489892e-05, "loss": 0.7195, "num_input_tokens_seen": 1781320, "step": 520 }, { "epoch": 1.322418136020151, "grad_norm": 1.4709012508392334, "learning_rate": 2.9622332593529563e-05, "loss": 0.5226, "num_input_tokens_seen": 1797760, "step": 525 }, { "epoch": 1.3350125944584383, "grad_norm": 2.4601175785064697, "learning_rate": 2.9297902694245542e-05, "loss": 1.2005, "num_input_tokens_seen": 1813224, "step": 530 }, { "epoch": 1.3476070528967254, "grad_norm": 2.9571943283081055, "learning_rate": 2.8972725199744033e-05, "loss": 0.7554, "num_input_tokens_seen": 1830728, "step": 535 }, { "epoch": 1.3602015113350125, "grad_norm": 1.610404372215271, "learning_rate": 2.864685667276201e-05, "loss": 1.0766, "num_input_tokens_seen": 1848816, "step": 540 }, { "epoch": 1.3727959697732999, "grad_norm": 2.7103452682495117, "learning_rate": 2.8320353796237553e-05, "loss": 0.8778, "num_input_tokens_seen": 1863224, "step": 545 }, { "epoch": 1.385390428211587, "grad_norm": 3.026928424835205, "learning_rate": 2.7993273363450184e-05, "loss": 0.6799, "num_input_tokens_seen": 1880288, "step": 550 }, { "epoch": 1.397984886649874, "grad_norm": 1.5982580184936523, "learning_rate": 2.7665672268141956e-05, "loss": 0.4951, "num_input_tokens_seen": 1896552, "step": 555 }, { "epoch": 1.4105793450881612, "grad_norm": 2.5049729347229004, "learning_rate": 2.7337607494621152e-05, "loss": 0.9428, "num_input_tokens_seen": 1915872, "step": 560 }, { "epoch": 1.4231738035264483, "grad_norm": 2.13607120513916, "learning_rate": 2.7009136107850185e-05, "loss": 0.8936, "num_input_tokens_seen": 1934704, "step": 565 }, { "epoch": 1.4357682619647356, "grad_norm": 3.7449216842651367, "learning_rate": 2.668031524351949e-05, "loss": 0.8481, "num_input_tokens_seen": 1951816, "step": 570 }, { "epoch": 1.4483627204030227, "grad_norm": 2.881800651550293, "learning_rate": 2.6351202098109083e-05, "loss": 1.1778, "num_input_tokens_seen": 1970640, "step": 575 }, { "epoch": 1.4609571788413098, "grad_norm": 3.845482110977173, "learning_rate": 2.6021853918939587e-05, "loss": 0.7675, "num_input_tokens_seen": 1986504, "step": 580 }, { "epoch": 1.473551637279597, "grad_norm": 2.6726672649383545, "learning_rate": 2.5692327994214383e-05, "loss": 0.8112, "num_input_tokens_seen": 2003440, "step": 585 }, { "epoch": 1.486146095717884, "grad_norm": 3.1416702270507812, "learning_rate": 2.536268164305465e-05, "loss": 1.3447, "num_input_tokens_seen": 2020568, "step": 590 }, { "epoch": 1.4987405541561714, "grad_norm": 1.6394110918045044, "learning_rate": 2.5032972205529044e-05, "loss": 0.9512, "num_input_tokens_seen": 2037096, "step": 595 }, { "epoch": 1.5113350125944585, "grad_norm": 4.283825874328613, "learning_rate": 2.4703257032679744e-05, "loss": 1.0526, "num_input_tokens_seen": 2053296, "step": 600 }, { "epoch": 1.5113350125944585, "eval_accuracy": 0.7689241678712095, "eval_loss": 1.0595225095748901, "eval_runtime": 536.6614, "eval_samples_per_second": 0.33, "eval_steps_per_second": 0.33, "num_input_tokens_seen": 2053296, "step": 600 }, { "epoch": 1.5239294710327456, "grad_norm": 2.488013505935669, "learning_rate": 2.437359347654655e-05, "loss": 0.7499, "num_input_tokens_seen": 2070840, "step": 605 }, { "epoch": 1.536523929471033, "grad_norm": 2.6227219104766846, "learning_rate": 2.4044038880190824e-05, "loss": 0.6816, "num_input_tokens_seen": 2087952, "step": 610 }, { "epoch": 1.5491183879093198, "grad_norm": 1.874306082725525, "learning_rate": 2.3714650567721016e-05, "loss": 0.7719, "num_input_tokens_seen": 2103824, "step": 615 }, { "epoch": 1.561712846347607, "grad_norm": 2.8610432147979736, "learning_rate": 2.338548583432144e-05, "loss": 1.1229, "num_input_tokens_seen": 2120240, "step": 620 }, { "epoch": 1.5743073047858942, "grad_norm": 6.437571048736572, "learning_rate": 2.305660193628618e-05, "loss": 1.1712, "num_input_tokens_seen": 2135416, "step": 625 }, { "epoch": 1.5869017632241813, "grad_norm": 2.076413154602051, "learning_rate": 2.272805608105958e-05, "loss": 0.6688, "num_input_tokens_seen": 2151904, "step": 630 }, { "epoch": 1.5994962216624686, "grad_norm": 2.047494888305664, "learning_rate": 2.2399905417285434e-05, "loss": 0.8043, "num_input_tokens_seen": 2168952, "step": 635 }, { "epoch": 1.6120906801007555, "grad_norm": 2.4243392944335938, "learning_rate": 2.2072207024866266e-05, "loss": 0.5582, "num_input_tokens_seen": 2185192, "step": 640 }, { "epoch": 1.6246851385390428, "grad_norm": 2.66733717918396, "learning_rate": 2.1745017905034625e-05, "loss": 1.1033, "num_input_tokens_seen": 2200856, "step": 645 }, { "epoch": 1.63727959697733, "grad_norm": 2.6124303340911865, "learning_rate": 2.141839497043806e-05, "loss": 0.8529, "num_input_tokens_seen": 2215080, "step": 650 }, { "epoch": 1.649874055415617, "grad_norm": 2.1553502082824707, "learning_rate": 2.1092395035239472e-05, "loss": 0.7331, "num_input_tokens_seen": 2229808, "step": 655 }, { "epoch": 1.6624685138539044, "grad_norm": 2.305765390396118, "learning_rate": 2.076707480523464e-05, "loss": 0.9966, "num_input_tokens_seen": 2247584, "step": 660 }, { "epoch": 1.6750629722921915, "grad_norm": 1.9919112920761108, "learning_rate": 2.0442490867988582e-05, "loss": 1.0719, "num_input_tokens_seen": 2264280, "step": 665 }, { "epoch": 1.6876574307304786, "grad_norm": 2.5872271060943604, "learning_rate": 2.011869968299245e-05, "loss": 0.8667, "num_input_tokens_seen": 2281600, "step": 670 }, { "epoch": 1.700251889168766, "grad_norm": 2.303976535797119, "learning_rate": 1.9795757571842744e-05, "loss": 0.559, "num_input_tokens_seen": 2298688, "step": 675 }, { "epoch": 1.7128463476070528, "grad_norm": 2.5704333782196045, "learning_rate": 1.947372070844452e-05, "loss": 0.899, "num_input_tokens_seen": 2318312, "step": 680 }, { "epoch": 1.7254408060453401, "grad_norm": 2.638066530227661, "learning_rate": 1.915264510924022e-05, "loss": 0.7052, "num_input_tokens_seen": 2334496, "step": 685 }, { "epoch": 1.7380352644836272, "grad_norm": 2.046243906021118, "learning_rate": 1.883258662346596e-05, "loss": 0.9922, "num_input_tokens_seen": 2355640, "step": 690 }, { "epoch": 1.7506297229219143, "grad_norm": 2.731766700744629, "learning_rate": 1.8513600923436923e-05, "loss": 0.9633, "num_input_tokens_seen": 2375368, "step": 695 }, { "epoch": 1.7632241813602016, "grad_norm": 4.190962791442871, "learning_rate": 1.8195743494863387e-05, "loss": 1.2255, "num_input_tokens_seen": 2392384, "step": 700 }, { "epoch": 1.7632241813602016, "eval_accuracy": 0.7701849009613134, "eval_loss": 1.0550936460494995, "eval_runtime": 535.3581, "eval_samples_per_second": 0.331, "eval_steps_per_second": 0.331, "num_input_tokens_seen": 2392384, "step": 700 }, { "epoch": 1.7758186397984885, "grad_norm": 6.717029094696045, "learning_rate": 1.787906962719939e-05, "loss": 0.7964, "num_input_tokens_seen": 2409672, "step": 705 }, { "epoch": 1.7884130982367759, "grad_norm": 2.1568119525909424, "learning_rate": 1.7563634404025414e-05, "loss": 0.7568, "num_input_tokens_seen": 2425144, "step": 710 }, { "epoch": 1.801007556675063, "grad_norm": 2.602142810821533, "learning_rate": 1.7249492693466934e-05, "loss": 0.739, "num_input_tokens_seen": 2445872, "step": 715 }, { "epoch": 1.81360201511335, "grad_norm": 2.0826470851898193, "learning_rate": 1.6936699138650397e-05, "loss": 0.7168, "num_input_tokens_seen": 2463232, "step": 720 }, { "epoch": 1.8261964735516374, "grad_norm": 3.3417489528656006, "learning_rate": 1.6625308148198413e-05, "loss": 0.7757, "num_input_tokens_seen": 2480816, "step": 725 }, { "epoch": 1.8387909319899243, "grad_norm": 2.6503310203552246, "learning_rate": 1.6315373886765646e-05, "loss": 0.6779, "num_input_tokens_seen": 2498488, "step": 730 }, { "epoch": 1.8513853904282116, "grad_norm": 3.9675190448760986, "learning_rate": 1.600695026561721e-05, "loss": 0.9367, "num_input_tokens_seen": 2516792, "step": 735 }, { "epoch": 1.8639798488664987, "grad_norm": 2.195193290710449, "learning_rate": 1.5700090933251115e-05, "loss": 0.447, "num_input_tokens_seen": 2533232, "step": 740 }, { "epoch": 1.8765743073047858, "grad_norm": 2.9985008239746094, "learning_rate": 1.5394849266066416e-05, "loss": 0.6294, "num_input_tokens_seen": 2552824, "step": 745 }, { "epoch": 1.8891687657430731, "grad_norm": 3.8788790702819824, "learning_rate": 1.509127835907872e-05, "loss": 1.3992, "num_input_tokens_seen": 2571512, "step": 750 }, { "epoch": 1.9017632241813602, "grad_norm": 1.619795799255371, "learning_rate": 1.4789431016684558e-05, "loss": 0.8268, "num_input_tokens_seen": 2588120, "step": 755 }, { "epoch": 1.9143576826196473, "grad_norm": 3.272700786590576, "learning_rate": 1.4489359743476461e-05, "loss": 0.6765, "num_input_tokens_seen": 2605248, "step": 760 }, { "epoch": 1.9269521410579347, "grad_norm": 2.847982168197632, "learning_rate": 1.4191116735110007e-05, "loss": 1.1278, "num_input_tokens_seen": 2621656, "step": 765 }, { "epoch": 1.9395465994962215, "grad_norm": 3.8507871627807617, "learning_rate": 1.3894753869224725e-05, "loss": 0.6863, "num_input_tokens_seen": 2639440, "step": 770 }, { "epoch": 1.9521410579345089, "grad_norm": 2.1526479721069336, "learning_rate": 1.3600322696420275e-05, "loss": 0.6884, "num_input_tokens_seen": 2657952, "step": 775 }, { "epoch": 1.964735516372796, "grad_norm": 2.447094202041626, "learning_rate": 1.330787443128953e-05, "loss": 0.6405, "num_input_tokens_seen": 2673752, "step": 780 }, { "epoch": 1.977329974811083, "grad_norm": 1.717159628868103, "learning_rate": 1.3017459943510084e-05, "loss": 0.6037, "num_input_tokens_seen": 2689440, "step": 785 }, { "epoch": 1.9899244332493704, "grad_norm": 3.210221529006958, "learning_rate": 1.2729129748995749e-05, "loss": 0.9172, "num_input_tokens_seen": 2706624, "step": 790 }, { "epoch": 2.0025188916876573, "grad_norm": 1.6862154006958008, "learning_rate": 1.2442934001109671e-05, "loss": 0.7949, "num_input_tokens_seen": 2723496, "step": 795 }, { "epoch": 2.0151133501259446, "grad_norm": 2.2537808418273926, "learning_rate": 1.2158922481940361e-05, "loss": 0.5226, "num_input_tokens_seen": 2738304, "step": 800 }, { "epoch": 2.0151133501259446, "eval_accuracy": 0.7704949500614456, "eval_loss": 1.0528764724731445, "eval_runtime": 535.8681, "eval_samples_per_second": 0.33, "eval_steps_per_second": 0.33, "num_input_tokens_seen": 2738304, "step": 800 }, { "epoch": 2.027707808564232, "grad_norm": 5.461361408233643, "learning_rate": 1.1877144593642439e-05, "loss": 0.9266, "num_input_tokens_seen": 2755120, "step": 805 }, { "epoch": 2.040302267002519, "grad_norm": 2.930683135986328, "learning_rate": 1.1597649349843413e-05, "loss": 0.9805, "num_input_tokens_seen": 2771080, "step": 810 }, { "epoch": 2.052896725440806, "grad_norm": 3.590545654296875, "learning_rate": 1.1320485367118017e-05, "loss": 1.1796, "num_input_tokens_seen": 2786752, "step": 815 }, { "epoch": 2.065491183879093, "grad_norm": 2.6750876903533936, "learning_rate": 1.1045700856531668e-05, "loss": 0.4839, "num_input_tokens_seen": 2804392, "step": 820 }, { "epoch": 2.0780856423173804, "grad_norm": 1.7390782833099365, "learning_rate": 1.0773343615254446e-05, "loss": 0.359, "num_input_tokens_seen": 2818640, "step": 825 }, { "epoch": 2.0906801007556677, "grad_norm": 3.9132091999053955, "learning_rate": 1.0503461018246977e-05, "loss": 1.0472, "num_input_tokens_seen": 2836256, "step": 830 }, { "epoch": 2.1032745591939546, "grad_norm": 3.056783437728882, "learning_rate": 1.0236100010019919e-05, "loss": 0.8781, "num_input_tokens_seen": 2855496, "step": 835 }, { "epoch": 2.115869017632242, "grad_norm": 3.0513293743133545, "learning_rate": 9.971307096468203e-06, "loss": 0.9041, "num_input_tokens_seen": 2872200, "step": 840 }, { "epoch": 2.1284634760705288, "grad_norm": 2.966620445251465, "learning_rate": 9.709128336781592e-06, "loss": 0.7187, "num_input_tokens_seen": 2888256, "step": 845 }, { "epoch": 2.141057934508816, "grad_norm": 2.2629482746124268, "learning_rate": 9.449609335432972e-06, "loss": 0.8696, "num_input_tokens_seen": 2905920, "step": 850 }, { "epoch": 2.1536523929471034, "grad_norm": 5.053290367126465, "learning_rate": 9.192795234245697e-06, "loss": 0.6862, "num_input_tokens_seen": 2921056, "step": 855 }, { "epoch": 2.1662468513853903, "grad_norm": 2.8935582637786865, "learning_rate": 8.938730704541473e-06, "loss": 0.702, "num_input_tokens_seen": 2942096, "step": 860 }, { "epoch": 2.1788413098236776, "grad_norm": 4.863999843597412, "learning_rate": 8.687459939369983e-06, "loss": 0.9868, "num_input_tokens_seen": 2959320, "step": 865 }, { "epoch": 2.1914357682619645, "grad_norm": 2.9665706157684326, "learning_rate": 8.439026645821802e-06, "loss": 0.5647, "num_input_tokens_seen": 2976800, "step": 870 }, { "epoch": 2.204030226700252, "grad_norm": 4.912269115447998, "learning_rate": 8.193474037425794e-06, "loss": 0.7983, "num_input_tokens_seen": 2994536, "step": 875 }, { "epoch": 2.216624685138539, "grad_norm": 3.118041515350342, "learning_rate": 7.950844826632373e-06, "loss": 0.9227, "num_input_tokens_seen": 3010504, "step": 880 }, { "epoch": 2.229219143576826, "grad_norm": 1.5938191413879395, "learning_rate": 7.711181217383896e-06, "loss": 0.9203, "num_input_tokens_seen": 3027240, "step": 885 }, { "epoch": 2.2418136020151134, "grad_norm": 2.1337177753448486, "learning_rate": 7.474524897773555e-06, "loss": 0.5222, "num_input_tokens_seen": 3044024, "step": 890 }, { "epoch": 2.2544080604534007, "grad_norm": 3.2635252475738525, "learning_rate": 7.240917032794003e-06, "loss": 0.5499, "num_input_tokens_seen": 3060632, "step": 895 }, { "epoch": 2.2670025188916876, "grad_norm": 2.3565027713775635, "learning_rate": 7.010398257176878e-06, "loss": 0.7812, "num_input_tokens_seen": 3075440, "step": 900 }, { "epoch": 2.2670025188916876, "eval_accuracy": 0.7686542323215242, "eval_loss": 1.077448844909668, "eval_runtime": 536.5671, "eval_samples_per_second": 0.33, "eval_steps_per_second": 0.33, "num_input_tokens_seen": 3075440, "step": 900 }, { "epoch": 2.279596977329975, "grad_norm": 2.752980947494507, "learning_rate": 6.78300866832467e-06, "loss": 0.9713, "num_input_tokens_seen": 3094664, "step": 905 }, { "epoch": 2.292191435768262, "grad_norm": 3.7448384761810303, "learning_rate": 6.558787819336002e-06, "loss": 0.6824, "num_input_tokens_seen": 3111856, "step": 910 }, { "epoch": 2.304785894206549, "grad_norm": 4.047982692718506, "learning_rate": 6.337774712125597e-06, "loss": 0.7068, "num_input_tokens_seen": 3128312, "step": 915 }, { "epoch": 2.3173803526448364, "grad_norm": 1.6455196142196655, "learning_rate": 6.120007790640123e-06, "loss": 0.7046, "num_input_tokens_seen": 3146240, "step": 920 }, { "epoch": 2.3299748110831233, "grad_norm": 3.6144297122955322, "learning_rate": 5.905524934171086e-06, "loss": 0.493, "num_input_tokens_seen": 3164896, "step": 925 }, { "epoch": 2.3425692695214106, "grad_norm": 2.6038029193878174, "learning_rate": 5.694363450765958e-06, "loss": 0.5957, "num_input_tokens_seen": 3180744, "step": 930 }, { "epoch": 2.355163727959698, "grad_norm": 2.931267023086548, "learning_rate": 5.486560070738647e-06, "loss": 0.5896, "num_input_tokens_seen": 3196144, "step": 935 }, { "epoch": 2.367758186397985, "grad_norm": 3.1074607372283936, "learning_rate": 5.282150940280481e-06, "loss": 0.7852, "num_input_tokens_seen": 3214888, "step": 940 }, { "epoch": 2.380352644836272, "grad_norm": 2.5608012676239014, "learning_rate": 5.081171615172781e-06, "loss": 0.6306, "num_input_tokens_seen": 3231664, "step": 945 }, { "epoch": 2.392947103274559, "grad_norm": 3.663986921310425, "learning_rate": 4.883657054602148e-06, "loss": 0.5392, "num_input_tokens_seen": 3249000, "step": 950 }, { "epoch": 2.4055415617128464, "grad_norm": 3.1661274433135986, "learning_rate": 4.689641615079499e-06, "loss": 0.6259, "num_input_tokens_seen": 3267528, "step": 955 }, { "epoch": 2.4181360201511337, "grad_norm": 3.51973819732666, "learning_rate": 4.499159044463983e-06, "loss": 1.0237, "num_input_tokens_seen": 3284096, "step": 960 }, { "epoch": 2.4307304785894206, "grad_norm": 3.3107500076293945, "learning_rate": 4.312242476092698e-06, "loss": 0.4255, "num_input_tokens_seen": 3300344, "step": 965 }, { "epoch": 2.443324937027708, "grad_norm": 2.484478712081909, "learning_rate": 4.1289244230173715e-06, "loss": 0.4904, "num_input_tokens_seen": 3317032, "step": 970 }, { "epoch": 2.455919395465995, "grad_norm": 2.773125410079956, "learning_rate": 3.9492367723488685e-06, "loss": 0.7385, "num_input_tokens_seen": 3335752, "step": 975 }, { "epoch": 2.468513853904282, "grad_norm": 4.129367351531982, "learning_rate": 3.773210779710662e-06, "loss": 0.5983, "num_input_tokens_seen": 3352968, "step": 980 }, { "epoch": 2.4811083123425695, "grad_norm": 2.1071560382843018, "learning_rate": 3.600877063802055e-06, "loss": 0.6822, "num_input_tokens_seen": 3366088, "step": 985 }, { "epoch": 2.4937027707808563, "grad_norm": 1.8715922832489014, "learning_rate": 3.4322656010722542e-06, "loss": 0.5104, "num_input_tokens_seen": 3382936, "step": 990 }, { "epoch": 2.5062972292191437, "grad_norm": 3.2055535316467285, "learning_rate": 3.267405720506156e-06, "loss": 0.6008, "num_input_tokens_seen": 3397648, "step": 995 }, { "epoch": 2.5188916876574305, "grad_norm": 3.4338555335998535, "learning_rate": 3.106326098522705e-06, "loss": 0.6864, "num_input_tokens_seen": 3416024, "step": 1000 }, { "epoch": 2.5188916876574305, "eval_accuracy": 0.7695559062966267, "eval_loss": 1.0843136310577393, "eval_runtime": 538.1124, "eval_samples_per_second": 0.329, "eval_steps_per_second": 0.329, "num_input_tokens_seen": 3416024, "step": 1000 } ], "logging_steps": 5, "max_steps": 1191, "num_input_tokens_seen": 3416024, "num_train_epochs": 3, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 1.5121076737651507e+17, "train_batch_size": 1, "trial_name": null, "trial_params": null }