{ "best_metric": 3.7249155044555664, "best_model_checkpoint": "./results/models/checkpoint-58650", "epoch": 30.0, "eval_steps": 500, "global_step": 58650, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.2557544757033248, "grad_norm": 0.047607421875, "learning_rate": 0.003979539641943734, "loss": 1.9283, "step": 500 }, { "epoch": 0.5115089514066496, "grad_norm": 0.05810546875, "learning_rate": 0.003959079283887468, "loss": 1.89, "step": 1000 }, { "epoch": 0.7672634271099744, "grad_norm": 0.048828125, "learning_rate": 0.003938618925831202, "loss": 1.8823, "step": 1500 }, { "epoch": 1.0, "eval_loss": 3.7650158405303955, "eval_runtime": 1.1856, "eval_samples_per_second": 421.713, "eval_steps_per_second": 0.843, "step": 1955 }, { "epoch": 1.0230179028132993, "grad_norm": 0.048583984375, "learning_rate": 0.003918158567774936, "loss": 1.8791, "step": 2000 }, { "epoch": 1.278772378516624, "grad_norm": 0.0458984375, "learning_rate": 0.00389769820971867, "loss": 1.8765, "step": 2500 }, { "epoch": 1.5345268542199488, "grad_norm": 0.04443359375, "learning_rate": 0.003877237851662404, "loss": 1.8758, "step": 3000 }, { "epoch": 1.7902813299232738, "grad_norm": 0.04638671875, "learning_rate": 0.003856777493606138, "loss": 1.874, "step": 3500 }, { "epoch": 2.0, "eval_loss": 3.755614757537842, "eval_runtime": 1.2689, "eval_samples_per_second": 394.03, "eval_steps_per_second": 0.788, "step": 3910 }, { "epoch": 2.0460358056265986, "grad_norm": 0.0322265625, "learning_rate": 0.0038363171355498722, "loss": 1.8728, "step": 4000 }, { "epoch": 2.3017902813299234, "grad_norm": 0.03857421875, "learning_rate": 0.0038158567774936062, "loss": 1.8721, "step": 4500 }, { "epoch": 2.557544757033248, "grad_norm": 0.051025390625, "learning_rate": 0.0037953964194373403, "loss": 1.8714, "step": 5000 }, { "epoch": 2.813299232736573, "grad_norm": 0.043701171875, "learning_rate": 0.0037749360613810743, "loss": 1.8704, "step": 5500 }, { "epoch": 3.0, "eval_loss": 3.749542474746704, "eval_runtime": 1.1996, "eval_samples_per_second": 416.822, "eval_steps_per_second": 0.834, "step": 5865 }, { "epoch": 3.0690537084398977, "grad_norm": 0.045654296875, "learning_rate": 0.0037544757033248083, "loss": 1.8699, "step": 6000 }, { "epoch": 3.3248081841432224, "grad_norm": 0.038818359375, "learning_rate": 0.0037340153452685423, "loss": 1.8695, "step": 6500 }, { "epoch": 3.580562659846547, "grad_norm": 0.04541015625, "learning_rate": 0.0037135549872122763, "loss": 1.8688, "step": 7000 }, { "epoch": 3.836317135549872, "grad_norm": 0.037109375, "learning_rate": 0.0036930946291560103, "loss": 1.8686, "step": 7500 }, { "epoch": 4.0, "eval_loss": 3.7466166019439697, "eval_runtime": 1.1726, "eval_samples_per_second": 426.391, "eval_steps_per_second": 0.853, "step": 7820 }, { "epoch": 4.092071611253197, "grad_norm": 0.04833984375, "learning_rate": 0.0036726342710997444, "loss": 1.8681, "step": 8000 }, { "epoch": 4.3478260869565215, "grad_norm": 0.047607421875, "learning_rate": 0.003652173913043478, "loss": 1.8678, "step": 8500 }, { "epoch": 4.603580562659847, "grad_norm": 0.056640625, "learning_rate": 0.0036317135549872124, "loss": 1.8674, "step": 9000 }, { "epoch": 4.859335038363171, "grad_norm": 0.0400390625, "learning_rate": 0.0036112531969309464, "loss": 1.867, "step": 9500 }, { "epoch": 5.0, "eval_loss": 3.7444469928741455, "eval_runtime": 1.1971, "eval_samples_per_second": 417.668, "eval_steps_per_second": 0.835, "step": 9775 }, { "epoch": 5.115089514066496, "grad_norm": 0.04736328125, "learning_rate": 0.0035907928388746804, "loss": 1.8668, "step": 10000 }, { "epoch": 5.370843989769821, "grad_norm": 0.03564453125, "learning_rate": 0.0035703324808184144, "loss": 1.8668, "step": 10500 }, { "epoch": 5.626598465473146, "grad_norm": 0.047119140625, "learning_rate": 0.003549872122762148, "loss": 1.866, "step": 11000 }, { "epoch": 5.882352941176471, "grad_norm": 0.04052734375, "learning_rate": 0.0035294117647058825, "loss": 1.8662, "step": 11500 }, { "epoch": 6.0, "eval_loss": 3.742023229598999, "eval_runtime": 1.1689, "eval_samples_per_second": 427.744, "eval_steps_per_second": 0.855, "step": 11730 }, { "epoch": 6.138107416879795, "grad_norm": 0.047119140625, "learning_rate": 0.0035089514066496165, "loss": 1.8652, "step": 12000 }, { "epoch": 6.3938618925831205, "grad_norm": 0.041259765625, "learning_rate": 0.0034884910485933505, "loss": 1.8656, "step": 12500 }, { "epoch": 6.649616368286445, "grad_norm": 0.040283203125, "learning_rate": 0.0034680306905370845, "loss": 1.8655, "step": 13000 }, { "epoch": 6.90537084398977, "grad_norm": 0.04736328125, "learning_rate": 0.003447570332480818, "loss": 1.8653, "step": 13500 }, { "epoch": 7.0, "eval_loss": 3.7410786151885986, "eval_runtime": 1.2028, "eval_samples_per_second": 415.68, "eval_steps_per_second": 0.831, "step": 13685 }, { "epoch": 7.161125319693094, "grad_norm": 0.0439453125, "learning_rate": 0.0034271099744245526, "loss": 1.865, "step": 14000 }, { "epoch": 7.41687979539642, "grad_norm": 0.04052734375, "learning_rate": 0.0034066496163682866, "loss": 1.8648, "step": 14500 }, { "epoch": 7.672634271099744, "grad_norm": 0.060302734375, "learning_rate": 0.0033861892583120206, "loss": 1.865, "step": 15000 }, { "epoch": 7.928388746803069, "grad_norm": 0.0400390625, "learning_rate": 0.0033657289002557546, "loss": 1.8641, "step": 15500 }, { "epoch": 8.0, "eval_loss": 3.7378389835357666, "eval_runtime": 1.1927, "eval_samples_per_second": 419.217, "eval_steps_per_second": 0.838, "step": 15640 }, { "epoch": 8.184143222506394, "grad_norm": 0.04345703125, "learning_rate": 0.003345268542199488, "loss": 1.8644, "step": 16000 }, { "epoch": 8.43989769820972, "grad_norm": 0.04736328125, "learning_rate": 0.0033248081841432226, "loss": 1.8638, "step": 16500 }, { "epoch": 8.695652173913043, "grad_norm": 0.05126953125, "learning_rate": 0.0033043478260869567, "loss": 1.8639, "step": 17000 }, { "epoch": 8.951406649616368, "grad_norm": 0.05224609375, "learning_rate": 0.0032838874680306907, "loss": 1.8641, "step": 17500 }, { "epoch": 9.0, "eval_loss": 3.7366719245910645, "eval_runtime": 1.2549, "eval_samples_per_second": 398.446, "eval_steps_per_second": 0.797, "step": 17595 }, { "epoch": 9.207161125319693, "grad_norm": 0.06396484375, "learning_rate": 0.0032634271099744247, "loss": 1.8636, "step": 18000 }, { "epoch": 9.462915601023019, "grad_norm": 0.03857421875, "learning_rate": 0.0032429667519181583, "loss": 1.8655, "step": 18500 }, { "epoch": 9.718670076726342, "grad_norm": 0.038818359375, "learning_rate": 0.0032225063938618927, "loss": 1.8655, "step": 19000 }, { "epoch": 9.974424552429667, "grad_norm": 0.04150390625, "learning_rate": 0.0032020460358056268, "loss": 1.8646, "step": 19500 }, { "epoch": 10.0, "eval_loss": 3.7368271350860596, "eval_runtime": 1.1654, "eval_samples_per_second": 429.02, "eval_steps_per_second": 0.858, "step": 19550 }, { "epoch": 10.230179028132993, "grad_norm": 0.051513671875, "learning_rate": 0.0031815856777493608, "loss": 1.8637, "step": 20000 }, { "epoch": 10.485933503836318, "grad_norm": 0.048095703125, "learning_rate": 0.003161125319693095, "loss": 1.8641, "step": 20500 }, { "epoch": 10.741687979539641, "grad_norm": 0.037841796875, "learning_rate": 0.0031406649616368284, "loss": 1.8633, "step": 21000 }, { "epoch": 10.997442455242966, "grad_norm": 0.04931640625, "learning_rate": 0.003120204603580563, "loss": 1.8633, "step": 21500 }, { "epoch": 11.0, "eval_loss": 3.736281156539917, "eval_runtime": 1.3206, "eval_samples_per_second": 378.604, "eval_steps_per_second": 0.757, "step": 21505 }, { "epoch": 11.253196930946292, "grad_norm": 0.04833984375, "learning_rate": 0.003099744245524297, "loss": 1.8636, "step": 22000 }, { "epoch": 11.508951406649617, "grad_norm": 0.0478515625, "learning_rate": 0.003079283887468031, "loss": 1.863, "step": 22500 }, { "epoch": 11.764705882352942, "grad_norm": 0.048583984375, "learning_rate": 0.0030588235294117644, "loss": 1.8629, "step": 23000 }, { "epoch": 12.0, "eval_loss": 3.734565496444702, "eval_runtime": 1.1835, "eval_samples_per_second": 422.479, "eval_steps_per_second": 0.845, "step": 23460 }, { "epoch": 12.020460358056265, "grad_norm": 0.044189453125, "learning_rate": 0.0030383631713554985, "loss": 1.8625, "step": 23500 }, { "epoch": 12.27621483375959, "grad_norm": 0.048828125, "learning_rate": 0.003017902813299233, "loss": 1.8629, "step": 24000 }, { "epoch": 12.531969309462916, "grad_norm": 0.04931640625, "learning_rate": 0.002997442455242967, "loss": 1.8624, "step": 24500 }, { "epoch": 12.787723785166241, "grad_norm": 0.04345703125, "learning_rate": 0.002976982097186701, "loss": 1.8621, "step": 25000 }, { "epoch": 13.0, "eval_loss": 3.7329983711242676, "eval_runtime": 1.3627, "eval_samples_per_second": 366.916, "eval_steps_per_second": 0.734, "step": 25415 }, { "epoch": 13.043478260869565, "grad_norm": 0.0546875, "learning_rate": 0.0029565217391304345, "loss": 1.8624, "step": 25500 }, { "epoch": 13.29923273657289, "grad_norm": 0.048095703125, "learning_rate": 0.002936061381074169, "loss": 1.8624, "step": 26000 }, { "epoch": 13.554987212276215, "grad_norm": 0.05517578125, "learning_rate": 0.002915601023017903, "loss": 1.8618, "step": 26500 }, { "epoch": 13.81074168797954, "grad_norm": 0.04443359375, "learning_rate": 0.002895140664961637, "loss": 1.8619, "step": 27000 }, { "epoch": 14.0, "eval_loss": 3.7325527667999268, "eval_runtime": 1.1565, "eval_samples_per_second": 432.347, "eval_steps_per_second": 0.865, "step": 27370 }, { "epoch": 14.066496163682864, "grad_norm": 0.043212890625, "learning_rate": 0.002874680306905371, "loss": 1.862, "step": 27500 }, { "epoch": 14.322250639386189, "grad_norm": 0.06640625, "learning_rate": 0.0028542199488491046, "loss": 1.8616, "step": 28000 }, { "epoch": 14.578005115089514, "grad_norm": 0.050048828125, "learning_rate": 0.002833759590792839, "loss": 1.8615, "step": 28500 }, { "epoch": 14.83375959079284, "grad_norm": 0.046875, "learning_rate": 0.002813299232736573, "loss": 1.8623, "step": 29000 }, { "epoch": 15.0, "eval_loss": 3.729966402053833, "eval_runtime": 1.1923, "eval_samples_per_second": 419.349, "eval_steps_per_second": 0.839, "step": 29325 }, { "epoch": 15.089514066496164, "grad_norm": 0.0458984375, "learning_rate": 0.002792838874680307, "loss": 1.8615, "step": 29500 }, { "epoch": 15.345268542199488, "grad_norm": 0.042236328125, "learning_rate": 0.002772378516624041, "loss": 1.8616, "step": 30000 }, { "epoch": 15.601023017902813, "grad_norm": 0.037353515625, "learning_rate": 0.0027519181585677747, "loss": 1.8618, "step": 30500 }, { "epoch": 15.856777493606138, "grad_norm": 0.06591796875, "learning_rate": 0.002731457800511509, "loss": 1.8613, "step": 31000 }, { "epoch": 16.0, "eval_loss": 3.7326087951660156, "eval_runtime": 1.1839, "eval_samples_per_second": 422.34, "eval_steps_per_second": 0.845, "step": 31280 }, { "epoch": 16.11253196930946, "grad_norm": 0.038818359375, "learning_rate": 0.002710997442455243, "loss": 1.8615, "step": 31500 }, { "epoch": 16.36828644501279, "grad_norm": 0.0458984375, "learning_rate": 0.002690537084398977, "loss": 1.8619, "step": 32000 }, { "epoch": 16.624040920716112, "grad_norm": 0.044677734375, "learning_rate": 0.002670076726342711, "loss": 1.8616, "step": 32500 }, { "epoch": 16.87979539641944, "grad_norm": 0.04541015625, "learning_rate": 0.0026496163682864448, "loss": 1.8616, "step": 33000 }, { "epoch": 17.0, "eval_loss": 3.729959011077881, "eval_runtime": 1.2063, "eval_samples_per_second": 414.478, "eval_steps_per_second": 0.829, "step": 33235 }, { "epoch": 17.135549872122763, "grad_norm": 0.04541015625, "learning_rate": 0.0026291560102301792, "loss": 1.8616, "step": 33500 }, { "epoch": 17.391304347826086, "grad_norm": 0.048583984375, "learning_rate": 0.0026086956521739132, "loss": 1.8608, "step": 34000 }, { "epoch": 17.647058823529413, "grad_norm": 0.047119140625, "learning_rate": 0.0025882352941176473, "loss": 1.8609, "step": 34500 }, { "epoch": 17.902813299232736, "grad_norm": 0.041259765625, "learning_rate": 0.002567774936061381, "loss": 1.8606, "step": 35000 }, { "epoch": 18.0, "eval_loss": 3.7307207584381104, "eval_runtime": 1.1724, "eval_samples_per_second": 426.468, "eval_steps_per_second": 0.853, "step": 35190 }, { "epoch": 18.15856777493606, "grad_norm": 0.04052734375, "learning_rate": 0.002547314578005115, "loss": 1.8613, "step": 35500 }, { "epoch": 18.414322250639387, "grad_norm": 0.042236328125, "learning_rate": 0.0025268542199488493, "loss": 1.8608, "step": 36000 }, { "epoch": 18.67007672634271, "grad_norm": 0.046142578125, "learning_rate": 0.0025063938618925833, "loss": 1.8607, "step": 36500 }, { "epoch": 18.925831202046037, "grad_norm": 0.03955078125, "learning_rate": 0.0024859335038363174, "loss": 1.8606, "step": 37000 }, { "epoch": 19.0, "eval_loss": 3.729092836380005, "eval_runtime": 1.1564, "eval_samples_per_second": 432.369, "eval_steps_per_second": 0.865, "step": 37145 }, { "epoch": 19.18158567774936, "grad_norm": 0.041259765625, "learning_rate": 0.002465473145780051, "loss": 1.8604, "step": 37500 }, { "epoch": 19.437340153452684, "grad_norm": 0.047607421875, "learning_rate": 0.002445012787723785, "loss": 1.8607, "step": 38000 }, { "epoch": 19.69309462915601, "grad_norm": 0.056396484375, "learning_rate": 0.0024245524296675194, "loss": 1.8605, "step": 38500 }, { "epoch": 19.948849104859335, "grad_norm": 0.05517578125, "learning_rate": 0.0024040920716112534, "loss": 1.8606, "step": 39000 }, { "epoch": 20.0, "eval_loss": 3.727680206298828, "eval_runtime": 1.1607, "eval_samples_per_second": 430.785, "eval_steps_per_second": 0.862, "step": 39100 }, { "epoch": 20.20460358056266, "grad_norm": 0.043212890625, "learning_rate": 0.0023836317135549874, "loss": 1.8605, "step": 39500 }, { "epoch": 20.460358056265985, "grad_norm": 0.04345703125, "learning_rate": 0.002363171355498721, "loss": 1.8605, "step": 40000 }, { "epoch": 20.71611253196931, "grad_norm": 0.037841796875, "learning_rate": 0.002342710997442455, "loss": 1.8606, "step": 40500 }, { "epoch": 20.971867007672635, "grad_norm": 0.051513671875, "learning_rate": 0.0023222506393861895, "loss": 1.8601, "step": 41000 }, { "epoch": 21.0, "eval_loss": 3.7270610332489014, "eval_runtime": 1.1618, "eval_samples_per_second": 430.348, "eval_steps_per_second": 0.861, "step": 41055 }, { "epoch": 21.22762148337596, "grad_norm": 0.05126953125, "learning_rate": 0.0023017902813299235, "loss": 1.8601, "step": 41500 }, { "epoch": 21.483375959079282, "grad_norm": 0.056884765625, "learning_rate": 0.0022813299232736575, "loss": 1.8601, "step": 42000 }, { "epoch": 21.73913043478261, "grad_norm": 0.045166015625, "learning_rate": 0.002260869565217391, "loss": 1.8602, "step": 42500 }, { "epoch": 21.994884910485933, "grad_norm": 0.036376953125, "learning_rate": 0.002240409207161125, "loss": 1.86, "step": 43000 }, { "epoch": 22.0, "eval_loss": 3.72686505317688, "eval_runtime": 1.2281, "eval_samples_per_second": 407.144, "eval_steps_per_second": 0.814, "step": 43010 }, { "epoch": 22.25063938618926, "grad_norm": 0.04931640625, "learning_rate": 0.0022199488491048596, "loss": 1.86, "step": 43500 }, { "epoch": 22.506393861892583, "grad_norm": 0.0380859375, "learning_rate": 0.0021994884910485936, "loss": 1.8596, "step": 44000 }, { "epoch": 22.762148337595907, "grad_norm": 0.039306640625, "learning_rate": 0.0021790281329923276, "loss": 1.8602, "step": 44500 }, { "epoch": 23.0, "eval_loss": 3.7282092571258545, "eval_runtime": 1.1696, "eval_samples_per_second": 427.495, "eval_steps_per_second": 0.855, "step": 44965 }, { "epoch": 23.017902813299234, "grad_norm": 0.046142578125, "learning_rate": 0.002158567774936061, "loss": 1.86, "step": 45000 }, { "epoch": 23.273657289002557, "grad_norm": 0.039306640625, "learning_rate": 0.002138107416879795, "loss": 1.8598, "step": 45500 }, { "epoch": 23.529411764705884, "grad_norm": 0.04296875, "learning_rate": 0.0021176470588235297, "loss": 1.86, "step": 46000 }, { "epoch": 23.785166240409207, "grad_norm": 0.054931640625, "learning_rate": 0.0020971867007672637, "loss": 1.8596, "step": 46500 }, { "epoch": 24.0, "eval_loss": 3.7264657020568848, "eval_runtime": 1.1709, "eval_samples_per_second": 427.022, "eval_steps_per_second": 0.854, "step": 46920 }, { "epoch": 24.04092071611253, "grad_norm": 0.0419921875, "learning_rate": 0.0020767263427109977, "loss": 1.8599, "step": 47000 }, { "epoch": 24.296675191815858, "grad_norm": 0.043701171875, "learning_rate": 0.0020562659846547313, "loss": 1.8603, "step": 47500 }, { "epoch": 24.55242966751918, "grad_norm": 0.06884765625, "learning_rate": 0.0020358056265984653, "loss": 1.8592, "step": 48000 }, { "epoch": 24.808184143222505, "grad_norm": 0.037353515625, "learning_rate": 0.0020153452685421997, "loss": 1.8596, "step": 48500 }, { "epoch": 25.0, "eval_loss": 3.7268757820129395, "eval_runtime": 1.2077, "eval_samples_per_second": 414.025, "eval_steps_per_second": 0.828, "step": 48875 }, { "epoch": 25.06393861892583, "grad_norm": 0.0498046875, "learning_rate": 0.0019948849104859333, "loss": 1.8595, "step": 49000 }, { "epoch": 25.319693094629155, "grad_norm": 0.0419921875, "learning_rate": 0.0019744245524296678, "loss": 1.8592, "step": 49500 }, { "epoch": 25.575447570332482, "grad_norm": 0.0478515625, "learning_rate": 0.001953964194373402, "loss": 1.8595, "step": 50000 }, { "epoch": 25.831202046035806, "grad_norm": 0.04052734375, "learning_rate": 0.0019335038363171356, "loss": 1.8598, "step": 50500 }, { "epoch": 26.0, "eval_loss": 3.7266719341278076, "eval_runtime": 1.1718, "eval_samples_per_second": 426.711, "eval_steps_per_second": 0.853, "step": 50830 }, { "epoch": 26.08695652173913, "grad_norm": 0.046142578125, "learning_rate": 0.0019130434782608696, "loss": 1.8595, "step": 51000 }, { "epoch": 26.342710997442456, "grad_norm": 0.043701171875, "learning_rate": 0.0018925831202046036, "loss": 1.8595, "step": 51500 }, { "epoch": 26.59846547314578, "grad_norm": 0.03369140625, "learning_rate": 0.0018721227621483376, "loss": 1.8593, "step": 52000 }, { "epoch": 26.854219948849106, "grad_norm": 0.03662109375, "learning_rate": 0.0018516624040920714, "loss": 1.8595, "step": 52500 }, { "epoch": 27.0, "eval_loss": 3.72537899017334, "eval_runtime": 1.2057, "eval_samples_per_second": 414.687, "eval_steps_per_second": 0.829, "step": 52785 }, { "epoch": 27.10997442455243, "grad_norm": 0.044921875, "learning_rate": 0.0018312020460358057, "loss": 1.8596, "step": 53000 }, { "epoch": 27.365728900255753, "grad_norm": 0.04248046875, "learning_rate": 0.0018107416879795397, "loss": 1.8592, "step": 53500 }, { "epoch": 27.62148337595908, "grad_norm": 0.04833984375, "learning_rate": 0.0017902813299232737, "loss": 1.8598, "step": 54000 }, { "epoch": 27.877237851662404, "grad_norm": 0.0400390625, "learning_rate": 0.0017698209718670077, "loss": 1.8592, "step": 54500 }, { "epoch": 28.0, "eval_loss": 3.7256040573120117, "eval_runtime": 1.1731, "eval_samples_per_second": 426.237, "eval_steps_per_second": 0.852, "step": 54740 }, { "epoch": 28.132992327365727, "grad_norm": 0.045654296875, "learning_rate": 0.0017493606138107415, "loss": 1.8593, "step": 55000 }, { "epoch": 28.388746803069054, "grad_norm": 0.0439453125, "learning_rate": 0.0017289002557544758, "loss": 1.8589, "step": 55500 }, { "epoch": 28.644501278772378, "grad_norm": 0.047607421875, "learning_rate": 0.0017084398976982098, "loss": 1.8591, "step": 56000 }, { "epoch": 28.900255754475705, "grad_norm": 0.05029296875, "learning_rate": 0.0016879795396419438, "loss": 1.8592, "step": 56500 }, { "epoch": 29.0, "eval_loss": 3.7257955074310303, "eval_runtime": 1.1794, "eval_samples_per_second": 423.928, "eval_steps_per_second": 0.848, "step": 56695 }, { "epoch": 29.156010230179028, "grad_norm": 0.045166015625, "learning_rate": 0.0016675191815856778, "loss": 1.8589, "step": 57000 }, { "epoch": 29.41176470588235, "grad_norm": 0.037109375, "learning_rate": 0.0016470588235294116, "loss": 1.8591, "step": 57500 }, { "epoch": 29.66751918158568, "grad_norm": 0.0390625, "learning_rate": 0.0016265984654731459, "loss": 1.8591, "step": 58000 }, { "epoch": 29.923273657289002, "grad_norm": 0.052734375, "learning_rate": 0.0016061381074168799, "loss": 1.8596, "step": 58500 }, { "epoch": 30.0, "eval_loss": 3.7249155044555664, "eval_runtime": 1.1467, "eval_samples_per_second": 436.021, "eval_steps_per_second": 0.872, "step": 58650 } ], "logging_steps": 500, "max_steps": 97750, "num_input_tokens_seen": 0, "num_train_epochs": 50, "save_steps": 500, "stateful_callbacks": { "EarlyStoppingCallback": { "args": { "early_stopping_patience": 3, "early_stopping_threshold": 0.0 }, "attributes": { "early_stopping_patience_counter": 0 } }, "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 8.390099826432e+16, "train_batch_size": 512, "trial_name": null, "trial_params": null }