|
{ |
|
"best_metric": 1.3216674327850342, |
|
"best_model_checkpoint": "model_training/reprover_proving/checkpoints-random-09-08-17-37/checkpoint-425", |
|
"epoch": 6.896551724137931, |
|
"eval_steps": 25, |
|
"global_step": 425, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.08113590263691683, |
|
"grad_norm": 23.381454467773438, |
|
"learning_rate": 5.102040816326531e-05, |
|
"loss": 13.7285, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.16227180527383367, |
|
"grad_norm": 14.872354507446289, |
|
"learning_rate": 0.00010204081632653062, |
|
"loss": 8.8328, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.2434077079107505, |
|
"grad_norm": 0.6985018849372864, |
|
"learning_rate": 0.00015306122448979594, |
|
"loss": 4.2009, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.32454361054766734, |
|
"grad_norm": 0.24331903457641602, |
|
"learning_rate": 0.00020408163265306123, |
|
"loss": 3.7482, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.4056795131845842, |
|
"grad_norm": 0.22634823620319366, |
|
"learning_rate": 0.00025510204081632655, |
|
"loss": 3.6305, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.4056795131845842, |
|
"eval_loss": 3.498396396636963, |
|
"eval_runtime": 42.1381, |
|
"eval_samples_per_second": 23.732, |
|
"eval_steps_per_second": 1.495, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.486815415821501, |
|
"grad_norm": 0.8556441068649292, |
|
"learning_rate": 0.0003061224489795919, |
|
"loss": 3.5821, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.5679513184584178, |
|
"grad_norm": 1.5031307935714722, |
|
"learning_rate": 0.00035714285714285714, |
|
"loss": 2.8239, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.6490872210953347, |
|
"grad_norm": 0.7835573554039001, |
|
"learning_rate": 0.00040816326530612246, |
|
"loss": 2.158, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.7302231237322515, |
|
"grad_norm": 0.49232250452041626, |
|
"learning_rate": 0.0004591836734693878, |
|
"loss": 1.8947, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.8113590263691683, |
|
"grad_norm": 0.5629563331604004, |
|
"learning_rate": 0.0004999935985425297, |
|
"loss": 1.7676, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.8113590263691683, |
|
"eval_loss": 1.5309563875198364, |
|
"eval_runtime": 41.744, |
|
"eval_samples_per_second": 23.956, |
|
"eval_steps_per_second": 1.509, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.8924949290060852, |
|
"grad_norm": 2.9327845573425293, |
|
"learning_rate": 0.0004997695819512612, |
|
"loss": 1.7081, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.973630831643002, |
|
"grad_norm": 0.5285796523094177, |
|
"learning_rate": 0.0004992258202402822, |
|
"loss": 1.6423, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 1.054766734279919, |
|
"grad_norm": 0.27397045493125916, |
|
"learning_rate": 0.0004983630095117843, |
|
"loss": 1.5808, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 1.1359026369168357, |
|
"grad_norm": 0.19275762140750885, |
|
"learning_rate": 0.0004971822543018662, |
|
"loss": 1.5512, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 1.2170385395537526, |
|
"grad_norm": 0.24138106405735016, |
|
"learning_rate": 0.0004956850661665511, |
|
"loss": 1.5823, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 1.2170385395537526, |
|
"eval_loss": 1.4183236360549927, |
|
"eval_runtime": 41.7019, |
|
"eval_samples_per_second": 23.98, |
|
"eval_steps_per_second": 1.511, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 1.2981744421906694, |
|
"grad_norm": 0.27136123180389404, |
|
"learning_rate": 0.0004938733617467517, |
|
"loss": 1.5641, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 1.3793103448275863, |
|
"grad_norm": 0.17292170226573944, |
|
"learning_rate": 0.0004917494603146632, |
|
"loss": 1.532, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 1.460446247464503, |
|
"grad_norm": 0.3078462481498718, |
|
"learning_rate": 0.0004893160808047222, |
|
"loss": 1.5332, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 1.5415821501014197, |
|
"grad_norm": 0.21065960824489594, |
|
"learning_rate": 0.00048657633833293557, |
|
"loss": 1.5567, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 1.6227180527383367, |
|
"grad_norm": 0.3646738827228546, |
|
"learning_rate": 0.0004835337402090316, |
|
"loss": 1.5243, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 1.6227180527383367, |
|
"eval_loss": 1.4111393690109253, |
|
"eval_runtime": 41.9674, |
|
"eval_samples_per_second": 23.828, |
|
"eval_steps_per_second": 1.501, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 1.7038539553752536, |
|
"grad_norm": 1.1830989122390747, |
|
"learning_rate": 0.0004801921814465414, |
|
"loss": 1.5343, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 1.7849898580121704, |
|
"grad_norm": 0.285373717546463, |
|
"learning_rate": 0.00047655593977655674, |
|
"loss": 1.5254, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 1.866125760649087, |
|
"grad_norm": 0.27493494749069214, |
|
"learning_rate": 0.0004726296701715489, |
|
"loss": 1.5083, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 1.947261663286004, |
|
"grad_norm": 0.16316387057304382, |
|
"learning_rate": 0.00046841839888625623, |
|
"loss": 1.4953, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 2.028397565922921, |
|
"grad_norm": 0.1423080563545227, |
|
"learning_rate": 0.0004639275170232734, |
|
"loss": 1.5183, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 2.028397565922921, |
|
"eval_loss": 1.3866875171661377, |
|
"eval_runtime": 41.7133, |
|
"eval_samples_per_second": 23.973, |
|
"eval_steps_per_second": 1.51, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 2.109533468559838, |
|
"grad_norm": 0.11375881731510162, |
|
"learning_rate": 0.0004591627736315743, |
|
"loss": 1.4912, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 2.1906693711967544, |
|
"grad_norm": 0.18000248074531555, |
|
"learning_rate": 0.0004541302683468084, |
|
"loss": 1.4759, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 2.2718052738336714, |
|
"grad_norm": 0.2928003668785095, |
|
"learning_rate": 0.0004488364435827881, |
|
"loss": 1.4897, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 2.3529411764705883, |
|
"grad_norm": 0.2769507169723511, |
|
"learning_rate": 0.00044328807628416644, |
|
"loss": 1.5346, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 2.4340770791075053, |
|
"grad_norm": 0.3308449685573578, |
|
"learning_rate": 0.0004374922692508611, |
|
"loss": 1.5186, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 2.4340770791075053, |
|
"eval_loss": 1.394946813583374, |
|
"eval_runtime": 41.7538, |
|
"eval_samples_per_second": 23.95, |
|
"eval_steps_per_second": 1.509, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 2.5152129817444218, |
|
"grad_norm": 0.6130169034004211, |
|
"learning_rate": 0.0004314564420453311, |
|
"loss": 1.5019, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 2.5963488843813387, |
|
"grad_norm": 0.2663266062736511, |
|
"learning_rate": 0.0004251883214943475, |
|
"loss": 1.4881, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 2.6774847870182557, |
|
"grad_norm": 0.20422033965587616, |
|
"learning_rate": 0.0004186959317974155, |
|
"loss": 1.4864, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 2.7586206896551726, |
|
"grad_norm": 0.1979188621044159, |
|
"learning_rate": 0.00041198758425451266, |
|
"loss": 1.4771, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 2.839756592292089, |
|
"grad_norm": 0.11998000741004944, |
|
"learning_rate": 0.00040507186662629185, |
|
"loss": 1.486, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 2.839756592292089, |
|
"eval_loss": 1.363732933998108, |
|
"eval_runtime": 42.0053, |
|
"eval_samples_per_second": 23.807, |
|
"eval_steps_per_second": 1.5, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 2.920892494929006, |
|
"grad_norm": 0.48267611861228943, |
|
"learning_rate": 0.0003979576321403705, |
|
"loss": 1.4653, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 3.002028397565923, |
|
"grad_norm": 0.9161086082458496, |
|
"learning_rate": 0.0003906539881577793, |
|
"loss": 1.5062, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 3.08316430020284, |
|
"grad_norm": 0.3533921241760254, |
|
"learning_rate": 0.0003831702845140801, |
|
"loss": 1.5027, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 3.1643002028397564, |
|
"grad_norm": 0.3786797821521759, |
|
"learning_rate": 0.00037551610155007613, |
|
"loss": 1.5009, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 3.2454361054766734, |
|
"grad_norm": 0.3626890778541565, |
|
"learning_rate": 0.00036770123784744027, |
|
"loss": 1.4689, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 3.2454361054766734, |
|
"eval_loss": 1.3820106983184814, |
|
"eval_runtime": 41.9752, |
|
"eval_samples_per_second": 23.824, |
|
"eval_steps_per_second": 1.501, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 3.3265720081135903, |
|
"grad_norm": 0.5597763657569885, |
|
"learning_rate": 0.00035973569768495855, |
|
"loss": 1.4783, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 3.4077079107505073, |
|
"grad_norm": 0.22316162288188934, |
|
"learning_rate": 0.0003516296782314491, |
|
"loss": 1.458, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 3.4888438133874238, |
|
"grad_norm": 0.2465822696685791, |
|
"learning_rate": 0.00034339355649175095, |
|
"loss": 1.4803, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 3.5699797160243407, |
|
"grad_norm": 0.1777486950159073, |
|
"learning_rate": 0.00033503787602249364, |
|
"loss": 1.4493, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 3.6511156186612577, |
|
"grad_norm": 0.12559564411640167, |
|
"learning_rate": 0.00032657333343465356, |
|
"loss": 1.4349, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 3.6511156186612577, |
|
"eval_loss": 1.344743251800537, |
|
"eval_runtime": 41.8458, |
|
"eval_samples_per_second": 23.897, |
|
"eval_steps_per_second": 1.506, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 3.732251521298174, |
|
"grad_norm": 0.1277971714735031, |
|
"learning_rate": 0.0003180107647001769, |
|
"loss": 1.4245, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 3.813387423935091, |
|
"grad_norm": 0.13526912033557892, |
|
"learning_rate": 0.0003093611312801979, |
|
"loss": 1.4433, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 3.894523326572008, |
|
"grad_norm": 0.16969378292560577, |
|
"learning_rate": 0.00030063550609261025, |
|
"loss": 1.4468, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 3.975659229208925, |
|
"grad_norm": 0.12927880883216858, |
|
"learning_rate": 0.000291845059336957, |
|
"loss": 1.4555, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 4.056795131845842, |
|
"grad_norm": 0.12625518441200256, |
|
"learning_rate": 0.0002830010441947834, |
|
"loss": 1.4352, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 4.056795131845842, |
|
"eval_loss": 1.3391929864883423, |
|
"eval_runtime": 41.9506, |
|
"eval_samples_per_second": 23.838, |
|
"eval_steps_per_second": 1.502, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 4.137931034482759, |
|
"grad_norm": 0.09084944427013397, |
|
"learning_rate": 0.00027411478242376017, |
|
"loss": 1.4307, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 4.219066937119676, |
|
"grad_norm": 0.0855906680226326, |
|
"learning_rate": 0.00026519764986401774, |
|
"loss": 1.4426, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 4.300202839756592, |
|
"grad_norm": 0.1407158076763153, |
|
"learning_rate": 0.000256261061875247, |
|
"loss": 1.4097, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 4.381338742393509, |
|
"grad_norm": 0.2571726441383362, |
|
"learning_rate": 0.0002473164587232079, |
|
"loss": 1.4414, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 4.462474645030426, |
|
"grad_norm": 0.2552030682563782, |
|
"learning_rate": 0.0002383752909343547, |
|
"loss": 1.4387, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 4.462474645030426, |
|
"eval_loss": 1.3391071557998657, |
|
"eval_runtime": 41.9602, |
|
"eval_samples_per_second": 23.832, |
|
"eval_steps_per_second": 1.501, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 4.543610547667343, |
|
"grad_norm": 0.22119635343551636, |
|
"learning_rate": 0.0002294490046373259, |
|
"loss": 1.4405, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 4.62474645030426, |
|
"grad_norm": 0.1968819946050644, |
|
"learning_rate": 0.00022054902691006405, |
|
"loss": 1.4416, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 4.705882352941177, |
|
"grad_norm": 0.27252131700515747, |
|
"learning_rate": 0.00021168675115132315, |
|
"loss": 1.4322, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 4.787018255578094, |
|
"grad_norm": 0.30559447407722473, |
|
"learning_rate": 0.00020287352249529153, |
|
"loss": 1.4417, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 4.8681541582150105, |
|
"grad_norm": 0.202793151140213, |
|
"learning_rate": 0.00019412062328800044, |
|
"loss": 1.4194, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 4.8681541582150105, |
|
"eval_loss": 1.3287636041641235, |
|
"eval_runtime": 41.6499, |
|
"eval_samples_per_second": 24.01, |
|
"eval_steps_per_second": 1.513, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 4.9492900608519275, |
|
"grad_norm": 0.1644078493118286, |
|
"learning_rate": 0.000185439258644112, |
|
"loss": 1.4281, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 5.0304259634888435, |
|
"grad_norm": 0.36123207211494446, |
|
"learning_rate": 0.00017684054210257517, |
|
"loss": 1.3923, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 5.1115618661257605, |
|
"grad_norm": 1.3023096323013306, |
|
"learning_rate": 0.00016833548139951395, |
|
"loss": 1.4163, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 5.192697768762677, |
|
"grad_norm": 0.2667650878429413, |
|
"learning_rate": 0.0001599349643765599, |
|
"loss": 1.4299, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 5.273833671399594, |
|
"grad_norm": 0.6209434866905212, |
|
"learning_rate": 0.0001516497450426686, |
|
"loss": 1.4546, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 5.273833671399594, |
|
"eval_loss": 1.3360823392868042, |
|
"eval_runtime": 41.6974, |
|
"eval_samples_per_second": 23.982, |
|
"eval_steps_per_second": 1.511, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 5.354969574036511, |
|
"grad_norm": 0.3910903334617615, |
|
"learning_rate": 0.00014349042980726362, |
|
"loss": 1.4003, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 5.436105476673428, |
|
"grad_norm": 0.1629013866186142, |
|
"learning_rate": 0.0001354674639023318, |
|
"loss": 1.4308, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 5.517241379310345, |
|
"grad_norm": 0.11925800144672394, |
|
"learning_rate": 0.00012759111801085066, |
|
"loss": 1.4154, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 5.598377281947261, |
|
"grad_norm": 0.17776216566562653, |
|
"learning_rate": 0.00011987147511866788, |
|
"loss": 1.4322, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 5.679513184584178, |
|
"grad_norm": 0.17757736146450043, |
|
"learning_rate": 0.00011231841760666186, |
|
"loss": 1.4136, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 5.679513184584178, |
|
"eval_loss": 1.3244247436523438, |
|
"eval_runtime": 41.8991, |
|
"eval_samples_per_second": 23.867, |
|
"eval_steps_per_second": 1.504, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 5.760649087221095, |
|
"grad_norm": 0.16734164953231812, |
|
"learning_rate": 0.0001049416145997094, |
|
"loss": 1.434, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 5.841784989858012, |
|
"grad_norm": 0.1208333671092987, |
|
"learning_rate": 9.775050958865584e-05, |
|
"loss": 1.4383, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 5.922920892494929, |
|
"grad_norm": 0.09565079212188721, |
|
"learning_rate": 9.075430834113152e-05, |
|
"loss": 1.3854, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 6.004056795131846, |
|
"grad_norm": 0.08396822959184647, |
|
"learning_rate": 8.396196711669335e-05, |
|
"loss": 1.4139, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 6.085192697768763, |
|
"grad_norm": 0.08815029263496399, |
|
"learning_rate": 7.738218120137671e-05, |
|
"loss": 1.3972, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 6.085192697768763, |
|
"eval_loss": 1.3242241144180298, |
|
"eval_runtime": 41.9432, |
|
"eval_samples_per_second": 23.842, |
|
"eval_steps_per_second": 1.502, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 6.16632860040568, |
|
"grad_norm": 0.07709522545337677, |
|
"learning_rate": 7.102337377633394e-05, |
|
"loss": 1.4098, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 6.247464503042596, |
|
"grad_norm": 0.09170571714639664, |
|
"learning_rate": 6.489368513481228e-05, |
|
"loss": 1.4198, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 6.328600405679513, |
|
"grad_norm": 0.08237478137016296, |
|
"learning_rate": 5.9000962261273136e-05, |
|
"loss": 1.4168, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 6.40973630831643, |
|
"grad_norm": 0.1285756379365921, |
|
"learning_rate": 5.3352748785993164e-05, |
|
"loss": 1.4374, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 6.490872210953347, |
|
"grad_norm": 0.07434211671352386, |
|
"learning_rate": 4.795627532800806e-05, |
|
"loss": 1.4049, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 6.490872210953347, |
|
"eval_loss": 1.3239072561264038, |
|
"eval_runtime": 41.9088, |
|
"eval_samples_per_second": 23.861, |
|
"eval_steps_per_second": 1.503, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 6.572008113590264, |
|
"grad_norm": 0.07087664306163788, |
|
"learning_rate": 4.281845023876074e-05, |
|
"loss": 1.3987, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 6.653144016227181, |
|
"grad_norm": 0.05918489769101143, |
|
"learning_rate": 3.794585075830329e-05, |
|
"loss": 1.3934, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 6.734279918864098, |
|
"grad_norm": 0.06427238136529922, |
|
"learning_rate": 3.334471459537497e-05, |
|
"loss": 1.4361, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 6.8154158215010145, |
|
"grad_norm": 0.058615412563085556, |
|
"learning_rate": 2.902093194213526e-05, |
|
"loss": 1.4155, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 6.896551724137931, |
|
"grad_norm": 0.07955154776573181, |
|
"learning_rate": 2.4980037933772488e-05, |
|
"loss": 1.437, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 6.896551724137931, |
|
"eval_loss": 1.3216674327850342, |
|
"eval_runtime": 41.9961, |
|
"eval_samples_per_second": 23.812, |
|
"eval_steps_per_second": 1.5, |
|
"step": 425 |
|
} |
|
], |
|
"logging_steps": 5, |
|
"max_steps": 488, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 8, |
|
"save_steps": 25, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1.024840513050624e+17, |
|
"train_batch_size": 16, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|