|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 2.1880341880341883, |
|
"eval_steps": 100, |
|
"global_step": 2800, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.07814407814407814, |
|
"grad_norm": 0.389704167842865, |
|
"learning_rate": 4.000000000000001e-06, |
|
"loss": 2.4124, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.07814407814407814, |
|
"eval_loss": 2.2181031703948975, |
|
"eval_runtime": 3367.0243, |
|
"eval_samples_per_second": 0.75, |
|
"eval_steps_per_second": 0.094, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.1562881562881563, |
|
"grad_norm": 0.27927058935165405, |
|
"learning_rate": 8.000000000000001e-06, |
|
"loss": 2.2928, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.1562881562881563, |
|
"eval_loss": 2.0591659545898438, |
|
"eval_runtime": 3366.419, |
|
"eval_samples_per_second": 0.75, |
|
"eval_steps_per_second": 0.094, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.23443223443223443, |
|
"grad_norm": 0.22626566886901855, |
|
"learning_rate": 1.2e-05, |
|
"loss": 2.1104, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.23443223443223443, |
|
"eval_loss": 1.8923490047454834, |
|
"eval_runtime": 3360.9485, |
|
"eval_samples_per_second": 0.752, |
|
"eval_steps_per_second": 0.094, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.3125763125763126, |
|
"grad_norm": 0.210145503282547, |
|
"learning_rate": 1.6000000000000003e-05, |
|
"loss": 1.9296, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.3125763125763126, |
|
"eval_loss": 1.7505871057510376, |
|
"eval_runtime": 3357.6413, |
|
"eval_samples_per_second": 0.752, |
|
"eval_steps_per_second": 0.094, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.3907203907203907, |
|
"grad_norm": 0.2200106978416443, |
|
"learning_rate": 2e-05, |
|
"loss": 1.8345, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.3907203907203907, |
|
"eval_loss": 1.7071024179458618, |
|
"eval_runtime": 3361.7985, |
|
"eval_samples_per_second": 0.751, |
|
"eval_steps_per_second": 0.094, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.46886446886446886, |
|
"grad_norm": 0.22381120920181274, |
|
"learning_rate": 1.9980028422948323e-05, |
|
"loss": 1.8103, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.46886446886446886, |
|
"eval_loss": 1.6838117837905884, |
|
"eval_runtime": 3360.7216, |
|
"eval_samples_per_second": 0.752, |
|
"eval_steps_per_second": 0.094, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.5470085470085471, |
|
"grad_norm": 0.2300158590078354, |
|
"learning_rate": 1.9920193464571277e-05, |
|
"loss": 1.7827, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.5470085470085471, |
|
"eval_loss": 1.6691471338272095, |
|
"eval_runtime": 3362.2, |
|
"eval_samples_per_second": 0.751, |
|
"eval_steps_per_second": 0.094, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.6251526251526252, |
|
"grad_norm": 0.25236302614212036, |
|
"learning_rate": 1.982073412456518e-05, |
|
"loss": 1.765, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.6251526251526252, |
|
"eval_loss": 1.6575521230697632, |
|
"eval_runtime": 3363.3293, |
|
"eval_samples_per_second": 0.751, |
|
"eval_steps_per_second": 0.094, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.7032967032967034, |
|
"grad_norm": 0.26079344749450684, |
|
"learning_rate": 1.9682047674904527e-05, |
|
"loss": 1.7425, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.7032967032967034, |
|
"eval_loss": 1.6489633321762085, |
|
"eval_runtime": 3359.1797, |
|
"eval_samples_per_second": 0.752, |
|
"eval_steps_per_second": 0.094, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.7814407814407814, |
|
"grad_norm": 0.2565068304538727, |
|
"learning_rate": 1.9504688073012397e-05, |
|
"loss": 1.7521, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.7814407814407814, |
|
"eval_loss": 1.6409879922866821, |
|
"eval_runtime": 3358.9412, |
|
"eval_samples_per_second": 0.752, |
|
"eval_steps_per_second": 0.094, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.8595848595848596, |
|
"grad_norm": 0.26189491152763367, |
|
"learning_rate": 1.9289363749079798e-05, |
|
"loss": 1.7326, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.8595848595848596, |
|
"eval_loss": 1.6336146593093872, |
|
"eval_runtime": 3363.4368, |
|
"eval_samples_per_second": 0.751, |
|
"eval_steps_per_second": 0.094, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.9377289377289377, |
|
"grad_norm": 0.2778639495372772, |
|
"learning_rate": 1.903693477637204e-05, |
|
"loss": 1.7298, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.9377289377289377, |
|
"eval_loss": 1.627331256866455, |
|
"eval_runtime": 3361.7893, |
|
"eval_samples_per_second": 0.751, |
|
"eval_steps_per_second": 0.094, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 1.0158730158730158, |
|
"grad_norm": 0.2954370975494385, |
|
"learning_rate": 1.874840943582482e-05, |
|
"loss": 1.7128, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 1.0158730158730158, |
|
"eval_loss": 1.62191641330719, |
|
"eval_runtime": 3358.2318, |
|
"eval_samples_per_second": 0.752, |
|
"eval_steps_per_second": 0.094, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 1.0940170940170941, |
|
"grad_norm": 0.2985123097896576, |
|
"learning_rate": 1.842494018865216e-05, |
|
"loss": 1.7265, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 1.0940170940170941, |
|
"eval_loss": 1.616808295249939, |
|
"eval_runtime": 3366.219, |
|
"eval_samples_per_second": 0.75, |
|
"eval_steps_per_second": 0.094, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 1.1721611721611722, |
|
"grad_norm": 0.3037821054458618, |
|
"learning_rate": 1.8067819073052813e-05, |
|
"loss": 1.7148, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 1.1721611721611722, |
|
"eval_loss": 1.612409234046936, |
|
"eval_runtime": 3364.8773, |
|
"eval_samples_per_second": 0.751, |
|
"eval_steps_per_second": 0.094, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 1.2503052503052503, |
|
"grad_norm": 0.3078468441963196, |
|
"learning_rate": 1.7678472543402166e-05, |
|
"loss": 1.7003, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 1.2503052503052503, |
|
"eval_loss": 1.607498049736023, |
|
"eval_runtime": 3366.2319, |
|
"eval_samples_per_second": 0.75, |
|
"eval_steps_per_second": 0.094, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 1.3284493284493284, |
|
"grad_norm": 0.3602350056171417, |
|
"learning_rate": 1.7258455772543573e-05, |
|
"loss": 1.6981, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 1.3284493284493284, |
|
"eval_loss": 1.6040035486221313, |
|
"eval_runtime": 3369.5225, |
|
"eval_samples_per_second": 0.75, |
|
"eval_steps_per_second": 0.094, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 1.4065934065934065, |
|
"grad_norm": 0.34948959946632385, |
|
"learning_rate": 1.6809446439937472e-05, |
|
"loss": 1.703, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 1.4065934065934065, |
|
"eval_loss": 1.6002304553985596, |
|
"eval_runtime": 3359.2322, |
|
"eval_samples_per_second": 0.752, |
|
"eval_steps_per_second": 0.094, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 1.4847374847374848, |
|
"grad_norm": 0.33941686153411865, |
|
"learning_rate": 1.6333238030480473e-05, |
|
"loss": 1.6842, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 1.4847374847374848, |
|
"eval_loss": 1.5976512432098389, |
|
"eval_runtime": 3363.9968, |
|
"eval_samples_per_second": 0.751, |
|
"eval_steps_per_second": 0.094, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 1.5628815628815629, |
|
"grad_norm": 0.3594741225242615, |
|
"learning_rate": 1.5831732670761e-05, |
|
"loss": 1.6813, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 1.5628815628815629, |
|
"eval_loss": 1.5942325592041016, |
|
"eval_runtime": 3356.6247, |
|
"eval_samples_per_second": 0.753, |
|
"eval_steps_per_second": 0.094, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 1.641025641025641, |
|
"grad_norm": 0.3513031303882599, |
|
"learning_rate": 1.5306933531365748e-05, |
|
"loss": 1.6747, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 1.641025641025641, |
|
"eval_loss": 1.591470718383789, |
|
"eval_runtime": 3360.2103, |
|
"eval_samples_per_second": 0.752, |
|
"eval_steps_per_second": 0.094, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 1.7191697191697193, |
|
"grad_norm": 0.3548774719238281, |
|
"learning_rate": 1.4760936825584535e-05, |
|
"loss": 1.6821, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 1.7191697191697193, |
|
"eval_loss": 1.5890487432479858, |
|
"eval_runtime": 3360.7245, |
|
"eval_samples_per_second": 0.752, |
|
"eval_steps_per_second": 0.094, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 1.7973137973137974, |
|
"grad_norm": 0.37473219633102417, |
|
"learning_rate": 1.4195923436473257e-05, |
|
"loss": 1.685, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 1.7973137973137974, |
|
"eval_loss": 1.586594820022583, |
|
"eval_runtime": 3365.8879, |
|
"eval_samples_per_second": 0.75, |
|
"eval_steps_per_second": 0.094, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 1.8754578754578755, |
|
"grad_norm": 0.37071430683135986, |
|
"learning_rate": 1.3614150205719086e-05, |
|
"loss": 1.6602, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 1.8754578754578755, |
|
"eval_loss": 1.5838119983673096, |
|
"eval_runtime": 3363.0708, |
|
"eval_samples_per_second": 0.751, |
|
"eval_steps_per_second": 0.094, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 1.9536019536019538, |
|
"grad_norm": 0.4014996588230133, |
|
"learning_rate": 1.3017940919102943e-05, |
|
"loss": 1.682, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 1.9536019536019538, |
|
"eval_loss": 1.5820069313049316, |
|
"eval_runtime": 3367.2545, |
|
"eval_samples_per_second": 0.75, |
|
"eval_steps_per_second": 0.094, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 2.0317460317460316, |
|
"grad_norm": 0.3767947256565094, |
|
"learning_rate": 1.2409677024566145e-05, |
|
"loss": 1.6595, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 2.0317460317460316, |
|
"eval_loss": 1.5797280073165894, |
|
"eval_runtime": 3364.0566, |
|
"eval_samples_per_second": 0.751, |
|
"eval_steps_per_second": 0.094, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 2.10989010989011, |
|
"grad_norm": 0.3852591812610626, |
|
"learning_rate": 1.1791788119956191e-05, |
|
"loss": 1.6606, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 2.10989010989011, |
|
"eval_loss": 1.578330636024475, |
|
"eval_runtime": 3359.3638, |
|
"eval_samples_per_second": 0.752, |
|
"eval_steps_per_second": 0.094, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 2.1880341880341883, |
|
"grad_norm": 0.3954846262931824, |
|
"learning_rate": 1.116674224844664e-05, |
|
"loss": 1.6611, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 2.1880341880341883, |
|
"eval_loss": 1.5769294500350952, |
|
"eval_runtime": 3365.2078, |
|
"eval_samples_per_second": 0.751, |
|
"eval_steps_per_second": 0.094, |
|
"step": 2800 |
|
} |
|
], |
|
"logging_steps": 100, |
|
"max_steps": 5470, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 5, |
|
"save_steps": 100, |
|
"total_flos": 1.141344360972288e+18, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|