|
{ |
|
"best_metric": 1.191159963607788, |
|
"best_model_checkpoint": "./output/training_results/C021_random_sample_Meta-Llama-3-8B_pretrain_20240726_033210/checkpoint-33984", |
|
"epoch": 4.0, |
|
"eval_steps": 3776, |
|
"global_step": 37752, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.00010595465140919687, |
|
"grad_norm": 0.0, |
|
"learning_rate": 0.0, |
|
"loss": 1.6347, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.20004238186056367, |
|
"grad_norm": 2.0841434899950975, |
|
"learning_rate": 1.9947033898305087e-06, |
|
"loss": 1.411, |
|
"step": 1888 |
|
}, |
|
{ |
|
"epoch": 0.40008476372112733, |
|
"grad_norm": 1.7712195190814046, |
|
"learning_rate": 2.23788694222373e-06, |
|
"loss": 1.3112, |
|
"step": 3776 |
|
}, |
|
{ |
|
"epoch": 0.40008476372112733, |
|
"eval_loss": 1.2742130756378174, |
|
"eval_runtime": 330.8268, |
|
"eval_samples_per_second": 202.868, |
|
"eval_steps_per_second": 1.587, |
|
"step": 3776 |
|
}, |
|
{ |
|
"epoch": 0.600127145581691, |
|
"grad_norm": 1.8384043597752484, |
|
"learning_rate": 1.2167168103937e-06, |
|
"loss": 1.2493, |
|
"step": 5664 |
|
}, |
|
{ |
|
"epoch": 0.8001695274422547, |
|
"grad_norm": 1.7725919716418708, |
|
"learning_rate": 6.493264000872854e-07, |
|
"loss": 1.2264, |
|
"step": 7552 |
|
}, |
|
{ |
|
"epoch": 0.8001695274422547, |
|
"eval_loss": 1.214762568473816, |
|
"eval_runtime": 327.7212, |
|
"eval_samples_per_second": 204.79, |
|
"eval_steps_per_second": 1.602, |
|
"step": 7552 |
|
}, |
|
{ |
|
"epoch": 1.0002119093028183, |
|
"grad_norm": 1.8972637627842048, |
|
"learning_rate": 3.4492111928598874e-07, |
|
"loss": 1.211, |
|
"step": 9440 |
|
}, |
|
{ |
|
"epoch": 1.200254291163382, |
|
"grad_norm": 1.8787635251660875, |
|
"learning_rate": 1.8831885638765917e-07, |
|
"loss": 1.1558, |
|
"step": 11328 |
|
}, |
|
{ |
|
"epoch": 1.200254291163382, |
|
"eval_loss": 1.2016733884811401, |
|
"eval_runtime": 327.5932, |
|
"eval_samples_per_second": 204.87, |
|
"eval_steps_per_second": 1.603, |
|
"step": 11328 |
|
}, |
|
{ |
|
"epoch": 1.4002966730239457, |
|
"grad_norm": 2.0582903963596952, |
|
"learning_rate": 1.1128443649930022e-07, |
|
"loss": 1.1568, |
|
"step": 13216 |
|
}, |
|
{ |
|
"epoch": 1.6003390548845093, |
|
"grad_norm": 1.8665084344236207, |
|
"learning_rate": 7.542962298572276e-08, |
|
"loss": 1.1544, |
|
"step": 15104 |
|
}, |
|
{ |
|
"epoch": 1.6003390548845093, |
|
"eval_loss": 1.1974143981933594, |
|
"eval_runtime": 327.7112, |
|
"eval_samples_per_second": 204.796, |
|
"eval_steps_per_second": 1.602, |
|
"step": 15104 |
|
}, |
|
{ |
|
"epoch": 1.8003814367450732, |
|
"grad_norm": 1.7528553673770804, |
|
"learning_rate": 5.978512044954537e-08, |
|
"loss": 1.1574, |
|
"step": 16992 |
|
}, |
|
{ |
|
"epoch": 2.0004238186056367, |
|
"grad_norm": 1.8861106343782263, |
|
"learning_rate": 5.343416604131081e-08, |
|
"loss": 1.1562, |
|
"step": 18880 |
|
}, |
|
{ |
|
"epoch": 2.0004238186056367, |
|
"eval_loss": 1.19540274143219, |
|
"eval_runtime": 328.044, |
|
"eval_samples_per_second": 204.588, |
|
"eval_steps_per_second": 1.6, |
|
"step": 18880 |
|
}, |
|
{ |
|
"epoch": 2.2004662004662006, |
|
"grad_norm": 1.9003730335704914, |
|
"learning_rate": 5.1080120038479373e-08, |
|
"loss": 1.1431, |
|
"step": 20768 |
|
}, |
|
{ |
|
"epoch": 2.400508582326764, |
|
"grad_norm": 1.941099899013768, |
|
"learning_rate": 5.029649877564949e-08, |
|
"loss": 1.1416, |
|
"step": 22656 |
|
}, |
|
{ |
|
"epoch": 2.400508582326764, |
|
"eval_loss": 1.1947613954544067, |
|
"eval_runtime": 327.5948, |
|
"eval_samples_per_second": 204.869, |
|
"eval_steps_per_second": 1.603, |
|
"step": 22656 |
|
}, |
|
{ |
|
"epoch": 2.600550964187328, |
|
"grad_norm": 1.9472750123018565, |
|
"learning_rate": 5.0068500018386117e-08, |
|
"loss": 1.1461, |
|
"step": 24544 |
|
}, |
|
{ |
|
"epoch": 2.8005933460478913, |
|
"grad_norm": 1.893076803756726, |
|
"learning_rate": 5.0012641588155837e-08, |
|
"loss": 1.1449, |
|
"step": 26432 |
|
}, |
|
{ |
|
"epoch": 2.8005933460478913, |
|
"eval_loss": 1.1934314966201782, |
|
"eval_runtime": 328.7941, |
|
"eval_samples_per_second": 204.122, |
|
"eval_steps_per_second": 1.597, |
|
"step": 26432 |
|
}, |
|
{ |
|
"epoch": 3.0006357279084552, |
|
"grad_norm": 2.082824554026489, |
|
"learning_rate": 5.0001709336153935e-08, |
|
"loss": 1.1464, |
|
"step": 28320 |
|
}, |
|
{ |
|
"epoch": 3.2006781097690187, |
|
"grad_norm": 1.9988580640926128, |
|
"learning_rate": 5.000014828801133e-08, |
|
"loss": 1.1401, |
|
"step": 30208 |
|
}, |
|
{ |
|
"epoch": 3.2006781097690187, |
|
"eval_loss": 1.1925288438796997, |
|
"eval_runtime": 327.8373, |
|
"eval_samples_per_second": 204.717, |
|
"eval_steps_per_second": 1.601, |
|
"step": 30208 |
|
}, |
|
{ |
|
"epoch": 3.4007204916295826, |
|
"grad_norm": 1.8907403603158888, |
|
"learning_rate": 5.0000006391422415e-08, |
|
"loss": 1.1397, |
|
"step": 32096 |
|
}, |
|
{ |
|
"epoch": 3.6007628734901465, |
|
"grad_norm": 1.9407599933229671, |
|
"learning_rate": 5.0000000076061e-08, |
|
"loss": 1.1382, |
|
"step": 33984 |
|
}, |
|
{ |
|
"epoch": 3.6007628734901465, |
|
"eval_loss": 1.191159963607788, |
|
"eval_runtime": 327.6723, |
|
"eval_samples_per_second": 204.82, |
|
"eval_steps_per_second": 1.602, |
|
"step": 33984 |
|
}, |
|
{ |
|
"epoch": 3.80080525535071, |
|
"grad_norm": 1.9516245761095283, |
|
"learning_rate": 5.000000000004096e-08, |
|
"loss": 1.1404, |
|
"step": 35872 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"step": 37752, |
|
"total_flos": 3947644028190720.0, |
|
"train_loss": 1.1804671342891009, |
|
"train_runtime": 53865.2221, |
|
"train_samples_per_second": 44.854, |
|
"train_steps_per_second": 0.701 |
|
} |
|
], |
|
"logging_steps": 1888, |
|
"max_steps": 37752, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 4, |
|
"save_steps": 3776, |
|
"total_flos": 3947644028190720.0, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|