|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 121.21212121212122, |
|
"eval_steps": 500, |
|
"global_step": 250, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 4.848484848484849, |
|
"grad_norm": 0.5543263554573059, |
|
"learning_rate": 0.0001992114701314478, |
|
"loss": 2.7902, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 9.696969696969697, |
|
"grad_norm": 0.6367977857589722, |
|
"learning_rate": 0.0001968583161128631, |
|
"loss": 2.3062, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 14.545454545454545, |
|
"grad_norm": 0.479856014251709, |
|
"learning_rate": 0.00019297764858882514, |
|
"loss": 1.7416, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 19.393939393939394, |
|
"grad_norm": 0.29826533794403076, |
|
"learning_rate": 0.00018763066800438636, |
|
"loss": 1.4342, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 24.242424242424242, |
|
"grad_norm": 0.27084779739379883, |
|
"learning_rate": 0.00018090169943749476, |
|
"loss": 1.3399, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 29.09090909090909, |
|
"grad_norm": 0.2398836314678192, |
|
"learning_rate": 0.00017289686274214118, |
|
"loss": 1.2648, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 33.93939393939394, |
|
"grad_norm": 0.22510196268558502, |
|
"learning_rate": 0.000163742398974869, |
|
"loss": 1.2083, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 38.78787878787879, |
|
"grad_norm": 0.2365349531173706, |
|
"learning_rate": 0.00015358267949789966, |
|
"loss": 1.1641, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 43.63636363636363, |
|
"grad_norm": 0.270672082901001, |
|
"learning_rate": 0.00014257792915650728, |
|
"loss": 1.1283, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 48.484848484848484, |
|
"grad_norm": 0.2681578993797302, |
|
"learning_rate": 0.00013090169943749476, |
|
"loss": 1.0801, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 53.333333333333336, |
|
"grad_norm": 0.31493374705314636, |
|
"learning_rate": 0.00011873813145857249, |
|
"loss": 1.0366, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 58.18181818181818, |
|
"grad_norm": 0.3526628315448761, |
|
"learning_rate": 0.00010627905195293135, |
|
"loss": 1.0097, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 63.03030303030303, |
|
"grad_norm": 0.31481245160102844, |
|
"learning_rate": 9.372094804706867e-05, |
|
"loss": 0.9806, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 67.87878787878788, |
|
"grad_norm": 0.3324221074581146, |
|
"learning_rate": 8.126186854142752e-05, |
|
"loss": 0.9474, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 72.72727272727273, |
|
"grad_norm": 0.32580453157424927, |
|
"learning_rate": 6.909830056250527e-05, |
|
"loss": 0.9228, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 77.57575757575758, |
|
"grad_norm": 0.3470986485481262, |
|
"learning_rate": 5.7422070843492734e-05, |
|
"loss": 0.9007, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 82.42424242424242, |
|
"grad_norm": 0.4177663326263428, |
|
"learning_rate": 4.6417320502100316e-05, |
|
"loss": 0.8839, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 87.27272727272727, |
|
"grad_norm": 0.4714958965778351, |
|
"learning_rate": 3.6257601025131026e-05, |
|
"loss": 0.8706, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 92.12121212121212, |
|
"grad_norm": 0.4196126163005829, |
|
"learning_rate": 2.7103137257858868e-05, |
|
"loss": 0.8541, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 96.96969696969697, |
|
"grad_norm": 0.35907700657844543, |
|
"learning_rate": 1.9098300562505266e-05, |
|
"loss": 0.8503, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 101.81818181818181, |
|
"grad_norm": 0.40164825320243835, |
|
"learning_rate": 1.2369331995613665e-05, |
|
"loss": 0.8449, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 106.66666666666667, |
|
"grad_norm": 0.37763774394989014, |
|
"learning_rate": 7.022351411174866e-06, |
|
"loss": 0.8384, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 111.51515151515152, |
|
"grad_norm": 0.3757176101207733, |
|
"learning_rate": 3.1416838871368924e-06, |
|
"loss": 0.8405, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 116.36363636363636, |
|
"grad_norm": 0.4176121950149536, |
|
"learning_rate": 7.885298685522235e-07, |
|
"loss": 0.8299, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 121.21212121212122, |
|
"grad_norm": 0.47404736280441284, |
|
"learning_rate": 0.0, |
|
"loss": 0.8342, |
|
"step": 250 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 250, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 125, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 6950976392371200.0, |
|
"train_batch_size": 4, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|