{ "best_metric": null, "best_model_checkpoint": null, "epoch": 121.21212121212122, "eval_steps": 500, "global_step": 250, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 4.848484848484849, "grad_norm": 0.5543263554573059, "learning_rate": 0.0001992114701314478, "loss": 2.7902, "step": 10 }, { "epoch": 9.696969696969697, "grad_norm": 0.6367977857589722, "learning_rate": 0.0001968583161128631, "loss": 2.3062, "step": 20 }, { "epoch": 14.545454545454545, "grad_norm": 0.479856014251709, "learning_rate": 0.00019297764858882514, "loss": 1.7416, "step": 30 }, { "epoch": 19.393939393939394, "grad_norm": 0.29826533794403076, "learning_rate": 0.00018763066800438636, "loss": 1.4342, "step": 40 }, { "epoch": 24.242424242424242, "grad_norm": 0.27084779739379883, "learning_rate": 0.00018090169943749476, "loss": 1.3399, "step": 50 }, { "epoch": 29.09090909090909, "grad_norm": 0.2398836314678192, "learning_rate": 0.00017289686274214118, "loss": 1.2648, "step": 60 }, { "epoch": 33.93939393939394, "grad_norm": 0.22510196268558502, "learning_rate": 0.000163742398974869, "loss": 1.2083, "step": 70 }, { "epoch": 38.78787878787879, "grad_norm": 0.2365349531173706, "learning_rate": 0.00015358267949789966, "loss": 1.1641, "step": 80 }, { "epoch": 43.63636363636363, "grad_norm": 0.270672082901001, "learning_rate": 0.00014257792915650728, "loss": 1.1283, "step": 90 }, { "epoch": 48.484848484848484, "grad_norm": 0.2681578993797302, "learning_rate": 0.00013090169943749476, "loss": 1.0801, "step": 100 }, { "epoch": 53.333333333333336, "grad_norm": 0.31493374705314636, "learning_rate": 0.00011873813145857249, "loss": 1.0366, "step": 110 }, { "epoch": 58.18181818181818, "grad_norm": 0.3526628315448761, "learning_rate": 0.00010627905195293135, "loss": 1.0097, "step": 120 }, { "epoch": 63.03030303030303, "grad_norm": 0.31481245160102844, "learning_rate": 9.372094804706867e-05, "loss": 0.9806, "step": 130 }, { "epoch": 67.87878787878788, "grad_norm": 0.3324221074581146, "learning_rate": 8.126186854142752e-05, "loss": 0.9474, "step": 140 }, { "epoch": 72.72727272727273, "grad_norm": 0.32580453157424927, "learning_rate": 6.909830056250527e-05, "loss": 0.9228, "step": 150 }, { "epoch": 77.57575757575758, "grad_norm": 0.3470986485481262, "learning_rate": 5.7422070843492734e-05, "loss": 0.9007, "step": 160 }, { "epoch": 82.42424242424242, "grad_norm": 0.4177663326263428, "learning_rate": 4.6417320502100316e-05, "loss": 0.8839, "step": 170 }, { "epoch": 87.27272727272727, "grad_norm": 0.4714958965778351, "learning_rate": 3.6257601025131026e-05, "loss": 0.8706, "step": 180 }, { "epoch": 92.12121212121212, "grad_norm": 0.4196126163005829, "learning_rate": 2.7103137257858868e-05, "loss": 0.8541, "step": 190 }, { "epoch": 96.96969696969697, "grad_norm": 0.35907700657844543, "learning_rate": 1.9098300562505266e-05, "loss": 0.8503, "step": 200 }, { "epoch": 101.81818181818181, "grad_norm": 0.40164825320243835, "learning_rate": 1.2369331995613665e-05, "loss": 0.8449, "step": 210 }, { "epoch": 106.66666666666667, "grad_norm": 0.37763774394989014, "learning_rate": 7.022351411174866e-06, "loss": 0.8384, "step": 220 }, { "epoch": 111.51515151515152, "grad_norm": 0.3757176101207733, "learning_rate": 3.1416838871368924e-06, "loss": 0.8405, "step": 230 }, { "epoch": 116.36363636363636, "grad_norm": 0.4176121950149536, "learning_rate": 7.885298685522235e-07, "loss": 0.8299, "step": 240 }, { "epoch": 121.21212121212122, "grad_norm": 0.47404736280441284, "learning_rate": 0.0, "loss": 0.8342, "step": 250 } ], "logging_steps": 10, "max_steps": 250, "num_input_tokens_seen": 0, "num_train_epochs": 125, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 6950976392371200.0, "train_batch_size": 4, "trial_name": null, "trial_params": null }