|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 2.2980251346499103, |
|
"eval_steps": 500, |
|
"global_step": 40, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.05745062836624776, |
|
"grad_norm": 3.1636736392974854, |
|
"learning_rate": 5e-06, |
|
"loss": 2.3037, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.11490125673249552, |
|
"grad_norm": 3.3778610229492188, |
|
"learning_rate": 1e-05, |
|
"loss": 2.3337, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.17235188509874327, |
|
"grad_norm": 3.0278897285461426, |
|
"learning_rate": 1.5e-05, |
|
"loss": 2.2715, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.22980251346499103, |
|
"grad_norm": 2.5405051708221436, |
|
"learning_rate": 2e-05, |
|
"loss": 2.2422, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.2872531418312388, |
|
"grad_norm": 1.789209246635437, |
|
"learning_rate": 2.5e-05, |
|
"loss": 2.2079, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.34470377019748655, |
|
"grad_norm": 1.653626561164856, |
|
"learning_rate": 3e-05, |
|
"loss": 2.2231, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.4021543985637343, |
|
"grad_norm": 2.006115436553955, |
|
"learning_rate": 3.5e-05, |
|
"loss": 2.2855, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.45960502692998206, |
|
"grad_norm": 2.322628974914551, |
|
"learning_rate": 4e-05, |
|
"loss": 2.2126, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.5170556552962298, |
|
"grad_norm": 2.069558620452881, |
|
"learning_rate": 4.5e-05, |
|
"loss": 2.2389, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.5745062836624776, |
|
"grad_norm": 1.7339693307876587, |
|
"learning_rate": 5e-05, |
|
"loss": 2.1964, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.6319569120287253, |
|
"grad_norm": 1.6608643531799316, |
|
"learning_rate": 4.992664502959351e-05, |
|
"loss": 2.2079, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.6894075403949731, |
|
"grad_norm": 1.642751932144165, |
|
"learning_rate": 4.970701059450872e-05, |
|
"loss": 2.2359, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.7468581687612208, |
|
"grad_norm": 1.6817108392715454, |
|
"learning_rate": 4.934238559694448e-05, |
|
"loss": 2.2229, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.8043087971274686, |
|
"grad_norm": 1.6007517576217651, |
|
"learning_rate": 4.8834909801373264e-05, |
|
"loss": 2.203, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.8617594254937163, |
|
"grad_norm": 1.457571029663086, |
|
"learning_rate": 4.8187561277552374e-05, |
|
"loss": 2.0994, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.9192100538599641, |
|
"grad_norm": 1.486559271812439, |
|
"learning_rate": 4.740413892402639e-05, |
|
"loss": 2.1777, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.9766606822262118, |
|
"grad_norm": 1.5398966073989868, |
|
"learning_rate": 4.648924017468003e-05, |
|
"loss": 2.2126, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 1.0341113105924595, |
|
"grad_norm": 1.4850305318832397, |
|
"learning_rate": 4.5448234019167945e-05, |
|
"loss": 2.1019, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 1.0915619389587075, |
|
"grad_norm": 1.4637010097503662, |
|
"learning_rate": 4.428722949554857e-05, |
|
"loss": 2.1119, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 1.1490125673249552, |
|
"grad_norm": 1.3282859325408936, |
|
"learning_rate": 4.301303984001967e-05, |
|
"loss": 2.0976, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 1.2064631956912029, |
|
"grad_norm": 1.2366828918457031, |
|
"learning_rate": 4.163314250413913e-05, |
|
"loss": 2.0682, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 1.2639138240574506, |
|
"grad_norm": 1.3221453428268433, |
|
"learning_rate": 4.015563527416595e-05, |
|
"loss": 2.0266, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 1.3213644524236985, |
|
"grad_norm": 1.269490122795105, |
|
"learning_rate": 3.858918875003053e-05, |
|
"loss": 2.011, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 1.3788150807899462, |
|
"grad_norm": 1.2723491191864014, |
|
"learning_rate": 3.694299546280657e-05, |
|
"loss": 2.0503, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 1.436265709156194, |
|
"grad_norm": 1.3157715797424316, |
|
"learning_rate": 3.5226715929283506e-05, |
|
"loss": 2.0297, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 1.4937163375224416, |
|
"grad_norm": 1.367810606956482, |
|
"learning_rate": 3.3450421960212566e-05, |
|
"loss": 2.03, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 1.5511669658886893, |
|
"grad_norm": 1.4925018548965454, |
|
"learning_rate": 3.162453755491655e-05, |
|
"loss": 2.0478, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 1.608617594254937, |
|
"grad_norm": 1.350865125656128, |
|
"learning_rate": 2.975977772911671e-05, |
|
"loss": 2.1047, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 1.666068222621185, |
|
"grad_norm": 1.3046050071716309, |
|
"learning_rate": 2.7867085634960016e-05, |
|
"loss": 2.0175, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 1.7235188509874326, |
|
"grad_norm": 1.3713139295578003, |
|
"learning_rate": 2.595756834225089e-05, |
|
"loss": 2.0437, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 1.7809694793536806, |
|
"grad_norm": 1.2804160118103027, |
|
"learning_rate": 2.4042431657749117e-05, |
|
"loss": 2.0573, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 1.8384201077199283, |
|
"grad_norm": 1.257683277130127, |
|
"learning_rate": 2.2132914365039993e-05, |
|
"loss": 2.0105, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 1.895870736086176, |
|
"grad_norm": 1.2399568557739258, |
|
"learning_rate": 2.0240222270883288e-05, |
|
"loss": 2.0335, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 1.9533213644524237, |
|
"grad_norm": 1.1331653594970703, |
|
"learning_rate": 1.8375462445083464e-05, |
|
"loss": 2.0217, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 2.0107719928186714, |
|
"grad_norm": 1.219581961631775, |
|
"learning_rate": 1.6549578039787436e-05, |
|
"loss": 2.0096, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 2.068222621184919, |
|
"grad_norm": 1.239498257637024, |
|
"learning_rate": 1.4773284070716503e-05, |
|
"loss": 2.0598, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 2.1256732495511668, |
|
"grad_norm": 1.2558954954147339, |
|
"learning_rate": 1.3057004537193423e-05, |
|
"loss": 2.0412, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 2.183123877917415, |
|
"grad_norm": 1.0945749282836914, |
|
"learning_rate": 1.1410811249969475e-05, |
|
"loss": 1.959, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 2.2405745062836626, |
|
"grad_norm": 1.0557724237442017, |
|
"learning_rate": 9.844364725834057e-06, |
|
"loss": 1.9716, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 2.2980251346499103, |
|
"grad_norm": 1.1883456707000732, |
|
"learning_rate": 8.36685749586087e-06, |
|
"loss": 1.9485, |
|
"step": 40 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 51, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 4, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 2.7513920113127424e+16, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|