|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.04622519577282749, |
|
"eval_steps": 76, |
|
"global_step": 152, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0003041131300843914, |
|
"grad_norm": NaN, |
|
"learning_rate": 4.000000000000001e-06, |
|
"loss": 0.0, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0003041131300843914, |
|
"eval_loss": NaN, |
|
"eval_runtime": 1407.213, |
|
"eval_samples_per_second": 3.935, |
|
"eval_steps_per_second": 0.984, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0006082262601687828, |
|
"grad_norm": NaN, |
|
"learning_rate": 8.000000000000001e-06, |
|
"loss": 0.0, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.0009123393902531741, |
|
"grad_norm": NaN, |
|
"learning_rate": 1.2e-05, |
|
"loss": 0.0, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.0012164525203375656, |
|
"grad_norm": NaN, |
|
"learning_rate": 1.6000000000000003e-05, |
|
"loss": 0.0, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.001520565650421957, |
|
"grad_norm": NaN, |
|
"learning_rate": 2e-05, |
|
"loss": 0.0, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.0018246787805063483, |
|
"grad_norm": NaN, |
|
"learning_rate": 2.4e-05, |
|
"loss": 0.0, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.0021287919105907396, |
|
"grad_norm": NaN, |
|
"learning_rate": 2.8000000000000003e-05, |
|
"loss": 0.0, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.002432905040675131, |
|
"grad_norm": NaN, |
|
"learning_rate": 3.2000000000000005e-05, |
|
"loss": 0.0, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.0027370181707595228, |
|
"grad_norm": NaN, |
|
"learning_rate": 3.6e-05, |
|
"loss": 0.0, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.003041131300843914, |
|
"grad_norm": NaN, |
|
"learning_rate": 4e-05, |
|
"loss": 0.0, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.0033452444309283054, |
|
"grad_norm": NaN, |
|
"learning_rate": 4.4000000000000006e-05, |
|
"loss": 0.0, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.0036493575610126966, |
|
"grad_norm": NaN, |
|
"learning_rate": 4.8e-05, |
|
"loss": 0.0, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.003953470691097088, |
|
"grad_norm": NaN, |
|
"learning_rate": 5.2000000000000004e-05, |
|
"loss": 0.0, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.004257583821181479, |
|
"grad_norm": NaN, |
|
"learning_rate": 5.6000000000000006e-05, |
|
"loss": 0.0, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.004561696951265871, |
|
"grad_norm": NaN, |
|
"learning_rate": 6e-05, |
|
"loss": 0.0, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.004865810081350262, |
|
"grad_norm": NaN, |
|
"learning_rate": 6.400000000000001e-05, |
|
"loss": 0.0, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.0051699232114346535, |
|
"grad_norm": NaN, |
|
"learning_rate": 6.800000000000001e-05, |
|
"loss": 0.0, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.0054740363415190455, |
|
"grad_norm": NaN, |
|
"learning_rate": 7.2e-05, |
|
"loss": 0.0, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.005778149471603437, |
|
"grad_norm": NaN, |
|
"learning_rate": 7.6e-05, |
|
"loss": 0.0, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.006082262601687828, |
|
"grad_norm": NaN, |
|
"learning_rate": 8e-05, |
|
"loss": 0.0, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.006386375731772219, |
|
"grad_norm": NaN, |
|
"learning_rate": 8.4e-05, |
|
"loss": 0.0, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.006690488861856611, |
|
"grad_norm": NaN, |
|
"learning_rate": 8.800000000000001e-05, |
|
"loss": 0.0, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.006994601991941002, |
|
"grad_norm": NaN, |
|
"learning_rate": 9.200000000000001e-05, |
|
"loss": 0.0, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.007298715122025393, |
|
"grad_norm": NaN, |
|
"learning_rate": 9.6e-05, |
|
"loss": 0.0, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.007602828252109785, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.0001, |
|
"loss": 0.0, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.007906941382194176, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00010400000000000001, |
|
"loss": 0.0, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.008211054512278567, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00010800000000000001, |
|
"loss": 0.0, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.008515167642362959, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00011200000000000001, |
|
"loss": 0.0, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.00881928077244735, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.000116, |
|
"loss": 0.0, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.009123393902531743, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00012, |
|
"loss": 0.0, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.009427507032616134, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.000124, |
|
"loss": 0.0, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.009731620162700525, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00012800000000000002, |
|
"loss": 0.0, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.010035733292784916, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.000132, |
|
"loss": 0.0, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.010339846422869307, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00013600000000000003, |
|
"loss": 0.0, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.010643959552953698, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00014, |
|
"loss": 0.0, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.010948072683038091, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.000144, |
|
"loss": 0.0, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.011252185813122482, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.000148, |
|
"loss": 0.0, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.011556298943206873, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.000152, |
|
"loss": 0.0, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.011860412073291264, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00015600000000000002, |
|
"loss": 0.0, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.012164525203375656, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00016, |
|
"loss": 0.0, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.012468638333460047, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.000164, |
|
"loss": 0.0, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.012772751463544438, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.000168, |
|
"loss": 0.0, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.01307686459362883, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.000172, |
|
"loss": 0.0, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.013380977723713222, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00017600000000000002, |
|
"loss": 0.0, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.013685090853797613, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00018, |
|
"loss": 0.0, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.013989203983882004, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00018400000000000003, |
|
"loss": 0.0, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.014293317113966395, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.000188, |
|
"loss": 0.0, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.014597430244050786, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.000192, |
|
"loss": 0.0, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.014901543374135177, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.000196, |
|
"loss": 0.0, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.01520565650421957, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.0002, |
|
"loss": 0.0, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.015509769634303961, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.0001999922905547776, |
|
"loss": 0.0, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.015813882764388353, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.0001999691634078213, |
|
"loss": 0.0, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.016117995894472745, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00019993062212508053, |
|
"loss": 0.0, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.016422109024557135, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.0001998766726491935, |
|
"loss": 0.0, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.016726222154641528, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00019980732329857076, |
|
"loss": 0.0, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.017030335284725917, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.0001997225847661127, |
|
"loss": 0.0, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.01733444841481031, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00019962247011756081, |
|
"loss": 0.0, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.0176385615448947, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00019950699478948309, |
|
"loss": 0.0, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.017942674674979092, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00019937617658689384, |
|
"loss": 0.0, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.018246787805063485, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00019923003568050844, |
|
"loss": 0.0, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.018550900935147874, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00019906859460363307, |
|
"loss": 0.0, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.018855014065232267, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.0001988918782486906, |
|
"loss": 0.0, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.019159127195316657, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.0001986999138633821, |
|
"loss": 0.0, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.01946324032540105, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00019849273104648592, |
|
"loss": 0.0, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.01976735345548544, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00019827036174329353, |
|
"loss": 0.0, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.020071466585569832, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00019803284024068427, |
|
"loss": 0.0, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.020375579715654225, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.0001977802031618383, |
|
"loss": 0.0, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.020679692845738614, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00019751248946059014, |
|
"loss": 0.0, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.020983805975823007, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00019722974041542203, |
|
"loss": 0.0, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.021287919105907396, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.0001969319996230995, |
|
"loss": 0.0, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.02159203223599179, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.0001966193129919491, |
|
"loss": 0.0, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.021896145366076182, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00019629172873477995, |
|
"loss": 0.0, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.02220025849616057, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00019594929736144976, |
|
"loss": 0.0, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.022504371626244964, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00019559207167107684, |
|
"loss": 0.0, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.022808484756329354, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.000195220106743899, |
|
"loss": 0.0, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.023112597886413747, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00019483345993278093, |
|
"loss": 0.0, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.023112597886413747, |
|
"eval_loss": NaN, |
|
"eval_runtime": 1405.4791, |
|
"eval_samples_per_second": 3.94, |
|
"eval_steps_per_second": 0.985, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.023416711016498136, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.0001944321908543708, |
|
"loss": 0.0, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.02372082414658253, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00019401636137990816, |
|
"loss": 0.0, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.02402493727666692, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00019358603562568416, |
|
"loss": 0.0, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.02432905040675131, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.0001931412799431554, |
|
"loss": 0.0, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.024633163536835704, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.0001926821629087133, |
|
"loss": 0.0, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.024937276666920093, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00019220875531311045, |
|
"loss": 0.0, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.025241389797004486, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00019172113015054532, |
|
"loss": 0.0, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.025545502927088876, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00019121936260740752, |
|
"loss": 0.0, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.02584961605717327, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00019070353005068484, |
|
"loss": 0.0, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.02615372918725766, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00019017371201603407, |
|
"loss": 0.0, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.02645784231734205, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00018962999019551754, |
|
"loss": 0.0, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.026761955447426444, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00018907244842500704, |
|
"loss": 0.0, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.027066068577510833, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00018850117267125738, |
|
"loss": 0.0, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.027370181707595226, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00018791625101865117, |
|
"loss": 0.0, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.027674294837679615, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.0001873177736556172, |
|
"loss": 0.0, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.027978407967764008, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00018670583286072443, |
|
"loss": 0.0, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.0282825210978484, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.0001860805229884536, |
|
"loss": 0.0, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.02858663422793279, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00018544194045464886, |
|
"loss": 0.0, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.028890747358017183, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.0001847901837216515, |
|
"loss": 0.0, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.029194860488101573, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00018412535328311814, |
|
"loss": 0.0, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.029498973618185965, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.0001834475516485257, |
|
"loss": 0.0, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.029803086748270355, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00018275688332736577, |
|
"loss": 0.0, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.030107199878354748, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00018205345481302998, |
|
"loss": 0.0, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.03041131300843914, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00018133737456639044, |
|
"loss": 0.0, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.03071542613852353, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.0001806087529990758, |
|
"loss": 0.0, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.031019539268607923, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.0001798677024564473, |
|
"loss": 0.0, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.031323652398692316, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00017911433720027624, |
|
"loss": 0.0, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.031627765528776705, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00017834877339112612, |
|
"loss": 0.0, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.031931878658861094, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.000177571129070442, |
|
"loss": 0.0, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.03223599178894549, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00017678152414234968, |
|
"loss": 0.0, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.03254010491902988, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.000175980080355168, |
|
"loss": 0.0, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.03284421804911427, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00017516692128263648, |
|
"loss": 0.0, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.03314833117919866, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00017434217230486164, |
|
"loss": 0.0, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.033452444309283055, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00017350596058898483, |
|
"loss": 0.0, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.033756557439367445, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.0001726584150695744, |
|
"loss": 0.0, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.034060670569451834, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00017179966642874543, |
|
"loss": 0.0, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.03436478369953623, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.0001709298470760101, |
|
"loss": 0.0, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.03466889682962062, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00017004909112786144, |
|
"loss": 0.0, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.03497300995970501, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00016915753438709417, |
|
"loss": 0.0, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.0352771230897894, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00016825531432186543, |
|
"loss": 0.0, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.035581236219873795, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00016734257004449862, |
|
"loss": 0.0, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.035885349349958184, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00016641944229003395, |
|
"loss": 0.0, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.036189462480042574, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00016548607339452853, |
|
"loss": 0.0, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.03649357561012697, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00016454260727310978, |
|
"loss": 0.0, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.03679768874021136, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00016358918939778536, |
|
"loss": 0.0, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.03710180187029575, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00016262596677501297, |
|
"loss": 0.0, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.03740591500038014, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.0001616530879230335, |
|
"loss": 0.0, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.037710028130464535, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00016067070284897137, |
|
"loss": 0.0, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.038014141260548924, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00015967896302570485, |
|
"loss": 0.0, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.03831825439063331, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.0001586780213685108, |
|
"loss": 0.0, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.03862236752071771, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00015766803221148673, |
|
"loss": 0.0, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.0389264806508021, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.0001566491512837543, |
|
"loss": 0.0, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.03923059378088649, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00015562153568544752, |
|
"loss": 0.0, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.03953470691097088, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00015458534386348966, |
|
"loss": 0.0, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.039838820041055274, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.0001535407355871626, |
|
"loss": 0.0, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.040142933171139664, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00015248787192347196, |
|
"loss": 0.0, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.04044704630122405, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00015142691521231267, |
|
"loss": 0.0, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.04075115943130845, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00015035802904143762, |
|
"loss": 0.0, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.04105527256139284, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00014928137822123452, |
|
"loss": 0.0, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.04135938569147723, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.0001481971287593138, |
|
"loss": 0.0, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.04166349882156162, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00014710544783491208, |
|
"loss": 0.0, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.041967611951646014, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00014600650377311522, |
|
"loss": 0.0, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.0422717250817304, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00014490046601890405, |
|
"loss": 0.0, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.04257583821181479, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00014378750511102826, |
|
"loss": 0.0, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.04287995134189919, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00014266779265571087, |
|
"loss": 0.0, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.04318406447198358, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00014154150130018866, |
|
"loss": 0.0, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.04348817760206797, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00014040880470609187, |
|
"loss": 0.0, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.043792290732152364, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00013926987752266735, |
|
"loss": 0.0, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.044096403862236754, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00013812489535984981, |
|
"loss": 0.0, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.04440051699232114, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00013697403476118454, |
|
"loss": 0.0, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.04470463012240553, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.0001358174731766064, |
|
"loss": 0.0, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.04500874325248993, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00013465538893507907, |
|
"loss": 0.0, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.04531285638257432, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00013348796121709862, |
|
"loss": 0.0, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.04561696951265871, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00013231537002706594, |
|
"loss": 0.0, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.045921082642743104, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.0001311377961655319, |
|
"loss": 0.0, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.04622519577282749, |
|
"grad_norm": NaN, |
|
"learning_rate": 0.00012995542120132017, |
|
"loss": 0.0, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.04622519577282749, |
|
"eval_loss": NaN, |
|
"eval_runtime": 1407.6583, |
|
"eval_samples_per_second": 3.934, |
|
"eval_steps_per_second": 0.984, |
|
"step": 152 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 303, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 76, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1.0424076796386345e+18, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|