|
{ |
|
"best_metric": 0.0008496911614201963, |
|
"best_model_checkpoint": "/home/paperspace/Data/models/akoul_whitehorseliquidity_25c/llm3br256/checkpoint-900", |
|
"epoch": 3.313131313131313, |
|
"eval_steps": 25, |
|
"global_step": 1025, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0032323232323232323, |
|
"grad_norm": 0.03562889248132706, |
|
"learning_rate": 1.2936610608020701e-07, |
|
"loss": 0.0088, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.006464646464646465, |
|
"grad_norm": 0.03539334237575531, |
|
"learning_rate": 2.5873221216041403e-07, |
|
"loss": 0.0098, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.009696969696969697, |
|
"grad_norm": 0.03888387605547905, |
|
"learning_rate": 3.8809831824062096e-07, |
|
"loss": 0.0103, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.01292929292929293, |
|
"grad_norm": 0.03283314406871796, |
|
"learning_rate": 5.174644243208281e-07, |
|
"loss": 0.0123, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.01616161616161616, |
|
"grad_norm": 0.033392902463674545, |
|
"learning_rate": 6.468305304010349e-07, |
|
"loss": 0.0104, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.019393939393939394, |
|
"grad_norm": 0.041360873728990555, |
|
"learning_rate": 7.761966364812419e-07, |
|
"loss": 0.0135, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.022626262626262626, |
|
"grad_norm": 0.03297152742743492, |
|
"learning_rate": 9.055627425614489e-07, |
|
"loss": 0.01, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.02585858585858586, |
|
"grad_norm": 0.03385542333126068, |
|
"learning_rate": 1.0349288486416561e-06, |
|
"loss": 0.0092, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.02909090909090909, |
|
"grad_norm": 0.038460493087768555, |
|
"learning_rate": 1.164294954721863e-06, |
|
"loss": 0.0132, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.03232323232323232, |
|
"grad_norm": 0.030043406412005424, |
|
"learning_rate": 1.2936610608020699e-06, |
|
"loss": 0.0088, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.035555555555555556, |
|
"grad_norm": 0.039398446679115295, |
|
"learning_rate": 1.423027166882277e-06, |
|
"loss": 0.0118, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.03878787878787879, |
|
"grad_norm": 0.03809528797864914, |
|
"learning_rate": 1.5523932729624839e-06, |
|
"loss": 0.0146, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.04202020202020202, |
|
"grad_norm": 0.0388377383351326, |
|
"learning_rate": 1.6817593790426907e-06, |
|
"loss": 0.0102, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.04525252525252525, |
|
"grad_norm": 0.043977439403533936, |
|
"learning_rate": 1.8111254851228978e-06, |
|
"loss": 0.0099, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.048484848484848485, |
|
"grad_norm": 0.036720160394907, |
|
"learning_rate": 1.9404915912031045e-06, |
|
"loss": 0.0143, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.05171717171717172, |
|
"grad_norm": 0.03513036668300629, |
|
"learning_rate": 2.0698576972833122e-06, |
|
"loss": 0.0092, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.05494949494949495, |
|
"grad_norm": 0.03713906928896904, |
|
"learning_rate": 2.199223803363519e-06, |
|
"loss": 0.0144, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.05818181818181818, |
|
"grad_norm": 0.0338338240981102, |
|
"learning_rate": 2.328589909443726e-06, |
|
"loss": 0.0133, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.061414141414141414, |
|
"grad_norm": 0.0263836532831192, |
|
"learning_rate": 2.457956015523933e-06, |
|
"loss": 0.0113, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.06464646464646465, |
|
"grad_norm": 0.01817137934267521, |
|
"learning_rate": 2.5873221216041398e-06, |
|
"loss": 0.0074, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.06787878787878789, |
|
"grad_norm": 0.018132060766220093, |
|
"learning_rate": 2.716688227684347e-06, |
|
"loss": 0.008, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.07111111111111111, |
|
"grad_norm": 0.02117246761918068, |
|
"learning_rate": 2.846054333764554e-06, |
|
"loss": 0.0091, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.07434343434343435, |
|
"grad_norm": 0.01761500909924507, |
|
"learning_rate": 2.975420439844761e-06, |
|
"loss": 0.008, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.07757575757575758, |
|
"grad_norm": 0.01908932812511921, |
|
"learning_rate": 3.1047865459249677e-06, |
|
"loss": 0.008, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.08080808080808081, |
|
"grad_norm": 0.01619753986597061, |
|
"learning_rate": 3.234152652005175e-06, |
|
"loss": 0.0078, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.08080808080808081, |
|
"eval_loss": 0.007939654402434826, |
|
"eval_runtime": 20.513, |
|
"eval_samples_per_second": 4.875, |
|
"eval_steps_per_second": 1.219, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.08404040404040404, |
|
"grad_norm": 0.01418386958539486, |
|
"learning_rate": 3.3635187580853815e-06, |
|
"loss": 0.0067, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.08727272727272728, |
|
"grad_norm": 0.018947051838040352, |
|
"learning_rate": 3.492884864165589e-06, |
|
"loss": 0.0086, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.0905050505050505, |
|
"grad_norm": 0.016921700909733772, |
|
"learning_rate": 3.6222509702457957e-06, |
|
"loss": 0.0085, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.09373737373737374, |
|
"grad_norm": 0.01551737543195486, |
|
"learning_rate": 3.751617076326003e-06, |
|
"loss": 0.0084, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.09696969696969697, |
|
"grad_norm": 0.014143792912364006, |
|
"learning_rate": 3.880983182406209e-06, |
|
"loss": 0.0078, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.10020202020202021, |
|
"grad_norm": 0.011397017166018486, |
|
"learning_rate": 4.010349288486417e-06, |
|
"loss": 0.0053, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.10343434343434343, |
|
"grad_norm": 0.015161341056227684, |
|
"learning_rate": 4.1397153945666245e-06, |
|
"loss": 0.0073, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.10666666666666667, |
|
"grad_norm": 0.012180107645690441, |
|
"learning_rate": 4.2690815006468305e-06, |
|
"loss": 0.0063, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.1098989898989899, |
|
"grad_norm": 0.012770959176123142, |
|
"learning_rate": 4.398447606727038e-06, |
|
"loss": 0.0079, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.11313131313131314, |
|
"grad_norm": 0.014727453701198101, |
|
"learning_rate": 4.527813712807244e-06, |
|
"loss": 0.0122, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.11636363636363636, |
|
"grad_norm": 0.012778300791978836, |
|
"learning_rate": 4.657179818887452e-06, |
|
"loss": 0.0071, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.1195959595959596, |
|
"grad_norm": 0.012685113586485386, |
|
"learning_rate": 4.786545924967659e-06, |
|
"loss": 0.0065, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.12282828282828283, |
|
"grad_norm": 0.014156874269247055, |
|
"learning_rate": 4.915912031047866e-06, |
|
"loss": 0.0082, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.12606060606060607, |
|
"grad_norm": 0.01115910243242979, |
|
"learning_rate": 5.045278137128073e-06, |
|
"loss": 0.007, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.1292929292929293, |
|
"grad_norm": 0.014089164324104786, |
|
"learning_rate": 5.1746442432082795e-06, |
|
"loss": 0.0073, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.13252525252525252, |
|
"grad_norm": 0.013440214097499847, |
|
"learning_rate": 5.304010349288486e-06, |
|
"loss": 0.0082, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.13575757575757577, |
|
"grad_norm": 0.01083499938249588, |
|
"learning_rate": 5.433376455368694e-06, |
|
"loss": 0.006, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.138989898989899, |
|
"grad_norm": 0.011082631535828114, |
|
"learning_rate": 5.5627425614489e-06, |
|
"loss": 0.007, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.14222222222222222, |
|
"grad_norm": 0.012490961700677872, |
|
"learning_rate": 5.692108667529108e-06, |
|
"loss": 0.0062, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.14545454545454545, |
|
"grad_norm": 0.012171825394034386, |
|
"learning_rate": 5.821474773609315e-06, |
|
"loss": 0.0066, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.1486868686868687, |
|
"grad_norm": 0.010148804634809494, |
|
"learning_rate": 5.950840879689522e-06, |
|
"loss": 0.0042, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.15191919191919193, |
|
"grad_norm": 0.013537143357098103, |
|
"learning_rate": 6.0802069857697286e-06, |
|
"loss": 0.0073, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.15515151515151515, |
|
"grad_norm": 0.011110929772257805, |
|
"learning_rate": 6.2095730918499354e-06, |
|
"loss": 0.0072, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.15838383838383838, |
|
"grad_norm": 0.00937278475612402, |
|
"learning_rate": 6.338939197930142e-06, |
|
"loss": 0.0054, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.16161616161616163, |
|
"grad_norm": 0.011824984103441238, |
|
"learning_rate": 6.46830530401035e-06, |
|
"loss": 0.0119, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.16161616161616163, |
|
"eval_loss": 0.005149205215275288, |
|
"eval_runtime": 18.6062, |
|
"eval_samples_per_second": 5.375, |
|
"eval_steps_per_second": 1.344, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.16484848484848486, |
|
"grad_norm": 0.009701536037027836, |
|
"learning_rate": 6.597671410090557e-06, |
|
"loss": 0.0054, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.16808080808080808, |
|
"grad_norm": 0.010364706628024578, |
|
"learning_rate": 6.727037516170763e-06, |
|
"loss": 0.0049, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.1713131313131313, |
|
"grad_norm": 0.009346776641905308, |
|
"learning_rate": 6.856403622250971e-06, |
|
"loss": 0.0054, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.17454545454545456, |
|
"grad_norm": 0.008815059438347816, |
|
"learning_rate": 6.985769728331178e-06, |
|
"loss": 0.0069, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.17777777777777778, |
|
"grad_norm": 0.009431697428226471, |
|
"learning_rate": 7.115135834411385e-06, |
|
"loss": 0.0083, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.181010101010101, |
|
"grad_norm": 0.012181616388261318, |
|
"learning_rate": 7.244501940491591e-06, |
|
"loss": 0.0065, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.18424242424242424, |
|
"grad_norm": 0.009227719157934189, |
|
"learning_rate": 7.373868046571798e-06, |
|
"loss": 0.0052, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.1874747474747475, |
|
"grad_norm": 0.009855546988546848, |
|
"learning_rate": 7.503234152652006e-06, |
|
"loss": 0.0082, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.1907070707070707, |
|
"grad_norm": 0.008454745635390282, |
|
"learning_rate": 7.632600258732213e-06, |
|
"loss": 0.0044, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.19393939393939394, |
|
"grad_norm": 0.010128876194357872, |
|
"learning_rate": 7.761966364812418e-06, |
|
"loss": 0.0092, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.19717171717171716, |
|
"grad_norm": 0.009267722256481647, |
|
"learning_rate": 7.891332470892627e-06, |
|
"loss": 0.0049, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.20040404040404042, |
|
"grad_norm": 0.009394655004143715, |
|
"learning_rate": 8.020698576972833e-06, |
|
"loss": 0.0059, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.20363636363636364, |
|
"grad_norm": 0.008983040228486061, |
|
"learning_rate": 8.15006468305304e-06, |
|
"loss": 0.0041, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.20686868686868687, |
|
"grad_norm": 0.009920783340930939, |
|
"learning_rate": 8.279430789133249e-06, |
|
"loss": 0.0058, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.2101010101010101, |
|
"grad_norm": 0.007694128435105085, |
|
"learning_rate": 8.408796895213454e-06, |
|
"loss": 0.0033, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.21333333333333335, |
|
"grad_norm": 0.01226822566241026, |
|
"learning_rate": 8.538163001293661e-06, |
|
"loss": 0.0099, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.21656565656565657, |
|
"grad_norm": 0.010597337037324905, |
|
"learning_rate": 8.66752910737387e-06, |
|
"loss": 0.0081, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.2197979797979798, |
|
"grad_norm": 0.009066218510270119, |
|
"learning_rate": 8.796895213454076e-06, |
|
"loss": 0.0044, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.22303030303030302, |
|
"grad_norm": 0.01163114607334137, |
|
"learning_rate": 8.926261319534282e-06, |
|
"loss": 0.0129, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.22626262626262628, |
|
"grad_norm": 0.012832598760724068, |
|
"learning_rate": 9.055627425614489e-06, |
|
"loss": 0.0071, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.2294949494949495, |
|
"grad_norm": 0.010620299726724625, |
|
"learning_rate": 9.184993531694697e-06, |
|
"loss": 0.004, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.23272727272727273, |
|
"grad_norm": 0.00913357175886631, |
|
"learning_rate": 9.314359637774904e-06, |
|
"loss": 0.0058, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.23595959595959595, |
|
"grad_norm": 0.011772734113037586, |
|
"learning_rate": 9.44372574385511e-06, |
|
"loss": 0.0036, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.2391919191919192, |
|
"grad_norm": 0.011326112784445286, |
|
"learning_rate": 9.573091849935318e-06, |
|
"loss": 0.0071, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.24242424242424243, |
|
"grad_norm": 0.010667567141354084, |
|
"learning_rate": 9.702457956015525e-06, |
|
"loss": 0.0036, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.24242424242424243, |
|
"eval_loss": 0.0031650287564843893, |
|
"eval_runtime": 18.6391, |
|
"eval_samples_per_second": 5.365, |
|
"eval_steps_per_second": 1.341, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.24565656565656566, |
|
"grad_norm": 0.007108451332896948, |
|
"learning_rate": 9.831824062095732e-06, |
|
"loss": 0.0023, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.24888888888888888, |
|
"grad_norm": 0.008278781548142433, |
|
"learning_rate": 9.961190168175938e-06, |
|
"loss": 0.0026, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.25212121212121213, |
|
"grad_norm": 0.010979422368109226, |
|
"learning_rate": 1.0090556274256145e-05, |
|
"loss": 0.0036, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.25535353535353533, |
|
"grad_norm": 0.007666402496397495, |
|
"learning_rate": 1.0219922380336352e-05, |
|
"loss": 0.0015, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.2585858585858586, |
|
"grad_norm": 0.009133332408964634, |
|
"learning_rate": 1.0349288486416559e-05, |
|
"loss": 0.0021, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.26181818181818184, |
|
"grad_norm": 0.009566927328705788, |
|
"learning_rate": 1.0478654592496766e-05, |
|
"loss": 0.0038, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.26505050505050504, |
|
"grad_norm": 0.011765814386308193, |
|
"learning_rate": 1.0608020698576973e-05, |
|
"loss": 0.0061, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.2682828282828283, |
|
"grad_norm": 0.009488740935921669, |
|
"learning_rate": 1.073738680465718e-05, |
|
"loss": 0.0023, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.27151515151515154, |
|
"grad_norm": 0.007137450389564037, |
|
"learning_rate": 1.0866752910737388e-05, |
|
"loss": 0.0021, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.27474747474747474, |
|
"grad_norm": 0.008984182961285114, |
|
"learning_rate": 1.0996119016817593e-05, |
|
"loss": 0.0028, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.277979797979798, |
|
"grad_norm": 0.010403821244835854, |
|
"learning_rate": 1.11254851228978e-05, |
|
"loss": 0.0057, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.2812121212121212, |
|
"grad_norm": 0.0071556540206074715, |
|
"learning_rate": 1.1254851228978009e-05, |
|
"loss": 0.0017, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.28444444444444444, |
|
"grad_norm": 0.012338937260210514, |
|
"learning_rate": 1.1384217335058216e-05, |
|
"loss": 0.0029, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.2876767676767677, |
|
"grad_norm": 0.010343515314161777, |
|
"learning_rate": 1.1513583441138421e-05, |
|
"loss": 0.0032, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.2909090909090909, |
|
"grad_norm": 0.009991390630602837, |
|
"learning_rate": 1.164294954721863e-05, |
|
"loss": 0.0035, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.29414141414141415, |
|
"grad_norm": 0.009020160883665085, |
|
"learning_rate": 1.1772315653298836e-05, |
|
"loss": 0.0033, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.2973737373737374, |
|
"grad_norm": 0.0067694829776883125, |
|
"learning_rate": 1.1901681759379043e-05, |
|
"loss": 0.0019, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.3006060606060606, |
|
"grad_norm": 0.013036763295531273, |
|
"learning_rate": 1.203104786545925e-05, |
|
"loss": 0.0041, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.30383838383838385, |
|
"grad_norm": 0.009171461686491966, |
|
"learning_rate": 1.2160413971539457e-05, |
|
"loss": 0.006, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.30707070707070705, |
|
"grad_norm": 0.007229423616081476, |
|
"learning_rate": 1.2289780077619664e-05, |
|
"loss": 0.002, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.3103030303030303, |
|
"grad_norm": 0.009065698832273483, |
|
"learning_rate": 1.2419146183699871e-05, |
|
"loss": 0.0028, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.31353535353535356, |
|
"grad_norm": 0.0068460931070148945, |
|
"learning_rate": 1.254851228978008e-05, |
|
"loss": 0.0024, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.31676767676767675, |
|
"grad_norm": 0.12772586941719055, |
|
"learning_rate": 1.2677878395860285e-05, |
|
"loss": 0.0046, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.008569791913032532, |
|
"learning_rate": 1.2807244501940493e-05, |
|
"loss": 0.0024, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.32323232323232326, |
|
"grad_norm": 0.010226168669760227, |
|
"learning_rate": 1.29366106080207e-05, |
|
"loss": 0.004, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.32323232323232326, |
|
"eval_loss": 0.002474932000041008, |
|
"eval_runtime": 18.5931, |
|
"eval_samples_per_second": 5.378, |
|
"eval_steps_per_second": 1.345, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.32646464646464646, |
|
"grad_norm": 0.010962710715830326, |
|
"learning_rate": 1.3065976714100905e-05, |
|
"loss": 0.0018, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.3296969696969697, |
|
"grad_norm": 0.01257998775690794, |
|
"learning_rate": 1.3195342820181114e-05, |
|
"loss": 0.0033, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.3329292929292929, |
|
"grad_norm": 0.013269903138279915, |
|
"learning_rate": 1.332470892626132e-05, |
|
"loss": 0.0036, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.33616161616161616, |
|
"grad_norm": 0.009509469382464886, |
|
"learning_rate": 1.3454075032341526e-05, |
|
"loss": 0.0034, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.3393939393939394, |
|
"grad_norm": 0.00902635883539915, |
|
"learning_rate": 1.3583441138421735e-05, |
|
"loss": 0.0028, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.3426262626262626, |
|
"grad_norm": 0.007760872133076191, |
|
"learning_rate": 1.3712807244501941e-05, |
|
"loss": 0.0024, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.34585858585858587, |
|
"grad_norm": 0.009263481013476849, |
|
"learning_rate": 1.384217335058215e-05, |
|
"loss": 0.0031, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.3490909090909091, |
|
"grad_norm": 0.00697364890947938, |
|
"learning_rate": 1.3971539456662355e-05, |
|
"loss": 0.0021, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.3523232323232323, |
|
"grad_norm": 0.007600918412208557, |
|
"learning_rate": 1.4100905562742562e-05, |
|
"loss": 0.0018, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.35555555555555557, |
|
"grad_norm": 0.00856352224946022, |
|
"learning_rate": 1.423027166882277e-05, |
|
"loss": 0.0031, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.35878787878787877, |
|
"grad_norm": 0.007947358302772045, |
|
"learning_rate": 1.4359637774902976e-05, |
|
"loss": 0.0025, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.362020202020202, |
|
"grad_norm": 0.007706194184720516, |
|
"learning_rate": 1.4489003880983183e-05, |
|
"loss": 0.0023, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.3652525252525253, |
|
"grad_norm": 0.008598407730460167, |
|
"learning_rate": 1.4618369987063391e-05, |
|
"loss": 0.0028, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.36848484848484847, |
|
"grad_norm": 0.00845644623041153, |
|
"learning_rate": 1.4747736093143596e-05, |
|
"loss": 0.0029, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.3717171717171717, |
|
"grad_norm": 0.008533057756721973, |
|
"learning_rate": 1.4877102199223805e-05, |
|
"loss": 0.0039, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.374949494949495, |
|
"grad_norm": 0.007429471705108881, |
|
"learning_rate": 1.5006468305304012e-05, |
|
"loss": 0.0024, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.3781818181818182, |
|
"grad_norm": 0.0069627161137759686, |
|
"learning_rate": 1.5135834411384217e-05, |
|
"loss": 0.0018, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.3814141414141414, |
|
"grad_norm": 0.007471222430467606, |
|
"learning_rate": 1.5265200517464426e-05, |
|
"loss": 0.0024, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.3846464646464646, |
|
"grad_norm": 0.006873416714370251, |
|
"learning_rate": 1.5394566623544633e-05, |
|
"loss": 0.0019, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.3878787878787879, |
|
"grad_norm": 0.009251467883586884, |
|
"learning_rate": 1.5523932729624836e-05, |
|
"loss": 0.0027, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.39111111111111113, |
|
"grad_norm": 0.007302634883671999, |
|
"learning_rate": 1.5653298835705046e-05, |
|
"loss": 0.0017, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.39434343434343433, |
|
"grad_norm": 0.006746005266904831, |
|
"learning_rate": 1.5782664941785253e-05, |
|
"loss": 0.0019, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.3975757575757576, |
|
"grad_norm": 0.008011849597096443, |
|
"learning_rate": 1.591203104786546e-05, |
|
"loss": 0.0023, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.40080808080808084, |
|
"grad_norm": 0.008799983188509941, |
|
"learning_rate": 1.6041397153945667e-05, |
|
"loss": 0.0023, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.40404040404040403, |
|
"grad_norm": 0.010617449879646301, |
|
"learning_rate": 1.6170763260025874e-05, |
|
"loss": 0.0019, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.40404040404040403, |
|
"eval_loss": 0.002037045545876026, |
|
"eval_runtime": 18.5899, |
|
"eval_samples_per_second": 5.379, |
|
"eval_steps_per_second": 1.345, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.4072727272727273, |
|
"grad_norm": 0.006759752053767443, |
|
"learning_rate": 1.630012936610608e-05, |
|
"loss": 0.0018, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.4105050505050505, |
|
"grad_norm": 0.008995631709694862, |
|
"learning_rate": 1.6429495472186288e-05, |
|
"loss": 0.0022, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.41373737373737374, |
|
"grad_norm": 0.00835257675498724, |
|
"learning_rate": 1.6558861578266498e-05, |
|
"loss": 0.0026, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.416969696969697, |
|
"grad_norm": 0.009772442281246185, |
|
"learning_rate": 1.66882276843467e-05, |
|
"loss": 0.0028, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.4202020202020202, |
|
"grad_norm": 0.006931178271770477, |
|
"learning_rate": 1.6817593790426908e-05, |
|
"loss": 0.002, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.42343434343434344, |
|
"grad_norm": 0.012796151451766491, |
|
"learning_rate": 1.694695989650712e-05, |
|
"loss": 0.0021, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.4266666666666667, |
|
"grad_norm": 0.007226724177598953, |
|
"learning_rate": 1.7076326002587322e-05, |
|
"loss": 0.0033, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.4298989898989899, |
|
"grad_norm": 0.0066243987530469894, |
|
"learning_rate": 1.720569210866753e-05, |
|
"loss": 0.002, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.43313131313131314, |
|
"grad_norm": 0.007128287572413683, |
|
"learning_rate": 1.733505821474774e-05, |
|
"loss": 0.0021, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.43636363636363634, |
|
"grad_norm": 0.007028148043900728, |
|
"learning_rate": 1.7464424320827943e-05, |
|
"loss": 0.0028, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.4395959595959596, |
|
"grad_norm": 0.00866060983389616, |
|
"learning_rate": 1.7593790426908153e-05, |
|
"loss": 0.0053, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.44282828282828285, |
|
"grad_norm": 0.005735491868108511, |
|
"learning_rate": 1.7723156532988356e-05, |
|
"loss": 0.0023, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.44606060606060605, |
|
"grad_norm": 0.005858046934008598, |
|
"learning_rate": 1.7852522639068563e-05, |
|
"loss": 0.0019, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.4492929292929293, |
|
"grad_norm": 0.007398343179374933, |
|
"learning_rate": 1.7981888745148774e-05, |
|
"loss": 0.0024, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.45252525252525255, |
|
"grad_norm": 0.009320907294750214, |
|
"learning_rate": 1.8111254851228977e-05, |
|
"loss": 0.0063, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.45575757575757575, |
|
"grad_norm": 0.008654161356389523, |
|
"learning_rate": 1.8240620957309184e-05, |
|
"loss": 0.0025, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.458989898989899, |
|
"grad_norm": 0.004506159573793411, |
|
"learning_rate": 1.8369987063389394e-05, |
|
"loss": 0.0013, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.4622222222222222, |
|
"grad_norm": 0.009703525342047215, |
|
"learning_rate": 1.8499353169469598e-05, |
|
"loss": 0.0034, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.46545454545454545, |
|
"grad_norm": 0.007734385784715414, |
|
"learning_rate": 1.8628719275549808e-05, |
|
"loss": 0.0029, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.4686868686868687, |
|
"grad_norm": 0.010392666794359684, |
|
"learning_rate": 1.8758085381630015e-05, |
|
"loss": 0.0028, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.4719191919191919, |
|
"grad_norm": 0.01011224091053009, |
|
"learning_rate": 1.888745148771022e-05, |
|
"loss": 0.0036, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.47515151515151516, |
|
"grad_norm": 0.005182855296880007, |
|
"learning_rate": 1.901681759379043e-05, |
|
"loss": 0.0014, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.4783838383838384, |
|
"grad_norm": 0.009993299841880798, |
|
"learning_rate": 1.9146183699870636e-05, |
|
"loss": 0.003, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.4816161616161616, |
|
"grad_norm": 0.008879208005964756, |
|
"learning_rate": 1.927554980595084e-05, |
|
"loss": 0.002, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.48484848484848486, |
|
"grad_norm": 0.006573604419827461, |
|
"learning_rate": 1.940491591203105e-05, |
|
"loss": 0.0021, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.48484848484848486, |
|
"eval_loss": 0.0017716821748763323, |
|
"eval_runtime": 18.6403, |
|
"eval_samples_per_second": 5.365, |
|
"eval_steps_per_second": 1.341, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.48808080808080806, |
|
"grad_norm": 0.005542725790292025, |
|
"learning_rate": 1.9534282018111256e-05, |
|
"loss": 0.0017, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.4913131313131313, |
|
"grad_norm": 0.008134805597364902, |
|
"learning_rate": 1.9663648124191463e-05, |
|
"loss": 0.002, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.49454545454545457, |
|
"grad_norm": 0.009408293291926384, |
|
"learning_rate": 1.979301423027167e-05, |
|
"loss": 0.0031, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.49777777777777776, |
|
"grad_norm": 0.007265687920153141, |
|
"learning_rate": 1.9922380336351877e-05, |
|
"loss": 0.0021, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.501010101010101, |
|
"grad_norm": 0.009999910369515419, |
|
"learning_rate": 2.0051746442432084e-05, |
|
"loss": 0.0014, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.5042424242424243, |
|
"grad_norm": 0.007147953379899263, |
|
"learning_rate": 2.018111254851229e-05, |
|
"loss": 0.0022, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.5074747474747475, |
|
"grad_norm": 0.0072359428741037846, |
|
"learning_rate": 2.0310478654592497e-05, |
|
"loss": 0.0034, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.5107070707070707, |
|
"grad_norm": 0.005821447819471359, |
|
"learning_rate": 2.0439844760672704e-05, |
|
"loss": 0.0018, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.5139393939393939, |
|
"grad_norm": 0.007256666664034128, |
|
"learning_rate": 2.056921086675291e-05, |
|
"loss": 0.0027, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.5171717171717172, |
|
"grad_norm": 0.00663521233946085, |
|
"learning_rate": 2.0698576972833118e-05, |
|
"loss": 0.0023, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.5204040404040404, |
|
"grad_norm": 0.007504627574235201, |
|
"learning_rate": 2.0827943078913325e-05, |
|
"loss": 0.004, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.5236363636363637, |
|
"grad_norm": 0.005167617462575436, |
|
"learning_rate": 2.0957309184993532e-05, |
|
"loss": 0.0015, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.5268686868686868, |
|
"grad_norm": 0.00587807409465313, |
|
"learning_rate": 2.108667529107374e-05, |
|
"loss": 0.0016, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.5301010101010101, |
|
"grad_norm": 0.005344127304852009, |
|
"learning_rate": 2.1216041397153946e-05, |
|
"loss": 0.0011, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.5333333333333333, |
|
"grad_norm": 0.005351161118596792, |
|
"learning_rate": 2.1345407503234156e-05, |
|
"loss": 0.0016, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.5365656565656566, |
|
"grad_norm": 0.006603873800486326, |
|
"learning_rate": 2.147477360931436e-05, |
|
"loss": 0.0017, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.5397979797979798, |
|
"grad_norm": 0.006440309341996908, |
|
"learning_rate": 2.1604139715394566e-05, |
|
"loss": 0.0016, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.5430303030303031, |
|
"grad_norm": 0.007731405086815357, |
|
"learning_rate": 2.1733505821474777e-05, |
|
"loss": 0.0024, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.5462626262626262, |
|
"grad_norm": 0.0070298160426318645, |
|
"learning_rate": 2.186287192755498e-05, |
|
"loss": 0.0019, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.5494949494949495, |
|
"grad_norm": 0.008382032625377178, |
|
"learning_rate": 2.1992238033635187e-05, |
|
"loss": 0.0017, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.5527272727272727, |
|
"grad_norm": 0.007320054341107607, |
|
"learning_rate": 2.2121604139715397e-05, |
|
"loss": 0.0017, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.555959595959596, |
|
"grad_norm": 0.008367806673049927, |
|
"learning_rate": 2.22509702457956e-05, |
|
"loss": 0.0036, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.5591919191919192, |
|
"grad_norm": 0.008995486423373222, |
|
"learning_rate": 2.238033635187581e-05, |
|
"loss": 0.0019, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.5624242424242424, |
|
"grad_norm": 0.007300530560314655, |
|
"learning_rate": 2.2509702457956018e-05, |
|
"loss": 0.0036, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.5656565656565656, |
|
"grad_norm": 0.005388753954321146, |
|
"learning_rate": 2.263906856403622e-05, |
|
"loss": 0.0016, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.5656565656565656, |
|
"eval_loss": 0.0016459682956337929, |
|
"eval_runtime": 18.6118, |
|
"eval_samples_per_second": 5.373, |
|
"eval_steps_per_second": 1.343, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.5688888888888889, |
|
"grad_norm": 0.005545719526708126, |
|
"learning_rate": 2.276843467011643e-05, |
|
"loss": 0.0016, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.5721212121212121, |
|
"grad_norm": 0.007394594140350819, |
|
"learning_rate": 2.289780077619664e-05, |
|
"loss": 0.0021, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.5753535353535354, |
|
"grad_norm": 0.005089296959340572, |
|
"learning_rate": 2.3027166882276842e-05, |
|
"loss": 0.0013, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.5785858585858585, |
|
"grad_norm": 0.00883107353001833, |
|
"learning_rate": 2.3156532988357052e-05, |
|
"loss": 0.0022, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.5818181818181818, |
|
"grad_norm": 0.012213426642119884, |
|
"learning_rate": 2.328589909443726e-05, |
|
"loss": 0.005, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.585050505050505, |
|
"grad_norm": 0.007255645003169775, |
|
"learning_rate": 2.3415265200517466e-05, |
|
"loss": 0.001, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.5882828282828283, |
|
"grad_norm": 0.006818380672484636, |
|
"learning_rate": 2.3544631306597673e-05, |
|
"loss": 0.0015, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.5915151515151515, |
|
"grad_norm": 0.006983195431530476, |
|
"learning_rate": 2.367399741267788e-05, |
|
"loss": 0.0027, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.5947474747474748, |
|
"grad_norm": 0.012661872431635857, |
|
"learning_rate": 2.3803363518758087e-05, |
|
"loss": 0.0053, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.597979797979798, |
|
"grad_norm": 0.005046526901423931, |
|
"learning_rate": 2.3932729624838294e-05, |
|
"loss": 0.0019, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.6012121212121212, |
|
"grad_norm": 0.004457033704966307, |
|
"learning_rate": 2.40620957309185e-05, |
|
"loss": 0.0012, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.6044444444444445, |
|
"grad_norm": 0.00732004176825285, |
|
"learning_rate": 2.4191461836998707e-05, |
|
"loss": 0.0017, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.6076767676767677, |
|
"grad_norm": 0.006394708063453436, |
|
"learning_rate": 2.4320827943078914e-05, |
|
"loss": 0.0019, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.610909090909091, |
|
"grad_norm": 0.007961345836520195, |
|
"learning_rate": 2.445019404915912e-05, |
|
"loss": 0.0034, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.6141414141414141, |
|
"grad_norm": 0.0068709347397089005, |
|
"learning_rate": 2.4579560155239328e-05, |
|
"loss": 0.0013, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.6173737373737374, |
|
"grad_norm": 0.009581703692674637, |
|
"learning_rate": 2.4708926261319535e-05, |
|
"loss": 0.0026, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.6206060606060606, |
|
"grad_norm": 0.007400548551231623, |
|
"learning_rate": 2.4838292367399742e-05, |
|
"loss": 0.0024, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.6238383838383839, |
|
"grad_norm": 0.00970857497304678, |
|
"learning_rate": 2.496765847347995e-05, |
|
"loss": 0.0013, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.6270707070707071, |
|
"grad_norm": 0.007368095684796572, |
|
"learning_rate": 2.509702457956016e-05, |
|
"loss": 0.0016, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.6303030303030303, |
|
"grad_norm": 0.007306639105081558, |
|
"learning_rate": 2.5226390685640362e-05, |
|
"loss": 0.0015, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.6335353535353535, |
|
"grad_norm": 0.008257709443569183, |
|
"learning_rate": 2.535575679172057e-05, |
|
"loss": 0.0017, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.6367676767676768, |
|
"grad_norm": 0.0050062090158462524, |
|
"learning_rate": 2.548512289780078e-05, |
|
"loss": 0.0012, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"grad_norm": 0.007328143808990717, |
|
"learning_rate": 2.5614489003880986e-05, |
|
"loss": 0.0025, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.6432323232323233, |
|
"grad_norm": 0.007282217964529991, |
|
"learning_rate": 2.574385510996119e-05, |
|
"loss": 0.0018, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.6464646464646465, |
|
"grad_norm": 0.007711055688560009, |
|
"learning_rate": 2.58732212160414e-05, |
|
"loss": 0.0013, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.6464646464646465, |
|
"eval_loss": 0.0015150802209973335, |
|
"eval_runtime": 18.6339, |
|
"eval_samples_per_second": 5.367, |
|
"eval_steps_per_second": 1.342, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.6496969696969697, |
|
"grad_norm": 0.006153900176286697, |
|
"learning_rate": 2.6002587322121607e-05, |
|
"loss": 0.0016, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 0.6529292929292929, |
|
"grad_norm": 0.0070861754938960075, |
|
"learning_rate": 2.613195342820181e-05, |
|
"loss": 0.002, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 0.6561616161616162, |
|
"grad_norm": 0.0048012156039476395, |
|
"learning_rate": 2.626131953428202e-05, |
|
"loss": 0.0012, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 0.6593939393939394, |
|
"grad_norm": 0.006167956627905369, |
|
"learning_rate": 2.6390685640362228e-05, |
|
"loss": 0.0018, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.6626262626262627, |
|
"grad_norm": 0.0051431735046207905, |
|
"learning_rate": 2.652005174644243e-05, |
|
"loss": 0.0013, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.6658585858585858, |
|
"grad_norm": 0.009899413213133812, |
|
"learning_rate": 2.664941785252264e-05, |
|
"loss": 0.0031, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 0.6690909090909091, |
|
"grad_norm": 0.007686274591833353, |
|
"learning_rate": 2.677878395860285e-05, |
|
"loss": 0.0022, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 0.6723232323232323, |
|
"grad_norm": 0.004816455766558647, |
|
"learning_rate": 2.6908150064683052e-05, |
|
"loss": 0.0011, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 0.6755555555555556, |
|
"grad_norm": 0.004873145837336779, |
|
"learning_rate": 2.7037516170763262e-05, |
|
"loss": 0.0011, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 0.6787878787878788, |
|
"grad_norm": 0.006136384792625904, |
|
"learning_rate": 2.716688227684347e-05, |
|
"loss": 0.0016, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.682020202020202, |
|
"grad_norm": 0.004508585669100285, |
|
"learning_rate": 2.7296248382923673e-05, |
|
"loss": 0.0012, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 0.6852525252525252, |
|
"grad_norm": 0.005723617039620876, |
|
"learning_rate": 2.7425614489003883e-05, |
|
"loss": 0.0016, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 0.6884848484848485, |
|
"grad_norm": 0.0072562843561172485, |
|
"learning_rate": 2.755498059508409e-05, |
|
"loss": 0.0015, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 0.6917171717171717, |
|
"grad_norm": 0.006079181097447872, |
|
"learning_rate": 2.76843467011643e-05, |
|
"loss": 0.0017, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 0.694949494949495, |
|
"grad_norm": 0.011977693997323513, |
|
"learning_rate": 2.7813712807244503e-05, |
|
"loss": 0.0048, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.6981818181818182, |
|
"grad_norm": 0.007406435441225767, |
|
"learning_rate": 2.794307891332471e-05, |
|
"loss": 0.0022, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.7014141414141414, |
|
"grad_norm": 0.006909268908202648, |
|
"learning_rate": 2.807244501940492e-05, |
|
"loss": 0.0022, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 0.7046464646464646, |
|
"grad_norm": 0.0051758866757154465, |
|
"learning_rate": 2.8201811125485124e-05, |
|
"loss": 0.0012, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 0.7078787878787879, |
|
"grad_norm": 0.00657995231449604, |
|
"learning_rate": 2.833117723156533e-05, |
|
"loss": 0.0014, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 0.7111111111111111, |
|
"grad_norm": 0.006561241112649441, |
|
"learning_rate": 2.846054333764554e-05, |
|
"loss": 0.0018, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.7143434343434344, |
|
"grad_norm": 0.005906842183321714, |
|
"learning_rate": 2.8589909443725745e-05, |
|
"loss": 0.0013, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 0.7175757575757575, |
|
"grad_norm": 0.004000976216048002, |
|
"learning_rate": 2.871927554980595e-05, |
|
"loss": 0.0009, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 0.7208080808080808, |
|
"grad_norm": 0.006313610821962357, |
|
"learning_rate": 2.8848641655886162e-05, |
|
"loss": 0.0015, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 0.724040404040404, |
|
"grad_norm": 0.005339731462299824, |
|
"learning_rate": 2.8978007761966365e-05, |
|
"loss": 0.0014, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 0.7272727272727273, |
|
"grad_norm": 0.00704893097281456, |
|
"learning_rate": 2.9107373868046572e-05, |
|
"loss": 0.0017, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.7272727272727273, |
|
"eval_loss": 0.0014526441227644682, |
|
"eval_runtime": 18.5916, |
|
"eval_samples_per_second": 5.379, |
|
"eval_steps_per_second": 1.345, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.7305050505050505, |
|
"grad_norm": 0.007863117381930351, |
|
"learning_rate": 2.9236739974126783e-05, |
|
"loss": 0.002, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 0.7337373737373737, |
|
"grad_norm": 0.007929647341370583, |
|
"learning_rate": 2.936610608020699e-05, |
|
"loss": 0.0014, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 0.7369696969696969, |
|
"grad_norm": 0.008834928274154663, |
|
"learning_rate": 2.9495472186287193e-05, |
|
"loss": 0.0033, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 0.7402020202020202, |
|
"grad_norm": 0.005479221232235432, |
|
"learning_rate": 2.9624838292367403e-05, |
|
"loss": 0.0013, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 0.7434343434343434, |
|
"grad_norm": 0.008690214715898037, |
|
"learning_rate": 2.975420439844761e-05, |
|
"loss": 0.0031, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.7466666666666667, |
|
"grad_norm": 0.006373909767717123, |
|
"learning_rate": 2.9883570504527814e-05, |
|
"loss": 0.0013, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 0.74989898989899, |
|
"grad_norm": 0.005924658849835396, |
|
"learning_rate": 3.0012936610608024e-05, |
|
"loss": 0.0014, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 0.7531313131313131, |
|
"grad_norm": 0.007938825525343418, |
|
"learning_rate": 3.014230271668823e-05, |
|
"loss": 0.0027, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 0.7563636363636363, |
|
"grad_norm": 0.007831827737390995, |
|
"learning_rate": 3.0271668822768434e-05, |
|
"loss": 0.0025, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 0.7595959595959596, |
|
"grad_norm": 0.008570835925638676, |
|
"learning_rate": 3.0401034928848644e-05, |
|
"loss": 0.003, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.7628282828282829, |
|
"grad_norm": 0.005871222820132971, |
|
"learning_rate": 3.053040103492885e-05, |
|
"loss": 0.0013, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 0.7660606060606061, |
|
"grad_norm": 0.007041990291327238, |
|
"learning_rate": 3.0659767141009055e-05, |
|
"loss": 0.0028, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 0.7692929292929293, |
|
"grad_norm": 0.0073891859501600266, |
|
"learning_rate": 3.0789133247089265e-05, |
|
"loss": 0.0023, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 0.7725252525252525, |
|
"grad_norm": 0.005527487024664879, |
|
"learning_rate": 3.0918499353169475e-05, |
|
"loss": 0.0014, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 0.7757575757575758, |
|
"grad_norm": 0.007551091257482767, |
|
"learning_rate": 3.104786545924967e-05, |
|
"loss": 0.002, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.778989898989899, |
|
"grad_norm": 0.005599314346909523, |
|
"learning_rate": 3.117723156532988e-05, |
|
"loss": 0.002, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 0.7822222222222223, |
|
"grad_norm": 0.007462210953235626, |
|
"learning_rate": 3.130659767141009e-05, |
|
"loss": 0.0019, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 0.7854545454545454, |
|
"grad_norm": 0.004866201896220446, |
|
"learning_rate": 3.14359637774903e-05, |
|
"loss": 0.0012, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 0.7886868686868687, |
|
"grad_norm": 0.005268635228276253, |
|
"learning_rate": 3.1565329883570506e-05, |
|
"loss": 0.0013, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 0.7919191919191919, |
|
"grad_norm": 0.007261104416102171, |
|
"learning_rate": 3.169469598965072e-05, |
|
"loss": 0.0031, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.7951515151515152, |
|
"grad_norm": 0.006308354903012514, |
|
"learning_rate": 3.182406209573092e-05, |
|
"loss": 0.0014, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 0.7983838383838384, |
|
"grad_norm": 0.0074616689234972, |
|
"learning_rate": 3.1953428201811124e-05, |
|
"loss": 0.0015, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 0.8016161616161617, |
|
"grad_norm": 0.007333788555115461, |
|
"learning_rate": 3.2082794307891334e-05, |
|
"loss": 0.0014, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 0.8048484848484848, |
|
"grad_norm": 0.006965926848351955, |
|
"learning_rate": 3.2212160413971544e-05, |
|
"loss": 0.0018, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 0.8080808080808081, |
|
"grad_norm": 0.00535194855183363, |
|
"learning_rate": 3.234152652005175e-05, |
|
"loss": 0.0015, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.8080808080808081, |
|
"eval_loss": 0.0013973835157230496, |
|
"eval_runtime": 18.6113, |
|
"eval_samples_per_second": 5.373, |
|
"eval_steps_per_second": 1.343, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.8113131313131313, |
|
"grad_norm": 0.006415734998881817, |
|
"learning_rate": 3.247089262613196e-05, |
|
"loss": 0.0014, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 0.8145454545454546, |
|
"grad_norm": 0.00617300346493721, |
|
"learning_rate": 3.260025873221216e-05, |
|
"loss": 0.0017, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 0.8177777777777778, |
|
"grad_norm": 0.007527848239988089, |
|
"learning_rate": 3.2729624838292365e-05, |
|
"loss": 0.0017, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 0.821010101010101, |
|
"grad_norm": 0.00747877499088645, |
|
"learning_rate": 3.2858990944372575e-05, |
|
"loss": 0.0014, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 0.8242424242424242, |
|
"grad_norm": 0.005451089236885309, |
|
"learning_rate": 3.2988357050452786e-05, |
|
"loss": 0.0012, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.8274747474747475, |
|
"grad_norm": 0.006235205102711916, |
|
"learning_rate": 3.3117723156532996e-05, |
|
"loss": 0.0014, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 0.8307070707070707, |
|
"grad_norm": 0.005764327012002468, |
|
"learning_rate": 3.324708926261319e-05, |
|
"loss": 0.0013, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 0.833939393939394, |
|
"grad_norm": 0.007408145349472761, |
|
"learning_rate": 3.33764553686934e-05, |
|
"loss": 0.0013, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 0.8371717171717171, |
|
"grad_norm": 0.004855205304920673, |
|
"learning_rate": 3.350582147477361e-05, |
|
"loss": 0.0009, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 0.8404040404040404, |
|
"grad_norm": 0.010664681904017925, |
|
"learning_rate": 3.3635187580853817e-05, |
|
"loss": 0.0016, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.8436363636363636, |
|
"grad_norm": 0.006055675912648439, |
|
"learning_rate": 3.376455368693403e-05, |
|
"loss": 0.0016, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 0.8468686868686869, |
|
"grad_norm": 0.01056759525090456, |
|
"learning_rate": 3.389391979301424e-05, |
|
"loss": 0.0016, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 0.8501010101010101, |
|
"grad_norm": 0.007450331002473831, |
|
"learning_rate": 3.4023285899094434e-05, |
|
"loss": 0.0016, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 0.8533333333333334, |
|
"grad_norm": 0.0053154826164245605, |
|
"learning_rate": 3.4152652005174644e-05, |
|
"loss": 0.0011, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 0.8565656565656565, |
|
"grad_norm": 0.0060371761210262775, |
|
"learning_rate": 3.4282018111254854e-05, |
|
"loss": 0.0013, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.8597979797979798, |
|
"grad_norm": 0.005466979928314686, |
|
"learning_rate": 3.441138421733506e-05, |
|
"loss": 0.0012, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 0.863030303030303, |
|
"grad_norm": 0.00831522885710001, |
|
"learning_rate": 3.454075032341527e-05, |
|
"loss": 0.0014, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 0.8662626262626263, |
|
"grad_norm": 0.007963420823216438, |
|
"learning_rate": 3.467011642949548e-05, |
|
"loss": 0.0025, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 0.8694949494949495, |
|
"grad_norm": 0.004937573801726103, |
|
"learning_rate": 3.4799482535575675e-05, |
|
"loss": 0.0011, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 0.8727272727272727, |
|
"grad_norm": 0.0058564417995512486, |
|
"learning_rate": 3.4928848641655885e-05, |
|
"loss": 0.0016, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.8759595959595959, |
|
"grad_norm": 0.007912451401352882, |
|
"learning_rate": 3.5058214747736096e-05, |
|
"loss": 0.0026, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 0.8791919191919192, |
|
"grad_norm": 0.003929544240236282, |
|
"learning_rate": 3.5187580853816306e-05, |
|
"loss": 0.0011, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 0.8824242424242424, |
|
"grad_norm": 0.008150831796228886, |
|
"learning_rate": 3.531694695989651e-05, |
|
"loss": 0.0018, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 0.8856565656565657, |
|
"grad_norm": 0.009232031181454659, |
|
"learning_rate": 3.544631306597671e-05, |
|
"loss": 0.0042, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 0.8888888888888888, |
|
"grad_norm": 0.007469587959349155, |
|
"learning_rate": 3.557567917205692e-05, |
|
"loss": 0.0023, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.8888888888888888, |
|
"eval_loss": 0.0013014678843319416, |
|
"eval_runtime": 18.7103, |
|
"eval_samples_per_second": 5.345, |
|
"eval_steps_per_second": 1.336, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.8921212121212121, |
|
"grad_norm": 0.006450500804930925, |
|
"learning_rate": 3.570504527813713e-05, |
|
"loss": 0.0014, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 0.8953535353535353, |
|
"grad_norm": 0.004984740167856216, |
|
"learning_rate": 3.583441138421734e-05, |
|
"loss": 0.0014, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 0.8985858585858586, |
|
"grad_norm": 0.006635539699345827, |
|
"learning_rate": 3.596377749029755e-05, |
|
"loss": 0.0016, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 0.9018181818181819, |
|
"grad_norm": 0.010307137854397297, |
|
"learning_rate": 3.609314359637775e-05, |
|
"loss": 0.0025, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 0.9050505050505051, |
|
"grad_norm": 0.008318118751049042, |
|
"learning_rate": 3.6222509702457954e-05, |
|
"loss": 0.0028, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.9082828282828282, |
|
"grad_norm": 0.007523260544985533, |
|
"learning_rate": 3.6351875808538164e-05, |
|
"loss": 0.0022, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 0.9115151515151515, |
|
"grad_norm": 0.00529656745493412, |
|
"learning_rate": 3.648124191461837e-05, |
|
"loss": 0.001, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 0.9147474747474748, |
|
"grad_norm": 0.0050112344324588776, |
|
"learning_rate": 3.661060802069858e-05, |
|
"loss": 0.0012, |
|
"step": 283 |
|
}, |
|
{ |
|
"epoch": 0.917979797979798, |
|
"grad_norm": 0.006318412255495787, |
|
"learning_rate": 3.673997412677879e-05, |
|
"loss": 0.0017, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 0.9212121212121213, |
|
"grad_norm": 0.006407878361642361, |
|
"learning_rate": 3.6869340232859e-05, |
|
"loss": 0.0022, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.9244444444444444, |
|
"grad_norm": 0.004356731195002794, |
|
"learning_rate": 3.6998706338939195e-05, |
|
"loss": 0.0007, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 0.9276767676767677, |
|
"grad_norm": 0.004966625012457371, |
|
"learning_rate": 3.7128072445019406e-05, |
|
"loss": 0.0012, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 0.9309090909090909, |
|
"grad_norm": 0.0055012330412864685, |
|
"learning_rate": 3.7257438551099616e-05, |
|
"loss": 0.0011, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 0.9341414141414142, |
|
"grad_norm": 0.006766812410205603, |
|
"learning_rate": 3.738680465717982e-05, |
|
"loss": 0.0017, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 0.9373737373737374, |
|
"grad_norm": 0.006011553108692169, |
|
"learning_rate": 3.751617076326003e-05, |
|
"loss": 0.0022, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.9406060606060606, |
|
"grad_norm": 0.0056586177088320255, |
|
"learning_rate": 3.764553686934023e-05, |
|
"loss": 0.0014, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 0.9438383838383838, |
|
"grad_norm": 0.0064941453747451305, |
|
"learning_rate": 3.777490297542044e-05, |
|
"loss": 0.0011, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 0.9470707070707071, |
|
"grad_norm": 0.006994554307311773, |
|
"learning_rate": 3.790426908150065e-05, |
|
"loss": 0.0021, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 0.9503030303030303, |
|
"grad_norm": 0.006325817201286554, |
|
"learning_rate": 3.803363518758086e-05, |
|
"loss": 0.0014, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 0.9535353535353536, |
|
"grad_norm": 0.006501290015876293, |
|
"learning_rate": 3.816300129366106e-05, |
|
"loss": 0.0017, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.9567676767676768, |
|
"grad_norm": 0.006014223676174879, |
|
"learning_rate": 3.829236739974127e-05, |
|
"loss": 0.0013, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 0.005024549085646868, |
|
"learning_rate": 3.8421733505821475e-05, |
|
"loss": 0.0013, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 0.9632323232323232, |
|
"grad_norm": 0.00840625911951065, |
|
"learning_rate": 3.855109961190168e-05, |
|
"loss": 0.0017, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 0.9664646464646465, |
|
"grad_norm": 0.005791939329355955, |
|
"learning_rate": 3.868046571798189e-05, |
|
"loss": 0.0014, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 0.9696969696969697, |
|
"grad_norm": 0.0047224657610058784, |
|
"learning_rate": 3.88098318240621e-05, |
|
"loss": 0.0012, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.9696969696969697, |
|
"eval_loss": 0.0012624365044757724, |
|
"eval_runtime": 18.7089, |
|
"eval_samples_per_second": 5.345, |
|
"eval_steps_per_second": 1.336, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.972929292929293, |
|
"grad_norm": 0.005727376788854599, |
|
"learning_rate": 3.893919793014231e-05, |
|
"loss": 0.0013, |
|
"step": 301 |
|
}, |
|
{ |
|
"epoch": 0.9761616161616161, |
|
"grad_norm": 0.007024977821856737, |
|
"learning_rate": 3.906856403622251e-05, |
|
"loss": 0.0018, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 0.9793939393939394, |
|
"grad_norm": 0.004745373502373695, |
|
"learning_rate": 3.9197930142302716e-05, |
|
"loss": 0.001, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 0.9826262626262626, |
|
"grad_norm": 0.0050310962833464146, |
|
"learning_rate": 3.9327296248382926e-05, |
|
"loss": 0.0013, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 0.9858585858585859, |
|
"grad_norm": 0.008561142720282078, |
|
"learning_rate": 3.945666235446313e-05, |
|
"loss": 0.0018, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.9890909090909091, |
|
"grad_norm": 0.0055971029214560986, |
|
"learning_rate": 3.958602846054334e-05, |
|
"loss": 0.0013, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 0.9923232323232323, |
|
"grad_norm": 0.007602418772876263, |
|
"learning_rate": 3.971539456662355e-05, |
|
"loss": 0.0014, |
|
"step": 307 |
|
}, |
|
{ |
|
"epoch": 0.9955555555555555, |
|
"grad_norm": 0.0054479725658893585, |
|
"learning_rate": 3.9844760672703754e-05, |
|
"loss": 0.0014, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 0.9987878787878788, |
|
"grad_norm": 0.006376450881361961, |
|
"learning_rate": 3.997412677878396e-05, |
|
"loss": 0.0013, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 1.002020202020202, |
|
"grad_norm": 0.00963524729013443, |
|
"learning_rate": 4.010349288486417e-05, |
|
"loss": 0.0025, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 1.0052525252525253, |
|
"grad_norm": 0.004950121510773897, |
|
"learning_rate": 4.023285899094437e-05, |
|
"loss": 0.0009, |
|
"step": 311 |
|
}, |
|
{ |
|
"epoch": 1.0084848484848485, |
|
"grad_norm": 0.0047010756097733974, |
|
"learning_rate": 4.036222509702458e-05, |
|
"loss": 0.0012, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 1.0117171717171718, |
|
"grad_norm": 0.004061982501298189, |
|
"learning_rate": 4.049159120310479e-05, |
|
"loss": 0.0008, |
|
"step": 313 |
|
}, |
|
{ |
|
"epoch": 1.014949494949495, |
|
"grad_norm": 0.006061443593353033, |
|
"learning_rate": 4.0620957309184995e-05, |
|
"loss": 0.0022, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 1.018181818181818, |
|
"grad_norm": 0.005154592916369438, |
|
"learning_rate": 4.07503234152652e-05, |
|
"loss": 0.0023, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 1.0214141414141413, |
|
"grad_norm": 0.004801798611879349, |
|
"learning_rate": 4.087968952134541e-05, |
|
"loss": 0.0012, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 1.0246464646464646, |
|
"grad_norm": 0.004275497514754534, |
|
"learning_rate": 4.100905562742562e-05, |
|
"loss": 0.0012, |
|
"step": 317 |
|
}, |
|
{ |
|
"epoch": 1.0278787878787878, |
|
"grad_norm": 0.006128865294158459, |
|
"learning_rate": 4.113842173350582e-05, |
|
"loss": 0.0022, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 1.031111111111111, |
|
"grad_norm": 0.004465613514184952, |
|
"learning_rate": 4.126778783958603e-05, |
|
"loss": 0.0019, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 1.0343434343434343, |
|
"grad_norm": 0.004612160846590996, |
|
"learning_rate": 4.1397153945666236e-05, |
|
"loss": 0.0013, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 1.0375757575757576, |
|
"grad_norm": 0.00512789748609066, |
|
"learning_rate": 4.152652005174644e-05, |
|
"loss": 0.0012, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 1.0408080808080808, |
|
"grad_norm": 0.006197880953550339, |
|
"learning_rate": 4.165588615782665e-05, |
|
"loss": 0.0015, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 1.044040404040404, |
|
"grad_norm": 0.005857696291059256, |
|
"learning_rate": 4.178525226390686e-05, |
|
"loss": 0.0014, |
|
"step": 323 |
|
}, |
|
{ |
|
"epoch": 1.0472727272727274, |
|
"grad_norm": 0.010157118551433086, |
|
"learning_rate": 4.1914618369987064e-05, |
|
"loss": 0.0015, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 1.0505050505050506, |
|
"grad_norm": 0.006577750667929649, |
|
"learning_rate": 4.2043984476067274e-05, |
|
"loss": 0.0011, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 1.0505050505050506, |
|
"eval_loss": 0.0012751913163810968, |
|
"eval_runtime": 18.6453, |
|
"eval_samples_per_second": 5.363, |
|
"eval_steps_per_second": 1.341, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 1.0537373737373736, |
|
"grad_norm": 0.004863920155912638, |
|
"learning_rate": 4.217335058214748e-05, |
|
"loss": 0.0009, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 1.056969696969697, |
|
"grad_norm": 0.005144843365997076, |
|
"learning_rate": 4.230271668822768e-05, |
|
"loss": 0.0013, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 1.0602020202020201, |
|
"grad_norm": 0.005289267282932997, |
|
"learning_rate": 4.243208279430789e-05, |
|
"loss": 0.0012, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 1.0634343434343434, |
|
"grad_norm": 0.005019436590373516, |
|
"learning_rate": 4.25614489003881e-05, |
|
"loss": 0.0011, |
|
"step": 329 |
|
}, |
|
{ |
|
"epoch": 1.0666666666666667, |
|
"grad_norm": 0.0038786993827670813, |
|
"learning_rate": 4.269081500646831e-05, |
|
"loss": 0.0008, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 1.06989898989899, |
|
"grad_norm": 0.005253227427601814, |
|
"learning_rate": 4.2820181112548515e-05, |
|
"loss": 0.0012, |
|
"step": 331 |
|
}, |
|
{ |
|
"epoch": 1.0731313131313132, |
|
"grad_norm": 0.005880107171833515, |
|
"learning_rate": 4.294954721862872e-05, |
|
"loss": 0.0016, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 1.0763636363636364, |
|
"grad_norm": 0.006426130421459675, |
|
"learning_rate": 4.307891332470893e-05, |
|
"loss": 0.0009, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 1.0795959595959597, |
|
"grad_norm": 0.0038312741089612246, |
|
"learning_rate": 4.320827943078913e-05, |
|
"loss": 0.0009, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 1.082828282828283, |
|
"grad_norm": 0.005139067303389311, |
|
"learning_rate": 4.333764553686934e-05, |
|
"loss": 0.0011, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 1.086060606060606, |
|
"grad_norm": 0.006610156502574682, |
|
"learning_rate": 4.346701164294955e-05, |
|
"loss": 0.0019, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 1.0892929292929292, |
|
"grad_norm": 0.003907322883605957, |
|
"learning_rate": 4.359637774902976e-05, |
|
"loss": 0.001, |
|
"step": 337 |
|
}, |
|
{ |
|
"epoch": 1.0925252525252525, |
|
"grad_norm": 0.006776052061468363, |
|
"learning_rate": 4.372574385510996e-05, |
|
"loss": 0.0011, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 1.0957575757575757, |
|
"grad_norm": 0.007625575177371502, |
|
"learning_rate": 4.385510996119017e-05, |
|
"loss": 0.0016, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 1.098989898989899, |
|
"grad_norm": 0.012086856178939342, |
|
"learning_rate": 4.3984476067270374e-05, |
|
"loss": 0.0035, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 1.1022222222222222, |
|
"grad_norm": 0.007610958535224199, |
|
"learning_rate": 4.4113842173350584e-05, |
|
"loss": 0.0014, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 1.1054545454545455, |
|
"grad_norm": 0.006693006958812475, |
|
"learning_rate": 4.4243208279430794e-05, |
|
"loss": 0.0013, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 1.1086868686868687, |
|
"grad_norm": 0.0045067970640957355, |
|
"learning_rate": 4.4372574385511e-05, |
|
"loss": 0.0011, |
|
"step": 343 |
|
}, |
|
{ |
|
"epoch": 1.111919191919192, |
|
"grad_norm": 0.006877266336232424, |
|
"learning_rate": 4.45019404915912e-05, |
|
"loss": 0.001, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 1.1151515151515152, |
|
"grad_norm": 0.0043833632953464985, |
|
"learning_rate": 4.463130659767141e-05, |
|
"loss": 0.0011, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 1.1183838383838385, |
|
"grad_norm": 0.004103951156139374, |
|
"learning_rate": 4.476067270375162e-05, |
|
"loss": 0.0009, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 1.1216161616161617, |
|
"grad_norm": 0.006397966295480728, |
|
"learning_rate": 4.4890038809831825e-05, |
|
"loss": 0.0008, |
|
"step": 347 |
|
}, |
|
{ |
|
"epoch": 1.1248484848484848, |
|
"grad_norm": 0.007373814936727285, |
|
"learning_rate": 4.5019404915912036e-05, |
|
"loss": 0.0016, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 1.128080808080808, |
|
"grad_norm": 0.005406413692981005, |
|
"learning_rate": 4.514877102199224e-05, |
|
"loss": 0.0012, |
|
"step": 349 |
|
}, |
|
{ |
|
"epoch": 1.1313131313131313, |
|
"grad_norm": 0.0059000300243496895, |
|
"learning_rate": 4.527813712807244e-05, |
|
"loss": 0.0011, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 1.1313131313131313, |
|
"eval_loss": 0.0012826790334656835, |
|
"eval_runtime": 18.7277, |
|
"eval_samples_per_second": 5.34, |
|
"eval_steps_per_second": 1.335, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 1.1345454545454545, |
|
"grad_norm": 0.007405865006148815, |
|
"learning_rate": 4.540750323415265e-05, |
|
"loss": 0.0026, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 1.1377777777777778, |
|
"grad_norm": 0.005577345844358206, |
|
"learning_rate": 4.553686934023286e-05, |
|
"loss": 0.0018, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 1.141010101010101, |
|
"grad_norm": 0.006122430320829153, |
|
"learning_rate": 4.566623544631307e-05, |
|
"loss": 0.0015, |
|
"step": 353 |
|
}, |
|
{ |
|
"epoch": 1.1442424242424243, |
|
"grad_norm": 0.007260697893798351, |
|
"learning_rate": 4.579560155239328e-05, |
|
"loss": 0.0012, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 1.1474747474747475, |
|
"grad_norm": 0.007470736745744944, |
|
"learning_rate": 4.592496765847348e-05, |
|
"loss": 0.0032, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 1.1507070707070708, |
|
"grad_norm": 0.0061160847544670105, |
|
"learning_rate": 4.6054333764553684e-05, |
|
"loss": 0.0013, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 1.1539393939393938, |
|
"grad_norm": 0.006069364957511425, |
|
"learning_rate": 4.6183699870633894e-05, |
|
"loss": 0.0011, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 1.157171717171717, |
|
"grad_norm": 0.004093985538929701, |
|
"learning_rate": 4.6313065976714105e-05, |
|
"loss": 0.0005, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 1.1604040404040403, |
|
"grad_norm": 0.0061876364052295685, |
|
"learning_rate": 4.6442432082794315e-05, |
|
"loss": 0.0017, |
|
"step": 359 |
|
}, |
|
{ |
|
"epoch": 1.1636363636363636, |
|
"grad_norm": 0.0048703462816774845, |
|
"learning_rate": 4.657179818887452e-05, |
|
"loss": 0.0009, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 1.1668686868686868, |
|
"grad_norm": 0.007618950214236975, |
|
"learning_rate": 4.670116429495472e-05, |
|
"loss": 0.0019, |
|
"step": 361 |
|
}, |
|
{ |
|
"epoch": 1.17010101010101, |
|
"grad_norm": 0.004160857293754816, |
|
"learning_rate": 4.683053040103493e-05, |
|
"loss": 0.001, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 1.1733333333333333, |
|
"grad_norm": 0.007051107473671436, |
|
"learning_rate": 4.6959896507115136e-05, |
|
"loss": 0.0019, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 1.1765656565656566, |
|
"grad_norm": 0.004829261917620897, |
|
"learning_rate": 4.7089262613195346e-05, |
|
"loss": 0.001, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 1.1797979797979798, |
|
"grad_norm": 0.005092508625239134, |
|
"learning_rate": 4.7218628719275556e-05, |
|
"loss": 0.0023, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 1.183030303030303, |
|
"grad_norm": 0.004417267628014088, |
|
"learning_rate": 4.734799482535576e-05, |
|
"loss": 0.001, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 1.1862626262626264, |
|
"grad_norm": 0.005750073119997978, |
|
"learning_rate": 4.747736093143596e-05, |
|
"loss": 0.0013, |
|
"step": 367 |
|
}, |
|
{ |
|
"epoch": 1.1894949494949496, |
|
"grad_norm": 0.0033691064454615116, |
|
"learning_rate": 4.760672703751617e-05, |
|
"loss": 0.0008, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 1.1927272727272726, |
|
"grad_norm": 0.0044763232581317425, |
|
"learning_rate": 4.773609314359638e-05, |
|
"loss": 0.0009, |
|
"step": 369 |
|
}, |
|
{ |
|
"epoch": 1.195959595959596, |
|
"grad_norm": 0.007102371193468571, |
|
"learning_rate": 4.786545924967659e-05, |
|
"loss": 0.0012, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 1.1991919191919191, |
|
"grad_norm": 0.0041382270865142345, |
|
"learning_rate": 4.79948253557568e-05, |
|
"loss": 0.001, |
|
"step": 371 |
|
}, |
|
{ |
|
"epoch": 1.2024242424242424, |
|
"grad_norm": 0.006266591139137745, |
|
"learning_rate": 4.8124191461837e-05, |
|
"loss": 0.0015, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 1.2056565656565656, |
|
"grad_norm": 0.00487930653616786, |
|
"learning_rate": 4.8253557567917204e-05, |
|
"loss": 0.0012, |
|
"step": 373 |
|
}, |
|
{ |
|
"epoch": 1.208888888888889, |
|
"grad_norm": 0.005311232525855303, |
|
"learning_rate": 4.8382923673997415e-05, |
|
"loss": 0.0014, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 1.2121212121212122, |
|
"grad_norm": 0.004289830103516579, |
|
"learning_rate": 4.8512289780077625e-05, |
|
"loss": 0.0009, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 1.2121212121212122, |
|
"eval_loss": 0.0011982680298388004, |
|
"eval_runtime": 18.7329, |
|
"eval_samples_per_second": 5.338, |
|
"eval_steps_per_second": 1.335, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 1.2153535353535354, |
|
"grad_norm": 0.0062402524054050446, |
|
"learning_rate": 4.864165588615783e-05, |
|
"loss": 0.0011, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 1.2185858585858587, |
|
"grad_norm": 0.0034229829907417297, |
|
"learning_rate": 4.877102199223804e-05, |
|
"loss": 0.0008, |
|
"step": 377 |
|
}, |
|
{ |
|
"epoch": 1.221818181818182, |
|
"grad_norm": 0.0063551426865160465, |
|
"learning_rate": 4.890038809831824e-05, |
|
"loss": 0.0027, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 1.225050505050505, |
|
"grad_norm": 0.004902714863419533, |
|
"learning_rate": 4.9029754204398446e-05, |
|
"loss": 0.0015, |
|
"step": 379 |
|
}, |
|
{ |
|
"epoch": 1.2282828282828282, |
|
"grad_norm": 0.0044431439600884914, |
|
"learning_rate": 4.9159120310478656e-05, |
|
"loss": 0.0005, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 1.2315151515151515, |
|
"grad_norm": 0.007968113757669926, |
|
"learning_rate": 4.9288486416558866e-05, |
|
"loss": 0.0013, |
|
"step": 381 |
|
}, |
|
{ |
|
"epoch": 1.2347474747474747, |
|
"grad_norm": 0.007964645512402058, |
|
"learning_rate": 4.941785252263907e-05, |
|
"loss": 0.0033, |
|
"step": 382 |
|
}, |
|
{ |
|
"epoch": 1.237979797979798, |
|
"grad_norm": 0.004984153900295496, |
|
"learning_rate": 4.954721862871928e-05, |
|
"loss": 0.0008, |
|
"step": 383 |
|
}, |
|
{ |
|
"epoch": 1.2412121212121212, |
|
"grad_norm": 0.005040554329752922, |
|
"learning_rate": 4.9676584734799483e-05, |
|
"loss": 0.0008, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 1.2444444444444445, |
|
"grad_norm": 0.005269059911370277, |
|
"learning_rate": 4.980595084087969e-05, |
|
"loss": 0.0016, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 1.2476767676767677, |
|
"grad_norm": 0.004820747766643763, |
|
"learning_rate": 4.99353169469599e-05, |
|
"loss": 0.0013, |
|
"step": 386 |
|
}, |
|
{ |
|
"epoch": 1.250909090909091, |
|
"grad_norm": 0.003650473430752754, |
|
"learning_rate": 5.006468305304011e-05, |
|
"loss": 0.001, |
|
"step": 387 |
|
}, |
|
{ |
|
"epoch": 1.2541414141414142, |
|
"grad_norm": 0.003410038072615862, |
|
"learning_rate": 5.019404915912032e-05, |
|
"loss": 0.0007, |
|
"step": 388 |
|
}, |
|
{ |
|
"epoch": 1.2573737373737375, |
|
"grad_norm": 0.00503958947956562, |
|
"learning_rate": 5.032341526520052e-05, |
|
"loss": 0.0013, |
|
"step": 389 |
|
}, |
|
{ |
|
"epoch": 1.2606060606060607, |
|
"grad_norm": 0.006551372352987528, |
|
"learning_rate": 5.0452781371280725e-05, |
|
"loss": 0.0017, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 1.2638383838383838, |
|
"grad_norm": 0.004784473218023777, |
|
"learning_rate": 5.058214747736093e-05, |
|
"loss": 0.0011, |
|
"step": 391 |
|
}, |
|
{ |
|
"epoch": 1.267070707070707, |
|
"grad_norm": 0.009283591993153095, |
|
"learning_rate": 5.071151358344114e-05, |
|
"loss": 0.0017, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 1.2703030303030303, |
|
"grad_norm": 0.005054951179772615, |
|
"learning_rate": 5.084087968952135e-05, |
|
"loss": 0.0014, |
|
"step": 393 |
|
}, |
|
{ |
|
"epoch": 1.2735353535353535, |
|
"grad_norm": 0.007423613220453262, |
|
"learning_rate": 5.097024579560156e-05, |
|
"loss": 0.0012, |
|
"step": 394 |
|
}, |
|
{ |
|
"epoch": 1.2767676767676768, |
|
"grad_norm": 0.004986994434148073, |
|
"learning_rate": 5.109961190168176e-05, |
|
"loss": 0.0017, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"grad_norm": 0.004347431473433971, |
|
"learning_rate": 5.122897800776197e-05, |
|
"loss": 0.0012, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 1.2832323232323233, |
|
"grad_norm": 0.006622066255658865, |
|
"learning_rate": 5.135834411384217e-05, |
|
"loss": 0.0015, |
|
"step": 397 |
|
}, |
|
{ |
|
"epoch": 1.2864646464646465, |
|
"grad_norm": 0.006094200070947409, |
|
"learning_rate": 5.148771021992238e-05, |
|
"loss": 0.0014, |
|
"step": 398 |
|
}, |
|
{ |
|
"epoch": 1.2896969696969696, |
|
"grad_norm": 0.004648863337934017, |
|
"learning_rate": 5.161707632600259e-05, |
|
"loss": 0.0012, |
|
"step": 399 |
|
}, |
|
{ |
|
"epoch": 1.2929292929292928, |
|
"grad_norm": 0.005023527424782515, |
|
"learning_rate": 5.17464424320828e-05, |
|
"loss": 0.0015, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 1.2929292929292928, |
|
"eval_loss": 0.001097235712222755, |
|
"eval_runtime": 18.6509, |
|
"eval_samples_per_second": 5.362, |
|
"eval_steps_per_second": 1.34, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 1.296161616161616, |
|
"grad_norm": 0.004932883661240339, |
|
"learning_rate": 5.1875808538163004e-05, |
|
"loss": 0.0012, |
|
"step": 401 |
|
}, |
|
{ |
|
"epoch": 1.2993939393939393, |
|
"grad_norm": 0.004028505180031061, |
|
"learning_rate": 5.2005174644243214e-05, |
|
"loss": 0.0013, |
|
"step": 402 |
|
}, |
|
{ |
|
"epoch": 1.3026262626262626, |
|
"grad_norm": 0.004800683818757534, |
|
"learning_rate": 5.213454075032341e-05, |
|
"loss": 0.0014, |
|
"step": 403 |
|
}, |
|
{ |
|
"epoch": 1.3058585858585858, |
|
"grad_norm": 0.0042937519028782845, |
|
"learning_rate": 5.226390685640362e-05, |
|
"loss": 0.001, |
|
"step": 404 |
|
}, |
|
{ |
|
"epoch": 1.309090909090909, |
|
"grad_norm": 0.004741652403026819, |
|
"learning_rate": 5.239327296248383e-05, |
|
"loss": 0.0011, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 1.3123232323232323, |
|
"grad_norm": 0.004214128013700247, |
|
"learning_rate": 5.252263906856404e-05, |
|
"loss": 0.0009, |
|
"step": 406 |
|
}, |
|
{ |
|
"epoch": 1.3155555555555556, |
|
"grad_norm": 0.004764182958751917, |
|
"learning_rate": 5.2652005174644245e-05, |
|
"loss": 0.0011, |
|
"step": 407 |
|
}, |
|
{ |
|
"epoch": 1.3187878787878788, |
|
"grad_norm": 0.005405279342085123, |
|
"learning_rate": 5.2781371280724455e-05, |
|
"loss": 0.0011, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 1.322020202020202, |
|
"grad_norm": 0.005355632398277521, |
|
"learning_rate": 5.2910737386804666e-05, |
|
"loss": 0.0012, |
|
"step": 409 |
|
}, |
|
{ |
|
"epoch": 1.3252525252525253, |
|
"grad_norm": 0.004391840659081936, |
|
"learning_rate": 5.304010349288486e-05, |
|
"loss": 0.0011, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 1.3284848484848486, |
|
"grad_norm": 0.002719137817621231, |
|
"learning_rate": 5.316946959896507e-05, |
|
"loss": 0.0005, |
|
"step": 411 |
|
}, |
|
{ |
|
"epoch": 1.3317171717171719, |
|
"grad_norm": 0.005910648964345455, |
|
"learning_rate": 5.329883570504528e-05, |
|
"loss": 0.0014, |
|
"step": 412 |
|
}, |
|
{ |
|
"epoch": 1.3349494949494949, |
|
"grad_norm": 0.004575495608150959, |
|
"learning_rate": 5.3428201811125486e-05, |
|
"loss": 0.0013, |
|
"step": 413 |
|
}, |
|
{ |
|
"epoch": 1.3381818181818181, |
|
"grad_norm": 0.004968920256942511, |
|
"learning_rate": 5.35575679172057e-05, |
|
"loss": 0.0012, |
|
"step": 414 |
|
}, |
|
{ |
|
"epoch": 1.3414141414141414, |
|
"grad_norm": 0.005680674687027931, |
|
"learning_rate": 5.368693402328591e-05, |
|
"loss": 0.0015, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 1.3446464646464646, |
|
"grad_norm": 0.005805244669318199, |
|
"learning_rate": 5.3816300129366104e-05, |
|
"loss": 0.0017, |
|
"step": 416 |
|
}, |
|
{ |
|
"epoch": 1.347878787878788, |
|
"grad_norm": 0.004365504253655672, |
|
"learning_rate": 5.3945666235446314e-05, |
|
"loss": 0.001, |
|
"step": 417 |
|
}, |
|
{ |
|
"epoch": 1.3511111111111112, |
|
"grad_norm": 0.004444981925189495, |
|
"learning_rate": 5.4075032341526524e-05, |
|
"loss": 0.0011, |
|
"step": 418 |
|
}, |
|
{ |
|
"epoch": 1.3543434343434344, |
|
"grad_norm": 0.0048050908371806145, |
|
"learning_rate": 5.420439844760673e-05, |
|
"loss": 0.0011, |
|
"step": 419 |
|
}, |
|
{ |
|
"epoch": 1.3575757575757577, |
|
"grad_norm": 0.0112074613571167, |
|
"learning_rate": 5.433376455368694e-05, |
|
"loss": 0.0009, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 1.3608080808080807, |
|
"grad_norm": 0.005805397406220436, |
|
"learning_rate": 5.446313065976715e-05, |
|
"loss": 0.001, |
|
"step": 421 |
|
}, |
|
{ |
|
"epoch": 1.364040404040404, |
|
"grad_norm": 0.006793915294110775, |
|
"learning_rate": 5.4592496765847345e-05, |
|
"loss": 0.0018, |
|
"step": 422 |
|
}, |
|
{ |
|
"epoch": 1.3672727272727272, |
|
"grad_norm": 0.005133500322699547, |
|
"learning_rate": 5.4721862871927555e-05, |
|
"loss": 0.0015, |
|
"step": 423 |
|
}, |
|
{ |
|
"epoch": 1.3705050505050504, |
|
"grad_norm": 0.0034578884951770306, |
|
"learning_rate": 5.4851228978007766e-05, |
|
"loss": 0.0008, |
|
"step": 424 |
|
}, |
|
{ |
|
"epoch": 1.3737373737373737, |
|
"grad_norm": 0.005579716991633177, |
|
"learning_rate": 5.498059508408797e-05, |
|
"loss": 0.0025, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 1.3737373737373737, |
|
"eval_loss": 0.0011115286033600569, |
|
"eval_runtime": 18.7263, |
|
"eval_samples_per_second": 5.34, |
|
"eval_steps_per_second": 1.335, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 1.376969696969697, |
|
"grad_norm": 0.005262942984700203, |
|
"learning_rate": 5.510996119016818e-05, |
|
"loss": 0.0011, |
|
"step": 426 |
|
}, |
|
{ |
|
"epoch": 1.3802020202020202, |
|
"grad_norm": 0.004813515581190586, |
|
"learning_rate": 5.523932729624839e-05, |
|
"loss": 0.0011, |
|
"step": 427 |
|
}, |
|
{ |
|
"epoch": 1.3834343434343435, |
|
"grad_norm": 0.005556623917073011, |
|
"learning_rate": 5.53686934023286e-05, |
|
"loss": 0.0014, |
|
"step": 428 |
|
}, |
|
{ |
|
"epoch": 1.3866666666666667, |
|
"grad_norm": 0.0028185530100017786, |
|
"learning_rate": 5.5498059508408797e-05, |
|
"loss": 0.0007, |
|
"step": 429 |
|
}, |
|
{ |
|
"epoch": 1.38989898989899, |
|
"grad_norm": 0.004296639934182167, |
|
"learning_rate": 5.562742561448901e-05, |
|
"loss": 0.001, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 1.3931313131313132, |
|
"grad_norm": 0.004092990420758724, |
|
"learning_rate": 5.575679172056921e-05, |
|
"loss": 0.001, |
|
"step": 431 |
|
}, |
|
{ |
|
"epoch": 1.3963636363636365, |
|
"grad_norm": 0.005713030230253935, |
|
"learning_rate": 5.588615782664942e-05, |
|
"loss": 0.0015, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 1.3995959595959597, |
|
"grad_norm": 0.005011020693928003, |
|
"learning_rate": 5.601552393272963e-05, |
|
"loss": 0.0012, |
|
"step": 433 |
|
}, |
|
{ |
|
"epoch": 1.4028282828282828, |
|
"grad_norm": 0.004393043462187052, |
|
"learning_rate": 5.614489003880984e-05, |
|
"loss": 0.0009, |
|
"step": 434 |
|
}, |
|
{ |
|
"epoch": 1.406060606060606, |
|
"grad_norm": 0.0047727166675031185, |
|
"learning_rate": 5.627425614489004e-05, |
|
"loss": 0.0011, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 1.4092929292929293, |
|
"grad_norm": 0.0034502504859119654, |
|
"learning_rate": 5.640362225097025e-05, |
|
"loss": 0.001, |
|
"step": 436 |
|
}, |
|
{ |
|
"epoch": 1.4125252525252525, |
|
"grad_norm": 0.004155455157160759, |
|
"learning_rate": 5.653298835705045e-05, |
|
"loss": 0.0008, |
|
"step": 437 |
|
}, |
|
{ |
|
"epoch": 1.4157575757575758, |
|
"grad_norm": 0.005455946549773216, |
|
"learning_rate": 5.666235446313066e-05, |
|
"loss": 0.0012, |
|
"step": 438 |
|
}, |
|
{ |
|
"epoch": 1.418989898989899, |
|
"grad_norm": 0.005093955434858799, |
|
"learning_rate": 5.679172056921087e-05, |
|
"loss": 0.0015, |
|
"step": 439 |
|
}, |
|
{ |
|
"epoch": 1.4222222222222223, |
|
"grad_norm": 0.004076706245541573, |
|
"learning_rate": 5.692108667529108e-05, |
|
"loss": 0.0012, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 1.4254545454545455, |
|
"grad_norm": 0.004360303282737732, |
|
"learning_rate": 5.7050452781371286e-05, |
|
"loss": 0.001, |
|
"step": 441 |
|
}, |
|
{ |
|
"epoch": 1.4286868686868686, |
|
"grad_norm": 0.005434586200863123, |
|
"learning_rate": 5.717981888745149e-05, |
|
"loss": 0.002, |
|
"step": 442 |
|
}, |
|
{ |
|
"epoch": 1.4319191919191918, |
|
"grad_norm": 0.006644332781434059, |
|
"learning_rate": 5.730918499353169e-05, |
|
"loss": 0.0033, |
|
"step": 443 |
|
}, |
|
{ |
|
"epoch": 1.435151515151515, |
|
"grad_norm": 0.006485591642558575, |
|
"learning_rate": 5.74385510996119e-05, |
|
"loss": 0.002, |
|
"step": 444 |
|
}, |
|
{ |
|
"epoch": 1.4383838383838383, |
|
"grad_norm": 0.006898663938045502, |
|
"learning_rate": 5.7567917205692113e-05, |
|
"loss": 0.001, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 1.4416161616161616, |
|
"grad_norm": 0.006443925201892853, |
|
"learning_rate": 5.7697283311772324e-05, |
|
"loss": 0.0015, |
|
"step": 446 |
|
}, |
|
{ |
|
"epoch": 1.4448484848484848, |
|
"grad_norm": 0.003106453223153949, |
|
"learning_rate": 5.782664941785253e-05, |
|
"loss": 0.0009, |
|
"step": 447 |
|
}, |
|
{ |
|
"epoch": 1.448080808080808, |
|
"grad_norm": 0.0036977077834308147, |
|
"learning_rate": 5.795601552393273e-05, |
|
"loss": 0.0009, |
|
"step": 448 |
|
}, |
|
{ |
|
"epoch": 1.4513131313131313, |
|
"grad_norm": 0.0050973836332559586, |
|
"learning_rate": 5.8085381630012934e-05, |
|
"loss": 0.0014, |
|
"step": 449 |
|
}, |
|
{ |
|
"epoch": 1.4545454545454546, |
|
"grad_norm": 0.0057764300145208836, |
|
"learning_rate": 5.8214747736093145e-05, |
|
"loss": 0.0016, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 1.4545454545454546, |
|
"eval_loss": 0.0010863245697692037, |
|
"eval_runtime": 18.7321, |
|
"eval_samples_per_second": 5.338, |
|
"eval_steps_per_second": 1.335, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 1.4577777777777778, |
|
"grad_norm": 0.004166171886026859, |
|
"learning_rate": 5.8344113842173355e-05, |
|
"loss": 0.0009, |
|
"step": 451 |
|
}, |
|
{ |
|
"epoch": 1.461010101010101, |
|
"grad_norm": 0.004900315310806036, |
|
"learning_rate": 5.8473479948253565e-05, |
|
"loss": 0.0013, |
|
"step": 452 |
|
}, |
|
{ |
|
"epoch": 1.4642424242424243, |
|
"grad_norm": 0.003387290518730879, |
|
"learning_rate": 5.860284605433377e-05, |
|
"loss": 0.0007, |
|
"step": 453 |
|
}, |
|
{ |
|
"epoch": 1.4674747474747476, |
|
"grad_norm": 0.004271840676665306, |
|
"learning_rate": 5.873221216041398e-05, |
|
"loss": 0.0011, |
|
"step": 454 |
|
}, |
|
{ |
|
"epoch": 1.4707070707070706, |
|
"grad_norm": 0.0032306541688740253, |
|
"learning_rate": 5.8861578266494176e-05, |
|
"loss": 0.0007, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 1.4739393939393939, |
|
"grad_norm": 0.0022131800651550293, |
|
"learning_rate": 5.8990944372574386e-05, |
|
"loss": 0.0006, |
|
"step": 456 |
|
}, |
|
{ |
|
"epoch": 1.4771717171717171, |
|
"grad_norm": 0.005533823277801275, |
|
"learning_rate": 5.9120310478654596e-05, |
|
"loss": 0.0017, |
|
"step": 457 |
|
}, |
|
{ |
|
"epoch": 1.4804040404040404, |
|
"grad_norm": 0.004672409035265446, |
|
"learning_rate": 5.9249676584734806e-05, |
|
"loss": 0.0014, |
|
"step": 458 |
|
}, |
|
{ |
|
"epoch": 1.4836363636363636, |
|
"grad_norm": 0.006192138884216547, |
|
"learning_rate": 5.937904269081501e-05, |
|
"loss": 0.0015, |
|
"step": 459 |
|
}, |
|
{ |
|
"epoch": 1.486868686868687, |
|
"grad_norm": 0.011036567389965057, |
|
"learning_rate": 5.950840879689522e-05, |
|
"loss": 0.0025, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 1.4901010101010101, |
|
"grad_norm": 0.0055129267275333405, |
|
"learning_rate": 5.963777490297542e-05, |
|
"loss": 0.0015, |
|
"step": 461 |
|
}, |
|
{ |
|
"epoch": 1.4933333333333334, |
|
"grad_norm": 0.005369866266846657, |
|
"learning_rate": 5.976714100905563e-05, |
|
"loss": 0.0027, |
|
"step": 462 |
|
}, |
|
{ |
|
"epoch": 1.4965656565656564, |
|
"grad_norm": 0.005017601884901524, |
|
"learning_rate": 5.989650711513584e-05, |
|
"loss": 0.0011, |
|
"step": 463 |
|
}, |
|
{ |
|
"epoch": 1.4997979797979797, |
|
"grad_norm": 0.00623415969312191, |
|
"learning_rate": 6.002587322121605e-05, |
|
"loss": 0.0028, |
|
"step": 464 |
|
}, |
|
{ |
|
"epoch": 1.503030303030303, |
|
"grad_norm": 0.005604151636362076, |
|
"learning_rate": 6.015523932729625e-05, |
|
"loss": 0.0023, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 1.5062626262626262, |
|
"grad_norm": 0.004366429056972265, |
|
"learning_rate": 6.028460543337646e-05, |
|
"loss": 0.001, |
|
"step": 466 |
|
}, |
|
{ |
|
"epoch": 1.5094949494949494, |
|
"grad_norm": 0.004982110112905502, |
|
"learning_rate": 6.041397153945667e-05, |
|
"loss": 0.0014, |
|
"step": 467 |
|
}, |
|
{ |
|
"epoch": 1.5127272727272727, |
|
"grad_norm": 0.005147982854396105, |
|
"learning_rate": 6.054333764553687e-05, |
|
"loss": 0.0013, |
|
"step": 468 |
|
}, |
|
{ |
|
"epoch": 1.515959595959596, |
|
"grad_norm": 0.004790551029145718, |
|
"learning_rate": 6.067270375161708e-05, |
|
"loss": 0.0011, |
|
"step": 469 |
|
}, |
|
{ |
|
"epoch": 1.5191919191919192, |
|
"grad_norm": 0.0038922505918890238, |
|
"learning_rate": 6.080206985769729e-05, |
|
"loss": 0.0011, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 1.5224242424242425, |
|
"grad_norm": 0.006303661502897739, |
|
"learning_rate": 6.093143596377749e-05, |
|
"loss": 0.0023, |
|
"step": 471 |
|
}, |
|
{ |
|
"epoch": 1.5256565656565657, |
|
"grad_norm": 0.003573813708499074, |
|
"learning_rate": 6.10608020698577e-05, |
|
"loss": 0.0011, |
|
"step": 472 |
|
}, |
|
{ |
|
"epoch": 1.528888888888889, |
|
"grad_norm": 0.005556274671107531, |
|
"learning_rate": 6.119016817593791e-05, |
|
"loss": 0.0033, |
|
"step": 473 |
|
}, |
|
{ |
|
"epoch": 1.5321212121212122, |
|
"grad_norm": 0.004455295857042074, |
|
"learning_rate": 6.131953428201811e-05, |
|
"loss": 0.0014, |
|
"step": 474 |
|
}, |
|
{ |
|
"epoch": 1.5353535353535355, |
|
"grad_norm": 0.003466435242444277, |
|
"learning_rate": 6.144890038809832e-05, |
|
"loss": 0.001, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 1.5353535353535355, |
|
"eval_loss": 0.0010681893909350038, |
|
"eval_runtime": 18.6082, |
|
"eval_samples_per_second": 5.374, |
|
"eval_steps_per_second": 1.343, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 1.5385858585858587, |
|
"grad_norm": 0.004641688894480467, |
|
"learning_rate": 6.157826649417853e-05, |
|
"loss": 0.0029, |
|
"step": 476 |
|
}, |
|
{ |
|
"epoch": 1.541818181818182, |
|
"grad_norm": 0.004081512801349163, |
|
"learning_rate": 6.170763260025874e-05, |
|
"loss": 0.0011, |
|
"step": 477 |
|
}, |
|
{ |
|
"epoch": 1.545050505050505, |
|
"grad_norm": 0.005640064366161823, |
|
"learning_rate": 6.183699870633895e-05, |
|
"loss": 0.0024, |
|
"step": 478 |
|
}, |
|
{ |
|
"epoch": 1.5482828282828283, |
|
"grad_norm": 0.0031765501480549574, |
|
"learning_rate": 6.196636481241915e-05, |
|
"loss": 0.0009, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 1.5515151515151515, |
|
"grad_norm": 0.0049113016575574875, |
|
"learning_rate": 6.209573091849934e-05, |
|
"loss": 0.0019, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 1.5547474747474748, |
|
"grad_norm": 0.003700861008837819, |
|
"learning_rate": 6.222509702457955e-05, |
|
"loss": 0.0012, |
|
"step": 481 |
|
}, |
|
{ |
|
"epoch": 1.557979797979798, |
|
"grad_norm": 0.004294991493225098, |
|
"learning_rate": 6.235446313065976e-05, |
|
"loss": 0.0014, |
|
"step": 482 |
|
}, |
|
{ |
|
"epoch": 1.561212121212121, |
|
"grad_norm": 0.003475453006103635, |
|
"learning_rate": 6.248382923673998e-05, |
|
"loss": 0.0007, |
|
"step": 483 |
|
}, |
|
{ |
|
"epoch": 1.5644444444444443, |
|
"grad_norm": 0.007765649352222681, |
|
"learning_rate": 6.261319534282019e-05, |
|
"loss": 0.001, |
|
"step": 484 |
|
}, |
|
{ |
|
"epoch": 1.5676767676767676, |
|
"grad_norm": 0.0032908658031374216, |
|
"learning_rate": 6.27425614489004e-05, |
|
"loss": 0.0008, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 1.5709090909090908, |
|
"grad_norm": 0.00453177560120821, |
|
"learning_rate": 6.28719275549806e-05, |
|
"loss": 0.0013, |
|
"step": 486 |
|
}, |
|
{ |
|
"epoch": 1.574141414141414, |
|
"grad_norm": 0.0038091707974672318, |
|
"learning_rate": 6.30012936610608e-05, |
|
"loss": 0.0008, |
|
"step": 487 |
|
}, |
|
{ |
|
"epoch": 1.5773737373737373, |
|
"grad_norm": 0.004123839549720287, |
|
"learning_rate": 6.313065976714101e-05, |
|
"loss": 0.0011, |
|
"step": 488 |
|
}, |
|
{ |
|
"epoch": 1.5806060606060606, |
|
"grad_norm": 0.003308449639007449, |
|
"learning_rate": 6.326002587322122e-05, |
|
"loss": 0.0012, |
|
"step": 489 |
|
}, |
|
{ |
|
"epoch": 1.5838383838383838, |
|
"grad_norm": 0.005438206251710653, |
|
"learning_rate": 6.338939197930143e-05, |
|
"loss": 0.0023, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 1.587070707070707, |
|
"grad_norm": 0.005823109764605761, |
|
"learning_rate": 6.351875808538163e-05, |
|
"loss": 0.0014, |
|
"step": 491 |
|
}, |
|
{ |
|
"epoch": 1.5903030303030303, |
|
"grad_norm": 0.0052726129069924355, |
|
"learning_rate": 6.364812419146184e-05, |
|
"loss": 0.0012, |
|
"step": 492 |
|
}, |
|
{ |
|
"epoch": 1.5935353535353536, |
|
"grad_norm": 0.003648497397080064, |
|
"learning_rate": 6.377749029754204e-05, |
|
"loss": 0.0013, |
|
"step": 493 |
|
}, |
|
{ |
|
"epoch": 1.5967676767676768, |
|
"grad_norm": 0.003007176099345088, |
|
"learning_rate": 6.390685640362225e-05, |
|
"loss": 0.0007, |
|
"step": 494 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"grad_norm": 0.004037541802972555, |
|
"learning_rate": 6.403622250970246e-05, |
|
"loss": 0.0011, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 1.6032323232323233, |
|
"grad_norm": 0.0036643114872276783, |
|
"learning_rate": 6.416558861578267e-05, |
|
"loss": 0.0009, |
|
"step": 496 |
|
}, |
|
{ |
|
"epoch": 1.6064646464646466, |
|
"grad_norm": 0.0038799517787992954, |
|
"learning_rate": 6.429495472186288e-05, |
|
"loss": 0.001, |
|
"step": 497 |
|
}, |
|
{ |
|
"epoch": 1.6096969696969698, |
|
"grad_norm": 0.002598103601485491, |
|
"learning_rate": 6.442432082794309e-05, |
|
"loss": 0.0005, |
|
"step": 498 |
|
}, |
|
{ |
|
"epoch": 1.6129292929292929, |
|
"grad_norm": 0.006756095215678215, |
|
"learning_rate": 6.45536869340233e-05, |
|
"loss": 0.0014, |
|
"step": 499 |
|
}, |
|
{ |
|
"epoch": 1.6161616161616161, |
|
"grad_norm": 0.0038995088543742895, |
|
"learning_rate": 6.46830530401035e-05, |
|
"loss": 0.0007, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 1.6161616161616161, |
|
"eval_loss": 0.0010785168269649148, |
|
"eval_runtime": 18.596, |
|
"eval_samples_per_second": 5.378, |
|
"eval_steps_per_second": 1.344, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 1.6193939393939394, |
|
"grad_norm": 0.004936838988214731, |
|
"learning_rate": 6.48124191461837e-05, |
|
"loss": 0.0012, |
|
"step": 501 |
|
}, |
|
{ |
|
"epoch": 1.6226262626262626, |
|
"grad_norm": 0.006074307020753622, |
|
"learning_rate": 6.494178525226392e-05, |
|
"loss": 0.0022, |
|
"step": 502 |
|
}, |
|
{ |
|
"epoch": 1.625858585858586, |
|
"grad_norm": 0.0044588991440832615, |
|
"learning_rate": 6.507115135834411e-05, |
|
"loss": 0.0011, |
|
"step": 503 |
|
}, |
|
{ |
|
"epoch": 1.6290909090909091, |
|
"grad_norm": 0.004507533740252256, |
|
"learning_rate": 6.520051746442432e-05, |
|
"loss": 0.0011, |
|
"step": 504 |
|
}, |
|
{ |
|
"epoch": 1.6323232323232322, |
|
"grad_norm": 0.0048494781367480755, |
|
"learning_rate": 6.532988357050453e-05, |
|
"loss": 0.0008, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 1.6355555555555554, |
|
"grad_norm": 0.0033034745138138533, |
|
"learning_rate": 6.545924967658473e-05, |
|
"loss": 0.0007, |
|
"step": 506 |
|
}, |
|
{ |
|
"epoch": 1.6387878787878787, |
|
"grad_norm": 0.0053002117201685905, |
|
"learning_rate": 6.558861578266494e-05, |
|
"loss": 0.0009, |
|
"step": 507 |
|
}, |
|
{ |
|
"epoch": 1.642020202020202, |
|
"grad_norm": 0.00404641218483448, |
|
"learning_rate": 6.571798188874515e-05, |
|
"loss": 0.0008, |
|
"step": 508 |
|
}, |
|
{ |
|
"epoch": 1.6452525252525252, |
|
"grad_norm": 0.0036740771029144526, |
|
"learning_rate": 6.584734799482536e-05, |
|
"loss": 0.0009, |
|
"step": 509 |
|
}, |
|
{ |
|
"epoch": 1.6484848484848484, |
|
"grad_norm": 0.005331697873771191, |
|
"learning_rate": 6.597671410090557e-05, |
|
"loss": 0.0014, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 1.6517171717171717, |
|
"grad_norm": 0.004965492524206638, |
|
"learning_rate": 6.610608020698578e-05, |
|
"loss": 0.0019, |
|
"step": 511 |
|
}, |
|
{ |
|
"epoch": 1.654949494949495, |
|
"grad_norm": 0.0037727411836385727, |
|
"learning_rate": 6.623544631306599e-05, |
|
"loss": 0.001, |
|
"step": 512 |
|
}, |
|
{ |
|
"epoch": 1.6581818181818182, |
|
"grad_norm": 0.0048839072696864605, |
|
"learning_rate": 6.636481241914619e-05, |
|
"loss": 0.0013, |
|
"step": 513 |
|
}, |
|
{ |
|
"epoch": 1.6614141414141415, |
|
"grad_norm": 0.004607974551618099, |
|
"learning_rate": 6.649417852522638e-05, |
|
"loss": 0.0017, |
|
"step": 514 |
|
}, |
|
{ |
|
"epoch": 1.6646464646464647, |
|
"grad_norm": 0.004647100809961557, |
|
"learning_rate": 6.66235446313066e-05, |
|
"loss": 0.0024, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 1.667878787878788, |
|
"grad_norm": 0.004482895601540804, |
|
"learning_rate": 6.67529107373868e-05, |
|
"loss": 0.0023, |
|
"step": 516 |
|
}, |
|
{ |
|
"epoch": 1.6711111111111112, |
|
"grad_norm": 0.004681951366364956, |
|
"learning_rate": 6.688227684346702e-05, |
|
"loss": 0.0031, |
|
"step": 517 |
|
}, |
|
{ |
|
"epoch": 1.6743434343434345, |
|
"grad_norm": 0.004061279818415642, |
|
"learning_rate": 6.701164294954723e-05, |
|
"loss": 0.0012, |
|
"step": 518 |
|
}, |
|
{ |
|
"epoch": 1.6775757575757577, |
|
"grad_norm": 0.005936305969953537, |
|
"learning_rate": 6.714100905562742e-05, |
|
"loss": 0.0018, |
|
"step": 519 |
|
}, |
|
{ |
|
"epoch": 1.680808080808081, |
|
"grad_norm": 0.003031841479241848, |
|
"learning_rate": 6.727037516170763e-05, |
|
"loss": 0.0009, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 1.684040404040404, |
|
"grad_norm": 0.0044912113808095455, |
|
"learning_rate": 6.739974126778784e-05, |
|
"loss": 0.0013, |
|
"step": 521 |
|
}, |
|
{ |
|
"epoch": 1.6872727272727273, |
|
"grad_norm": 0.003882101271301508, |
|
"learning_rate": 6.752910737386805e-05, |
|
"loss": 0.0014, |
|
"step": 522 |
|
}, |
|
{ |
|
"epoch": 1.6905050505050505, |
|
"grad_norm": 0.0031570433638989925, |
|
"learning_rate": 6.765847347994826e-05, |
|
"loss": 0.0011, |
|
"step": 523 |
|
}, |
|
{ |
|
"epoch": 1.6937373737373738, |
|
"grad_norm": 0.004184515681117773, |
|
"learning_rate": 6.778783958602847e-05, |
|
"loss": 0.0015, |
|
"step": 524 |
|
}, |
|
{ |
|
"epoch": 1.696969696969697, |
|
"grad_norm": 0.002800683258101344, |
|
"learning_rate": 6.791720569210867e-05, |
|
"loss": 0.0008, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 1.696969696969697, |
|
"eval_loss": 0.0010542384115979075, |
|
"eval_runtime": 18.5989, |
|
"eval_samples_per_second": 5.377, |
|
"eval_steps_per_second": 1.344, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 1.70020202020202, |
|
"grad_norm": 0.003550174878910184, |
|
"learning_rate": 6.804657179818887e-05, |
|
"loss": 0.0014, |
|
"step": 526 |
|
}, |
|
{ |
|
"epoch": 1.7034343434343433, |
|
"grad_norm": 0.0030985972844064236, |
|
"learning_rate": 6.817593790426908e-05, |
|
"loss": 0.0008, |
|
"step": 527 |
|
}, |
|
{ |
|
"epoch": 1.7066666666666666, |
|
"grad_norm": 0.0048317620530724525, |
|
"learning_rate": 6.830530401034929e-05, |
|
"loss": 0.0018, |
|
"step": 528 |
|
}, |
|
{ |
|
"epoch": 1.7098989898989898, |
|
"grad_norm": 0.005726094823330641, |
|
"learning_rate": 6.84346701164295e-05, |
|
"loss": 0.0025, |
|
"step": 529 |
|
}, |
|
{ |
|
"epoch": 1.713131313131313, |
|
"grad_norm": 0.0024808107409626245, |
|
"learning_rate": 6.856403622250971e-05, |
|
"loss": 0.0007, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 1.7163636363636363, |
|
"grad_norm": 0.003422652604058385, |
|
"learning_rate": 6.869340232858992e-05, |
|
"loss": 0.001, |
|
"step": 531 |
|
}, |
|
{ |
|
"epoch": 1.7195959595959596, |
|
"grad_norm": 0.0037957008462399244, |
|
"learning_rate": 6.882276843467012e-05, |
|
"loss": 0.0009, |
|
"step": 532 |
|
}, |
|
{ |
|
"epoch": 1.7228282828282828, |
|
"grad_norm": 0.0028711955528706312, |
|
"learning_rate": 6.895213454075033e-05, |
|
"loss": 0.0008, |
|
"step": 533 |
|
}, |
|
{ |
|
"epoch": 1.726060606060606, |
|
"grad_norm": 0.002812835620716214, |
|
"learning_rate": 6.908150064683054e-05, |
|
"loss": 0.0009, |
|
"step": 534 |
|
}, |
|
{ |
|
"epoch": 1.7292929292929293, |
|
"grad_norm": 0.003739473642781377, |
|
"learning_rate": 6.921086675291075e-05, |
|
"loss": 0.0016, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 1.7325252525252526, |
|
"grad_norm": 0.004807054530829191, |
|
"learning_rate": 6.934023285899096e-05, |
|
"loss": 0.0023, |
|
"step": 536 |
|
}, |
|
{ |
|
"epoch": 1.7357575757575758, |
|
"grad_norm": 0.004576352424919605, |
|
"learning_rate": 6.946959896507115e-05, |
|
"loss": 0.0024, |
|
"step": 537 |
|
}, |
|
{ |
|
"epoch": 1.738989898989899, |
|
"grad_norm": 0.0030553669203072786, |
|
"learning_rate": 6.959896507115135e-05, |
|
"loss": 0.0009, |
|
"step": 538 |
|
}, |
|
{ |
|
"epoch": 1.7422222222222223, |
|
"grad_norm": 0.003879109164699912, |
|
"learning_rate": 6.972833117723156e-05, |
|
"loss": 0.0012, |
|
"step": 539 |
|
}, |
|
{ |
|
"epoch": 1.7454545454545456, |
|
"grad_norm": 0.0031888741068542004, |
|
"learning_rate": 6.985769728331177e-05, |
|
"loss": 0.0013, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 1.7486868686868688, |
|
"grad_norm": 0.0037957336753606796, |
|
"learning_rate": 6.998706338939198e-05, |
|
"loss": 0.0008, |
|
"step": 541 |
|
}, |
|
{ |
|
"epoch": 1.7519191919191919, |
|
"grad_norm": 0.002901956904679537, |
|
"learning_rate": 7.011642949547219e-05, |
|
"loss": 0.0007, |
|
"step": 542 |
|
}, |
|
{ |
|
"epoch": 1.7551515151515151, |
|
"grad_norm": 0.0029531833715736866, |
|
"learning_rate": 7.02457956015524e-05, |
|
"loss": 0.0008, |
|
"step": 543 |
|
}, |
|
{ |
|
"epoch": 1.7583838383838384, |
|
"grad_norm": 0.004063300788402557, |
|
"learning_rate": 7.037516170763261e-05, |
|
"loss": 0.0012, |
|
"step": 544 |
|
}, |
|
{ |
|
"epoch": 1.7616161616161616, |
|
"grad_norm": 0.0041914028115570545, |
|
"learning_rate": 7.050452781371281e-05, |
|
"loss": 0.001, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 1.7648484848484849, |
|
"grad_norm": 0.004884886089712381, |
|
"learning_rate": 7.063389391979302e-05, |
|
"loss": 0.0021, |
|
"step": 546 |
|
}, |
|
{ |
|
"epoch": 1.768080808080808, |
|
"grad_norm": 0.005683641415089369, |
|
"learning_rate": 7.076326002587323e-05, |
|
"loss": 0.0024, |
|
"step": 547 |
|
}, |
|
{ |
|
"epoch": 1.7713131313131312, |
|
"grad_norm": 0.0042356885969638824, |
|
"learning_rate": 7.089262613195343e-05, |
|
"loss": 0.0015, |
|
"step": 548 |
|
}, |
|
{ |
|
"epoch": 1.7745454545454544, |
|
"grad_norm": 0.005650636274367571, |
|
"learning_rate": 7.102199223803364e-05, |
|
"loss": 0.0013, |
|
"step": 549 |
|
}, |
|
{ |
|
"epoch": 1.7777777777777777, |
|
"grad_norm": 0.00460450816899538, |
|
"learning_rate": 7.115135834411385e-05, |
|
"loss": 0.001, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 1.7777777777777777, |
|
"eval_loss": 0.0010012522106990218, |
|
"eval_runtime": 18.6807, |
|
"eval_samples_per_second": 5.353, |
|
"eval_steps_per_second": 1.338, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 1.781010101010101, |
|
"grad_norm": 0.004403329454362392, |
|
"learning_rate": 7.128072445019404e-05, |
|
"loss": 0.0012, |
|
"step": 551 |
|
}, |
|
{ |
|
"epoch": 1.7842424242424242, |
|
"grad_norm": 0.0032920974772423506, |
|
"learning_rate": 7.141009055627425e-05, |
|
"loss": 0.0008, |
|
"step": 552 |
|
}, |
|
{ |
|
"epoch": 1.7874747474747474, |
|
"grad_norm": 0.0038138548843562603, |
|
"learning_rate": 7.153945666235446e-05, |
|
"loss": 0.0011, |
|
"step": 553 |
|
}, |
|
{ |
|
"epoch": 1.7907070707070707, |
|
"grad_norm": 0.00296188285574317, |
|
"learning_rate": 7.166882276843467e-05, |
|
"loss": 0.0008, |
|
"step": 554 |
|
}, |
|
{ |
|
"epoch": 1.793939393939394, |
|
"grad_norm": 0.0031317227985709906, |
|
"learning_rate": 7.179818887451488e-05, |
|
"loss": 0.001, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 1.7971717171717172, |
|
"grad_norm": 0.003491588868200779, |
|
"learning_rate": 7.19275549805951e-05, |
|
"loss": 0.0007, |
|
"step": 556 |
|
}, |
|
{ |
|
"epoch": 1.8004040404040405, |
|
"grad_norm": 0.0031343346927314997, |
|
"learning_rate": 7.20569210866753e-05, |
|
"loss": 0.0005, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 1.8036363636363637, |
|
"grad_norm": 0.004438635893166065, |
|
"learning_rate": 7.21862871927555e-05, |
|
"loss": 0.0011, |
|
"step": 558 |
|
}, |
|
{ |
|
"epoch": 1.806868686868687, |
|
"grad_norm": 0.00411741528660059, |
|
"learning_rate": 7.231565329883571e-05, |
|
"loss": 0.0008, |
|
"step": 559 |
|
}, |
|
{ |
|
"epoch": 1.8101010101010102, |
|
"grad_norm": 0.004590868018567562, |
|
"learning_rate": 7.244501940491591e-05, |
|
"loss": 0.0009, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 1.8133333333333335, |
|
"grad_norm": 0.00454489141702652, |
|
"learning_rate": 7.257438551099612e-05, |
|
"loss": 0.001, |
|
"step": 561 |
|
}, |
|
{ |
|
"epoch": 1.8165656565656567, |
|
"grad_norm": 0.004015767015516758, |
|
"learning_rate": 7.270375161707633e-05, |
|
"loss": 0.0007, |
|
"step": 562 |
|
}, |
|
{ |
|
"epoch": 1.8197979797979797, |
|
"grad_norm": 0.003563391976058483, |
|
"learning_rate": 7.283311772315654e-05, |
|
"loss": 0.0008, |
|
"step": 563 |
|
}, |
|
{ |
|
"epoch": 1.823030303030303, |
|
"grad_norm": 0.0037166401743888855, |
|
"learning_rate": 7.296248382923674e-05, |
|
"loss": 0.0009, |
|
"step": 564 |
|
}, |
|
{ |
|
"epoch": 1.8262626262626263, |
|
"grad_norm": 0.004116456024348736, |
|
"learning_rate": 7.309184993531695e-05, |
|
"loss": 0.0009, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 1.8294949494949495, |
|
"grad_norm": 0.004104538355022669, |
|
"learning_rate": 7.322121604139716e-05, |
|
"loss": 0.0013, |
|
"step": 566 |
|
}, |
|
{ |
|
"epoch": 1.8327272727272728, |
|
"grad_norm": 0.005509045906364918, |
|
"learning_rate": 7.335058214747737e-05, |
|
"loss": 0.0023, |
|
"step": 567 |
|
}, |
|
{ |
|
"epoch": 1.835959595959596, |
|
"grad_norm": 0.0035526566207408905, |
|
"learning_rate": 7.347994825355758e-05, |
|
"loss": 0.0008, |
|
"step": 568 |
|
}, |
|
{ |
|
"epoch": 1.839191919191919, |
|
"grad_norm": 0.003440872300416231, |
|
"learning_rate": 7.360931435963779e-05, |
|
"loss": 0.0007, |
|
"step": 569 |
|
}, |
|
{ |
|
"epoch": 1.8424242424242423, |
|
"grad_norm": 0.0029369727708399296, |
|
"learning_rate": 7.3738680465718e-05, |
|
"loss": 0.0008, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 1.8456565656565656, |
|
"grad_norm": 0.004200720228254795, |
|
"learning_rate": 7.38680465717982e-05, |
|
"loss": 0.0011, |
|
"step": 571 |
|
}, |
|
{ |
|
"epoch": 1.8488888888888888, |
|
"grad_norm": 0.005561929661780596, |
|
"learning_rate": 7.399741267787839e-05, |
|
"loss": 0.0018, |
|
"step": 572 |
|
}, |
|
{ |
|
"epoch": 1.852121212121212, |
|
"grad_norm": 0.0033285904210060835, |
|
"learning_rate": 7.41267787839586e-05, |
|
"loss": 0.0009, |
|
"step": 573 |
|
}, |
|
{ |
|
"epoch": 1.8553535353535353, |
|
"grad_norm": 0.004428067244589329, |
|
"learning_rate": 7.425614489003881e-05, |
|
"loss": 0.0015, |
|
"step": 574 |
|
}, |
|
{ |
|
"epoch": 1.8585858585858586, |
|
"grad_norm": 0.003229588968679309, |
|
"learning_rate": 7.438551099611902e-05, |
|
"loss": 0.0007, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 1.8585858585858586, |
|
"eval_loss": 0.0009781663538888097, |
|
"eval_runtime": 18.7511, |
|
"eval_samples_per_second": 5.333, |
|
"eval_steps_per_second": 1.333, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 1.8618181818181818, |
|
"grad_norm": 0.003516717813909054, |
|
"learning_rate": 7.451487710219923e-05, |
|
"loss": 0.0011, |
|
"step": 576 |
|
}, |
|
{ |
|
"epoch": 1.865050505050505, |
|
"grad_norm": 0.002692221663892269, |
|
"learning_rate": 7.464424320827943e-05, |
|
"loss": 0.0008, |
|
"step": 577 |
|
}, |
|
{ |
|
"epoch": 1.8682828282828283, |
|
"grad_norm": 0.003864881582558155, |
|
"learning_rate": 7.477360931435964e-05, |
|
"loss": 0.0013, |
|
"step": 578 |
|
}, |
|
{ |
|
"epoch": 1.8715151515151516, |
|
"grad_norm": 0.003865283913910389, |
|
"learning_rate": 7.490297542043985e-05, |
|
"loss": 0.0015, |
|
"step": 579 |
|
}, |
|
{ |
|
"epoch": 1.8747474747474748, |
|
"grad_norm": 0.0035206254106014967, |
|
"learning_rate": 7.503234152652006e-05, |
|
"loss": 0.0012, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 1.877979797979798, |
|
"grad_norm": 0.004602121654897928, |
|
"learning_rate": 7.516170763260027e-05, |
|
"loss": 0.0018, |
|
"step": 581 |
|
}, |
|
{ |
|
"epoch": 1.8812121212121213, |
|
"grad_norm": 0.0029543524142354727, |
|
"learning_rate": 7.529107373868047e-05, |
|
"loss": 0.0009, |
|
"step": 582 |
|
}, |
|
{ |
|
"epoch": 1.8844444444444446, |
|
"grad_norm": 0.0023640701547265053, |
|
"learning_rate": 7.542043984476068e-05, |
|
"loss": 0.0007, |
|
"step": 583 |
|
}, |
|
{ |
|
"epoch": 1.8876767676767678, |
|
"grad_norm": 0.0040098559111356735, |
|
"learning_rate": 7.554980595084087e-05, |
|
"loss": 0.0012, |
|
"step": 584 |
|
}, |
|
{ |
|
"epoch": 1.8909090909090909, |
|
"grad_norm": 0.004274715203791857, |
|
"learning_rate": 7.567917205692108e-05, |
|
"loss": 0.0015, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 1.8941414141414141, |
|
"grad_norm": 0.0031054418068379164, |
|
"learning_rate": 7.58085381630013e-05, |
|
"loss": 0.0008, |
|
"step": 586 |
|
}, |
|
{ |
|
"epoch": 1.8973737373737374, |
|
"grad_norm": 0.003954428713768721, |
|
"learning_rate": 7.59379042690815e-05, |
|
"loss": 0.0009, |
|
"step": 587 |
|
}, |
|
{ |
|
"epoch": 1.9006060606060606, |
|
"grad_norm": 0.003786911489441991, |
|
"learning_rate": 7.606727037516171e-05, |
|
"loss": 0.0006, |
|
"step": 588 |
|
}, |
|
{ |
|
"epoch": 1.9038383838383839, |
|
"grad_norm": 0.0041464087553322315, |
|
"learning_rate": 7.619663648124192e-05, |
|
"loss": 0.0011, |
|
"step": 589 |
|
}, |
|
{ |
|
"epoch": 1.907070707070707, |
|
"grad_norm": 0.004568018950521946, |
|
"learning_rate": 7.632600258732212e-05, |
|
"loss": 0.0019, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 1.9103030303030302, |
|
"grad_norm": 0.003288006642833352, |
|
"learning_rate": 7.645536869340233e-05, |
|
"loss": 0.0009, |
|
"step": 591 |
|
}, |
|
{ |
|
"epoch": 1.9135353535353534, |
|
"grad_norm": 0.004277890548110008, |
|
"learning_rate": 7.658473479948254e-05, |
|
"loss": 0.0008, |
|
"step": 592 |
|
}, |
|
{ |
|
"epoch": 1.9167676767676767, |
|
"grad_norm": 0.004379922058433294, |
|
"learning_rate": 7.671410090556275e-05, |
|
"loss": 0.001, |
|
"step": 593 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"grad_norm": 0.003021504729986191, |
|
"learning_rate": 7.684346701164295e-05, |
|
"loss": 0.0006, |
|
"step": 594 |
|
}, |
|
{ |
|
"epoch": 1.9232323232323232, |
|
"grad_norm": 0.0032772128470242023, |
|
"learning_rate": 7.697283311772316e-05, |
|
"loss": 0.0008, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 1.9264646464646464, |
|
"grad_norm": 0.004533900413662195, |
|
"learning_rate": 7.710219922380336e-05, |
|
"loss": 0.0015, |
|
"step": 596 |
|
}, |
|
{ |
|
"epoch": 1.9296969696969697, |
|
"grad_norm": 0.005284740123897791, |
|
"learning_rate": 7.723156532988357e-05, |
|
"loss": 0.0011, |
|
"step": 597 |
|
}, |
|
{ |
|
"epoch": 1.932929292929293, |
|
"grad_norm": 0.005875818431377411, |
|
"learning_rate": 7.736093143596378e-05, |
|
"loss": 0.0013, |
|
"step": 598 |
|
}, |
|
{ |
|
"epoch": 1.9361616161616162, |
|
"grad_norm": 0.004483737051486969, |
|
"learning_rate": 7.749029754204399e-05, |
|
"loss": 0.0009, |
|
"step": 599 |
|
}, |
|
{ |
|
"epoch": 1.9393939393939394, |
|
"grad_norm": 0.004435641225427389, |
|
"learning_rate": 7.76196636481242e-05, |
|
"loss": 0.0013, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 1.9393939393939394, |
|
"eval_loss": 0.0009266917477361858, |
|
"eval_runtime": 18.7646, |
|
"eval_samples_per_second": 5.329, |
|
"eval_steps_per_second": 1.332, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 1.9426262626262627, |
|
"grad_norm": 0.003876154311001301, |
|
"learning_rate": 7.774902975420441e-05, |
|
"loss": 0.001, |
|
"step": 601 |
|
}, |
|
{ |
|
"epoch": 1.945858585858586, |
|
"grad_norm": 0.004851337987929583, |
|
"learning_rate": 7.787839586028462e-05, |
|
"loss": 0.0014, |
|
"step": 602 |
|
}, |
|
{ |
|
"epoch": 1.9490909090909092, |
|
"grad_norm": 0.003333737375214696, |
|
"learning_rate": 7.800776196636481e-05, |
|
"loss": 0.001, |
|
"step": 603 |
|
}, |
|
{ |
|
"epoch": 1.9523232323232325, |
|
"grad_norm": 0.0032944250851869583, |
|
"learning_rate": 7.813712807244502e-05, |
|
"loss": 0.0007, |
|
"step": 604 |
|
}, |
|
{ |
|
"epoch": 1.9555555555555557, |
|
"grad_norm": 0.004762569442391396, |
|
"learning_rate": 7.826649417852523e-05, |
|
"loss": 0.0027, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 1.9587878787878787, |
|
"grad_norm": 0.0037507396191358566, |
|
"learning_rate": 7.839586028460543e-05, |
|
"loss": 0.0014, |
|
"step": 606 |
|
}, |
|
{ |
|
"epoch": 1.962020202020202, |
|
"grad_norm": 0.005207899492233992, |
|
"learning_rate": 7.852522639068564e-05, |
|
"loss": 0.0031, |
|
"step": 607 |
|
}, |
|
{ |
|
"epoch": 1.9652525252525253, |
|
"grad_norm": 0.0035236128605902195, |
|
"learning_rate": 7.865459249676585e-05, |
|
"loss": 0.001, |
|
"step": 608 |
|
}, |
|
{ |
|
"epoch": 1.9684848484848485, |
|
"grad_norm": 0.004186858423054218, |
|
"learning_rate": 7.878395860284605e-05, |
|
"loss": 0.0012, |
|
"step": 609 |
|
}, |
|
{ |
|
"epoch": 1.9717171717171718, |
|
"grad_norm": 0.0038515774067491293, |
|
"learning_rate": 7.891332470892626e-05, |
|
"loss": 0.0022, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 1.9749494949494948, |
|
"grad_norm": 0.0030782243702560663, |
|
"learning_rate": 7.904269081500647e-05, |
|
"loss": 0.0006, |
|
"step": 611 |
|
}, |
|
{ |
|
"epoch": 1.978181818181818, |
|
"grad_norm": 0.005162171553820372, |
|
"learning_rate": 7.917205692108668e-05, |
|
"loss": 0.0013, |
|
"step": 612 |
|
}, |
|
{ |
|
"epoch": 1.9814141414141413, |
|
"grad_norm": 0.004226456396281719, |
|
"learning_rate": 7.930142302716689e-05, |
|
"loss": 0.0009, |
|
"step": 613 |
|
}, |
|
{ |
|
"epoch": 1.9846464646464645, |
|
"grad_norm": 0.003615841967985034, |
|
"learning_rate": 7.94307891332471e-05, |
|
"loss": 0.0022, |
|
"step": 614 |
|
}, |
|
{ |
|
"epoch": 1.9878787878787878, |
|
"grad_norm": 0.00408519571647048, |
|
"learning_rate": 7.956015523932731e-05, |
|
"loss": 0.0015, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 1.991111111111111, |
|
"grad_norm": 0.0042512728832662106, |
|
"learning_rate": 7.968952134540751e-05, |
|
"loss": 0.0014, |
|
"step": 616 |
|
}, |
|
{ |
|
"epoch": 1.9943434343434343, |
|
"grad_norm": 0.003178700339049101, |
|
"learning_rate": 7.981888745148772e-05, |
|
"loss": 0.0008, |
|
"step": 617 |
|
}, |
|
{ |
|
"epoch": 1.9975757575757576, |
|
"grad_norm": 0.003822662867605686, |
|
"learning_rate": 7.994825355756791e-05, |
|
"loss": 0.001, |
|
"step": 618 |
|
}, |
|
{ |
|
"epoch": 2.000808080808081, |
|
"grad_norm": 0.0046024019829928875, |
|
"learning_rate": 8.007761966364812e-05, |
|
"loss": 0.0011, |
|
"step": 619 |
|
}, |
|
{ |
|
"epoch": 2.004040404040404, |
|
"grad_norm": 0.0038011916913092136, |
|
"learning_rate": 8.020698576972833e-05, |
|
"loss": 0.0008, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 2.0072727272727273, |
|
"grad_norm": 0.006909184157848358, |
|
"learning_rate": 8.033635187580855e-05, |
|
"loss": 0.0026, |
|
"step": 621 |
|
}, |
|
{ |
|
"epoch": 2.0105050505050506, |
|
"grad_norm": 0.005042599979788065, |
|
"learning_rate": 8.046571798188874e-05, |
|
"loss": 0.0011, |
|
"step": 622 |
|
}, |
|
{ |
|
"epoch": 2.013737373737374, |
|
"grad_norm": 0.004002322442829609, |
|
"learning_rate": 8.059508408796895e-05, |
|
"loss": 0.0011, |
|
"step": 623 |
|
}, |
|
{ |
|
"epoch": 2.016969696969697, |
|
"grad_norm": 0.0031209783628582954, |
|
"learning_rate": 8.072445019404916e-05, |
|
"loss": 0.0009, |
|
"step": 624 |
|
}, |
|
{ |
|
"epoch": 2.0202020202020203, |
|
"grad_norm": 0.0026158462278544903, |
|
"learning_rate": 8.085381630012937e-05, |
|
"loss": 0.0007, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 2.0202020202020203, |
|
"eval_loss": 0.0009721739334054291, |
|
"eval_runtime": 18.7204, |
|
"eval_samples_per_second": 5.342, |
|
"eval_steps_per_second": 1.335, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 2.0234343434343436, |
|
"grad_norm": 0.003943873103708029, |
|
"learning_rate": 8.098318240620958e-05, |
|
"loss": 0.001, |
|
"step": 626 |
|
}, |
|
{ |
|
"epoch": 2.026666666666667, |
|
"grad_norm": 0.0038578941021114588, |
|
"learning_rate": 8.111254851228979e-05, |
|
"loss": 0.0017, |
|
"step": 627 |
|
}, |
|
{ |
|
"epoch": 2.02989898989899, |
|
"grad_norm": 0.002991423010826111, |
|
"learning_rate": 8.124191461836999e-05, |
|
"loss": 0.0007, |
|
"step": 628 |
|
}, |
|
{ |
|
"epoch": 2.0331313131313133, |
|
"grad_norm": 0.002815463813021779, |
|
"learning_rate": 8.13712807244502e-05, |
|
"loss": 0.0009, |
|
"step": 629 |
|
}, |
|
{ |
|
"epoch": 2.036363636363636, |
|
"grad_norm": 0.0025943731889128685, |
|
"learning_rate": 8.15006468305304e-05, |
|
"loss": 0.0006, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 2.0395959595959594, |
|
"grad_norm": 0.0029918155632913113, |
|
"learning_rate": 8.163001293661061e-05, |
|
"loss": 0.0008, |
|
"step": 631 |
|
}, |
|
{ |
|
"epoch": 2.0428282828282827, |
|
"grad_norm": 0.00398919777944684, |
|
"learning_rate": 8.175937904269082e-05, |
|
"loss": 0.0011, |
|
"step": 632 |
|
}, |
|
{ |
|
"epoch": 2.046060606060606, |
|
"grad_norm": 0.003045483957976103, |
|
"learning_rate": 8.188874514877103e-05, |
|
"loss": 0.0007, |
|
"step": 633 |
|
}, |
|
{ |
|
"epoch": 2.049292929292929, |
|
"grad_norm": 0.005803203675895929, |
|
"learning_rate": 8.201811125485124e-05, |
|
"loss": 0.0008, |
|
"step": 634 |
|
}, |
|
{ |
|
"epoch": 2.0525252525252524, |
|
"grad_norm": 0.003953828942030668, |
|
"learning_rate": 8.214747736093143e-05, |
|
"loss": 0.001, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 2.0557575757575757, |
|
"grad_norm": 0.003973971586674452, |
|
"learning_rate": 8.227684346701164e-05, |
|
"loss": 0.0006, |
|
"step": 636 |
|
}, |
|
{ |
|
"epoch": 2.058989898989899, |
|
"grad_norm": 0.0025888276286423206, |
|
"learning_rate": 8.240620957309186e-05, |
|
"loss": 0.0005, |
|
"step": 637 |
|
}, |
|
{ |
|
"epoch": 2.062222222222222, |
|
"grad_norm": 0.003989651799201965, |
|
"learning_rate": 8.253557567917207e-05, |
|
"loss": 0.0008, |
|
"step": 638 |
|
}, |
|
{ |
|
"epoch": 2.0654545454545454, |
|
"grad_norm": 0.0030330433510243893, |
|
"learning_rate": 8.266494178525228e-05, |
|
"loss": 0.0009, |
|
"step": 639 |
|
}, |
|
{ |
|
"epoch": 2.0686868686868687, |
|
"grad_norm": 0.0044325897470116615, |
|
"learning_rate": 8.279430789133247e-05, |
|
"loss": 0.0016, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 2.071919191919192, |
|
"grad_norm": 0.002264161128550768, |
|
"learning_rate": 8.292367399741268e-05, |
|
"loss": 0.0006, |
|
"step": 641 |
|
}, |
|
{ |
|
"epoch": 2.075151515151515, |
|
"grad_norm": 0.0033110452350229025, |
|
"learning_rate": 8.305304010349288e-05, |
|
"loss": 0.001, |
|
"step": 642 |
|
}, |
|
{ |
|
"epoch": 2.0783838383838384, |
|
"grad_norm": 0.0029199772980064154, |
|
"learning_rate": 8.318240620957309e-05, |
|
"loss": 0.0006, |
|
"step": 643 |
|
}, |
|
{ |
|
"epoch": 2.0816161616161617, |
|
"grad_norm": 0.0031028217636048794, |
|
"learning_rate": 8.33117723156533e-05, |
|
"loss": 0.0007, |
|
"step": 644 |
|
}, |
|
{ |
|
"epoch": 2.084848484848485, |
|
"grad_norm": 0.004011626821011305, |
|
"learning_rate": 8.344113842173351e-05, |
|
"loss": 0.0009, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 2.088080808080808, |
|
"grad_norm": 0.003449072130024433, |
|
"learning_rate": 8.357050452781372e-05, |
|
"loss": 0.0009, |
|
"step": 646 |
|
}, |
|
{ |
|
"epoch": 2.0913131313131315, |
|
"grad_norm": 0.003367091529071331, |
|
"learning_rate": 8.369987063389393e-05, |
|
"loss": 0.001, |
|
"step": 647 |
|
}, |
|
{ |
|
"epoch": 2.0945454545454547, |
|
"grad_norm": 0.003321894910186529, |
|
"learning_rate": 8.382923673997413e-05, |
|
"loss": 0.0016, |
|
"step": 648 |
|
}, |
|
{ |
|
"epoch": 2.097777777777778, |
|
"grad_norm": 0.003046546597033739, |
|
"learning_rate": 8.395860284605434e-05, |
|
"loss": 0.0008, |
|
"step": 649 |
|
}, |
|
{ |
|
"epoch": 2.101010101010101, |
|
"grad_norm": 0.0023006205447018147, |
|
"learning_rate": 8.408796895213455e-05, |
|
"loss": 0.0006, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 2.101010101010101, |
|
"eval_loss": 0.0009433354716747999, |
|
"eval_runtime": 18.7374, |
|
"eval_samples_per_second": 5.337, |
|
"eval_steps_per_second": 1.334, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 2.1042424242424245, |
|
"grad_norm": 0.004013998433947563, |
|
"learning_rate": 8.421733505821476e-05, |
|
"loss": 0.001, |
|
"step": 651 |
|
}, |
|
{ |
|
"epoch": 2.1074747474747473, |
|
"grad_norm": 0.0035437876358628273, |
|
"learning_rate": 8.434670116429496e-05, |
|
"loss": 0.0007, |
|
"step": 652 |
|
}, |
|
{ |
|
"epoch": 2.1107070707070705, |
|
"grad_norm": 0.004152386449277401, |
|
"learning_rate": 8.447606727037517e-05, |
|
"loss": 0.001, |
|
"step": 653 |
|
}, |
|
{ |
|
"epoch": 2.113939393939394, |
|
"grad_norm": 0.0031009165104478598, |
|
"learning_rate": 8.460543337645536e-05, |
|
"loss": 0.0007, |
|
"step": 654 |
|
}, |
|
{ |
|
"epoch": 2.117171717171717, |
|
"grad_norm": 0.002030389616265893, |
|
"learning_rate": 8.473479948253557e-05, |
|
"loss": 0.0005, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 2.1204040404040403, |
|
"grad_norm": 0.003986351191997528, |
|
"learning_rate": 8.486416558861578e-05, |
|
"loss": 0.0011, |
|
"step": 656 |
|
}, |
|
{ |
|
"epoch": 2.1236363636363635, |
|
"grad_norm": 0.0037633986212313175, |
|
"learning_rate": 8.499353169469599e-05, |
|
"loss": 0.0017, |
|
"step": 657 |
|
}, |
|
{ |
|
"epoch": 2.126868686868687, |
|
"grad_norm": 0.003191509749740362, |
|
"learning_rate": 8.51228978007762e-05, |
|
"loss": 0.0009, |
|
"step": 658 |
|
}, |
|
{ |
|
"epoch": 2.13010101010101, |
|
"grad_norm": 0.003234416712075472, |
|
"learning_rate": 8.525226390685641e-05, |
|
"loss": 0.0015, |
|
"step": 659 |
|
}, |
|
{ |
|
"epoch": 2.1333333333333333, |
|
"grad_norm": 0.00314782140776515, |
|
"learning_rate": 8.538163001293662e-05, |
|
"loss": 0.0008, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 2.1365656565656566, |
|
"grad_norm": 0.004626644309610128, |
|
"learning_rate": 8.551099611901682e-05, |
|
"loss": 0.0017, |
|
"step": 661 |
|
}, |
|
{ |
|
"epoch": 2.13979797979798, |
|
"grad_norm": 0.0037515638396143913, |
|
"learning_rate": 8.564036222509703e-05, |
|
"loss": 0.001, |
|
"step": 662 |
|
}, |
|
{ |
|
"epoch": 2.143030303030303, |
|
"grad_norm": 0.0028777304105460644, |
|
"learning_rate": 8.576972833117724e-05, |
|
"loss": 0.0009, |
|
"step": 663 |
|
}, |
|
{ |
|
"epoch": 2.1462626262626263, |
|
"grad_norm": 0.0035881204530596733, |
|
"learning_rate": 8.589909443725744e-05, |
|
"loss": 0.0008, |
|
"step": 664 |
|
}, |
|
{ |
|
"epoch": 2.1494949494949496, |
|
"grad_norm": 0.002268304582685232, |
|
"learning_rate": 8.602846054333765e-05, |
|
"loss": 0.0005, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 2.152727272727273, |
|
"grad_norm": 0.003100323723629117, |
|
"learning_rate": 8.615782664941786e-05, |
|
"loss": 0.0008, |
|
"step": 666 |
|
}, |
|
{ |
|
"epoch": 2.155959595959596, |
|
"grad_norm": 0.003520137397572398, |
|
"learning_rate": 8.628719275549805e-05, |
|
"loss": 0.0009, |
|
"step": 667 |
|
}, |
|
{ |
|
"epoch": 2.1591919191919193, |
|
"grad_norm": 0.0042273253202438354, |
|
"learning_rate": 8.641655886157827e-05, |
|
"loss": 0.0015, |
|
"step": 668 |
|
}, |
|
{ |
|
"epoch": 2.1624242424242426, |
|
"grad_norm": 0.0030146201606839895, |
|
"learning_rate": 8.654592496765848e-05, |
|
"loss": 0.0009, |
|
"step": 669 |
|
}, |
|
{ |
|
"epoch": 2.165656565656566, |
|
"grad_norm": 0.0055503263138234615, |
|
"learning_rate": 8.667529107373869e-05, |
|
"loss": 0.0025, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 2.168888888888889, |
|
"grad_norm": 0.0036939766723662615, |
|
"learning_rate": 8.68046571798189e-05, |
|
"loss": 0.002, |
|
"step": 671 |
|
}, |
|
{ |
|
"epoch": 2.172121212121212, |
|
"grad_norm": 0.0032185425516217947, |
|
"learning_rate": 8.69340232858991e-05, |
|
"loss": 0.0009, |
|
"step": 672 |
|
}, |
|
{ |
|
"epoch": 2.175353535353535, |
|
"grad_norm": 0.0030356363859027624, |
|
"learning_rate": 8.706338939197932e-05, |
|
"loss": 0.0007, |
|
"step": 673 |
|
}, |
|
{ |
|
"epoch": 2.1785858585858584, |
|
"grad_norm": 0.006692877039313316, |
|
"learning_rate": 8.719275549805951e-05, |
|
"loss": 0.0008, |
|
"step": 674 |
|
}, |
|
{ |
|
"epoch": 2.1818181818181817, |
|
"grad_norm": 0.0026474855840206146, |
|
"learning_rate": 8.732212160413972e-05, |
|
"loss": 0.0007, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 2.1818181818181817, |
|
"eval_loss": 0.0009410877246409655, |
|
"eval_runtime": 18.7748, |
|
"eval_samples_per_second": 5.326, |
|
"eval_steps_per_second": 1.332, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 2.185050505050505, |
|
"grad_norm": 0.003672214224934578, |
|
"learning_rate": 8.745148771021992e-05, |
|
"loss": 0.0018, |
|
"step": 676 |
|
}, |
|
{ |
|
"epoch": 2.188282828282828, |
|
"grad_norm": 0.0031880387105047703, |
|
"learning_rate": 8.758085381630013e-05, |
|
"loss": 0.0012, |
|
"step": 677 |
|
}, |
|
{ |
|
"epoch": 2.1915151515151514, |
|
"grad_norm": 0.0033309224527329206, |
|
"learning_rate": 8.771021992238034e-05, |
|
"loss": 0.001, |
|
"step": 678 |
|
}, |
|
{ |
|
"epoch": 2.1947474747474747, |
|
"grad_norm": 0.003027291502803564, |
|
"learning_rate": 8.783958602846055e-05, |
|
"loss": 0.0008, |
|
"step": 679 |
|
}, |
|
{ |
|
"epoch": 2.197979797979798, |
|
"grad_norm": 0.0036887172609567642, |
|
"learning_rate": 8.796895213454075e-05, |
|
"loss": 0.001, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 2.201212121212121, |
|
"grad_norm": 0.003997828811407089, |
|
"learning_rate": 8.809831824062096e-05, |
|
"loss": 0.0011, |
|
"step": 681 |
|
}, |
|
{ |
|
"epoch": 2.2044444444444444, |
|
"grad_norm": 0.003314135130494833, |
|
"learning_rate": 8.822768434670117e-05, |
|
"loss": 0.0017, |
|
"step": 682 |
|
}, |
|
{ |
|
"epoch": 2.2076767676767677, |
|
"grad_norm": 0.0036913359072059393, |
|
"learning_rate": 8.835705045278138e-05, |
|
"loss": 0.0009, |
|
"step": 683 |
|
}, |
|
{ |
|
"epoch": 2.210909090909091, |
|
"grad_norm": 0.0030470637138932943, |
|
"learning_rate": 8.848641655886159e-05, |
|
"loss": 0.0008, |
|
"step": 684 |
|
}, |
|
{ |
|
"epoch": 2.214141414141414, |
|
"grad_norm": 0.004008392803370953, |
|
"learning_rate": 8.86157826649418e-05, |
|
"loss": 0.0012, |
|
"step": 685 |
|
}, |
|
{ |
|
"epoch": 2.2173737373737374, |
|
"grad_norm": 0.0055717285722494125, |
|
"learning_rate": 8.8745148771022e-05, |
|
"loss": 0.0019, |
|
"step": 686 |
|
}, |
|
{ |
|
"epoch": 2.2206060606060607, |
|
"grad_norm": 0.002990501932799816, |
|
"learning_rate": 8.88745148771022e-05, |
|
"loss": 0.0009, |
|
"step": 687 |
|
}, |
|
{ |
|
"epoch": 2.223838383838384, |
|
"grad_norm": 0.002291604643687606, |
|
"learning_rate": 8.90038809831824e-05, |
|
"loss": 0.0006, |
|
"step": 688 |
|
}, |
|
{ |
|
"epoch": 2.227070707070707, |
|
"grad_norm": 0.002231605350971222, |
|
"learning_rate": 8.913324708926261e-05, |
|
"loss": 0.0006, |
|
"step": 689 |
|
}, |
|
{ |
|
"epoch": 2.2303030303030305, |
|
"grad_norm": 0.002650222275406122, |
|
"learning_rate": 8.926261319534282e-05, |
|
"loss": 0.0007, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 2.2335353535353537, |
|
"grad_norm": 0.0019864251371473074, |
|
"learning_rate": 8.939197930142303e-05, |
|
"loss": 0.0006, |
|
"step": 691 |
|
}, |
|
{ |
|
"epoch": 2.236767676767677, |
|
"grad_norm": 0.0028250280302017927, |
|
"learning_rate": 8.952134540750324e-05, |
|
"loss": 0.0006, |
|
"step": 692 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"grad_norm": 0.0035562533885240555, |
|
"learning_rate": 8.965071151358344e-05, |
|
"loss": 0.0018, |
|
"step": 693 |
|
}, |
|
{ |
|
"epoch": 2.2432323232323235, |
|
"grad_norm": 0.0030060771387070417, |
|
"learning_rate": 8.978007761966365e-05, |
|
"loss": 0.0007, |
|
"step": 694 |
|
}, |
|
{ |
|
"epoch": 2.2464646464646463, |
|
"grad_norm": 0.002113576978445053, |
|
"learning_rate": 8.990944372574386e-05, |
|
"loss": 0.0005, |
|
"step": 695 |
|
}, |
|
{ |
|
"epoch": 2.2496969696969695, |
|
"grad_norm": 0.002114512724801898, |
|
"learning_rate": 9.003880983182407e-05, |
|
"loss": 0.0006, |
|
"step": 696 |
|
}, |
|
{ |
|
"epoch": 2.252929292929293, |
|
"grad_norm": 0.0029473064932972193, |
|
"learning_rate": 9.016817593790428e-05, |
|
"loss": 0.0006, |
|
"step": 697 |
|
}, |
|
{ |
|
"epoch": 2.256161616161616, |
|
"grad_norm": 0.004192824941128492, |
|
"learning_rate": 9.029754204398448e-05, |
|
"loss": 0.001, |
|
"step": 698 |
|
}, |
|
{ |
|
"epoch": 2.2593939393939393, |
|
"grad_norm": 0.003509392263367772, |
|
"learning_rate": 9.042690815006469e-05, |
|
"loss": 0.0009, |
|
"step": 699 |
|
}, |
|
{ |
|
"epoch": 2.2626262626262625, |
|
"grad_norm": 0.004787352867424488, |
|
"learning_rate": 9.055627425614489e-05, |
|
"loss": 0.001, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 2.2626262626262625, |
|
"eval_loss": 0.000903558568097651, |
|
"eval_runtime": 18.694, |
|
"eval_samples_per_second": 5.349, |
|
"eval_steps_per_second": 1.337, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 2.265858585858586, |
|
"grad_norm": 0.0027677167672663927, |
|
"learning_rate": 9.06856403622251e-05, |
|
"loss": 0.0009, |
|
"step": 701 |
|
}, |
|
{ |
|
"epoch": 2.269090909090909, |
|
"grad_norm": 0.0026491908356547356, |
|
"learning_rate": 9.08150064683053e-05, |
|
"loss": 0.0007, |
|
"step": 702 |
|
}, |
|
{ |
|
"epoch": 2.2723232323232323, |
|
"grad_norm": 0.0025004090275615454, |
|
"learning_rate": 9.094437257438552e-05, |
|
"loss": 0.0007, |
|
"step": 703 |
|
}, |
|
{ |
|
"epoch": 2.2755555555555556, |
|
"grad_norm": 0.0028262247797101736, |
|
"learning_rate": 9.107373868046573e-05, |
|
"loss": 0.0008, |
|
"step": 704 |
|
}, |
|
{ |
|
"epoch": 2.278787878787879, |
|
"grad_norm": 0.003089368110522628, |
|
"learning_rate": 9.120310478654594e-05, |
|
"loss": 0.0009, |
|
"step": 705 |
|
}, |
|
{ |
|
"epoch": 2.282020202020202, |
|
"grad_norm": 0.002824244322255254, |
|
"learning_rate": 9.133247089262613e-05, |
|
"loss": 0.0008, |
|
"step": 706 |
|
}, |
|
{ |
|
"epoch": 2.2852525252525253, |
|
"grad_norm": 0.0023714362177997828, |
|
"learning_rate": 9.146183699870634e-05, |
|
"loss": 0.0007, |
|
"step": 707 |
|
}, |
|
{ |
|
"epoch": 2.2884848484848486, |
|
"grad_norm": 0.0033874395303428173, |
|
"learning_rate": 9.159120310478655e-05, |
|
"loss": 0.0017, |
|
"step": 708 |
|
}, |
|
{ |
|
"epoch": 2.291717171717172, |
|
"grad_norm": 0.0033185749780386686, |
|
"learning_rate": 9.172056921086676e-05, |
|
"loss": 0.0017, |
|
"step": 709 |
|
}, |
|
{ |
|
"epoch": 2.294949494949495, |
|
"grad_norm": 0.00360478856600821, |
|
"learning_rate": 9.184993531694696e-05, |
|
"loss": 0.0009, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 2.2981818181818183, |
|
"grad_norm": 0.0032836326863616705, |
|
"learning_rate": 9.197930142302717e-05, |
|
"loss": 0.0009, |
|
"step": 711 |
|
}, |
|
{ |
|
"epoch": 2.3014141414141416, |
|
"grad_norm": 0.0029353760182857513, |
|
"learning_rate": 9.210866752910737e-05, |
|
"loss": 0.0008, |
|
"step": 712 |
|
}, |
|
{ |
|
"epoch": 2.304646464646465, |
|
"grad_norm": 0.004789955448359251, |
|
"learning_rate": 9.223803363518758e-05, |
|
"loss": 0.0015, |
|
"step": 713 |
|
}, |
|
{ |
|
"epoch": 2.3078787878787876, |
|
"grad_norm": 0.00258410326205194, |
|
"learning_rate": 9.236739974126779e-05, |
|
"loss": 0.0007, |
|
"step": 714 |
|
}, |
|
{ |
|
"epoch": 2.311111111111111, |
|
"grad_norm": 0.003211255418136716, |
|
"learning_rate": 9.2496765847348e-05, |
|
"loss": 0.0009, |
|
"step": 715 |
|
}, |
|
{ |
|
"epoch": 2.314343434343434, |
|
"grad_norm": 0.003458111546933651, |
|
"learning_rate": 9.262613195342821e-05, |
|
"loss": 0.0008, |
|
"step": 716 |
|
}, |
|
{ |
|
"epoch": 2.3175757575757574, |
|
"grad_norm": 0.0020549860782921314, |
|
"learning_rate": 9.275549805950842e-05, |
|
"loss": 0.0005, |
|
"step": 717 |
|
}, |
|
{ |
|
"epoch": 2.3208080808080807, |
|
"grad_norm": 0.0034006584901362658, |
|
"learning_rate": 9.288486416558863e-05, |
|
"loss": 0.0011, |
|
"step": 718 |
|
}, |
|
{ |
|
"epoch": 2.324040404040404, |
|
"grad_norm": 0.003893098793923855, |
|
"learning_rate": 9.301423027166883e-05, |
|
"loss": 0.0008, |
|
"step": 719 |
|
}, |
|
{ |
|
"epoch": 2.327272727272727, |
|
"grad_norm": 0.004374553449451923, |
|
"learning_rate": 9.314359637774904e-05, |
|
"loss": 0.0014, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 2.3305050505050504, |
|
"grad_norm": 0.004447360523045063, |
|
"learning_rate": 9.327296248382925e-05, |
|
"loss": 0.0016, |
|
"step": 721 |
|
}, |
|
{ |
|
"epoch": 2.3337373737373737, |
|
"grad_norm": 0.003019913798198104, |
|
"learning_rate": 9.340232858990944e-05, |
|
"loss": 0.0007, |
|
"step": 722 |
|
}, |
|
{ |
|
"epoch": 2.336969696969697, |
|
"grad_norm": 0.00266702170483768, |
|
"learning_rate": 9.353169469598965e-05, |
|
"loss": 0.0006, |
|
"step": 723 |
|
}, |
|
{ |
|
"epoch": 2.34020202020202, |
|
"grad_norm": 0.004037540405988693, |
|
"learning_rate": 9.366106080206986e-05, |
|
"loss": 0.001, |
|
"step": 724 |
|
}, |
|
{ |
|
"epoch": 2.3434343434343434, |
|
"grad_norm": 0.003830693429335952, |
|
"learning_rate": 9.379042690815006e-05, |
|
"loss": 0.0015, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 2.3434343434343434, |
|
"eval_loss": 0.0009221473592333496, |
|
"eval_runtime": 18.682, |
|
"eval_samples_per_second": 5.353, |
|
"eval_steps_per_second": 1.338, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 2.3466666666666667, |
|
"grad_norm": 0.002714785747230053, |
|
"learning_rate": 9.391979301423027e-05, |
|
"loss": 0.0007, |
|
"step": 726 |
|
}, |
|
{ |
|
"epoch": 2.34989898989899, |
|
"grad_norm": 0.0034484020434319973, |
|
"learning_rate": 9.404915912031048e-05, |
|
"loss": 0.0021, |
|
"step": 727 |
|
}, |
|
{ |
|
"epoch": 2.353131313131313, |
|
"grad_norm": 0.0027550137601792812, |
|
"learning_rate": 9.417852522639069e-05, |
|
"loss": 0.0014, |
|
"step": 728 |
|
}, |
|
{ |
|
"epoch": 2.3563636363636364, |
|
"grad_norm": 0.004323527216911316, |
|
"learning_rate": 9.43078913324709e-05, |
|
"loss": 0.0021, |
|
"step": 729 |
|
}, |
|
{ |
|
"epoch": 2.3595959595959597, |
|
"grad_norm": 0.0036909414920955896, |
|
"learning_rate": 9.443725743855111e-05, |
|
"loss": 0.0009, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 2.362828282828283, |
|
"grad_norm": 0.0026773421559482813, |
|
"learning_rate": 9.456662354463132e-05, |
|
"loss": 0.0007, |
|
"step": 731 |
|
}, |
|
{ |
|
"epoch": 2.366060606060606, |
|
"grad_norm": 0.0027335931081324816, |
|
"learning_rate": 9.469598965071152e-05, |
|
"loss": 0.0011, |
|
"step": 732 |
|
}, |
|
{ |
|
"epoch": 2.3692929292929295, |
|
"grad_norm": 0.0037821868900209665, |
|
"learning_rate": 9.482535575679173e-05, |
|
"loss": 0.0009, |
|
"step": 733 |
|
}, |
|
{ |
|
"epoch": 2.3725252525252527, |
|
"grad_norm": 0.0027637695893645287, |
|
"learning_rate": 9.495472186287193e-05, |
|
"loss": 0.0008, |
|
"step": 734 |
|
}, |
|
{ |
|
"epoch": 2.375757575757576, |
|
"grad_norm": 0.0028088942635804415, |
|
"learning_rate": 9.508408796895214e-05, |
|
"loss": 0.0008, |
|
"step": 735 |
|
}, |
|
{ |
|
"epoch": 2.378989898989899, |
|
"grad_norm": 0.004683246370404959, |
|
"learning_rate": 9.521345407503235e-05, |
|
"loss": 0.0011, |
|
"step": 736 |
|
}, |
|
{ |
|
"epoch": 2.3822222222222225, |
|
"grad_norm": 0.0035344541538506746, |
|
"learning_rate": 9.534282018111256e-05, |
|
"loss": 0.001, |
|
"step": 737 |
|
}, |
|
{ |
|
"epoch": 2.3854545454545453, |
|
"grad_norm": 0.003326730104163289, |
|
"learning_rate": 9.547218628719275e-05, |
|
"loss": 0.001, |
|
"step": 738 |
|
}, |
|
{ |
|
"epoch": 2.3886868686868685, |
|
"grad_norm": 0.0027331975288689137, |
|
"learning_rate": 9.560155239327296e-05, |
|
"loss": 0.001, |
|
"step": 739 |
|
}, |
|
{ |
|
"epoch": 2.391919191919192, |
|
"grad_norm": 0.0023180183488875628, |
|
"learning_rate": 9.573091849935317e-05, |
|
"loss": 0.0005, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 2.395151515151515, |
|
"grad_norm": 0.00545561034232378, |
|
"learning_rate": 9.586028460543338e-05, |
|
"loss": 0.0019, |
|
"step": 741 |
|
}, |
|
{ |
|
"epoch": 2.3983838383838383, |
|
"grad_norm": 0.002278296509757638, |
|
"learning_rate": 9.59896507115136e-05, |
|
"loss": 0.0008, |
|
"step": 742 |
|
}, |
|
{ |
|
"epoch": 2.4016161616161615, |
|
"grad_norm": 0.0036029706243425608, |
|
"learning_rate": 9.61190168175938e-05, |
|
"loss": 0.0011, |
|
"step": 743 |
|
}, |
|
{ |
|
"epoch": 2.404848484848485, |
|
"grad_norm": 0.0028302932623773813, |
|
"learning_rate": 9.6248382923674e-05, |
|
"loss": 0.001, |
|
"step": 744 |
|
}, |
|
{ |
|
"epoch": 2.408080808080808, |
|
"grad_norm": 0.005623773206025362, |
|
"learning_rate": 9.63777490297542e-05, |
|
"loss": 0.0025, |
|
"step": 745 |
|
}, |
|
{ |
|
"epoch": 2.4113131313131313, |
|
"grad_norm": 0.003791616763919592, |
|
"learning_rate": 9.650711513583441e-05, |
|
"loss": 0.001, |
|
"step": 746 |
|
}, |
|
{ |
|
"epoch": 2.4145454545454546, |
|
"grad_norm": 0.003847538959234953, |
|
"learning_rate": 9.663648124191462e-05, |
|
"loss": 0.002, |
|
"step": 747 |
|
}, |
|
{ |
|
"epoch": 2.417777777777778, |
|
"grad_norm": 0.003144781803712249, |
|
"learning_rate": 9.676584734799483e-05, |
|
"loss": 0.0008, |
|
"step": 748 |
|
}, |
|
{ |
|
"epoch": 2.421010101010101, |
|
"grad_norm": 0.0030476360116153955, |
|
"learning_rate": 9.689521345407504e-05, |
|
"loss": 0.002, |
|
"step": 749 |
|
}, |
|
{ |
|
"epoch": 2.4242424242424243, |
|
"grad_norm": 0.023542512208223343, |
|
"learning_rate": 9.702457956015525e-05, |
|
"loss": 0.0012, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 2.4242424242424243, |
|
"eval_loss": 0.0009755383944138885, |
|
"eval_runtime": 18.7174, |
|
"eval_samples_per_second": 5.343, |
|
"eval_steps_per_second": 1.336, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 2.4274747474747476, |
|
"grad_norm": 0.005616334266960621, |
|
"learning_rate": 9.715394566623545e-05, |
|
"loss": 0.0009, |
|
"step": 751 |
|
}, |
|
{ |
|
"epoch": 2.430707070707071, |
|
"grad_norm": 0.007717492058873177, |
|
"learning_rate": 9.728331177231566e-05, |
|
"loss": 0.0008, |
|
"step": 752 |
|
}, |
|
{ |
|
"epoch": 2.433939393939394, |
|
"grad_norm": 0.008256220258772373, |
|
"learning_rate": 9.741267787839587e-05, |
|
"loss": 0.0011, |
|
"step": 753 |
|
}, |
|
{ |
|
"epoch": 2.4371717171717173, |
|
"grad_norm": 0.005972269922494888, |
|
"learning_rate": 9.754204398447608e-05, |
|
"loss": 0.0004, |
|
"step": 754 |
|
}, |
|
{ |
|
"epoch": 2.4404040404040406, |
|
"grad_norm": 0.006538939196616411, |
|
"learning_rate": 9.767141009055629e-05, |
|
"loss": 0.0009, |
|
"step": 755 |
|
}, |
|
{ |
|
"epoch": 2.443636363636364, |
|
"grad_norm": 0.003234976204112172, |
|
"learning_rate": 9.780077619663648e-05, |
|
"loss": 0.0011, |
|
"step": 756 |
|
}, |
|
{ |
|
"epoch": 2.4468686868686866, |
|
"grad_norm": 0.0033180294558405876, |
|
"learning_rate": 9.793014230271668e-05, |
|
"loss": 0.0009, |
|
"step": 757 |
|
}, |
|
{ |
|
"epoch": 2.45010101010101, |
|
"grad_norm": 0.0037402757443487644, |
|
"learning_rate": 9.805950840879689e-05, |
|
"loss": 0.0007, |
|
"step": 758 |
|
}, |
|
{ |
|
"epoch": 2.453333333333333, |
|
"grad_norm": 0.0028039240278303623, |
|
"learning_rate": 9.81888745148771e-05, |
|
"loss": 0.0008, |
|
"step": 759 |
|
}, |
|
{ |
|
"epoch": 2.4565656565656564, |
|
"grad_norm": 0.0029848841950297356, |
|
"learning_rate": 9.831824062095731e-05, |
|
"loss": 0.0007, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 2.4597979797979797, |
|
"grad_norm": 0.002809871220961213, |
|
"learning_rate": 9.844760672703752e-05, |
|
"loss": 0.0007, |
|
"step": 761 |
|
}, |
|
{ |
|
"epoch": 2.463030303030303, |
|
"grad_norm": 0.004293316043913364, |
|
"learning_rate": 9.857697283311773e-05, |
|
"loss": 0.0012, |
|
"step": 762 |
|
}, |
|
{ |
|
"epoch": 2.466262626262626, |
|
"grad_norm": 0.003171822987496853, |
|
"learning_rate": 9.870633893919794e-05, |
|
"loss": 0.0008, |
|
"step": 763 |
|
}, |
|
{ |
|
"epoch": 2.4694949494949494, |
|
"grad_norm": 0.0035177527461200953, |
|
"learning_rate": 9.883570504527814e-05, |
|
"loss": 0.0008, |
|
"step": 764 |
|
}, |
|
{ |
|
"epoch": 2.4727272727272727, |
|
"grad_norm": 0.0037470462266355753, |
|
"learning_rate": 9.896507115135835e-05, |
|
"loss": 0.0011, |
|
"step": 765 |
|
}, |
|
{ |
|
"epoch": 2.475959595959596, |
|
"grad_norm": 0.004131955560296774, |
|
"learning_rate": 9.909443725743856e-05, |
|
"loss": 0.0015, |
|
"step": 766 |
|
}, |
|
{ |
|
"epoch": 2.479191919191919, |
|
"grad_norm": 0.002823730930685997, |
|
"learning_rate": 9.922380336351877e-05, |
|
"loss": 0.0007, |
|
"step": 767 |
|
}, |
|
{ |
|
"epoch": 2.4824242424242424, |
|
"grad_norm": 0.0037240665405988693, |
|
"learning_rate": 9.935316946959897e-05, |
|
"loss": 0.001, |
|
"step": 768 |
|
}, |
|
{ |
|
"epoch": 2.4856565656565657, |
|
"grad_norm": 0.0038744148332625628, |
|
"learning_rate": 9.948253557567918e-05, |
|
"loss": 0.0019, |
|
"step": 769 |
|
}, |
|
{ |
|
"epoch": 2.488888888888889, |
|
"grad_norm": 0.003629435319453478, |
|
"learning_rate": 9.961190168175937e-05, |
|
"loss": 0.0008, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 2.492121212121212, |
|
"grad_norm": 0.004831044003367424, |
|
"learning_rate": 9.974126778783958e-05, |
|
"loss": 0.0011, |
|
"step": 771 |
|
}, |
|
{ |
|
"epoch": 2.4953535353535354, |
|
"grad_norm": 0.003547330852597952, |
|
"learning_rate": 9.98706338939198e-05, |
|
"loss": 0.0009, |
|
"step": 772 |
|
}, |
|
{ |
|
"epoch": 2.4985858585858587, |
|
"grad_norm": 0.003263116115704179, |
|
"learning_rate": 0.0001, |
|
"loss": 0.0008, |
|
"step": 773 |
|
}, |
|
{ |
|
"epoch": 2.501818181818182, |
|
"grad_norm": 0.0030386094003915787, |
|
"learning_rate": 9.999999489471233e-05, |
|
"loss": 0.0007, |
|
"step": 774 |
|
}, |
|
{ |
|
"epoch": 2.505050505050505, |
|
"grad_norm": 0.0036390256136655807, |
|
"learning_rate": 9.99999795788503e-05, |
|
"loss": 0.0012, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 2.505050505050505, |
|
"eval_loss": 0.000891694042365998, |
|
"eval_runtime": 18.7365, |
|
"eval_samples_per_second": 5.337, |
|
"eval_steps_per_second": 1.334, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 2.5082828282828284, |
|
"grad_norm": 0.0027272645384073257, |
|
"learning_rate": 9.99999540524171e-05, |
|
"loss": 0.0006, |
|
"step": 776 |
|
}, |
|
{ |
|
"epoch": 2.5115151515151517, |
|
"grad_norm": 0.004212843254208565, |
|
"learning_rate": 9.999991831541789e-05, |
|
"loss": 0.0023, |
|
"step": 777 |
|
}, |
|
{ |
|
"epoch": 2.514747474747475, |
|
"grad_norm": 0.0034173503518104553, |
|
"learning_rate": 9.999987236786e-05, |
|
"loss": 0.0008, |
|
"step": 778 |
|
}, |
|
{ |
|
"epoch": 2.517979797979798, |
|
"grad_norm": 0.0028145266696810722, |
|
"learning_rate": 9.999981620975281e-05, |
|
"loss": 0.0007, |
|
"step": 779 |
|
}, |
|
{ |
|
"epoch": 2.5212121212121215, |
|
"grad_norm": 0.0030887024477124214, |
|
"learning_rate": 9.999974984110779e-05, |
|
"loss": 0.0009, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 2.5244444444444447, |
|
"grad_norm": 0.003968552686274052, |
|
"learning_rate": 9.999967326193847e-05, |
|
"loss": 0.0009, |
|
"step": 781 |
|
}, |
|
{ |
|
"epoch": 2.5276767676767675, |
|
"grad_norm": 0.00264795683324337, |
|
"learning_rate": 9.999958647226049e-05, |
|
"loss": 0.0006, |
|
"step": 782 |
|
}, |
|
{ |
|
"epoch": 2.5309090909090908, |
|
"grad_norm": 0.004550054203718901, |
|
"learning_rate": 9.999948947209162e-05, |
|
"loss": 0.002, |
|
"step": 783 |
|
}, |
|
{ |
|
"epoch": 2.534141414141414, |
|
"grad_norm": 0.0030430385377258062, |
|
"learning_rate": 9.999938226145161e-05, |
|
"loss": 0.0007, |
|
"step": 784 |
|
}, |
|
{ |
|
"epoch": 2.5373737373737373, |
|
"grad_norm": 0.003766052657738328, |
|
"learning_rate": 9.999926484036237e-05, |
|
"loss": 0.001, |
|
"step": 785 |
|
}, |
|
{ |
|
"epoch": 2.5406060606060605, |
|
"grad_norm": 0.003455114783719182, |
|
"learning_rate": 9.999913720884791e-05, |
|
"loss": 0.002, |
|
"step": 786 |
|
}, |
|
{ |
|
"epoch": 2.543838383838384, |
|
"grad_norm": 0.002563537796959281, |
|
"learning_rate": 9.999899936693426e-05, |
|
"loss": 0.0008, |
|
"step": 787 |
|
}, |
|
{ |
|
"epoch": 2.547070707070707, |
|
"grad_norm": 0.0025122894439846277, |
|
"learning_rate": 9.99988513146496e-05, |
|
"loss": 0.0006, |
|
"step": 788 |
|
}, |
|
{ |
|
"epoch": 2.5503030303030303, |
|
"grad_norm": 0.004589984659105539, |
|
"learning_rate": 9.999869305202412e-05, |
|
"loss": 0.0015, |
|
"step": 789 |
|
}, |
|
{ |
|
"epoch": 2.5535353535353535, |
|
"grad_norm": 0.0041996450163424015, |
|
"learning_rate": 9.999852457909018e-05, |
|
"loss": 0.0019, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 2.556767676767677, |
|
"grad_norm": 0.0033383311238139868, |
|
"learning_rate": 9.999834589588217e-05, |
|
"loss": 0.0013, |
|
"step": 791 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"grad_norm": 0.0018662047805264592, |
|
"learning_rate": 9.999815700243656e-05, |
|
"loss": 0.0006, |
|
"step": 792 |
|
}, |
|
{ |
|
"epoch": 2.5632323232323233, |
|
"grad_norm": 0.0032810529228299856, |
|
"learning_rate": 9.999795789879196e-05, |
|
"loss": 0.0009, |
|
"step": 793 |
|
}, |
|
{ |
|
"epoch": 2.5664646464646466, |
|
"grad_norm": 0.003441553795710206, |
|
"learning_rate": 9.9997748584989e-05, |
|
"loss": 0.0013, |
|
"step": 794 |
|
}, |
|
{ |
|
"epoch": 2.56969696969697, |
|
"grad_norm": 0.0025200124364346266, |
|
"learning_rate": 9.999752906107042e-05, |
|
"loss": 0.0008, |
|
"step": 795 |
|
}, |
|
{ |
|
"epoch": 2.572929292929293, |
|
"grad_norm": 0.0030850358307361603, |
|
"learning_rate": 9.999729932708109e-05, |
|
"loss": 0.0009, |
|
"step": 796 |
|
}, |
|
{ |
|
"epoch": 2.5761616161616163, |
|
"grad_norm": 0.0029367755632847548, |
|
"learning_rate": 9.999705938306789e-05, |
|
"loss": 0.0007, |
|
"step": 797 |
|
}, |
|
{ |
|
"epoch": 2.579393939393939, |
|
"grad_norm": 0.003588201245293021, |
|
"learning_rate": 9.999680922907982e-05, |
|
"loss": 0.0026, |
|
"step": 798 |
|
}, |
|
{ |
|
"epoch": 2.5826262626262624, |
|
"grad_norm": 0.0031253646593540907, |
|
"learning_rate": 9.999654886516798e-05, |
|
"loss": 0.0009, |
|
"step": 799 |
|
}, |
|
{ |
|
"epoch": 2.5858585858585856, |
|
"grad_norm": 0.0036176196299493313, |
|
"learning_rate": 9.999627829138554e-05, |
|
"loss": 0.0015, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 2.5858585858585856, |
|
"eval_loss": 0.0009534513228572905, |
|
"eval_runtime": 18.6223, |
|
"eval_samples_per_second": 5.37, |
|
"eval_steps_per_second": 1.342, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 2.589090909090909, |
|
"grad_norm": 0.00325636169873178, |
|
"learning_rate": 9.999599750778772e-05, |
|
"loss": 0.0007, |
|
"step": 801 |
|
}, |
|
{ |
|
"epoch": 2.592323232323232, |
|
"grad_norm": 0.003968099132180214, |
|
"learning_rate": 9.999570651443191e-05, |
|
"loss": 0.0011, |
|
"step": 802 |
|
}, |
|
{ |
|
"epoch": 2.5955555555555554, |
|
"grad_norm": 0.00244711060076952, |
|
"learning_rate": 9.99954053113775e-05, |
|
"loss": 0.0006, |
|
"step": 803 |
|
}, |
|
{ |
|
"epoch": 2.5987878787878786, |
|
"grad_norm": 0.003342804964631796, |
|
"learning_rate": 9.9995093898686e-05, |
|
"loss": 0.0014, |
|
"step": 804 |
|
}, |
|
{ |
|
"epoch": 2.602020202020202, |
|
"grad_norm": 0.006182527635246515, |
|
"learning_rate": 9.999477227642103e-05, |
|
"loss": 0.0025, |
|
"step": 805 |
|
}, |
|
{ |
|
"epoch": 2.605252525252525, |
|
"grad_norm": 0.0032223479356616735, |
|
"learning_rate": 9.999444044464823e-05, |
|
"loss": 0.0007, |
|
"step": 806 |
|
}, |
|
{ |
|
"epoch": 2.6084848484848484, |
|
"grad_norm": 0.002208409830927849, |
|
"learning_rate": 9.999409840343539e-05, |
|
"loss": 0.0007, |
|
"step": 807 |
|
}, |
|
{ |
|
"epoch": 2.6117171717171717, |
|
"grad_norm": 0.002247196389362216, |
|
"learning_rate": 9.999374615285236e-05, |
|
"loss": 0.0007, |
|
"step": 808 |
|
}, |
|
{ |
|
"epoch": 2.614949494949495, |
|
"grad_norm": 0.0024588643573224545, |
|
"learning_rate": 9.999338369297106e-05, |
|
"loss": 0.0005, |
|
"step": 809 |
|
}, |
|
{ |
|
"epoch": 2.618181818181818, |
|
"grad_norm": 0.003213444259017706, |
|
"learning_rate": 9.999301102386553e-05, |
|
"loss": 0.0008, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 2.6214141414141414, |
|
"grad_norm": 0.0028623330872505903, |
|
"learning_rate": 9.999262814561185e-05, |
|
"loss": 0.0009, |
|
"step": 811 |
|
}, |
|
{ |
|
"epoch": 2.6246464646464647, |
|
"grad_norm": 0.0031820102594792843, |
|
"learning_rate": 9.999223505828821e-05, |
|
"loss": 0.001, |
|
"step": 812 |
|
}, |
|
{ |
|
"epoch": 2.627878787878788, |
|
"grad_norm": 0.0017483988776803017, |
|
"learning_rate": 9.999183176197491e-05, |
|
"loss": 0.0006, |
|
"step": 813 |
|
}, |
|
{ |
|
"epoch": 2.631111111111111, |
|
"grad_norm": 0.0020935633219778538, |
|
"learning_rate": 9.999141825675426e-05, |
|
"loss": 0.0006, |
|
"step": 814 |
|
}, |
|
{ |
|
"epoch": 2.6343434343434344, |
|
"grad_norm": 0.003117109416052699, |
|
"learning_rate": 9.999099454271074e-05, |
|
"loss": 0.0008, |
|
"step": 815 |
|
}, |
|
{ |
|
"epoch": 2.6375757575757577, |
|
"grad_norm": 0.0037443467881530523, |
|
"learning_rate": 9.999056061993089e-05, |
|
"loss": 0.001, |
|
"step": 816 |
|
}, |
|
{ |
|
"epoch": 2.640808080808081, |
|
"grad_norm": 0.0032844438683241606, |
|
"learning_rate": 9.999011648850329e-05, |
|
"loss": 0.0009, |
|
"step": 817 |
|
}, |
|
{ |
|
"epoch": 2.644040404040404, |
|
"grad_norm": 0.0027299323119223118, |
|
"learning_rate": 9.998966214851864e-05, |
|
"loss": 0.001, |
|
"step": 818 |
|
}, |
|
{ |
|
"epoch": 2.6472727272727274, |
|
"grad_norm": 0.003157002152875066, |
|
"learning_rate": 9.998919760006972e-05, |
|
"loss": 0.001, |
|
"step": 819 |
|
}, |
|
{ |
|
"epoch": 2.6505050505050507, |
|
"grad_norm": 0.003021983429789543, |
|
"learning_rate": 9.998872284325142e-05, |
|
"loss": 0.0011, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 2.653737373737374, |
|
"grad_norm": 0.0036499700509011745, |
|
"learning_rate": 9.998823787816066e-05, |
|
"loss": 0.001, |
|
"step": 821 |
|
}, |
|
{ |
|
"epoch": 2.656969696969697, |
|
"grad_norm": 0.0031615635380148888, |
|
"learning_rate": 9.99877427048965e-05, |
|
"loss": 0.0009, |
|
"step": 822 |
|
}, |
|
{ |
|
"epoch": 2.6602020202020205, |
|
"grad_norm": 0.0032300378661602736, |
|
"learning_rate": 9.998723732356006e-05, |
|
"loss": 0.0009, |
|
"step": 823 |
|
}, |
|
{ |
|
"epoch": 2.6634343434343437, |
|
"grad_norm": 0.00247188750654459, |
|
"learning_rate": 9.998672173425452e-05, |
|
"loss": 0.0008, |
|
"step": 824 |
|
}, |
|
{ |
|
"epoch": 2.6666666666666665, |
|
"grad_norm": 0.004531141836196184, |
|
"learning_rate": 9.998619593708518e-05, |
|
"loss": 0.0011, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 2.6666666666666665, |
|
"eval_loss": 0.0008804297540336847, |
|
"eval_runtime": 18.6211, |
|
"eval_samples_per_second": 5.37, |
|
"eval_steps_per_second": 1.343, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 2.6698989898989898, |
|
"grad_norm": 0.0034359728451818228, |
|
"learning_rate": 9.998565993215943e-05, |
|
"loss": 0.0009, |
|
"step": 826 |
|
}, |
|
{ |
|
"epoch": 2.673131313131313, |
|
"grad_norm": 0.002785489894449711, |
|
"learning_rate": 9.998511371958672e-05, |
|
"loss": 0.0007, |
|
"step": 827 |
|
}, |
|
{ |
|
"epoch": 2.6763636363636363, |
|
"grad_norm": 0.0025934847071766853, |
|
"learning_rate": 9.998455729947858e-05, |
|
"loss": 0.0008, |
|
"step": 828 |
|
}, |
|
{ |
|
"epoch": 2.6795959595959595, |
|
"grad_norm": 0.0026367492973804474, |
|
"learning_rate": 9.998399067194864e-05, |
|
"loss": 0.0008, |
|
"step": 829 |
|
}, |
|
{ |
|
"epoch": 2.682828282828283, |
|
"grad_norm": 0.003330858191475272, |
|
"learning_rate": 9.998341383711263e-05, |
|
"loss": 0.0007, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 2.686060606060606, |
|
"grad_norm": 0.003155591432005167, |
|
"learning_rate": 9.998282679508835e-05, |
|
"loss": 0.0017, |
|
"step": 831 |
|
}, |
|
{ |
|
"epoch": 2.6892929292929293, |
|
"grad_norm": 0.0030347639694809914, |
|
"learning_rate": 9.998222954599563e-05, |
|
"loss": 0.0007, |
|
"step": 832 |
|
}, |
|
{ |
|
"epoch": 2.6925252525252525, |
|
"grad_norm": 0.0032036558259278536, |
|
"learning_rate": 9.99816220899565e-05, |
|
"loss": 0.0009, |
|
"step": 833 |
|
}, |
|
{ |
|
"epoch": 2.695757575757576, |
|
"grad_norm": 0.003590774955227971, |
|
"learning_rate": 9.998100442709497e-05, |
|
"loss": 0.0009, |
|
"step": 834 |
|
}, |
|
{ |
|
"epoch": 2.698989898989899, |
|
"grad_norm": 0.002222040668129921, |
|
"learning_rate": 9.998037655753717e-05, |
|
"loss": 0.0009, |
|
"step": 835 |
|
}, |
|
{ |
|
"epoch": 2.7022222222222223, |
|
"grad_norm": 0.0027599751483649015, |
|
"learning_rate": 9.997973848141137e-05, |
|
"loss": 0.0009, |
|
"step": 836 |
|
}, |
|
{ |
|
"epoch": 2.7054545454545456, |
|
"grad_norm": 0.00284950970672071, |
|
"learning_rate": 9.997909019884781e-05, |
|
"loss": 0.0008, |
|
"step": 837 |
|
}, |
|
{ |
|
"epoch": 2.708686868686869, |
|
"grad_norm": 0.0024073810782283545, |
|
"learning_rate": 9.99784317099789e-05, |
|
"loss": 0.001, |
|
"step": 838 |
|
}, |
|
{ |
|
"epoch": 2.711919191919192, |
|
"grad_norm": 0.0033184473868459463, |
|
"learning_rate": 9.997776301493914e-05, |
|
"loss": 0.0016, |
|
"step": 839 |
|
}, |
|
{ |
|
"epoch": 2.7151515151515153, |
|
"grad_norm": 0.0023880742955952883, |
|
"learning_rate": 9.997708411386501e-05, |
|
"loss": 0.0015, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 2.718383838383838, |
|
"grad_norm": 0.0031732122879475355, |
|
"learning_rate": 9.997639500689523e-05, |
|
"loss": 0.0007, |
|
"step": 841 |
|
}, |
|
{ |
|
"epoch": 2.7216161616161614, |
|
"grad_norm": 0.0036892895586788654, |
|
"learning_rate": 9.997569569417049e-05, |
|
"loss": 0.002, |
|
"step": 842 |
|
}, |
|
{ |
|
"epoch": 2.7248484848484846, |
|
"grad_norm": 0.002280700486153364, |
|
"learning_rate": 9.997498617583358e-05, |
|
"loss": 0.0018, |
|
"step": 843 |
|
}, |
|
{ |
|
"epoch": 2.728080808080808, |
|
"grad_norm": 0.003923532087355852, |
|
"learning_rate": 9.997426645202943e-05, |
|
"loss": 0.0006, |
|
"step": 844 |
|
}, |
|
{ |
|
"epoch": 2.731313131313131, |
|
"grad_norm": 0.0025437732692807913, |
|
"learning_rate": 9.9973536522905e-05, |
|
"loss": 0.0009, |
|
"step": 845 |
|
}, |
|
{ |
|
"epoch": 2.7345454545454544, |
|
"grad_norm": 0.002573117846623063, |
|
"learning_rate": 9.997279638860933e-05, |
|
"loss": 0.0005, |
|
"step": 846 |
|
}, |
|
{ |
|
"epoch": 2.7377777777777776, |
|
"grad_norm": 0.0021492561791092157, |
|
"learning_rate": 9.99720460492936e-05, |
|
"loss": 0.0008, |
|
"step": 847 |
|
}, |
|
{ |
|
"epoch": 2.741010101010101, |
|
"grad_norm": 0.0021558962762355804, |
|
"learning_rate": 9.997128550511099e-05, |
|
"loss": 0.0006, |
|
"step": 848 |
|
}, |
|
{ |
|
"epoch": 2.744242424242424, |
|
"grad_norm": 0.0031675281934440136, |
|
"learning_rate": 9.997051475621687e-05, |
|
"loss": 0.001, |
|
"step": 849 |
|
}, |
|
{ |
|
"epoch": 2.7474747474747474, |
|
"grad_norm": 0.003528911853209138, |
|
"learning_rate": 9.996973380276857e-05, |
|
"loss": 0.0007, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 2.7474747474747474, |
|
"eval_loss": 0.000889240182004869, |
|
"eval_runtime": 18.6302, |
|
"eval_samples_per_second": 5.368, |
|
"eval_steps_per_second": 1.342, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 2.7507070707070707, |
|
"grad_norm": 0.003921836614608765, |
|
"learning_rate": 9.996894264492563e-05, |
|
"loss": 0.001, |
|
"step": 851 |
|
}, |
|
{ |
|
"epoch": 2.753939393939394, |
|
"grad_norm": 0.002436129143461585, |
|
"learning_rate": 9.99681412828496e-05, |
|
"loss": 0.0007, |
|
"step": 852 |
|
}, |
|
{ |
|
"epoch": 2.757171717171717, |
|
"grad_norm": 0.002964465180411935, |
|
"learning_rate": 9.996732971670408e-05, |
|
"loss": 0.0008, |
|
"step": 853 |
|
}, |
|
{ |
|
"epoch": 2.7604040404040404, |
|
"grad_norm": 0.0034572421573102474, |
|
"learning_rate": 9.996650794665487e-05, |
|
"loss": 0.0007, |
|
"step": 854 |
|
}, |
|
{ |
|
"epoch": 2.7636363636363637, |
|
"grad_norm": 0.0021866310853511095, |
|
"learning_rate": 9.996567597286974e-05, |
|
"loss": 0.0005, |
|
"step": 855 |
|
}, |
|
{ |
|
"epoch": 2.766868686868687, |
|
"grad_norm": 0.0026556740049272776, |
|
"learning_rate": 9.996483379551861e-05, |
|
"loss": 0.0014, |
|
"step": 856 |
|
}, |
|
{ |
|
"epoch": 2.77010101010101, |
|
"grad_norm": 0.003023721743375063, |
|
"learning_rate": 9.996398141477344e-05, |
|
"loss": 0.0007, |
|
"step": 857 |
|
}, |
|
{ |
|
"epoch": 2.7733333333333334, |
|
"grad_norm": 0.0022941052448004484, |
|
"learning_rate": 9.996311883080832e-05, |
|
"loss": 0.0006, |
|
"step": 858 |
|
}, |
|
{ |
|
"epoch": 2.7765656565656567, |
|
"grad_norm": 0.003141100751236081, |
|
"learning_rate": 9.996224604379938e-05, |
|
"loss": 0.0007, |
|
"step": 859 |
|
}, |
|
{ |
|
"epoch": 2.77979797979798, |
|
"grad_norm": 0.002024593763053417, |
|
"learning_rate": 9.996136305392487e-05, |
|
"loss": 0.0006, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 2.783030303030303, |
|
"grad_norm": 0.004872876685112715, |
|
"learning_rate": 9.996046986136509e-05, |
|
"loss": 0.0026, |
|
"step": 861 |
|
}, |
|
{ |
|
"epoch": 2.7862626262626264, |
|
"grad_norm": 0.002601329004392028, |
|
"learning_rate": 9.995956646630246e-05, |
|
"loss": 0.0009, |
|
"step": 862 |
|
}, |
|
{ |
|
"epoch": 2.7894949494949497, |
|
"grad_norm": 0.0028132593724876642, |
|
"learning_rate": 9.995865286892145e-05, |
|
"loss": 0.0009, |
|
"step": 863 |
|
}, |
|
{ |
|
"epoch": 2.792727272727273, |
|
"grad_norm": 0.002662686165422201, |
|
"learning_rate": 9.995772906940864e-05, |
|
"loss": 0.0007, |
|
"step": 864 |
|
}, |
|
{ |
|
"epoch": 2.795959595959596, |
|
"grad_norm": 0.0027195930015295744, |
|
"learning_rate": 9.995679506795264e-05, |
|
"loss": 0.0007, |
|
"step": 865 |
|
}, |
|
{ |
|
"epoch": 2.7991919191919195, |
|
"grad_norm": 0.003179864026606083, |
|
"learning_rate": 9.995585086474424e-05, |
|
"loss": 0.0014, |
|
"step": 866 |
|
}, |
|
{ |
|
"epoch": 2.8024242424242423, |
|
"grad_norm": 0.0034151843283325434, |
|
"learning_rate": 9.995489645997622e-05, |
|
"loss": 0.0009, |
|
"step": 867 |
|
}, |
|
{ |
|
"epoch": 2.8056565656565655, |
|
"grad_norm": 0.0026823675725609064, |
|
"learning_rate": 9.99539318538435e-05, |
|
"loss": 0.0007, |
|
"step": 868 |
|
}, |
|
{ |
|
"epoch": 2.8088888888888888, |
|
"grad_norm": 0.0021664374507963657, |
|
"learning_rate": 9.995295704654304e-05, |
|
"loss": 0.0012, |
|
"step": 869 |
|
}, |
|
{ |
|
"epoch": 2.812121212121212, |
|
"grad_norm": 0.003440326079726219, |
|
"learning_rate": 9.995197203827393e-05, |
|
"loss": 0.0009, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 2.8153535353535353, |
|
"grad_norm": 0.003980646841228008, |
|
"learning_rate": 9.995097682923733e-05, |
|
"loss": 0.0018, |
|
"step": 871 |
|
}, |
|
{ |
|
"epoch": 2.8185858585858585, |
|
"grad_norm": 0.003389423480257392, |
|
"learning_rate": 9.994997141963644e-05, |
|
"loss": 0.001, |
|
"step": 872 |
|
}, |
|
{ |
|
"epoch": 2.821818181818182, |
|
"grad_norm": 0.0027842579875141382, |
|
"learning_rate": 9.994895580967658e-05, |
|
"loss": 0.0007, |
|
"step": 873 |
|
}, |
|
{ |
|
"epoch": 2.825050505050505, |
|
"grad_norm": 0.0033308009151369333, |
|
"learning_rate": 9.994792999956518e-05, |
|
"loss": 0.0011, |
|
"step": 874 |
|
}, |
|
{ |
|
"epoch": 2.8282828282828283, |
|
"grad_norm": 0.002514815656468272, |
|
"learning_rate": 9.994689398951169e-05, |
|
"loss": 0.0009, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 2.8282828282828283, |
|
"eval_loss": 0.0008539481204934418, |
|
"eval_runtime": 18.6059, |
|
"eval_samples_per_second": 5.375, |
|
"eval_steps_per_second": 1.344, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 2.8315151515151515, |
|
"grad_norm": 0.0024293591268360615, |
|
"learning_rate": 9.994584777972769e-05, |
|
"loss": 0.0008, |
|
"step": 876 |
|
}, |
|
{ |
|
"epoch": 2.834747474747475, |
|
"grad_norm": 0.0035905223339796066, |
|
"learning_rate": 9.994479137042683e-05, |
|
"loss": 0.002, |
|
"step": 877 |
|
}, |
|
{ |
|
"epoch": 2.837979797979798, |
|
"grad_norm": 0.002918825950473547, |
|
"learning_rate": 9.994372476182484e-05, |
|
"loss": 0.0009, |
|
"step": 878 |
|
}, |
|
{ |
|
"epoch": 2.8412121212121213, |
|
"grad_norm": 0.0031527606770396233, |
|
"learning_rate": 9.994264795413953e-05, |
|
"loss": 0.0008, |
|
"step": 879 |
|
}, |
|
{ |
|
"epoch": 2.8444444444444446, |
|
"grad_norm": 0.0035410267300903797, |
|
"learning_rate": 9.99415609475908e-05, |
|
"loss": 0.0021, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 2.847676767676768, |
|
"grad_norm": 0.0020397889893501997, |
|
"learning_rate": 9.994046374240062e-05, |
|
"loss": 0.0007, |
|
"step": 881 |
|
}, |
|
{ |
|
"epoch": 2.850909090909091, |
|
"grad_norm": 0.002342833438888192, |
|
"learning_rate": 9.993935633879306e-05, |
|
"loss": 0.0007, |
|
"step": 882 |
|
}, |
|
{ |
|
"epoch": 2.854141414141414, |
|
"grad_norm": 0.003542246064171195, |
|
"learning_rate": 9.993823873699426e-05, |
|
"loss": 0.0008, |
|
"step": 883 |
|
}, |
|
{ |
|
"epoch": 2.857373737373737, |
|
"grad_norm": 0.0024878752883523703, |
|
"learning_rate": 9.993711093723245e-05, |
|
"loss": 0.0008, |
|
"step": 884 |
|
}, |
|
{ |
|
"epoch": 2.8606060606060604, |
|
"grad_norm": 0.0030815114732831717, |
|
"learning_rate": 9.993597293973796e-05, |
|
"loss": 0.0018, |
|
"step": 885 |
|
}, |
|
{ |
|
"epoch": 2.8638383838383836, |
|
"grad_norm": 0.004479612223803997, |
|
"learning_rate": 9.993482474474314e-05, |
|
"loss": 0.0009, |
|
"step": 886 |
|
}, |
|
{ |
|
"epoch": 2.867070707070707, |
|
"grad_norm": 0.0033372335601598024, |
|
"learning_rate": 9.99336663524825e-05, |
|
"loss": 0.0009, |
|
"step": 887 |
|
}, |
|
{ |
|
"epoch": 2.87030303030303, |
|
"grad_norm": 0.0021939794532954693, |
|
"learning_rate": 9.993249776319258e-05, |
|
"loss": 0.0006, |
|
"step": 888 |
|
}, |
|
{ |
|
"epoch": 2.8735353535353534, |
|
"grad_norm": 0.0028632464818656445, |
|
"learning_rate": 9.993131897711202e-05, |
|
"loss": 0.0009, |
|
"step": 889 |
|
}, |
|
{ |
|
"epoch": 2.8767676767676766, |
|
"grad_norm": 0.002615751465782523, |
|
"learning_rate": 9.993012999448154e-05, |
|
"loss": 0.0009, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"grad_norm": 0.0017363366205245256, |
|
"learning_rate": 9.992893081554397e-05, |
|
"loss": 0.0005, |
|
"step": 891 |
|
}, |
|
{ |
|
"epoch": 2.883232323232323, |
|
"grad_norm": 0.0032067778520286083, |
|
"learning_rate": 9.992772144054415e-05, |
|
"loss": 0.0009, |
|
"step": 892 |
|
}, |
|
{ |
|
"epoch": 2.8864646464646464, |
|
"grad_norm": 0.0025459511671215296, |
|
"learning_rate": 9.992650186972909e-05, |
|
"loss": 0.0008, |
|
"step": 893 |
|
}, |
|
{ |
|
"epoch": 2.8896969696969697, |
|
"grad_norm": 0.001871303771622479, |
|
"learning_rate": 9.99252721033478e-05, |
|
"loss": 0.0008, |
|
"step": 894 |
|
}, |
|
{ |
|
"epoch": 2.892929292929293, |
|
"grad_norm": 0.001998218474909663, |
|
"learning_rate": 9.992403214165147e-05, |
|
"loss": 0.0005, |
|
"step": 895 |
|
}, |
|
{ |
|
"epoch": 2.896161616161616, |
|
"grad_norm": 0.0023115125950425863, |
|
"learning_rate": 9.992278198489327e-05, |
|
"loss": 0.0007, |
|
"step": 896 |
|
}, |
|
{ |
|
"epoch": 2.8993939393939394, |
|
"grad_norm": 0.0019588919822126627, |
|
"learning_rate": 9.99215216333285e-05, |
|
"loss": 0.0007, |
|
"step": 897 |
|
}, |
|
{ |
|
"epoch": 2.9026262626262627, |
|
"grad_norm": 0.001903701457194984, |
|
"learning_rate": 9.992025108721454e-05, |
|
"loss": 0.0004, |
|
"step": 898 |
|
}, |
|
{ |
|
"epoch": 2.905858585858586, |
|
"grad_norm": 0.0023024296388030052, |
|
"learning_rate": 9.991897034681087e-05, |
|
"loss": 0.0006, |
|
"step": 899 |
|
}, |
|
{ |
|
"epoch": 2.909090909090909, |
|
"grad_norm": 0.002626319881528616, |
|
"learning_rate": 9.9917679412379e-05, |
|
"loss": 0.0009, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 2.909090909090909, |
|
"eval_loss": 0.0008496911614201963, |
|
"eval_runtime": 18.6252, |
|
"eval_samples_per_second": 5.369, |
|
"eval_steps_per_second": 1.342, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 2.9123232323232324, |
|
"grad_norm": 0.0029552706982940435, |
|
"learning_rate": 9.99163782841826e-05, |
|
"loss": 0.001, |
|
"step": 901 |
|
}, |
|
{ |
|
"epoch": 2.9155555555555557, |
|
"grad_norm": 0.002366641303524375, |
|
"learning_rate": 9.991506696248731e-05, |
|
"loss": 0.0006, |
|
"step": 902 |
|
}, |
|
{ |
|
"epoch": 2.918787878787879, |
|
"grad_norm": 0.0030002701096236706, |
|
"learning_rate": 9.991374544756098e-05, |
|
"loss": 0.0009, |
|
"step": 903 |
|
}, |
|
{ |
|
"epoch": 2.922020202020202, |
|
"grad_norm": 0.002418822841718793, |
|
"learning_rate": 9.991241373967344e-05, |
|
"loss": 0.0005, |
|
"step": 904 |
|
}, |
|
{ |
|
"epoch": 2.9252525252525254, |
|
"grad_norm": 0.0027566729113459587, |
|
"learning_rate": 9.991107183909664e-05, |
|
"loss": 0.002, |
|
"step": 905 |
|
}, |
|
{ |
|
"epoch": 2.9284848484848487, |
|
"grad_norm": 0.003200074890628457, |
|
"learning_rate": 9.990971974610466e-05, |
|
"loss": 0.0009, |
|
"step": 906 |
|
}, |
|
{ |
|
"epoch": 2.931717171717172, |
|
"grad_norm": 0.00368666322901845, |
|
"learning_rate": 9.990835746097356e-05, |
|
"loss": 0.0008, |
|
"step": 907 |
|
}, |
|
{ |
|
"epoch": 2.934949494949495, |
|
"grad_norm": 0.0036671683192253113, |
|
"learning_rate": 9.990698498398155e-05, |
|
"loss": 0.001, |
|
"step": 908 |
|
}, |
|
{ |
|
"epoch": 2.9381818181818184, |
|
"grad_norm": 0.0037938845343887806, |
|
"learning_rate": 9.990560231540889e-05, |
|
"loss": 0.0007, |
|
"step": 909 |
|
}, |
|
{ |
|
"epoch": 2.9414141414141413, |
|
"grad_norm": 0.0027589588426053524, |
|
"learning_rate": 9.990420945553797e-05, |
|
"loss": 0.0006, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 2.9446464646464645, |
|
"grad_norm": 0.003152150195091963, |
|
"learning_rate": 9.990280640465321e-05, |
|
"loss": 0.0008, |
|
"step": 911 |
|
}, |
|
{ |
|
"epoch": 2.9478787878787878, |
|
"grad_norm": 0.002495568012818694, |
|
"learning_rate": 9.990139316304112e-05, |
|
"loss": 0.0007, |
|
"step": 912 |
|
}, |
|
{ |
|
"epoch": 2.951111111111111, |
|
"grad_norm": 0.0025980276986956596, |
|
"learning_rate": 9.989996973099032e-05, |
|
"loss": 0.0008, |
|
"step": 913 |
|
}, |
|
{ |
|
"epoch": 2.9543434343434343, |
|
"grad_norm": 0.0013909428380429745, |
|
"learning_rate": 9.989853610879147e-05, |
|
"loss": 0.0004, |
|
"step": 914 |
|
}, |
|
{ |
|
"epoch": 2.9575757575757575, |
|
"grad_norm": 0.0025712710339576006, |
|
"learning_rate": 9.989709229673736e-05, |
|
"loss": 0.0008, |
|
"step": 915 |
|
}, |
|
{ |
|
"epoch": 2.9608080808080808, |
|
"grad_norm": 0.0029310788959264755, |
|
"learning_rate": 9.98956382951228e-05, |
|
"loss": 0.0007, |
|
"step": 916 |
|
}, |
|
{ |
|
"epoch": 2.964040404040404, |
|
"grad_norm": 0.0041606370359659195, |
|
"learning_rate": 9.989417410424475e-05, |
|
"loss": 0.0027, |
|
"step": 917 |
|
}, |
|
{ |
|
"epoch": 2.9672727272727273, |
|
"grad_norm": 0.001812944421544671, |
|
"learning_rate": 9.98926997244022e-05, |
|
"loss": 0.0006, |
|
"step": 918 |
|
}, |
|
{ |
|
"epoch": 2.9705050505050505, |
|
"grad_norm": 0.0026405269745737314, |
|
"learning_rate": 9.989121515589622e-05, |
|
"loss": 0.0011, |
|
"step": 919 |
|
}, |
|
{ |
|
"epoch": 2.973737373737374, |
|
"grad_norm": 0.003111011115834117, |
|
"learning_rate": 9.988972039902997e-05, |
|
"loss": 0.0009, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 2.976969696969697, |
|
"grad_norm": 0.0038792812265455723, |
|
"learning_rate": 9.988821545410874e-05, |
|
"loss": 0.0018, |
|
"step": 921 |
|
}, |
|
{ |
|
"epoch": 2.9802020202020203, |
|
"grad_norm": 0.0031774002127349377, |
|
"learning_rate": 9.988670032143981e-05, |
|
"loss": 0.0011, |
|
"step": 922 |
|
}, |
|
{ |
|
"epoch": 2.9834343434343435, |
|
"grad_norm": 0.002835791325196624, |
|
"learning_rate": 9.988517500133262e-05, |
|
"loss": 0.0007, |
|
"step": 923 |
|
}, |
|
{ |
|
"epoch": 2.986666666666667, |
|
"grad_norm": 0.003195343306288123, |
|
"learning_rate": 9.988363949409865e-05, |
|
"loss": 0.0008, |
|
"step": 924 |
|
}, |
|
{ |
|
"epoch": 2.98989898989899, |
|
"grad_norm": 0.004220789764076471, |
|
"learning_rate": 9.988209380005144e-05, |
|
"loss": 0.001, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 2.98989898989899, |
|
"eval_loss": 0.000867326685693115, |
|
"eval_runtime": 18.6533, |
|
"eval_samples_per_second": 5.361, |
|
"eval_steps_per_second": 1.34, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 2.993131313131313, |
|
"grad_norm": 0.0027963484171777964, |
|
"learning_rate": 9.98805379195067e-05, |
|
"loss": 0.001, |
|
"step": 926 |
|
}, |
|
{ |
|
"epoch": 2.996363636363636, |
|
"grad_norm": 0.002367103472352028, |
|
"learning_rate": 9.987897185278208e-05, |
|
"loss": 0.0006, |
|
"step": 927 |
|
}, |
|
{ |
|
"epoch": 2.9995959595959594, |
|
"grad_norm": 0.008955973200500011, |
|
"learning_rate": 9.987739560019746e-05, |
|
"loss": 0.0042, |
|
"step": 928 |
|
}, |
|
{ |
|
"epoch": 3.0028282828282826, |
|
"grad_norm": 0.0027045756578445435, |
|
"learning_rate": 9.987580916207468e-05, |
|
"loss": 0.0019, |
|
"step": 929 |
|
}, |
|
{ |
|
"epoch": 3.006060606060606, |
|
"grad_norm": 0.002136245835572481, |
|
"learning_rate": 9.987421253873775e-05, |
|
"loss": 0.0013, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 3.009292929292929, |
|
"grad_norm": 0.0018668243428692222, |
|
"learning_rate": 9.987260573051269e-05, |
|
"loss": 0.0006, |
|
"step": 931 |
|
}, |
|
{ |
|
"epoch": 3.0125252525252524, |
|
"grad_norm": 0.0018371654441580176, |
|
"learning_rate": 9.987098873772763e-05, |
|
"loss": 0.0004, |
|
"step": 932 |
|
}, |
|
{ |
|
"epoch": 3.0157575757575756, |
|
"grad_norm": 0.0020080492831766605, |
|
"learning_rate": 9.986936156071278e-05, |
|
"loss": 0.0007, |
|
"step": 933 |
|
}, |
|
{ |
|
"epoch": 3.018989898989899, |
|
"grad_norm": 0.0019176624482497573, |
|
"learning_rate": 9.986772419980044e-05, |
|
"loss": 0.0006, |
|
"step": 934 |
|
}, |
|
{ |
|
"epoch": 3.022222222222222, |
|
"grad_norm": 0.0024317919742316008, |
|
"learning_rate": 9.986607665532497e-05, |
|
"loss": 0.0006, |
|
"step": 935 |
|
}, |
|
{ |
|
"epoch": 3.0254545454545454, |
|
"grad_norm": 0.002591727999970317, |
|
"learning_rate": 9.986441892762281e-05, |
|
"loss": 0.0006, |
|
"step": 936 |
|
}, |
|
{ |
|
"epoch": 3.0286868686868686, |
|
"grad_norm": 0.0026875538751482964, |
|
"learning_rate": 9.98627510170325e-05, |
|
"loss": 0.0008, |
|
"step": 937 |
|
}, |
|
{ |
|
"epoch": 3.031919191919192, |
|
"grad_norm": 0.003272917354479432, |
|
"learning_rate": 9.986107292389464e-05, |
|
"loss": 0.0007, |
|
"step": 938 |
|
}, |
|
{ |
|
"epoch": 3.035151515151515, |
|
"grad_norm": 0.002936862874776125, |
|
"learning_rate": 9.985938464855191e-05, |
|
"loss": 0.0008, |
|
"step": 939 |
|
}, |
|
{ |
|
"epoch": 3.0383838383838384, |
|
"grad_norm": 0.0021727578714489937, |
|
"learning_rate": 9.985768619134909e-05, |
|
"loss": 0.0007, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 3.0416161616161617, |
|
"grad_norm": 0.002251671627163887, |
|
"learning_rate": 9.985597755263302e-05, |
|
"loss": 0.0006, |
|
"step": 941 |
|
}, |
|
{ |
|
"epoch": 3.044848484848485, |
|
"grad_norm": 0.0035938192158937454, |
|
"learning_rate": 9.985425873275263e-05, |
|
"loss": 0.0011, |
|
"step": 942 |
|
}, |
|
{ |
|
"epoch": 3.048080808080808, |
|
"grad_norm": 0.0024208389222621918, |
|
"learning_rate": 9.98525297320589e-05, |
|
"loss": 0.0007, |
|
"step": 943 |
|
}, |
|
{ |
|
"epoch": 3.0513131313131314, |
|
"grad_norm": 0.0019874447025358677, |
|
"learning_rate": 9.985079055090493e-05, |
|
"loss": 0.0005, |
|
"step": 944 |
|
}, |
|
{ |
|
"epoch": 3.0545454545454547, |
|
"grad_norm": 0.0014314191648736596, |
|
"learning_rate": 9.984904118964588e-05, |
|
"loss": 0.0004, |
|
"step": 945 |
|
}, |
|
{ |
|
"epoch": 3.057777777777778, |
|
"grad_norm": 0.0019143620738759637, |
|
"learning_rate": 9.984728164863898e-05, |
|
"loss": 0.0006, |
|
"step": 946 |
|
}, |
|
{ |
|
"epoch": 3.061010101010101, |
|
"grad_norm": 0.003406926291063428, |
|
"learning_rate": 9.984551192824355e-05, |
|
"loss": 0.0011, |
|
"step": 947 |
|
}, |
|
{ |
|
"epoch": 3.0642424242424244, |
|
"grad_norm": 0.0015149825485423207, |
|
"learning_rate": 9.9843732028821e-05, |
|
"loss": 0.0004, |
|
"step": 948 |
|
}, |
|
{ |
|
"epoch": 3.0674747474747477, |
|
"grad_norm": 0.0028568189591169357, |
|
"learning_rate": 9.98419419507348e-05, |
|
"loss": 0.0008, |
|
"step": 949 |
|
}, |
|
{ |
|
"epoch": 3.0707070707070705, |
|
"grad_norm": 0.00383683480322361, |
|
"learning_rate": 9.98401416943505e-05, |
|
"loss": 0.0006, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 3.0707070707070705, |
|
"eval_loss": 0.0009233049931935966, |
|
"eval_runtime": 18.6291, |
|
"eval_samples_per_second": 5.368, |
|
"eval_steps_per_second": 1.342, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 3.0739393939393937, |
|
"grad_norm": 0.004385147709399462, |
|
"learning_rate": 9.983833126003572e-05, |
|
"loss": 0.0009, |
|
"step": 951 |
|
}, |
|
{ |
|
"epoch": 3.077171717171717, |
|
"grad_norm": 0.0024896373506635427, |
|
"learning_rate": 9.98365106481602e-05, |
|
"loss": 0.0006, |
|
"step": 952 |
|
}, |
|
{ |
|
"epoch": 3.0804040404040403, |
|
"grad_norm": 0.002700845478102565, |
|
"learning_rate": 9.983467985909573e-05, |
|
"loss": 0.0006, |
|
"step": 953 |
|
}, |
|
{ |
|
"epoch": 3.0836363636363635, |
|
"grad_norm": 0.003025912679731846, |
|
"learning_rate": 9.983283889321615e-05, |
|
"loss": 0.0008, |
|
"step": 954 |
|
}, |
|
{ |
|
"epoch": 3.0868686868686868, |
|
"grad_norm": 0.003908303566277027, |
|
"learning_rate": 9.983098775089742e-05, |
|
"loss": 0.0017, |
|
"step": 955 |
|
}, |
|
{ |
|
"epoch": 3.09010101010101, |
|
"grad_norm": 0.0035844456870108843, |
|
"learning_rate": 9.982912643251757e-05, |
|
"loss": 0.0024, |
|
"step": 956 |
|
}, |
|
{ |
|
"epoch": 3.0933333333333333, |
|
"grad_norm": 0.0022256888914853334, |
|
"learning_rate": 9.98272549384567e-05, |
|
"loss": 0.0005, |
|
"step": 957 |
|
}, |
|
{ |
|
"epoch": 3.0965656565656565, |
|
"grad_norm": 0.0024845930747687817, |
|
"learning_rate": 9.982537326909697e-05, |
|
"loss": 0.0007, |
|
"step": 958 |
|
}, |
|
{ |
|
"epoch": 3.0997979797979798, |
|
"grad_norm": 0.0021472277585417032, |
|
"learning_rate": 9.982348142482269e-05, |
|
"loss": 0.0006, |
|
"step": 959 |
|
}, |
|
{ |
|
"epoch": 3.103030303030303, |
|
"grad_norm": 0.0027109517250210047, |
|
"learning_rate": 9.982157940602014e-05, |
|
"loss": 0.0007, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 3.1062626262626263, |
|
"grad_norm": 0.0016069613629952073, |
|
"learning_rate": 9.981966721307778e-05, |
|
"loss": 0.0005, |
|
"step": 961 |
|
}, |
|
{ |
|
"epoch": 3.1094949494949495, |
|
"grad_norm": 0.002237373497337103, |
|
"learning_rate": 9.981774484638606e-05, |
|
"loss": 0.0005, |
|
"step": 962 |
|
}, |
|
{ |
|
"epoch": 3.112727272727273, |
|
"grad_norm": 0.002959214383736253, |
|
"learning_rate": 9.981581230633758e-05, |
|
"loss": 0.001, |
|
"step": 963 |
|
}, |
|
{ |
|
"epoch": 3.115959595959596, |
|
"grad_norm": 0.0031158181373029947, |
|
"learning_rate": 9.981386959332697e-05, |
|
"loss": 0.0008, |
|
"step": 964 |
|
}, |
|
{ |
|
"epoch": 3.1191919191919193, |
|
"grad_norm": 0.007239287253469229, |
|
"learning_rate": 9.981191670775097e-05, |
|
"loss": 0.0037, |
|
"step": 965 |
|
}, |
|
{ |
|
"epoch": 3.1224242424242425, |
|
"grad_norm": 0.0029164832085371017, |
|
"learning_rate": 9.980995365000836e-05, |
|
"loss": 0.0006, |
|
"step": 966 |
|
}, |
|
{ |
|
"epoch": 3.125656565656566, |
|
"grad_norm": 0.0018486580811440945, |
|
"learning_rate": 9.980798042050004e-05, |
|
"loss": 0.0004, |
|
"step": 967 |
|
}, |
|
{ |
|
"epoch": 3.128888888888889, |
|
"grad_norm": 0.0032664432656019926, |
|
"learning_rate": 9.980599701962896e-05, |
|
"loss": 0.0009, |
|
"step": 968 |
|
}, |
|
{ |
|
"epoch": 3.1321212121212123, |
|
"grad_norm": 0.0037746732123196125, |
|
"learning_rate": 9.980400344780015e-05, |
|
"loss": 0.0015, |
|
"step": 969 |
|
}, |
|
{ |
|
"epoch": 3.1353535353535356, |
|
"grad_norm": 0.002780639799311757, |
|
"learning_rate": 9.98019997054207e-05, |
|
"loss": 0.0006, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 3.1385858585858584, |
|
"grad_norm": 0.0026836313772946596, |
|
"learning_rate": 9.979998579289984e-05, |
|
"loss": 0.0006, |
|
"step": 971 |
|
}, |
|
{ |
|
"epoch": 3.1418181818181816, |
|
"grad_norm": 0.002338019199669361, |
|
"learning_rate": 9.979796171064881e-05, |
|
"loss": 0.0007, |
|
"step": 972 |
|
}, |
|
{ |
|
"epoch": 3.145050505050505, |
|
"grad_norm": 0.0024902368895709515, |
|
"learning_rate": 9.979592745908095e-05, |
|
"loss": 0.0011, |
|
"step": 973 |
|
}, |
|
{ |
|
"epoch": 3.148282828282828, |
|
"grad_norm": 0.0033587226644158363, |
|
"learning_rate": 9.979388303861169e-05, |
|
"loss": 0.0006, |
|
"step": 974 |
|
}, |
|
{ |
|
"epoch": 3.1515151515151514, |
|
"grad_norm": 0.002750332234427333, |
|
"learning_rate": 9.97918284496585e-05, |
|
"loss": 0.0006, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 3.1515151515151514, |
|
"eval_loss": 0.0009188714902848005, |
|
"eval_runtime": 18.6436, |
|
"eval_samples_per_second": 5.364, |
|
"eval_steps_per_second": 1.341, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 3.1547474747474746, |
|
"grad_norm": 0.004522048868238926, |
|
"learning_rate": 9.978976369264098e-05, |
|
"loss": 0.0009, |
|
"step": 976 |
|
}, |
|
{ |
|
"epoch": 3.157979797979798, |
|
"grad_norm": 0.0020558724645525217, |
|
"learning_rate": 9.978768876798075e-05, |
|
"loss": 0.0004, |
|
"step": 977 |
|
}, |
|
{ |
|
"epoch": 3.161212121212121, |
|
"grad_norm": 0.0017784403171390295, |
|
"learning_rate": 9.978560367610156e-05, |
|
"loss": 0.0004, |
|
"step": 978 |
|
}, |
|
{ |
|
"epoch": 3.1644444444444444, |
|
"grad_norm": 0.0039355335757136345, |
|
"learning_rate": 9.978350841742919e-05, |
|
"loss": 0.0017, |
|
"step": 979 |
|
}, |
|
{ |
|
"epoch": 3.1676767676767676, |
|
"grad_norm": 0.0031672257464379072, |
|
"learning_rate": 9.978140299239152e-05, |
|
"loss": 0.0017, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 3.170909090909091, |
|
"grad_norm": 0.0048868078738451, |
|
"learning_rate": 9.977928740141851e-05, |
|
"loss": 0.001, |
|
"step": 981 |
|
}, |
|
{ |
|
"epoch": 3.174141414141414, |
|
"grad_norm": 0.002355673350393772, |
|
"learning_rate": 9.977716164494217e-05, |
|
"loss": 0.0005, |
|
"step": 982 |
|
}, |
|
{ |
|
"epoch": 3.1773737373737374, |
|
"grad_norm": 0.0035993049386888742, |
|
"learning_rate": 9.977502572339664e-05, |
|
"loss": 0.001, |
|
"step": 983 |
|
}, |
|
{ |
|
"epoch": 3.1806060606060607, |
|
"grad_norm": 0.00291255209594965, |
|
"learning_rate": 9.977287963721804e-05, |
|
"loss": 0.0007, |
|
"step": 984 |
|
}, |
|
{ |
|
"epoch": 3.183838383838384, |
|
"grad_norm": 0.002027621492743492, |
|
"learning_rate": 9.977072338684469e-05, |
|
"loss": 0.0006, |
|
"step": 985 |
|
}, |
|
{ |
|
"epoch": 3.187070707070707, |
|
"grad_norm": 0.0021479360293596983, |
|
"learning_rate": 9.976855697271689e-05, |
|
"loss": 0.0006, |
|
"step": 986 |
|
}, |
|
{ |
|
"epoch": 3.1903030303030304, |
|
"grad_norm": 0.002248652745038271, |
|
"learning_rate": 9.976638039527704e-05, |
|
"loss": 0.0005, |
|
"step": 987 |
|
}, |
|
{ |
|
"epoch": 3.1935353535353537, |
|
"grad_norm": 0.002659781137481332, |
|
"learning_rate": 9.976419365496963e-05, |
|
"loss": 0.0008, |
|
"step": 988 |
|
}, |
|
{ |
|
"epoch": 3.196767676767677, |
|
"grad_norm": 0.0037245461717247963, |
|
"learning_rate": 9.976199675224123e-05, |
|
"loss": 0.001, |
|
"step": 989 |
|
}, |
|
{ |
|
"epoch": 3.2, |
|
"grad_norm": 0.004409831948578358, |
|
"learning_rate": 9.975978968754045e-05, |
|
"loss": 0.0022, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 3.2032323232323234, |
|
"grad_norm": 0.0030354245100170374, |
|
"learning_rate": 9.975757246131803e-05, |
|
"loss": 0.0006, |
|
"step": 991 |
|
}, |
|
{ |
|
"epoch": 3.2064646464646467, |
|
"grad_norm": 0.0026450392324477434, |
|
"learning_rate": 9.975534507402671e-05, |
|
"loss": 0.0008, |
|
"step": 992 |
|
}, |
|
{ |
|
"epoch": 3.2096969696969695, |
|
"grad_norm": 0.003626378020271659, |
|
"learning_rate": 9.975310752612137e-05, |
|
"loss": 0.0004, |
|
"step": 993 |
|
}, |
|
{ |
|
"epoch": 3.2129292929292927, |
|
"grad_norm": 0.0026495913043618202, |
|
"learning_rate": 9.975085981805897e-05, |
|
"loss": 0.0018, |
|
"step": 994 |
|
}, |
|
{ |
|
"epoch": 3.216161616161616, |
|
"grad_norm": 0.0032507823780179024, |
|
"learning_rate": 9.974860195029847e-05, |
|
"loss": 0.0012, |
|
"step": 995 |
|
}, |
|
{ |
|
"epoch": 3.2193939393939393, |
|
"grad_norm": 0.002977850381284952, |
|
"learning_rate": 9.974633392330097e-05, |
|
"loss": 0.0006, |
|
"step": 996 |
|
}, |
|
{ |
|
"epoch": 3.2226262626262625, |
|
"grad_norm": 0.002434935886412859, |
|
"learning_rate": 9.974405573752965e-05, |
|
"loss": 0.0006, |
|
"step": 997 |
|
}, |
|
{ |
|
"epoch": 3.2258585858585858, |
|
"grad_norm": 0.002726243808865547, |
|
"learning_rate": 9.974176739344971e-05, |
|
"loss": 0.0007, |
|
"step": 998 |
|
}, |
|
{ |
|
"epoch": 3.229090909090909, |
|
"grad_norm": 0.003465299028903246, |
|
"learning_rate": 9.973946889152847e-05, |
|
"loss": 0.0006, |
|
"step": 999 |
|
}, |
|
{ |
|
"epoch": 3.2323232323232323, |
|
"grad_norm": 0.002784241922199726, |
|
"learning_rate": 9.973716023223531e-05, |
|
"loss": 0.0007, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 3.2323232323232323, |
|
"eval_loss": 0.0009049187647178769, |
|
"eval_runtime": 18.6293, |
|
"eval_samples_per_second": 5.368, |
|
"eval_steps_per_second": 1.342, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 3.2355555555555555, |
|
"grad_norm": 0.0021909528877586126, |
|
"learning_rate": 9.97348414160417e-05, |
|
"loss": 0.0005, |
|
"step": 1001 |
|
}, |
|
{ |
|
"epoch": 3.2387878787878788, |
|
"grad_norm": 0.003046189434826374, |
|
"learning_rate": 9.973251244342114e-05, |
|
"loss": 0.0009, |
|
"step": 1002 |
|
}, |
|
{ |
|
"epoch": 3.242020202020202, |
|
"grad_norm": 0.004089404363185167, |
|
"learning_rate": 9.973017331484926e-05, |
|
"loss": 0.0019, |
|
"step": 1003 |
|
}, |
|
{ |
|
"epoch": 3.2452525252525253, |
|
"grad_norm": 0.002897520549595356, |
|
"learning_rate": 9.972782403080372e-05, |
|
"loss": 0.0009, |
|
"step": 1004 |
|
}, |
|
{ |
|
"epoch": 3.2484848484848485, |
|
"grad_norm": 0.0027832165360450745, |
|
"learning_rate": 9.972546459176425e-05, |
|
"loss": 0.0006, |
|
"step": 1005 |
|
}, |
|
{ |
|
"epoch": 3.251717171717172, |
|
"grad_norm": 0.0021642649080604315, |
|
"learning_rate": 9.972309499821273e-05, |
|
"loss": 0.0005, |
|
"step": 1006 |
|
}, |
|
{ |
|
"epoch": 3.254949494949495, |
|
"grad_norm": 0.0030184059869498014, |
|
"learning_rate": 9.972071525063303e-05, |
|
"loss": 0.0019, |
|
"step": 1007 |
|
}, |
|
{ |
|
"epoch": 3.2581818181818183, |
|
"grad_norm": 0.0022592165041714907, |
|
"learning_rate": 9.971832534951108e-05, |
|
"loss": 0.0005, |
|
"step": 1008 |
|
}, |
|
{ |
|
"epoch": 3.2614141414141415, |
|
"grad_norm": 0.0013180490350350738, |
|
"learning_rate": 9.9715925295335e-05, |
|
"loss": 0.0004, |
|
"step": 1009 |
|
}, |
|
{ |
|
"epoch": 3.264646464646465, |
|
"grad_norm": 0.00403400557115674, |
|
"learning_rate": 9.971351508859488e-05, |
|
"loss": 0.0007, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 3.267878787878788, |
|
"grad_norm": 0.003028259379789233, |
|
"learning_rate": 9.971109472978288e-05, |
|
"loss": 0.0013, |
|
"step": 1011 |
|
}, |
|
{ |
|
"epoch": 3.2711111111111113, |
|
"grad_norm": 0.004238371271640062, |
|
"learning_rate": 9.97086642193933e-05, |
|
"loss": 0.0008, |
|
"step": 1012 |
|
}, |
|
{ |
|
"epoch": 3.274343434343434, |
|
"grad_norm": 0.003420765744522214, |
|
"learning_rate": 9.970622355792247e-05, |
|
"loss": 0.002, |
|
"step": 1013 |
|
}, |
|
{ |
|
"epoch": 3.2775757575757574, |
|
"grad_norm": 0.002874986035749316, |
|
"learning_rate": 9.970377274586879e-05, |
|
"loss": 0.0007, |
|
"step": 1014 |
|
}, |
|
{ |
|
"epoch": 3.2808080808080806, |
|
"grad_norm": 0.001124391914345324, |
|
"learning_rate": 9.970131178373277e-05, |
|
"loss": 0.0003, |
|
"step": 1015 |
|
}, |
|
{ |
|
"epoch": 3.284040404040404, |
|
"grad_norm": 0.0029891314916312695, |
|
"learning_rate": 9.969884067201695e-05, |
|
"loss": 0.0007, |
|
"step": 1016 |
|
}, |
|
{ |
|
"epoch": 3.287272727272727, |
|
"grad_norm": 0.0026401756331324577, |
|
"learning_rate": 9.969635941122595e-05, |
|
"loss": 0.0007, |
|
"step": 1017 |
|
}, |
|
{ |
|
"epoch": 3.2905050505050504, |
|
"grad_norm": 0.001897217589430511, |
|
"learning_rate": 9.969386800186649e-05, |
|
"loss": 0.0004, |
|
"step": 1018 |
|
}, |
|
{ |
|
"epoch": 3.2937373737373736, |
|
"grad_norm": 0.002726042177528143, |
|
"learning_rate": 9.969136644444731e-05, |
|
"loss": 0.0008, |
|
"step": 1019 |
|
}, |
|
{ |
|
"epoch": 3.296969696969697, |
|
"grad_norm": 0.0030468441545963287, |
|
"learning_rate": 9.968885473947932e-05, |
|
"loss": 0.0008, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 3.30020202020202, |
|
"grad_norm": 0.0016252384521067142, |
|
"learning_rate": 9.968633288747539e-05, |
|
"loss": 0.0004, |
|
"step": 1021 |
|
}, |
|
{ |
|
"epoch": 3.3034343434343434, |
|
"grad_norm": 0.002642759820446372, |
|
"learning_rate": 9.968380088895052e-05, |
|
"loss": 0.0007, |
|
"step": 1022 |
|
}, |
|
{ |
|
"epoch": 3.3066666666666666, |
|
"grad_norm": 0.003615351626649499, |
|
"learning_rate": 9.968125874442179e-05, |
|
"loss": 0.0008, |
|
"step": 1023 |
|
}, |
|
{ |
|
"epoch": 3.30989898989899, |
|
"grad_norm": 0.0026971250772476196, |
|
"learning_rate": 9.96787064544083e-05, |
|
"loss": 0.0007, |
|
"step": 1024 |
|
}, |
|
{ |
|
"epoch": 3.313131313131313, |
|
"grad_norm": 0.001658923109062016, |
|
"learning_rate": 9.96761440194313e-05, |
|
"loss": 0.0004, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 3.313131313131313, |
|
"eval_loss": 0.0008755004382692277, |
|
"eval_runtime": 18.6451, |
|
"eval_samples_per_second": 5.363, |
|
"eval_steps_per_second": 1.341, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 3.313131313131313, |
|
"step": 1025, |
|
"total_flos": 2.5838521494131835e+18, |
|
"train_loss": 0.001856716921518943, |
|
"train_runtime": 20010.9872, |
|
"train_samples_per_second": 12.368, |
|
"train_steps_per_second": 0.386 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 7725, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 25, |
|
"save_steps": 100, |
|
"stateful_callbacks": { |
|
"EarlyStoppingCallback": { |
|
"args": { |
|
"early_stopping_patience": 5, |
|
"early_stopping_threshold": 0.0 |
|
}, |
|
"attributes": { |
|
"early_stopping_patience_counter": 4 |
|
} |
|
}, |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 2.5838521494131835e+18, |
|
"train_batch_size": 4, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|