{ "best_metric": 0.0008496911614201963, "best_model_checkpoint": "/home/paperspace/Data/models/akoul_whitehorseliquidity_25c/llm3br256/checkpoint-900", "epoch": 3.313131313131313, "eval_steps": 25, "global_step": 1025, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0032323232323232323, "grad_norm": 0.03562889248132706, "learning_rate": 1.2936610608020701e-07, "loss": 0.0088, "step": 1 }, { "epoch": 0.006464646464646465, "grad_norm": 0.03539334237575531, "learning_rate": 2.5873221216041403e-07, "loss": 0.0098, "step": 2 }, { "epoch": 0.009696969696969697, "grad_norm": 0.03888387605547905, "learning_rate": 3.8809831824062096e-07, "loss": 0.0103, "step": 3 }, { "epoch": 0.01292929292929293, "grad_norm": 0.03283314406871796, "learning_rate": 5.174644243208281e-07, "loss": 0.0123, "step": 4 }, { "epoch": 0.01616161616161616, "grad_norm": 0.033392902463674545, "learning_rate": 6.468305304010349e-07, "loss": 0.0104, "step": 5 }, { "epoch": 0.019393939393939394, "grad_norm": 0.041360873728990555, "learning_rate": 7.761966364812419e-07, "loss": 0.0135, "step": 6 }, { "epoch": 0.022626262626262626, "grad_norm": 0.03297152742743492, "learning_rate": 9.055627425614489e-07, "loss": 0.01, "step": 7 }, { "epoch": 0.02585858585858586, "grad_norm": 0.03385542333126068, "learning_rate": 1.0349288486416561e-06, "loss": 0.0092, "step": 8 }, { "epoch": 0.02909090909090909, "grad_norm": 0.038460493087768555, "learning_rate": 1.164294954721863e-06, "loss": 0.0132, "step": 9 }, { "epoch": 0.03232323232323232, "grad_norm": 0.030043406412005424, "learning_rate": 1.2936610608020699e-06, "loss": 0.0088, "step": 10 }, { "epoch": 0.035555555555555556, "grad_norm": 0.039398446679115295, "learning_rate": 1.423027166882277e-06, "loss": 0.0118, "step": 11 }, { "epoch": 0.03878787878787879, "grad_norm": 0.03809528797864914, "learning_rate": 1.5523932729624839e-06, "loss": 0.0146, "step": 12 }, { "epoch": 0.04202020202020202, "grad_norm": 0.0388377383351326, "learning_rate": 1.6817593790426907e-06, "loss": 0.0102, "step": 13 }, { "epoch": 0.04525252525252525, "grad_norm": 0.043977439403533936, "learning_rate": 1.8111254851228978e-06, "loss": 0.0099, "step": 14 }, { "epoch": 0.048484848484848485, "grad_norm": 0.036720160394907, "learning_rate": 1.9404915912031045e-06, "loss": 0.0143, "step": 15 }, { "epoch": 0.05171717171717172, "grad_norm": 0.03513036668300629, "learning_rate": 2.0698576972833122e-06, "loss": 0.0092, "step": 16 }, { "epoch": 0.05494949494949495, "grad_norm": 0.03713906928896904, "learning_rate": 2.199223803363519e-06, "loss": 0.0144, "step": 17 }, { "epoch": 0.05818181818181818, "grad_norm": 0.0338338240981102, "learning_rate": 2.328589909443726e-06, "loss": 0.0133, "step": 18 }, { "epoch": 0.061414141414141414, "grad_norm": 0.0263836532831192, "learning_rate": 2.457956015523933e-06, "loss": 0.0113, "step": 19 }, { "epoch": 0.06464646464646465, "grad_norm": 0.01817137934267521, "learning_rate": 2.5873221216041398e-06, "loss": 0.0074, "step": 20 }, { "epoch": 0.06787878787878789, "grad_norm": 0.018132060766220093, "learning_rate": 2.716688227684347e-06, "loss": 0.008, "step": 21 }, { "epoch": 0.07111111111111111, "grad_norm": 0.02117246761918068, "learning_rate": 2.846054333764554e-06, "loss": 0.0091, "step": 22 }, { "epoch": 0.07434343434343435, "grad_norm": 0.01761500909924507, "learning_rate": 2.975420439844761e-06, "loss": 0.008, "step": 23 }, { "epoch": 0.07757575757575758, "grad_norm": 0.01908932812511921, "learning_rate": 3.1047865459249677e-06, "loss": 0.008, "step": 24 }, { "epoch": 0.08080808080808081, "grad_norm": 0.01619753986597061, "learning_rate": 3.234152652005175e-06, "loss": 0.0078, "step": 25 }, { "epoch": 0.08080808080808081, "eval_loss": 0.007939654402434826, "eval_runtime": 20.513, "eval_samples_per_second": 4.875, "eval_steps_per_second": 1.219, "step": 25 }, { "epoch": 0.08404040404040404, "grad_norm": 0.01418386958539486, "learning_rate": 3.3635187580853815e-06, "loss": 0.0067, "step": 26 }, { "epoch": 0.08727272727272728, "grad_norm": 0.018947051838040352, "learning_rate": 3.492884864165589e-06, "loss": 0.0086, "step": 27 }, { "epoch": 0.0905050505050505, "grad_norm": 0.016921700909733772, "learning_rate": 3.6222509702457957e-06, "loss": 0.0085, "step": 28 }, { "epoch": 0.09373737373737374, "grad_norm": 0.01551737543195486, "learning_rate": 3.751617076326003e-06, "loss": 0.0084, "step": 29 }, { "epoch": 0.09696969696969697, "grad_norm": 0.014143792912364006, "learning_rate": 3.880983182406209e-06, "loss": 0.0078, "step": 30 }, { "epoch": 0.10020202020202021, "grad_norm": 0.011397017166018486, "learning_rate": 4.010349288486417e-06, "loss": 0.0053, "step": 31 }, { "epoch": 0.10343434343434343, "grad_norm": 0.015161341056227684, "learning_rate": 4.1397153945666245e-06, "loss": 0.0073, "step": 32 }, { "epoch": 0.10666666666666667, "grad_norm": 0.012180107645690441, "learning_rate": 4.2690815006468305e-06, "loss": 0.0063, "step": 33 }, { "epoch": 0.1098989898989899, "grad_norm": 0.012770959176123142, "learning_rate": 4.398447606727038e-06, "loss": 0.0079, "step": 34 }, { "epoch": 0.11313131313131314, "grad_norm": 0.014727453701198101, "learning_rate": 4.527813712807244e-06, "loss": 0.0122, "step": 35 }, { "epoch": 0.11636363636363636, "grad_norm": 0.012778300791978836, "learning_rate": 4.657179818887452e-06, "loss": 0.0071, "step": 36 }, { "epoch": 0.1195959595959596, "grad_norm": 0.012685113586485386, "learning_rate": 4.786545924967659e-06, "loss": 0.0065, "step": 37 }, { "epoch": 0.12282828282828283, "grad_norm": 0.014156874269247055, "learning_rate": 4.915912031047866e-06, "loss": 0.0082, "step": 38 }, { "epoch": 0.12606060606060607, "grad_norm": 0.01115910243242979, "learning_rate": 5.045278137128073e-06, "loss": 0.007, "step": 39 }, { "epoch": 0.1292929292929293, "grad_norm": 0.014089164324104786, "learning_rate": 5.1746442432082795e-06, "loss": 0.0073, "step": 40 }, { "epoch": 0.13252525252525252, "grad_norm": 0.013440214097499847, "learning_rate": 5.304010349288486e-06, "loss": 0.0082, "step": 41 }, { "epoch": 0.13575757575757577, "grad_norm": 0.01083499938249588, "learning_rate": 5.433376455368694e-06, "loss": 0.006, "step": 42 }, { "epoch": 0.138989898989899, "grad_norm": 0.011082631535828114, "learning_rate": 5.5627425614489e-06, "loss": 0.007, "step": 43 }, { "epoch": 0.14222222222222222, "grad_norm": 0.012490961700677872, "learning_rate": 5.692108667529108e-06, "loss": 0.0062, "step": 44 }, { "epoch": 0.14545454545454545, "grad_norm": 0.012171825394034386, "learning_rate": 5.821474773609315e-06, "loss": 0.0066, "step": 45 }, { "epoch": 0.1486868686868687, "grad_norm": 0.010148804634809494, "learning_rate": 5.950840879689522e-06, "loss": 0.0042, "step": 46 }, { "epoch": 0.15191919191919193, "grad_norm": 0.013537143357098103, "learning_rate": 6.0802069857697286e-06, "loss": 0.0073, "step": 47 }, { "epoch": 0.15515151515151515, "grad_norm": 0.011110929772257805, "learning_rate": 6.2095730918499354e-06, "loss": 0.0072, "step": 48 }, { "epoch": 0.15838383838383838, "grad_norm": 0.00937278475612402, "learning_rate": 6.338939197930142e-06, "loss": 0.0054, "step": 49 }, { "epoch": 0.16161616161616163, "grad_norm": 0.011824984103441238, "learning_rate": 6.46830530401035e-06, "loss": 0.0119, "step": 50 }, { "epoch": 0.16161616161616163, "eval_loss": 0.005149205215275288, "eval_runtime": 18.6062, "eval_samples_per_second": 5.375, "eval_steps_per_second": 1.344, "step": 50 }, { "epoch": 0.16484848484848486, "grad_norm": 0.009701536037027836, "learning_rate": 6.597671410090557e-06, "loss": 0.0054, "step": 51 }, { "epoch": 0.16808080808080808, "grad_norm": 0.010364706628024578, "learning_rate": 6.727037516170763e-06, "loss": 0.0049, "step": 52 }, { "epoch": 0.1713131313131313, "grad_norm": 0.009346776641905308, "learning_rate": 6.856403622250971e-06, "loss": 0.0054, "step": 53 }, { "epoch": 0.17454545454545456, "grad_norm": 0.008815059438347816, "learning_rate": 6.985769728331178e-06, "loss": 0.0069, "step": 54 }, { "epoch": 0.17777777777777778, "grad_norm": 0.009431697428226471, "learning_rate": 7.115135834411385e-06, "loss": 0.0083, "step": 55 }, { "epoch": 0.181010101010101, "grad_norm": 0.012181616388261318, "learning_rate": 7.244501940491591e-06, "loss": 0.0065, "step": 56 }, { "epoch": 0.18424242424242424, "grad_norm": 0.009227719157934189, "learning_rate": 7.373868046571798e-06, "loss": 0.0052, "step": 57 }, { "epoch": 0.1874747474747475, "grad_norm": 0.009855546988546848, "learning_rate": 7.503234152652006e-06, "loss": 0.0082, "step": 58 }, { "epoch": 0.1907070707070707, "grad_norm": 0.008454745635390282, "learning_rate": 7.632600258732213e-06, "loss": 0.0044, "step": 59 }, { "epoch": 0.19393939393939394, "grad_norm": 0.010128876194357872, "learning_rate": 7.761966364812418e-06, "loss": 0.0092, "step": 60 }, { "epoch": 0.19717171717171716, "grad_norm": 0.009267722256481647, "learning_rate": 7.891332470892627e-06, "loss": 0.0049, "step": 61 }, { "epoch": 0.20040404040404042, "grad_norm": 0.009394655004143715, "learning_rate": 8.020698576972833e-06, "loss": 0.0059, "step": 62 }, { "epoch": 0.20363636363636364, "grad_norm": 0.008983040228486061, "learning_rate": 8.15006468305304e-06, "loss": 0.0041, "step": 63 }, { "epoch": 0.20686868686868687, "grad_norm": 0.009920783340930939, "learning_rate": 8.279430789133249e-06, "loss": 0.0058, "step": 64 }, { "epoch": 0.2101010101010101, "grad_norm": 0.007694128435105085, "learning_rate": 8.408796895213454e-06, "loss": 0.0033, "step": 65 }, { "epoch": 0.21333333333333335, "grad_norm": 0.01226822566241026, "learning_rate": 8.538163001293661e-06, "loss": 0.0099, "step": 66 }, { "epoch": 0.21656565656565657, "grad_norm": 0.010597337037324905, "learning_rate": 8.66752910737387e-06, "loss": 0.0081, "step": 67 }, { "epoch": 0.2197979797979798, "grad_norm": 0.009066218510270119, "learning_rate": 8.796895213454076e-06, "loss": 0.0044, "step": 68 }, { "epoch": 0.22303030303030302, "grad_norm": 0.01163114607334137, "learning_rate": 8.926261319534282e-06, "loss": 0.0129, "step": 69 }, { "epoch": 0.22626262626262628, "grad_norm": 0.012832598760724068, "learning_rate": 9.055627425614489e-06, "loss": 0.0071, "step": 70 }, { "epoch": 0.2294949494949495, "grad_norm": 0.010620299726724625, "learning_rate": 9.184993531694697e-06, "loss": 0.004, "step": 71 }, { "epoch": 0.23272727272727273, "grad_norm": 0.00913357175886631, "learning_rate": 9.314359637774904e-06, "loss": 0.0058, "step": 72 }, { "epoch": 0.23595959595959595, "grad_norm": 0.011772734113037586, "learning_rate": 9.44372574385511e-06, "loss": 0.0036, "step": 73 }, { "epoch": 0.2391919191919192, "grad_norm": 0.011326112784445286, "learning_rate": 9.573091849935318e-06, "loss": 0.0071, "step": 74 }, { "epoch": 0.24242424242424243, "grad_norm": 0.010667567141354084, "learning_rate": 9.702457956015525e-06, "loss": 0.0036, "step": 75 }, { "epoch": 0.24242424242424243, "eval_loss": 0.0031650287564843893, "eval_runtime": 18.6391, "eval_samples_per_second": 5.365, "eval_steps_per_second": 1.341, "step": 75 }, { "epoch": 0.24565656565656566, "grad_norm": 0.007108451332896948, "learning_rate": 9.831824062095732e-06, "loss": 0.0023, "step": 76 }, { "epoch": 0.24888888888888888, "grad_norm": 0.008278781548142433, "learning_rate": 9.961190168175938e-06, "loss": 0.0026, "step": 77 }, { "epoch": 0.25212121212121213, "grad_norm": 0.010979422368109226, "learning_rate": 1.0090556274256145e-05, "loss": 0.0036, "step": 78 }, { "epoch": 0.25535353535353533, "grad_norm": 0.007666402496397495, "learning_rate": 1.0219922380336352e-05, "loss": 0.0015, "step": 79 }, { "epoch": 0.2585858585858586, "grad_norm": 0.009133332408964634, "learning_rate": 1.0349288486416559e-05, "loss": 0.0021, "step": 80 }, { "epoch": 0.26181818181818184, "grad_norm": 0.009566927328705788, "learning_rate": 1.0478654592496766e-05, "loss": 0.0038, "step": 81 }, { "epoch": 0.26505050505050504, "grad_norm": 0.011765814386308193, "learning_rate": 1.0608020698576973e-05, "loss": 0.0061, "step": 82 }, { "epoch": 0.2682828282828283, "grad_norm": 0.009488740935921669, "learning_rate": 1.073738680465718e-05, "loss": 0.0023, "step": 83 }, { "epoch": 0.27151515151515154, "grad_norm": 0.007137450389564037, "learning_rate": 1.0866752910737388e-05, "loss": 0.0021, "step": 84 }, { "epoch": 0.27474747474747474, "grad_norm": 0.008984182961285114, "learning_rate": 1.0996119016817593e-05, "loss": 0.0028, "step": 85 }, { "epoch": 0.277979797979798, "grad_norm": 0.010403821244835854, "learning_rate": 1.11254851228978e-05, "loss": 0.0057, "step": 86 }, { "epoch": 0.2812121212121212, "grad_norm": 0.0071556540206074715, "learning_rate": 1.1254851228978009e-05, "loss": 0.0017, "step": 87 }, { "epoch": 0.28444444444444444, "grad_norm": 0.012338937260210514, "learning_rate": 1.1384217335058216e-05, "loss": 0.0029, "step": 88 }, { "epoch": 0.2876767676767677, "grad_norm": 0.010343515314161777, "learning_rate": 1.1513583441138421e-05, "loss": 0.0032, "step": 89 }, { "epoch": 0.2909090909090909, "grad_norm": 0.009991390630602837, "learning_rate": 1.164294954721863e-05, "loss": 0.0035, "step": 90 }, { "epoch": 0.29414141414141415, "grad_norm": 0.009020160883665085, "learning_rate": 1.1772315653298836e-05, "loss": 0.0033, "step": 91 }, { "epoch": 0.2973737373737374, "grad_norm": 0.0067694829776883125, "learning_rate": 1.1901681759379043e-05, "loss": 0.0019, "step": 92 }, { "epoch": 0.3006060606060606, "grad_norm": 0.013036763295531273, "learning_rate": 1.203104786545925e-05, "loss": 0.0041, "step": 93 }, { "epoch": 0.30383838383838385, "grad_norm": 0.009171461686491966, "learning_rate": 1.2160413971539457e-05, "loss": 0.006, "step": 94 }, { "epoch": 0.30707070707070705, "grad_norm": 0.007229423616081476, "learning_rate": 1.2289780077619664e-05, "loss": 0.002, "step": 95 }, { "epoch": 0.3103030303030303, "grad_norm": 0.009065698832273483, "learning_rate": 1.2419146183699871e-05, "loss": 0.0028, "step": 96 }, { "epoch": 0.31353535353535356, "grad_norm": 0.0068460931070148945, "learning_rate": 1.254851228978008e-05, "loss": 0.0024, "step": 97 }, { "epoch": 0.31676767676767675, "grad_norm": 0.12772586941719055, "learning_rate": 1.2677878395860285e-05, "loss": 0.0046, "step": 98 }, { "epoch": 0.32, "grad_norm": 0.008569791913032532, "learning_rate": 1.2807244501940493e-05, "loss": 0.0024, "step": 99 }, { "epoch": 0.32323232323232326, "grad_norm": 0.010226168669760227, "learning_rate": 1.29366106080207e-05, "loss": 0.004, "step": 100 }, { "epoch": 0.32323232323232326, "eval_loss": 0.002474932000041008, "eval_runtime": 18.5931, "eval_samples_per_second": 5.378, "eval_steps_per_second": 1.345, "step": 100 }, { "epoch": 0.32646464646464646, "grad_norm": 0.010962710715830326, "learning_rate": 1.3065976714100905e-05, "loss": 0.0018, "step": 101 }, { "epoch": 0.3296969696969697, "grad_norm": 0.01257998775690794, "learning_rate": 1.3195342820181114e-05, "loss": 0.0033, "step": 102 }, { "epoch": 0.3329292929292929, "grad_norm": 0.013269903138279915, "learning_rate": 1.332470892626132e-05, "loss": 0.0036, "step": 103 }, { "epoch": 0.33616161616161616, "grad_norm": 0.009509469382464886, "learning_rate": 1.3454075032341526e-05, "loss": 0.0034, "step": 104 }, { "epoch": 0.3393939393939394, "grad_norm": 0.00902635883539915, "learning_rate": 1.3583441138421735e-05, "loss": 0.0028, "step": 105 }, { "epoch": 0.3426262626262626, "grad_norm": 0.007760872133076191, "learning_rate": 1.3712807244501941e-05, "loss": 0.0024, "step": 106 }, { "epoch": 0.34585858585858587, "grad_norm": 0.009263481013476849, "learning_rate": 1.384217335058215e-05, "loss": 0.0031, "step": 107 }, { "epoch": 0.3490909090909091, "grad_norm": 0.00697364890947938, "learning_rate": 1.3971539456662355e-05, "loss": 0.0021, "step": 108 }, { "epoch": 0.3523232323232323, "grad_norm": 0.007600918412208557, "learning_rate": 1.4100905562742562e-05, "loss": 0.0018, "step": 109 }, { "epoch": 0.35555555555555557, "grad_norm": 0.00856352224946022, "learning_rate": 1.423027166882277e-05, "loss": 0.0031, "step": 110 }, { "epoch": 0.35878787878787877, "grad_norm": 0.007947358302772045, "learning_rate": 1.4359637774902976e-05, "loss": 0.0025, "step": 111 }, { "epoch": 0.362020202020202, "grad_norm": 0.007706194184720516, "learning_rate": 1.4489003880983183e-05, "loss": 0.0023, "step": 112 }, { "epoch": 0.3652525252525253, "grad_norm": 0.008598407730460167, "learning_rate": 1.4618369987063391e-05, "loss": 0.0028, "step": 113 }, { "epoch": 0.36848484848484847, "grad_norm": 0.00845644623041153, "learning_rate": 1.4747736093143596e-05, "loss": 0.0029, "step": 114 }, { "epoch": 0.3717171717171717, "grad_norm": 0.008533057756721973, "learning_rate": 1.4877102199223805e-05, "loss": 0.0039, "step": 115 }, { "epoch": 0.374949494949495, "grad_norm": 0.007429471705108881, "learning_rate": 1.5006468305304012e-05, "loss": 0.0024, "step": 116 }, { "epoch": 0.3781818181818182, "grad_norm": 0.0069627161137759686, "learning_rate": 1.5135834411384217e-05, "loss": 0.0018, "step": 117 }, { "epoch": 0.3814141414141414, "grad_norm": 0.007471222430467606, "learning_rate": 1.5265200517464426e-05, "loss": 0.0024, "step": 118 }, { "epoch": 0.3846464646464646, "grad_norm": 0.006873416714370251, "learning_rate": 1.5394566623544633e-05, "loss": 0.0019, "step": 119 }, { "epoch": 0.3878787878787879, "grad_norm": 0.009251467883586884, "learning_rate": 1.5523932729624836e-05, "loss": 0.0027, "step": 120 }, { "epoch": 0.39111111111111113, "grad_norm": 0.007302634883671999, "learning_rate": 1.5653298835705046e-05, "loss": 0.0017, "step": 121 }, { "epoch": 0.39434343434343433, "grad_norm": 0.006746005266904831, "learning_rate": 1.5782664941785253e-05, "loss": 0.0019, "step": 122 }, { "epoch": 0.3975757575757576, "grad_norm": 0.008011849597096443, "learning_rate": 1.591203104786546e-05, "loss": 0.0023, "step": 123 }, { "epoch": 0.40080808080808084, "grad_norm": 0.008799983188509941, "learning_rate": 1.6041397153945667e-05, "loss": 0.0023, "step": 124 }, { "epoch": 0.40404040404040403, "grad_norm": 0.010617449879646301, "learning_rate": 1.6170763260025874e-05, "loss": 0.0019, "step": 125 }, { "epoch": 0.40404040404040403, "eval_loss": 0.002037045545876026, "eval_runtime": 18.5899, "eval_samples_per_second": 5.379, "eval_steps_per_second": 1.345, "step": 125 }, { "epoch": 0.4072727272727273, "grad_norm": 0.006759752053767443, "learning_rate": 1.630012936610608e-05, "loss": 0.0018, "step": 126 }, { "epoch": 0.4105050505050505, "grad_norm": 0.008995631709694862, "learning_rate": 1.6429495472186288e-05, "loss": 0.0022, "step": 127 }, { "epoch": 0.41373737373737374, "grad_norm": 0.00835257675498724, "learning_rate": 1.6558861578266498e-05, "loss": 0.0026, "step": 128 }, { "epoch": 0.416969696969697, "grad_norm": 0.009772442281246185, "learning_rate": 1.66882276843467e-05, "loss": 0.0028, "step": 129 }, { "epoch": 0.4202020202020202, "grad_norm": 0.006931178271770477, "learning_rate": 1.6817593790426908e-05, "loss": 0.002, "step": 130 }, { "epoch": 0.42343434343434344, "grad_norm": 0.012796151451766491, "learning_rate": 1.694695989650712e-05, "loss": 0.0021, "step": 131 }, { "epoch": 0.4266666666666667, "grad_norm": 0.007226724177598953, "learning_rate": 1.7076326002587322e-05, "loss": 0.0033, "step": 132 }, { "epoch": 0.4298989898989899, "grad_norm": 0.0066243987530469894, "learning_rate": 1.720569210866753e-05, "loss": 0.002, "step": 133 }, { "epoch": 0.43313131313131314, "grad_norm": 0.007128287572413683, "learning_rate": 1.733505821474774e-05, "loss": 0.0021, "step": 134 }, { "epoch": 0.43636363636363634, "grad_norm": 0.007028148043900728, "learning_rate": 1.7464424320827943e-05, "loss": 0.0028, "step": 135 }, { "epoch": 0.4395959595959596, "grad_norm": 0.00866060983389616, "learning_rate": 1.7593790426908153e-05, "loss": 0.0053, "step": 136 }, { "epoch": 0.44282828282828285, "grad_norm": 0.005735491868108511, "learning_rate": 1.7723156532988356e-05, "loss": 0.0023, "step": 137 }, { "epoch": 0.44606060606060605, "grad_norm": 0.005858046934008598, "learning_rate": 1.7852522639068563e-05, "loss": 0.0019, "step": 138 }, { "epoch": 0.4492929292929293, "grad_norm": 0.007398343179374933, "learning_rate": 1.7981888745148774e-05, "loss": 0.0024, "step": 139 }, { "epoch": 0.45252525252525255, "grad_norm": 0.009320907294750214, "learning_rate": 1.8111254851228977e-05, "loss": 0.0063, "step": 140 }, { "epoch": 0.45575757575757575, "grad_norm": 0.008654161356389523, "learning_rate": 1.8240620957309184e-05, "loss": 0.0025, "step": 141 }, { "epoch": 0.458989898989899, "grad_norm": 0.004506159573793411, "learning_rate": 1.8369987063389394e-05, "loss": 0.0013, "step": 142 }, { "epoch": 0.4622222222222222, "grad_norm": 0.009703525342047215, "learning_rate": 1.8499353169469598e-05, "loss": 0.0034, "step": 143 }, { "epoch": 0.46545454545454545, "grad_norm": 0.007734385784715414, "learning_rate": 1.8628719275549808e-05, "loss": 0.0029, "step": 144 }, { "epoch": 0.4686868686868687, "grad_norm": 0.010392666794359684, "learning_rate": 1.8758085381630015e-05, "loss": 0.0028, "step": 145 }, { "epoch": 0.4719191919191919, "grad_norm": 0.01011224091053009, "learning_rate": 1.888745148771022e-05, "loss": 0.0036, "step": 146 }, { "epoch": 0.47515151515151516, "grad_norm": 0.005182855296880007, "learning_rate": 1.901681759379043e-05, "loss": 0.0014, "step": 147 }, { "epoch": 0.4783838383838384, "grad_norm": 0.009993299841880798, "learning_rate": 1.9146183699870636e-05, "loss": 0.003, "step": 148 }, { "epoch": 0.4816161616161616, "grad_norm": 0.008879208005964756, "learning_rate": 1.927554980595084e-05, "loss": 0.002, "step": 149 }, { "epoch": 0.48484848484848486, "grad_norm": 0.006573604419827461, "learning_rate": 1.940491591203105e-05, "loss": 0.0021, "step": 150 }, { "epoch": 0.48484848484848486, "eval_loss": 0.0017716821748763323, "eval_runtime": 18.6403, "eval_samples_per_second": 5.365, "eval_steps_per_second": 1.341, "step": 150 }, { "epoch": 0.48808080808080806, "grad_norm": 0.005542725790292025, "learning_rate": 1.9534282018111256e-05, "loss": 0.0017, "step": 151 }, { "epoch": 0.4913131313131313, "grad_norm": 0.008134805597364902, "learning_rate": 1.9663648124191463e-05, "loss": 0.002, "step": 152 }, { "epoch": 0.49454545454545457, "grad_norm": 0.009408293291926384, "learning_rate": 1.979301423027167e-05, "loss": 0.0031, "step": 153 }, { "epoch": 0.49777777777777776, "grad_norm": 0.007265687920153141, "learning_rate": 1.9922380336351877e-05, "loss": 0.0021, "step": 154 }, { "epoch": 0.501010101010101, "grad_norm": 0.009999910369515419, "learning_rate": 2.0051746442432084e-05, "loss": 0.0014, "step": 155 }, { "epoch": 0.5042424242424243, "grad_norm": 0.007147953379899263, "learning_rate": 2.018111254851229e-05, "loss": 0.0022, "step": 156 }, { "epoch": 0.5074747474747475, "grad_norm": 0.0072359428741037846, "learning_rate": 2.0310478654592497e-05, "loss": 0.0034, "step": 157 }, { "epoch": 0.5107070707070707, "grad_norm": 0.005821447819471359, "learning_rate": 2.0439844760672704e-05, "loss": 0.0018, "step": 158 }, { "epoch": 0.5139393939393939, "grad_norm": 0.007256666664034128, "learning_rate": 2.056921086675291e-05, "loss": 0.0027, "step": 159 }, { "epoch": 0.5171717171717172, "grad_norm": 0.00663521233946085, "learning_rate": 2.0698576972833118e-05, "loss": 0.0023, "step": 160 }, { "epoch": 0.5204040404040404, "grad_norm": 0.007504627574235201, "learning_rate": 2.0827943078913325e-05, "loss": 0.004, "step": 161 }, { "epoch": 0.5236363636363637, "grad_norm": 0.005167617462575436, "learning_rate": 2.0957309184993532e-05, "loss": 0.0015, "step": 162 }, { "epoch": 0.5268686868686868, "grad_norm": 0.00587807409465313, "learning_rate": 2.108667529107374e-05, "loss": 0.0016, "step": 163 }, { "epoch": 0.5301010101010101, "grad_norm": 0.005344127304852009, "learning_rate": 2.1216041397153946e-05, "loss": 0.0011, "step": 164 }, { "epoch": 0.5333333333333333, "grad_norm": 0.005351161118596792, "learning_rate": 2.1345407503234156e-05, "loss": 0.0016, "step": 165 }, { "epoch": 0.5365656565656566, "grad_norm": 0.006603873800486326, "learning_rate": 2.147477360931436e-05, "loss": 0.0017, "step": 166 }, { "epoch": 0.5397979797979798, "grad_norm": 0.006440309341996908, "learning_rate": 2.1604139715394566e-05, "loss": 0.0016, "step": 167 }, { "epoch": 0.5430303030303031, "grad_norm": 0.007731405086815357, "learning_rate": 2.1733505821474777e-05, "loss": 0.0024, "step": 168 }, { "epoch": 0.5462626262626262, "grad_norm": 0.0070298160426318645, "learning_rate": 2.186287192755498e-05, "loss": 0.0019, "step": 169 }, { "epoch": 0.5494949494949495, "grad_norm": 0.008382032625377178, "learning_rate": 2.1992238033635187e-05, "loss": 0.0017, "step": 170 }, { "epoch": 0.5527272727272727, "grad_norm": 0.007320054341107607, "learning_rate": 2.2121604139715397e-05, "loss": 0.0017, "step": 171 }, { "epoch": 0.555959595959596, "grad_norm": 0.008367806673049927, "learning_rate": 2.22509702457956e-05, "loss": 0.0036, "step": 172 }, { "epoch": 0.5591919191919192, "grad_norm": 0.008995486423373222, "learning_rate": 2.238033635187581e-05, "loss": 0.0019, "step": 173 }, { "epoch": 0.5624242424242424, "grad_norm": 0.007300530560314655, "learning_rate": 2.2509702457956018e-05, "loss": 0.0036, "step": 174 }, { "epoch": 0.5656565656565656, "grad_norm": 0.005388753954321146, "learning_rate": 2.263906856403622e-05, "loss": 0.0016, "step": 175 }, { "epoch": 0.5656565656565656, "eval_loss": 0.0016459682956337929, "eval_runtime": 18.6118, "eval_samples_per_second": 5.373, "eval_steps_per_second": 1.343, "step": 175 }, { "epoch": 0.5688888888888889, "grad_norm": 0.005545719526708126, "learning_rate": 2.276843467011643e-05, "loss": 0.0016, "step": 176 }, { "epoch": 0.5721212121212121, "grad_norm": 0.007394594140350819, "learning_rate": 2.289780077619664e-05, "loss": 0.0021, "step": 177 }, { "epoch": 0.5753535353535354, "grad_norm": 0.005089296959340572, "learning_rate": 2.3027166882276842e-05, "loss": 0.0013, "step": 178 }, { "epoch": 0.5785858585858585, "grad_norm": 0.00883107353001833, "learning_rate": 2.3156532988357052e-05, "loss": 0.0022, "step": 179 }, { "epoch": 0.5818181818181818, "grad_norm": 0.012213426642119884, "learning_rate": 2.328589909443726e-05, "loss": 0.005, "step": 180 }, { "epoch": 0.585050505050505, "grad_norm": 0.007255645003169775, "learning_rate": 2.3415265200517466e-05, "loss": 0.001, "step": 181 }, { "epoch": 0.5882828282828283, "grad_norm": 0.006818380672484636, "learning_rate": 2.3544631306597673e-05, "loss": 0.0015, "step": 182 }, { "epoch": 0.5915151515151515, "grad_norm": 0.006983195431530476, "learning_rate": 2.367399741267788e-05, "loss": 0.0027, "step": 183 }, { "epoch": 0.5947474747474748, "grad_norm": 0.012661872431635857, "learning_rate": 2.3803363518758087e-05, "loss": 0.0053, "step": 184 }, { "epoch": 0.597979797979798, "grad_norm": 0.005046526901423931, "learning_rate": 2.3932729624838294e-05, "loss": 0.0019, "step": 185 }, { "epoch": 0.6012121212121212, "grad_norm": 0.004457033704966307, "learning_rate": 2.40620957309185e-05, "loss": 0.0012, "step": 186 }, { "epoch": 0.6044444444444445, "grad_norm": 0.00732004176825285, "learning_rate": 2.4191461836998707e-05, "loss": 0.0017, "step": 187 }, { "epoch": 0.6076767676767677, "grad_norm": 0.006394708063453436, "learning_rate": 2.4320827943078914e-05, "loss": 0.0019, "step": 188 }, { "epoch": 0.610909090909091, "grad_norm": 0.007961345836520195, "learning_rate": 2.445019404915912e-05, "loss": 0.0034, "step": 189 }, { "epoch": 0.6141414141414141, "grad_norm": 0.0068709347397089005, "learning_rate": 2.4579560155239328e-05, "loss": 0.0013, "step": 190 }, { "epoch": 0.6173737373737374, "grad_norm": 0.009581703692674637, "learning_rate": 2.4708926261319535e-05, "loss": 0.0026, "step": 191 }, { "epoch": 0.6206060606060606, "grad_norm": 0.007400548551231623, "learning_rate": 2.4838292367399742e-05, "loss": 0.0024, "step": 192 }, { "epoch": 0.6238383838383839, "grad_norm": 0.00970857497304678, "learning_rate": 2.496765847347995e-05, "loss": 0.0013, "step": 193 }, { "epoch": 0.6270707070707071, "grad_norm": 0.007368095684796572, "learning_rate": 2.509702457956016e-05, "loss": 0.0016, "step": 194 }, { "epoch": 0.6303030303030303, "grad_norm": 0.007306639105081558, "learning_rate": 2.5226390685640362e-05, "loss": 0.0015, "step": 195 }, { "epoch": 0.6335353535353535, "grad_norm": 0.008257709443569183, "learning_rate": 2.535575679172057e-05, "loss": 0.0017, "step": 196 }, { "epoch": 0.6367676767676768, "grad_norm": 0.0050062090158462524, "learning_rate": 2.548512289780078e-05, "loss": 0.0012, "step": 197 }, { "epoch": 0.64, "grad_norm": 0.007328143808990717, "learning_rate": 2.5614489003880986e-05, "loss": 0.0025, "step": 198 }, { "epoch": 0.6432323232323233, "grad_norm": 0.007282217964529991, "learning_rate": 2.574385510996119e-05, "loss": 0.0018, "step": 199 }, { "epoch": 0.6464646464646465, "grad_norm": 0.007711055688560009, "learning_rate": 2.58732212160414e-05, "loss": 0.0013, "step": 200 }, { "epoch": 0.6464646464646465, "eval_loss": 0.0015150802209973335, "eval_runtime": 18.6339, "eval_samples_per_second": 5.367, "eval_steps_per_second": 1.342, "step": 200 }, { "epoch": 0.6496969696969697, "grad_norm": 0.006153900176286697, "learning_rate": 2.6002587322121607e-05, "loss": 0.0016, "step": 201 }, { "epoch": 0.6529292929292929, "grad_norm": 0.0070861754938960075, "learning_rate": 2.613195342820181e-05, "loss": 0.002, "step": 202 }, { "epoch": 0.6561616161616162, "grad_norm": 0.0048012156039476395, "learning_rate": 2.626131953428202e-05, "loss": 0.0012, "step": 203 }, { "epoch": 0.6593939393939394, "grad_norm": 0.006167956627905369, "learning_rate": 2.6390685640362228e-05, "loss": 0.0018, "step": 204 }, { "epoch": 0.6626262626262627, "grad_norm": 0.0051431735046207905, "learning_rate": 2.652005174644243e-05, "loss": 0.0013, "step": 205 }, { "epoch": 0.6658585858585858, "grad_norm": 0.009899413213133812, "learning_rate": 2.664941785252264e-05, "loss": 0.0031, "step": 206 }, { "epoch": 0.6690909090909091, "grad_norm": 0.007686274591833353, "learning_rate": 2.677878395860285e-05, "loss": 0.0022, "step": 207 }, { "epoch": 0.6723232323232323, "grad_norm": 0.004816455766558647, "learning_rate": 2.6908150064683052e-05, "loss": 0.0011, "step": 208 }, { "epoch": 0.6755555555555556, "grad_norm": 0.004873145837336779, "learning_rate": 2.7037516170763262e-05, "loss": 0.0011, "step": 209 }, { "epoch": 0.6787878787878788, "grad_norm": 0.006136384792625904, "learning_rate": 2.716688227684347e-05, "loss": 0.0016, "step": 210 }, { "epoch": 0.682020202020202, "grad_norm": 0.004508585669100285, "learning_rate": 2.7296248382923673e-05, "loss": 0.0012, "step": 211 }, { "epoch": 0.6852525252525252, "grad_norm": 0.005723617039620876, "learning_rate": 2.7425614489003883e-05, "loss": 0.0016, "step": 212 }, { "epoch": 0.6884848484848485, "grad_norm": 0.0072562843561172485, "learning_rate": 2.755498059508409e-05, "loss": 0.0015, "step": 213 }, { "epoch": 0.6917171717171717, "grad_norm": 0.006079181097447872, "learning_rate": 2.76843467011643e-05, "loss": 0.0017, "step": 214 }, { "epoch": 0.694949494949495, "grad_norm": 0.011977693997323513, "learning_rate": 2.7813712807244503e-05, "loss": 0.0048, "step": 215 }, { "epoch": 0.6981818181818182, "grad_norm": 0.007406435441225767, "learning_rate": 2.794307891332471e-05, "loss": 0.0022, "step": 216 }, { "epoch": 0.7014141414141414, "grad_norm": 0.006909268908202648, "learning_rate": 2.807244501940492e-05, "loss": 0.0022, "step": 217 }, { "epoch": 0.7046464646464646, "grad_norm": 0.0051758866757154465, "learning_rate": 2.8201811125485124e-05, "loss": 0.0012, "step": 218 }, { "epoch": 0.7078787878787879, "grad_norm": 0.00657995231449604, "learning_rate": 2.833117723156533e-05, "loss": 0.0014, "step": 219 }, { "epoch": 0.7111111111111111, "grad_norm": 0.006561241112649441, "learning_rate": 2.846054333764554e-05, "loss": 0.0018, "step": 220 }, { "epoch": 0.7143434343434344, "grad_norm": 0.005906842183321714, "learning_rate": 2.8589909443725745e-05, "loss": 0.0013, "step": 221 }, { "epoch": 0.7175757575757575, "grad_norm": 0.004000976216048002, "learning_rate": 2.871927554980595e-05, "loss": 0.0009, "step": 222 }, { "epoch": 0.7208080808080808, "grad_norm": 0.006313610821962357, "learning_rate": 2.8848641655886162e-05, "loss": 0.0015, "step": 223 }, { "epoch": 0.724040404040404, "grad_norm": 0.005339731462299824, "learning_rate": 2.8978007761966365e-05, "loss": 0.0014, "step": 224 }, { "epoch": 0.7272727272727273, "grad_norm": 0.00704893097281456, "learning_rate": 2.9107373868046572e-05, "loss": 0.0017, "step": 225 }, { "epoch": 0.7272727272727273, "eval_loss": 0.0014526441227644682, "eval_runtime": 18.5916, "eval_samples_per_second": 5.379, "eval_steps_per_second": 1.345, "step": 225 }, { "epoch": 0.7305050505050505, "grad_norm": 0.007863117381930351, "learning_rate": 2.9236739974126783e-05, "loss": 0.002, "step": 226 }, { "epoch": 0.7337373737373737, "grad_norm": 0.007929647341370583, "learning_rate": 2.936610608020699e-05, "loss": 0.0014, "step": 227 }, { "epoch": 0.7369696969696969, "grad_norm": 0.008834928274154663, "learning_rate": 2.9495472186287193e-05, "loss": 0.0033, "step": 228 }, { "epoch": 0.7402020202020202, "grad_norm": 0.005479221232235432, "learning_rate": 2.9624838292367403e-05, "loss": 0.0013, "step": 229 }, { "epoch": 0.7434343434343434, "grad_norm": 0.008690214715898037, "learning_rate": 2.975420439844761e-05, "loss": 0.0031, "step": 230 }, { "epoch": 0.7466666666666667, "grad_norm": 0.006373909767717123, "learning_rate": 2.9883570504527814e-05, "loss": 0.0013, "step": 231 }, { "epoch": 0.74989898989899, "grad_norm": 0.005924658849835396, "learning_rate": 3.0012936610608024e-05, "loss": 0.0014, "step": 232 }, { "epoch": 0.7531313131313131, "grad_norm": 0.007938825525343418, "learning_rate": 3.014230271668823e-05, "loss": 0.0027, "step": 233 }, { "epoch": 0.7563636363636363, "grad_norm": 0.007831827737390995, "learning_rate": 3.0271668822768434e-05, "loss": 0.0025, "step": 234 }, { "epoch": 0.7595959595959596, "grad_norm": 0.008570835925638676, "learning_rate": 3.0401034928848644e-05, "loss": 0.003, "step": 235 }, { "epoch": 0.7628282828282829, "grad_norm": 0.005871222820132971, "learning_rate": 3.053040103492885e-05, "loss": 0.0013, "step": 236 }, { "epoch": 0.7660606060606061, "grad_norm": 0.007041990291327238, "learning_rate": 3.0659767141009055e-05, "loss": 0.0028, "step": 237 }, { "epoch": 0.7692929292929293, "grad_norm": 0.0073891859501600266, "learning_rate": 3.0789133247089265e-05, "loss": 0.0023, "step": 238 }, { "epoch": 0.7725252525252525, "grad_norm": 0.005527487024664879, "learning_rate": 3.0918499353169475e-05, "loss": 0.0014, "step": 239 }, { "epoch": 0.7757575757575758, "grad_norm": 0.007551091257482767, "learning_rate": 3.104786545924967e-05, "loss": 0.002, "step": 240 }, { "epoch": 0.778989898989899, "grad_norm": 0.005599314346909523, "learning_rate": 3.117723156532988e-05, "loss": 0.002, "step": 241 }, { "epoch": 0.7822222222222223, "grad_norm": 0.007462210953235626, "learning_rate": 3.130659767141009e-05, "loss": 0.0019, "step": 242 }, { "epoch": 0.7854545454545454, "grad_norm": 0.004866201896220446, "learning_rate": 3.14359637774903e-05, "loss": 0.0012, "step": 243 }, { "epoch": 0.7886868686868687, "grad_norm": 0.005268635228276253, "learning_rate": 3.1565329883570506e-05, "loss": 0.0013, "step": 244 }, { "epoch": 0.7919191919191919, "grad_norm": 0.007261104416102171, "learning_rate": 3.169469598965072e-05, "loss": 0.0031, "step": 245 }, { "epoch": 0.7951515151515152, "grad_norm": 0.006308354903012514, "learning_rate": 3.182406209573092e-05, "loss": 0.0014, "step": 246 }, { "epoch": 0.7983838383838384, "grad_norm": 0.0074616689234972, "learning_rate": 3.1953428201811124e-05, "loss": 0.0015, "step": 247 }, { "epoch": 0.8016161616161617, "grad_norm": 0.007333788555115461, "learning_rate": 3.2082794307891334e-05, "loss": 0.0014, "step": 248 }, { "epoch": 0.8048484848484848, "grad_norm": 0.006965926848351955, "learning_rate": 3.2212160413971544e-05, "loss": 0.0018, "step": 249 }, { "epoch": 0.8080808080808081, "grad_norm": 0.00535194855183363, "learning_rate": 3.234152652005175e-05, "loss": 0.0015, "step": 250 }, { "epoch": 0.8080808080808081, "eval_loss": 0.0013973835157230496, "eval_runtime": 18.6113, "eval_samples_per_second": 5.373, "eval_steps_per_second": 1.343, "step": 250 }, { "epoch": 0.8113131313131313, "grad_norm": 0.006415734998881817, "learning_rate": 3.247089262613196e-05, "loss": 0.0014, "step": 251 }, { "epoch": 0.8145454545454546, "grad_norm": 0.00617300346493721, "learning_rate": 3.260025873221216e-05, "loss": 0.0017, "step": 252 }, { "epoch": 0.8177777777777778, "grad_norm": 0.007527848239988089, "learning_rate": 3.2729624838292365e-05, "loss": 0.0017, "step": 253 }, { "epoch": 0.821010101010101, "grad_norm": 0.00747877499088645, "learning_rate": 3.2858990944372575e-05, "loss": 0.0014, "step": 254 }, { "epoch": 0.8242424242424242, "grad_norm": 0.005451089236885309, "learning_rate": 3.2988357050452786e-05, "loss": 0.0012, "step": 255 }, { "epoch": 0.8274747474747475, "grad_norm": 0.006235205102711916, "learning_rate": 3.3117723156532996e-05, "loss": 0.0014, "step": 256 }, { "epoch": 0.8307070707070707, "grad_norm": 0.005764327012002468, "learning_rate": 3.324708926261319e-05, "loss": 0.0013, "step": 257 }, { "epoch": 0.833939393939394, "grad_norm": 0.007408145349472761, "learning_rate": 3.33764553686934e-05, "loss": 0.0013, "step": 258 }, { "epoch": 0.8371717171717171, "grad_norm": 0.004855205304920673, "learning_rate": 3.350582147477361e-05, "loss": 0.0009, "step": 259 }, { "epoch": 0.8404040404040404, "grad_norm": 0.010664681904017925, "learning_rate": 3.3635187580853817e-05, "loss": 0.0016, "step": 260 }, { "epoch": 0.8436363636363636, "grad_norm": 0.006055675912648439, "learning_rate": 3.376455368693403e-05, "loss": 0.0016, "step": 261 }, { "epoch": 0.8468686868686869, "grad_norm": 0.01056759525090456, "learning_rate": 3.389391979301424e-05, "loss": 0.0016, "step": 262 }, { "epoch": 0.8501010101010101, "grad_norm": 0.007450331002473831, "learning_rate": 3.4023285899094434e-05, "loss": 0.0016, "step": 263 }, { "epoch": 0.8533333333333334, "grad_norm": 0.0053154826164245605, "learning_rate": 3.4152652005174644e-05, "loss": 0.0011, "step": 264 }, { "epoch": 0.8565656565656565, "grad_norm": 0.0060371761210262775, "learning_rate": 3.4282018111254854e-05, "loss": 0.0013, "step": 265 }, { "epoch": 0.8597979797979798, "grad_norm": 0.005466979928314686, "learning_rate": 3.441138421733506e-05, "loss": 0.0012, "step": 266 }, { "epoch": 0.863030303030303, "grad_norm": 0.00831522885710001, "learning_rate": 3.454075032341527e-05, "loss": 0.0014, "step": 267 }, { "epoch": 0.8662626262626263, "grad_norm": 0.007963420823216438, "learning_rate": 3.467011642949548e-05, "loss": 0.0025, "step": 268 }, { "epoch": 0.8694949494949495, "grad_norm": 0.004937573801726103, "learning_rate": 3.4799482535575675e-05, "loss": 0.0011, "step": 269 }, { "epoch": 0.8727272727272727, "grad_norm": 0.0058564417995512486, "learning_rate": 3.4928848641655885e-05, "loss": 0.0016, "step": 270 }, { "epoch": 0.8759595959595959, "grad_norm": 0.007912451401352882, "learning_rate": 3.5058214747736096e-05, "loss": 0.0026, "step": 271 }, { "epoch": 0.8791919191919192, "grad_norm": 0.003929544240236282, "learning_rate": 3.5187580853816306e-05, "loss": 0.0011, "step": 272 }, { "epoch": 0.8824242424242424, "grad_norm": 0.008150831796228886, "learning_rate": 3.531694695989651e-05, "loss": 0.0018, "step": 273 }, { "epoch": 0.8856565656565657, "grad_norm": 0.009232031181454659, "learning_rate": 3.544631306597671e-05, "loss": 0.0042, "step": 274 }, { "epoch": 0.8888888888888888, "grad_norm": 0.007469587959349155, "learning_rate": 3.557567917205692e-05, "loss": 0.0023, "step": 275 }, { "epoch": 0.8888888888888888, "eval_loss": 0.0013014678843319416, "eval_runtime": 18.7103, "eval_samples_per_second": 5.345, "eval_steps_per_second": 1.336, "step": 275 }, { "epoch": 0.8921212121212121, "grad_norm": 0.006450500804930925, "learning_rate": 3.570504527813713e-05, "loss": 0.0014, "step": 276 }, { "epoch": 0.8953535353535353, "grad_norm": 0.004984740167856216, "learning_rate": 3.583441138421734e-05, "loss": 0.0014, "step": 277 }, { "epoch": 0.8985858585858586, "grad_norm": 0.006635539699345827, "learning_rate": 3.596377749029755e-05, "loss": 0.0016, "step": 278 }, { "epoch": 0.9018181818181819, "grad_norm": 0.010307137854397297, "learning_rate": 3.609314359637775e-05, "loss": 0.0025, "step": 279 }, { "epoch": 0.9050505050505051, "grad_norm": 0.008318118751049042, "learning_rate": 3.6222509702457954e-05, "loss": 0.0028, "step": 280 }, { "epoch": 0.9082828282828282, "grad_norm": 0.007523260544985533, "learning_rate": 3.6351875808538164e-05, "loss": 0.0022, "step": 281 }, { "epoch": 0.9115151515151515, "grad_norm": 0.00529656745493412, "learning_rate": 3.648124191461837e-05, "loss": 0.001, "step": 282 }, { "epoch": 0.9147474747474748, "grad_norm": 0.0050112344324588776, "learning_rate": 3.661060802069858e-05, "loss": 0.0012, "step": 283 }, { "epoch": 0.917979797979798, "grad_norm": 0.006318412255495787, "learning_rate": 3.673997412677879e-05, "loss": 0.0017, "step": 284 }, { "epoch": 0.9212121212121213, "grad_norm": 0.006407878361642361, "learning_rate": 3.6869340232859e-05, "loss": 0.0022, "step": 285 }, { "epoch": 0.9244444444444444, "grad_norm": 0.004356731195002794, "learning_rate": 3.6998706338939195e-05, "loss": 0.0007, "step": 286 }, { "epoch": 0.9276767676767677, "grad_norm": 0.004966625012457371, "learning_rate": 3.7128072445019406e-05, "loss": 0.0012, "step": 287 }, { "epoch": 0.9309090909090909, "grad_norm": 0.0055012330412864685, "learning_rate": 3.7257438551099616e-05, "loss": 0.0011, "step": 288 }, { "epoch": 0.9341414141414142, "grad_norm": 0.006766812410205603, "learning_rate": 3.738680465717982e-05, "loss": 0.0017, "step": 289 }, { "epoch": 0.9373737373737374, "grad_norm": 0.006011553108692169, "learning_rate": 3.751617076326003e-05, "loss": 0.0022, "step": 290 }, { "epoch": 0.9406060606060606, "grad_norm": 0.0056586177088320255, "learning_rate": 3.764553686934023e-05, "loss": 0.0014, "step": 291 }, { "epoch": 0.9438383838383838, "grad_norm": 0.0064941453747451305, "learning_rate": 3.777490297542044e-05, "loss": 0.0011, "step": 292 }, { "epoch": 0.9470707070707071, "grad_norm": 0.006994554307311773, "learning_rate": 3.790426908150065e-05, "loss": 0.0021, "step": 293 }, { "epoch": 0.9503030303030303, "grad_norm": 0.006325817201286554, "learning_rate": 3.803363518758086e-05, "loss": 0.0014, "step": 294 }, { "epoch": 0.9535353535353536, "grad_norm": 0.006501290015876293, "learning_rate": 3.816300129366106e-05, "loss": 0.0017, "step": 295 }, { "epoch": 0.9567676767676768, "grad_norm": 0.006014223676174879, "learning_rate": 3.829236739974127e-05, "loss": 0.0013, "step": 296 }, { "epoch": 0.96, "grad_norm": 0.005024549085646868, "learning_rate": 3.8421733505821475e-05, "loss": 0.0013, "step": 297 }, { "epoch": 0.9632323232323232, "grad_norm": 0.00840625911951065, "learning_rate": 3.855109961190168e-05, "loss": 0.0017, "step": 298 }, { "epoch": 0.9664646464646465, "grad_norm": 0.005791939329355955, "learning_rate": 3.868046571798189e-05, "loss": 0.0014, "step": 299 }, { "epoch": 0.9696969696969697, "grad_norm": 0.0047224657610058784, "learning_rate": 3.88098318240621e-05, "loss": 0.0012, "step": 300 }, { "epoch": 0.9696969696969697, "eval_loss": 0.0012624365044757724, "eval_runtime": 18.7089, "eval_samples_per_second": 5.345, "eval_steps_per_second": 1.336, "step": 300 }, { "epoch": 0.972929292929293, "grad_norm": 0.005727376788854599, "learning_rate": 3.893919793014231e-05, "loss": 0.0013, "step": 301 }, { "epoch": 0.9761616161616161, "grad_norm": 0.007024977821856737, "learning_rate": 3.906856403622251e-05, "loss": 0.0018, "step": 302 }, { "epoch": 0.9793939393939394, "grad_norm": 0.004745373502373695, "learning_rate": 3.9197930142302716e-05, "loss": 0.001, "step": 303 }, { "epoch": 0.9826262626262626, "grad_norm": 0.0050310962833464146, "learning_rate": 3.9327296248382926e-05, "loss": 0.0013, "step": 304 }, { "epoch": 0.9858585858585859, "grad_norm": 0.008561142720282078, "learning_rate": 3.945666235446313e-05, "loss": 0.0018, "step": 305 }, { "epoch": 0.9890909090909091, "grad_norm": 0.0055971029214560986, "learning_rate": 3.958602846054334e-05, "loss": 0.0013, "step": 306 }, { "epoch": 0.9923232323232323, "grad_norm": 0.007602418772876263, "learning_rate": 3.971539456662355e-05, "loss": 0.0014, "step": 307 }, { "epoch": 0.9955555555555555, "grad_norm": 0.0054479725658893585, "learning_rate": 3.9844760672703754e-05, "loss": 0.0014, "step": 308 }, { "epoch": 0.9987878787878788, "grad_norm": 0.006376450881361961, "learning_rate": 3.997412677878396e-05, "loss": 0.0013, "step": 309 }, { "epoch": 1.002020202020202, "grad_norm": 0.00963524729013443, "learning_rate": 4.010349288486417e-05, "loss": 0.0025, "step": 310 }, { "epoch": 1.0052525252525253, "grad_norm": 0.004950121510773897, "learning_rate": 4.023285899094437e-05, "loss": 0.0009, "step": 311 }, { "epoch": 1.0084848484848485, "grad_norm": 0.0047010756097733974, "learning_rate": 4.036222509702458e-05, "loss": 0.0012, "step": 312 }, { "epoch": 1.0117171717171718, "grad_norm": 0.004061982501298189, "learning_rate": 4.049159120310479e-05, "loss": 0.0008, "step": 313 }, { "epoch": 1.014949494949495, "grad_norm": 0.006061443593353033, "learning_rate": 4.0620957309184995e-05, "loss": 0.0022, "step": 314 }, { "epoch": 1.018181818181818, "grad_norm": 0.005154592916369438, "learning_rate": 4.07503234152652e-05, "loss": 0.0023, "step": 315 }, { "epoch": 1.0214141414141413, "grad_norm": 0.004801798611879349, "learning_rate": 4.087968952134541e-05, "loss": 0.0012, "step": 316 }, { "epoch": 1.0246464646464646, "grad_norm": 0.004275497514754534, "learning_rate": 4.100905562742562e-05, "loss": 0.0012, "step": 317 }, { "epoch": 1.0278787878787878, "grad_norm": 0.006128865294158459, "learning_rate": 4.113842173350582e-05, "loss": 0.0022, "step": 318 }, { "epoch": 1.031111111111111, "grad_norm": 0.004465613514184952, "learning_rate": 4.126778783958603e-05, "loss": 0.0019, "step": 319 }, { "epoch": 1.0343434343434343, "grad_norm": 0.004612160846590996, "learning_rate": 4.1397153945666236e-05, "loss": 0.0013, "step": 320 }, { "epoch": 1.0375757575757576, "grad_norm": 0.00512789748609066, "learning_rate": 4.152652005174644e-05, "loss": 0.0012, "step": 321 }, { "epoch": 1.0408080808080808, "grad_norm": 0.006197880953550339, "learning_rate": 4.165588615782665e-05, "loss": 0.0015, "step": 322 }, { "epoch": 1.044040404040404, "grad_norm": 0.005857696291059256, "learning_rate": 4.178525226390686e-05, "loss": 0.0014, "step": 323 }, { "epoch": 1.0472727272727274, "grad_norm": 0.010157118551433086, "learning_rate": 4.1914618369987064e-05, "loss": 0.0015, "step": 324 }, { "epoch": 1.0505050505050506, "grad_norm": 0.006577750667929649, "learning_rate": 4.2043984476067274e-05, "loss": 0.0011, "step": 325 }, { "epoch": 1.0505050505050506, "eval_loss": 0.0012751913163810968, "eval_runtime": 18.6453, "eval_samples_per_second": 5.363, "eval_steps_per_second": 1.341, "step": 325 }, { "epoch": 1.0537373737373736, "grad_norm": 0.004863920155912638, "learning_rate": 4.217335058214748e-05, "loss": 0.0009, "step": 326 }, { "epoch": 1.056969696969697, "grad_norm": 0.005144843365997076, "learning_rate": 4.230271668822768e-05, "loss": 0.0013, "step": 327 }, { "epoch": 1.0602020202020201, "grad_norm": 0.005289267282932997, "learning_rate": 4.243208279430789e-05, "loss": 0.0012, "step": 328 }, { "epoch": 1.0634343434343434, "grad_norm": 0.005019436590373516, "learning_rate": 4.25614489003881e-05, "loss": 0.0011, "step": 329 }, { "epoch": 1.0666666666666667, "grad_norm": 0.0038786993827670813, "learning_rate": 4.269081500646831e-05, "loss": 0.0008, "step": 330 }, { "epoch": 1.06989898989899, "grad_norm": 0.005253227427601814, "learning_rate": 4.2820181112548515e-05, "loss": 0.0012, "step": 331 }, { "epoch": 1.0731313131313132, "grad_norm": 0.005880107171833515, "learning_rate": 4.294954721862872e-05, "loss": 0.0016, "step": 332 }, { "epoch": 1.0763636363636364, "grad_norm": 0.006426130421459675, "learning_rate": 4.307891332470893e-05, "loss": 0.0009, "step": 333 }, { "epoch": 1.0795959595959597, "grad_norm": 0.0038312741089612246, "learning_rate": 4.320827943078913e-05, "loss": 0.0009, "step": 334 }, { "epoch": 1.082828282828283, "grad_norm": 0.005139067303389311, "learning_rate": 4.333764553686934e-05, "loss": 0.0011, "step": 335 }, { "epoch": 1.086060606060606, "grad_norm": 0.006610156502574682, "learning_rate": 4.346701164294955e-05, "loss": 0.0019, "step": 336 }, { "epoch": 1.0892929292929292, "grad_norm": 0.003907322883605957, "learning_rate": 4.359637774902976e-05, "loss": 0.001, "step": 337 }, { "epoch": 1.0925252525252525, "grad_norm": 0.006776052061468363, "learning_rate": 4.372574385510996e-05, "loss": 0.0011, "step": 338 }, { "epoch": 1.0957575757575757, "grad_norm": 0.007625575177371502, "learning_rate": 4.385510996119017e-05, "loss": 0.0016, "step": 339 }, { "epoch": 1.098989898989899, "grad_norm": 0.012086856178939342, "learning_rate": 4.3984476067270374e-05, "loss": 0.0035, "step": 340 }, { "epoch": 1.1022222222222222, "grad_norm": 0.007610958535224199, "learning_rate": 4.4113842173350584e-05, "loss": 0.0014, "step": 341 }, { "epoch": 1.1054545454545455, "grad_norm": 0.006693006958812475, "learning_rate": 4.4243208279430794e-05, "loss": 0.0013, "step": 342 }, { "epoch": 1.1086868686868687, "grad_norm": 0.0045067970640957355, "learning_rate": 4.4372574385511e-05, "loss": 0.0011, "step": 343 }, { "epoch": 1.111919191919192, "grad_norm": 0.006877266336232424, "learning_rate": 4.45019404915912e-05, "loss": 0.001, "step": 344 }, { "epoch": 1.1151515151515152, "grad_norm": 0.0043833632953464985, "learning_rate": 4.463130659767141e-05, "loss": 0.0011, "step": 345 }, { "epoch": 1.1183838383838385, "grad_norm": 0.004103951156139374, "learning_rate": 4.476067270375162e-05, "loss": 0.0009, "step": 346 }, { "epoch": 1.1216161616161617, "grad_norm": 0.006397966295480728, "learning_rate": 4.4890038809831825e-05, "loss": 0.0008, "step": 347 }, { "epoch": 1.1248484848484848, "grad_norm": 0.007373814936727285, "learning_rate": 4.5019404915912036e-05, "loss": 0.0016, "step": 348 }, { "epoch": 1.128080808080808, "grad_norm": 0.005406413692981005, "learning_rate": 4.514877102199224e-05, "loss": 0.0012, "step": 349 }, { "epoch": 1.1313131313131313, "grad_norm": 0.0059000300243496895, "learning_rate": 4.527813712807244e-05, "loss": 0.0011, "step": 350 }, { "epoch": 1.1313131313131313, "eval_loss": 0.0012826790334656835, "eval_runtime": 18.7277, "eval_samples_per_second": 5.34, "eval_steps_per_second": 1.335, "step": 350 }, { "epoch": 1.1345454545454545, "grad_norm": 0.007405865006148815, "learning_rate": 4.540750323415265e-05, "loss": 0.0026, "step": 351 }, { "epoch": 1.1377777777777778, "grad_norm": 0.005577345844358206, "learning_rate": 4.553686934023286e-05, "loss": 0.0018, "step": 352 }, { "epoch": 1.141010101010101, "grad_norm": 0.006122430320829153, "learning_rate": 4.566623544631307e-05, "loss": 0.0015, "step": 353 }, { "epoch": 1.1442424242424243, "grad_norm": 0.007260697893798351, "learning_rate": 4.579560155239328e-05, "loss": 0.0012, "step": 354 }, { "epoch": 1.1474747474747475, "grad_norm": 0.007470736745744944, "learning_rate": 4.592496765847348e-05, "loss": 0.0032, "step": 355 }, { "epoch": 1.1507070707070708, "grad_norm": 0.0061160847544670105, "learning_rate": 4.6054333764553684e-05, "loss": 0.0013, "step": 356 }, { "epoch": 1.1539393939393938, "grad_norm": 0.006069364957511425, "learning_rate": 4.6183699870633894e-05, "loss": 0.0011, "step": 357 }, { "epoch": 1.157171717171717, "grad_norm": 0.004093985538929701, "learning_rate": 4.6313065976714105e-05, "loss": 0.0005, "step": 358 }, { "epoch": 1.1604040404040403, "grad_norm": 0.0061876364052295685, "learning_rate": 4.6442432082794315e-05, "loss": 0.0017, "step": 359 }, { "epoch": 1.1636363636363636, "grad_norm": 0.0048703462816774845, "learning_rate": 4.657179818887452e-05, "loss": 0.0009, "step": 360 }, { "epoch": 1.1668686868686868, "grad_norm": 0.007618950214236975, "learning_rate": 4.670116429495472e-05, "loss": 0.0019, "step": 361 }, { "epoch": 1.17010101010101, "grad_norm": 0.004160857293754816, "learning_rate": 4.683053040103493e-05, "loss": 0.001, "step": 362 }, { "epoch": 1.1733333333333333, "grad_norm": 0.007051107473671436, "learning_rate": 4.6959896507115136e-05, "loss": 0.0019, "step": 363 }, { "epoch": 1.1765656565656566, "grad_norm": 0.004829261917620897, "learning_rate": 4.7089262613195346e-05, "loss": 0.001, "step": 364 }, { "epoch": 1.1797979797979798, "grad_norm": 0.005092508625239134, "learning_rate": 4.7218628719275556e-05, "loss": 0.0023, "step": 365 }, { "epoch": 1.183030303030303, "grad_norm": 0.004417267628014088, "learning_rate": 4.734799482535576e-05, "loss": 0.001, "step": 366 }, { "epoch": 1.1862626262626264, "grad_norm": 0.005750073119997978, "learning_rate": 4.747736093143596e-05, "loss": 0.0013, "step": 367 }, { "epoch": 1.1894949494949496, "grad_norm": 0.0033691064454615116, "learning_rate": 4.760672703751617e-05, "loss": 0.0008, "step": 368 }, { "epoch": 1.1927272727272726, "grad_norm": 0.0044763232581317425, "learning_rate": 4.773609314359638e-05, "loss": 0.0009, "step": 369 }, { "epoch": 1.195959595959596, "grad_norm": 0.007102371193468571, "learning_rate": 4.786545924967659e-05, "loss": 0.0012, "step": 370 }, { "epoch": 1.1991919191919191, "grad_norm": 0.0041382270865142345, "learning_rate": 4.79948253557568e-05, "loss": 0.001, "step": 371 }, { "epoch": 1.2024242424242424, "grad_norm": 0.006266591139137745, "learning_rate": 4.8124191461837e-05, "loss": 0.0015, "step": 372 }, { "epoch": 1.2056565656565656, "grad_norm": 0.00487930653616786, "learning_rate": 4.8253557567917204e-05, "loss": 0.0012, "step": 373 }, { "epoch": 1.208888888888889, "grad_norm": 0.005311232525855303, "learning_rate": 4.8382923673997415e-05, "loss": 0.0014, "step": 374 }, { "epoch": 1.2121212121212122, "grad_norm": 0.004289830103516579, "learning_rate": 4.8512289780077625e-05, "loss": 0.0009, "step": 375 }, { "epoch": 1.2121212121212122, "eval_loss": 0.0011982680298388004, "eval_runtime": 18.7329, "eval_samples_per_second": 5.338, "eval_steps_per_second": 1.335, "step": 375 }, { "epoch": 1.2153535353535354, "grad_norm": 0.0062402524054050446, "learning_rate": 4.864165588615783e-05, "loss": 0.0011, "step": 376 }, { "epoch": 1.2185858585858587, "grad_norm": 0.0034229829907417297, "learning_rate": 4.877102199223804e-05, "loss": 0.0008, "step": 377 }, { "epoch": 1.221818181818182, "grad_norm": 0.0063551426865160465, "learning_rate": 4.890038809831824e-05, "loss": 0.0027, "step": 378 }, { "epoch": 1.225050505050505, "grad_norm": 0.004902714863419533, "learning_rate": 4.9029754204398446e-05, "loss": 0.0015, "step": 379 }, { "epoch": 1.2282828282828282, "grad_norm": 0.0044431439600884914, "learning_rate": 4.9159120310478656e-05, "loss": 0.0005, "step": 380 }, { "epoch": 1.2315151515151515, "grad_norm": 0.007968113757669926, "learning_rate": 4.9288486416558866e-05, "loss": 0.0013, "step": 381 }, { "epoch": 1.2347474747474747, "grad_norm": 0.007964645512402058, "learning_rate": 4.941785252263907e-05, "loss": 0.0033, "step": 382 }, { "epoch": 1.237979797979798, "grad_norm": 0.004984153900295496, "learning_rate": 4.954721862871928e-05, "loss": 0.0008, "step": 383 }, { "epoch": 1.2412121212121212, "grad_norm": 0.005040554329752922, "learning_rate": 4.9676584734799483e-05, "loss": 0.0008, "step": 384 }, { "epoch": 1.2444444444444445, "grad_norm": 0.005269059911370277, "learning_rate": 4.980595084087969e-05, "loss": 0.0016, "step": 385 }, { "epoch": 1.2476767676767677, "grad_norm": 0.004820747766643763, "learning_rate": 4.99353169469599e-05, "loss": 0.0013, "step": 386 }, { "epoch": 1.250909090909091, "grad_norm": 0.003650473430752754, "learning_rate": 5.006468305304011e-05, "loss": 0.001, "step": 387 }, { "epoch": 1.2541414141414142, "grad_norm": 0.003410038072615862, "learning_rate": 5.019404915912032e-05, "loss": 0.0007, "step": 388 }, { "epoch": 1.2573737373737375, "grad_norm": 0.00503958947956562, "learning_rate": 5.032341526520052e-05, "loss": 0.0013, "step": 389 }, { "epoch": 1.2606060606060607, "grad_norm": 0.006551372352987528, "learning_rate": 5.0452781371280725e-05, "loss": 0.0017, "step": 390 }, { "epoch": 1.2638383838383838, "grad_norm": 0.004784473218023777, "learning_rate": 5.058214747736093e-05, "loss": 0.0011, "step": 391 }, { "epoch": 1.267070707070707, "grad_norm": 0.009283591993153095, "learning_rate": 5.071151358344114e-05, "loss": 0.0017, "step": 392 }, { "epoch": 1.2703030303030303, "grad_norm": 0.005054951179772615, "learning_rate": 5.084087968952135e-05, "loss": 0.0014, "step": 393 }, { "epoch": 1.2735353535353535, "grad_norm": 0.007423613220453262, "learning_rate": 5.097024579560156e-05, "loss": 0.0012, "step": 394 }, { "epoch": 1.2767676767676768, "grad_norm": 0.004986994434148073, "learning_rate": 5.109961190168176e-05, "loss": 0.0017, "step": 395 }, { "epoch": 1.28, "grad_norm": 0.004347431473433971, "learning_rate": 5.122897800776197e-05, "loss": 0.0012, "step": 396 }, { "epoch": 1.2832323232323233, "grad_norm": 0.006622066255658865, "learning_rate": 5.135834411384217e-05, "loss": 0.0015, "step": 397 }, { "epoch": 1.2864646464646465, "grad_norm": 0.006094200070947409, "learning_rate": 5.148771021992238e-05, "loss": 0.0014, "step": 398 }, { "epoch": 1.2896969696969696, "grad_norm": 0.004648863337934017, "learning_rate": 5.161707632600259e-05, "loss": 0.0012, "step": 399 }, { "epoch": 1.2929292929292928, "grad_norm": 0.005023527424782515, "learning_rate": 5.17464424320828e-05, "loss": 0.0015, "step": 400 }, { "epoch": 1.2929292929292928, "eval_loss": 0.001097235712222755, "eval_runtime": 18.6509, "eval_samples_per_second": 5.362, "eval_steps_per_second": 1.34, "step": 400 }, { "epoch": 1.296161616161616, "grad_norm": 0.004932883661240339, "learning_rate": 5.1875808538163004e-05, "loss": 0.0012, "step": 401 }, { "epoch": 1.2993939393939393, "grad_norm": 0.004028505180031061, "learning_rate": 5.2005174644243214e-05, "loss": 0.0013, "step": 402 }, { "epoch": 1.3026262626262626, "grad_norm": 0.004800683818757534, "learning_rate": 5.213454075032341e-05, "loss": 0.0014, "step": 403 }, { "epoch": 1.3058585858585858, "grad_norm": 0.0042937519028782845, "learning_rate": 5.226390685640362e-05, "loss": 0.001, "step": 404 }, { "epoch": 1.309090909090909, "grad_norm": 0.004741652403026819, "learning_rate": 5.239327296248383e-05, "loss": 0.0011, "step": 405 }, { "epoch": 1.3123232323232323, "grad_norm": 0.004214128013700247, "learning_rate": 5.252263906856404e-05, "loss": 0.0009, "step": 406 }, { "epoch": 1.3155555555555556, "grad_norm": 0.004764182958751917, "learning_rate": 5.2652005174644245e-05, "loss": 0.0011, "step": 407 }, { "epoch": 1.3187878787878788, "grad_norm": 0.005405279342085123, "learning_rate": 5.2781371280724455e-05, "loss": 0.0011, "step": 408 }, { "epoch": 1.322020202020202, "grad_norm": 0.005355632398277521, "learning_rate": 5.2910737386804666e-05, "loss": 0.0012, "step": 409 }, { "epoch": 1.3252525252525253, "grad_norm": 0.004391840659081936, "learning_rate": 5.304010349288486e-05, "loss": 0.0011, "step": 410 }, { "epoch": 1.3284848484848486, "grad_norm": 0.002719137817621231, "learning_rate": 5.316946959896507e-05, "loss": 0.0005, "step": 411 }, { "epoch": 1.3317171717171719, "grad_norm": 0.005910648964345455, "learning_rate": 5.329883570504528e-05, "loss": 0.0014, "step": 412 }, { "epoch": 1.3349494949494949, "grad_norm": 0.004575495608150959, "learning_rate": 5.3428201811125486e-05, "loss": 0.0013, "step": 413 }, { "epoch": 1.3381818181818181, "grad_norm": 0.004968920256942511, "learning_rate": 5.35575679172057e-05, "loss": 0.0012, "step": 414 }, { "epoch": 1.3414141414141414, "grad_norm": 0.005680674687027931, "learning_rate": 5.368693402328591e-05, "loss": 0.0015, "step": 415 }, { "epoch": 1.3446464646464646, "grad_norm": 0.005805244669318199, "learning_rate": 5.3816300129366104e-05, "loss": 0.0017, "step": 416 }, { "epoch": 1.347878787878788, "grad_norm": 0.004365504253655672, "learning_rate": 5.3945666235446314e-05, "loss": 0.001, "step": 417 }, { "epoch": 1.3511111111111112, "grad_norm": 0.004444981925189495, "learning_rate": 5.4075032341526524e-05, "loss": 0.0011, "step": 418 }, { "epoch": 1.3543434343434344, "grad_norm": 0.0048050908371806145, "learning_rate": 5.420439844760673e-05, "loss": 0.0011, "step": 419 }, { "epoch": 1.3575757575757577, "grad_norm": 0.0112074613571167, "learning_rate": 5.433376455368694e-05, "loss": 0.0009, "step": 420 }, { "epoch": 1.3608080808080807, "grad_norm": 0.005805397406220436, "learning_rate": 5.446313065976715e-05, "loss": 0.001, "step": 421 }, { "epoch": 1.364040404040404, "grad_norm": 0.006793915294110775, "learning_rate": 5.4592496765847345e-05, "loss": 0.0018, "step": 422 }, { "epoch": 1.3672727272727272, "grad_norm": 0.005133500322699547, "learning_rate": 5.4721862871927555e-05, "loss": 0.0015, "step": 423 }, { "epoch": 1.3705050505050504, "grad_norm": 0.0034578884951770306, "learning_rate": 5.4851228978007766e-05, "loss": 0.0008, "step": 424 }, { "epoch": 1.3737373737373737, "grad_norm": 0.005579716991633177, "learning_rate": 5.498059508408797e-05, "loss": 0.0025, "step": 425 }, { "epoch": 1.3737373737373737, "eval_loss": 0.0011115286033600569, "eval_runtime": 18.7263, "eval_samples_per_second": 5.34, "eval_steps_per_second": 1.335, "step": 425 }, { "epoch": 1.376969696969697, "grad_norm": 0.005262942984700203, "learning_rate": 5.510996119016818e-05, "loss": 0.0011, "step": 426 }, { "epoch": 1.3802020202020202, "grad_norm": 0.004813515581190586, "learning_rate": 5.523932729624839e-05, "loss": 0.0011, "step": 427 }, { "epoch": 1.3834343434343435, "grad_norm": 0.005556623917073011, "learning_rate": 5.53686934023286e-05, "loss": 0.0014, "step": 428 }, { "epoch": 1.3866666666666667, "grad_norm": 0.0028185530100017786, "learning_rate": 5.5498059508408797e-05, "loss": 0.0007, "step": 429 }, { "epoch": 1.38989898989899, "grad_norm": 0.004296639934182167, "learning_rate": 5.562742561448901e-05, "loss": 0.001, "step": 430 }, { "epoch": 1.3931313131313132, "grad_norm": 0.004092990420758724, "learning_rate": 5.575679172056921e-05, "loss": 0.001, "step": 431 }, { "epoch": 1.3963636363636365, "grad_norm": 0.005713030230253935, "learning_rate": 5.588615782664942e-05, "loss": 0.0015, "step": 432 }, { "epoch": 1.3995959595959597, "grad_norm": 0.005011020693928003, "learning_rate": 5.601552393272963e-05, "loss": 0.0012, "step": 433 }, { "epoch": 1.4028282828282828, "grad_norm": 0.004393043462187052, "learning_rate": 5.614489003880984e-05, "loss": 0.0009, "step": 434 }, { "epoch": 1.406060606060606, "grad_norm": 0.0047727166675031185, "learning_rate": 5.627425614489004e-05, "loss": 0.0011, "step": 435 }, { "epoch": 1.4092929292929293, "grad_norm": 0.0034502504859119654, "learning_rate": 5.640362225097025e-05, "loss": 0.001, "step": 436 }, { "epoch": 1.4125252525252525, "grad_norm": 0.004155455157160759, "learning_rate": 5.653298835705045e-05, "loss": 0.0008, "step": 437 }, { "epoch": 1.4157575757575758, "grad_norm": 0.005455946549773216, "learning_rate": 5.666235446313066e-05, "loss": 0.0012, "step": 438 }, { "epoch": 1.418989898989899, "grad_norm": 0.005093955434858799, "learning_rate": 5.679172056921087e-05, "loss": 0.0015, "step": 439 }, { "epoch": 1.4222222222222223, "grad_norm": 0.004076706245541573, "learning_rate": 5.692108667529108e-05, "loss": 0.0012, "step": 440 }, { "epoch": 1.4254545454545455, "grad_norm": 0.004360303282737732, "learning_rate": 5.7050452781371286e-05, "loss": 0.001, "step": 441 }, { "epoch": 1.4286868686868686, "grad_norm": 0.005434586200863123, "learning_rate": 5.717981888745149e-05, "loss": 0.002, "step": 442 }, { "epoch": 1.4319191919191918, "grad_norm": 0.006644332781434059, "learning_rate": 5.730918499353169e-05, "loss": 0.0033, "step": 443 }, { "epoch": 1.435151515151515, "grad_norm": 0.006485591642558575, "learning_rate": 5.74385510996119e-05, "loss": 0.002, "step": 444 }, { "epoch": 1.4383838383838383, "grad_norm": 0.006898663938045502, "learning_rate": 5.7567917205692113e-05, "loss": 0.001, "step": 445 }, { "epoch": 1.4416161616161616, "grad_norm": 0.006443925201892853, "learning_rate": 5.7697283311772324e-05, "loss": 0.0015, "step": 446 }, { "epoch": 1.4448484848484848, "grad_norm": 0.003106453223153949, "learning_rate": 5.782664941785253e-05, "loss": 0.0009, "step": 447 }, { "epoch": 1.448080808080808, "grad_norm": 0.0036977077834308147, "learning_rate": 5.795601552393273e-05, "loss": 0.0009, "step": 448 }, { "epoch": 1.4513131313131313, "grad_norm": 0.0050973836332559586, "learning_rate": 5.8085381630012934e-05, "loss": 0.0014, "step": 449 }, { "epoch": 1.4545454545454546, "grad_norm": 0.0057764300145208836, "learning_rate": 5.8214747736093145e-05, "loss": 0.0016, "step": 450 }, { "epoch": 1.4545454545454546, "eval_loss": 0.0010863245697692037, "eval_runtime": 18.7321, "eval_samples_per_second": 5.338, "eval_steps_per_second": 1.335, "step": 450 }, { "epoch": 1.4577777777777778, "grad_norm": 0.004166171886026859, "learning_rate": 5.8344113842173355e-05, "loss": 0.0009, "step": 451 }, { "epoch": 1.461010101010101, "grad_norm": 0.004900315310806036, "learning_rate": 5.8473479948253565e-05, "loss": 0.0013, "step": 452 }, { "epoch": 1.4642424242424243, "grad_norm": 0.003387290518730879, "learning_rate": 5.860284605433377e-05, "loss": 0.0007, "step": 453 }, { "epoch": 1.4674747474747476, "grad_norm": 0.004271840676665306, "learning_rate": 5.873221216041398e-05, "loss": 0.0011, "step": 454 }, { "epoch": 1.4707070707070706, "grad_norm": 0.0032306541688740253, "learning_rate": 5.8861578266494176e-05, "loss": 0.0007, "step": 455 }, { "epoch": 1.4739393939393939, "grad_norm": 0.0022131800651550293, "learning_rate": 5.8990944372574386e-05, "loss": 0.0006, "step": 456 }, { "epoch": 1.4771717171717171, "grad_norm": 0.005533823277801275, "learning_rate": 5.9120310478654596e-05, "loss": 0.0017, "step": 457 }, { "epoch": 1.4804040404040404, "grad_norm": 0.004672409035265446, "learning_rate": 5.9249676584734806e-05, "loss": 0.0014, "step": 458 }, { "epoch": 1.4836363636363636, "grad_norm": 0.006192138884216547, "learning_rate": 5.937904269081501e-05, "loss": 0.0015, "step": 459 }, { "epoch": 1.486868686868687, "grad_norm": 0.011036567389965057, "learning_rate": 5.950840879689522e-05, "loss": 0.0025, "step": 460 }, { "epoch": 1.4901010101010101, "grad_norm": 0.0055129267275333405, "learning_rate": 5.963777490297542e-05, "loss": 0.0015, "step": 461 }, { "epoch": 1.4933333333333334, "grad_norm": 0.005369866266846657, "learning_rate": 5.976714100905563e-05, "loss": 0.0027, "step": 462 }, { "epoch": 1.4965656565656564, "grad_norm": 0.005017601884901524, "learning_rate": 5.989650711513584e-05, "loss": 0.0011, "step": 463 }, { "epoch": 1.4997979797979797, "grad_norm": 0.00623415969312191, "learning_rate": 6.002587322121605e-05, "loss": 0.0028, "step": 464 }, { "epoch": 1.503030303030303, "grad_norm": 0.005604151636362076, "learning_rate": 6.015523932729625e-05, "loss": 0.0023, "step": 465 }, { "epoch": 1.5062626262626262, "grad_norm": 0.004366429056972265, "learning_rate": 6.028460543337646e-05, "loss": 0.001, "step": 466 }, { "epoch": 1.5094949494949494, "grad_norm": 0.004982110112905502, "learning_rate": 6.041397153945667e-05, "loss": 0.0014, "step": 467 }, { "epoch": 1.5127272727272727, "grad_norm": 0.005147982854396105, "learning_rate": 6.054333764553687e-05, "loss": 0.0013, "step": 468 }, { "epoch": 1.515959595959596, "grad_norm": 0.004790551029145718, "learning_rate": 6.067270375161708e-05, "loss": 0.0011, "step": 469 }, { "epoch": 1.5191919191919192, "grad_norm": 0.0038922505918890238, "learning_rate": 6.080206985769729e-05, "loss": 0.0011, "step": 470 }, { "epoch": 1.5224242424242425, "grad_norm": 0.006303661502897739, "learning_rate": 6.093143596377749e-05, "loss": 0.0023, "step": 471 }, { "epoch": 1.5256565656565657, "grad_norm": 0.003573813708499074, "learning_rate": 6.10608020698577e-05, "loss": 0.0011, "step": 472 }, { "epoch": 1.528888888888889, "grad_norm": 0.005556274671107531, "learning_rate": 6.119016817593791e-05, "loss": 0.0033, "step": 473 }, { "epoch": 1.5321212121212122, "grad_norm": 0.004455295857042074, "learning_rate": 6.131953428201811e-05, "loss": 0.0014, "step": 474 }, { "epoch": 1.5353535353535355, "grad_norm": 0.003466435242444277, "learning_rate": 6.144890038809832e-05, "loss": 0.001, "step": 475 }, { "epoch": 1.5353535353535355, "eval_loss": 0.0010681893909350038, "eval_runtime": 18.6082, "eval_samples_per_second": 5.374, "eval_steps_per_second": 1.343, "step": 475 }, { "epoch": 1.5385858585858587, "grad_norm": 0.004641688894480467, "learning_rate": 6.157826649417853e-05, "loss": 0.0029, "step": 476 }, { "epoch": 1.541818181818182, "grad_norm": 0.004081512801349163, "learning_rate": 6.170763260025874e-05, "loss": 0.0011, "step": 477 }, { "epoch": 1.545050505050505, "grad_norm": 0.005640064366161823, "learning_rate": 6.183699870633895e-05, "loss": 0.0024, "step": 478 }, { "epoch": 1.5482828282828283, "grad_norm": 0.0031765501480549574, "learning_rate": 6.196636481241915e-05, "loss": 0.0009, "step": 479 }, { "epoch": 1.5515151515151515, "grad_norm": 0.0049113016575574875, "learning_rate": 6.209573091849934e-05, "loss": 0.0019, "step": 480 }, { "epoch": 1.5547474747474748, "grad_norm": 0.003700861008837819, "learning_rate": 6.222509702457955e-05, "loss": 0.0012, "step": 481 }, { "epoch": 1.557979797979798, "grad_norm": 0.004294991493225098, "learning_rate": 6.235446313065976e-05, "loss": 0.0014, "step": 482 }, { "epoch": 1.561212121212121, "grad_norm": 0.003475453006103635, "learning_rate": 6.248382923673998e-05, "loss": 0.0007, "step": 483 }, { "epoch": 1.5644444444444443, "grad_norm": 0.007765649352222681, "learning_rate": 6.261319534282019e-05, "loss": 0.001, "step": 484 }, { "epoch": 1.5676767676767676, "grad_norm": 0.0032908658031374216, "learning_rate": 6.27425614489004e-05, "loss": 0.0008, "step": 485 }, { "epoch": 1.5709090909090908, "grad_norm": 0.00453177560120821, "learning_rate": 6.28719275549806e-05, "loss": 0.0013, "step": 486 }, { "epoch": 1.574141414141414, "grad_norm": 0.0038091707974672318, "learning_rate": 6.30012936610608e-05, "loss": 0.0008, "step": 487 }, { "epoch": 1.5773737373737373, "grad_norm": 0.004123839549720287, "learning_rate": 6.313065976714101e-05, "loss": 0.0011, "step": 488 }, { "epoch": 1.5806060606060606, "grad_norm": 0.003308449639007449, "learning_rate": 6.326002587322122e-05, "loss": 0.0012, "step": 489 }, { "epoch": 1.5838383838383838, "grad_norm": 0.005438206251710653, "learning_rate": 6.338939197930143e-05, "loss": 0.0023, "step": 490 }, { "epoch": 1.587070707070707, "grad_norm": 0.005823109764605761, "learning_rate": 6.351875808538163e-05, "loss": 0.0014, "step": 491 }, { "epoch": 1.5903030303030303, "grad_norm": 0.0052726129069924355, "learning_rate": 6.364812419146184e-05, "loss": 0.0012, "step": 492 }, { "epoch": 1.5935353535353536, "grad_norm": 0.003648497397080064, "learning_rate": 6.377749029754204e-05, "loss": 0.0013, "step": 493 }, { "epoch": 1.5967676767676768, "grad_norm": 0.003007176099345088, "learning_rate": 6.390685640362225e-05, "loss": 0.0007, "step": 494 }, { "epoch": 1.6, "grad_norm": 0.004037541802972555, "learning_rate": 6.403622250970246e-05, "loss": 0.0011, "step": 495 }, { "epoch": 1.6032323232323233, "grad_norm": 0.0036643114872276783, "learning_rate": 6.416558861578267e-05, "loss": 0.0009, "step": 496 }, { "epoch": 1.6064646464646466, "grad_norm": 0.0038799517787992954, "learning_rate": 6.429495472186288e-05, "loss": 0.001, "step": 497 }, { "epoch": 1.6096969696969698, "grad_norm": 0.002598103601485491, "learning_rate": 6.442432082794309e-05, "loss": 0.0005, "step": 498 }, { "epoch": 1.6129292929292929, "grad_norm": 0.006756095215678215, "learning_rate": 6.45536869340233e-05, "loss": 0.0014, "step": 499 }, { "epoch": 1.6161616161616161, "grad_norm": 0.0038995088543742895, "learning_rate": 6.46830530401035e-05, "loss": 0.0007, "step": 500 }, { "epoch": 1.6161616161616161, "eval_loss": 0.0010785168269649148, "eval_runtime": 18.596, "eval_samples_per_second": 5.378, "eval_steps_per_second": 1.344, "step": 500 }, { "epoch": 1.6193939393939394, "grad_norm": 0.004936838988214731, "learning_rate": 6.48124191461837e-05, "loss": 0.0012, "step": 501 }, { "epoch": 1.6226262626262626, "grad_norm": 0.006074307020753622, "learning_rate": 6.494178525226392e-05, "loss": 0.0022, "step": 502 }, { "epoch": 1.625858585858586, "grad_norm": 0.0044588991440832615, "learning_rate": 6.507115135834411e-05, "loss": 0.0011, "step": 503 }, { "epoch": 1.6290909090909091, "grad_norm": 0.004507533740252256, "learning_rate": 6.520051746442432e-05, "loss": 0.0011, "step": 504 }, { "epoch": 1.6323232323232322, "grad_norm": 0.0048494781367480755, "learning_rate": 6.532988357050453e-05, "loss": 0.0008, "step": 505 }, { "epoch": 1.6355555555555554, "grad_norm": 0.0033034745138138533, "learning_rate": 6.545924967658473e-05, "loss": 0.0007, "step": 506 }, { "epoch": 1.6387878787878787, "grad_norm": 0.0053002117201685905, "learning_rate": 6.558861578266494e-05, "loss": 0.0009, "step": 507 }, { "epoch": 1.642020202020202, "grad_norm": 0.00404641218483448, "learning_rate": 6.571798188874515e-05, "loss": 0.0008, "step": 508 }, { "epoch": 1.6452525252525252, "grad_norm": 0.0036740771029144526, "learning_rate": 6.584734799482536e-05, "loss": 0.0009, "step": 509 }, { "epoch": 1.6484848484848484, "grad_norm": 0.005331697873771191, "learning_rate": 6.597671410090557e-05, "loss": 0.0014, "step": 510 }, { "epoch": 1.6517171717171717, "grad_norm": 0.004965492524206638, "learning_rate": 6.610608020698578e-05, "loss": 0.0019, "step": 511 }, { "epoch": 1.654949494949495, "grad_norm": 0.0037727411836385727, "learning_rate": 6.623544631306599e-05, "loss": 0.001, "step": 512 }, { "epoch": 1.6581818181818182, "grad_norm": 0.0048839072696864605, "learning_rate": 6.636481241914619e-05, "loss": 0.0013, "step": 513 }, { "epoch": 1.6614141414141415, "grad_norm": 0.004607974551618099, "learning_rate": 6.649417852522638e-05, "loss": 0.0017, "step": 514 }, { "epoch": 1.6646464646464647, "grad_norm": 0.004647100809961557, "learning_rate": 6.66235446313066e-05, "loss": 0.0024, "step": 515 }, { "epoch": 1.667878787878788, "grad_norm": 0.004482895601540804, "learning_rate": 6.67529107373868e-05, "loss": 0.0023, "step": 516 }, { "epoch": 1.6711111111111112, "grad_norm": 0.004681951366364956, "learning_rate": 6.688227684346702e-05, "loss": 0.0031, "step": 517 }, { "epoch": 1.6743434343434345, "grad_norm": 0.004061279818415642, "learning_rate": 6.701164294954723e-05, "loss": 0.0012, "step": 518 }, { "epoch": 1.6775757575757577, "grad_norm": 0.005936305969953537, "learning_rate": 6.714100905562742e-05, "loss": 0.0018, "step": 519 }, { "epoch": 1.680808080808081, "grad_norm": 0.003031841479241848, "learning_rate": 6.727037516170763e-05, "loss": 0.0009, "step": 520 }, { "epoch": 1.684040404040404, "grad_norm": 0.0044912113808095455, "learning_rate": 6.739974126778784e-05, "loss": 0.0013, "step": 521 }, { "epoch": 1.6872727272727273, "grad_norm": 0.003882101271301508, "learning_rate": 6.752910737386805e-05, "loss": 0.0014, "step": 522 }, { "epoch": 1.6905050505050505, "grad_norm": 0.0031570433638989925, "learning_rate": 6.765847347994826e-05, "loss": 0.0011, "step": 523 }, { "epoch": 1.6937373737373738, "grad_norm": 0.004184515681117773, "learning_rate": 6.778783958602847e-05, "loss": 0.0015, "step": 524 }, { "epoch": 1.696969696969697, "grad_norm": 0.002800683258101344, "learning_rate": 6.791720569210867e-05, "loss": 0.0008, "step": 525 }, { "epoch": 1.696969696969697, "eval_loss": 0.0010542384115979075, "eval_runtime": 18.5989, "eval_samples_per_second": 5.377, "eval_steps_per_second": 1.344, "step": 525 }, { "epoch": 1.70020202020202, "grad_norm": 0.003550174878910184, "learning_rate": 6.804657179818887e-05, "loss": 0.0014, "step": 526 }, { "epoch": 1.7034343434343433, "grad_norm": 0.0030985972844064236, "learning_rate": 6.817593790426908e-05, "loss": 0.0008, "step": 527 }, { "epoch": 1.7066666666666666, "grad_norm": 0.0048317620530724525, "learning_rate": 6.830530401034929e-05, "loss": 0.0018, "step": 528 }, { "epoch": 1.7098989898989898, "grad_norm": 0.005726094823330641, "learning_rate": 6.84346701164295e-05, "loss": 0.0025, "step": 529 }, { "epoch": 1.713131313131313, "grad_norm": 0.0024808107409626245, "learning_rate": 6.856403622250971e-05, "loss": 0.0007, "step": 530 }, { "epoch": 1.7163636363636363, "grad_norm": 0.003422652604058385, "learning_rate": 6.869340232858992e-05, "loss": 0.001, "step": 531 }, { "epoch": 1.7195959595959596, "grad_norm": 0.0037957008462399244, "learning_rate": 6.882276843467012e-05, "loss": 0.0009, "step": 532 }, { "epoch": 1.7228282828282828, "grad_norm": 0.0028711955528706312, "learning_rate": 6.895213454075033e-05, "loss": 0.0008, "step": 533 }, { "epoch": 1.726060606060606, "grad_norm": 0.002812835620716214, "learning_rate": 6.908150064683054e-05, "loss": 0.0009, "step": 534 }, { "epoch": 1.7292929292929293, "grad_norm": 0.003739473642781377, "learning_rate": 6.921086675291075e-05, "loss": 0.0016, "step": 535 }, { "epoch": 1.7325252525252526, "grad_norm": 0.004807054530829191, "learning_rate": 6.934023285899096e-05, "loss": 0.0023, "step": 536 }, { "epoch": 1.7357575757575758, "grad_norm": 0.004576352424919605, "learning_rate": 6.946959896507115e-05, "loss": 0.0024, "step": 537 }, { "epoch": 1.738989898989899, "grad_norm": 0.0030553669203072786, "learning_rate": 6.959896507115135e-05, "loss": 0.0009, "step": 538 }, { "epoch": 1.7422222222222223, "grad_norm": 0.003879109164699912, "learning_rate": 6.972833117723156e-05, "loss": 0.0012, "step": 539 }, { "epoch": 1.7454545454545456, "grad_norm": 0.0031888741068542004, "learning_rate": 6.985769728331177e-05, "loss": 0.0013, "step": 540 }, { "epoch": 1.7486868686868688, "grad_norm": 0.0037957336753606796, "learning_rate": 6.998706338939198e-05, "loss": 0.0008, "step": 541 }, { "epoch": 1.7519191919191919, "grad_norm": 0.002901956904679537, "learning_rate": 7.011642949547219e-05, "loss": 0.0007, "step": 542 }, { "epoch": 1.7551515151515151, "grad_norm": 0.0029531833715736866, "learning_rate": 7.02457956015524e-05, "loss": 0.0008, "step": 543 }, { "epoch": 1.7583838383838384, "grad_norm": 0.004063300788402557, "learning_rate": 7.037516170763261e-05, "loss": 0.0012, "step": 544 }, { "epoch": 1.7616161616161616, "grad_norm": 0.0041914028115570545, "learning_rate": 7.050452781371281e-05, "loss": 0.001, "step": 545 }, { "epoch": 1.7648484848484849, "grad_norm": 0.004884886089712381, "learning_rate": 7.063389391979302e-05, "loss": 0.0021, "step": 546 }, { "epoch": 1.768080808080808, "grad_norm": 0.005683641415089369, "learning_rate": 7.076326002587323e-05, "loss": 0.0024, "step": 547 }, { "epoch": 1.7713131313131312, "grad_norm": 0.0042356885969638824, "learning_rate": 7.089262613195343e-05, "loss": 0.0015, "step": 548 }, { "epoch": 1.7745454545454544, "grad_norm": 0.005650636274367571, "learning_rate": 7.102199223803364e-05, "loss": 0.0013, "step": 549 }, { "epoch": 1.7777777777777777, "grad_norm": 0.00460450816899538, "learning_rate": 7.115135834411385e-05, "loss": 0.001, "step": 550 }, { "epoch": 1.7777777777777777, "eval_loss": 0.0010012522106990218, "eval_runtime": 18.6807, "eval_samples_per_second": 5.353, "eval_steps_per_second": 1.338, "step": 550 }, { "epoch": 1.781010101010101, "grad_norm": 0.004403329454362392, "learning_rate": 7.128072445019404e-05, "loss": 0.0012, "step": 551 }, { "epoch": 1.7842424242424242, "grad_norm": 0.0032920974772423506, "learning_rate": 7.141009055627425e-05, "loss": 0.0008, "step": 552 }, { "epoch": 1.7874747474747474, "grad_norm": 0.0038138548843562603, "learning_rate": 7.153945666235446e-05, "loss": 0.0011, "step": 553 }, { "epoch": 1.7907070707070707, "grad_norm": 0.00296188285574317, "learning_rate": 7.166882276843467e-05, "loss": 0.0008, "step": 554 }, { "epoch": 1.793939393939394, "grad_norm": 0.0031317227985709906, "learning_rate": 7.179818887451488e-05, "loss": 0.001, "step": 555 }, { "epoch": 1.7971717171717172, "grad_norm": 0.003491588868200779, "learning_rate": 7.19275549805951e-05, "loss": 0.0007, "step": 556 }, { "epoch": 1.8004040404040405, "grad_norm": 0.0031343346927314997, "learning_rate": 7.20569210866753e-05, "loss": 0.0005, "step": 557 }, { "epoch": 1.8036363636363637, "grad_norm": 0.004438635893166065, "learning_rate": 7.21862871927555e-05, "loss": 0.0011, "step": 558 }, { "epoch": 1.806868686868687, "grad_norm": 0.00411741528660059, "learning_rate": 7.231565329883571e-05, "loss": 0.0008, "step": 559 }, { "epoch": 1.8101010101010102, "grad_norm": 0.004590868018567562, "learning_rate": 7.244501940491591e-05, "loss": 0.0009, "step": 560 }, { "epoch": 1.8133333333333335, "grad_norm": 0.00454489141702652, "learning_rate": 7.257438551099612e-05, "loss": 0.001, "step": 561 }, { "epoch": 1.8165656565656567, "grad_norm": 0.004015767015516758, "learning_rate": 7.270375161707633e-05, "loss": 0.0007, "step": 562 }, { "epoch": 1.8197979797979797, "grad_norm": 0.003563391976058483, "learning_rate": 7.283311772315654e-05, "loss": 0.0008, "step": 563 }, { "epoch": 1.823030303030303, "grad_norm": 0.0037166401743888855, "learning_rate": 7.296248382923674e-05, "loss": 0.0009, "step": 564 }, { "epoch": 1.8262626262626263, "grad_norm": 0.004116456024348736, "learning_rate": 7.309184993531695e-05, "loss": 0.0009, "step": 565 }, { "epoch": 1.8294949494949495, "grad_norm": 0.004104538355022669, "learning_rate": 7.322121604139716e-05, "loss": 0.0013, "step": 566 }, { "epoch": 1.8327272727272728, "grad_norm": 0.005509045906364918, "learning_rate": 7.335058214747737e-05, "loss": 0.0023, "step": 567 }, { "epoch": 1.835959595959596, "grad_norm": 0.0035526566207408905, "learning_rate": 7.347994825355758e-05, "loss": 0.0008, "step": 568 }, { "epoch": 1.839191919191919, "grad_norm": 0.003440872300416231, "learning_rate": 7.360931435963779e-05, "loss": 0.0007, "step": 569 }, { "epoch": 1.8424242424242423, "grad_norm": 0.0029369727708399296, "learning_rate": 7.3738680465718e-05, "loss": 0.0008, "step": 570 }, { "epoch": 1.8456565656565656, "grad_norm": 0.004200720228254795, "learning_rate": 7.38680465717982e-05, "loss": 0.0011, "step": 571 }, { "epoch": 1.8488888888888888, "grad_norm": 0.005561929661780596, "learning_rate": 7.399741267787839e-05, "loss": 0.0018, "step": 572 }, { "epoch": 1.852121212121212, "grad_norm": 0.0033285904210060835, "learning_rate": 7.41267787839586e-05, "loss": 0.0009, "step": 573 }, { "epoch": 1.8553535353535353, "grad_norm": 0.004428067244589329, "learning_rate": 7.425614489003881e-05, "loss": 0.0015, "step": 574 }, { "epoch": 1.8585858585858586, "grad_norm": 0.003229588968679309, "learning_rate": 7.438551099611902e-05, "loss": 0.0007, "step": 575 }, { "epoch": 1.8585858585858586, "eval_loss": 0.0009781663538888097, "eval_runtime": 18.7511, "eval_samples_per_second": 5.333, "eval_steps_per_second": 1.333, "step": 575 }, { "epoch": 1.8618181818181818, "grad_norm": 0.003516717813909054, "learning_rate": 7.451487710219923e-05, "loss": 0.0011, "step": 576 }, { "epoch": 1.865050505050505, "grad_norm": 0.002692221663892269, "learning_rate": 7.464424320827943e-05, "loss": 0.0008, "step": 577 }, { "epoch": 1.8682828282828283, "grad_norm": 0.003864881582558155, "learning_rate": 7.477360931435964e-05, "loss": 0.0013, "step": 578 }, { "epoch": 1.8715151515151516, "grad_norm": 0.003865283913910389, "learning_rate": 7.490297542043985e-05, "loss": 0.0015, "step": 579 }, { "epoch": 1.8747474747474748, "grad_norm": 0.0035206254106014967, "learning_rate": 7.503234152652006e-05, "loss": 0.0012, "step": 580 }, { "epoch": 1.877979797979798, "grad_norm": 0.004602121654897928, "learning_rate": 7.516170763260027e-05, "loss": 0.0018, "step": 581 }, { "epoch": 1.8812121212121213, "grad_norm": 0.0029543524142354727, "learning_rate": 7.529107373868047e-05, "loss": 0.0009, "step": 582 }, { "epoch": 1.8844444444444446, "grad_norm": 0.0023640701547265053, "learning_rate": 7.542043984476068e-05, "loss": 0.0007, "step": 583 }, { "epoch": 1.8876767676767678, "grad_norm": 0.0040098559111356735, "learning_rate": 7.554980595084087e-05, "loss": 0.0012, "step": 584 }, { "epoch": 1.8909090909090909, "grad_norm": 0.004274715203791857, "learning_rate": 7.567917205692108e-05, "loss": 0.0015, "step": 585 }, { "epoch": 1.8941414141414141, "grad_norm": 0.0031054418068379164, "learning_rate": 7.58085381630013e-05, "loss": 0.0008, "step": 586 }, { "epoch": 1.8973737373737374, "grad_norm": 0.003954428713768721, "learning_rate": 7.59379042690815e-05, "loss": 0.0009, "step": 587 }, { "epoch": 1.9006060606060606, "grad_norm": 0.003786911489441991, "learning_rate": 7.606727037516171e-05, "loss": 0.0006, "step": 588 }, { "epoch": 1.9038383838383839, "grad_norm": 0.0041464087553322315, "learning_rate": 7.619663648124192e-05, "loss": 0.0011, "step": 589 }, { "epoch": 1.907070707070707, "grad_norm": 0.004568018950521946, "learning_rate": 7.632600258732212e-05, "loss": 0.0019, "step": 590 }, { "epoch": 1.9103030303030302, "grad_norm": 0.003288006642833352, "learning_rate": 7.645536869340233e-05, "loss": 0.0009, "step": 591 }, { "epoch": 1.9135353535353534, "grad_norm": 0.004277890548110008, "learning_rate": 7.658473479948254e-05, "loss": 0.0008, "step": 592 }, { "epoch": 1.9167676767676767, "grad_norm": 0.004379922058433294, "learning_rate": 7.671410090556275e-05, "loss": 0.001, "step": 593 }, { "epoch": 1.92, "grad_norm": 0.003021504729986191, "learning_rate": 7.684346701164295e-05, "loss": 0.0006, "step": 594 }, { "epoch": 1.9232323232323232, "grad_norm": 0.0032772128470242023, "learning_rate": 7.697283311772316e-05, "loss": 0.0008, "step": 595 }, { "epoch": 1.9264646464646464, "grad_norm": 0.004533900413662195, "learning_rate": 7.710219922380336e-05, "loss": 0.0015, "step": 596 }, { "epoch": 1.9296969696969697, "grad_norm": 0.005284740123897791, "learning_rate": 7.723156532988357e-05, "loss": 0.0011, "step": 597 }, { "epoch": 1.932929292929293, "grad_norm": 0.005875818431377411, "learning_rate": 7.736093143596378e-05, "loss": 0.0013, "step": 598 }, { "epoch": 1.9361616161616162, "grad_norm": 0.004483737051486969, "learning_rate": 7.749029754204399e-05, "loss": 0.0009, "step": 599 }, { "epoch": 1.9393939393939394, "grad_norm": 0.004435641225427389, "learning_rate": 7.76196636481242e-05, "loss": 0.0013, "step": 600 }, { "epoch": 1.9393939393939394, "eval_loss": 0.0009266917477361858, "eval_runtime": 18.7646, "eval_samples_per_second": 5.329, "eval_steps_per_second": 1.332, "step": 600 }, { "epoch": 1.9426262626262627, "grad_norm": 0.003876154311001301, "learning_rate": 7.774902975420441e-05, "loss": 0.001, "step": 601 }, { "epoch": 1.945858585858586, "grad_norm": 0.004851337987929583, "learning_rate": 7.787839586028462e-05, "loss": 0.0014, "step": 602 }, { "epoch": 1.9490909090909092, "grad_norm": 0.003333737375214696, "learning_rate": 7.800776196636481e-05, "loss": 0.001, "step": 603 }, { "epoch": 1.9523232323232325, "grad_norm": 0.0032944250851869583, "learning_rate": 7.813712807244502e-05, "loss": 0.0007, "step": 604 }, { "epoch": 1.9555555555555557, "grad_norm": 0.004762569442391396, "learning_rate": 7.826649417852523e-05, "loss": 0.0027, "step": 605 }, { "epoch": 1.9587878787878787, "grad_norm": 0.0037507396191358566, "learning_rate": 7.839586028460543e-05, "loss": 0.0014, "step": 606 }, { "epoch": 1.962020202020202, "grad_norm": 0.005207899492233992, "learning_rate": 7.852522639068564e-05, "loss": 0.0031, "step": 607 }, { "epoch": 1.9652525252525253, "grad_norm": 0.0035236128605902195, "learning_rate": 7.865459249676585e-05, "loss": 0.001, "step": 608 }, { "epoch": 1.9684848484848485, "grad_norm": 0.004186858423054218, "learning_rate": 7.878395860284605e-05, "loss": 0.0012, "step": 609 }, { "epoch": 1.9717171717171718, "grad_norm": 0.0038515774067491293, "learning_rate": 7.891332470892626e-05, "loss": 0.0022, "step": 610 }, { "epoch": 1.9749494949494948, "grad_norm": 0.0030782243702560663, "learning_rate": 7.904269081500647e-05, "loss": 0.0006, "step": 611 }, { "epoch": 1.978181818181818, "grad_norm": 0.005162171553820372, "learning_rate": 7.917205692108668e-05, "loss": 0.0013, "step": 612 }, { "epoch": 1.9814141414141413, "grad_norm": 0.004226456396281719, "learning_rate": 7.930142302716689e-05, "loss": 0.0009, "step": 613 }, { "epoch": 1.9846464646464645, "grad_norm": 0.003615841967985034, "learning_rate": 7.94307891332471e-05, "loss": 0.0022, "step": 614 }, { "epoch": 1.9878787878787878, "grad_norm": 0.00408519571647048, "learning_rate": 7.956015523932731e-05, "loss": 0.0015, "step": 615 }, { "epoch": 1.991111111111111, "grad_norm": 0.0042512728832662106, "learning_rate": 7.968952134540751e-05, "loss": 0.0014, "step": 616 }, { "epoch": 1.9943434343434343, "grad_norm": 0.003178700339049101, "learning_rate": 7.981888745148772e-05, "loss": 0.0008, "step": 617 }, { "epoch": 1.9975757575757576, "grad_norm": 0.003822662867605686, "learning_rate": 7.994825355756791e-05, "loss": 0.001, "step": 618 }, { "epoch": 2.000808080808081, "grad_norm": 0.0046024019829928875, "learning_rate": 8.007761966364812e-05, "loss": 0.0011, "step": 619 }, { "epoch": 2.004040404040404, "grad_norm": 0.0038011916913092136, "learning_rate": 8.020698576972833e-05, "loss": 0.0008, "step": 620 }, { "epoch": 2.0072727272727273, "grad_norm": 0.006909184157848358, "learning_rate": 8.033635187580855e-05, "loss": 0.0026, "step": 621 }, { "epoch": 2.0105050505050506, "grad_norm": 0.005042599979788065, "learning_rate": 8.046571798188874e-05, "loss": 0.0011, "step": 622 }, { "epoch": 2.013737373737374, "grad_norm": 0.004002322442829609, "learning_rate": 8.059508408796895e-05, "loss": 0.0011, "step": 623 }, { "epoch": 2.016969696969697, "grad_norm": 0.0031209783628582954, "learning_rate": 8.072445019404916e-05, "loss": 0.0009, "step": 624 }, { "epoch": 2.0202020202020203, "grad_norm": 0.0026158462278544903, "learning_rate": 8.085381630012937e-05, "loss": 0.0007, "step": 625 }, { "epoch": 2.0202020202020203, "eval_loss": 0.0009721739334054291, "eval_runtime": 18.7204, "eval_samples_per_second": 5.342, "eval_steps_per_second": 1.335, "step": 625 }, { "epoch": 2.0234343434343436, "grad_norm": 0.003943873103708029, "learning_rate": 8.098318240620958e-05, "loss": 0.001, "step": 626 }, { "epoch": 2.026666666666667, "grad_norm": 0.0038578941021114588, "learning_rate": 8.111254851228979e-05, "loss": 0.0017, "step": 627 }, { "epoch": 2.02989898989899, "grad_norm": 0.002991423010826111, "learning_rate": 8.124191461836999e-05, "loss": 0.0007, "step": 628 }, { "epoch": 2.0331313131313133, "grad_norm": 0.002815463813021779, "learning_rate": 8.13712807244502e-05, "loss": 0.0009, "step": 629 }, { "epoch": 2.036363636363636, "grad_norm": 0.0025943731889128685, "learning_rate": 8.15006468305304e-05, "loss": 0.0006, "step": 630 }, { "epoch": 2.0395959595959594, "grad_norm": 0.0029918155632913113, "learning_rate": 8.163001293661061e-05, "loss": 0.0008, "step": 631 }, { "epoch": 2.0428282828282827, "grad_norm": 0.00398919777944684, "learning_rate": 8.175937904269082e-05, "loss": 0.0011, "step": 632 }, { "epoch": 2.046060606060606, "grad_norm": 0.003045483957976103, "learning_rate": 8.188874514877103e-05, "loss": 0.0007, "step": 633 }, { "epoch": 2.049292929292929, "grad_norm": 0.005803203675895929, "learning_rate": 8.201811125485124e-05, "loss": 0.0008, "step": 634 }, { "epoch": 2.0525252525252524, "grad_norm": 0.003953828942030668, "learning_rate": 8.214747736093143e-05, "loss": 0.001, "step": 635 }, { "epoch": 2.0557575757575757, "grad_norm": 0.003973971586674452, "learning_rate": 8.227684346701164e-05, "loss": 0.0006, "step": 636 }, { "epoch": 2.058989898989899, "grad_norm": 0.0025888276286423206, "learning_rate": 8.240620957309186e-05, "loss": 0.0005, "step": 637 }, { "epoch": 2.062222222222222, "grad_norm": 0.003989651799201965, "learning_rate": 8.253557567917207e-05, "loss": 0.0008, "step": 638 }, { "epoch": 2.0654545454545454, "grad_norm": 0.0030330433510243893, "learning_rate": 8.266494178525228e-05, "loss": 0.0009, "step": 639 }, { "epoch": 2.0686868686868687, "grad_norm": 0.0044325897470116615, "learning_rate": 8.279430789133247e-05, "loss": 0.0016, "step": 640 }, { "epoch": 2.071919191919192, "grad_norm": 0.002264161128550768, "learning_rate": 8.292367399741268e-05, "loss": 0.0006, "step": 641 }, { "epoch": 2.075151515151515, "grad_norm": 0.0033110452350229025, "learning_rate": 8.305304010349288e-05, "loss": 0.001, "step": 642 }, { "epoch": 2.0783838383838384, "grad_norm": 0.0029199772980064154, "learning_rate": 8.318240620957309e-05, "loss": 0.0006, "step": 643 }, { "epoch": 2.0816161616161617, "grad_norm": 0.0031028217636048794, "learning_rate": 8.33117723156533e-05, "loss": 0.0007, "step": 644 }, { "epoch": 2.084848484848485, "grad_norm": 0.004011626821011305, "learning_rate": 8.344113842173351e-05, "loss": 0.0009, "step": 645 }, { "epoch": 2.088080808080808, "grad_norm": 0.003449072130024433, "learning_rate": 8.357050452781372e-05, "loss": 0.0009, "step": 646 }, { "epoch": 2.0913131313131315, "grad_norm": 0.003367091529071331, "learning_rate": 8.369987063389393e-05, "loss": 0.001, "step": 647 }, { "epoch": 2.0945454545454547, "grad_norm": 0.003321894910186529, "learning_rate": 8.382923673997413e-05, "loss": 0.0016, "step": 648 }, { "epoch": 2.097777777777778, "grad_norm": 0.003046546597033739, "learning_rate": 8.395860284605434e-05, "loss": 0.0008, "step": 649 }, { "epoch": 2.101010101010101, "grad_norm": 0.0023006205447018147, "learning_rate": 8.408796895213455e-05, "loss": 0.0006, "step": 650 }, { "epoch": 2.101010101010101, "eval_loss": 0.0009433354716747999, "eval_runtime": 18.7374, "eval_samples_per_second": 5.337, "eval_steps_per_second": 1.334, "step": 650 }, { "epoch": 2.1042424242424245, "grad_norm": 0.004013998433947563, "learning_rate": 8.421733505821476e-05, "loss": 0.001, "step": 651 }, { "epoch": 2.1074747474747473, "grad_norm": 0.0035437876358628273, "learning_rate": 8.434670116429496e-05, "loss": 0.0007, "step": 652 }, { "epoch": 2.1107070707070705, "grad_norm": 0.004152386449277401, "learning_rate": 8.447606727037517e-05, "loss": 0.001, "step": 653 }, { "epoch": 2.113939393939394, "grad_norm": 0.0031009165104478598, "learning_rate": 8.460543337645536e-05, "loss": 0.0007, "step": 654 }, { "epoch": 2.117171717171717, "grad_norm": 0.002030389616265893, "learning_rate": 8.473479948253557e-05, "loss": 0.0005, "step": 655 }, { "epoch": 2.1204040404040403, "grad_norm": 0.003986351191997528, "learning_rate": 8.486416558861578e-05, "loss": 0.0011, "step": 656 }, { "epoch": 2.1236363636363635, "grad_norm": 0.0037633986212313175, "learning_rate": 8.499353169469599e-05, "loss": 0.0017, "step": 657 }, { "epoch": 2.126868686868687, "grad_norm": 0.003191509749740362, "learning_rate": 8.51228978007762e-05, "loss": 0.0009, "step": 658 }, { "epoch": 2.13010101010101, "grad_norm": 0.003234416712075472, "learning_rate": 8.525226390685641e-05, "loss": 0.0015, "step": 659 }, { "epoch": 2.1333333333333333, "grad_norm": 0.00314782140776515, "learning_rate": 8.538163001293662e-05, "loss": 0.0008, "step": 660 }, { "epoch": 2.1365656565656566, "grad_norm": 0.004626644309610128, "learning_rate": 8.551099611901682e-05, "loss": 0.0017, "step": 661 }, { "epoch": 2.13979797979798, "grad_norm": 0.0037515638396143913, "learning_rate": 8.564036222509703e-05, "loss": 0.001, "step": 662 }, { "epoch": 2.143030303030303, "grad_norm": 0.0028777304105460644, "learning_rate": 8.576972833117724e-05, "loss": 0.0009, "step": 663 }, { "epoch": 2.1462626262626263, "grad_norm": 0.0035881204530596733, "learning_rate": 8.589909443725744e-05, "loss": 0.0008, "step": 664 }, { "epoch": 2.1494949494949496, "grad_norm": 0.002268304582685232, "learning_rate": 8.602846054333765e-05, "loss": 0.0005, "step": 665 }, { "epoch": 2.152727272727273, "grad_norm": 0.003100323723629117, "learning_rate": 8.615782664941786e-05, "loss": 0.0008, "step": 666 }, { "epoch": 2.155959595959596, "grad_norm": 0.003520137397572398, "learning_rate": 8.628719275549805e-05, "loss": 0.0009, "step": 667 }, { "epoch": 2.1591919191919193, "grad_norm": 0.0042273253202438354, "learning_rate": 8.641655886157827e-05, "loss": 0.0015, "step": 668 }, { "epoch": 2.1624242424242426, "grad_norm": 0.0030146201606839895, "learning_rate": 8.654592496765848e-05, "loss": 0.0009, "step": 669 }, { "epoch": 2.165656565656566, "grad_norm": 0.0055503263138234615, "learning_rate": 8.667529107373869e-05, "loss": 0.0025, "step": 670 }, { "epoch": 2.168888888888889, "grad_norm": 0.0036939766723662615, "learning_rate": 8.68046571798189e-05, "loss": 0.002, "step": 671 }, { "epoch": 2.172121212121212, "grad_norm": 0.0032185425516217947, "learning_rate": 8.69340232858991e-05, "loss": 0.0009, "step": 672 }, { "epoch": 2.175353535353535, "grad_norm": 0.0030356363859027624, "learning_rate": 8.706338939197932e-05, "loss": 0.0007, "step": 673 }, { "epoch": 2.1785858585858584, "grad_norm": 0.006692877039313316, "learning_rate": 8.719275549805951e-05, "loss": 0.0008, "step": 674 }, { "epoch": 2.1818181818181817, "grad_norm": 0.0026474855840206146, "learning_rate": 8.732212160413972e-05, "loss": 0.0007, "step": 675 }, { "epoch": 2.1818181818181817, "eval_loss": 0.0009410877246409655, "eval_runtime": 18.7748, "eval_samples_per_second": 5.326, "eval_steps_per_second": 1.332, "step": 675 }, { "epoch": 2.185050505050505, "grad_norm": 0.003672214224934578, "learning_rate": 8.745148771021992e-05, "loss": 0.0018, "step": 676 }, { "epoch": 2.188282828282828, "grad_norm": 0.0031880387105047703, "learning_rate": 8.758085381630013e-05, "loss": 0.0012, "step": 677 }, { "epoch": 2.1915151515151514, "grad_norm": 0.0033309224527329206, "learning_rate": 8.771021992238034e-05, "loss": 0.001, "step": 678 }, { "epoch": 2.1947474747474747, "grad_norm": 0.003027291502803564, "learning_rate": 8.783958602846055e-05, "loss": 0.0008, "step": 679 }, { "epoch": 2.197979797979798, "grad_norm": 0.0036887172609567642, "learning_rate": 8.796895213454075e-05, "loss": 0.001, "step": 680 }, { "epoch": 2.201212121212121, "grad_norm": 0.003997828811407089, "learning_rate": 8.809831824062096e-05, "loss": 0.0011, "step": 681 }, { "epoch": 2.2044444444444444, "grad_norm": 0.003314135130494833, "learning_rate": 8.822768434670117e-05, "loss": 0.0017, "step": 682 }, { "epoch": 2.2076767676767677, "grad_norm": 0.0036913359072059393, "learning_rate": 8.835705045278138e-05, "loss": 0.0009, "step": 683 }, { "epoch": 2.210909090909091, "grad_norm": 0.0030470637138932943, "learning_rate": 8.848641655886159e-05, "loss": 0.0008, "step": 684 }, { "epoch": 2.214141414141414, "grad_norm": 0.004008392803370953, "learning_rate": 8.86157826649418e-05, "loss": 0.0012, "step": 685 }, { "epoch": 2.2173737373737374, "grad_norm": 0.0055717285722494125, "learning_rate": 8.8745148771022e-05, "loss": 0.0019, "step": 686 }, { "epoch": 2.2206060606060607, "grad_norm": 0.002990501932799816, "learning_rate": 8.88745148771022e-05, "loss": 0.0009, "step": 687 }, { "epoch": 2.223838383838384, "grad_norm": 0.002291604643687606, "learning_rate": 8.90038809831824e-05, "loss": 0.0006, "step": 688 }, { "epoch": 2.227070707070707, "grad_norm": 0.002231605350971222, "learning_rate": 8.913324708926261e-05, "loss": 0.0006, "step": 689 }, { "epoch": 2.2303030303030305, "grad_norm": 0.002650222275406122, "learning_rate": 8.926261319534282e-05, "loss": 0.0007, "step": 690 }, { "epoch": 2.2335353535353537, "grad_norm": 0.0019864251371473074, "learning_rate": 8.939197930142303e-05, "loss": 0.0006, "step": 691 }, { "epoch": 2.236767676767677, "grad_norm": 0.0028250280302017927, "learning_rate": 8.952134540750324e-05, "loss": 0.0006, "step": 692 }, { "epoch": 2.24, "grad_norm": 0.0035562533885240555, "learning_rate": 8.965071151358344e-05, "loss": 0.0018, "step": 693 }, { "epoch": 2.2432323232323235, "grad_norm": 0.0030060771387070417, "learning_rate": 8.978007761966365e-05, "loss": 0.0007, "step": 694 }, { "epoch": 2.2464646464646463, "grad_norm": 0.002113576978445053, "learning_rate": 8.990944372574386e-05, "loss": 0.0005, "step": 695 }, { "epoch": 2.2496969696969695, "grad_norm": 0.002114512724801898, "learning_rate": 9.003880983182407e-05, "loss": 0.0006, "step": 696 }, { "epoch": 2.252929292929293, "grad_norm": 0.0029473064932972193, "learning_rate": 9.016817593790428e-05, "loss": 0.0006, "step": 697 }, { "epoch": 2.256161616161616, "grad_norm": 0.004192824941128492, "learning_rate": 9.029754204398448e-05, "loss": 0.001, "step": 698 }, { "epoch": 2.2593939393939393, "grad_norm": 0.003509392263367772, "learning_rate": 9.042690815006469e-05, "loss": 0.0009, "step": 699 }, { "epoch": 2.2626262626262625, "grad_norm": 0.004787352867424488, "learning_rate": 9.055627425614489e-05, "loss": 0.001, "step": 700 }, { "epoch": 2.2626262626262625, "eval_loss": 0.000903558568097651, "eval_runtime": 18.694, "eval_samples_per_second": 5.349, "eval_steps_per_second": 1.337, "step": 700 }, { "epoch": 2.265858585858586, "grad_norm": 0.0027677167672663927, "learning_rate": 9.06856403622251e-05, "loss": 0.0009, "step": 701 }, { "epoch": 2.269090909090909, "grad_norm": 0.0026491908356547356, "learning_rate": 9.08150064683053e-05, "loss": 0.0007, "step": 702 }, { "epoch": 2.2723232323232323, "grad_norm": 0.0025004090275615454, "learning_rate": 9.094437257438552e-05, "loss": 0.0007, "step": 703 }, { "epoch": 2.2755555555555556, "grad_norm": 0.0028262247797101736, "learning_rate": 9.107373868046573e-05, "loss": 0.0008, "step": 704 }, { "epoch": 2.278787878787879, "grad_norm": 0.003089368110522628, "learning_rate": 9.120310478654594e-05, "loss": 0.0009, "step": 705 }, { "epoch": 2.282020202020202, "grad_norm": 0.002824244322255254, "learning_rate": 9.133247089262613e-05, "loss": 0.0008, "step": 706 }, { "epoch": 2.2852525252525253, "grad_norm": 0.0023714362177997828, "learning_rate": 9.146183699870634e-05, "loss": 0.0007, "step": 707 }, { "epoch": 2.2884848484848486, "grad_norm": 0.0033874395303428173, "learning_rate": 9.159120310478655e-05, "loss": 0.0017, "step": 708 }, { "epoch": 2.291717171717172, "grad_norm": 0.0033185749780386686, "learning_rate": 9.172056921086676e-05, "loss": 0.0017, "step": 709 }, { "epoch": 2.294949494949495, "grad_norm": 0.00360478856600821, "learning_rate": 9.184993531694696e-05, "loss": 0.0009, "step": 710 }, { "epoch": 2.2981818181818183, "grad_norm": 0.0032836326863616705, "learning_rate": 9.197930142302717e-05, "loss": 0.0009, "step": 711 }, { "epoch": 2.3014141414141416, "grad_norm": 0.0029353760182857513, "learning_rate": 9.210866752910737e-05, "loss": 0.0008, "step": 712 }, { "epoch": 2.304646464646465, "grad_norm": 0.004789955448359251, "learning_rate": 9.223803363518758e-05, "loss": 0.0015, "step": 713 }, { "epoch": 2.3078787878787876, "grad_norm": 0.00258410326205194, "learning_rate": 9.236739974126779e-05, "loss": 0.0007, "step": 714 }, { "epoch": 2.311111111111111, "grad_norm": 0.003211255418136716, "learning_rate": 9.2496765847348e-05, "loss": 0.0009, "step": 715 }, { "epoch": 2.314343434343434, "grad_norm": 0.003458111546933651, "learning_rate": 9.262613195342821e-05, "loss": 0.0008, "step": 716 }, { "epoch": 2.3175757575757574, "grad_norm": 0.0020549860782921314, "learning_rate": 9.275549805950842e-05, "loss": 0.0005, "step": 717 }, { "epoch": 2.3208080808080807, "grad_norm": 0.0034006584901362658, "learning_rate": 9.288486416558863e-05, "loss": 0.0011, "step": 718 }, { "epoch": 2.324040404040404, "grad_norm": 0.003893098793923855, "learning_rate": 9.301423027166883e-05, "loss": 0.0008, "step": 719 }, { "epoch": 2.327272727272727, "grad_norm": 0.004374553449451923, "learning_rate": 9.314359637774904e-05, "loss": 0.0014, "step": 720 }, { "epoch": 2.3305050505050504, "grad_norm": 0.004447360523045063, "learning_rate": 9.327296248382925e-05, "loss": 0.0016, "step": 721 }, { "epoch": 2.3337373737373737, "grad_norm": 0.003019913798198104, "learning_rate": 9.340232858990944e-05, "loss": 0.0007, "step": 722 }, { "epoch": 2.336969696969697, "grad_norm": 0.00266702170483768, "learning_rate": 9.353169469598965e-05, "loss": 0.0006, "step": 723 }, { "epoch": 2.34020202020202, "grad_norm": 0.004037540405988693, "learning_rate": 9.366106080206986e-05, "loss": 0.001, "step": 724 }, { "epoch": 2.3434343434343434, "grad_norm": 0.003830693429335952, "learning_rate": 9.379042690815006e-05, "loss": 0.0015, "step": 725 }, { "epoch": 2.3434343434343434, "eval_loss": 0.0009221473592333496, "eval_runtime": 18.682, "eval_samples_per_second": 5.353, "eval_steps_per_second": 1.338, "step": 725 }, { "epoch": 2.3466666666666667, "grad_norm": 0.002714785747230053, "learning_rate": 9.391979301423027e-05, "loss": 0.0007, "step": 726 }, { "epoch": 2.34989898989899, "grad_norm": 0.0034484020434319973, "learning_rate": 9.404915912031048e-05, "loss": 0.0021, "step": 727 }, { "epoch": 2.353131313131313, "grad_norm": 0.0027550137601792812, "learning_rate": 9.417852522639069e-05, "loss": 0.0014, "step": 728 }, { "epoch": 2.3563636363636364, "grad_norm": 0.004323527216911316, "learning_rate": 9.43078913324709e-05, "loss": 0.0021, "step": 729 }, { "epoch": 2.3595959595959597, "grad_norm": 0.0036909414920955896, "learning_rate": 9.443725743855111e-05, "loss": 0.0009, "step": 730 }, { "epoch": 2.362828282828283, "grad_norm": 0.0026773421559482813, "learning_rate": 9.456662354463132e-05, "loss": 0.0007, "step": 731 }, { "epoch": 2.366060606060606, "grad_norm": 0.0027335931081324816, "learning_rate": 9.469598965071152e-05, "loss": 0.0011, "step": 732 }, { "epoch": 2.3692929292929295, "grad_norm": 0.0037821868900209665, "learning_rate": 9.482535575679173e-05, "loss": 0.0009, "step": 733 }, { "epoch": 2.3725252525252527, "grad_norm": 0.0027637695893645287, "learning_rate": 9.495472186287193e-05, "loss": 0.0008, "step": 734 }, { "epoch": 2.375757575757576, "grad_norm": 0.0028088942635804415, "learning_rate": 9.508408796895214e-05, "loss": 0.0008, "step": 735 }, { "epoch": 2.378989898989899, "grad_norm": 0.004683246370404959, "learning_rate": 9.521345407503235e-05, "loss": 0.0011, "step": 736 }, { "epoch": 2.3822222222222225, "grad_norm": 0.0035344541538506746, "learning_rate": 9.534282018111256e-05, "loss": 0.001, "step": 737 }, { "epoch": 2.3854545454545453, "grad_norm": 0.003326730104163289, "learning_rate": 9.547218628719275e-05, "loss": 0.001, "step": 738 }, { "epoch": 2.3886868686868685, "grad_norm": 0.0027331975288689137, "learning_rate": 9.560155239327296e-05, "loss": 0.001, "step": 739 }, { "epoch": 2.391919191919192, "grad_norm": 0.0023180183488875628, "learning_rate": 9.573091849935317e-05, "loss": 0.0005, "step": 740 }, { "epoch": 2.395151515151515, "grad_norm": 0.00545561034232378, "learning_rate": 9.586028460543338e-05, "loss": 0.0019, "step": 741 }, { "epoch": 2.3983838383838383, "grad_norm": 0.002278296509757638, "learning_rate": 9.59896507115136e-05, "loss": 0.0008, "step": 742 }, { "epoch": 2.4016161616161615, "grad_norm": 0.0036029706243425608, "learning_rate": 9.61190168175938e-05, "loss": 0.0011, "step": 743 }, { "epoch": 2.404848484848485, "grad_norm": 0.0028302932623773813, "learning_rate": 9.6248382923674e-05, "loss": 0.001, "step": 744 }, { "epoch": 2.408080808080808, "grad_norm": 0.005623773206025362, "learning_rate": 9.63777490297542e-05, "loss": 0.0025, "step": 745 }, { "epoch": 2.4113131313131313, "grad_norm": 0.003791616763919592, "learning_rate": 9.650711513583441e-05, "loss": 0.001, "step": 746 }, { "epoch": 2.4145454545454546, "grad_norm": 0.003847538959234953, "learning_rate": 9.663648124191462e-05, "loss": 0.002, "step": 747 }, { "epoch": 2.417777777777778, "grad_norm": 0.003144781803712249, "learning_rate": 9.676584734799483e-05, "loss": 0.0008, "step": 748 }, { "epoch": 2.421010101010101, "grad_norm": 0.0030476360116153955, "learning_rate": 9.689521345407504e-05, "loss": 0.002, "step": 749 }, { "epoch": 2.4242424242424243, "grad_norm": 0.023542512208223343, "learning_rate": 9.702457956015525e-05, "loss": 0.0012, "step": 750 }, { "epoch": 2.4242424242424243, "eval_loss": 0.0009755383944138885, "eval_runtime": 18.7174, "eval_samples_per_second": 5.343, "eval_steps_per_second": 1.336, "step": 750 }, { "epoch": 2.4274747474747476, "grad_norm": 0.005616334266960621, "learning_rate": 9.715394566623545e-05, "loss": 0.0009, "step": 751 }, { "epoch": 2.430707070707071, "grad_norm": 0.007717492058873177, "learning_rate": 9.728331177231566e-05, "loss": 0.0008, "step": 752 }, { "epoch": 2.433939393939394, "grad_norm": 0.008256220258772373, "learning_rate": 9.741267787839587e-05, "loss": 0.0011, "step": 753 }, { "epoch": 2.4371717171717173, "grad_norm": 0.005972269922494888, "learning_rate": 9.754204398447608e-05, "loss": 0.0004, "step": 754 }, { "epoch": 2.4404040404040406, "grad_norm": 0.006538939196616411, "learning_rate": 9.767141009055629e-05, "loss": 0.0009, "step": 755 }, { "epoch": 2.443636363636364, "grad_norm": 0.003234976204112172, "learning_rate": 9.780077619663648e-05, "loss": 0.0011, "step": 756 }, { "epoch": 2.4468686868686866, "grad_norm": 0.0033180294558405876, "learning_rate": 9.793014230271668e-05, "loss": 0.0009, "step": 757 }, { "epoch": 2.45010101010101, "grad_norm": 0.0037402757443487644, "learning_rate": 9.805950840879689e-05, "loss": 0.0007, "step": 758 }, { "epoch": 2.453333333333333, "grad_norm": 0.0028039240278303623, "learning_rate": 9.81888745148771e-05, "loss": 0.0008, "step": 759 }, { "epoch": 2.4565656565656564, "grad_norm": 0.0029848841950297356, "learning_rate": 9.831824062095731e-05, "loss": 0.0007, "step": 760 }, { "epoch": 2.4597979797979797, "grad_norm": 0.002809871220961213, "learning_rate": 9.844760672703752e-05, "loss": 0.0007, "step": 761 }, { "epoch": 2.463030303030303, "grad_norm": 0.004293316043913364, "learning_rate": 9.857697283311773e-05, "loss": 0.0012, "step": 762 }, { "epoch": 2.466262626262626, "grad_norm": 0.003171822987496853, "learning_rate": 9.870633893919794e-05, "loss": 0.0008, "step": 763 }, { "epoch": 2.4694949494949494, "grad_norm": 0.0035177527461200953, "learning_rate": 9.883570504527814e-05, "loss": 0.0008, "step": 764 }, { "epoch": 2.4727272727272727, "grad_norm": 0.0037470462266355753, "learning_rate": 9.896507115135835e-05, "loss": 0.0011, "step": 765 }, { "epoch": 2.475959595959596, "grad_norm": 0.004131955560296774, "learning_rate": 9.909443725743856e-05, "loss": 0.0015, "step": 766 }, { "epoch": 2.479191919191919, "grad_norm": 0.002823730930685997, "learning_rate": 9.922380336351877e-05, "loss": 0.0007, "step": 767 }, { "epoch": 2.4824242424242424, "grad_norm": 0.0037240665405988693, "learning_rate": 9.935316946959897e-05, "loss": 0.001, "step": 768 }, { "epoch": 2.4856565656565657, "grad_norm": 0.0038744148332625628, "learning_rate": 9.948253557567918e-05, "loss": 0.0019, "step": 769 }, { "epoch": 2.488888888888889, "grad_norm": 0.003629435319453478, "learning_rate": 9.961190168175937e-05, "loss": 0.0008, "step": 770 }, { "epoch": 2.492121212121212, "grad_norm": 0.004831044003367424, "learning_rate": 9.974126778783958e-05, "loss": 0.0011, "step": 771 }, { "epoch": 2.4953535353535354, "grad_norm": 0.003547330852597952, "learning_rate": 9.98706338939198e-05, "loss": 0.0009, "step": 772 }, { "epoch": 2.4985858585858587, "grad_norm": 0.003263116115704179, "learning_rate": 0.0001, "loss": 0.0008, "step": 773 }, { "epoch": 2.501818181818182, "grad_norm": 0.0030386094003915787, "learning_rate": 9.999999489471233e-05, "loss": 0.0007, "step": 774 }, { "epoch": 2.505050505050505, "grad_norm": 0.0036390256136655807, "learning_rate": 9.99999795788503e-05, "loss": 0.0012, "step": 775 }, { "epoch": 2.505050505050505, "eval_loss": 0.000891694042365998, "eval_runtime": 18.7365, "eval_samples_per_second": 5.337, "eval_steps_per_second": 1.334, "step": 775 }, { "epoch": 2.5082828282828284, "grad_norm": 0.0027272645384073257, "learning_rate": 9.99999540524171e-05, "loss": 0.0006, "step": 776 }, { "epoch": 2.5115151515151517, "grad_norm": 0.004212843254208565, "learning_rate": 9.999991831541789e-05, "loss": 0.0023, "step": 777 }, { "epoch": 2.514747474747475, "grad_norm": 0.0034173503518104553, "learning_rate": 9.999987236786e-05, "loss": 0.0008, "step": 778 }, { "epoch": 2.517979797979798, "grad_norm": 0.0028145266696810722, "learning_rate": 9.999981620975281e-05, "loss": 0.0007, "step": 779 }, { "epoch": 2.5212121212121215, "grad_norm": 0.0030887024477124214, "learning_rate": 9.999974984110779e-05, "loss": 0.0009, "step": 780 }, { "epoch": 2.5244444444444447, "grad_norm": 0.003968552686274052, "learning_rate": 9.999967326193847e-05, "loss": 0.0009, "step": 781 }, { "epoch": 2.5276767676767675, "grad_norm": 0.00264795683324337, "learning_rate": 9.999958647226049e-05, "loss": 0.0006, "step": 782 }, { "epoch": 2.5309090909090908, "grad_norm": 0.004550054203718901, "learning_rate": 9.999948947209162e-05, "loss": 0.002, "step": 783 }, { "epoch": 2.534141414141414, "grad_norm": 0.0030430385377258062, "learning_rate": 9.999938226145161e-05, "loss": 0.0007, "step": 784 }, { "epoch": 2.5373737373737373, "grad_norm": 0.003766052657738328, "learning_rate": 9.999926484036237e-05, "loss": 0.001, "step": 785 }, { "epoch": 2.5406060606060605, "grad_norm": 0.003455114783719182, "learning_rate": 9.999913720884791e-05, "loss": 0.002, "step": 786 }, { "epoch": 2.543838383838384, "grad_norm": 0.002563537796959281, "learning_rate": 9.999899936693426e-05, "loss": 0.0008, "step": 787 }, { "epoch": 2.547070707070707, "grad_norm": 0.0025122894439846277, "learning_rate": 9.99988513146496e-05, "loss": 0.0006, "step": 788 }, { "epoch": 2.5503030303030303, "grad_norm": 0.004589984659105539, "learning_rate": 9.999869305202412e-05, "loss": 0.0015, "step": 789 }, { "epoch": 2.5535353535353535, "grad_norm": 0.0041996450163424015, "learning_rate": 9.999852457909018e-05, "loss": 0.0019, "step": 790 }, { "epoch": 2.556767676767677, "grad_norm": 0.0033383311238139868, "learning_rate": 9.999834589588217e-05, "loss": 0.0013, "step": 791 }, { "epoch": 2.56, "grad_norm": 0.0018662047805264592, "learning_rate": 9.999815700243656e-05, "loss": 0.0006, "step": 792 }, { "epoch": 2.5632323232323233, "grad_norm": 0.0032810529228299856, "learning_rate": 9.999795789879196e-05, "loss": 0.0009, "step": 793 }, { "epoch": 2.5664646464646466, "grad_norm": 0.003441553795710206, "learning_rate": 9.9997748584989e-05, "loss": 0.0013, "step": 794 }, { "epoch": 2.56969696969697, "grad_norm": 0.0025200124364346266, "learning_rate": 9.999752906107042e-05, "loss": 0.0008, "step": 795 }, { "epoch": 2.572929292929293, "grad_norm": 0.0030850358307361603, "learning_rate": 9.999729932708109e-05, "loss": 0.0009, "step": 796 }, { "epoch": 2.5761616161616163, "grad_norm": 0.0029367755632847548, "learning_rate": 9.999705938306789e-05, "loss": 0.0007, "step": 797 }, { "epoch": 2.579393939393939, "grad_norm": 0.003588201245293021, "learning_rate": 9.999680922907982e-05, "loss": 0.0026, "step": 798 }, { "epoch": 2.5826262626262624, "grad_norm": 0.0031253646593540907, "learning_rate": 9.999654886516798e-05, "loss": 0.0009, "step": 799 }, { "epoch": 2.5858585858585856, "grad_norm": 0.0036176196299493313, "learning_rate": 9.999627829138554e-05, "loss": 0.0015, "step": 800 }, { "epoch": 2.5858585858585856, "eval_loss": 0.0009534513228572905, "eval_runtime": 18.6223, "eval_samples_per_second": 5.37, "eval_steps_per_second": 1.342, "step": 800 }, { "epoch": 2.589090909090909, "grad_norm": 0.00325636169873178, "learning_rate": 9.999599750778772e-05, "loss": 0.0007, "step": 801 }, { "epoch": 2.592323232323232, "grad_norm": 0.003968099132180214, "learning_rate": 9.999570651443191e-05, "loss": 0.0011, "step": 802 }, { "epoch": 2.5955555555555554, "grad_norm": 0.00244711060076952, "learning_rate": 9.99954053113775e-05, "loss": 0.0006, "step": 803 }, { "epoch": 2.5987878787878786, "grad_norm": 0.003342804964631796, "learning_rate": 9.9995093898686e-05, "loss": 0.0014, "step": 804 }, { "epoch": 2.602020202020202, "grad_norm": 0.006182527635246515, "learning_rate": 9.999477227642103e-05, "loss": 0.0025, "step": 805 }, { "epoch": 2.605252525252525, "grad_norm": 0.0032223479356616735, "learning_rate": 9.999444044464823e-05, "loss": 0.0007, "step": 806 }, { "epoch": 2.6084848484848484, "grad_norm": 0.002208409830927849, "learning_rate": 9.999409840343539e-05, "loss": 0.0007, "step": 807 }, { "epoch": 2.6117171717171717, "grad_norm": 0.002247196389362216, "learning_rate": 9.999374615285236e-05, "loss": 0.0007, "step": 808 }, { "epoch": 2.614949494949495, "grad_norm": 0.0024588643573224545, "learning_rate": 9.999338369297106e-05, "loss": 0.0005, "step": 809 }, { "epoch": 2.618181818181818, "grad_norm": 0.003213444259017706, "learning_rate": 9.999301102386553e-05, "loss": 0.0008, "step": 810 }, { "epoch": 2.6214141414141414, "grad_norm": 0.0028623330872505903, "learning_rate": 9.999262814561185e-05, "loss": 0.0009, "step": 811 }, { "epoch": 2.6246464646464647, "grad_norm": 0.0031820102594792843, "learning_rate": 9.999223505828821e-05, "loss": 0.001, "step": 812 }, { "epoch": 2.627878787878788, "grad_norm": 0.0017483988776803017, "learning_rate": 9.999183176197491e-05, "loss": 0.0006, "step": 813 }, { "epoch": 2.631111111111111, "grad_norm": 0.0020935633219778538, "learning_rate": 9.999141825675426e-05, "loss": 0.0006, "step": 814 }, { "epoch": 2.6343434343434344, "grad_norm": 0.003117109416052699, "learning_rate": 9.999099454271074e-05, "loss": 0.0008, "step": 815 }, { "epoch": 2.6375757575757577, "grad_norm": 0.0037443467881530523, "learning_rate": 9.999056061993089e-05, "loss": 0.001, "step": 816 }, { "epoch": 2.640808080808081, "grad_norm": 0.0032844438683241606, "learning_rate": 9.999011648850329e-05, "loss": 0.0009, "step": 817 }, { "epoch": 2.644040404040404, "grad_norm": 0.0027299323119223118, "learning_rate": 9.998966214851864e-05, "loss": 0.001, "step": 818 }, { "epoch": 2.6472727272727274, "grad_norm": 0.003157002152875066, "learning_rate": 9.998919760006972e-05, "loss": 0.001, "step": 819 }, { "epoch": 2.6505050505050507, "grad_norm": 0.003021983429789543, "learning_rate": 9.998872284325142e-05, "loss": 0.0011, "step": 820 }, { "epoch": 2.653737373737374, "grad_norm": 0.0036499700509011745, "learning_rate": 9.998823787816066e-05, "loss": 0.001, "step": 821 }, { "epoch": 2.656969696969697, "grad_norm": 0.0031615635380148888, "learning_rate": 9.99877427048965e-05, "loss": 0.0009, "step": 822 }, { "epoch": 2.6602020202020205, "grad_norm": 0.0032300378661602736, "learning_rate": 9.998723732356006e-05, "loss": 0.0009, "step": 823 }, { "epoch": 2.6634343434343437, "grad_norm": 0.00247188750654459, "learning_rate": 9.998672173425452e-05, "loss": 0.0008, "step": 824 }, { "epoch": 2.6666666666666665, "grad_norm": 0.004531141836196184, "learning_rate": 9.998619593708518e-05, "loss": 0.0011, "step": 825 }, { "epoch": 2.6666666666666665, "eval_loss": 0.0008804297540336847, "eval_runtime": 18.6211, "eval_samples_per_second": 5.37, "eval_steps_per_second": 1.343, "step": 825 }, { "epoch": 2.6698989898989898, "grad_norm": 0.0034359728451818228, "learning_rate": 9.998565993215943e-05, "loss": 0.0009, "step": 826 }, { "epoch": 2.673131313131313, "grad_norm": 0.002785489894449711, "learning_rate": 9.998511371958672e-05, "loss": 0.0007, "step": 827 }, { "epoch": 2.6763636363636363, "grad_norm": 0.0025934847071766853, "learning_rate": 9.998455729947858e-05, "loss": 0.0008, "step": 828 }, { "epoch": 2.6795959595959595, "grad_norm": 0.0026367492973804474, "learning_rate": 9.998399067194864e-05, "loss": 0.0008, "step": 829 }, { "epoch": 2.682828282828283, "grad_norm": 0.003330858191475272, "learning_rate": 9.998341383711263e-05, "loss": 0.0007, "step": 830 }, { "epoch": 2.686060606060606, "grad_norm": 0.003155591432005167, "learning_rate": 9.998282679508835e-05, "loss": 0.0017, "step": 831 }, { "epoch": 2.6892929292929293, "grad_norm": 0.0030347639694809914, "learning_rate": 9.998222954599563e-05, "loss": 0.0007, "step": 832 }, { "epoch": 2.6925252525252525, "grad_norm": 0.0032036558259278536, "learning_rate": 9.99816220899565e-05, "loss": 0.0009, "step": 833 }, { "epoch": 2.695757575757576, "grad_norm": 0.003590774955227971, "learning_rate": 9.998100442709497e-05, "loss": 0.0009, "step": 834 }, { "epoch": 2.698989898989899, "grad_norm": 0.002222040668129921, "learning_rate": 9.998037655753717e-05, "loss": 0.0009, "step": 835 }, { "epoch": 2.7022222222222223, "grad_norm": 0.0027599751483649015, "learning_rate": 9.997973848141137e-05, "loss": 0.0009, "step": 836 }, { "epoch": 2.7054545454545456, "grad_norm": 0.00284950970672071, "learning_rate": 9.997909019884781e-05, "loss": 0.0008, "step": 837 }, { "epoch": 2.708686868686869, "grad_norm": 0.0024073810782283545, "learning_rate": 9.99784317099789e-05, "loss": 0.001, "step": 838 }, { "epoch": 2.711919191919192, "grad_norm": 0.0033184473868459463, "learning_rate": 9.997776301493914e-05, "loss": 0.0016, "step": 839 }, { "epoch": 2.7151515151515153, "grad_norm": 0.0023880742955952883, "learning_rate": 9.997708411386501e-05, "loss": 0.0015, "step": 840 }, { "epoch": 2.718383838383838, "grad_norm": 0.0031732122879475355, "learning_rate": 9.997639500689523e-05, "loss": 0.0007, "step": 841 }, { "epoch": 2.7216161616161614, "grad_norm": 0.0036892895586788654, "learning_rate": 9.997569569417049e-05, "loss": 0.002, "step": 842 }, { "epoch": 2.7248484848484846, "grad_norm": 0.002280700486153364, "learning_rate": 9.997498617583358e-05, "loss": 0.0018, "step": 843 }, { "epoch": 2.728080808080808, "grad_norm": 0.003923532087355852, "learning_rate": 9.997426645202943e-05, "loss": 0.0006, "step": 844 }, { "epoch": 2.731313131313131, "grad_norm": 0.0025437732692807913, "learning_rate": 9.9973536522905e-05, "loss": 0.0009, "step": 845 }, { "epoch": 2.7345454545454544, "grad_norm": 0.002573117846623063, "learning_rate": 9.997279638860933e-05, "loss": 0.0005, "step": 846 }, { "epoch": 2.7377777777777776, "grad_norm": 0.0021492561791092157, "learning_rate": 9.99720460492936e-05, "loss": 0.0008, "step": 847 }, { "epoch": 2.741010101010101, "grad_norm": 0.0021558962762355804, "learning_rate": 9.997128550511099e-05, "loss": 0.0006, "step": 848 }, { "epoch": 2.744242424242424, "grad_norm": 0.0031675281934440136, "learning_rate": 9.997051475621687e-05, "loss": 0.001, "step": 849 }, { "epoch": 2.7474747474747474, "grad_norm": 0.003528911853209138, "learning_rate": 9.996973380276857e-05, "loss": 0.0007, "step": 850 }, { "epoch": 2.7474747474747474, "eval_loss": 0.000889240182004869, "eval_runtime": 18.6302, "eval_samples_per_second": 5.368, "eval_steps_per_second": 1.342, "step": 850 }, { "epoch": 2.7507070707070707, "grad_norm": 0.003921836614608765, "learning_rate": 9.996894264492563e-05, "loss": 0.001, "step": 851 }, { "epoch": 2.753939393939394, "grad_norm": 0.002436129143461585, "learning_rate": 9.99681412828496e-05, "loss": 0.0007, "step": 852 }, { "epoch": 2.757171717171717, "grad_norm": 0.002964465180411935, "learning_rate": 9.996732971670408e-05, "loss": 0.0008, "step": 853 }, { "epoch": 2.7604040404040404, "grad_norm": 0.0034572421573102474, "learning_rate": 9.996650794665487e-05, "loss": 0.0007, "step": 854 }, { "epoch": 2.7636363636363637, "grad_norm": 0.0021866310853511095, "learning_rate": 9.996567597286974e-05, "loss": 0.0005, "step": 855 }, { "epoch": 2.766868686868687, "grad_norm": 0.0026556740049272776, "learning_rate": 9.996483379551861e-05, "loss": 0.0014, "step": 856 }, { "epoch": 2.77010101010101, "grad_norm": 0.003023721743375063, "learning_rate": 9.996398141477344e-05, "loss": 0.0007, "step": 857 }, { "epoch": 2.7733333333333334, "grad_norm": 0.0022941052448004484, "learning_rate": 9.996311883080832e-05, "loss": 0.0006, "step": 858 }, { "epoch": 2.7765656565656567, "grad_norm": 0.003141100751236081, "learning_rate": 9.996224604379938e-05, "loss": 0.0007, "step": 859 }, { "epoch": 2.77979797979798, "grad_norm": 0.002024593763053417, "learning_rate": 9.996136305392487e-05, "loss": 0.0006, "step": 860 }, { "epoch": 2.783030303030303, "grad_norm": 0.004872876685112715, "learning_rate": 9.996046986136509e-05, "loss": 0.0026, "step": 861 }, { "epoch": 2.7862626262626264, "grad_norm": 0.002601329004392028, "learning_rate": 9.995956646630246e-05, "loss": 0.0009, "step": 862 }, { "epoch": 2.7894949494949497, "grad_norm": 0.0028132593724876642, "learning_rate": 9.995865286892145e-05, "loss": 0.0009, "step": 863 }, { "epoch": 2.792727272727273, "grad_norm": 0.002662686165422201, "learning_rate": 9.995772906940864e-05, "loss": 0.0007, "step": 864 }, { "epoch": 2.795959595959596, "grad_norm": 0.0027195930015295744, "learning_rate": 9.995679506795264e-05, "loss": 0.0007, "step": 865 }, { "epoch": 2.7991919191919195, "grad_norm": 0.003179864026606083, "learning_rate": 9.995585086474424e-05, "loss": 0.0014, "step": 866 }, { "epoch": 2.8024242424242423, "grad_norm": 0.0034151843283325434, "learning_rate": 9.995489645997622e-05, "loss": 0.0009, "step": 867 }, { "epoch": 2.8056565656565655, "grad_norm": 0.0026823675725609064, "learning_rate": 9.99539318538435e-05, "loss": 0.0007, "step": 868 }, { "epoch": 2.8088888888888888, "grad_norm": 0.0021664374507963657, "learning_rate": 9.995295704654304e-05, "loss": 0.0012, "step": 869 }, { "epoch": 2.812121212121212, "grad_norm": 0.003440326079726219, "learning_rate": 9.995197203827393e-05, "loss": 0.0009, "step": 870 }, { "epoch": 2.8153535353535353, "grad_norm": 0.003980646841228008, "learning_rate": 9.995097682923733e-05, "loss": 0.0018, "step": 871 }, { "epoch": 2.8185858585858585, "grad_norm": 0.003389423480257392, "learning_rate": 9.994997141963644e-05, "loss": 0.001, "step": 872 }, { "epoch": 2.821818181818182, "grad_norm": 0.0027842579875141382, "learning_rate": 9.994895580967658e-05, "loss": 0.0007, "step": 873 }, { "epoch": 2.825050505050505, "grad_norm": 0.0033308009151369333, "learning_rate": 9.994792999956518e-05, "loss": 0.0011, "step": 874 }, { "epoch": 2.8282828282828283, "grad_norm": 0.002514815656468272, "learning_rate": 9.994689398951169e-05, "loss": 0.0009, "step": 875 }, { "epoch": 2.8282828282828283, "eval_loss": 0.0008539481204934418, "eval_runtime": 18.6059, "eval_samples_per_second": 5.375, "eval_steps_per_second": 1.344, "step": 875 }, { "epoch": 2.8315151515151515, "grad_norm": 0.0024293591268360615, "learning_rate": 9.994584777972769e-05, "loss": 0.0008, "step": 876 }, { "epoch": 2.834747474747475, "grad_norm": 0.0035905223339796066, "learning_rate": 9.994479137042683e-05, "loss": 0.002, "step": 877 }, { "epoch": 2.837979797979798, "grad_norm": 0.002918825950473547, "learning_rate": 9.994372476182484e-05, "loss": 0.0009, "step": 878 }, { "epoch": 2.8412121212121213, "grad_norm": 0.0031527606770396233, "learning_rate": 9.994264795413953e-05, "loss": 0.0008, "step": 879 }, { "epoch": 2.8444444444444446, "grad_norm": 0.0035410267300903797, "learning_rate": 9.99415609475908e-05, "loss": 0.0021, "step": 880 }, { "epoch": 2.847676767676768, "grad_norm": 0.0020397889893501997, "learning_rate": 9.994046374240062e-05, "loss": 0.0007, "step": 881 }, { "epoch": 2.850909090909091, "grad_norm": 0.002342833438888192, "learning_rate": 9.993935633879306e-05, "loss": 0.0007, "step": 882 }, { "epoch": 2.854141414141414, "grad_norm": 0.003542246064171195, "learning_rate": 9.993823873699426e-05, "loss": 0.0008, "step": 883 }, { "epoch": 2.857373737373737, "grad_norm": 0.0024878752883523703, "learning_rate": 9.993711093723245e-05, "loss": 0.0008, "step": 884 }, { "epoch": 2.8606060606060604, "grad_norm": 0.0030815114732831717, "learning_rate": 9.993597293973796e-05, "loss": 0.0018, "step": 885 }, { "epoch": 2.8638383838383836, "grad_norm": 0.004479612223803997, "learning_rate": 9.993482474474314e-05, "loss": 0.0009, "step": 886 }, { "epoch": 2.867070707070707, "grad_norm": 0.0033372335601598024, "learning_rate": 9.99336663524825e-05, "loss": 0.0009, "step": 887 }, { "epoch": 2.87030303030303, "grad_norm": 0.0021939794532954693, "learning_rate": 9.993249776319258e-05, "loss": 0.0006, "step": 888 }, { "epoch": 2.8735353535353534, "grad_norm": 0.0028632464818656445, "learning_rate": 9.993131897711202e-05, "loss": 0.0009, "step": 889 }, { "epoch": 2.8767676767676766, "grad_norm": 0.002615751465782523, "learning_rate": 9.993012999448154e-05, "loss": 0.0009, "step": 890 }, { "epoch": 2.88, "grad_norm": 0.0017363366205245256, "learning_rate": 9.992893081554397e-05, "loss": 0.0005, "step": 891 }, { "epoch": 2.883232323232323, "grad_norm": 0.0032067778520286083, "learning_rate": 9.992772144054415e-05, "loss": 0.0009, "step": 892 }, { "epoch": 2.8864646464646464, "grad_norm": 0.0025459511671215296, "learning_rate": 9.992650186972909e-05, "loss": 0.0008, "step": 893 }, { "epoch": 2.8896969696969697, "grad_norm": 0.001871303771622479, "learning_rate": 9.99252721033478e-05, "loss": 0.0008, "step": 894 }, { "epoch": 2.892929292929293, "grad_norm": 0.001998218474909663, "learning_rate": 9.992403214165147e-05, "loss": 0.0005, "step": 895 }, { "epoch": 2.896161616161616, "grad_norm": 0.0023115125950425863, "learning_rate": 9.992278198489327e-05, "loss": 0.0007, "step": 896 }, { "epoch": 2.8993939393939394, "grad_norm": 0.0019588919822126627, "learning_rate": 9.99215216333285e-05, "loss": 0.0007, "step": 897 }, { "epoch": 2.9026262626262627, "grad_norm": 0.001903701457194984, "learning_rate": 9.992025108721454e-05, "loss": 0.0004, "step": 898 }, { "epoch": 2.905858585858586, "grad_norm": 0.0023024296388030052, "learning_rate": 9.991897034681087e-05, "loss": 0.0006, "step": 899 }, { "epoch": 2.909090909090909, "grad_norm": 0.002626319881528616, "learning_rate": 9.9917679412379e-05, "loss": 0.0009, "step": 900 }, { "epoch": 2.909090909090909, "eval_loss": 0.0008496911614201963, "eval_runtime": 18.6252, "eval_samples_per_second": 5.369, "eval_steps_per_second": 1.342, "step": 900 }, { "epoch": 2.9123232323232324, "grad_norm": 0.0029552706982940435, "learning_rate": 9.99163782841826e-05, "loss": 0.001, "step": 901 }, { "epoch": 2.9155555555555557, "grad_norm": 0.002366641303524375, "learning_rate": 9.991506696248731e-05, "loss": 0.0006, "step": 902 }, { "epoch": 2.918787878787879, "grad_norm": 0.0030002701096236706, "learning_rate": 9.991374544756098e-05, "loss": 0.0009, "step": 903 }, { "epoch": 2.922020202020202, "grad_norm": 0.002418822841718793, "learning_rate": 9.991241373967344e-05, "loss": 0.0005, "step": 904 }, { "epoch": 2.9252525252525254, "grad_norm": 0.0027566729113459587, "learning_rate": 9.991107183909664e-05, "loss": 0.002, "step": 905 }, { "epoch": 2.9284848484848487, "grad_norm": 0.003200074890628457, "learning_rate": 9.990971974610466e-05, "loss": 0.0009, "step": 906 }, { "epoch": 2.931717171717172, "grad_norm": 0.00368666322901845, "learning_rate": 9.990835746097356e-05, "loss": 0.0008, "step": 907 }, { "epoch": 2.934949494949495, "grad_norm": 0.0036671683192253113, "learning_rate": 9.990698498398155e-05, "loss": 0.001, "step": 908 }, { "epoch": 2.9381818181818184, "grad_norm": 0.0037938845343887806, "learning_rate": 9.990560231540889e-05, "loss": 0.0007, "step": 909 }, { "epoch": 2.9414141414141413, "grad_norm": 0.0027589588426053524, "learning_rate": 9.990420945553797e-05, "loss": 0.0006, "step": 910 }, { "epoch": 2.9446464646464645, "grad_norm": 0.003152150195091963, "learning_rate": 9.990280640465321e-05, "loss": 0.0008, "step": 911 }, { "epoch": 2.9478787878787878, "grad_norm": 0.002495568012818694, "learning_rate": 9.990139316304112e-05, "loss": 0.0007, "step": 912 }, { "epoch": 2.951111111111111, "grad_norm": 0.0025980276986956596, "learning_rate": 9.989996973099032e-05, "loss": 0.0008, "step": 913 }, { "epoch": 2.9543434343434343, "grad_norm": 0.0013909428380429745, "learning_rate": 9.989853610879147e-05, "loss": 0.0004, "step": 914 }, { "epoch": 2.9575757575757575, "grad_norm": 0.0025712710339576006, "learning_rate": 9.989709229673736e-05, "loss": 0.0008, "step": 915 }, { "epoch": 2.9608080808080808, "grad_norm": 0.0029310788959264755, "learning_rate": 9.98956382951228e-05, "loss": 0.0007, "step": 916 }, { "epoch": 2.964040404040404, "grad_norm": 0.0041606370359659195, "learning_rate": 9.989417410424475e-05, "loss": 0.0027, "step": 917 }, { "epoch": 2.9672727272727273, "grad_norm": 0.001812944421544671, "learning_rate": 9.98926997244022e-05, "loss": 0.0006, "step": 918 }, { "epoch": 2.9705050505050505, "grad_norm": 0.0026405269745737314, "learning_rate": 9.989121515589622e-05, "loss": 0.0011, "step": 919 }, { "epoch": 2.973737373737374, "grad_norm": 0.003111011115834117, "learning_rate": 9.988972039902997e-05, "loss": 0.0009, "step": 920 }, { "epoch": 2.976969696969697, "grad_norm": 0.0038792812265455723, "learning_rate": 9.988821545410874e-05, "loss": 0.0018, "step": 921 }, { "epoch": 2.9802020202020203, "grad_norm": 0.0031774002127349377, "learning_rate": 9.988670032143981e-05, "loss": 0.0011, "step": 922 }, { "epoch": 2.9834343434343435, "grad_norm": 0.002835791325196624, "learning_rate": 9.988517500133262e-05, "loss": 0.0007, "step": 923 }, { "epoch": 2.986666666666667, "grad_norm": 0.003195343306288123, "learning_rate": 9.988363949409865e-05, "loss": 0.0008, "step": 924 }, { "epoch": 2.98989898989899, "grad_norm": 0.004220789764076471, "learning_rate": 9.988209380005144e-05, "loss": 0.001, "step": 925 }, { "epoch": 2.98989898989899, "eval_loss": 0.000867326685693115, "eval_runtime": 18.6533, "eval_samples_per_second": 5.361, "eval_steps_per_second": 1.34, "step": 925 }, { "epoch": 2.993131313131313, "grad_norm": 0.0027963484171777964, "learning_rate": 9.98805379195067e-05, "loss": 0.001, "step": 926 }, { "epoch": 2.996363636363636, "grad_norm": 0.002367103472352028, "learning_rate": 9.987897185278208e-05, "loss": 0.0006, "step": 927 }, { "epoch": 2.9995959595959594, "grad_norm": 0.008955973200500011, "learning_rate": 9.987739560019746e-05, "loss": 0.0042, "step": 928 }, { "epoch": 3.0028282828282826, "grad_norm": 0.0027045756578445435, "learning_rate": 9.987580916207468e-05, "loss": 0.0019, "step": 929 }, { "epoch": 3.006060606060606, "grad_norm": 0.002136245835572481, "learning_rate": 9.987421253873775e-05, "loss": 0.0013, "step": 930 }, { "epoch": 3.009292929292929, "grad_norm": 0.0018668243428692222, "learning_rate": 9.987260573051269e-05, "loss": 0.0006, "step": 931 }, { "epoch": 3.0125252525252524, "grad_norm": 0.0018371654441580176, "learning_rate": 9.987098873772763e-05, "loss": 0.0004, "step": 932 }, { "epoch": 3.0157575757575756, "grad_norm": 0.0020080492831766605, "learning_rate": 9.986936156071278e-05, "loss": 0.0007, "step": 933 }, { "epoch": 3.018989898989899, "grad_norm": 0.0019176624482497573, "learning_rate": 9.986772419980044e-05, "loss": 0.0006, "step": 934 }, { "epoch": 3.022222222222222, "grad_norm": 0.0024317919742316008, "learning_rate": 9.986607665532497e-05, "loss": 0.0006, "step": 935 }, { "epoch": 3.0254545454545454, "grad_norm": 0.002591727999970317, "learning_rate": 9.986441892762281e-05, "loss": 0.0006, "step": 936 }, { "epoch": 3.0286868686868686, "grad_norm": 0.0026875538751482964, "learning_rate": 9.98627510170325e-05, "loss": 0.0008, "step": 937 }, { "epoch": 3.031919191919192, "grad_norm": 0.003272917354479432, "learning_rate": 9.986107292389464e-05, "loss": 0.0007, "step": 938 }, { "epoch": 3.035151515151515, "grad_norm": 0.002936862874776125, "learning_rate": 9.985938464855191e-05, "loss": 0.0008, "step": 939 }, { "epoch": 3.0383838383838384, "grad_norm": 0.0021727578714489937, "learning_rate": 9.985768619134909e-05, "loss": 0.0007, "step": 940 }, { "epoch": 3.0416161616161617, "grad_norm": 0.002251671627163887, "learning_rate": 9.985597755263302e-05, "loss": 0.0006, "step": 941 }, { "epoch": 3.044848484848485, "grad_norm": 0.0035938192158937454, "learning_rate": 9.985425873275263e-05, "loss": 0.0011, "step": 942 }, { "epoch": 3.048080808080808, "grad_norm": 0.0024208389222621918, "learning_rate": 9.98525297320589e-05, "loss": 0.0007, "step": 943 }, { "epoch": 3.0513131313131314, "grad_norm": 0.0019874447025358677, "learning_rate": 9.985079055090493e-05, "loss": 0.0005, "step": 944 }, { "epoch": 3.0545454545454547, "grad_norm": 0.0014314191648736596, "learning_rate": 9.984904118964588e-05, "loss": 0.0004, "step": 945 }, { "epoch": 3.057777777777778, "grad_norm": 0.0019143620738759637, "learning_rate": 9.984728164863898e-05, "loss": 0.0006, "step": 946 }, { "epoch": 3.061010101010101, "grad_norm": 0.003406926291063428, "learning_rate": 9.984551192824355e-05, "loss": 0.0011, "step": 947 }, { "epoch": 3.0642424242424244, "grad_norm": 0.0015149825485423207, "learning_rate": 9.9843732028821e-05, "loss": 0.0004, "step": 948 }, { "epoch": 3.0674747474747477, "grad_norm": 0.0028568189591169357, "learning_rate": 9.98419419507348e-05, "loss": 0.0008, "step": 949 }, { "epoch": 3.0707070707070705, "grad_norm": 0.00383683480322361, "learning_rate": 9.98401416943505e-05, "loss": 0.0006, "step": 950 }, { "epoch": 3.0707070707070705, "eval_loss": 0.0009233049931935966, "eval_runtime": 18.6291, "eval_samples_per_second": 5.368, "eval_steps_per_second": 1.342, "step": 950 }, { "epoch": 3.0739393939393937, "grad_norm": 0.004385147709399462, "learning_rate": 9.983833126003572e-05, "loss": 0.0009, "step": 951 }, { "epoch": 3.077171717171717, "grad_norm": 0.0024896373506635427, "learning_rate": 9.98365106481602e-05, "loss": 0.0006, "step": 952 }, { "epoch": 3.0804040404040403, "grad_norm": 0.002700845478102565, "learning_rate": 9.983467985909573e-05, "loss": 0.0006, "step": 953 }, { "epoch": 3.0836363636363635, "grad_norm": 0.003025912679731846, "learning_rate": 9.983283889321615e-05, "loss": 0.0008, "step": 954 }, { "epoch": 3.0868686868686868, "grad_norm": 0.003908303566277027, "learning_rate": 9.983098775089742e-05, "loss": 0.0017, "step": 955 }, { "epoch": 3.09010101010101, "grad_norm": 0.0035844456870108843, "learning_rate": 9.982912643251757e-05, "loss": 0.0024, "step": 956 }, { "epoch": 3.0933333333333333, "grad_norm": 0.0022256888914853334, "learning_rate": 9.98272549384567e-05, "loss": 0.0005, "step": 957 }, { "epoch": 3.0965656565656565, "grad_norm": 0.0024845930747687817, "learning_rate": 9.982537326909697e-05, "loss": 0.0007, "step": 958 }, { "epoch": 3.0997979797979798, "grad_norm": 0.0021472277585417032, "learning_rate": 9.982348142482269e-05, "loss": 0.0006, "step": 959 }, { "epoch": 3.103030303030303, "grad_norm": 0.0027109517250210047, "learning_rate": 9.982157940602014e-05, "loss": 0.0007, "step": 960 }, { "epoch": 3.1062626262626263, "grad_norm": 0.0016069613629952073, "learning_rate": 9.981966721307778e-05, "loss": 0.0005, "step": 961 }, { "epoch": 3.1094949494949495, "grad_norm": 0.002237373497337103, "learning_rate": 9.981774484638606e-05, "loss": 0.0005, "step": 962 }, { "epoch": 3.112727272727273, "grad_norm": 0.002959214383736253, "learning_rate": 9.981581230633758e-05, "loss": 0.001, "step": 963 }, { "epoch": 3.115959595959596, "grad_norm": 0.0031158181373029947, "learning_rate": 9.981386959332697e-05, "loss": 0.0008, "step": 964 }, { "epoch": 3.1191919191919193, "grad_norm": 0.007239287253469229, "learning_rate": 9.981191670775097e-05, "loss": 0.0037, "step": 965 }, { "epoch": 3.1224242424242425, "grad_norm": 0.0029164832085371017, "learning_rate": 9.980995365000836e-05, "loss": 0.0006, "step": 966 }, { "epoch": 3.125656565656566, "grad_norm": 0.0018486580811440945, "learning_rate": 9.980798042050004e-05, "loss": 0.0004, "step": 967 }, { "epoch": 3.128888888888889, "grad_norm": 0.0032664432656019926, "learning_rate": 9.980599701962896e-05, "loss": 0.0009, "step": 968 }, { "epoch": 3.1321212121212123, "grad_norm": 0.0037746732123196125, "learning_rate": 9.980400344780015e-05, "loss": 0.0015, "step": 969 }, { "epoch": 3.1353535353535356, "grad_norm": 0.002780639799311757, "learning_rate": 9.98019997054207e-05, "loss": 0.0006, "step": 970 }, { "epoch": 3.1385858585858584, "grad_norm": 0.0026836313772946596, "learning_rate": 9.979998579289984e-05, "loss": 0.0006, "step": 971 }, { "epoch": 3.1418181818181816, "grad_norm": 0.002338019199669361, "learning_rate": 9.979796171064881e-05, "loss": 0.0007, "step": 972 }, { "epoch": 3.145050505050505, "grad_norm": 0.0024902368895709515, "learning_rate": 9.979592745908095e-05, "loss": 0.0011, "step": 973 }, { "epoch": 3.148282828282828, "grad_norm": 0.0033587226644158363, "learning_rate": 9.979388303861169e-05, "loss": 0.0006, "step": 974 }, { "epoch": 3.1515151515151514, "grad_norm": 0.002750332234427333, "learning_rate": 9.97918284496585e-05, "loss": 0.0006, "step": 975 }, { "epoch": 3.1515151515151514, "eval_loss": 0.0009188714902848005, "eval_runtime": 18.6436, "eval_samples_per_second": 5.364, "eval_steps_per_second": 1.341, "step": 975 }, { "epoch": 3.1547474747474746, "grad_norm": 0.004522048868238926, "learning_rate": 9.978976369264098e-05, "loss": 0.0009, "step": 976 }, { "epoch": 3.157979797979798, "grad_norm": 0.0020558724645525217, "learning_rate": 9.978768876798075e-05, "loss": 0.0004, "step": 977 }, { "epoch": 3.161212121212121, "grad_norm": 0.0017784403171390295, "learning_rate": 9.978560367610156e-05, "loss": 0.0004, "step": 978 }, { "epoch": 3.1644444444444444, "grad_norm": 0.0039355335757136345, "learning_rate": 9.978350841742919e-05, "loss": 0.0017, "step": 979 }, { "epoch": 3.1676767676767676, "grad_norm": 0.0031672257464379072, "learning_rate": 9.978140299239152e-05, "loss": 0.0017, "step": 980 }, { "epoch": 3.170909090909091, "grad_norm": 0.0048868078738451, "learning_rate": 9.977928740141851e-05, "loss": 0.001, "step": 981 }, { "epoch": 3.174141414141414, "grad_norm": 0.002355673350393772, "learning_rate": 9.977716164494217e-05, "loss": 0.0005, "step": 982 }, { "epoch": 3.1773737373737374, "grad_norm": 0.0035993049386888742, "learning_rate": 9.977502572339664e-05, "loss": 0.001, "step": 983 }, { "epoch": 3.1806060606060607, "grad_norm": 0.00291255209594965, "learning_rate": 9.977287963721804e-05, "loss": 0.0007, "step": 984 }, { "epoch": 3.183838383838384, "grad_norm": 0.002027621492743492, "learning_rate": 9.977072338684469e-05, "loss": 0.0006, "step": 985 }, { "epoch": 3.187070707070707, "grad_norm": 0.0021479360293596983, "learning_rate": 9.976855697271689e-05, "loss": 0.0006, "step": 986 }, { "epoch": 3.1903030303030304, "grad_norm": 0.002248652745038271, "learning_rate": 9.976638039527704e-05, "loss": 0.0005, "step": 987 }, { "epoch": 3.1935353535353537, "grad_norm": 0.002659781137481332, "learning_rate": 9.976419365496963e-05, "loss": 0.0008, "step": 988 }, { "epoch": 3.196767676767677, "grad_norm": 0.0037245461717247963, "learning_rate": 9.976199675224123e-05, "loss": 0.001, "step": 989 }, { "epoch": 3.2, "grad_norm": 0.004409831948578358, "learning_rate": 9.975978968754045e-05, "loss": 0.0022, "step": 990 }, { "epoch": 3.2032323232323234, "grad_norm": 0.0030354245100170374, "learning_rate": 9.975757246131803e-05, "loss": 0.0006, "step": 991 }, { "epoch": 3.2064646464646467, "grad_norm": 0.0026450392324477434, "learning_rate": 9.975534507402671e-05, "loss": 0.0008, "step": 992 }, { "epoch": 3.2096969696969695, "grad_norm": 0.003626378020271659, "learning_rate": 9.975310752612137e-05, "loss": 0.0004, "step": 993 }, { "epoch": 3.2129292929292927, "grad_norm": 0.0026495913043618202, "learning_rate": 9.975085981805897e-05, "loss": 0.0018, "step": 994 }, { "epoch": 3.216161616161616, "grad_norm": 0.0032507823780179024, "learning_rate": 9.974860195029847e-05, "loss": 0.0012, "step": 995 }, { "epoch": 3.2193939393939393, "grad_norm": 0.002977850381284952, "learning_rate": 9.974633392330097e-05, "loss": 0.0006, "step": 996 }, { "epoch": 3.2226262626262625, "grad_norm": 0.002434935886412859, "learning_rate": 9.974405573752965e-05, "loss": 0.0006, "step": 997 }, { "epoch": 3.2258585858585858, "grad_norm": 0.002726243808865547, "learning_rate": 9.974176739344971e-05, "loss": 0.0007, "step": 998 }, { "epoch": 3.229090909090909, "grad_norm": 0.003465299028903246, "learning_rate": 9.973946889152847e-05, "loss": 0.0006, "step": 999 }, { "epoch": 3.2323232323232323, "grad_norm": 0.002784241922199726, "learning_rate": 9.973716023223531e-05, "loss": 0.0007, "step": 1000 }, { "epoch": 3.2323232323232323, "eval_loss": 0.0009049187647178769, "eval_runtime": 18.6293, "eval_samples_per_second": 5.368, "eval_steps_per_second": 1.342, "step": 1000 }, { "epoch": 3.2355555555555555, "grad_norm": 0.0021909528877586126, "learning_rate": 9.97348414160417e-05, "loss": 0.0005, "step": 1001 }, { "epoch": 3.2387878787878788, "grad_norm": 0.003046189434826374, "learning_rate": 9.973251244342114e-05, "loss": 0.0009, "step": 1002 }, { "epoch": 3.242020202020202, "grad_norm": 0.004089404363185167, "learning_rate": 9.973017331484926e-05, "loss": 0.0019, "step": 1003 }, { "epoch": 3.2452525252525253, "grad_norm": 0.002897520549595356, "learning_rate": 9.972782403080372e-05, "loss": 0.0009, "step": 1004 }, { "epoch": 3.2484848484848485, "grad_norm": 0.0027832165360450745, "learning_rate": 9.972546459176425e-05, "loss": 0.0006, "step": 1005 }, { "epoch": 3.251717171717172, "grad_norm": 0.0021642649080604315, "learning_rate": 9.972309499821273e-05, "loss": 0.0005, "step": 1006 }, { "epoch": 3.254949494949495, "grad_norm": 0.0030184059869498014, "learning_rate": 9.972071525063303e-05, "loss": 0.0019, "step": 1007 }, { "epoch": 3.2581818181818183, "grad_norm": 0.0022592165041714907, "learning_rate": 9.971832534951108e-05, "loss": 0.0005, "step": 1008 }, { "epoch": 3.2614141414141415, "grad_norm": 0.0013180490350350738, "learning_rate": 9.9715925295335e-05, "loss": 0.0004, "step": 1009 }, { "epoch": 3.264646464646465, "grad_norm": 0.00403400557115674, "learning_rate": 9.971351508859488e-05, "loss": 0.0007, "step": 1010 }, { "epoch": 3.267878787878788, "grad_norm": 0.003028259379789233, "learning_rate": 9.971109472978288e-05, "loss": 0.0013, "step": 1011 }, { "epoch": 3.2711111111111113, "grad_norm": 0.004238371271640062, "learning_rate": 9.97086642193933e-05, "loss": 0.0008, "step": 1012 }, { "epoch": 3.274343434343434, "grad_norm": 0.003420765744522214, "learning_rate": 9.970622355792247e-05, "loss": 0.002, "step": 1013 }, { "epoch": 3.2775757575757574, "grad_norm": 0.002874986035749316, "learning_rate": 9.970377274586879e-05, "loss": 0.0007, "step": 1014 }, { "epoch": 3.2808080808080806, "grad_norm": 0.001124391914345324, "learning_rate": 9.970131178373277e-05, "loss": 0.0003, "step": 1015 }, { "epoch": 3.284040404040404, "grad_norm": 0.0029891314916312695, "learning_rate": 9.969884067201695e-05, "loss": 0.0007, "step": 1016 }, { "epoch": 3.287272727272727, "grad_norm": 0.0026401756331324577, "learning_rate": 9.969635941122595e-05, "loss": 0.0007, "step": 1017 }, { "epoch": 3.2905050505050504, "grad_norm": 0.001897217589430511, "learning_rate": 9.969386800186649e-05, "loss": 0.0004, "step": 1018 }, { "epoch": 3.2937373737373736, "grad_norm": 0.002726042177528143, "learning_rate": 9.969136644444731e-05, "loss": 0.0008, "step": 1019 }, { "epoch": 3.296969696969697, "grad_norm": 0.0030468441545963287, "learning_rate": 9.968885473947932e-05, "loss": 0.0008, "step": 1020 }, { "epoch": 3.30020202020202, "grad_norm": 0.0016252384521067142, "learning_rate": 9.968633288747539e-05, "loss": 0.0004, "step": 1021 }, { "epoch": 3.3034343434343434, "grad_norm": 0.002642759820446372, "learning_rate": 9.968380088895052e-05, "loss": 0.0007, "step": 1022 }, { "epoch": 3.3066666666666666, "grad_norm": 0.003615351626649499, "learning_rate": 9.968125874442179e-05, "loss": 0.0008, "step": 1023 }, { "epoch": 3.30989898989899, "grad_norm": 0.0026971250772476196, "learning_rate": 9.96787064544083e-05, "loss": 0.0007, "step": 1024 }, { "epoch": 3.313131313131313, "grad_norm": 0.001658923109062016, "learning_rate": 9.96761440194313e-05, "loss": 0.0004, "step": 1025 }, { "epoch": 3.313131313131313, "eval_loss": 0.0008755004382692277, "eval_runtime": 18.6451, "eval_samples_per_second": 5.363, "eval_steps_per_second": 1.341, "step": 1025 }, { "epoch": 3.313131313131313, "step": 1025, "total_flos": 2.5838521494131835e+18, "train_loss": 0.001856716921518943, "train_runtime": 20010.9872, "train_samples_per_second": 12.368, "train_steps_per_second": 0.386 } ], "logging_steps": 1, "max_steps": 7725, "num_input_tokens_seen": 0, "num_train_epochs": 25, "save_steps": 100, "stateful_callbacks": { "EarlyStoppingCallback": { "args": { "early_stopping_patience": 5, "early_stopping_threshold": 0.0 }, "attributes": { "early_stopping_patience_counter": 4 } }, "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 2.5838521494131835e+18, "train_batch_size": 4, "trial_name": null, "trial_params": null }