{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.05060984867655246, "eval_steps": 200, "global_step": 2000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 5.060984867655246e-05, "grad_norm": 5.005204677581787, "learning_rate": 4.999915650252206e-05, "loss": 0.722, "step": 2 }, { "epoch": 0.00010121969735310492, "grad_norm": 4.560483932495117, "learning_rate": 4.9998313005044116e-05, "loss": 0.3625, "step": 4 }, { "epoch": 0.00015182954602965736, "grad_norm": 9.806407928466797, "learning_rate": 4.9997469507566176e-05, "loss": 0.786, "step": 6 }, { "epoch": 0.00020243939470620983, "grad_norm": 11.181573867797852, "learning_rate": 4.999662601008823e-05, "loss": 0.6063, "step": 8 }, { "epoch": 0.0002530492433827623, "grad_norm": 3.7916452884674072, "learning_rate": 4.999578251261029e-05, "loss": 0.4681, "step": 10 }, { "epoch": 0.0003036590920593147, "grad_norm": 10.196318626403809, "learning_rate": 4.999493901513235e-05, "loss": 0.8192, "step": 12 }, { "epoch": 0.0003542689407358672, "grad_norm": 8.210321426391602, "learning_rate": 4.9994095517654403e-05, "loss": 0.7953, "step": 14 }, { "epoch": 0.00040487878941241966, "grad_norm": 8.648380279541016, "learning_rate": 4.9993252020176464e-05, "loss": 0.5404, "step": 16 }, { "epoch": 0.0004554886380889721, "grad_norm": NaN, "learning_rate": 4.999283027143749e-05, "loss": 0.7557, "step": 18 }, { "epoch": 0.0005060984867655246, "grad_norm": 3.2002220153808594, "learning_rate": 4.999198677395955e-05, "loss": 0.5948, "step": 20 }, { "epoch": 0.000556708335442077, "grad_norm": 6.236138343811035, "learning_rate": 4.9991143276481604e-05, "loss": 1.1657, "step": 22 }, { "epoch": 0.0006073181841186294, "grad_norm": 8.113347053527832, "learning_rate": 4.9990299779003664e-05, "loss": 0.6126, "step": 24 }, { "epoch": 0.0006579280327951819, "grad_norm": 6.6634979248046875, "learning_rate": 4.998945628152572e-05, "loss": 0.5794, "step": 26 }, { "epoch": 0.0007085378814717344, "grad_norm": 7.940774440765381, "learning_rate": 4.998861278404778e-05, "loss": 0.3316, "step": 28 }, { "epoch": 0.0007591477301482868, "grad_norm": 7.296051025390625, "learning_rate": 4.998776928656984e-05, "loss": 0.8218, "step": 30 }, { "epoch": 0.0008097575788248393, "grad_norm": 6.318060398101807, "learning_rate": 4.998692578909189e-05, "loss": 0.7266, "step": 32 }, { "epoch": 0.0008603674275013917, "grad_norm": 13.293371200561523, "learning_rate": 4.998608229161395e-05, "loss": 0.96, "step": 34 }, { "epoch": 0.0009109772761779442, "grad_norm": NaN, "learning_rate": 4.998566054287498e-05, "loss": 0.9009, "step": 36 }, { "epoch": 0.0009615871248544967, "grad_norm": 4.929239749908447, "learning_rate": 4.998481704539704e-05, "loss": 0.7153, "step": 38 }, { "epoch": 0.001012196973531049, "grad_norm": 8.825164794921875, "learning_rate": 4.998397354791909e-05, "loss": 0.4486, "step": 40 }, { "epoch": 0.0010628068222076016, "grad_norm": 8.393811225891113, "learning_rate": 4.998313005044115e-05, "loss": 0.7293, "step": 42 }, { "epoch": 0.001113416670884154, "grad_norm": 4.109868049621582, "learning_rate": 4.9982286552963206e-05, "loss": 0.5638, "step": 44 }, { "epoch": 0.0011640265195607066, "grad_norm": 13.631953239440918, "learning_rate": 4.9981443055485266e-05, "loss": 1.0084, "step": 46 }, { "epoch": 0.0012146363682372589, "grad_norm": 9.624829292297363, "learning_rate": 4.9980599558007326e-05, "loss": 0.9545, "step": 48 }, { "epoch": 0.0012652462169138114, "grad_norm": 5.674628257751465, "learning_rate": 4.997975606052938e-05, "loss": 0.5188, "step": 50 }, { "epoch": 0.0013158560655903639, "grad_norm": 6.339033603668213, "learning_rate": 4.997891256305144e-05, "loss": 0.5308, "step": 52 }, { "epoch": 0.0013664659142669164, "grad_norm": 3.895756244659424, "learning_rate": 4.997806906557349e-05, "loss": 0.6042, "step": 54 }, { "epoch": 0.0014170757629434689, "grad_norm": 4.451659679412842, "learning_rate": 4.997722556809555e-05, "loss": 0.4957, "step": 56 }, { "epoch": 0.0014676856116200212, "grad_norm": 5.246840476989746, "learning_rate": 4.9976382070617613e-05, "loss": 0.7031, "step": 58 }, { "epoch": 0.0015182954602965737, "grad_norm": 7.467830657958984, "learning_rate": 4.997553857313967e-05, "loss": 0.8567, "step": 60 }, { "epoch": 0.0015689053089731262, "grad_norm": 6.088858127593994, "learning_rate": 4.997469507566173e-05, "loss": 0.5998, "step": 62 }, { "epoch": 0.0016195151576496787, "grad_norm": 10.719757080078125, "learning_rate": 4.997385157818378e-05, "loss": 0.8178, "step": 64 }, { "epoch": 0.0016701250063262312, "grad_norm": 5.923684120178223, "learning_rate": 4.997300808070584e-05, "loss": 0.9103, "step": 66 }, { "epoch": 0.0017207348550027834, "grad_norm": 8.114771842956543, "learning_rate": 4.99721645832279e-05, "loss": 0.4154, "step": 68 }, { "epoch": 0.001771344703679336, "grad_norm": 9.67381477355957, "learning_rate": 4.9971321085749954e-05, "loss": 0.6067, "step": 70 }, { "epoch": 0.0018219545523558884, "grad_norm": 7.255108833312988, "learning_rate": 4.9970477588272014e-05, "loss": 0.4104, "step": 72 }, { "epoch": 0.001872564401032441, "grad_norm": 6.330812454223633, "learning_rate": 4.996963409079407e-05, "loss": 0.7361, "step": 74 }, { "epoch": 0.0019231742497089934, "grad_norm": 10.727275848388672, "learning_rate": 4.996879059331613e-05, "loss": 0.8339, "step": 76 }, { "epoch": 0.001973784098385546, "grad_norm": 5.6855950355529785, "learning_rate": 4.996794709583819e-05, "loss": 0.6211, "step": 78 }, { "epoch": 0.002024393947062098, "grad_norm": 9.52706527709961, "learning_rate": 4.996710359836024e-05, "loss": 0.6736, "step": 80 }, { "epoch": 0.002075003795738651, "grad_norm": 9.556534767150879, "learning_rate": 4.99662601008823e-05, "loss": 0.5151, "step": 82 }, { "epoch": 0.002125613644415203, "grad_norm": 4.778113842010498, "learning_rate": 4.9965416603404355e-05, "loss": 0.5348, "step": 84 }, { "epoch": 0.0021762234930917555, "grad_norm": 10.250826835632324, "learning_rate": 4.9964573105926416e-05, "loss": 0.5674, "step": 86 }, { "epoch": 0.002226833341768308, "grad_norm": 6.03010892868042, "learning_rate": 4.9963729608448476e-05, "loss": 0.4641, "step": 88 }, { "epoch": 0.0022774431904448605, "grad_norm": 10.152463912963867, "learning_rate": 4.996288611097053e-05, "loss": 1.0214, "step": 90 }, { "epoch": 0.002328053039121413, "grad_norm": 8.930192947387695, "learning_rate": 4.996204261349259e-05, "loss": 0.7772, "step": 92 }, { "epoch": 0.0023786628877979655, "grad_norm": 10.251153945922852, "learning_rate": 4.996119911601464e-05, "loss": 0.9428, "step": 94 }, { "epoch": 0.0024292727364745178, "grad_norm": 6.223593711853027, "learning_rate": 4.99603556185367e-05, "loss": 0.6318, "step": 96 }, { "epoch": 0.0024798825851510705, "grad_norm": 2.4227373600006104, "learning_rate": 4.995951212105876e-05, "loss": 0.436, "step": 98 }, { "epoch": 0.0025304924338276228, "grad_norm": 6.225770473480225, "learning_rate": 4.9958668623580817e-05, "loss": 0.7753, "step": 100 }, { "epoch": 0.0025811022825041755, "grad_norm": 5.370884895324707, "learning_rate": 4.995782512610288e-05, "loss": 0.5137, "step": 102 }, { "epoch": 0.0026317121311807278, "grad_norm": 5.137220859527588, "learning_rate": 4.995698162862493e-05, "loss": 0.5953, "step": 104 }, { "epoch": 0.00268232197985728, "grad_norm": 5.909286975860596, "learning_rate": 4.995613813114699e-05, "loss": 0.8687, "step": 106 }, { "epoch": 0.0027329318285338328, "grad_norm": 6.003445148468018, "learning_rate": 4.995529463366905e-05, "loss": 0.53, "step": 108 }, { "epoch": 0.002783541677210385, "grad_norm": 3.8742566108703613, "learning_rate": 4.9954451136191104e-05, "loss": 0.5512, "step": 110 }, { "epoch": 0.0028341515258869378, "grad_norm": 3.5072569847106934, "learning_rate": 4.9953607638713164e-05, "loss": 0.3383, "step": 112 }, { "epoch": 0.00288476137456349, "grad_norm": 5.679248809814453, "learning_rate": 4.995276414123522e-05, "loss": 0.7801, "step": 114 }, { "epoch": 0.0029353712232400423, "grad_norm": 5.291722297668457, "learning_rate": 4.995192064375728e-05, "loss": 0.6538, "step": 116 }, { "epoch": 0.002985981071916595, "grad_norm": 4.748198986053467, "learning_rate": 4.995107714627934e-05, "loss": 0.398, "step": 118 }, { "epoch": 0.0030365909205931473, "grad_norm": 4.336450576782227, "learning_rate": 4.995023364880139e-05, "loss": 0.3351, "step": 120 }, { "epoch": 0.0030872007692697, "grad_norm": 5.251417636871338, "learning_rate": 4.994939015132345e-05, "loss": 0.4014, "step": 122 }, { "epoch": 0.0031378106179462523, "grad_norm": 7.911617755889893, "learning_rate": 4.9948546653845505e-05, "loss": 0.3151, "step": 124 }, { "epoch": 0.0031884204666228046, "grad_norm": 6.031822681427002, "learning_rate": 4.9947703156367565e-05, "loss": 0.9222, "step": 126 }, { "epoch": 0.0032390303152993573, "grad_norm": 10.08216381072998, "learning_rate": 4.994685965888962e-05, "loss": 0.6676, "step": 128 }, { "epoch": 0.0032896401639759096, "grad_norm": 3.475123167037964, "learning_rate": 4.994601616141168e-05, "loss": 0.2461, "step": 130 }, { "epoch": 0.0033402500126524623, "grad_norm": 4.408476829528809, "learning_rate": 4.994517266393374e-05, "loss": 0.26, "step": 132 }, { "epoch": 0.0033908598613290146, "grad_norm": 6.929357528686523, "learning_rate": 4.994432916645579e-05, "loss": 0.9818, "step": 134 }, { "epoch": 0.003441469710005567, "grad_norm": 6.456033229827881, "learning_rate": 4.994348566897785e-05, "loss": 0.4414, "step": 136 }, { "epoch": 0.0034920795586821196, "grad_norm": 26.87461280822754, "learning_rate": 4.9942642171499906e-05, "loss": 0.7951, "step": 138 }, { "epoch": 0.003542689407358672, "grad_norm": 7.420943737030029, "learning_rate": 4.9941798674021966e-05, "loss": 0.5827, "step": 140 }, { "epoch": 0.0035932992560352246, "grad_norm": 5.9366583824157715, "learning_rate": 4.9940955176544027e-05, "loss": 0.6941, "step": 142 }, { "epoch": 0.003643909104711777, "grad_norm": 5.8037543296813965, "learning_rate": 4.994011167906608e-05, "loss": 0.6301, "step": 144 }, { "epoch": 0.003694518953388329, "grad_norm": 4.243365287780762, "learning_rate": 4.993926818158814e-05, "loss": 0.6995, "step": 146 }, { "epoch": 0.003745128802064882, "grad_norm": 7.5515522956848145, "learning_rate": 4.9938424684110194e-05, "loss": 0.5537, "step": 148 }, { "epoch": 0.003795738650741434, "grad_norm": 8.966812133789062, "learning_rate": 4.9937581186632254e-05, "loss": 0.5416, "step": 150 }, { "epoch": 0.003846348499417987, "grad_norm": 7.70654296875, "learning_rate": 4.9936737689154314e-05, "loss": 0.6907, "step": 152 }, { "epoch": 0.003896958348094539, "grad_norm": 6.371860504150391, "learning_rate": 4.993589419167637e-05, "loss": 0.468, "step": 154 }, { "epoch": 0.003947568196771092, "grad_norm": 6.1000542640686035, "learning_rate": 4.993505069419843e-05, "loss": 0.6523, "step": 156 }, { "epoch": 0.003998178045447644, "grad_norm": 5.390421390533447, "learning_rate": 4.993420719672048e-05, "loss": 1.0799, "step": 158 }, { "epoch": 0.004048787894124196, "grad_norm": 7.502729415893555, "learning_rate": 4.993336369924254e-05, "loss": 0.7921, "step": 160 }, { "epoch": 0.004099397742800749, "grad_norm": 3.19514536857605, "learning_rate": 4.99325202017646e-05, "loss": 0.2293, "step": 162 }, { "epoch": 0.004150007591477302, "grad_norm": 7.2197585105896, "learning_rate": 4.9931676704286655e-05, "loss": 0.6778, "step": 164 }, { "epoch": 0.004200617440153854, "grad_norm": 4.662430286407471, "learning_rate": 4.9930833206808715e-05, "loss": 0.4956, "step": 166 }, { "epoch": 0.004251227288830406, "grad_norm": 5.545823574066162, "learning_rate": 4.992998970933077e-05, "loss": 0.4587, "step": 168 }, { "epoch": 0.004301837137506959, "grad_norm": 8.578537940979004, "learning_rate": 4.992914621185283e-05, "loss": 0.2751, "step": 170 }, { "epoch": 0.004352446986183511, "grad_norm": 4.3413472175598145, "learning_rate": 4.992830271437489e-05, "loss": 0.3712, "step": 172 }, { "epoch": 0.004403056834860064, "grad_norm": 3.3616743087768555, "learning_rate": 4.992745921689694e-05, "loss": 0.4654, "step": 174 }, { "epoch": 0.004453666683536616, "grad_norm": 3.7586441040039062, "learning_rate": 4.9926615719419e-05, "loss": 0.5338, "step": 176 }, { "epoch": 0.004504276532213168, "grad_norm": 4.674256801605225, "learning_rate": 4.9925772221941056e-05, "loss": 0.3885, "step": 178 }, { "epoch": 0.004554886380889721, "grad_norm": 6.164566993713379, "learning_rate": 4.9924928724463116e-05, "loss": 0.3949, "step": 180 }, { "epoch": 0.004605496229566274, "grad_norm": 12.5003023147583, "learning_rate": 4.9924085226985176e-05, "loss": 0.6829, "step": 182 }, { "epoch": 0.004656106078242826, "grad_norm": 8.045877456665039, "learning_rate": 4.992324172950723e-05, "loss": 0.7711, "step": 184 }, { "epoch": 0.004706715926919378, "grad_norm": 5.4890570640563965, "learning_rate": 4.992239823202929e-05, "loss": 0.6287, "step": 186 }, { "epoch": 0.004757325775595931, "grad_norm": 3.246534585952759, "learning_rate": 4.992155473455134e-05, "loss": 0.3895, "step": 188 }, { "epoch": 0.004807935624272484, "grad_norm": 7.648909568786621, "learning_rate": 4.9920711237073403e-05, "loss": 0.7327, "step": 190 }, { "epoch": 0.0048585454729490355, "grad_norm": 5.417934894561768, "learning_rate": 4.9919867739595464e-05, "loss": 0.5529, "step": 192 }, { "epoch": 0.004909155321625588, "grad_norm": 18.588844299316406, "learning_rate": 4.991902424211752e-05, "loss": 1.1181, "step": 194 }, { "epoch": 0.004959765170302141, "grad_norm": 11.765092849731445, "learning_rate": 4.991818074463958e-05, "loss": 0.7821, "step": 196 }, { "epoch": 0.005010375018978693, "grad_norm": 6.446319580078125, "learning_rate": 4.991733724716163e-05, "loss": 0.7071, "step": 198 }, { "epoch": 0.0050609848676552455, "grad_norm": 7.347968101501465, "learning_rate": 4.991649374968369e-05, "loss": 0.4164, "step": 200 }, { "epoch": 0.0050609848676552455, "eval_cer": 0.23587196123692475, "eval_loss": 0.37637796998023987, "eval_runtime": 2775.7084, "eval_samples_per_second": 5.647, "eval_steps_per_second": 0.353, "step": 200 }, { "epoch": 0.005111594716331798, "grad_norm": 10.311556816101074, "learning_rate": 4.991565025220575e-05, "loss": 1.1246, "step": 202 }, { "epoch": 0.005162204565008351, "grad_norm": 10.519822120666504, "learning_rate": 4.9914806754727805e-05, "loss": 0.6127, "step": 204 }, { "epoch": 0.005212814413684903, "grad_norm": 6.434566974639893, "learning_rate": 4.9913963257249865e-05, "loss": 0.5269, "step": 206 }, { "epoch": 0.0052634242623614555, "grad_norm": 8.08187198638916, "learning_rate": 4.991311975977192e-05, "loss": 1.0978, "step": 208 }, { "epoch": 0.005314034111038008, "grad_norm": 7.574239730834961, "learning_rate": 4.991227626229398e-05, "loss": 0.7327, "step": 210 }, { "epoch": 0.00536464395971456, "grad_norm": 7.921819686889648, "learning_rate": 4.991143276481603e-05, "loss": 1.194, "step": 212 }, { "epoch": 0.005415253808391113, "grad_norm": 9.017950057983398, "learning_rate": 4.991058926733809e-05, "loss": 0.9906, "step": 214 }, { "epoch": 0.0054658636570676655, "grad_norm": 6.131129741668701, "learning_rate": 4.990974576986015e-05, "loss": 0.8681, "step": 216 }, { "epoch": 0.005516473505744218, "grad_norm": 6.411858081817627, "learning_rate": 4.9908902272382206e-05, "loss": 0.6271, "step": 218 }, { "epoch": 0.00556708335442077, "grad_norm": 6.6365227699279785, "learning_rate": 4.9908058774904266e-05, "loss": 0.9051, "step": 220 }, { "epoch": 0.005617693203097323, "grad_norm": 7.575653553009033, "learning_rate": 4.990721527742632e-05, "loss": 0.9199, "step": 222 }, { "epoch": 0.0056683030517738755, "grad_norm": 8.839277267456055, "learning_rate": 4.990637177994838e-05, "loss": 0.909, "step": 224 }, { "epoch": 0.005718912900450427, "grad_norm": 8.077840805053711, "learning_rate": 4.990552828247044e-05, "loss": 0.6063, "step": 226 }, { "epoch": 0.00576952274912698, "grad_norm": 8.73996639251709, "learning_rate": 4.990468478499249e-05, "loss": 0.9447, "step": 228 }, { "epoch": 0.005820132597803533, "grad_norm": 6.638965129852295, "learning_rate": 4.990384128751455e-05, "loss": 0.6174, "step": 230 }, { "epoch": 0.005870742446480085, "grad_norm": 5.351879119873047, "learning_rate": 4.990299779003661e-05, "loss": 0.4632, "step": 232 }, { "epoch": 0.005921352295156637, "grad_norm": 6.497000694274902, "learning_rate": 4.990215429255867e-05, "loss": 0.5266, "step": 234 }, { "epoch": 0.00597196214383319, "grad_norm": 10.24178695678711, "learning_rate": 4.990131079508073e-05, "loss": 0.8037, "step": 236 }, { "epoch": 0.006022571992509743, "grad_norm": 5.579001426696777, "learning_rate": 4.990046729760278e-05, "loss": 0.6132, "step": 238 }, { "epoch": 0.006073181841186295, "grad_norm": 3.613511562347412, "learning_rate": 4.989962380012484e-05, "loss": 0.3918, "step": 240 }, { "epoch": 0.006123791689862847, "grad_norm": 9.699933052062988, "learning_rate": 4.9898780302646894e-05, "loss": 0.8956, "step": 242 }, { "epoch": 0.0061744015385394, "grad_norm": 10.87297534942627, "learning_rate": 4.9897936805168954e-05, "loss": 1.1579, "step": 244 }, { "epoch": 0.006225011387215952, "grad_norm": 6.29569673538208, "learning_rate": 4.9897093307691014e-05, "loss": 0.4237, "step": 246 }, { "epoch": 0.006275621235892505, "grad_norm": 9.018596649169922, "learning_rate": 4.989624981021307e-05, "loss": 0.7854, "step": 248 }, { "epoch": 0.006326231084569057, "grad_norm": 9.707599639892578, "learning_rate": 4.989540631273513e-05, "loss": 1.3102, "step": 250 }, { "epoch": 0.006376840933245609, "grad_norm": 5.1023430824279785, "learning_rate": 4.989456281525718e-05, "loss": 0.5906, "step": 252 }, { "epoch": 0.006427450781922162, "grad_norm": 2.3029062747955322, "learning_rate": 4.989371931777924e-05, "loss": 0.7183, "step": 254 }, { "epoch": 0.006478060630598715, "grad_norm": 11.459485054016113, "learning_rate": 4.98928758203013e-05, "loss": 0.5505, "step": 256 }, { "epoch": 0.006528670479275267, "grad_norm": 7.620695114135742, "learning_rate": 4.9892032322823355e-05, "loss": 0.9374, "step": 258 }, { "epoch": 0.006579280327951819, "grad_norm": 5.305224895477295, "learning_rate": 4.9891188825345416e-05, "loss": 0.7362, "step": 260 }, { "epoch": 0.006629890176628372, "grad_norm": 7.901217460632324, "learning_rate": 4.989034532786747e-05, "loss": 0.9518, "step": 262 }, { "epoch": 0.006680500025304925, "grad_norm": 7.463654041290283, "learning_rate": 4.988950183038953e-05, "loss": 0.9789, "step": 264 }, { "epoch": 0.0067311098739814765, "grad_norm": 4.86021089553833, "learning_rate": 4.988865833291159e-05, "loss": 0.8441, "step": 266 }, { "epoch": 0.006781719722658029, "grad_norm": 5.242839813232422, "learning_rate": 4.988781483543364e-05, "loss": 0.8258, "step": 268 }, { "epoch": 0.006832329571334582, "grad_norm": 3.524228572845459, "learning_rate": 4.98869713379557e-05, "loss": 0.3924, "step": 270 }, { "epoch": 0.006882939420011134, "grad_norm": 3.8527095317840576, "learning_rate": 4.9886127840477756e-05, "loss": 0.5253, "step": 272 }, { "epoch": 0.0069335492686876865, "grad_norm": 3.4836299419403076, "learning_rate": 4.9885284342999817e-05, "loss": 0.5859, "step": 274 }, { "epoch": 0.006984159117364239, "grad_norm": 9.650673866271973, "learning_rate": 4.988444084552188e-05, "loss": 0.7122, "step": 276 }, { "epoch": 0.007034768966040792, "grad_norm": 5.5393829345703125, "learning_rate": 4.988359734804393e-05, "loss": 0.5434, "step": 278 }, { "epoch": 0.007085378814717344, "grad_norm": 11.345093727111816, "learning_rate": 4.988275385056599e-05, "loss": 0.9166, "step": 280 }, { "epoch": 0.0071359886633938965, "grad_norm": 4.247881889343262, "learning_rate": 4.9881910353088044e-05, "loss": 0.6162, "step": 282 }, { "epoch": 0.007186598512070449, "grad_norm": 5.472718238830566, "learning_rate": 4.9881066855610104e-05, "loss": 0.8378, "step": 284 }, { "epoch": 0.007237208360747001, "grad_norm": 6.520363807678223, "learning_rate": 4.9880223358132164e-05, "loss": 0.6258, "step": 286 }, { "epoch": 0.007287818209423554, "grad_norm": 4.629631996154785, "learning_rate": 4.987937986065422e-05, "loss": 0.5166, "step": 288 }, { "epoch": 0.0073384280581001065, "grad_norm": 9.533284187316895, "learning_rate": 4.987853636317628e-05, "loss": 1.1145, "step": 290 }, { "epoch": 0.007389037906776658, "grad_norm": 7.794739246368408, "learning_rate": 4.987769286569833e-05, "loss": 1.0134, "step": 292 }, { "epoch": 0.007439647755453211, "grad_norm": 8.504875183105469, "learning_rate": 4.987684936822039e-05, "loss": 0.8261, "step": 294 }, { "epoch": 0.007490257604129764, "grad_norm": 6.671535015106201, "learning_rate": 4.9876005870742445e-05, "loss": 0.7141, "step": 296 }, { "epoch": 0.0075408674528063165, "grad_norm": 11.959641456604004, "learning_rate": 4.9875162373264505e-05, "loss": 0.8247, "step": 298 }, { "epoch": 0.007591477301482868, "grad_norm": 4.4155378341674805, "learning_rate": 4.9874318875786565e-05, "loss": 0.348, "step": 300 }, { "epoch": 0.007642087150159421, "grad_norm": 5.849681377410889, "learning_rate": 4.987347537830862e-05, "loss": 0.8355, "step": 302 }, { "epoch": 0.007692696998835974, "grad_norm": 10.244682312011719, "learning_rate": 4.987263188083068e-05, "loss": 0.7634, "step": 304 }, { "epoch": 0.007743306847512526, "grad_norm": 3.5256447792053223, "learning_rate": 4.987178838335273e-05, "loss": 0.451, "step": 306 }, { "epoch": 0.007793916696189078, "grad_norm": 7.318663120269775, "learning_rate": 4.987094488587479e-05, "loss": 0.7009, "step": 308 }, { "epoch": 0.007844526544865631, "grad_norm": 5.343123912811279, "learning_rate": 4.987010138839685e-05, "loss": 0.4291, "step": 310 }, { "epoch": 0.007895136393542184, "grad_norm": 6.702011585235596, "learning_rate": 4.9869257890918906e-05, "loss": 0.9008, "step": 312 }, { "epoch": 0.007945746242218736, "grad_norm": 9.040878295898438, "learning_rate": 4.9868414393440966e-05, "loss": 1.0295, "step": 314 }, { "epoch": 0.007996356090895287, "grad_norm": 5.2994585037231445, "learning_rate": 4.986757089596302e-05, "loss": 0.6427, "step": 316 }, { "epoch": 0.00804696593957184, "grad_norm": 8.381704330444336, "learning_rate": 4.986672739848508e-05, "loss": 0.8675, "step": 318 }, { "epoch": 0.008097575788248393, "grad_norm": 9.162700653076172, "learning_rate": 4.986588390100714e-05, "loss": 1.1891, "step": 320 }, { "epoch": 0.008148185636924946, "grad_norm": 10.618518829345703, "learning_rate": 4.9865040403529194e-05, "loss": 0.7131, "step": 322 }, { "epoch": 0.008198795485601498, "grad_norm": 7.559556484222412, "learning_rate": 4.9864196906051254e-05, "loss": 0.7327, "step": 324 }, { "epoch": 0.008249405334278051, "grad_norm": 7.372714519500732, "learning_rate": 4.986335340857331e-05, "loss": 0.5496, "step": 326 }, { "epoch": 0.008300015182954604, "grad_norm": 5.887473106384277, "learning_rate": 4.986250991109537e-05, "loss": 0.4605, "step": 328 }, { "epoch": 0.008350625031631155, "grad_norm": 3.8664021492004395, "learning_rate": 4.986166641361743e-05, "loss": 0.8836, "step": 330 }, { "epoch": 0.008401234880307707, "grad_norm": 6.713327407836914, "learning_rate": 4.986082291613948e-05, "loss": 0.7846, "step": 332 }, { "epoch": 0.00845184472898426, "grad_norm": 6.036464691162109, "learning_rate": 4.985997941866154e-05, "loss": 0.8055, "step": 334 }, { "epoch": 0.008502454577660813, "grad_norm": 4.087986946105957, "learning_rate": 4.9859135921183595e-05, "loss": 0.7829, "step": 336 }, { "epoch": 0.008553064426337366, "grad_norm": 9.335679054260254, "learning_rate": 4.9858292423705655e-05, "loss": 1.279, "step": 338 }, { "epoch": 0.008603674275013918, "grad_norm": 6.914140701293945, "learning_rate": 4.9857448926227715e-05, "loss": 0.8773, "step": 340 }, { "epoch": 0.008654284123690471, "grad_norm": 5.106595993041992, "learning_rate": 4.985660542874977e-05, "loss": 0.7357, "step": 342 }, { "epoch": 0.008704893972367022, "grad_norm": 5.217001914978027, "learning_rate": 4.985576193127183e-05, "loss": 1.0202, "step": 344 }, { "epoch": 0.008755503821043575, "grad_norm": 7.956677436828613, "learning_rate": 4.985491843379388e-05, "loss": 0.8517, "step": 346 }, { "epoch": 0.008806113669720127, "grad_norm": 6.432283878326416, "learning_rate": 4.985407493631594e-05, "loss": 0.7963, "step": 348 }, { "epoch": 0.00885672351839668, "grad_norm": 25.808292388916016, "learning_rate": 4.9853231438838e-05, "loss": 0.5176, "step": 350 }, { "epoch": 0.008907333367073233, "grad_norm": 6.786406517028809, "learning_rate": 4.9852387941360056e-05, "loss": 0.9981, "step": 352 }, { "epoch": 0.008957943215749786, "grad_norm": 13.001777648925781, "learning_rate": 4.9851544443882116e-05, "loss": 0.6546, "step": 354 }, { "epoch": 0.009008553064426337, "grad_norm": 5.085973262786865, "learning_rate": 4.985070094640417e-05, "loss": 0.4445, "step": 356 }, { "epoch": 0.00905916291310289, "grad_norm": 10.445619583129883, "learning_rate": 4.984985744892623e-05, "loss": 0.519, "step": 358 }, { "epoch": 0.009109772761779442, "grad_norm": 7.5755510330200195, "learning_rate": 4.984901395144829e-05, "loss": 0.9253, "step": 360 }, { "epoch": 0.009160382610455995, "grad_norm": 12.85741901397705, "learning_rate": 4.984817045397034e-05, "loss": 0.7553, "step": 362 }, { "epoch": 0.009210992459132547, "grad_norm": 4.291009426116943, "learning_rate": 4.9847326956492404e-05, "loss": 0.7912, "step": 364 }, { "epoch": 0.0092616023078091, "grad_norm": 9.602625846862793, "learning_rate": 4.984648345901446e-05, "loss": 0.7954, "step": 366 }, { "epoch": 0.009312212156485653, "grad_norm": 10.099987030029297, "learning_rate": 4.984563996153652e-05, "loss": 0.6241, "step": 368 }, { "epoch": 0.009362822005162204, "grad_norm": 9.353877067565918, "learning_rate": 4.984479646405858e-05, "loss": 0.7845, "step": 370 }, { "epoch": 0.009413431853838757, "grad_norm": 5.5436506271362305, "learning_rate": 4.984395296658063e-05, "loss": 0.6337, "step": 372 }, { "epoch": 0.00946404170251531, "grad_norm": 6.538369655609131, "learning_rate": 4.984310946910269e-05, "loss": 0.4818, "step": 374 }, { "epoch": 0.009514651551191862, "grad_norm": 11.956756591796875, "learning_rate": 4.9842265971624744e-05, "loss": 0.9237, "step": 376 }, { "epoch": 0.009565261399868415, "grad_norm": 11.44477367401123, "learning_rate": 4.9841422474146805e-05, "loss": 1.0572, "step": 378 }, { "epoch": 0.009615871248544967, "grad_norm": 9.988565444946289, "learning_rate": 4.984057897666886e-05, "loss": 0.6858, "step": 380 }, { "epoch": 0.00966648109722152, "grad_norm": 4.8977813720703125, "learning_rate": 4.983973547919092e-05, "loss": 0.5965, "step": 382 }, { "epoch": 0.009717090945898071, "grad_norm": 6.308709144592285, "learning_rate": 4.983889198171298e-05, "loss": 0.4817, "step": 384 }, { "epoch": 0.009767700794574624, "grad_norm": 5.661224842071533, "learning_rate": 4.983804848423503e-05, "loss": 0.5738, "step": 386 }, { "epoch": 0.009818310643251177, "grad_norm": 5.926614284515381, "learning_rate": 4.983720498675709e-05, "loss": 0.5391, "step": 388 }, { "epoch": 0.00986892049192773, "grad_norm": 4.607276916503906, "learning_rate": 4.9836361489279145e-05, "loss": 0.4223, "step": 390 }, { "epoch": 0.009919530340604282, "grad_norm": 14.739141464233398, "learning_rate": 4.9835517991801206e-05, "loss": 1.0047, "step": 392 }, { "epoch": 0.009970140189280835, "grad_norm": 4.335879325866699, "learning_rate": 4.9834674494323266e-05, "loss": 0.7172, "step": 394 }, { "epoch": 0.010020750037957386, "grad_norm": 8.950626373291016, "learning_rate": 4.983383099684532e-05, "loss": 0.8787, "step": 396 }, { "epoch": 0.010071359886633938, "grad_norm": 5.630717754364014, "learning_rate": 4.983298749936738e-05, "loss": 0.5813, "step": 398 }, { "epoch": 0.010121969735310491, "grad_norm": 9.189420700073242, "learning_rate": 4.983214400188943e-05, "loss": 0.9012, "step": 400 }, { "epoch": 0.010121969735310491, "eval_cer": 0.22362332855033584, "eval_loss": 0.3546978533267975, "eval_runtime": 2758.4684, "eval_samples_per_second": 5.682, "eval_steps_per_second": 0.355, "step": 400 }, { "epoch": 0.010172579583987044, "grad_norm": 3.5027172565460205, "learning_rate": 4.983130050441149e-05, "loss": 0.5343, "step": 402 }, { "epoch": 0.010223189432663597, "grad_norm": 7.1599836349487305, "learning_rate": 4.983045700693355e-05, "loss": 0.64, "step": 404 }, { "epoch": 0.01027379928134015, "grad_norm": 6.179046630859375, "learning_rate": 4.982961350945561e-05, "loss": 0.6958, "step": 406 }, { "epoch": 0.010324409130016702, "grad_norm": 6.452561855316162, "learning_rate": 4.982877001197767e-05, "loss": 0.9579, "step": 408 }, { "epoch": 0.010375018978693253, "grad_norm": 6.45066499710083, "learning_rate": 4.982792651449972e-05, "loss": 0.8967, "step": 410 }, { "epoch": 0.010425628827369806, "grad_norm": 8.554861068725586, "learning_rate": 4.982708301702178e-05, "loss": 1.1857, "step": 412 }, { "epoch": 0.010476238676046358, "grad_norm": 3.6959705352783203, "learning_rate": 4.982623951954384e-05, "loss": 0.4889, "step": 414 }, { "epoch": 0.010526848524722911, "grad_norm": 12.848132133483887, "learning_rate": 4.9825396022065894e-05, "loss": 0.64, "step": 416 }, { "epoch": 0.010577458373399464, "grad_norm": 6.260450839996338, "learning_rate": 4.9824552524587954e-05, "loss": 1.0134, "step": 418 }, { "epoch": 0.010628068222076016, "grad_norm": 6.700595378875732, "learning_rate": 4.982370902711001e-05, "loss": 0.8566, "step": 420 }, { "epoch": 0.01067867807075257, "grad_norm": 6.211639404296875, "learning_rate": 4.982286552963207e-05, "loss": 0.9721, "step": 422 }, { "epoch": 0.01072928791942912, "grad_norm": 7.872774600982666, "learning_rate": 4.982202203215413e-05, "loss": 0.4064, "step": 424 }, { "epoch": 0.010779897768105673, "grad_norm": 9.730685234069824, "learning_rate": 4.982117853467618e-05, "loss": 1.044, "step": 426 }, { "epoch": 0.010830507616782226, "grad_norm": 4.908615589141846, "learning_rate": 4.982033503719824e-05, "loss": 0.6847, "step": 428 }, { "epoch": 0.010881117465458778, "grad_norm": 7.32842493057251, "learning_rate": 4.9819491539720295e-05, "loss": 0.8148, "step": 430 }, { "epoch": 0.010931727314135331, "grad_norm": 4.537167072296143, "learning_rate": 4.9818648042242355e-05, "loss": 0.4907, "step": 432 }, { "epoch": 0.010982337162811884, "grad_norm": 5.290032863616943, "learning_rate": 4.9817804544764416e-05, "loss": 0.6806, "step": 434 }, { "epoch": 0.011032947011488436, "grad_norm": 6.21295166015625, "learning_rate": 4.981696104728647e-05, "loss": 0.6636, "step": 436 }, { "epoch": 0.011083556860164987, "grad_norm": 5.210526466369629, "learning_rate": 4.981611754980853e-05, "loss": 0.7971, "step": 438 }, { "epoch": 0.01113416670884154, "grad_norm": 40.482147216796875, "learning_rate": 4.981527405233058e-05, "loss": 0.7819, "step": 440 }, { "epoch": 0.011184776557518093, "grad_norm": 6.940312385559082, "learning_rate": 4.981443055485264e-05, "loss": 0.8548, "step": 442 }, { "epoch": 0.011235386406194646, "grad_norm": 10.026724815368652, "learning_rate": 4.98135870573747e-05, "loss": 0.5857, "step": 444 }, { "epoch": 0.011285996254871198, "grad_norm": 10.645378112792969, "learning_rate": 4.9812743559896756e-05, "loss": 0.7426, "step": 446 }, { "epoch": 0.011336606103547751, "grad_norm": 17.70381736755371, "learning_rate": 4.981190006241882e-05, "loss": 1.0465, "step": 448 }, { "epoch": 0.011387215952224302, "grad_norm": 4.017563819885254, "learning_rate": 4.981105656494087e-05, "loss": 0.6183, "step": 450 }, { "epoch": 0.011437825800900855, "grad_norm": 7.969531059265137, "learning_rate": 4.981021306746293e-05, "loss": 0.8661, "step": 452 }, { "epoch": 0.011488435649577407, "grad_norm": 12.08060359954834, "learning_rate": 4.980936956998499e-05, "loss": 0.7461, "step": 454 }, { "epoch": 0.01153904549825396, "grad_norm": 6.960525989532471, "learning_rate": 4.9808526072507044e-05, "loss": 0.5767, "step": 456 }, { "epoch": 0.011589655346930513, "grad_norm": 10.034902572631836, "learning_rate": 4.9807682575029104e-05, "loss": 0.7012, "step": 458 }, { "epoch": 0.011640265195607066, "grad_norm": 11.939443588256836, "learning_rate": 4.980683907755116e-05, "loss": 0.6601, "step": 460 }, { "epoch": 0.011690875044283618, "grad_norm": 5.776790618896484, "learning_rate": 4.980599558007322e-05, "loss": 0.7514, "step": 462 }, { "epoch": 0.01174148489296017, "grad_norm": 9.15820598602295, "learning_rate": 4.980515208259527e-05, "loss": 0.8779, "step": 464 }, { "epoch": 0.011792094741636722, "grad_norm": 7.228218078613281, "learning_rate": 4.980430858511733e-05, "loss": 1.0643, "step": 466 }, { "epoch": 0.011842704590313275, "grad_norm": 4.174966812133789, "learning_rate": 4.980346508763939e-05, "loss": 0.3511, "step": 468 }, { "epoch": 0.011893314438989827, "grad_norm": 9.501602172851562, "learning_rate": 4.9802621590161445e-05, "loss": 0.8323, "step": 470 }, { "epoch": 0.01194392428766638, "grad_norm": 4.438223361968994, "learning_rate": 4.9801778092683505e-05, "loss": 0.3112, "step": 472 }, { "epoch": 0.011994534136342933, "grad_norm": 8.708006858825684, "learning_rate": 4.980093459520556e-05, "loss": 0.6483, "step": 474 }, { "epoch": 0.012045143985019486, "grad_norm": 4.928430557250977, "learning_rate": 4.980009109772762e-05, "loss": 0.7263, "step": 476 }, { "epoch": 0.012095753833696037, "grad_norm": 11.870718955993652, "learning_rate": 4.979924760024968e-05, "loss": 0.9143, "step": 478 }, { "epoch": 0.01214636368237259, "grad_norm": 6.654867649078369, "learning_rate": 4.979840410277173e-05, "loss": 0.5257, "step": 480 }, { "epoch": 0.012196973531049142, "grad_norm": 4.725414752960205, "learning_rate": 4.979756060529379e-05, "loss": 0.5919, "step": 482 }, { "epoch": 0.012247583379725695, "grad_norm": 5.607127666473389, "learning_rate": 4.9796717107815846e-05, "loss": 0.5198, "step": 484 }, { "epoch": 0.012298193228402247, "grad_norm": 5.8716864585876465, "learning_rate": 4.9795873610337906e-05, "loss": 0.7554, "step": 486 }, { "epoch": 0.0123488030770788, "grad_norm": 7.410449504852295, "learning_rate": 4.9795030112859966e-05, "loss": 0.8967, "step": 488 }, { "epoch": 0.012399412925755351, "grad_norm": 10.435539245605469, "learning_rate": 4.979418661538202e-05, "loss": 0.7446, "step": 490 }, { "epoch": 0.012450022774431904, "grad_norm": 4.085461139678955, "learning_rate": 4.979334311790408e-05, "loss": 0.56, "step": 492 }, { "epoch": 0.012500632623108457, "grad_norm": 7.522024154663086, "learning_rate": 4.9792499620426133e-05, "loss": 0.7217, "step": 494 }, { "epoch": 0.01255124247178501, "grad_norm": 6.347813606262207, "learning_rate": 4.9791656122948194e-05, "loss": 0.5247, "step": 496 }, { "epoch": 0.012601852320461562, "grad_norm": 8.447410583496094, "learning_rate": 4.9790812625470254e-05, "loss": 0.74, "step": 498 }, { "epoch": 0.012652462169138115, "grad_norm": 5.143301486968994, "learning_rate": 4.978996912799231e-05, "loss": 0.5553, "step": 500 }, { "epoch": 0.012703072017814667, "grad_norm": 5.69724702835083, "learning_rate": 4.978912563051437e-05, "loss": 0.9869, "step": 502 }, { "epoch": 0.012753681866491218, "grad_norm": 3.5676660537719727, "learning_rate": 4.978828213303642e-05, "loss": 0.423, "step": 504 }, { "epoch": 0.012804291715167771, "grad_norm": 5.91968297958374, "learning_rate": 4.978743863555848e-05, "loss": 0.7037, "step": 506 }, { "epoch": 0.012854901563844324, "grad_norm": 9.412429809570312, "learning_rate": 4.978659513808054e-05, "loss": 1.1408, "step": 508 }, { "epoch": 0.012905511412520877, "grad_norm": 8.342994689941406, "learning_rate": 4.9785751640602595e-05, "loss": 0.8661, "step": 510 }, { "epoch": 0.01295612126119743, "grad_norm": 7.621118068695068, "learning_rate": 4.9784908143124655e-05, "loss": 0.677, "step": 512 }, { "epoch": 0.013006731109873982, "grad_norm": 4.431066036224365, "learning_rate": 4.978406464564671e-05, "loss": 0.637, "step": 514 }, { "epoch": 0.013057340958550535, "grad_norm": 9.607247352600098, "learning_rate": 4.978322114816877e-05, "loss": 0.5467, "step": 516 }, { "epoch": 0.013107950807227086, "grad_norm": 8.333694458007812, "learning_rate": 4.978237765069083e-05, "loss": 1.0044, "step": 518 }, { "epoch": 0.013158560655903638, "grad_norm": 3.5092146396636963, "learning_rate": 4.978153415321288e-05, "loss": 0.3496, "step": 520 }, { "epoch": 0.013209170504580191, "grad_norm": 5.619466781616211, "learning_rate": 4.978069065573494e-05, "loss": 0.5608, "step": 522 }, { "epoch": 0.013259780353256744, "grad_norm": 10.299610137939453, "learning_rate": 4.9779847158256996e-05, "loss": 0.7437, "step": 524 }, { "epoch": 0.013310390201933297, "grad_norm": 9.14527416229248, "learning_rate": 4.9779003660779056e-05, "loss": 0.7961, "step": 526 }, { "epoch": 0.01336100005060985, "grad_norm": 4.311985015869141, "learning_rate": 4.9778160163301116e-05, "loss": 0.5069, "step": 528 }, { "epoch": 0.013411609899286402, "grad_norm": 7.6014862060546875, "learning_rate": 4.977731666582317e-05, "loss": 0.8133, "step": 530 }, { "epoch": 0.013462219747962953, "grad_norm": 5.695495128631592, "learning_rate": 4.977647316834523e-05, "loss": 0.4957, "step": 532 }, { "epoch": 0.013512829596639506, "grad_norm": 8.789469718933105, "learning_rate": 4.977562967086728e-05, "loss": 0.9009, "step": 534 }, { "epoch": 0.013563439445316058, "grad_norm": 6.521892547607422, "learning_rate": 4.977478617338934e-05, "loss": 0.4392, "step": 536 }, { "epoch": 0.013614049293992611, "grad_norm": 8.529374122619629, "learning_rate": 4.9773942675911404e-05, "loss": 0.507, "step": 538 }, { "epoch": 0.013664659142669164, "grad_norm": 5.722716808319092, "learning_rate": 4.977309917843346e-05, "loss": 0.5071, "step": 540 }, { "epoch": 0.013715268991345717, "grad_norm": 4.3265156745910645, "learning_rate": 4.977225568095552e-05, "loss": 0.7658, "step": 542 }, { "epoch": 0.013765878840022268, "grad_norm": 5.346278190612793, "learning_rate": 4.977141218347757e-05, "loss": 0.5658, "step": 544 }, { "epoch": 0.01381648868869882, "grad_norm": 7.274371147155762, "learning_rate": 4.977056868599963e-05, "loss": 0.8856, "step": 546 }, { "epoch": 0.013867098537375373, "grad_norm": 5.354006767272949, "learning_rate": 4.9769725188521684e-05, "loss": 0.5224, "step": 548 }, { "epoch": 0.013917708386051926, "grad_norm": 8.813762664794922, "learning_rate": 4.9768881691043744e-05, "loss": 1.1558, "step": 550 }, { "epoch": 0.013968318234728478, "grad_norm": 6.957085132598877, "learning_rate": 4.9768038193565805e-05, "loss": 0.4142, "step": 552 }, { "epoch": 0.014018928083405031, "grad_norm": 12.203534126281738, "learning_rate": 4.976719469608786e-05, "loss": 0.5945, "step": 554 }, { "epoch": 0.014069537932081584, "grad_norm": 6.845212459564209, "learning_rate": 4.976635119860992e-05, "loss": 0.4359, "step": 556 }, { "epoch": 0.014120147780758135, "grad_norm": 6.074872970581055, "learning_rate": 4.976550770113197e-05, "loss": 0.7704, "step": 558 }, { "epoch": 0.014170757629434688, "grad_norm": 5.776919364929199, "learning_rate": 4.976466420365403e-05, "loss": 0.7799, "step": 560 }, { "epoch": 0.01422136747811124, "grad_norm": 11.90626335144043, "learning_rate": 4.976382070617609e-05, "loss": 0.9279, "step": 562 }, { "epoch": 0.014271977326787793, "grad_norm": 4.0126848220825195, "learning_rate": 4.9762977208698145e-05, "loss": 0.8148, "step": 564 }, { "epoch": 0.014322587175464346, "grad_norm": 5.540651798248291, "learning_rate": 4.9762133711220206e-05, "loss": 0.528, "step": 566 }, { "epoch": 0.014373197024140898, "grad_norm": 4.119876861572266, "learning_rate": 4.976129021374226e-05, "loss": 0.8705, "step": 568 }, { "epoch": 0.014423806872817451, "grad_norm": 11.188005447387695, "learning_rate": 4.976044671626432e-05, "loss": 0.467, "step": 570 }, { "epoch": 0.014474416721494002, "grad_norm": 6.426743030548096, "learning_rate": 4.975960321878638e-05, "loss": 0.5936, "step": 572 }, { "epoch": 0.014525026570170555, "grad_norm": 6.039766788482666, "learning_rate": 4.975875972130843e-05, "loss": 0.7222, "step": 574 }, { "epoch": 0.014575636418847107, "grad_norm": 12.856761932373047, "learning_rate": 4.975791622383049e-05, "loss": 0.501, "step": 576 }, { "epoch": 0.01462624626752366, "grad_norm": 13.36010456085205, "learning_rate": 4.9757072726352547e-05, "loss": 0.473, "step": 578 }, { "epoch": 0.014676856116200213, "grad_norm": 13.758210182189941, "learning_rate": 4.975622922887461e-05, "loss": 1.1569, "step": 580 }, { "epoch": 0.014727465964876766, "grad_norm": 5.978826522827148, "learning_rate": 4.975538573139667e-05, "loss": 0.5853, "step": 582 }, { "epoch": 0.014778075813553317, "grad_norm": 4.91432523727417, "learning_rate": 4.975454223391872e-05, "loss": 0.5073, "step": 584 }, { "epoch": 0.01482868566222987, "grad_norm": 4.914419174194336, "learning_rate": 4.975369873644078e-05, "loss": 0.5638, "step": 586 }, { "epoch": 0.014879295510906422, "grad_norm": 6.7935967445373535, "learning_rate": 4.9752855238962834e-05, "loss": 0.773, "step": 588 }, { "epoch": 0.014929905359582975, "grad_norm": 5.500904083251953, "learning_rate": 4.9752011741484894e-05, "loss": 0.6985, "step": 590 }, { "epoch": 0.014980515208259527, "grad_norm": 10.455467224121094, "learning_rate": 4.9751168244006954e-05, "loss": 0.5856, "step": 592 }, { "epoch": 0.01503112505693608, "grad_norm": 14.375707626342773, "learning_rate": 4.975032474652901e-05, "loss": 0.8323, "step": 594 }, { "epoch": 0.015081734905612633, "grad_norm": 8.137022972106934, "learning_rate": 4.974948124905107e-05, "loss": 0.7872, "step": 596 }, { "epoch": 0.015132344754289184, "grad_norm": 4.684526443481445, "learning_rate": 4.974863775157312e-05, "loss": 0.7216, "step": 598 }, { "epoch": 0.015182954602965737, "grad_norm": 6.151386737823486, "learning_rate": 4.974779425409518e-05, "loss": 0.7315, "step": 600 }, { "epoch": 0.015182954602965737, "eval_cer": 0.21082101191590244, "eval_loss": 0.33135300874710083, "eval_runtime": 2751.8944, "eval_samples_per_second": 5.696, "eval_steps_per_second": 0.356, "step": 600 }, { "epoch": 0.01523356445164229, "grad_norm": 7.473769664764404, "learning_rate": 4.974695075661724e-05, "loss": 0.3884, "step": 602 }, { "epoch": 0.015284174300318842, "grad_norm": 6.925198078155518, "learning_rate": 4.9746107259139295e-05, "loss": 0.6214, "step": 604 }, { "epoch": 0.015334784148995395, "grad_norm": 7.6508684158325195, "learning_rate": 4.9745263761661355e-05, "loss": 0.3224, "step": 606 }, { "epoch": 0.015385393997671947, "grad_norm": 8.625561714172363, "learning_rate": 4.974442026418341e-05, "loss": 1.1576, "step": 608 }, { "epoch": 0.0154360038463485, "grad_norm": 4.799185276031494, "learning_rate": 4.974357676670547e-05, "loss": 0.4953, "step": 610 }, { "epoch": 0.015486613695025051, "grad_norm": 6.022134780883789, "learning_rate": 4.974273326922753e-05, "loss": 0.417, "step": 612 }, { "epoch": 0.015537223543701604, "grad_norm": 6.1436333656311035, "learning_rate": 4.974188977174958e-05, "loss": 0.7479, "step": 614 }, { "epoch": 0.015587833392378157, "grad_norm": 11.182157516479492, "learning_rate": 4.974104627427164e-05, "loss": 1.1078, "step": 616 }, { "epoch": 0.01563844324105471, "grad_norm": 7.718019485473633, "learning_rate": 4.9740202776793696e-05, "loss": 0.7953, "step": 618 }, { "epoch": 0.015689053089731262, "grad_norm": 6.890071392059326, "learning_rate": 4.9739359279315756e-05, "loss": 0.8054, "step": 620 }, { "epoch": 0.015739662938407815, "grad_norm": 9.046427726745605, "learning_rate": 4.973851578183782e-05, "loss": 0.8913, "step": 622 }, { "epoch": 0.015790272787084367, "grad_norm": 7.701432704925537, "learning_rate": 4.973767228435987e-05, "loss": 1.2825, "step": 624 }, { "epoch": 0.01584088263576092, "grad_norm": 6.6972808837890625, "learning_rate": 4.973682878688193e-05, "loss": 0.7063, "step": 626 }, { "epoch": 0.015891492484437473, "grad_norm": 7.926259517669678, "learning_rate": 4.9735985289403984e-05, "loss": 0.6431, "step": 628 }, { "epoch": 0.015942102333114026, "grad_norm": 6.506617546081543, "learning_rate": 4.9735141791926044e-05, "loss": 0.7283, "step": 630 }, { "epoch": 0.015992712181790575, "grad_norm": 5.657668590545654, "learning_rate": 4.97342982944481e-05, "loss": 0.7906, "step": 632 }, { "epoch": 0.016043322030467128, "grad_norm": 5.814841270446777, "learning_rate": 4.973345479697016e-05, "loss": 0.7872, "step": 634 }, { "epoch": 0.01609393187914368, "grad_norm": 4.063594818115234, "learning_rate": 4.973261129949222e-05, "loss": 0.7104, "step": 636 }, { "epoch": 0.016144541727820233, "grad_norm": 2.756787061691284, "learning_rate": 4.973176780201427e-05, "loss": 0.3609, "step": 638 }, { "epoch": 0.016195151576496786, "grad_norm": 6.163934707641602, "learning_rate": 4.973092430453633e-05, "loss": 0.7657, "step": 640 }, { "epoch": 0.01624576142517334, "grad_norm": 7.708083629608154, "learning_rate": 4.9730080807058385e-05, "loss": 0.7486, "step": 642 }, { "epoch": 0.01629637127384989, "grad_norm": 6.693125247955322, "learning_rate": 4.9729237309580445e-05, "loss": 0.799, "step": 644 }, { "epoch": 0.016346981122526444, "grad_norm": 8.108495712280273, "learning_rate": 4.9728393812102505e-05, "loss": 0.8542, "step": 646 }, { "epoch": 0.016397590971202997, "grad_norm": 13.466230392456055, "learning_rate": 4.972755031462456e-05, "loss": 0.9259, "step": 648 }, { "epoch": 0.01644820081987955, "grad_norm": 7.1850056648254395, "learning_rate": 4.972670681714662e-05, "loss": 0.4547, "step": 650 }, { "epoch": 0.016498810668556102, "grad_norm": 10.240496635437012, "learning_rate": 4.972586331966867e-05, "loss": 0.7534, "step": 652 }, { "epoch": 0.016549420517232655, "grad_norm": 7.284756183624268, "learning_rate": 4.972501982219073e-05, "loss": 0.7111, "step": 654 }, { "epoch": 0.016600030365909207, "grad_norm": 4.935502052307129, "learning_rate": 4.972417632471279e-05, "loss": 0.587, "step": 656 }, { "epoch": 0.016650640214585757, "grad_norm": 9.42276382446289, "learning_rate": 4.9723332827234846e-05, "loss": 0.6914, "step": 658 }, { "epoch": 0.01670125006326231, "grad_norm": 8.303751945495605, "learning_rate": 4.9722489329756906e-05, "loss": 0.7424, "step": 660 }, { "epoch": 0.016751859911938862, "grad_norm": 5.425510883331299, "learning_rate": 4.972164583227896e-05, "loss": 0.4959, "step": 662 }, { "epoch": 0.016802469760615415, "grad_norm": 8.581243515014648, "learning_rate": 4.972080233480102e-05, "loss": 0.5844, "step": 664 }, { "epoch": 0.016853079609291968, "grad_norm": 7.702110290527344, "learning_rate": 4.971995883732308e-05, "loss": 0.6744, "step": 666 }, { "epoch": 0.01690368945796852, "grad_norm": 5.569732189178467, "learning_rate": 4.9719115339845133e-05, "loss": 0.5237, "step": 668 }, { "epoch": 0.016954299306645073, "grad_norm": 7.927607536315918, "learning_rate": 4.9718271842367194e-05, "loss": 0.6969, "step": 670 }, { "epoch": 0.017004909155321626, "grad_norm": 6.792113304138184, "learning_rate": 4.971742834488925e-05, "loss": 0.4953, "step": 672 }, { "epoch": 0.01705551900399818, "grad_norm": 41.776954650878906, "learning_rate": 4.971658484741131e-05, "loss": 0.5495, "step": 674 }, { "epoch": 0.01710612885267473, "grad_norm": 5.085512638092041, "learning_rate": 4.971574134993337e-05, "loss": 0.6303, "step": 676 }, { "epoch": 0.017156738701351284, "grad_norm": 5.323431491851807, "learning_rate": 4.971489785245542e-05, "loss": 0.7973, "step": 678 }, { "epoch": 0.017207348550027837, "grad_norm": 5.844650745391846, "learning_rate": 4.971405435497748e-05, "loss": 1.0397, "step": 680 }, { "epoch": 0.01725795839870439, "grad_norm": 1.8864692449569702, "learning_rate": 4.9713210857499535e-05, "loss": 0.3274, "step": 682 }, { "epoch": 0.017308568247380942, "grad_norm": 3.4474141597747803, "learning_rate": 4.9712367360021595e-05, "loss": 0.5992, "step": 684 }, { "epoch": 0.01735917809605749, "grad_norm": 5.180785655975342, "learning_rate": 4.9711523862543655e-05, "loss": 0.6462, "step": 686 }, { "epoch": 0.017409787944734044, "grad_norm": 5.768479347229004, "learning_rate": 4.971068036506571e-05, "loss": 0.6288, "step": 688 }, { "epoch": 0.017460397793410597, "grad_norm": 5.524656772613525, "learning_rate": 4.970983686758777e-05, "loss": 0.6045, "step": 690 }, { "epoch": 0.01751100764208715, "grad_norm": 6.222484111785889, "learning_rate": 4.970899337010982e-05, "loss": 0.4882, "step": 692 }, { "epoch": 0.017561617490763702, "grad_norm": 2.960340738296509, "learning_rate": 4.970814987263188e-05, "loss": 0.3803, "step": 694 }, { "epoch": 0.017612227339440255, "grad_norm": 3.2236385345458984, "learning_rate": 4.970730637515394e-05, "loss": 0.5617, "step": 696 }, { "epoch": 0.017662837188116808, "grad_norm": 8.282496452331543, "learning_rate": 4.9706462877675996e-05, "loss": 0.6835, "step": 698 }, { "epoch": 0.01771344703679336, "grad_norm": 7.472006797790527, "learning_rate": 4.9705619380198056e-05, "loss": 0.4747, "step": 700 }, { "epoch": 0.017764056885469913, "grad_norm": 6.8331780433654785, "learning_rate": 4.970477588272011e-05, "loss": 0.4765, "step": 702 }, { "epoch": 0.017814666734146466, "grad_norm": 4.116718292236328, "learning_rate": 4.970393238524217e-05, "loss": 0.4177, "step": 704 }, { "epoch": 0.01786527658282302, "grad_norm": 14.779850959777832, "learning_rate": 4.970308888776423e-05, "loss": 0.5939, "step": 706 }, { "epoch": 0.01791588643149957, "grad_norm": 10.066930770874023, "learning_rate": 4.970224539028628e-05, "loss": 0.7392, "step": 708 }, { "epoch": 0.017966496280176124, "grad_norm": 5.495025634765625, "learning_rate": 4.9701401892808343e-05, "loss": 0.6957, "step": 710 }, { "epoch": 0.018017106128852673, "grad_norm": 9.583708763122559, "learning_rate": 4.97005583953304e-05, "loss": 0.8032, "step": 712 }, { "epoch": 0.018067715977529226, "grad_norm": 6.409923076629639, "learning_rate": 4.969971489785246e-05, "loss": 0.9505, "step": 714 }, { "epoch": 0.01811832582620578, "grad_norm": 8.952066421508789, "learning_rate": 4.969887140037451e-05, "loss": 0.7739, "step": 716 }, { "epoch": 0.01816893567488233, "grad_norm": 6.334309101104736, "learning_rate": 4.969802790289657e-05, "loss": 1.0117, "step": 718 }, { "epoch": 0.018219545523558884, "grad_norm": 8.889698028564453, "learning_rate": 4.969718440541863e-05, "loss": 0.6535, "step": 720 }, { "epoch": 0.018270155372235437, "grad_norm": 2.9195971488952637, "learning_rate": 4.9696340907940684e-05, "loss": 0.3021, "step": 722 }, { "epoch": 0.01832076522091199, "grad_norm": 8.154908180236816, "learning_rate": 4.9695497410462744e-05, "loss": 0.8477, "step": 724 }, { "epoch": 0.018371375069588542, "grad_norm": 3.179515838623047, "learning_rate": 4.96946539129848e-05, "loss": 0.5856, "step": 726 }, { "epoch": 0.018421984918265095, "grad_norm": 7.491313457489014, "learning_rate": 4.969381041550686e-05, "loss": 0.8113, "step": 728 }, { "epoch": 0.018472594766941648, "grad_norm": 6.968503475189209, "learning_rate": 4.969296691802892e-05, "loss": 0.6095, "step": 730 }, { "epoch": 0.0185232046156182, "grad_norm": 7.4488348960876465, "learning_rate": 4.969212342055097e-05, "loss": 1.0515, "step": 732 }, { "epoch": 0.018573814464294753, "grad_norm": 2.959810733795166, "learning_rate": 4.969127992307303e-05, "loss": 0.5783, "step": 734 }, { "epoch": 0.018624424312971306, "grad_norm": 9.092787742614746, "learning_rate": 4.9690436425595085e-05, "loss": 0.878, "step": 736 }, { "epoch": 0.01867503416164786, "grad_norm": 7.564940929412842, "learning_rate": 4.9689592928117146e-05, "loss": 0.8122, "step": 738 }, { "epoch": 0.018725644010324408, "grad_norm": 7.550844192504883, "learning_rate": 4.9688749430639206e-05, "loss": 0.4964, "step": 740 }, { "epoch": 0.01877625385900096, "grad_norm": 11.271805763244629, "learning_rate": 4.968790593316126e-05, "loss": 0.9631, "step": 742 }, { "epoch": 0.018826863707677513, "grad_norm": 13.20101547241211, "learning_rate": 4.968706243568332e-05, "loss": 0.6274, "step": 744 }, { "epoch": 0.018877473556354066, "grad_norm": 5.18681526184082, "learning_rate": 4.968621893820537e-05, "loss": 0.6049, "step": 746 }, { "epoch": 0.01892808340503062, "grad_norm": 3.5832290649414062, "learning_rate": 4.968537544072743e-05, "loss": 0.363, "step": 748 }, { "epoch": 0.01897869325370717, "grad_norm": 4.764203071594238, "learning_rate": 4.968453194324949e-05, "loss": 0.4824, "step": 750 }, { "epoch": 0.019029303102383724, "grad_norm": 3.9694879055023193, "learning_rate": 4.9683688445771547e-05, "loss": 0.6074, "step": 752 }, { "epoch": 0.019079912951060277, "grad_norm": 4.219832897186279, "learning_rate": 4.968284494829361e-05, "loss": 0.3649, "step": 754 }, { "epoch": 0.01913052279973683, "grad_norm": 25.03742027282715, "learning_rate": 4.968200145081566e-05, "loss": 0.742, "step": 756 }, { "epoch": 0.019181132648413382, "grad_norm": 8.844772338867188, "learning_rate": 4.968115795333772e-05, "loss": 0.6859, "step": 758 }, { "epoch": 0.019231742497089935, "grad_norm": 7.3545918464660645, "learning_rate": 4.968031445585978e-05, "loss": 0.5395, "step": 760 }, { "epoch": 0.019282352345766487, "grad_norm": 3.499608278274536, "learning_rate": 4.9679470958381834e-05, "loss": 0.613, "step": 762 }, { "epoch": 0.01933296219444304, "grad_norm": 4.7259440422058105, "learning_rate": 4.9678627460903894e-05, "loss": 0.7538, "step": 764 }, { "epoch": 0.01938357204311959, "grad_norm": 5.475869178771973, "learning_rate": 4.967778396342595e-05, "loss": 0.7714, "step": 766 }, { "epoch": 0.019434181891796142, "grad_norm": 18.313730239868164, "learning_rate": 4.967694046594801e-05, "loss": 0.9697, "step": 768 }, { "epoch": 0.019484791740472695, "grad_norm": 8.159904479980469, "learning_rate": 4.967609696847007e-05, "loss": 0.5248, "step": 770 }, { "epoch": 0.019535401589149248, "grad_norm": 5.177513122558594, "learning_rate": 4.967525347099212e-05, "loss": 0.4911, "step": 772 }, { "epoch": 0.0195860114378258, "grad_norm": 4.758183479309082, "learning_rate": 4.967440997351418e-05, "loss": 0.7151, "step": 774 }, { "epoch": 0.019636621286502353, "grad_norm": 8.288613319396973, "learning_rate": 4.9673566476036235e-05, "loss": 0.49, "step": 776 }, { "epoch": 0.019687231135178906, "grad_norm": 8.448434829711914, "learning_rate": 4.9672722978558295e-05, "loss": 0.7352, "step": 778 }, { "epoch": 0.01973784098385546, "grad_norm": 11.22861385345459, "learning_rate": 4.9671879481080355e-05, "loss": 0.4209, "step": 780 }, { "epoch": 0.01978845083253201, "grad_norm": 8.725863456726074, "learning_rate": 4.967103598360241e-05, "loss": 0.5673, "step": 782 }, { "epoch": 0.019839060681208564, "grad_norm": 6.322774410247803, "learning_rate": 4.967019248612447e-05, "loss": 0.6288, "step": 784 }, { "epoch": 0.019889670529885117, "grad_norm": 3.711097240447998, "learning_rate": 4.966934898864652e-05, "loss": 0.6343, "step": 786 }, { "epoch": 0.01994028037856167, "grad_norm": 7.519350528717041, "learning_rate": 4.966850549116858e-05, "loss": 0.7308, "step": 788 }, { "epoch": 0.019990890227238222, "grad_norm": 7.28798246383667, "learning_rate": 4.966766199369064e-05, "loss": 0.5802, "step": 790 }, { "epoch": 0.02004150007591477, "grad_norm": 5.8284783363342285, "learning_rate": 4.9666818496212696e-05, "loss": 0.5121, "step": 792 }, { "epoch": 0.020092109924591324, "grad_norm": 6.361229419708252, "learning_rate": 4.9665974998734757e-05, "loss": 0.7215, "step": 794 }, { "epoch": 0.020142719773267877, "grad_norm": 5.134431838989258, "learning_rate": 4.966513150125681e-05, "loss": 0.4415, "step": 796 }, { "epoch": 0.02019332962194443, "grad_norm": 6.047237873077393, "learning_rate": 4.966428800377887e-05, "loss": 0.8233, "step": 798 }, { "epoch": 0.020243939470620982, "grad_norm": 3.95216965675354, "learning_rate": 4.9663444506300924e-05, "loss": 0.3292, "step": 800 }, { "epoch": 0.020243939470620982, "eval_cer": 0.21470199846112256, "eval_loss": 0.33481982350349426, "eval_runtime": 2771.1883, "eval_samples_per_second": 5.656, "eval_steps_per_second": 0.354, "step": 800 }, { "epoch": 0.020294549319297535, "grad_norm": 11.878438949584961, "learning_rate": 4.9662601008822984e-05, "loss": 0.6688, "step": 802 }, { "epoch": 0.020345159167974088, "grad_norm": 4.3031392097473145, "learning_rate": 4.9661757511345044e-05, "loss": 0.5529, "step": 804 }, { "epoch": 0.02039576901665064, "grad_norm": 3.877115249633789, "learning_rate": 4.96609140138671e-05, "loss": 0.4497, "step": 806 }, { "epoch": 0.020446378865327193, "grad_norm": 4.481441497802734, "learning_rate": 4.966007051638916e-05, "loss": 0.5141, "step": 808 }, { "epoch": 0.020496988714003746, "grad_norm": 8.904651641845703, "learning_rate": 4.965922701891121e-05, "loss": 0.5108, "step": 810 }, { "epoch": 0.0205475985626803, "grad_norm": 7.119351863861084, "learning_rate": 4.965838352143327e-05, "loss": 1.0192, "step": 812 }, { "epoch": 0.02059820841135685, "grad_norm": 9.75696849822998, "learning_rate": 4.965754002395533e-05, "loss": 0.8766, "step": 814 }, { "epoch": 0.020648818260033404, "grad_norm": 3.441471576690674, "learning_rate": 4.9656696526477385e-05, "loss": 0.4397, "step": 816 }, { "epoch": 0.020699428108709957, "grad_norm": 5.399097442626953, "learning_rate": 4.9655853028999445e-05, "loss": 0.8645, "step": 818 }, { "epoch": 0.020750037957386506, "grad_norm": 8.58830738067627, "learning_rate": 4.96550095315215e-05, "loss": 0.6172, "step": 820 }, { "epoch": 0.02080064780606306, "grad_norm": 4.000862121582031, "learning_rate": 4.965416603404356e-05, "loss": 0.4244, "step": 822 }, { "epoch": 0.02085125765473961, "grad_norm": 5.518575668334961, "learning_rate": 4.965332253656562e-05, "loss": 0.7934, "step": 824 }, { "epoch": 0.020901867503416164, "grad_norm": 5.466125011444092, "learning_rate": 4.965247903908767e-05, "loss": 0.5228, "step": 826 }, { "epoch": 0.020952477352092717, "grad_norm": 8.060519218444824, "learning_rate": 4.965163554160973e-05, "loss": 0.7465, "step": 828 }, { "epoch": 0.02100308720076927, "grad_norm": 7.969659805297852, "learning_rate": 4.9650792044131786e-05, "loss": 0.7079, "step": 830 }, { "epoch": 0.021053697049445822, "grad_norm": 5.721604347229004, "learning_rate": 4.9649948546653846e-05, "loss": 0.5805, "step": 832 }, { "epoch": 0.021104306898122375, "grad_norm": 5.849686145782471, "learning_rate": 4.9649105049175906e-05, "loss": 0.7729, "step": 834 }, { "epoch": 0.021154916746798928, "grad_norm": 3.3855998516082764, "learning_rate": 4.964826155169796e-05, "loss": 0.7193, "step": 836 }, { "epoch": 0.02120552659547548, "grad_norm": 3.7108068466186523, "learning_rate": 4.964741805422002e-05, "loss": 0.5297, "step": 838 }, { "epoch": 0.021256136444152033, "grad_norm": 7.068669319152832, "learning_rate": 4.964657455674207e-05, "loss": 0.8856, "step": 840 }, { "epoch": 0.021306746292828586, "grad_norm": 3.230989694595337, "learning_rate": 4.9645731059264133e-05, "loss": 0.4015, "step": 842 }, { "epoch": 0.02135735614150514, "grad_norm": 3.4090418815612793, "learning_rate": 4.9644887561786194e-05, "loss": 0.2911, "step": 844 }, { "epoch": 0.021407965990181688, "grad_norm": 4.298464298248291, "learning_rate": 4.964404406430825e-05, "loss": 0.6512, "step": 846 }, { "epoch": 0.02145857583885824, "grad_norm": 2.939181327819824, "learning_rate": 4.964320056683031e-05, "loss": 0.4267, "step": 848 }, { "epoch": 0.021509185687534793, "grad_norm": 6.4090728759765625, "learning_rate": 4.964235706935236e-05, "loss": 0.5932, "step": 850 }, { "epoch": 0.021559795536211346, "grad_norm": 10.887248992919922, "learning_rate": 4.964151357187442e-05, "loss": 0.8047, "step": 852 }, { "epoch": 0.0216104053848879, "grad_norm": 5.774937629699707, "learning_rate": 4.964067007439648e-05, "loss": 0.7478, "step": 854 }, { "epoch": 0.02166101523356445, "grad_norm": 4.135079860687256, "learning_rate": 4.9639826576918535e-05, "loss": 0.5991, "step": 856 }, { "epoch": 0.021711625082241004, "grad_norm": 5.668390274047852, "learning_rate": 4.9638983079440595e-05, "loss": 0.4716, "step": 858 }, { "epoch": 0.021762234930917557, "grad_norm": 2.0646257400512695, "learning_rate": 4.963813958196265e-05, "loss": 0.3783, "step": 860 }, { "epoch": 0.02181284477959411, "grad_norm": 5.5543532371521, "learning_rate": 4.963729608448471e-05, "loss": 0.8937, "step": 862 }, { "epoch": 0.021863454628270662, "grad_norm": 4.02618408203125, "learning_rate": 4.963645258700677e-05, "loss": 0.4268, "step": 864 }, { "epoch": 0.021914064476947215, "grad_norm": 5.774693012237549, "learning_rate": 4.963560908952882e-05, "loss": 0.7588, "step": 866 }, { "epoch": 0.021964674325623768, "grad_norm": 3.130143880844116, "learning_rate": 4.963476559205088e-05, "loss": 0.4623, "step": 868 }, { "epoch": 0.02201528417430032, "grad_norm": 7.7838521003723145, "learning_rate": 4.9633922094572936e-05, "loss": 0.8172, "step": 870 }, { "epoch": 0.022065894022976873, "grad_norm": 7.278140544891357, "learning_rate": 4.9633078597094996e-05, "loss": 0.4937, "step": 872 }, { "epoch": 0.022116503871653422, "grad_norm": 3.3574812412261963, "learning_rate": 4.9632235099617056e-05, "loss": 0.4263, "step": 874 }, { "epoch": 0.022167113720329975, "grad_norm": 5.792145729064941, "learning_rate": 4.963139160213911e-05, "loss": 0.6564, "step": 876 }, { "epoch": 0.022217723569006528, "grad_norm": 4.801455020904541, "learning_rate": 4.963054810466117e-05, "loss": 0.4392, "step": 878 }, { "epoch": 0.02226833341768308, "grad_norm": 5.278634548187256, "learning_rate": 4.962970460718322e-05, "loss": 0.6453, "step": 880 }, { "epoch": 0.022318943266359633, "grad_norm": 4.173251628875732, "learning_rate": 4.962886110970528e-05, "loss": 0.5477, "step": 882 }, { "epoch": 0.022369553115036186, "grad_norm": 3.603672981262207, "learning_rate": 4.962801761222734e-05, "loss": 0.8842, "step": 884 }, { "epoch": 0.02242016296371274, "grad_norm": 18.358938217163086, "learning_rate": 4.96271741147494e-05, "loss": 0.8393, "step": 886 }, { "epoch": 0.02247077281238929, "grad_norm": 6.532278537750244, "learning_rate": 4.962633061727146e-05, "loss": 0.5507, "step": 888 }, { "epoch": 0.022521382661065844, "grad_norm": 6.95924711227417, "learning_rate": 4.962548711979351e-05, "loss": 0.6536, "step": 890 }, { "epoch": 0.022571992509742397, "grad_norm": 3.4727678298950195, "learning_rate": 4.962464362231557e-05, "loss": 0.435, "step": 892 }, { "epoch": 0.02262260235841895, "grad_norm": 5.473514080047607, "learning_rate": 4.9623800124837624e-05, "loss": 0.3681, "step": 894 }, { "epoch": 0.022673212207095502, "grad_norm": 7.168368339538574, "learning_rate": 4.9622956627359684e-05, "loss": 0.6088, "step": 896 }, { "epoch": 0.022723822055772055, "grad_norm": 9.777496337890625, "learning_rate": 4.9622113129881744e-05, "loss": 0.8791, "step": 898 }, { "epoch": 0.022774431904448604, "grad_norm": 4.331769943237305, "learning_rate": 4.96212696324038e-05, "loss": 0.4703, "step": 900 }, { "epoch": 0.022825041753125157, "grad_norm": 11.051033973693848, "learning_rate": 4.962042613492586e-05, "loss": 0.6771, "step": 902 }, { "epoch": 0.02287565160180171, "grad_norm": 4.339256763458252, "learning_rate": 4.961958263744791e-05, "loss": 0.3235, "step": 904 }, { "epoch": 0.022926261450478262, "grad_norm": 9.73657512664795, "learning_rate": 4.961873913996997e-05, "loss": 0.8238, "step": 906 }, { "epoch": 0.022976871299154815, "grad_norm": 8.267867088317871, "learning_rate": 4.961789564249203e-05, "loss": 0.4584, "step": 908 }, { "epoch": 0.023027481147831368, "grad_norm": 4.065835952758789, "learning_rate": 4.9617052145014085e-05, "loss": 0.5364, "step": 910 }, { "epoch": 0.02307809099650792, "grad_norm": 9.213961601257324, "learning_rate": 4.9616208647536146e-05, "loss": 0.7232, "step": 912 }, { "epoch": 0.023128700845184473, "grad_norm": 3.7316653728485107, "learning_rate": 4.96153651500582e-05, "loss": 0.417, "step": 914 }, { "epoch": 0.023179310693861026, "grad_norm": 4.829885959625244, "learning_rate": 4.961452165258026e-05, "loss": 0.4171, "step": 916 }, { "epoch": 0.02322992054253758, "grad_norm": 13.912676811218262, "learning_rate": 4.961367815510232e-05, "loss": 0.5439, "step": 918 }, { "epoch": 0.02328053039121413, "grad_norm": 6.918741226196289, "learning_rate": 4.961283465762437e-05, "loss": 0.6434, "step": 920 }, { "epoch": 0.023331140239890684, "grad_norm": 7.197675704956055, "learning_rate": 4.961199116014643e-05, "loss": 0.7114, "step": 922 }, { "epoch": 0.023381750088567237, "grad_norm": 6.958329200744629, "learning_rate": 4.9611147662668486e-05, "loss": 0.6549, "step": 924 }, { "epoch": 0.02343235993724379, "grad_norm": 7.052434921264648, "learning_rate": 4.9610304165190547e-05, "loss": 0.8423, "step": 926 }, { "epoch": 0.02348296978592034, "grad_norm": 4.447729110717773, "learning_rate": 4.960946066771261e-05, "loss": 0.4996, "step": 928 }, { "epoch": 0.02353357963459689, "grad_norm": 8.922036170959473, "learning_rate": 4.960861717023466e-05, "loss": 0.7087, "step": 930 }, { "epoch": 0.023584189483273444, "grad_norm": 5.072062969207764, "learning_rate": 4.960777367275672e-05, "loss": 0.6203, "step": 932 }, { "epoch": 0.023634799331949997, "grad_norm": 2.5875844955444336, "learning_rate": 4.9606930175278774e-05, "loss": 0.3663, "step": 934 }, { "epoch": 0.02368540918062655, "grad_norm": 5.000091552734375, "learning_rate": 4.9606086677800834e-05, "loss": 0.2929, "step": 936 }, { "epoch": 0.023736019029303102, "grad_norm": 5.237270355224609, "learning_rate": 4.9605243180322894e-05, "loss": 0.632, "step": 938 }, { "epoch": 0.023786628877979655, "grad_norm": 9.747302055358887, "learning_rate": 4.960439968284495e-05, "loss": 0.7382, "step": 940 }, { "epoch": 0.023837238726656208, "grad_norm": 7.886781215667725, "learning_rate": 4.960355618536701e-05, "loss": 0.6935, "step": 942 }, { "epoch": 0.02388784857533276, "grad_norm": 6.9037885665893555, "learning_rate": 4.960271268788906e-05, "loss": 0.631, "step": 944 }, { "epoch": 0.023938458424009313, "grad_norm": 4.556064128875732, "learning_rate": 4.960186919041112e-05, "loss": 0.5162, "step": 946 }, { "epoch": 0.023989068272685866, "grad_norm": 6.1615519523620605, "learning_rate": 4.960102569293318e-05, "loss": 0.6383, "step": 948 }, { "epoch": 0.02403967812136242, "grad_norm": 142.0865020751953, "learning_rate": 4.9600182195455235e-05, "loss": 0.7586, "step": 950 }, { "epoch": 0.02409028797003897, "grad_norm": 10.337366104125977, "learning_rate": 4.9599338697977295e-05, "loss": 0.5753, "step": 952 }, { "epoch": 0.02414089781871552, "grad_norm": 6.216306686401367, "learning_rate": 4.959849520049935e-05, "loss": 0.6553, "step": 954 }, { "epoch": 0.024191507667392073, "grad_norm": 2.8385775089263916, "learning_rate": 4.959765170302141e-05, "loss": 0.3002, "step": 956 }, { "epoch": 0.024242117516068626, "grad_norm": 5.520168304443359, "learning_rate": 4.959680820554347e-05, "loss": 0.6907, "step": 958 }, { "epoch": 0.02429272736474518, "grad_norm": 3.9114444255828857, "learning_rate": 4.959596470806552e-05, "loss": 0.419, "step": 960 }, { "epoch": 0.02434333721342173, "grad_norm": 6.165211200714111, "learning_rate": 4.959512121058758e-05, "loss": 0.6344, "step": 962 }, { "epoch": 0.024393947062098284, "grad_norm": 4.263137340545654, "learning_rate": 4.9594277713109636e-05, "loss": 0.5867, "step": 964 }, { "epoch": 0.024444556910774837, "grad_norm": 2.720306873321533, "learning_rate": 4.9593434215631696e-05, "loss": 0.3657, "step": 966 }, { "epoch": 0.02449516675945139, "grad_norm": 6.873605251312256, "learning_rate": 4.959259071815375e-05, "loss": 0.6407, "step": 968 }, { "epoch": 0.024545776608127942, "grad_norm": 16.427352905273438, "learning_rate": 4.959174722067581e-05, "loss": 0.8508, "step": 970 }, { "epoch": 0.024596386456804495, "grad_norm": 8.866124153137207, "learning_rate": 4.959090372319787e-05, "loss": 0.803, "step": 972 }, { "epoch": 0.024646996305481048, "grad_norm": 5.541032791137695, "learning_rate": 4.9590060225719924e-05, "loss": 0.6485, "step": 974 }, { "epoch": 0.0246976061541576, "grad_norm": 2.8217546939849854, "learning_rate": 4.9589216728241984e-05, "loss": 0.3261, "step": 976 }, { "epoch": 0.024748216002834153, "grad_norm": 6.076355457305908, "learning_rate": 4.958837323076404e-05, "loss": 0.5189, "step": 978 }, { "epoch": 0.024798825851510702, "grad_norm": 2.687714099884033, "learning_rate": 4.95875297332861e-05, "loss": 0.3242, "step": 980 }, { "epoch": 0.024849435700187255, "grad_norm": 6.32763671875, "learning_rate": 4.958668623580816e-05, "loss": 0.6136, "step": 982 }, { "epoch": 0.024900045548863808, "grad_norm": 11.270010948181152, "learning_rate": 4.958584273833021e-05, "loss": 0.8674, "step": 984 }, { "epoch": 0.02495065539754036, "grad_norm": 5.883991718292236, "learning_rate": 4.958499924085227e-05, "loss": 0.5242, "step": 986 }, { "epoch": 0.025001265246216913, "grad_norm": 6.999844551086426, "learning_rate": 4.9584155743374325e-05, "loss": 0.5351, "step": 988 }, { "epoch": 0.025051875094893466, "grad_norm": 6.340963363647461, "learning_rate": 4.9583312245896385e-05, "loss": 0.5624, "step": 990 }, { "epoch": 0.02510248494357002, "grad_norm": 7.818021774291992, "learning_rate": 4.9582468748418445e-05, "loss": 0.9044, "step": 992 }, { "epoch": 0.02515309479224657, "grad_norm": 6.447050094604492, "learning_rate": 4.95816252509405e-05, "loss": 0.7352, "step": 994 }, { "epoch": 0.025203704640923124, "grad_norm": 4.902888774871826, "learning_rate": 4.958078175346256e-05, "loss": 0.4516, "step": 996 }, { "epoch": 0.025254314489599677, "grad_norm": 71.4312973022461, "learning_rate": 4.957993825598461e-05, "loss": 0.6593, "step": 998 }, { "epoch": 0.02530492433827623, "grad_norm": 5.481997489929199, "learning_rate": 4.957909475850667e-05, "loss": 0.4801, "step": 1000 }, { "epoch": 0.02530492433827623, "eval_cer": 0.20989040697069894, "eval_loss": 0.3260483741760254, "eval_runtime": 2649.7137, "eval_samples_per_second": 5.915, "eval_steps_per_second": 0.37, "step": 1000 }, { "epoch": 0.025355534186952782, "grad_norm": 13.54730224609375, "learning_rate": 4.957825126102873e-05, "loss": 0.5685, "step": 1002 }, { "epoch": 0.025406144035629335, "grad_norm": 8.101643562316895, "learning_rate": 4.9577407763550786e-05, "loss": 0.6376, "step": 1004 }, { "epoch": 0.025456753884305888, "grad_norm": 5.483541965484619, "learning_rate": 4.9576564266072846e-05, "loss": 0.5417, "step": 1006 }, { "epoch": 0.025507363732982437, "grad_norm": 5.6926493644714355, "learning_rate": 4.95757207685949e-05, "loss": 0.5145, "step": 1008 }, { "epoch": 0.02555797358165899, "grad_norm": 4.287508487701416, "learning_rate": 4.957487727111696e-05, "loss": 0.5817, "step": 1010 }, { "epoch": 0.025608583430335542, "grad_norm": 3.4692742824554443, "learning_rate": 4.957403377363902e-05, "loss": 0.4781, "step": 1012 }, { "epoch": 0.025659193279012095, "grad_norm": 6.066647052764893, "learning_rate": 4.957319027616107e-05, "loss": 0.5475, "step": 1014 }, { "epoch": 0.025709803127688648, "grad_norm": 3.1936872005462646, "learning_rate": 4.9572346778683134e-05, "loss": 0.6294, "step": 1016 }, { "epoch": 0.0257604129763652, "grad_norm": 10.364377975463867, "learning_rate": 4.957150328120519e-05, "loss": 0.7561, "step": 1018 }, { "epoch": 0.025811022825041753, "grad_norm": 5.162211894989014, "learning_rate": 4.957065978372725e-05, "loss": 0.5821, "step": 1020 }, { "epoch": 0.025861632673718306, "grad_norm": 8.685132026672363, "learning_rate": 4.956981628624931e-05, "loss": 0.7048, "step": 1022 }, { "epoch": 0.02591224252239486, "grad_norm": 4.4980316162109375, "learning_rate": 4.956897278877136e-05, "loss": 0.5405, "step": 1024 }, { "epoch": 0.02596285237107141, "grad_norm": 5.45166540145874, "learning_rate": 4.956812929129342e-05, "loss": 0.3885, "step": 1026 }, { "epoch": 0.026013462219747964, "grad_norm": 7.379059314727783, "learning_rate": 4.9567285793815474e-05, "loss": 0.5148, "step": 1028 }, { "epoch": 0.026064072068424517, "grad_norm": 4.766157150268555, "learning_rate": 4.9566442296337535e-05, "loss": 0.7287, "step": 1030 }, { "epoch": 0.02611468191710107, "grad_norm": 7.131243705749512, "learning_rate": 4.9565598798859595e-05, "loss": 0.6868, "step": 1032 }, { "epoch": 0.02616529176577762, "grad_norm": 9.707527160644531, "learning_rate": 4.956475530138165e-05, "loss": 0.6592, "step": 1034 }, { "epoch": 0.02621590161445417, "grad_norm": 5.630888938903809, "learning_rate": 4.956391180390371e-05, "loss": 0.5028, "step": 1036 }, { "epoch": 0.026266511463130724, "grad_norm": 3.7485404014587402, "learning_rate": 4.956306830642576e-05, "loss": 0.3452, "step": 1038 }, { "epoch": 0.026317121311807277, "grad_norm": 6.3416643142700195, "learning_rate": 4.956222480894782e-05, "loss": 0.4316, "step": 1040 }, { "epoch": 0.02636773116048383, "grad_norm": 8.458013534545898, "learning_rate": 4.956138131146988e-05, "loss": 0.8142, "step": 1042 }, { "epoch": 0.026418341009160382, "grad_norm": 9.704322814941406, "learning_rate": 4.9560537813991936e-05, "loss": 0.7614, "step": 1044 }, { "epoch": 0.026468950857836935, "grad_norm": 3.311298370361328, "learning_rate": 4.9559694316513996e-05, "loss": 0.5417, "step": 1046 }, { "epoch": 0.026519560706513488, "grad_norm": 7.961212635040283, "learning_rate": 4.955885081903605e-05, "loss": 0.6401, "step": 1048 }, { "epoch": 0.02657017055519004, "grad_norm": 6.631721496582031, "learning_rate": 4.955800732155811e-05, "loss": 0.5357, "step": 1050 }, { "epoch": 0.026620780403866593, "grad_norm": 5.143334865570068, "learning_rate": 4.955716382408016e-05, "loss": 0.545, "step": 1052 }, { "epoch": 0.026671390252543146, "grad_norm": 8.77175235748291, "learning_rate": 4.955632032660222e-05, "loss": 0.5938, "step": 1054 }, { "epoch": 0.0267220001012197, "grad_norm": 10.350188255310059, "learning_rate": 4.955547682912428e-05, "loss": 0.9662, "step": 1056 }, { "epoch": 0.02677260994989625, "grad_norm": 6.271733283996582, "learning_rate": 4.955463333164634e-05, "loss": 0.5715, "step": 1058 }, { "epoch": 0.026823219798572804, "grad_norm": 5.548452854156494, "learning_rate": 4.95537898341684e-05, "loss": 0.5213, "step": 1060 }, { "epoch": 0.026873829647249353, "grad_norm": 5.460413455963135, "learning_rate": 4.955294633669045e-05, "loss": 0.3746, "step": 1062 }, { "epoch": 0.026924439495925906, "grad_norm": 10.801025390625, "learning_rate": 4.955210283921251e-05, "loss": 0.66, "step": 1064 }, { "epoch": 0.02697504934460246, "grad_norm": 35.335445404052734, "learning_rate": 4.955125934173457e-05, "loss": 0.5122, "step": 1066 }, { "epoch": 0.02702565919327901, "grad_norm": 3.974865436553955, "learning_rate": 4.9550415844256624e-05, "loss": 0.3642, "step": 1068 }, { "epoch": 0.027076269041955564, "grad_norm": 6.508527755737305, "learning_rate": 4.9549572346778684e-05, "loss": 0.6126, "step": 1070 }, { "epoch": 0.027126878890632117, "grad_norm": 11.70407772064209, "learning_rate": 4.954872884930074e-05, "loss": 0.4939, "step": 1072 }, { "epoch": 0.02717748873930867, "grad_norm": 9.648119926452637, "learning_rate": 4.95478853518228e-05, "loss": 0.6462, "step": 1074 }, { "epoch": 0.027228098587985222, "grad_norm": 8.660693168640137, "learning_rate": 4.954704185434486e-05, "loss": 0.5923, "step": 1076 }, { "epoch": 0.027278708436661775, "grad_norm": 17.504438400268555, "learning_rate": 4.954619835686691e-05, "loss": 0.9972, "step": 1078 }, { "epoch": 0.027329318285338328, "grad_norm": 6.019506454467773, "learning_rate": 4.954535485938897e-05, "loss": 0.5884, "step": 1080 }, { "epoch": 0.02737992813401488, "grad_norm": 11.403207778930664, "learning_rate": 4.9544511361911025e-05, "loss": 0.7651, "step": 1082 }, { "epoch": 0.027430537982691433, "grad_norm": 3.842545747756958, "learning_rate": 4.9543667864433085e-05, "loss": 0.3572, "step": 1084 }, { "epoch": 0.027481147831367986, "grad_norm": 4.691946029663086, "learning_rate": 4.9542824366955146e-05, "loss": 0.556, "step": 1086 }, { "epoch": 0.027531757680044535, "grad_norm": 8.138900756835938, "learning_rate": 4.95419808694772e-05, "loss": 0.545, "step": 1088 }, { "epoch": 0.027582367528721088, "grad_norm": 7.7400431632995605, "learning_rate": 4.954113737199926e-05, "loss": 0.6399, "step": 1090 }, { "epoch": 0.02763297737739764, "grad_norm": 18.444286346435547, "learning_rate": 4.954029387452131e-05, "loss": 0.9819, "step": 1092 }, { "epoch": 0.027683587226074193, "grad_norm": 4.818946838378906, "learning_rate": 4.953945037704337e-05, "loss": 0.5091, "step": 1094 }, { "epoch": 0.027734197074750746, "grad_norm": 6.969218730926514, "learning_rate": 4.953860687956543e-05, "loss": 0.6144, "step": 1096 }, { "epoch": 0.0277848069234273, "grad_norm": 5.701696395874023, "learning_rate": 4.9537763382087486e-05, "loss": 0.6378, "step": 1098 }, { "epoch": 0.02783541677210385, "grad_norm": 4.948043346405029, "learning_rate": 4.953691988460955e-05, "loss": 0.61, "step": 1100 }, { "epoch": 0.027886026620780404, "grad_norm": 6.133516788482666, "learning_rate": 4.95360763871316e-05, "loss": 0.5396, "step": 1102 }, { "epoch": 0.027936636469456957, "grad_norm": 4.206554889678955, "learning_rate": 4.953523288965366e-05, "loss": 0.3066, "step": 1104 }, { "epoch": 0.02798724631813351, "grad_norm": 3.7985496520996094, "learning_rate": 4.953438939217572e-05, "loss": 0.5389, "step": 1106 }, { "epoch": 0.028037856166810062, "grad_norm": 6.991200923919678, "learning_rate": 4.9533545894697774e-05, "loss": 0.7457, "step": 1108 }, { "epoch": 0.028088466015486615, "grad_norm": 5.731369495391846, "learning_rate": 4.9532702397219834e-05, "loss": 0.3976, "step": 1110 }, { "epoch": 0.028139075864163168, "grad_norm": 7.29412317276001, "learning_rate": 4.953185889974189e-05, "loss": 0.56, "step": 1112 }, { "epoch": 0.02818968571283972, "grad_norm": 10.055305480957031, "learning_rate": 4.953101540226395e-05, "loss": 0.8518, "step": 1114 }, { "epoch": 0.02824029556151627, "grad_norm": 5.1682047843933105, "learning_rate": 4.953017190478601e-05, "loss": 0.3921, "step": 1116 }, { "epoch": 0.028290905410192822, "grad_norm": 7.5293049812316895, "learning_rate": 4.952932840730806e-05, "loss": 0.732, "step": 1118 }, { "epoch": 0.028341515258869375, "grad_norm": 7.173330783843994, "learning_rate": 4.952848490983012e-05, "loss": 0.6786, "step": 1120 }, { "epoch": 0.028392125107545928, "grad_norm": 7.453824520111084, "learning_rate": 4.9527641412352175e-05, "loss": 0.6836, "step": 1122 }, { "epoch": 0.02844273495622248, "grad_norm": 16.70603370666504, "learning_rate": 4.9526797914874235e-05, "loss": 0.8831, "step": 1124 }, { "epoch": 0.028493344804899033, "grad_norm": 8.2377290725708, "learning_rate": 4.952595441739629e-05, "loss": 0.8214, "step": 1126 }, { "epoch": 0.028543954653575586, "grad_norm": 4.513237953186035, "learning_rate": 4.952511091991835e-05, "loss": 0.7878, "step": 1128 }, { "epoch": 0.02859456450225214, "grad_norm": 6.024347305297852, "learning_rate": 4.952426742244041e-05, "loss": 0.2515, "step": 1130 }, { "epoch": 0.02864517435092869, "grad_norm": 7.672776699066162, "learning_rate": 4.952342392496246e-05, "loss": 0.9578, "step": 1132 }, { "epoch": 0.028695784199605244, "grad_norm": 11.622359275817871, "learning_rate": 4.952258042748452e-05, "loss": 0.6932, "step": 1134 }, { "epoch": 0.028746394048281797, "grad_norm": 6.2994232177734375, "learning_rate": 4.9521736930006576e-05, "loss": 0.5317, "step": 1136 }, { "epoch": 0.02879700389695835, "grad_norm": 8.886155128479004, "learning_rate": 4.9520893432528636e-05, "loss": 0.9717, "step": 1138 }, { "epoch": 0.028847613745634902, "grad_norm": 3.2561750411987305, "learning_rate": 4.9520049935050696e-05, "loss": 0.3471, "step": 1140 }, { "epoch": 0.02889822359431145, "grad_norm": 6.805208683013916, "learning_rate": 4.951920643757275e-05, "loss": 0.4706, "step": 1142 }, { "epoch": 0.028948833442988004, "grad_norm": 6.713986396789551, "learning_rate": 4.951836294009481e-05, "loss": 0.7055, "step": 1144 }, { "epoch": 0.028999443291664557, "grad_norm": 5.203835964202881, "learning_rate": 4.9517519442616863e-05, "loss": 0.6585, "step": 1146 }, { "epoch": 0.02905005314034111, "grad_norm": 3.168962001800537, "learning_rate": 4.9516675945138924e-05, "loss": 0.4794, "step": 1148 }, { "epoch": 0.029100662989017662, "grad_norm": 11.134737014770508, "learning_rate": 4.9515832447660984e-05, "loss": 0.899, "step": 1150 }, { "epoch": 0.029151272837694215, "grad_norm": 5.522861480712891, "learning_rate": 4.951498895018304e-05, "loss": 0.4002, "step": 1152 }, { "epoch": 0.029201882686370768, "grad_norm": 2.908127784729004, "learning_rate": 4.95141454527051e-05, "loss": 0.4008, "step": 1154 }, { "epoch": 0.02925249253504732, "grad_norm": 11.76309871673584, "learning_rate": 4.951330195522715e-05, "loss": 0.3493, "step": 1156 }, { "epoch": 0.029303102383723873, "grad_norm": 4.505433082580566, "learning_rate": 4.951245845774921e-05, "loss": 0.3181, "step": 1158 }, { "epoch": 0.029353712232400426, "grad_norm": 4.662505149841309, "learning_rate": 4.951161496027127e-05, "loss": 0.6274, "step": 1160 }, { "epoch": 0.02940432208107698, "grad_norm": 4.870098114013672, "learning_rate": 4.9510771462793325e-05, "loss": 0.5994, "step": 1162 }, { "epoch": 0.02945493192975353, "grad_norm": 4.153433322906494, "learning_rate": 4.9509927965315385e-05, "loss": 0.3769, "step": 1164 }, { "epoch": 0.029505541778430084, "grad_norm": 8.398072242736816, "learning_rate": 4.950908446783744e-05, "loss": 0.8386, "step": 1166 }, { "epoch": 0.029556151627106633, "grad_norm": 5.86764669418335, "learning_rate": 4.95082409703595e-05, "loss": 0.3906, "step": 1168 }, { "epoch": 0.029606761475783186, "grad_norm": 4.7525200843811035, "learning_rate": 4.950739747288156e-05, "loss": 0.5805, "step": 1170 }, { "epoch": 0.02965737132445974, "grad_norm": 14.163674354553223, "learning_rate": 4.950655397540361e-05, "loss": 0.714, "step": 1172 }, { "epoch": 0.02970798117313629, "grad_norm": 6.36476469039917, "learning_rate": 4.950571047792567e-05, "loss": 0.7714, "step": 1174 }, { "epoch": 0.029758591021812844, "grad_norm": 3.8334286212921143, "learning_rate": 4.9504866980447726e-05, "loss": 0.4451, "step": 1176 }, { "epoch": 0.029809200870489397, "grad_norm": 3.1083927154541016, "learning_rate": 4.9504023482969786e-05, "loss": 0.3391, "step": 1178 }, { "epoch": 0.02985981071916595, "grad_norm": 11.239181518554688, "learning_rate": 4.9503179985491846e-05, "loss": 0.6345, "step": 1180 }, { "epoch": 0.029910420567842502, "grad_norm": 6.135115623474121, "learning_rate": 4.95023364880139e-05, "loss": 0.8366, "step": 1182 }, { "epoch": 0.029961030416519055, "grad_norm": 3.8036446571350098, "learning_rate": 4.950149299053596e-05, "loss": 0.4256, "step": 1184 }, { "epoch": 0.030011640265195608, "grad_norm": 11.396409034729004, "learning_rate": 4.950064949305801e-05, "loss": 0.5988, "step": 1186 }, { "epoch": 0.03006225011387216, "grad_norm": 6.251826286315918, "learning_rate": 4.949980599558007e-05, "loss": 0.4033, "step": 1188 }, { "epoch": 0.030112859962548713, "grad_norm": 6.115148544311523, "learning_rate": 4.9498962498102134e-05, "loss": 0.4699, "step": 1190 }, { "epoch": 0.030163469811225266, "grad_norm": 8.004677772521973, "learning_rate": 4.949811900062419e-05, "loss": 0.7723, "step": 1192 }, { "epoch": 0.03021407965990182, "grad_norm": 8.397205352783203, "learning_rate": 4.949727550314625e-05, "loss": 0.4565, "step": 1194 }, { "epoch": 0.030264689508578368, "grad_norm": 7.252643585205078, "learning_rate": 4.94964320056683e-05, "loss": 0.7132, "step": 1196 }, { "epoch": 0.03031529935725492, "grad_norm": 5.881059169769287, "learning_rate": 4.949558850819036e-05, "loss": 0.7198, "step": 1198 }, { "epoch": 0.030365909205931473, "grad_norm": 11.146910667419434, "learning_rate": 4.949474501071242e-05, "loss": 0.5668, "step": 1200 }, { "epoch": 0.030365909205931473, "eval_cer": 0.2020166573086281, "eval_loss": 0.3112943768501282, "eval_runtime": 2732.7913, "eval_samples_per_second": 5.736, "eval_steps_per_second": 0.359, "step": 1200 }, { "epoch": 0.030416519054608026, "grad_norm": 5.239530563354492, "learning_rate": 4.9493901513234474e-05, "loss": 0.8572, "step": 1202 }, { "epoch": 0.03046712890328458, "grad_norm": 6.763473987579346, "learning_rate": 4.9493058015756535e-05, "loss": 0.4804, "step": 1204 }, { "epoch": 0.03051773875196113, "grad_norm": 15.63022232055664, "learning_rate": 4.949221451827859e-05, "loss": 0.6519, "step": 1206 }, { "epoch": 0.030568348600637684, "grad_norm": 6.332169055938721, "learning_rate": 4.949137102080065e-05, "loss": 0.4457, "step": 1208 }, { "epoch": 0.030618958449314237, "grad_norm": 8.737252235412598, "learning_rate": 4.94905275233227e-05, "loss": 0.6122, "step": 1210 }, { "epoch": 0.03066956829799079, "grad_norm": 9.297114372253418, "learning_rate": 4.948968402584476e-05, "loss": 0.7155, "step": 1212 }, { "epoch": 0.030720178146667342, "grad_norm": 7.846271991729736, "learning_rate": 4.948884052836682e-05, "loss": 0.7046, "step": 1214 }, { "epoch": 0.030770787995343895, "grad_norm": 9.104436874389648, "learning_rate": 4.9487997030888875e-05, "loss": 0.5831, "step": 1216 }, { "epoch": 0.030821397844020448, "grad_norm": 4.265629768371582, "learning_rate": 4.9487153533410936e-05, "loss": 0.2498, "step": 1218 }, { "epoch": 0.030872007692697, "grad_norm": 3.7431282997131348, "learning_rate": 4.948631003593299e-05, "loss": 0.6634, "step": 1220 }, { "epoch": 0.03092261754137355, "grad_norm": 5.563877105712891, "learning_rate": 4.948546653845505e-05, "loss": 0.7814, "step": 1222 }, { "epoch": 0.030973227390050102, "grad_norm": 5.606101036071777, "learning_rate": 4.948462304097711e-05, "loss": 0.7189, "step": 1224 }, { "epoch": 0.031023837238726655, "grad_norm": 9.97654914855957, "learning_rate": 4.948377954349916e-05, "loss": 0.7406, "step": 1226 }, { "epoch": 0.031074447087403208, "grad_norm": 10.628059387207031, "learning_rate": 4.948293604602122e-05, "loss": 0.904, "step": 1228 }, { "epoch": 0.03112505693607976, "grad_norm": 2.7047297954559326, "learning_rate": 4.9482092548543277e-05, "loss": 0.3665, "step": 1230 }, { "epoch": 0.031175666784756313, "grad_norm": 8.546875953674316, "learning_rate": 4.948124905106534e-05, "loss": 0.5705, "step": 1232 }, { "epoch": 0.031226276633432866, "grad_norm": 6.664468765258789, "learning_rate": 4.94804055535874e-05, "loss": 0.518, "step": 1234 }, { "epoch": 0.03127688648210942, "grad_norm": 8.334696769714355, "learning_rate": 4.947956205610945e-05, "loss": 0.5282, "step": 1236 }, { "epoch": 0.03132749633078597, "grad_norm": 7.652597427368164, "learning_rate": 4.947871855863151e-05, "loss": 0.6678, "step": 1238 }, { "epoch": 0.031378106179462524, "grad_norm": 5.162440776824951, "learning_rate": 4.9477875061153564e-05, "loss": 0.4045, "step": 1240 }, { "epoch": 0.03142871602813908, "grad_norm": 6.496342658996582, "learning_rate": 4.9477031563675624e-05, "loss": 0.4663, "step": 1242 }, { "epoch": 0.03147932587681563, "grad_norm": 7.366580963134766, "learning_rate": 4.9476188066197684e-05, "loss": 0.6894, "step": 1244 }, { "epoch": 0.03152993572549218, "grad_norm": 13.762931823730469, "learning_rate": 4.947534456871974e-05, "loss": 0.7944, "step": 1246 }, { "epoch": 0.031580545574168735, "grad_norm": 12.880459785461426, "learning_rate": 4.94745010712418e-05, "loss": 0.7428, "step": 1248 }, { "epoch": 0.03163115542284529, "grad_norm": 6.380914211273193, "learning_rate": 4.947365757376385e-05, "loss": 0.6661, "step": 1250 }, { "epoch": 0.03168176527152184, "grad_norm": 7.456797122955322, "learning_rate": 4.947281407628591e-05, "loss": 0.7024, "step": 1252 }, { "epoch": 0.03173237512019839, "grad_norm": 4.948566913604736, "learning_rate": 4.947197057880797e-05, "loss": 0.5951, "step": 1254 }, { "epoch": 0.031782984968874946, "grad_norm": 7.60888147354126, "learning_rate": 4.9471127081330025e-05, "loss": 0.4111, "step": 1256 }, { "epoch": 0.0318335948175515, "grad_norm": 5.659879684448242, "learning_rate": 4.9470283583852085e-05, "loss": 0.6262, "step": 1258 }, { "epoch": 0.03188420466622805, "grad_norm": 6.134880542755127, "learning_rate": 4.946944008637414e-05, "loss": 0.5249, "step": 1260 }, { "epoch": 0.0319348145149046, "grad_norm": 6.531559467315674, "learning_rate": 4.94685965888962e-05, "loss": 0.5775, "step": 1262 }, { "epoch": 0.03198542436358115, "grad_norm": 12.462188720703125, "learning_rate": 4.946775309141826e-05, "loss": 0.6286, "step": 1264 }, { "epoch": 0.0320360342122577, "grad_norm": 2.809241533279419, "learning_rate": 4.946690959394031e-05, "loss": 0.3842, "step": 1266 }, { "epoch": 0.032086644060934255, "grad_norm": 5.521834850311279, "learning_rate": 4.946606609646237e-05, "loss": 0.4794, "step": 1268 }, { "epoch": 0.03213725390961081, "grad_norm": 5.101696491241455, "learning_rate": 4.9465222598984426e-05, "loss": 0.4713, "step": 1270 }, { "epoch": 0.03218786375828736, "grad_norm": 6.392932891845703, "learning_rate": 4.9464379101506486e-05, "loss": 0.6618, "step": 1272 }, { "epoch": 0.03223847360696391, "grad_norm": 4.8280768394470215, "learning_rate": 4.946353560402855e-05, "loss": 0.3516, "step": 1274 }, { "epoch": 0.032289083455640466, "grad_norm": 3.3262429237365723, "learning_rate": 4.94626921065506e-05, "loss": 0.5898, "step": 1276 }, { "epoch": 0.03233969330431702, "grad_norm": 5.624804496765137, "learning_rate": 4.946184860907266e-05, "loss": 0.5552, "step": 1278 }, { "epoch": 0.03239030315299357, "grad_norm": 3.1129558086395264, "learning_rate": 4.9461005111594714e-05, "loss": 0.6172, "step": 1280 }, { "epoch": 0.032440913001670124, "grad_norm": 3.5634653568267822, "learning_rate": 4.9460161614116774e-05, "loss": 0.3858, "step": 1282 }, { "epoch": 0.03249152285034668, "grad_norm": 7.236937046051025, "learning_rate": 4.9459318116638834e-05, "loss": 0.6032, "step": 1284 }, { "epoch": 0.03254213269902323, "grad_norm": 3.344967842102051, "learning_rate": 4.945847461916089e-05, "loss": 0.5774, "step": 1286 }, { "epoch": 0.03259274254769978, "grad_norm": 3.4330480098724365, "learning_rate": 4.945763112168295e-05, "loss": 0.7142, "step": 1288 }, { "epoch": 0.032643352396376335, "grad_norm": 5.767812728881836, "learning_rate": 4.9456787624205e-05, "loss": 0.5658, "step": 1290 }, { "epoch": 0.03269396224505289, "grad_norm": 5.187238693237305, "learning_rate": 4.945594412672707e-05, "loss": 0.346, "step": 1292 }, { "epoch": 0.03274457209372944, "grad_norm": 5.828567028045654, "learning_rate": 4.945510062924912e-05, "loss": 0.3802, "step": 1294 }, { "epoch": 0.03279518194240599, "grad_norm": 6.394260406494141, "learning_rate": 4.945425713177118e-05, "loss": 1.0168, "step": 1296 }, { "epoch": 0.032845791791082546, "grad_norm": 4.123929977416992, "learning_rate": 4.945341363429324e-05, "loss": 0.4564, "step": 1298 }, { "epoch": 0.0328964016397591, "grad_norm": 6.843530178070068, "learning_rate": 4.9452570136815295e-05, "loss": 0.6159, "step": 1300 }, { "epoch": 0.03294701148843565, "grad_norm": 10.560795783996582, "learning_rate": 4.9451726639337356e-05, "loss": 0.517, "step": 1302 }, { "epoch": 0.032997621337112204, "grad_norm": 2.8675217628479004, "learning_rate": 4.945088314185941e-05, "loss": 0.6848, "step": 1304 }, { "epoch": 0.03304823118578876, "grad_norm": 3.9155211448669434, "learning_rate": 4.945003964438147e-05, "loss": 0.6188, "step": 1306 }, { "epoch": 0.03309884103446531, "grad_norm": 6.229773998260498, "learning_rate": 4.944919614690352e-05, "loss": 0.4295, "step": 1308 }, { "epoch": 0.03314945088314186, "grad_norm": 4.002541542053223, "learning_rate": 4.944835264942558e-05, "loss": 0.5816, "step": 1310 }, { "epoch": 0.033200060731818415, "grad_norm": 6.393809795379639, "learning_rate": 4.944750915194764e-05, "loss": 0.8652, "step": 1312 }, { "epoch": 0.03325067058049497, "grad_norm": 3.2906885147094727, "learning_rate": 4.9446665654469696e-05, "loss": 0.6938, "step": 1314 }, { "epoch": 0.03330128042917151, "grad_norm": 2.870098352432251, "learning_rate": 4.944582215699176e-05, "loss": 0.2623, "step": 1316 }, { "epoch": 0.033351890277848066, "grad_norm": 5.652560710906982, "learning_rate": 4.944497865951381e-05, "loss": 0.5474, "step": 1318 }, { "epoch": 0.03340250012652462, "grad_norm": 12.31924819946289, "learning_rate": 4.944413516203587e-05, "loss": 0.7774, "step": 1320 }, { "epoch": 0.03345310997520117, "grad_norm": 4.865198135375977, "learning_rate": 4.944329166455793e-05, "loss": 0.2116, "step": 1322 }, { "epoch": 0.033503719823877724, "grad_norm": 4.536335468292236, "learning_rate": 4.9442448167079984e-05, "loss": 0.541, "step": 1324 }, { "epoch": 0.03355432967255428, "grad_norm": 14.479270935058594, "learning_rate": 4.9441604669602044e-05, "loss": 0.3704, "step": 1326 }, { "epoch": 0.03360493952123083, "grad_norm": 7.071706295013428, "learning_rate": 4.94407611721241e-05, "loss": 0.5868, "step": 1328 }, { "epoch": 0.03365554936990738, "grad_norm": 12.812637329101562, "learning_rate": 4.943991767464616e-05, "loss": 0.7385, "step": 1330 }, { "epoch": 0.033706159218583935, "grad_norm": 10.205668449401855, "learning_rate": 4.943907417716822e-05, "loss": 0.5933, "step": 1332 }, { "epoch": 0.03375676906726049, "grad_norm": 6.436938762664795, "learning_rate": 4.943823067969027e-05, "loss": 1.0516, "step": 1334 }, { "epoch": 0.03380737891593704, "grad_norm": 4.715915679931641, "learning_rate": 4.943738718221233e-05, "loss": 0.5055, "step": 1336 }, { "epoch": 0.03385798876461359, "grad_norm": 5.58711051940918, "learning_rate": 4.9436543684734385e-05, "loss": 0.4184, "step": 1338 }, { "epoch": 0.033908598613290146, "grad_norm": 16.581960678100586, "learning_rate": 4.9435700187256445e-05, "loss": 0.478, "step": 1340 }, { "epoch": 0.0339592084619667, "grad_norm": 7.326168060302734, "learning_rate": 4.9434856689778505e-05, "loss": 0.7105, "step": 1342 }, { "epoch": 0.03400981831064325, "grad_norm": 4.85835075378418, "learning_rate": 4.943401319230056e-05, "loss": 0.7862, "step": 1344 }, { "epoch": 0.034060428159319804, "grad_norm": 3.9227964878082275, "learning_rate": 4.943316969482262e-05, "loss": 0.5298, "step": 1346 }, { "epoch": 0.03411103800799636, "grad_norm": 6.2520904541015625, "learning_rate": 4.943232619734467e-05, "loss": 0.5795, "step": 1348 }, { "epoch": 0.03416164785667291, "grad_norm": 6.502024173736572, "learning_rate": 4.943148269986673e-05, "loss": 0.4504, "step": 1350 }, { "epoch": 0.03421225770534946, "grad_norm": 6.299930095672607, "learning_rate": 4.943063920238879e-05, "loss": 0.8499, "step": 1352 }, { "epoch": 0.034262867554026015, "grad_norm": 3.0065736770629883, "learning_rate": 4.9429795704910846e-05, "loss": 0.3785, "step": 1354 }, { "epoch": 0.03431347740270257, "grad_norm": 6.053264617919922, "learning_rate": 4.9428952207432906e-05, "loss": 0.4405, "step": 1356 }, { "epoch": 0.03436408725137912, "grad_norm": 2.8478128910064697, "learning_rate": 4.942810870995496e-05, "loss": 0.3576, "step": 1358 }, { "epoch": 0.03441469710005567, "grad_norm": 4.3910369873046875, "learning_rate": 4.942726521247702e-05, "loss": 0.6643, "step": 1360 }, { "epoch": 0.034465306948732226, "grad_norm": 8.032602310180664, "learning_rate": 4.942642171499908e-05, "loss": 0.3172, "step": 1362 }, { "epoch": 0.03451591679740878, "grad_norm": 4.423160552978516, "learning_rate": 4.9425578217521134e-05, "loss": 0.5107, "step": 1364 }, { "epoch": 0.03456652664608533, "grad_norm": 4.310211181640625, "learning_rate": 4.9424734720043194e-05, "loss": 0.7117, "step": 1366 }, { "epoch": 0.034617136494761884, "grad_norm": 7.594289779663086, "learning_rate": 4.942389122256525e-05, "loss": 0.3758, "step": 1368 }, { "epoch": 0.03466774634343843, "grad_norm": 10.414362907409668, "learning_rate": 4.942304772508731e-05, "loss": 0.9048, "step": 1370 }, { "epoch": 0.03471835619211498, "grad_norm": 5.809612274169922, "learning_rate": 4.942220422760937e-05, "loss": 0.4148, "step": 1372 }, { "epoch": 0.034768966040791535, "grad_norm": 7.8530449867248535, "learning_rate": 4.942136073013142e-05, "loss": 0.6252, "step": 1374 }, { "epoch": 0.03481957588946809, "grad_norm": 9.151339530944824, "learning_rate": 4.942051723265348e-05, "loss": 0.6158, "step": 1376 }, { "epoch": 0.03487018573814464, "grad_norm": 20.164710998535156, "learning_rate": 4.9419673735175535e-05, "loss": 0.6702, "step": 1378 }, { "epoch": 0.03492079558682119, "grad_norm": 6.756007194519043, "learning_rate": 4.9418830237697595e-05, "loss": 0.2371, "step": 1380 }, { "epoch": 0.034971405435497746, "grad_norm": 6.241485595703125, "learning_rate": 4.9417986740219655e-05, "loss": 0.7635, "step": 1382 }, { "epoch": 0.0350220152841743, "grad_norm": 10.091927528381348, "learning_rate": 4.941714324274171e-05, "loss": 0.8506, "step": 1384 }, { "epoch": 0.03507262513285085, "grad_norm": 7.4213080406188965, "learning_rate": 4.941629974526377e-05, "loss": 0.5124, "step": 1386 }, { "epoch": 0.035123234981527404, "grad_norm": 11.214619636535645, "learning_rate": 4.941545624778582e-05, "loss": 0.4202, "step": 1388 }, { "epoch": 0.03517384483020396, "grad_norm": 8.220820426940918, "learning_rate": 4.941461275030788e-05, "loss": 0.9393, "step": 1390 }, { "epoch": 0.03522445467888051, "grad_norm": 8.862229347229004, "learning_rate": 4.9413769252829936e-05, "loss": 0.813, "step": 1392 }, { "epoch": 0.03527506452755706, "grad_norm": 3.892012596130371, "learning_rate": 4.9412925755351996e-05, "loss": 0.4279, "step": 1394 }, { "epoch": 0.035325674376233615, "grad_norm": 13.525033950805664, "learning_rate": 4.9412082257874056e-05, "loss": 0.4542, "step": 1396 }, { "epoch": 0.03537628422491017, "grad_norm": 7.098147869110107, "learning_rate": 4.941123876039611e-05, "loss": 0.613, "step": 1398 }, { "epoch": 0.03542689407358672, "grad_norm": 6.280513286590576, "learning_rate": 4.941039526291817e-05, "loss": 0.8402, "step": 1400 }, { "epoch": 0.03542689407358672, "eval_cer": 0.19522168153555014, "eval_loss": 0.2994668483734131, "eval_runtime": 2672.9422, "eval_samples_per_second": 5.864, "eval_steps_per_second": 0.367, "step": 1400 }, { "epoch": 0.03547750392226327, "grad_norm": 20.038259506225586, "learning_rate": 4.940955176544022e-05, "loss": 0.4221, "step": 1402 }, { "epoch": 0.035528113770939826, "grad_norm": 4.178603649139404, "learning_rate": 4.940870826796228e-05, "loss": 0.5002, "step": 1404 }, { "epoch": 0.03557872361961638, "grad_norm": 4.7133469581604, "learning_rate": 4.9407864770484344e-05, "loss": 0.7616, "step": 1406 }, { "epoch": 0.03562933346829293, "grad_norm": 5.610054969787598, "learning_rate": 4.94070212730064e-05, "loss": 0.5404, "step": 1408 }, { "epoch": 0.035679943316969484, "grad_norm": 12.473624229431152, "learning_rate": 4.940617777552846e-05, "loss": 0.4731, "step": 1410 }, { "epoch": 0.03573055316564604, "grad_norm": 5.456740856170654, "learning_rate": 4.940533427805051e-05, "loss": 0.3353, "step": 1412 }, { "epoch": 0.03578116301432259, "grad_norm": 8.414679527282715, "learning_rate": 4.940449078057257e-05, "loss": 0.5136, "step": 1414 }, { "epoch": 0.03583177286299914, "grad_norm": 6.136534214019775, "learning_rate": 4.940364728309463e-05, "loss": 0.4605, "step": 1416 }, { "epoch": 0.035882382711675695, "grad_norm": 6.461178779602051, "learning_rate": 4.9402803785616684e-05, "loss": 0.395, "step": 1418 }, { "epoch": 0.03593299256035225, "grad_norm": 8.947699546813965, "learning_rate": 4.9401960288138745e-05, "loss": 0.8262, "step": 1420 }, { "epoch": 0.0359836024090288, "grad_norm": 7.892368793487549, "learning_rate": 4.94011167906608e-05, "loss": 0.529, "step": 1422 }, { "epoch": 0.036034212257705346, "grad_norm": 6.520657062530518, "learning_rate": 4.940027329318286e-05, "loss": 0.6126, "step": 1424 }, { "epoch": 0.0360848221063819, "grad_norm": 17.974655151367188, "learning_rate": 4.939942979570492e-05, "loss": 0.748, "step": 1426 }, { "epoch": 0.03613543195505845, "grad_norm": 3.9267520904541016, "learning_rate": 4.939858629822697e-05, "loss": 0.3001, "step": 1428 }, { "epoch": 0.036186041803735004, "grad_norm": 14.933813095092773, "learning_rate": 4.939774280074903e-05, "loss": 0.766, "step": 1430 }, { "epoch": 0.03623665165241156, "grad_norm": 9.554229736328125, "learning_rate": 4.9396899303271085e-05, "loss": 0.5409, "step": 1432 }, { "epoch": 0.03628726150108811, "grad_norm": 8.939488410949707, "learning_rate": 4.9396055805793146e-05, "loss": 0.4732, "step": 1434 }, { "epoch": 0.03633787134976466, "grad_norm": 3.313821315765381, "learning_rate": 4.9395212308315206e-05, "loss": 0.3792, "step": 1436 }, { "epoch": 0.036388481198441215, "grad_norm": 15.510714530944824, "learning_rate": 4.939436881083726e-05, "loss": 0.7043, "step": 1438 }, { "epoch": 0.03643909104711777, "grad_norm": 8.56005573272705, "learning_rate": 4.939352531335932e-05, "loss": 0.4035, "step": 1440 }, { "epoch": 0.03648970089579432, "grad_norm": 5.8778252601623535, "learning_rate": 4.939268181588137e-05, "loss": 0.3744, "step": 1442 }, { "epoch": 0.03654031074447087, "grad_norm": 6.815622806549072, "learning_rate": 4.939183831840343e-05, "loss": 0.5826, "step": 1444 }, { "epoch": 0.036590920593147426, "grad_norm": 7.3577680587768555, "learning_rate": 4.939099482092549e-05, "loss": 0.505, "step": 1446 }, { "epoch": 0.03664153044182398, "grad_norm": 8.071341514587402, "learning_rate": 4.939015132344755e-05, "loss": 0.7999, "step": 1448 }, { "epoch": 0.03669214029050053, "grad_norm": 10.3243989944458, "learning_rate": 4.938930782596961e-05, "loss": 0.6247, "step": 1450 }, { "epoch": 0.036742750139177084, "grad_norm": 4.544474124908447, "learning_rate": 4.938846432849166e-05, "loss": 0.5807, "step": 1452 }, { "epoch": 0.03679335998785364, "grad_norm": 9.284850120544434, "learning_rate": 4.938762083101372e-05, "loss": 0.4313, "step": 1454 }, { "epoch": 0.03684396983653019, "grad_norm": 7.1159749031066895, "learning_rate": 4.938677733353578e-05, "loss": 0.4875, "step": 1456 }, { "epoch": 0.03689457968520674, "grad_norm": 7.782135009765625, "learning_rate": 4.9385933836057834e-05, "loss": 0.8014, "step": 1458 }, { "epoch": 0.036945189533883295, "grad_norm": 9.787236213684082, "learning_rate": 4.9385090338579894e-05, "loss": 0.9414, "step": 1460 }, { "epoch": 0.03699579938255985, "grad_norm": 5.5786848068237305, "learning_rate": 4.938424684110195e-05, "loss": 0.6361, "step": 1462 }, { "epoch": 0.0370464092312364, "grad_norm": 4.597846508026123, "learning_rate": 4.938340334362401e-05, "loss": 0.2795, "step": 1464 }, { "epoch": 0.03709701907991295, "grad_norm": 5.9186553955078125, "learning_rate": 4.938255984614607e-05, "loss": 0.4051, "step": 1466 }, { "epoch": 0.037147628928589506, "grad_norm": 5.021423816680908, "learning_rate": 4.938171634866812e-05, "loss": 0.4298, "step": 1468 }, { "epoch": 0.03719823877726606, "grad_norm": 3.111738681793213, "learning_rate": 4.938087285119018e-05, "loss": 0.3101, "step": 1470 }, { "epoch": 0.03724884862594261, "grad_norm": 6.111041069030762, "learning_rate": 4.9380029353712235e-05, "loss": 0.6118, "step": 1472 }, { "epoch": 0.037299458474619164, "grad_norm": 30.16288185119629, "learning_rate": 4.9379185856234295e-05, "loss": 0.6206, "step": 1474 }, { "epoch": 0.03735006832329572, "grad_norm": 9.460359573364258, "learning_rate": 4.937834235875635e-05, "loss": 0.3322, "step": 1476 }, { "epoch": 0.03740067817197226, "grad_norm": 7.554260730743408, "learning_rate": 4.937749886127841e-05, "loss": 0.5655, "step": 1478 }, { "epoch": 0.037451288020648815, "grad_norm": 8.392956733703613, "learning_rate": 4.937665536380047e-05, "loss": 0.6669, "step": 1480 }, { "epoch": 0.03750189786932537, "grad_norm": 7.463123321533203, "learning_rate": 4.937581186632252e-05, "loss": 0.6082, "step": 1482 }, { "epoch": 0.03755250771800192, "grad_norm": 2.9462671279907227, "learning_rate": 4.937496836884458e-05, "loss": 0.7547, "step": 1484 }, { "epoch": 0.03760311756667847, "grad_norm": 9.454707145690918, "learning_rate": 4.9374124871366636e-05, "loss": 0.5665, "step": 1486 }, { "epoch": 0.037653727415355026, "grad_norm": 2.5308687686920166, "learning_rate": 4.9373281373888696e-05, "loss": 0.4683, "step": 1488 }, { "epoch": 0.03770433726403158, "grad_norm": 3.8949882984161377, "learning_rate": 4.937243787641076e-05, "loss": 0.3981, "step": 1490 }, { "epoch": 0.03775494711270813, "grad_norm": 3.255460739135742, "learning_rate": 4.937159437893281e-05, "loss": 0.3254, "step": 1492 }, { "epoch": 0.037805556961384684, "grad_norm": 10.862367630004883, "learning_rate": 4.937075088145487e-05, "loss": 0.8337, "step": 1494 }, { "epoch": 0.03785616681006124, "grad_norm": 12.926139831542969, "learning_rate": 4.9369907383976924e-05, "loss": 0.8218, "step": 1496 }, { "epoch": 0.03790677665873779, "grad_norm": 5.084683418273926, "learning_rate": 4.9369063886498984e-05, "loss": 0.4073, "step": 1498 }, { "epoch": 0.03795738650741434, "grad_norm": 5.375326633453369, "learning_rate": 4.9368220389021044e-05, "loss": 0.4468, "step": 1500 }, { "epoch": 0.038007996356090895, "grad_norm": 2.891730785369873, "learning_rate": 4.93673768915431e-05, "loss": 0.4952, "step": 1502 }, { "epoch": 0.03805860620476745, "grad_norm": 6.7384161949157715, "learning_rate": 4.936653339406516e-05, "loss": 0.6817, "step": 1504 }, { "epoch": 0.038109216053444, "grad_norm": 4.95111608505249, "learning_rate": 4.936568989658721e-05, "loss": 0.7322, "step": 1506 }, { "epoch": 0.03815982590212055, "grad_norm": 4.689101219177246, "learning_rate": 4.936484639910927e-05, "loss": 0.6041, "step": 1508 }, { "epoch": 0.038210435750797106, "grad_norm": 6.9700798988342285, "learning_rate": 4.936400290163133e-05, "loss": 0.4521, "step": 1510 }, { "epoch": 0.03826104559947366, "grad_norm": 5.602224826812744, "learning_rate": 4.9363159404153385e-05, "loss": 0.6519, "step": 1512 }, { "epoch": 0.03831165544815021, "grad_norm": 5.881453037261963, "learning_rate": 4.9362315906675445e-05, "loss": 0.6069, "step": 1514 }, { "epoch": 0.038362265296826764, "grad_norm": 5.395936012268066, "learning_rate": 4.93614724091975e-05, "loss": 0.3001, "step": 1516 }, { "epoch": 0.03841287514550332, "grad_norm": 5.4586663246154785, "learning_rate": 4.936062891171956e-05, "loss": 0.6286, "step": 1518 }, { "epoch": 0.03846348499417987, "grad_norm": 5.338792324066162, "learning_rate": 4.935978541424162e-05, "loss": 0.4526, "step": 1520 }, { "epoch": 0.03851409484285642, "grad_norm": 2.985135555267334, "learning_rate": 4.935894191676367e-05, "loss": 0.2444, "step": 1522 }, { "epoch": 0.038564704691532975, "grad_norm": 7.952456474304199, "learning_rate": 4.935809841928573e-05, "loss": 0.5033, "step": 1524 }, { "epoch": 0.03861531454020953, "grad_norm": 7.913976192474365, "learning_rate": 4.9357254921807786e-05, "loss": 0.5823, "step": 1526 }, { "epoch": 0.03866592438888608, "grad_norm": 7.781038284301758, "learning_rate": 4.9356411424329846e-05, "loss": 0.6915, "step": 1528 }, { "epoch": 0.03871653423756263, "grad_norm": 5.074882984161377, "learning_rate": 4.9355567926851906e-05, "loss": 0.8664, "step": 1530 }, { "epoch": 0.03876714408623918, "grad_norm": 3.08144474029541, "learning_rate": 4.935472442937396e-05, "loss": 0.3723, "step": 1532 }, { "epoch": 0.03881775393491573, "grad_norm": 6.625777244567871, "learning_rate": 4.935388093189602e-05, "loss": 0.5908, "step": 1534 }, { "epoch": 0.038868363783592284, "grad_norm": 8.451767921447754, "learning_rate": 4.9353037434418073e-05, "loss": 0.8605, "step": 1536 }, { "epoch": 0.03891897363226884, "grad_norm": 5.070049285888672, "learning_rate": 4.9352193936940134e-05, "loss": 0.3042, "step": 1538 }, { "epoch": 0.03896958348094539, "grad_norm": 7.37528133392334, "learning_rate": 4.9351350439462194e-05, "loss": 0.7376, "step": 1540 }, { "epoch": 0.03902019332962194, "grad_norm": 3.3460285663604736, "learning_rate": 4.935050694198425e-05, "loss": 0.6662, "step": 1542 }, { "epoch": 0.039070803178298495, "grad_norm": 3.999098062515259, "learning_rate": 4.934966344450631e-05, "loss": 0.3612, "step": 1544 }, { "epoch": 0.03912141302697505, "grad_norm": 6.4491868019104, "learning_rate": 4.934881994702836e-05, "loss": 0.7708, "step": 1546 }, { "epoch": 0.0391720228756516, "grad_norm": 4.037872314453125, "learning_rate": 4.934797644955042e-05, "loss": 0.4529, "step": 1548 }, { "epoch": 0.03922263272432815, "grad_norm": 7.757660865783691, "learning_rate": 4.9347132952072475e-05, "loss": 0.6572, "step": 1550 }, { "epoch": 0.039273242573004706, "grad_norm": 10.399569511413574, "learning_rate": 4.9346289454594535e-05, "loss": 0.971, "step": 1552 }, { "epoch": 0.03932385242168126, "grad_norm": 4.098683834075928, "learning_rate": 4.9345445957116595e-05, "loss": 0.5687, "step": 1554 }, { "epoch": 0.03937446227035781, "grad_norm": 4.029207706451416, "learning_rate": 4.934460245963865e-05, "loss": 0.4164, "step": 1556 }, { "epoch": 0.039425072119034364, "grad_norm": 9.3286771774292, "learning_rate": 4.934375896216071e-05, "loss": 0.8485, "step": 1558 }, { "epoch": 0.03947568196771092, "grad_norm": 8.822758674621582, "learning_rate": 4.934291546468276e-05, "loss": 0.5747, "step": 1560 }, { "epoch": 0.03952629181638747, "grad_norm": 6.740304946899414, "learning_rate": 4.934207196720482e-05, "loss": 0.5014, "step": 1562 }, { "epoch": 0.03957690166506402, "grad_norm": 4.9694929122924805, "learning_rate": 4.934122846972688e-05, "loss": 0.5403, "step": 1564 }, { "epoch": 0.039627511513740575, "grad_norm": 7.032242298126221, "learning_rate": 4.9340384972248936e-05, "loss": 0.3896, "step": 1566 }, { "epoch": 0.03967812136241713, "grad_norm": 3.1985256671905518, "learning_rate": 4.9339541474770996e-05, "loss": 0.4138, "step": 1568 }, { "epoch": 0.03972873121109368, "grad_norm": 4.911308288574219, "learning_rate": 4.933869797729305e-05, "loss": 0.6613, "step": 1570 }, { "epoch": 0.03977934105977023, "grad_norm": 10.836835861206055, "learning_rate": 4.933785447981511e-05, "loss": 0.368, "step": 1572 }, { "epoch": 0.039829950908446786, "grad_norm": 7.5843377113342285, "learning_rate": 4.933701098233717e-05, "loss": 0.8429, "step": 1574 }, { "epoch": 0.03988056075712334, "grad_norm": 4.915452003479004, "learning_rate": 4.933616748485922e-05, "loss": 0.4179, "step": 1576 }, { "epoch": 0.03993117060579989, "grad_norm": 4.324033260345459, "learning_rate": 4.9335323987381283e-05, "loss": 0.5588, "step": 1578 }, { "epoch": 0.039981780454476444, "grad_norm": 5.227353096008301, "learning_rate": 4.933448048990334e-05, "loss": 0.4188, "step": 1580 }, { "epoch": 0.040032390303153, "grad_norm": 6.144440650939941, "learning_rate": 4.93336369924254e-05, "loss": 0.7091, "step": 1582 }, { "epoch": 0.04008300015182954, "grad_norm": 9.614253044128418, "learning_rate": 4.933279349494746e-05, "loss": 0.5816, "step": 1584 }, { "epoch": 0.040133610000506095, "grad_norm": 5.539832592010498, "learning_rate": 4.933194999746951e-05, "loss": 0.2897, "step": 1586 }, { "epoch": 0.04018421984918265, "grad_norm": 4.285924911499023, "learning_rate": 4.933110649999157e-05, "loss": 0.4647, "step": 1588 }, { "epoch": 0.0402348296978592, "grad_norm": 5.418667316436768, "learning_rate": 4.9330263002513624e-05, "loss": 0.377, "step": 1590 }, { "epoch": 0.04028543954653575, "grad_norm": 7.248344421386719, "learning_rate": 4.9329419505035684e-05, "loss": 0.5599, "step": 1592 }, { "epoch": 0.040336049395212306, "grad_norm": 7.9554643630981445, "learning_rate": 4.9328576007557745e-05, "loss": 0.8153, "step": 1594 }, { "epoch": 0.04038665924388886, "grad_norm": 8.028340339660645, "learning_rate": 4.93277325100798e-05, "loss": 0.5364, "step": 1596 }, { "epoch": 0.04043726909256541, "grad_norm": 7.2708964347839355, "learning_rate": 4.932688901260186e-05, "loss": 0.8303, "step": 1598 }, { "epoch": 0.040487878941241964, "grad_norm": 5.855560779571533, "learning_rate": 4.932604551512391e-05, "loss": 0.4462, "step": 1600 }, { "epoch": 0.040487878941241964, "eval_cer": 0.18224000249547695, "eval_loss": 0.28801780939102173, "eval_runtime": 2605.2542, "eval_samples_per_second": 6.016, "eval_steps_per_second": 0.376, "step": 1600 }, { "epoch": 0.04053848878991852, "grad_norm": 10.296344757080078, "learning_rate": 4.932520201764597e-05, "loss": 0.5008, "step": 1602 }, { "epoch": 0.04058909863859507, "grad_norm": 8.924449920654297, "learning_rate": 4.932435852016803e-05, "loss": 0.3729, "step": 1604 }, { "epoch": 0.04063970848727162, "grad_norm": 5.110000133514404, "learning_rate": 4.9323515022690085e-05, "loss": 0.5821, "step": 1606 }, { "epoch": 0.040690318335948175, "grad_norm": 5.40463399887085, "learning_rate": 4.9322671525212146e-05, "loss": 0.6681, "step": 1608 }, { "epoch": 0.04074092818462473, "grad_norm": 10.994048118591309, "learning_rate": 4.93218280277342e-05, "loss": 0.453, "step": 1610 }, { "epoch": 0.04079153803330128, "grad_norm": 6.396229267120361, "learning_rate": 4.932098453025626e-05, "loss": 0.4864, "step": 1612 }, { "epoch": 0.04084214788197783, "grad_norm": 6.884303569793701, "learning_rate": 4.932014103277832e-05, "loss": 0.9021, "step": 1614 }, { "epoch": 0.040892757730654386, "grad_norm": 15.925917625427246, "learning_rate": 4.931929753530037e-05, "loss": 0.6345, "step": 1616 }, { "epoch": 0.04094336757933094, "grad_norm": 3.3225505352020264, "learning_rate": 4.931845403782243e-05, "loss": 0.5429, "step": 1618 }, { "epoch": 0.04099397742800749, "grad_norm": 5.548943519592285, "learning_rate": 4.9317610540344487e-05, "loss": 0.5081, "step": 1620 }, { "epoch": 0.041044587276684044, "grad_norm": 8.253005027770996, "learning_rate": 4.931676704286655e-05, "loss": 0.4432, "step": 1622 }, { "epoch": 0.0410951971253606, "grad_norm": 10.242237091064453, "learning_rate": 4.931592354538861e-05, "loss": 0.6442, "step": 1624 }, { "epoch": 0.04114580697403715, "grad_norm": 4.957180976867676, "learning_rate": 4.931508004791066e-05, "loss": 0.4146, "step": 1626 }, { "epoch": 0.0411964168227137, "grad_norm": 7.067080497741699, "learning_rate": 4.931423655043272e-05, "loss": 0.424, "step": 1628 }, { "epoch": 0.041247026671390255, "grad_norm": 5.9971842765808105, "learning_rate": 4.9313393052954774e-05, "loss": 0.4941, "step": 1630 }, { "epoch": 0.04129763652006681, "grad_norm": 5.412482261657715, "learning_rate": 4.9312549555476834e-05, "loss": 0.4065, "step": 1632 }, { "epoch": 0.04134824636874336, "grad_norm": 13.345954895019531, "learning_rate": 4.931170605799889e-05, "loss": 0.8215, "step": 1634 }, { "epoch": 0.04139885621741991, "grad_norm": 9.993549346923828, "learning_rate": 4.931086256052095e-05, "loss": 0.8707, "step": 1636 }, { "epoch": 0.04144946606609646, "grad_norm": 27.219308853149414, "learning_rate": 4.931001906304301e-05, "loss": 0.5455, "step": 1638 }, { "epoch": 0.04150007591477301, "grad_norm": 5.13199520111084, "learning_rate": 4.930917556556506e-05, "loss": 0.5479, "step": 1640 }, { "epoch": 0.041550685763449564, "grad_norm": 4.877132892608643, "learning_rate": 4.930833206808712e-05, "loss": 0.5966, "step": 1642 }, { "epoch": 0.04160129561212612, "grad_norm": 3.1169495582580566, "learning_rate": 4.9307488570609175e-05, "loss": 0.5371, "step": 1644 }, { "epoch": 0.04165190546080267, "grad_norm": 4.428645133972168, "learning_rate": 4.9306645073131235e-05, "loss": 0.3974, "step": 1646 }, { "epoch": 0.04170251530947922, "grad_norm": 12.36322021484375, "learning_rate": 4.9305801575653295e-05, "loss": 0.6203, "step": 1648 }, { "epoch": 0.041753125158155775, "grad_norm": 4.944615840911865, "learning_rate": 4.930495807817535e-05, "loss": 0.6994, "step": 1650 }, { "epoch": 0.04180373500683233, "grad_norm": 3.7057061195373535, "learning_rate": 4.930411458069741e-05, "loss": 0.2836, "step": 1652 }, { "epoch": 0.04185434485550888, "grad_norm": 5.559707164764404, "learning_rate": 4.930327108321946e-05, "loss": 0.7595, "step": 1654 }, { "epoch": 0.04190495470418543, "grad_norm": 6.805633544921875, "learning_rate": 4.930242758574152e-05, "loss": 0.497, "step": 1656 }, { "epoch": 0.041955564552861986, "grad_norm": 6.29421329498291, "learning_rate": 4.930158408826358e-05, "loss": 0.5615, "step": 1658 }, { "epoch": 0.04200617440153854, "grad_norm": 3.3329832553863525, "learning_rate": 4.9300740590785636e-05, "loss": 0.2914, "step": 1660 }, { "epoch": 0.04205678425021509, "grad_norm": 5.0592827796936035, "learning_rate": 4.9299897093307696e-05, "loss": 0.3799, "step": 1662 }, { "epoch": 0.042107394098891644, "grad_norm": 4.176607131958008, "learning_rate": 4.929905359582975e-05, "loss": 0.3987, "step": 1664 }, { "epoch": 0.0421580039475682, "grad_norm": 6.9350175857543945, "learning_rate": 4.929821009835181e-05, "loss": 0.337, "step": 1666 }, { "epoch": 0.04220861379624475, "grad_norm": 12.642129898071289, "learning_rate": 4.929736660087387e-05, "loss": 0.5035, "step": 1668 }, { "epoch": 0.0422592236449213, "grad_norm": 3.18143892288208, "learning_rate": 4.9296523103395924e-05, "loss": 0.2602, "step": 1670 }, { "epoch": 0.042309833493597855, "grad_norm": 7.448511600494385, "learning_rate": 4.9295679605917984e-05, "loss": 0.6351, "step": 1672 }, { "epoch": 0.04236044334227441, "grad_norm": 9.632100105285645, "learning_rate": 4.929483610844004e-05, "loss": 0.6639, "step": 1674 }, { "epoch": 0.04241105319095096, "grad_norm": 4.993129730224609, "learning_rate": 4.92939926109621e-05, "loss": 0.5902, "step": 1676 }, { "epoch": 0.04246166303962751, "grad_norm": 8.809192657470703, "learning_rate": 4.929314911348416e-05, "loss": 0.7136, "step": 1678 }, { "epoch": 0.042512272888304066, "grad_norm": 4.831273078918457, "learning_rate": 4.929230561600621e-05, "loss": 0.702, "step": 1680 }, { "epoch": 0.04256288273698062, "grad_norm": 3.406562089920044, "learning_rate": 4.929146211852827e-05, "loss": 0.7114, "step": 1682 }, { "epoch": 0.04261349258565717, "grad_norm": 4.17376184463501, "learning_rate": 4.9290618621050325e-05, "loss": 0.5236, "step": 1684 }, { "epoch": 0.042664102434333724, "grad_norm": 3.0841195583343506, "learning_rate": 4.9289775123572385e-05, "loss": 0.4595, "step": 1686 }, { "epoch": 0.04271471228301028, "grad_norm": 6.880139350891113, "learning_rate": 4.9288931626094445e-05, "loss": 0.6339, "step": 1688 }, { "epoch": 0.04276532213168683, "grad_norm": 3.7621781826019287, "learning_rate": 4.92880881286165e-05, "loss": 0.4469, "step": 1690 }, { "epoch": 0.042815931980363375, "grad_norm": 9.83583927154541, "learning_rate": 4.928724463113856e-05, "loss": 0.7447, "step": 1692 }, { "epoch": 0.04286654182903993, "grad_norm": 3.6700632572174072, "learning_rate": 4.928640113366061e-05, "loss": 0.3185, "step": 1694 }, { "epoch": 0.04291715167771648, "grad_norm": 5.93882417678833, "learning_rate": 4.928555763618267e-05, "loss": 0.5691, "step": 1696 }, { "epoch": 0.042967761526393033, "grad_norm": 2.9145689010620117, "learning_rate": 4.928471413870473e-05, "loss": 0.3396, "step": 1698 }, { "epoch": 0.043018371375069586, "grad_norm": 4.243858814239502, "learning_rate": 4.9283870641226786e-05, "loss": 0.3018, "step": 1700 }, { "epoch": 0.04306898122374614, "grad_norm": 4.957646369934082, "learning_rate": 4.9283027143748846e-05, "loss": 0.4555, "step": 1702 }, { "epoch": 0.04311959107242269, "grad_norm": 5.5672831535339355, "learning_rate": 4.92821836462709e-05, "loss": 0.4606, "step": 1704 }, { "epoch": 0.043170200921099244, "grad_norm": 4.047713756561279, "learning_rate": 4.928134014879296e-05, "loss": 0.3793, "step": 1706 }, { "epoch": 0.0432208107697758, "grad_norm": 7.607228755950928, "learning_rate": 4.928049665131502e-05, "loss": 0.6637, "step": 1708 }, { "epoch": 0.04327142061845235, "grad_norm": 11.504854202270508, "learning_rate": 4.9279653153837073e-05, "loss": 0.6826, "step": 1710 }, { "epoch": 0.0433220304671289, "grad_norm": 5.851675510406494, "learning_rate": 4.9278809656359134e-05, "loss": 0.526, "step": 1712 }, { "epoch": 0.043372640315805455, "grad_norm": 14.779943466186523, "learning_rate": 4.927796615888119e-05, "loss": 0.6321, "step": 1714 }, { "epoch": 0.04342325016448201, "grad_norm": 9.94005012512207, "learning_rate": 4.927712266140325e-05, "loss": 0.7401, "step": 1716 }, { "epoch": 0.04347386001315856, "grad_norm": 4.635720729827881, "learning_rate": 4.92762791639253e-05, "loss": 0.5283, "step": 1718 }, { "epoch": 0.04352446986183511, "grad_norm": 3.568844795227051, "learning_rate": 4.927543566644736e-05, "loss": 0.4053, "step": 1720 }, { "epoch": 0.043575079710511666, "grad_norm": 6.748814582824707, "learning_rate": 4.927459216896942e-05, "loss": 0.5624, "step": 1722 }, { "epoch": 0.04362568955918822, "grad_norm": 7.1615424156188965, "learning_rate": 4.9273748671491475e-05, "loss": 0.8786, "step": 1724 }, { "epoch": 0.04367629940786477, "grad_norm": 5.897001266479492, "learning_rate": 4.9272905174013535e-05, "loss": 0.7447, "step": 1726 }, { "epoch": 0.043726909256541324, "grad_norm": 6.555644512176514, "learning_rate": 4.927206167653559e-05, "loss": 0.4951, "step": 1728 }, { "epoch": 0.04377751910521788, "grad_norm": 8.615190505981445, "learning_rate": 4.927121817905765e-05, "loss": 0.6687, "step": 1730 }, { "epoch": 0.04382812895389443, "grad_norm": 6.572555065155029, "learning_rate": 4.927037468157971e-05, "loss": 0.4269, "step": 1732 }, { "epoch": 0.04387873880257098, "grad_norm": 6.60528564453125, "learning_rate": 4.926953118410176e-05, "loss": 0.4845, "step": 1734 }, { "epoch": 0.043929348651247535, "grad_norm": 5.96574592590332, "learning_rate": 4.926868768662382e-05, "loss": 0.6266, "step": 1736 }, { "epoch": 0.04397995849992409, "grad_norm": 2.911107063293457, "learning_rate": 4.9267844189145876e-05, "loss": 0.312, "step": 1738 }, { "epoch": 0.04403056834860064, "grad_norm": 6.024972438812256, "learning_rate": 4.9267000691667936e-05, "loss": 0.5205, "step": 1740 }, { "epoch": 0.04408117819727719, "grad_norm": 12.015813827514648, "learning_rate": 4.9266157194189996e-05, "loss": 0.621, "step": 1742 }, { "epoch": 0.044131788045953746, "grad_norm": 17.426437377929688, "learning_rate": 4.926531369671205e-05, "loss": 0.7822, "step": 1744 }, { "epoch": 0.04418239789463029, "grad_norm": 8.361766815185547, "learning_rate": 4.926447019923411e-05, "loss": 0.8237, "step": 1746 }, { "epoch": 0.044233007743306844, "grad_norm": 6.699423789978027, "learning_rate": 4.926362670175616e-05, "loss": 0.5673, "step": 1748 }, { "epoch": 0.0442836175919834, "grad_norm": 3.933328151702881, "learning_rate": 4.926278320427822e-05, "loss": 0.6946, "step": 1750 }, { "epoch": 0.04433422744065995, "grad_norm": 5.092313766479492, "learning_rate": 4.9261939706800283e-05, "loss": 0.6024, "step": 1752 }, { "epoch": 0.0443848372893365, "grad_norm": 16.1398868560791, "learning_rate": 4.926109620932234e-05, "loss": 0.7255, "step": 1754 }, { "epoch": 0.044435447138013055, "grad_norm": 4.958022117614746, "learning_rate": 4.92602527118444e-05, "loss": 0.4381, "step": 1756 }, { "epoch": 0.04448605698668961, "grad_norm": 6.70586633682251, "learning_rate": 4.925940921436645e-05, "loss": 0.5228, "step": 1758 }, { "epoch": 0.04453666683536616, "grad_norm": 3.8433001041412354, "learning_rate": 4.925856571688851e-05, "loss": 0.3635, "step": 1760 }, { "epoch": 0.04458727668404271, "grad_norm": 3.4007515907287598, "learning_rate": 4.925772221941057e-05, "loss": 0.2397, "step": 1762 }, { "epoch": 0.044637886532719266, "grad_norm": 12.362958908081055, "learning_rate": 4.9256878721932624e-05, "loss": 0.5092, "step": 1764 }, { "epoch": 0.04468849638139582, "grad_norm": 3.942413568496704, "learning_rate": 4.9256035224454684e-05, "loss": 0.4432, "step": 1766 }, { "epoch": 0.04473910623007237, "grad_norm": 5.850658893585205, "learning_rate": 4.925519172697674e-05, "loss": 0.8487, "step": 1768 }, { "epoch": 0.044789716078748924, "grad_norm": 3.8962783813476562, "learning_rate": 4.92543482294988e-05, "loss": 0.6375, "step": 1770 }, { "epoch": 0.04484032592742548, "grad_norm": 10.159829139709473, "learning_rate": 4.925350473202086e-05, "loss": 0.5135, "step": 1772 }, { "epoch": 0.04489093577610203, "grad_norm": 2.924187183380127, "learning_rate": 4.925266123454291e-05, "loss": 0.2508, "step": 1774 }, { "epoch": 0.04494154562477858, "grad_norm": 6.0558342933654785, "learning_rate": 4.925181773706497e-05, "loss": 0.3182, "step": 1776 }, { "epoch": 0.044992155473455135, "grad_norm": 5.556451320648193, "learning_rate": 4.9250974239587025e-05, "loss": 0.492, "step": 1778 }, { "epoch": 0.04504276532213169, "grad_norm": 4.185290813446045, "learning_rate": 4.9250130742109086e-05, "loss": 0.4214, "step": 1780 }, { "epoch": 0.04509337517080824, "grad_norm": 5.0443196296691895, "learning_rate": 4.9249287244631146e-05, "loss": 0.6803, "step": 1782 }, { "epoch": 0.04514398501948479, "grad_norm": 10.363567352294922, "learning_rate": 4.92484437471532e-05, "loss": 0.5701, "step": 1784 }, { "epoch": 0.045194594868161346, "grad_norm": 6.5281243324279785, "learning_rate": 4.924760024967526e-05, "loss": 0.417, "step": 1786 }, { "epoch": 0.0452452047168379, "grad_norm": 9.699246406555176, "learning_rate": 4.924675675219731e-05, "loss": 0.7565, "step": 1788 }, { "epoch": 0.04529581456551445, "grad_norm": 13.578716278076172, "learning_rate": 4.924591325471937e-05, "loss": 0.5528, "step": 1790 }, { "epoch": 0.045346424414191004, "grad_norm": 6.151464462280273, "learning_rate": 4.924506975724143e-05, "loss": 0.655, "step": 1792 }, { "epoch": 0.04539703426286756, "grad_norm": 11.576047897338867, "learning_rate": 4.9244226259763487e-05, "loss": 0.8649, "step": 1794 }, { "epoch": 0.04544764411154411, "grad_norm": 6.639153003692627, "learning_rate": 4.924338276228555e-05, "loss": 0.553, "step": 1796 }, { "epoch": 0.04549825396022066, "grad_norm": 4.5320563316345215, "learning_rate": 4.92425392648076e-05, "loss": 0.4832, "step": 1798 }, { "epoch": 0.04554886380889721, "grad_norm": 3.708400249481201, "learning_rate": 4.924169576732966e-05, "loss": 0.5646, "step": 1800 }, { "epoch": 0.04554886380889721, "eval_cer": 0.19456141992638343, "eval_loss": 0.29710084199905396, "eval_runtime": 2630.6631, "eval_samples_per_second": 5.958, "eval_steps_per_second": 0.373, "step": 1800 }, { "epoch": 0.04559947365757376, "grad_norm": 4.805381774902344, "learning_rate": 4.9240852269851714e-05, "loss": 0.5711, "step": 1802 }, { "epoch": 0.045650083506250314, "grad_norm": 3.130221366882324, "learning_rate": 4.9240008772373774e-05, "loss": 0.4231, "step": 1804 }, { "epoch": 0.045700693354926866, "grad_norm": 4.455588340759277, "learning_rate": 4.9239165274895834e-05, "loss": 0.3281, "step": 1806 }, { "epoch": 0.04575130320360342, "grad_norm": 2.951422929763794, "learning_rate": 4.923832177741789e-05, "loss": 0.4477, "step": 1808 }, { "epoch": 0.04580191305227997, "grad_norm": 5.397090435028076, "learning_rate": 4.923747827993995e-05, "loss": 0.4032, "step": 1810 }, { "epoch": 0.045852522900956524, "grad_norm": 5.767248630523682, "learning_rate": 4.9236634782462e-05, "loss": 0.6842, "step": 1812 }, { "epoch": 0.04590313274963308, "grad_norm": 3.4613723754882812, "learning_rate": 4.923579128498406e-05, "loss": 0.2943, "step": 1814 }, { "epoch": 0.04595374259830963, "grad_norm": 4.715557098388672, "learning_rate": 4.923494778750612e-05, "loss": 0.3851, "step": 1816 }, { "epoch": 0.04600435244698618, "grad_norm": 4.709720611572266, "learning_rate": 4.9234104290028175e-05, "loss": 0.2846, "step": 1818 }, { "epoch": 0.046054962295662735, "grad_norm": 2.6926629543304443, "learning_rate": 4.9233260792550235e-05, "loss": 0.474, "step": 1820 }, { "epoch": 0.04610557214433929, "grad_norm": 7.4501752853393555, "learning_rate": 4.923241729507229e-05, "loss": 0.6805, "step": 1822 }, { "epoch": 0.04615618199301584, "grad_norm": 4.941989421844482, "learning_rate": 4.923157379759435e-05, "loss": 0.4549, "step": 1824 }, { "epoch": 0.04620679184169239, "grad_norm": 7.671848297119141, "learning_rate": 4.923073030011641e-05, "loss": 0.4181, "step": 1826 }, { "epoch": 0.046257401690368946, "grad_norm": 1.92340886592865, "learning_rate": 4.922988680263846e-05, "loss": 0.2794, "step": 1828 }, { "epoch": 0.0463080115390455, "grad_norm": 4.5162200927734375, "learning_rate": 4.922904330516052e-05, "loss": 0.4803, "step": 1830 }, { "epoch": 0.04635862138772205, "grad_norm": 6.522688388824463, "learning_rate": 4.9228199807682576e-05, "loss": 0.5851, "step": 1832 }, { "epoch": 0.046409231236398604, "grad_norm": 3.49273419380188, "learning_rate": 4.9227356310204636e-05, "loss": 0.5751, "step": 1834 }, { "epoch": 0.04645984108507516, "grad_norm": 14.866829872131348, "learning_rate": 4.9226512812726697e-05, "loss": 0.5211, "step": 1836 }, { "epoch": 0.04651045093375171, "grad_norm": 6.658441066741943, "learning_rate": 4.922566931524875e-05, "loss": 0.5325, "step": 1838 }, { "epoch": 0.04656106078242826, "grad_norm": 3.8592264652252197, "learning_rate": 4.922482581777081e-05, "loss": 0.5917, "step": 1840 }, { "epoch": 0.046611670631104815, "grad_norm": 10.734745025634766, "learning_rate": 4.9223982320292864e-05, "loss": 0.3665, "step": 1842 }, { "epoch": 0.04666228047978137, "grad_norm": 7.1689534187316895, "learning_rate": 4.9223138822814924e-05, "loss": 0.386, "step": 1844 }, { "epoch": 0.04671289032845792, "grad_norm": 4.963681697845459, "learning_rate": 4.9222295325336984e-05, "loss": 0.4437, "step": 1846 }, { "epoch": 0.04676350017713447, "grad_norm": 2.5440425872802734, "learning_rate": 4.922145182785904e-05, "loss": 0.3268, "step": 1848 }, { "epoch": 0.046814110025811026, "grad_norm": 14.03139877319336, "learning_rate": 4.92206083303811e-05, "loss": 0.7161, "step": 1850 }, { "epoch": 0.04686471987448758, "grad_norm": 7.766716480255127, "learning_rate": 4.921976483290315e-05, "loss": 0.4935, "step": 1852 }, { "epoch": 0.046915329723164124, "grad_norm": 13.12921142578125, "learning_rate": 4.921892133542521e-05, "loss": 0.4105, "step": 1854 }, { "epoch": 0.04696593957184068, "grad_norm": 3.441727876663208, "learning_rate": 4.921807783794727e-05, "loss": 0.8493, "step": 1856 }, { "epoch": 0.04701654942051723, "grad_norm": 6.923675060272217, "learning_rate": 4.9217234340469325e-05, "loss": 0.819, "step": 1858 }, { "epoch": 0.04706715926919378, "grad_norm": 8.306994438171387, "learning_rate": 4.9216390842991385e-05, "loss": 0.6548, "step": 1860 }, { "epoch": 0.047117769117870335, "grad_norm": 3.314638614654541, "learning_rate": 4.921554734551344e-05, "loss": 0.6389, "step": 1862 }, { "epoch": 0.04716837896654689, "grad_norm": 8.720133781433105, "learning_rate": 4.92147038480355e-05, "loss": 0.5129, "step": 1864 }, { "epoch": 0.04721898881522344, "grad_norm": 15.914939880371094, "learning_rate": 4.921386035055756e-05, "loss": 0.5298, "step": 1866 }, { "epoch": 0.047269598663899993, "grad_norm": 4.567658424377441, "learning_rate": 4.921301685307961e-05, "loss": 0.426, "step": 1868 }, { "epoch": 0.047320208512576546, "grad_norm": 2.2042078971862793, "learning_rate": 4.921217335560167e-05, "loss": 0.2168, "step": 1870 }, { "epoch": 0.0473708183612531, "grad_norm": 2.475092649459839, "learning_rate": 4.9211329858123726e-05, "loss": 0.5226, "step": 1872 }, { "epoch": 0.04742142820992965, "grad_norm": 3.972266435623169, "learning_rate": 4.9210486360645786e-05, "loss": 0.5745, "step": 1874 }, { "epoch": 0.047472038058606204, "grad_norm": 4.532939910888672, "learning_rate": 4.9209642863167846e-05, "loss": 0.6833, "step": 1876 }, { "epoch": 0.04752264790728276, "grad_norm": 7.014007568359375, "learning_rate": 4.92087993656899e-05, "loss": 0.5458, "step": 1878 }, { "epoch": 0.04757325775595931, "grad_norm": 8.639058113098145, "learning_rate": 4.920795586821196e-05, "loss": 0.698, "step": 1880 }, { "epoch": 0.04762386760463586, "grad_norm": 4.487074375152588, "learning_rate": 4.920711237073401e-05, "loss": 0.4481, "step": 1882 }, { "epoch": 0.047674477453312415, "grad_norm": 7.813543319702148, "learning_rate": 4.9206268873256074e-05, "loss": 0.5677, "step": 1884 }, { "epoch": 0.04772508730198897, "grad_norm": 14.59656047821045, "learning_rate": 4.920542537577813e-05, "loss": 0.5587, "step": 1886 }, { "epoch": 0.04777569715066552, "grad_norm": 7.641439437866211, "learning_rate": 4.920458187830019e-05, "loss": 0.9166, "step": 1888 }, { "epoch": 0.04782630699934207, "grad_norm": 11.967202186584473, "learning_rate": 4.920373838082225e-05, "loss": 0.6675, "step": 1890 }, { "epoch": 0.047876916848018626, "grad_norm": 3.4140710830688477, "learning_rate": 4.92028948833443e-05, "loss": 0.5182, "step": 1892 }, { "epoch": 0.04792752669669518, "grad_norm": 11.876256942749023, "learning_rate": 4.920205138586636e-05, "loss": 0.5383, "step": 1894 }, { "epoch": 0.04797813654537173, "grad_norm": 3.3766562938690186, "learning_rate": 4.9201207888388414e-05, "loss": 0.4677, "step": 1896 }, { "epoch": 0.048028746394048284, "grad_norm": 4.277167320251465, "learning_rate": 4.9200364390910475e-05, "loss": 0.4177, "step": 1898 }, { "epoch": 0.04807935624272484, "grad_norm": 5.8467936515808105, "learning_rate": 4.9199520893432535e-05, "loss": 0.5603, "step": 1900 }, { "epoch": 0.04812996609140139, "grad_norm": 5.7047038078308105, "learning_rate": 4.919867739595459e-05, "loss": 0.416, "step": 1902 }, { "epoch": 0.04818057594007794, "grad_norm": 8.13704776763916, "learning_rate": 4.919783389847665e-05, "loss": 0.2622, "step": 1904 }, { "epoch": 0.048231185788754495, "grad_norm": 7.6252923011779785, "learning_rate": 4.91969904009987e-05, "loss": 0.9887, "step": 1906 }, { "epoch": 0.04828179563743104, "grad_norm": 6.5191497802734375, "learning_rate": 4.919614690352076e-05, "loss": 0.6701, "step": 1908 }, { "epoch": 0.048332405486107594, "grad_norm": 98.16575622558594, "learning_rate": 4.919530340604282e-05, "loss": 1.0357, "step": 1910 }, { "epoch": 0.048383015334784146, "grad_norm": 3.221017837524414, "learning_rate": 4.9194459908564876e-05, "loss": 0.396, "step": 1912 }, { "epoch": 0.0484336251834607, "grad_norm": 4.679788112640381, "learning_rate": 4.9193616411086936e-05, "loss": 0.3766, "step": 1914 }, { "epoch": 0.04848423503213725, "grad_norm": 7.324192047119141, "learning_rate": 4.919277291360899e-05, "loss": 0.6765, "step": 1916 }, { "epoch": 0.048534844880813804, "grad_norm": 5.692389488220215, "learning_rate": 4.919192941613105e-05, "loss": 0.7648, "step": 1918 }, { "epoch": 0.04858545472949036, "grad_norm": 8.046274185180664, "learning_rate": 4.919108591865311e-05, "loss": 0.717, "step": 1920 }, { "epoch": 0.04863606457816691, "grad_norm": 8.431305885314941, "learning_rate": 4.919024242117516e-05, "loss": 0.5176, "step": 1922 }, { "epoch": 0.04868667442684346, "grad_norm": 4.852190017700195, "learning_rate": 4.918939892369722e-05, "loss": 0.4792, "step": 1924 }, { "epoch": 0.048737284275520015, "grad_norm": 3.7859811782836914, "learning_rate": 4.918855542621928e-05, "loss": 0.4534, "step": 1926 }, { "epoch": 0.04878789412419657, "grad_norm": 3.5033421516418457, "learning_rate": 4.918771192874134e-05, "loss": 0.5004, "step": 1928 }, { "epoch": 0.04883850397287312, "grad_norm": 19.82146453857422, "learning_rate": 4.91868684312634e-05, "loss": 0.7809, "step": 1930 }, { "epoch": 0.04888911382154967, "grad_norm": 7.877125263214111, "learning_rate": 4.918602493378545e-05, "loss": 0.4266, "step": 1932 }, { "epoch": 0.048939723670226226, "grad_norm": 9.093170166015625, "learning_rate": 4.918518143630751e-05, "loss": 0.7279, "step": 1934 }, { "epoch": 0.04899033351890278, "grad_norm": 6.574698448181152, "learning_rate": 4.9184337938829564e-05, "loss": 0.4833, "step": 1936 }, { "epoch": 0.04904094336757933, "grad_norm": 4.608344078063965, "learning_rate": 4.9183494441351624e-05, "loss": 0.5268, "step": 1938 }, { "epoch": 0.049091553216255884, "grad_norm": 8.013690948486328, "learning_rate": 4.9182650943873685e-05, "loss": 0.5434, "step": 1940 }, { "epoch": 0.04914216306493244, "grad_norm": 4.216137409210205, "learning_rate": 4.918180744639574e-05, "loss": 0.4351, "step": 1942 }, { "epoch": 0.04919277291360899, "grad_norm": 7.3226776123046875, "learning_rate": 4.91809639489178e-05, "loss": 0.3895, "step": 1944 }, { "epoch": 0.04924338276228554, "grad_norm": 4.8168840408325195, "learning_rate": 4.918012045143985e-05, "loss": 0.3544, "step": 1946 }, { "epoch": 0.049293992610962095, "grad_norm": 8.331216812133789, "learning_rate": 4.917927695396191e-05, "loss": 0.2721, "step": 1948 }, { "epoch": 0.04934460245963865, "grad_norm": 10.902067184448242, "learning_rate": 4.917843345648397e-05, "loss": 0.4973, "step": 1950 }, { "epoch": 0.0493952123083152, "grad_norm": 5.606088161468506, "learning_rate": 4.9177589959006025e-05, "loss": 0.5057, "step": 1952 }, { "epoch": 0.04944582215699175, "grad_norm": 14.297886848449707, "learning_rate": 4.9176746461528086e-05, "loss": 0.6632, "step": 1954 }, { "epoch": 0.049496432005668306, "grad_norm": 5.292407989501953, "learning_rate": 4.917590296405014e-05, "loss": 0.8784, "step": 1956 }, { "epoch": 0.04954704185434486, "grad_norm": 2.2982068061828613, "learning_rate": 4.91750594665722e-05, "loss": 0.3005, "step": 1958 }, { "epoch": 0.049597651703021405, "grad_norm": 5.021894931793213, "learning_rate": 4.917421596909426e-05, "loss": 0.5095, "step": 1960 }, { "epoch": 0.04964826155169796, "grad_norm": 6.863503932952881, "learning_rate": 4.917337247161631e-05, "loss": 0.5115, "step": 1962 }, { "epoch": 0.04969887140037451, "grad_norm": 8.457030296325684, "learning_rate": 4.917252897413837e-05, "loss": 0.6833, "step": 1964 }, { "epoch": 0.04974948124905106, "grad_norm": 7.3646674156188965, "learning_rate": 4.9171685476660426e-05, "loss": 0.6812, "step": 1966 }, { "epoch": 0.049800091097727615, "grad_norm": 4.478794574737549, "learning_rate": 4.917084197918249e-05, "loss": 0.3277, "step": 1968 }, { "epoch": 0.04985070094640417, "grad_norm": 8.332141876220703, "learning_rate": 4.916999848170454e-05, "loss": 0.3486, "step": 1970 }, { "epoch": 0.04990131079508072, "grad_norm": 5.418234348297119, "learning_rate": 4.91691549842266e-05, "loss": 0.5811, "step": 1972 }, { "epoch": 0.049951920643757274, "grad_norm": 7.849460124969482, "learning_rate": 4.916831148674866e-05, "loss": 0.9629, "step": 1974 }, { "epoch": 0.050002530492433826, "grad_norm": 4.812062740325928, "learning_rate": 4.9167467989270714e-05, "loss": 0.4392, "step": 1976 }, { "epoch": 0.05005314034111038, "grad_norm": 7.80704402923584, "learning_rate": 4.9166624491792774e-05, "loss": 0.9341, "step": 1978 }, { "epoch": 0.05010375018978693, "grad_norm": 4.2226243019104, "learning_rate": 4.916578099431483e-05, "loss": 0.8323, "step": 1980 }, { "epoch": 0.050154360038463484, "grad_norm": 5.1408867835998535, "learning_rate": 4.916493749683689e-05, "loss": 0.3631, "step": 1982 }, { "epoch": 0.05020496988714004, "grad_norm": 6.943533897399902, "learning_rate": 4.916409399935895e-05, "loss": 0.4674, "step": 1984 }, { "epoch": 0.05025557973581659, "grad_norm": 2.5346226692199707, "learning_rate": 4.9163250501881e-05, "loss": 0.3056, "step": 1986 }, { "epoch": 0.05030618958449314, "grad_norm": 7.4098944664001465, "learning_rate": 4.916240700440306e-05, "loss": 0.7281, "step": 1988 }, { "epoch": 0.050356799433169695, "grad_norm": 5.00679349899292, "learning_rate": 4.9161563506925115e-05, "loss": 0.5875, "step": 1990 }, { "epoch": 0.05040740928184625, "grad_norm": 7.140769004821777, "learning_rate": 4.9160720009447175e-05, "loss": 0.4304, "step": 1992 }, { "epoch": 0.0504580191305228, "grad_norm": 4.408178806304932, "learning_rate": 4.9159876511969235e-05, "loss": 0.3655, "step": 1994 }, { "epoch": 0.05050862897919935, "grad_norm": 6.3958353996276855, "learning_rate": 4.915903301449129e-05, "loss": 0.5034, "step": 1996 }, { "epoch": 0.050559238827875906, "grad_norm": 11.404850959777832, "learning_rate": 4.915818951701335e-05, "loss": 0.6704, "step": 1998 }, { "epoch": 0.05060984867655246, "grad_norm": 22.483257293701172, "learning_rate": 4.91573460195354e-05, "loss": 0.5152, "step": 2000 }, { "epoch": 0.05060984867655246, "eval_cer": 0.1947745752490278, "eval_loss": 0.306193083524704, "eval_runtime": 2641.4126, "eval_samples_per_second": 5.934, "eval_steps_per_second": 0.371, "step": 2000 } ], "logging_steps": 2, "max_steps": 118554, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 1000, "total_flos": 4.299636354947482e+19, "train_batch_size": 16, "trial_name": null, "trial_params": null }