{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.9998107374929026, "eval_steps": 500, "global_step": 3962, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0025235000946312535, "grad_norm": 3.572803497314453, "learning_rate": 1.2594458438287156e-06, "loss": 1.0672, "step": 10 }, { "epoch": 0.005047000189262507, "grad_norm": 1.6470932960510254, "learning_rate": 2.518891687657431e-06, "loss": 1.0001, "step": 20 }, { "epoch": 0.007570500283893761, "grad_norm": 1.1262171268463135, "learning_rate": 3.7783375314861467e-06, "loss": 0.9414, "step": 30 }, { "epoch": 0.010094000378525014, "grad_norm": 0.8495129346847534, "learning_rate": 5.037783375314862e-06, "loss": 0.9321, "step": 40 }, { "epoch": 0.012617500473156268, "grad_norm": 0.8612141013145447, "learning_rate": 6.297229219143577e-06, "loss": 0.8746, "step": 50 }, { "epoch": 0.015141000567787522, "grad_norm": 0.8412306308746338, "learning_rate": 7.556675062972293e-06, "loss": 0.9044, "step": 60 }, { "epoch": 0.017664500662418776, "grad_norm": 0.8401440978050232, "learning_rate": 8.816120906801008e-06, "loss": 0.9022, "step": 70 }, { "epoch": 0.020188000757050028, "grad_norm": 0.86940997838974, "learning_rate": 1.0075566750629725e-05, "loss": 0.8887, "step": 80 }, { "epoch": 0.022711500851681284, "grad_norm": 0.8858376741409302, "learning_rate": 1.133501259445844e-05, "loss": 0.8715, "step": 90 }, { "epoch": 0.025235000946312536, "grad_norm": 0.8635324239730835, "learning_rate": 1.2594458438287154e-05, "loss": 0.8709, "step": 100 }, { "epoch": 0.027758501040943788, "grad_norm": 0.9026916027069092, "learning_rate": 1.385390428211587e-05, "loss": 0.9428, "step": 110 }, { "epoch": 0.030282001135575044, "grad_norm": 0.6949167847633362, "learning_rate": 1.5113350125944587e-05, "loss": 0.9041, "step": 120 }, { "epoch": 0.0328055012302063, "grad_norm": 0.9495663046836853, "learning_rate": 1.63727959697733e-05, "loss": 0.8636, "step": 130 }, { "epoch": 0.03532900132483755, "grad_norm": 0.8648976683616638, "learning_rate": 1.7632241813602016e-05, "loss": 0.8471, "step": 140 }, { "epoch": 0.037852501419468804, "grad_norm": 0.8415878415107727, "learning_rate": 1.8891687657430733e-05, "loss": 0.9108, "step": 150 }, { "epoch": 0.040376001514100056, "grad_norm": 0.8484784364700317, "learning_rate": 2.015113350125945e-05, "loss": 0.8883, "step": 160 }, { "epoch": 0.04289950160873131, "grad_norm": 0.8620557188987732, "learning_rate": 2.1410579345088162e-05, "loss": 0.8657, "step": 170 }, { "epoch": 0.04542300170336257, "grad_norm": 0.8222241401672363, "learning_rate": 2.267002518891688e-05, "loss": 0.9174, "step": 180 }, { "epoch": 0.04794650179799382, "grad_norm": 0.8526296019554138, "learning_rate": 2.392947103274559e-05, "loss": 0.8997, "step": 190 }, { "epoch": 0.05047000189262507, "grad_norm": 0.8018633723258972, "learning_rate": 2.5188916876574308e-05, "loss": 0.9076, "step": 200 }, { "epoch": 0.052993501987256324, "grad_norm": 0.859157145023346, "learning_rate": 2.6448362720403024e-05, "loss": 0.8891, "step": 210 }, { "epoch": 0.055517002081887576, "grad_norm": 0.7516281604766846, "learning_rate": 2.770780856423174e-05, "loss": 0.9026, "step": 220 }, { "epoch": 0.058040502176518835, "grad_norm": 0.8353524804115295, "learning_rate": 2.8967254408060457e-05, "loss": 0.8393, "step": 230 }, { "epoch": 0.06056400227115009, "grad_norm": 0.7622519731521606, "learning_rate": 3.0226700251889174e-05, "loss": 0.8524, "step": 240 }, { "epoch": 0.06308750236578134, "grad_norm": 0.8780621290206909, "learning_rate": 3.148614609571788e-05, "loss": 0.9286, "step": 250 }, { "epoch": 0.0656110024604126, "grad_norm": 0.9684115052223206, "learning_rate": 3.27455919395466e-05, "loss": 0.8974, "step": 260 }, { "epoch": 0.06813450255504384, "grad_norm": 0.8870697617530823, "learning_rate": 3.4005037783375316e-05, "loss": 0.8969, "step": 270 }, { "epoch": 0.0706580026496751, "grad_norm": 0.8952546119689941, "learning_rate": 3.526448362720403e-05, "loss": 0.8689, "step": 280 }, { "epoch": 0.07318150274430635, "grad_norm": 0.9221532344818115, "learning_rate": 3.652392947103275e-05, "loss": 0.883, "step": 290 }, { "epoch": 0.07570500283893761, "grad_norm": 0.9585578441619873, "learning_rate": 3.7783375314861465e-05, "loss": 0.8508, "step": 300 }, { "epoch": 0.07822850293356887, "grad_norm": 0.9648734331130981, "learning_rate": 3.904282115869018e-05, "loss": 0.8982, "step": 310 }, { "epoch": 0.08075200302820011, "grad_norm": 0.8147997260093689, "learning_rate": 4.03022670025189e-05, "loss": 0.8694, "step": 320 }, { "epoch": 0.08327550312283137, "grad_norm": 0.8099369406700134, "learning_rate": 4.1561712846347615e-05, "loss": 0.8929, "step": 330 }, { "epoch": 0.08579900321746262, "grad_norm": 0.8512017130851746, "learning_rate": 4.2821158690176324e-05, "loss": 0.86, "step": 340 }, { "epoch": 0.08832250331209388, "grad_norm": 0.8499712347984314, "learning_rate": 4.408060453400504e-05, "loss": 0.8423, "step": 350 }, { "epoch": 0.09084600340672513, "grad_norm": 0.8530069589614868, "learning_rate": 4.534005037783376e-05, "loss": 0.8852, "step": 360 }, { "epoch": 0.09336950350135638, "grad_norm": 0.8837921023368835, "learning_rate": 4.659949622166247e-05, "loss": 0.8864, "step": 370 }, { "epoch": 0.09589300359598764, "grad_norm": 0.8840718865394592, "learning_rate": 4.785894206549118e-05, "loss": 0.8999, "step": 380 }, { "epoch": 0.09841650369061888, "grad_norm": 0.7395176887512207, "learning_rate": 4.91183879093199e-05, "loss": 0.8954, "step": 390 }, { "epoch": 0.10094000378525014, "grad_norm": 0.8697903752326965, "learning_rate": 4.999991263591223e-05, "loss": 0.8353, "step": 400 }, { "epoch": 0.1034635038798814, "grad_norm": 0.8527745008468628, "learning_rate": 4.9998359513560176e-05, "loss": 0.8591, "step": 410 }, { "epoch": 0.10598700397451265, "grad_norm": 0.806920051574707, "learning_rate": 4.999486510586282e-05, "loss": 0.9076, "step": 420 }, { "epoch": 0.1085105040691439, "grad_norm": 0.8428360223770142, "learning_rate": 4.9989429684183686e-05, "loss": 0.9032, "step": 430 }, { "epoch": 0.11103400416377515, "grad_norm": 0.8436294198036194, "learning_rate": 4.9982053670618626e-05, "loss": 0.8894, "step": 440 }, { "epoch": 0.11355750425840641, "grad_norm": 0.7664760947227478, "learning_rate": 4.997273763796312e-05, "loss": 0.8732, "step": 450 }, { "epoch": 0.11608100435303767, "grad_norm": 0.9134059548377991, "learning_rate": 4.996148230966775e-05, "loss": 0.8438, "step": 460 }, { "epoch": 0.11860450444766892, "grad_norm": 0.849233090877533, "learning_rate": 4.994828855978202e-05, "loss": 0.9281, "step": 470 }, { "epoch": 0.12112800454230017, "grad_norm": 0.8473492860794067, "learning_rate": 4.99331574128865e-05, "loss": 0.8368, "step": 480 }, { "epoch": 0.12365150463693142, "grad_norm": 0.8474506735801697, "learning_rate": 4.991609004401324e-05, "loss": 0.8852, "step": 490 }, { "epoch": 0.12617500473156268, "grad_norm": 0.7737945318222046, "learning_rate": 4.989708777855453e-05, "loss": 0.8881, "step": 500 }, { "epoch": 0.12869850482619394, "grad_norm": 0.6961573958396912, "learning_rate": 4.9876152092159994e-05, "loss": 0.9173, "step": 510 }, { "epoch": 0.1312220049208252, "grad_norm": 0.7320950031280518, "learning_rate": 4.985328461062195e-05, "loss": 0.8899, "step": 520 }, { "epoch": 0.13374550501545643, "grad_norm": 0.7261886596679688, "learning_rate": 4.98284871097492e-05, "loss": 0.8855, "step": 530 }, { "epoch": 0.1362690051100877, "grad_norm": 0.7850842475891113, "learning_rate": 4.98017615152291e-05, "loss": 0.884, "step": 540 }, { "epoch": 0.13879250520471895, "grad_norm": 0.8015512228012085, "learning_rate": 4.977310990247807e-05, "loss": 0.8767, "step": 550 }, { "epoch": 0.1413160052993502, "grad_norm": 0.8083379864692688, "learning_rate": 4.974253449648031e-05, "loss": 0.8861, "step": 560 }, { "epoch": 0.14383950539398146, "grad_norm": 0.7726438045501709, "learning_rate": 4.971003767161516e-05, "loss": 0.8747, "step": 570 }, { "epoch": 0.1463630054886127, "grad_norm": 0.7719607949256897, "learning_rate": 4.9675621951472584e-05, "loss": 0.8862, "step": 580 }, { "epoch": 0.14888650558324396, "grad_norm": 0.7348030209541321, "learning_rate": 4.9639290008657304e-05, "loss": 0.8915, "step": 590 }, { "epoch": 0.15141000567787521, "grad_norm": 0.7903275489807129, "learning_rate": 4.960104466458118e-05, "loss": 0.8916, "step": 600 }, { "epoch": 0.15393350577250647, "grad_norm": 0.778893232345581, "learning_rate": 4.956088888924414e-05, "loss": 0.8674, "step": 610 }, { "epoch": 0.15645700586713773, "grad_norm": 0.7972787618637085, "learning_rate": 4.951882580100353e-05, "loss": 0.8908, "step": 620 }, { "epoch": 0.15898050596176896, "grad_norm": 0.741663932800293, "learning_rate": 4.947485866633199e-05, "loss": 0.8876, "step": 630 }, { "epoch": 0.16150400605640022, "grad_norm": 0.8059448003768921, "learning_rate": 4.94289908995637e-05, "loss": 0.8164, "step": 640 }, { "epoch": 0.16402750615103148, "grad_norm": 0.8004572987556458, "learning_rate": 4.938122606262936e-05, "loss": 0.9075, "step": 650 }, { "epoch": 0.16655100624566274, "grad_norm": 1.5734790563583374, "learning_rate": 4.9331567864779457e-05, "loss": 0.9146, "step": 660 }, { "epoch": 0.169074506340294, "grad_norm": 0.7215042114257812, "learning_rate": 4.928002016229634e-05, "loss": 0.885, "step": 670 }, { "epoch": 0.17159800643492523, "grad_norm": 0.6652220487594604, "learning_rate": 4.9226586958194647e-05, "loss": 0.9085, "step": 680 }, { "epoch": 0.1741215065295565, "grad_norm": 0.676682710647583, "learning_rate": 4.9171272401910504e-05, "loss": 0.837, "step": 690 }, { "epoch": 0.17664500662418775, "grad_norm": 0.7034597396850586, "learning_rate": 4.9114080788979284e-05, "loss": 0.8905, "step": 700 }, { "epoch": 0.179168506718819, "grad_norm": 0.7657853960990906, "learning_rate": 4.905501656070202e-05, "loss": 0.8945, "step": 710 }, { "epoch": 0.18169200681345027, "grad_norm": 0.7395844459533691, "learning_rate": 4.8994084303800525e-05, "loss": 0.8762, "step": 720 }, { "epoch": 0.1842155069080815, "grad_norm": 0.7073786854743958, "learning_rate": 4.89312887500612e-05, "loss": 0.8824, "step": 730 }, { "epoch": 0.18673900700271276, "grad_norm": 0.7239210605621338, "learning_rate": 4.8866634775967544e-05, "loss": 0.8855, "step": 740 }, { "epoch": 0.18926250709734402, "grad_norm": 0.6406372785568237, "learning_rate": 4.880012740232154e-05, "loss": 0.8775, "step": 750 }, { "epoch": 0.19178600719197528, "grad_norm": 0.76404869556427, "learning_rate": 4.873177179385368e-05, "loss": 0.862, "step": 760 }, { "epoch": 0.19430950728660654, "grad_norm": 0.7401562929153442, "learning_rate": 4.866157325882192e-05, "loss": 0.8734, "step": 770 }, { "epoch": 0.19683300738123777, "grad_norm": 0.7563286423683167, "learning_rate": 4.858953724859948e-05, "loss": 0.8652, "step": 780 }, { "epoch": 0.19935650747586903, "grad_norm": 0.7244860529899597, "learning_rate": 4.851566935725147e-05, "loss": 0.8436, "step": 790 }, { "epoch": 0.20188000757050029, "grad_norm": 0.7061064839363098, "learning_rate": 4.843997532110051e-05, "loss": 0.8717, "step": 800 }, { "epoch": 0.20440350766513155, "grad_norm": 0.9287751913070679, "learning_rate": 4.836246101828124e-05, "loss": 0.884, "step": 810 }, { "epoch": 0.2069270077597628, "grad_norm": 0.6689929366111755, "learning_rate": 4.828313246828386e-05, "loss": 0.8871, "step": 820 }, { "epoch": 0.20945050785439404, "grad_norm": 0.7176743149757385, "learning_rate": 4.820199583148667e-05, "loss": 0.8799, "step": 830 }, { "epoch": 0.2119740079490253, "grad_norm": 0.6979175209999084, "learning_rate": 4.811905740867769e-05, "loss": 0.8874, "step": 840 }, { "epoch": 0.21449750804365655, "grad_norm": 0.780451774597168, "learning_rate": 4.803432364056535e-05, "loss": 0.8843, "step": 850 }, { "epoch": 0.2170210081382878, "grad_norm": 0.7175182700157166, "learning_rate": 4.794780110727832e-05, "loss": 0.8578, "step": 860 }, { "epoch": 0.21954450823291907, "grad_norm": 0.724116861820221, "learning_rate": 4.785949652785453e-05, "loss": 0.8869, "step": 870 }, { "epoch": 0.2220680083275503, "grad_norm": 0.8724785447120667, "learning_rate": 4.776941675971941e-05, "loss": 0.8648, "step": 880 }, { "epoch": 0.22459150842218156, "grad_norm": 0.7354777455329895, "learning_rate": 4.767756879815334e-05, "loss": 0.8683, "step": 890 }, { "epoch": 0.22711500851681282, "grad_norm": 0.7593517899513245, "learning_rate": 4.758395977574841e-05, "loss": 0.9101, "step": 900 }, { "epoch": 0.22963850861144408, "grad_norm": 0.7243201732635498, "learning_rate": 4.748859696185458e-05, "loss": 0.8945, "step": 910 }, { "epoch": 0.23216200870607534, "grad_norm": 0.6870005130767822, "learning_rate": 4.739148776201512e-05, "loss": 0.8158, "step": 920 }, { "epoch": 0.23468550880070657, "grad_norm": 0.7116649746894836, "learning_rate": 4.729263971739154e-05, "loss": 0.8855, "step": 930 }, { "epoch": 0.23720900889533783, "grad_norm": 0.6931096911430359, "learning_rate": 4.719206050417796e-05, "loss": 0.8674, "step": 940 }, { "epoch": 0.2397325089899691, "grad_norm": 0.7314079999923706, "learning_rate": 4.7089757933005016e-05, "loss": 0.8743, "step": 950 }, { "epoch": 0.24225600908460035, "grad_norm": 0.7538678646087646, "learning_rate": 4.698573994833332e-05, "loss": 0.866, "step": 960 }, { "epoch": 0.2447795091792316, "grad_norm": 0.6961751580238342, "learning_rate": 4.688001462783648e-05, "loss": 0.8528, "step": 970 }, { "epoch": 0.24730300927386284, "grad_norm": 0.7808176875114441, "learning_rate": 4.6772590181773866e-05, "loss": 0.8315, "step": 980 }, { "epoch": 0.2498265093684941, "grad_norm": 0.716074526309967, "learning_rate": 4.6663474952353004e-05, "loss": 0.8372, "step": 990 }, { "epoch": 0.25235000946312536, "grad_norm": 0.8192372918128967, "learning_rate": 4.6552677413081756e-05, "loss": 0.902, "step": 1000 }, { "epoch": 0.2548735095577566, "grad_norm": 0.7088383436203003, "learning_rate": 4.644020616811029e-05, "loss": 0.8847, "step": 1010 }, { "epoch": 0.2573970096523879, "grad_norm": 0.8579234480857849, "learning_rate": 4.6326069951562924e-05, "loss": 0.9071, "step": 1020 }, { "epoch": 0.25992050974701914, "grad_norm": 0.6537004709243774, "learning_rate": 4.6210277626859856e-05, "loss": 0.8187, "step": 1030 }, { "epoch": 0.2624440098416504, "grad_norm": 0.6265996694564819, "learning_rate": 4.609283818602884e-05, "loss": 0.8744, "step": 1040 }, { "epoch": 0.2649675099362816, "grad_norm": 0.7445203065872192, "learning_rate": 4.5973760749006963e-05, "loss": 0.8831, "step": 1050 }, { "epoch": 0.26749101003091286, "grad_norm": 0.7054116129875183, "learning_rate": 4.585305456293235e-05, "loss": 0.9171, "step": 1060 }, { "epoch": 0.2700145101255441, "grad_norm": 1.429075837135315, "learning_rate": 4.5730729001426083e-05, "loss": 0.8894, "step": 1070 }, { "epoch": 0.2725380102201754, "grad_norm": 0.6793610453605652, "learning_rate": 4.5606793563864316e-05, "loss": 0.8629, "step": 1080 }, { "epoch": 0.27506151031480663, "grad_norm": 0.6932589411735535, "learning_rate": 4.548125787464054e-05, "loss": 0.8564, "step": 1090 }, { "epoch": 0.2775850104094379, "grad_norm": 0.6600730419158936, "learning_rate": 4.535413168241821e-05, "loss": 0.8685, "step": 1100 }, { "epoch": 0.28010851050406915, "grad_norm": 0.6784124970436096, "learning_rate": 4.522542485937369e-05, "loss": 0.9024, "step": 1110 }, { "epoch": 0.2826320105987004, "grad_norm": 0.6841257214546204, "learning_rate": 4.509514740042962e-05, "loss": 0.8698, "step": 1120 }, { "epoch": 0.28515551069333167, "grad_norm": 0.7785212397575378, "learning_rate": 4.496330942247873e-05, "loss": 0.8731, "step": 1130 }, { "epoch": 0.28767901078796293, "grad_norm": 0.730110228061676, "learning_rate": 4.482992116359824e-05, "loss": 0.8542, "step": 1140 }, { "epoch": 0.29020251088259413, "grad_norm": 0.6644122004508972, "learning_rate": 4.469499298225473e-05, "loss": 0.8246, "step": 1150 }, { "epoch": 0.2927260109772254, "grad_norm": 0.7170603275299072, "learning_rate": 4.455853535649984e-05, "loss": 0.8576, "step": 1160 }, { "epoch": 0.29524951107185665, "grad_norm": 0.6883527040481567, "learning_rate": 4.442055888315646e-05, "loss": 0.8639, "step": 1170 }, { "epoch": 0.2977730111664879, "grad_norm": 0.6971318125724792, "learning_rate": 4.4281074276995936e-05, "loss": 0.8218, "step": 1180 }, { "epoch": 0.30029651126111917, "grad_norm": 0.7020850777626038, "learning_rate": 4.4140092369905914e-05, "loss": 0.8376, "step": 1190 }, { "epoch": 0.30282001135575043, "grad_norm": 0.6218104362487793, "learning_rate": 4.399762411004922e-05, "loss": 0.8741, "step": 1200 }, { "epoch": 0.3053435114503817, "grad_norm": 0.8031836152076721, "learning_rate": 4.3853680561013647e-05, "loss": 0.8977, "step": 1210 }, { "epoch": 0.30786701154501295, "grad_norm": 0.6999651789665222, "learning_rate": 4.370827290095277e-05, "loss": 0.8628, "step": 1220 }, { "epoch": 0.3103905116396442, "grad_norm": 0.6727817058563232, "learning_rate": 4.356141242171795e-05, "loss": 0.8674, "step": 1230 }, { "epoch": 0.31291401173427547, "grad_norm": 0.6965411305427551, "learning_rate": 4.3413110527981406e-05, "loss": 0.8416, "step": 1240 }, { "epoch": 0.31543751182890667, "grad_norm": 0.7655733823776245, "learning_rate": 4.3263378736350566e-05, "loss": 0.8662, "step": 1250 }, { "epoch": 0.31796101192353793, "grad_norm": 0.7115268111228943, "learning_rate": 4.311222867447375e-05, "loss": 0.9022, "step": 1260 }, { "epoch": 0.3204845120181692, "grad_norm": 0.7572771310806274, "learning_rate": 4.295967208013717e-05, "loss": 0.8649, "step": 1270 }, { "epoch": 0.32300801211280045, "grad_norm": 0.6894986629486084, "learning_rate": 4.280572080035348e-05, "loss": 0.8659, "step": 1280 }, { "epoch": 0.3255315122074317, "grad_norm": 0.6966748833656311, "learning_rate": 4.2650386790441696e-05, "loss": 0.8558, "step": 1290 }, { "epoch": 0.32805501230206296, "grad_norm": 0.7241553664207458, "learning_rate": 4.2493682113098855e-05, "loss": 0.8666, "step": 1300 }, { "epoch": 0.3305785123966942, "grad_norm": 0.6839144825935364, "learning_rate": 4.233561893746323e-05, "loss": 0.8879, "step": 1310 }, { "epoch": 0.3331020124913255, "grad_norm": 0.6955851912498474, "learning_rate": 4.217620953816935e-05, "loss": 0.8446, "step": 1320 }, { "epoch": 0.33562551258595674, "grad_norm": 0.6097539067268372, "learning_rate": 4.2015466294394756e-05, "loss": 0.8816, "step": 1330 }, { "epoch": 0.338149012680588, "grad_norm": 0.7663230299949646, "learning_rate": 4.185340168889868e-05, "loss": 0.8518, "step": 1340 }, { "epoch": 0.3406725127752192, "grad_norm": 0.6563027501106262, "learning_rate": 4.169002830705274e-05, "loss": 0.8516, "step": 1350 }, { "epoch": 0.34319601286985046, "grad_norm": 0.639011025428772, "learning_rate": 4.152535883586352e-05, "loss": 0.8324, "step": 1360 }, { "epoch": 0.3457195129644817, "grad_norm": 0.7072712779045105, "learning_rate": 4.135940606298738e-05, "loss": 0.8445, "step": 1370 }, { "epoch": 0.348243013059113, "grad_norm": 0.6532591581344604, "learning_rate": 4.119218287573743e-05, "loss": 0.8293, "step": 1380 }, { "epoch": 0.35076651315374424, "grad_norm": 0.6421136260032654, "learning_rate": 4.102370226008271e-05, "loss": 0.8809, "step": 1390 }, { "epoch": 0.3532900132483755, "grad_norm": 0.6466293931007385, "learning_rate": 4.085397729963976e-05, "loss": 0.8478, "step": 1400 }, { "epoch": 0.35581351334300676, "grad_norm": 0.7026222348213196, "learning_rate": 4.06830211746566e-05, "loss": 0.8855, "step": 1410 }, { "epoch": 0.358337013437638, "grad_norm": 0.7792401313781738, "learning_rate": 4.051084716098921e-05, "loss": 0.8523, "step": 1420 }, { "epoch": 0.3608605135322693, "grad_norm": 0.641736626625061, "learning_rate": 4.0337468629070496e-05, "loss": 0.8605, "step": 1430 }, { "epoch": 0.36338401362690054, "grad_norm": 0.6911234855651855, "learning_rate": 4.016289904287212e-05, "loss": 0.8492, "step": 1440 }, { "epoch": 0.36590751372153174, "grad_norm": 0.7274027466773987, "learning_rate": 3.9987151958858794e-05, "loss": 0.8623, "step": 1450 }, { "epoch": 0.368431013816163, "grad_norm": 0.6672956347465515, "learning_rate": 3.981024102493566e-05, "loss": 0.8309, "step": 1460 }, { "epoch": 0.37095451391079426, "grad_norm": 0.7280237078666687, "learning_rate": 3.963217997938834e-05, "loss": 0.8633, "step": 1470 }, { "epoch": 0.3734780140054255, "grad_norm": 0.749769389629364, "learning_rate": 3.945298264981614e-05, "loss": 0.8433, "step": 1480 }, { "epoch": 0.3760015141000568, "grad_norm": 0.7026387453079224, "learning_rate": 3.927266295205818e-05, "loss": 0.8665, "step": 1490 }, { "epoch": 0.37852501419468804, "grad_norm": 0.6626182794570923, "learning_rate": 3.9091234889112815e-05, "loss": 0.8597, "step": 1500 }, { "epoch": 0.3810485142893193, "grad_norm": 0.6502306461334229, "learning_rate": 3.8908712550050154e-05, "loss": 0.8652, "step": 1510 }, { "epoch": 0.38357201438395055, "grad_norm": 0.6474471688270569, "learning_rate": 3.8725110108917975e-05, "loss": 0.8258, "step": 1520 }, { "epoch": 0.3860955144785818, "grad_norm": 0.6739810109138489, "learning_rate": 3.854044182364098e-05, "loss": 0.8578, "step": 1530 }, { "epoch": 0.3886190145732131, "grad_norm": 0.7030637264251709, "learning_rate": 3.835472203491367e-05, "loss": 0.8468, "step": 1540 }, { "epoch": 0.3911425146678443, "grad_norm": 0.6305805444717407, "learning_rate": 3.816796516508658e-05, "loss": 0.8476, "step": 1550 }, { "epoch": 0.39366601476247554, "grad_norm": 0.6209976077079773, "learning_rate": 3.798018571704638e-05, "loss": 0.8376, "step": 1560 }, { "epoch": 0.3961895148571068, "grad_norm": 0.6698387265205383, "learning_rate": 3.779139827308956e-05, "loss": 0.8744, "step": 1570 }, { "epoch": 0.39871301495173805, "grad_norm": 0.7300374507904053, "learning_rate": 3.760161749379008e-05, "loss": 0.8609, "step": 1580 }, { "epoch": 0.4012365150463693, "grad_norm": 0.6837272047996521, "learning_rate": 3.7410858116860836e-05, "loss": 0.837, "step": 1590 }, { "epoch": 0.40376001514100057, "grad_norm": 0.6649072170257568, "learning_rate": 3.721913495600923e-05, "loss": 0.8694, "step": 1600 }, { "epoch": 0.40628351523563183, "grad_norm": 0.5960752367973328, "learning_rate": 3.7026462899786726e-05, "loss": 0.8129, "step": 1610 }, { "epoch": 0.4088070153302631, "grad_norm": 0.6648868322372437, "learning_rate": 3.683285691043272e-05, "loss": 0.8634, "step": 1620 }, { "epoch": 0.41133051542489435, "grad_norm": 0.7035058736801147, "learning_rate": 3.663833202271257e-05, "loss": 0.8685, "step": 1630 }, { "epoch": 0.4138540155195256, "grad_norm": 0.6673656702041626, "learning_rate": 3.6442903342750084e-05, "loss": 0.8063, "step": 1640 }, { "epoch": 0.4163775156141568, "grad_norm": 0.6990562081336975, "learning_rate": 3.624658604685443e-05, "loss": 0.8335, "step": 1650 }, { "epoch": 0.41890101570878807, "grad_norm": 0.7190445065498352, "learning_rate": 3.604939538034158e-05, "loss": 0.8509, "step": 1660 }, { "epoch": 0.42142451580341933, "grad_norm": 0.7450734376907349, "learning_rate": 3.585134665635041e-05, "loss": 0.8446, "step": 1670 }, { "epoch": 0.4239480158980506, "grad_norm": 0.6475887298583984, "learning_rate": 3.565245525465355e-05, "loss": 0.8836, "step": 1680 }, { "epoch": 0.42647151599268185, "grad_norm": 0.6419990658760071, "learning_rate": 3.5452736620463064e-05, "loss": 0.8428, "step": 1690 }, { "epoch": 0.4289950160873131, "grad_norm": 0.7428763508796692, "learning_rate": 3.525220626323097e-05, "loss": 0.8247, "step": 1700 }, { "epoch": 0.43151851618194437, "grad_norm": 0.6717978119850159, "learning_rate": 3.5050879755444877e-05, "loss": 0.881, "step": 1710 }, { "epoch": 0.4340420162765756, "grad_norm": 0.6862205862998962, "learning_rate": 3.484877273141866e-05, "loss": 0.8511, "step": 1720 }, { "epoch": 0.4365655163712069, "grad_norm": 0.6874988079071045, "learning_rate": 3.464590088607839e-05, "loss": 0.8649, "step": 1730 }, { "epoch": 0.43908901646583814, "grad_norm": 0.6635965704917908, "learning_rate": 3.444227997374345e-05, "loss": 0.8719, "step": 1740 }, { "epoch": 0.44161251656046935, "grad_norm": 0.7285788655281067, "learning_rate": 3.4237925806903184e-05, "loss": 0.8534, "step": 1750 }, { "epoch": 0.4441360166551006, "grad_norm": 0.6177170872688293, "learning_rate": 3.403285425498889e-05, "loss": 0.8516, "step": 1760 }, { "epoch": 0.44665951674973187, "grad_norm": 0.7633406519889832, "learning_rate": 3.3827081243141534e-05, "loss": 0.8193, "step": 1770 }, { "epoch": 0.4491830168443631, "grad_norm": 0.6661052107810974, "learning_rate": 3.362062275097496e-05, "loss": 0.8745, "step": 1780 }, { "epoch": 0.4517065169389944, "grad_norm": 0.7744668126106262, "learning_rate": 3.341349481133507e-05, "loss": 0.8158, "step": 1790 }, { "epoch": 0.45423001703362564, "grad_norm": 0.6634140014648438, "learning_rate": 3.320571350905466e-05, "loss": 0.8574, "step": 1800 }, { "epoch": 0.4567535171282569, "grad_norm": 0.7289906740188599, "learning_rate": 3.299729497970444e-05, "loss": 0.8776, "step": 1810 }, { "epoch": 0.45927701722288816, "grad_norm": 0.6595107913017273, "learning_rate": 3.278825540833995e-05, "loss": 0.8416, "step": 1820 }, { "epoch": 0.4618005173175194, "grad_norm": 0.6596432328224182, "learning_rate": 3.2578611028244656e-05, "loss": 0.8295, "step": 1830 }, { "epoch": 0.4643240174121507, "grad_norm": 0.7007511258125305, "learning_rate": 3.2368378119669363e-05, "loss": 0.8075, "step": 1840 }, { "epoch": 0.4668475175067819, "grad_norm": 0.5890100598335266, "learning_rate": 3.215757300856796e-05, "loss": 0.8331, "step": 1850 }, { "epoch": 0.46937101760141314, "grad_norm": 0.670438826084137, "learning_rate": 3.194621206532957e-05, "loss": 0.8739, "step": 1860 }, { "epoch": 0.4718945176960444, "grad_norm": 0.6237263083457947, "learning_rate": 3.173431170350732e-05, "loss": 0.8377, "step": 1870 }, { "epoch": 0.47441801779067566, "grad_norm": 0.7160887122154236, "learning_rate": 3.152188837854369e-05, "loss": 0.8708, "step": 1880 }, { "epoch": 0.4769415178853069, "grad_norm": 0.6525737643241882, "learning_rate": 3.130895858649264e-05, "loss": 0.8207, "step": 1890 }, { "epoch": 0.4794650179799382, "grad_norm": 0.7249549627304077, "learning_rate": 3.109553886273863e-05, "loss": 0.8516, "step": 1900 }, { "epoch": 0.48198851807456944, "grad_norm": 0.6668533682823181, "learning_rate": 3.088164578071246e-05, "loss": 0.8275, "step": 1910 }, { "epoch": 0.4845120181692007, "grad_norm": 0.7262100577354431, "learning_rate": 3.066729595060431e-05, "loss": 0.8147, "step": 1920 }, { "epoch": 0.48703551826383196, "grad_norm": 0.7166665196418762, "learning_rate": 3.0452506018073833e-05, "loss": 0.8514, "step": 1930 }, { "epoch": 0.4895590183584632, "grad_norm": 0.6810010075569153, "learning_rate": 3.0237292662957473e-05, "loss": 0.8323, "step": 1940 }, { "epoch": 0.4920825184530944, "grad_norm": 0.6473044157028198, "learning_rate": 3.0021672597973207e-05, "loss": 0.8265, "step": 1950 }, { "epoch": 0.4946060185477257, "grad_norm": 0.6784878969192505, "learning_rate": 2.9805662567422676e-05, "loss": 0.8636, "step": 1960 }, { "epoch": 0.49712951864235694, "grad_norm": 0.7378344535827637, "learning_rate": 2.9589279345890895e-05, "loss": 0.8483, "step": 1970 }, { "epoch": 0.4996530187369882, "grad_norm": 0.5715174078941345, "learning_rate": 2.9372539736943577e-05, "loss": 0.8434, "step": 1980 }, { "epoch": 0.5021765188316195, "grad_norm": 0.5842220783233643, "learning_rate": 2.9155460571822245e-05, "loss": 0.8305, "step": 1990 }, { "epoch": 0.5047000189262507, "grad_norm": 0.7206842303276062, "learning_rate": 2.893805870813717e-05, "loss": 0.8127, "step": 2000 }, { "epoch": 0.5072235190208819, "grad_norm": 0.6641551852226257, "learning_rate": 2.872035102855826e-05, "loss": 0.8272, "step": 2010 }, { "epoch": 0.5097470191155132, "grad_norm": 0.6917135119438171, "learning_rate": 2.850235443950402e-05, "loss": 0.7998, "step": 2020 }, { "epoch": 0.5122705192101444, "grad_norm": 0.6133066415786743, "learning_rate": 2.8284085869828665e-05, "loss": 0.8413, "step": 2030 }, { "epoch": 0.5147940193047758, "grad_norm": 0.6827579140663147, "learning_rate": 2.8065562269507463e-05, "loss": 0.8452, "step": 2040 }, { "epoch": 0.517317519399407, "grad_norm": 0.7090153694152832, "learning_rate": 2.7846800608320485e-05, "loss": 0.8293, "step": 2050 }, { "epoch": 0.5198410194940383, "grad_norm": 0.6256769299507141, "learning_rate": 2.7627817874534762e-05, "loss": 0.8159, "step": 2060 }, { "epoch": 0.5223645195886695, "grad_norm": 0.6957070231437683, "learning_rate": 2.7408631073585068e-05, "loss": 0.8023, "step": 2070 }, { "epoch": 0.5248880196833008, "grad_norm": 0.6817536950111389, "learning_rate": 2.7189257226753305e-05, "loss": 0.8334, "step": 2080 }, { "epoch": 0.527411519777932, "grad_norm": 0.6535147428512573, "learning_rate": 2.696971336984672e-05, "loss": 0.8558, "step": 2090 }, { "epoch": 0.5299350198725632, "grad_norm": 0.7457418441772461, "learning_rate": 2.6750016551874945e-05, "loss": 0.8244, "step": 2100 }, { "epoch": 0.5324585199671945, "grad_norm": 0.6570724248886108, "learning_rate": 2.6530183833726025e-05, "loss": 0.8283, "step": 2110 }, { "epoch": 0.5349820200618257, "grad_norm": 0.7065024375915527, "learning_rate": 2.6310232286841546e-05, "loss": 0.8565, "step": 2120 }, { "epoch": 0.537505520156457, "grad_norm": 0.671667218208313, "learning_rate": 2.609017899189092e-05, "loss": 0.8447, "step": 2130 }, { "epoch": 0.5400290202510882, "grad_norm": 0.6672875285148621, "learning_rate": 2.587004103744495e-05, "loss": 0.7912, "step": 2140 }, { "epoch": 0.5425525203457195, "grad_norm": 0.6282544732093811, "learning_rate": 2.564983551864882e-05, "loss": 0.8079, "step": 2150 }, { "epoch": 0.5450760204403508, "grad_norm": 0.7435926795005798, "learning_rate": 2.54295795358945e-05, "loss": 0.8342, "step": 2160 }, { "epoch": 0.5475995205349821, "grad_norm": 0.6785821318626404, "learning_rate": 2.5209290193492834e-05, "loss": 0.8281, "step": 2170 }, { "epoch": 0.5501230206296133, "grad_norm": 0.6483226418495178, "learning_rate": 2.4988984598345247e-05, "loss": 0.79, "step": 2180 }, { "epoch": 0.5526465207242445, "grad_norm": 0.6465590000152588, "learning_rate": 2.4768679858615304e-05, "loss": 0.841, "step": 2190 }, { "epoch": 0.5551700208188758, "grad_norm": 0.7468442916870117, "learning_rate": 2.454839308240014e-05, "loss": 0.8717, "step": 2200 }, { "epoch": 0.557693520913507, "grad_norm": 0.6535473465919495, "learning_rate": 2.4328141376401903e-05, "loss": 0.826, "step": 2210 }, { "epoch": 0.5602170210081383, "grad_norm": 0.6404563188552856, "learning_rate": 2.4107941844599312e-05, "loss": 0.8062, "step": 2220 }, { "epoch": 0.5627405211027695, "grad_norm": 0.6602795720100403, "learning_rate": 2.3887811586919424e-05, "loss": 0.8418, "step": 2230 }, { "epoch": 0.5652640211974008, "grad_norm": 0.6988357305526733, "learning_rate": 2.3667767697909694e-05, "loss": 0.8177, "step": 2240 }, { "epoch": 0.567787521292032, "grad_norm": 0.6755298376083374, "learning_rate": 2.3447827265410517e-05, "loss": 0.8653, "step": 2250 }, { "epoch": 0.5703110213866633, "grad_norm": 0.72756028175354, "learning_rate": 2.3228007369228178e-05, "loss": 0.8896, "step": 2260 }, { "epoch": 0.5728345214812945, "grad_norm": 0.6584864854812622, "learning_rate": 2.3008325079808576e-05, "loss": 0.8393, "step": 2270 }, { "epoch": 0.5753580215759259, "grad_norm": 0.6699262857437134, "learning_rate": 2.2788797456911503e-05, "loss": 0.7976, "step": 2280 }, { "epoch": 0.5778815216705571, "grad_norm": 0.7463390827178955, "learning_rate": 2.2569441548285934e-05, "loss": 0.8321, "step": 2290 }, { "epoch": 0.5804050217651883, "grad_norm": 0.542870283126831, "learning_rate": 2.2350274388346064e-05, "loss": 0.786, "step": 2300 }, { "epoch": 0.5829285218598196, "grad_norm": 0.652056872844696, "learning_rate": 2.213131299684858e-05, "loss": 0.848, "step": 2310 }, { "epoch": 0.5854520219544508, "grad_norm": 0.7307469248771667, "learning_rate": 2.191257437757086e-05, "loss": 0.8117, "step": 2320 }, { "epoch": 0.5879755220490821, "grad_norm": 0.6336262822151184, "learning_rate": 2.16940755169906e-05, "loss": 0.8417, "step": 2330 }, { "epoch": 0.5904990221437133, "grad_norm": 0.7636166214942932, "learning_rate": 2.1475833382966647e-05, "loss": 0.8786, "step": 2340 }, { "epoch": 0.5930225222383446, "grad_norm": 0.6622100472450256, "learning_rate": 2.1257864923421404e-05, "loss": 0.8629, "step": 2350 }, { "epoch": 0.5955460223329758, "grad_norm": 0.602483332157135, "learning_rate": 2.1040187065024605e-05, "loss": 0.7786, "step": 2360 }, { "epoch": 0.5980695224276071, "grad_norm": 0.6503065824508667, "learning_rate": 2.0822816711878978e-05, "loss": 0.8445, "step": 2370 }, { "epoch": 0.6005930225222383, "grad_norm": 0.6901794672012329, "learning_rate": 2.0605770744207413e-05, "loss": 0.8259, "step": 2380 }, { "epoch": 0.6031165226168695, "grad_norm": 0.7173271179199219, "learning_rate": 2.0389066017042192e-05, "loss": 0.802, "step": 2390 }, { "epoch": 0.6056400227115009, "grad_norm": 0.7431663870811462, "learning_rate": 2.0172719358916042e-05, "loss": 0.8092, "step": 2400 }, { "epoch": 0.6081635228061321, "grad_norm": 0.7227687239646912, "learning_rate": 1.9956747570555288e-05, "loss": 0.8563, "step": 2410 }, { "epoch": 0.6106870229007634, "grad_norm": 0.6300061345100403, "learning_rate": 1.9741167423575186e-05, "loss": 0.7849, "step": 2420 }, { "epoch": 0.6132105229953946, "grad_norm": 0.6208367347717285, "learning_rate": 1.9525995659177484e-05, "loss": 0.8239, "step": 2430 }, { "epoch": 0.6157340230900259, "grad_norm": 0.6272019147872925, "learning_rate": 1.9311248986850365e-05, "loss": 0.8102, "step": 2440 }, { "epoch": 0.6182575231846571, "grad_norm": 0.6594968438148499, "learning_rate": 1.9096944083070866e-05, "loss": 0.8266, "step": 2450 }, { "epoch": 0.6207810232792884, "grad_norm": 0.673553466796875, "learning_rate": 1.8883097590009775e-05, "loss": 0.8375, "step": 2460 }, { "epoch": 0.6233045233739196, "grad_norm": 0.7199084162712097, "learning_rate": 1.866972611423936e-05, "loss": 0.8188, "step": 2470 }, { "epoch": 0.6258280234685509, "grad_norm": 0.697413444519043, "learning_rate": 1.8456846225443648e-05, "loss": 0.7709, "step": 2480 }, { "epoch": 0.6283515235631821, "grad_norm": 0.6711037158966064, "learning_rate": 1.8244474455131792e-05, "loss": 0.8156, "step": 2490 }, { "epoch": 0.6308750236578133, "grad_norm": 0.7030087113380432, "learning_rate": 1.8032627295354183e-05, "loss": 0.8125, "step": 2500 }, { "epoch": 0.6333985237524447, "grad_norm": 0.7960418462753296, "learning_rate": 1.7821321197421837e-05, "loss": 0.8604, "step": 2510 }, { "epoch": 0.6359220238470759, "grad_norm": 0.6948102116584778, "learning_rate": 1.761057257062876e-05, "loss": 0.8301, "step": 2520 }, { "epoch": 0.6384455239417072, "grad_norm": 0.5919877290725708, "learning_rate": 1.740039778097772e-05, "loss": 0.7821, "step": 2530 }, { "epoch": 0.6409690240363384, "grad_norm": 0.6569110751152039, "learning_rate": 1.7190813149909274e-05, "loss": 0.8213, "step": 2540 }, { "epoch": 0.6434925241309697, "grad_norm": 0.677099347114563, "learning_rate": 1.6981834953034344e-05, "loss": 0.8278, "step": 2550 }, { "epoch": 0.6460160242256009, "grad_norm": 0.7233052253723145, "learning_rate": 1.677347941887028e-05, "loss": 0.7919, "step": 2560 }, { "epoch": 0.6485395243202322, "grad_norm": 0.7088631987571716, "learning_rate": 1.656576272758061e-05, "loss": 0.8444, "step": 2570 }, { "epoch": 0.6510630244148634, "grad_norm": 0.6909515857696533, "learning_rate": 1.6358701009718577e-05, "loss": 0.8222, "step": 2580 }, { "epoch": 0.6535865245094946, "grad_norm": 0.5979318618774414, "learning_rate": 1.615231034497444e-05, "loss": 0.8376, "step": 2590 }, { "epoch": 0.6561100246041259, "grad_norm": 0.7273426055908203, "learning_rate": 1.5946606760926865e-05, "loss": 0.8037, "step": 2600 }, { "epoch": 0.6586335246987571, "grad_norm": 0.719450056552887, "learning_rate": 1.574160623179816e-05, "loss": 0.8268, "step": 2610 }, { "epoch": 0.6611570247933884, "grad_norm": 0.7163055539131165, "learning_rate": 1.553732467721392e-05, "loss": 0.7853, "step": 2620 }, { "epoch": 0.6636805248880197, "grad_norm": 0.6172025799751282, "learning_rate": 1.5333777960966616e-05, "loss": 0.7926, "step": 2630 }, { "epoch": 0.666204024982651, "grad_norm": 0.6272744536399841, "learning_rate": 1.5130981889783795e-05, "loss": 0.7982, "step": 2640 }, { "epoch": 0.6687275250772822, "grad_norm": 0.680596649646759, "learning_rate": 1.4928952212100483e-05, "loss": 0.8312, "step": 2650 }, { "epoch": 0.6712510251719135, "grad_norm": 0.6080834865570068, "learning_rate": 1.4727704616836296e-05, "loss": 0.8273, "step": 2660 }, { "epoch": 0.6737745252665447, "grad_norm": 0.6613759398460388, "learning_rate": 1.4527254732177043e-05, "loss": 0.8141, "step": 2670 }, { "epoch": 0.676298025361176, "grad_norm": 0.6180728077888489, "learning_rate": 1.4327618124361114e-05, "loss": 0.8231, "step": 2680 }, { "epoch": 0.6788215254558072, "grad_norm": 0.6406080722808838, "learning_rate": 1.412881029647065e-05, "loss": 0.7876, "step": 2690 }, { "epoch": 0.6813450255504384, "grad_norm": 0.6109746098518372, "learning_rate": 1.3930846687227664e-05, "loss": 0.7957, "step": 2700 }, { "epoch": 0.6838685256450697, "grad_norm": 0.6827517747879028, "learning_rate": 1.3733742669795049e-05, "loss": 0.8428, "step": 2710 }, { "epoch": 0.6863920257397009, "grad_norm": 0.7277110815048218, "learning_rate": 1.3537513550582853e-05, "loss": 0.8326, "step": 2720 }, { "epoch": 0.6889155258343322, "grad_norm": 0.597568154335022, "learning_rate": 1.3342174568059527e-05, "loss": 0.7998, "step": 2730 }, { "epoch": 0.6914390259289634, "grad_norm": 0.6378962993621826, "learning_rate": 1.3147740891568661e-05, "loss": 0.785, "step": 2740 }, { "epoch": 0.6939625260235948, "grad_norm": 0.6579405069351196, "learning_rate": 1.2954227620150904e-05, "loss": 0.8332, "step": 2750 }, { "epoch": 0.696486026118226, "grad_norm": 0.6977427005767822, "learning_rate": 1.2761649781371479e-05, "loss": 0.8095, "step": 2760 }, { "epoch": 0.6990095262128573, "grad_norm": 0.6410185098648071, "learning_rate": 1.257002233015318e-05, "loss": 0.8341, "step": 2770 }, { "epoch": 0.7015330263074885, "grad_norm": 0.6869609355926514, "learning_rate": 1.2379360147614994e-05, "loss": 0.8023, "step": 2780 }, { "epoch": 0.7040565264021197, "grad_norm": 0.6658973097801208, "learning_rate": 1.2189678039916532e-05, "loss": 0.7755, "step": 2790 }, { "epoch": 0.706580026496751, "grad_norm": 0.6188139915466309, "learning_rate": 1.2000990737108225e-05, "loss": 0.796, "step": 2800 }, { "epoch": 0.7091035265913822, "grad_norm": 0.7432144284248352, "learning_rate": 1.1813312891987392e-05, "loss": 0.8381, "step": 2810 }, { "epoch": 0.7116270266860135, "grad_norm": 0.6776263117790222, "learning_rate": 1.1626659078960424e-05, "loss": 0.8087, "step": 2820 }, { "epoch": 0.7141505267806447, "grad_norm": 0.6468738913536072, "learning_rate": 1.1441043792910936e-05, "loss": 0.8032, "step": 2830 }, { "epoch": 0.716674026875276, "grad_norm": 0.7177358865737915, "learning_rate": 1.1256481448074179e-05, "loss": 0.8039, "step": 2840 }, { "epoch": 0.7191975269699072, "grad_norm": 0.6401441693305969, "learning_rate": 1.1072986376917638e-05, "loss": 0.8135, "step": 2850 }, { "epoch": 0.7217210270645386, "grad_norm": 0.6511224508285522, "learning_rate": 1.0890572829028087e-05, "loss": 0.8496, "step": 2860 }, { "epoch": 0.7242445271591698, "grad_norm": 0.632625162601471, "learning_rate": 1.0709254970004937e-05, "loss": 0.7964, "step": 2870 }, { "epoch": 0.7267680272538011, "grad_norm": 0.5535660982131958, "learning_rate": 1.0529046880360263e-05, "loss": 0.7932, "step": 2880 }, { "epoch": 0.7292915273484323, "grad_norm": 0.5996463298797607, "learning_rate": 1.034996255442529e-05, "loss": 0.8437, "step": 2890 }, { "epoch": 0.7318150274430635, "grad_norm": 0.6257640719413757, "learning_rate": 1.0172015899263712e-05, "loss": 0.8069, "step": 2900 }, { "epoch": 0.7343385275376948, "grad_norm": 0.6533858776092529, "learning_rate": 9.995220733591639e-06, "loss": 0.7921, "step": 2910 }, { "epoch": 0.736862027632326, "grad_norm": 0.6002010107040405, "learning_rate": 9.819590786704572e-06, "loss": 0.8307, "step": 2920 }, { "epoch": 0.7393855277269573, "grad_norm": 0.6418666243553162, "learning_rate": 9.645139697411149e-06, "loss": 0.8036, "step": 2930 }, { "epoch": 0.7419090278215885, "grad_norm": 0.6554102897644043, "learning_rate": 9.471881012974071e-06, "loss": 0.8285, "step": 2940 }, { "epoch": 0.7444325279162198, "grad_norm": 0.6879960894584656, "learning_rate": 9.299828188058013e-06, "loss": 0.8154, "step": 2950 }, { "epoch": 0.746956028010851, "grad_norm": 0.6418633460998535, "learning_rate": 9.128994583684838e-06, "loss": 0.7945, "step": 2960 }, { "epoch": 0.7494795281054824, "grad_norm": 0.6467211246490479, "learning_rate": 8.959393466195972e-06, "loss": 0.8464, "step": 2970 }, { "epoch": 0.7520030282001136, "grad_norm": 0.6477042436599731, "learning_rate": 8.791038006222233e-06, "loss": 0.8235, "step": 2980 }, { "epoch": 0.7545265282947448, "grad_norm": 0.6426742672920227, "learning_rate": 8.623941277660994e-06, "loss": 0.8001, "step": 2990 }, { "epoch": 0.7570500283893761, "grad_norm": 0.7026243805885315, "learning_rate": 8.458116256660981e-06, "loss": 0.842, "step": 3000 }, { "epoch": 0.7595735284840073, "grad_norm": 0.6429437398910522, "learning_rate": 8.293575820614508e-06, "loss": 0.8143, "step": 3010 }, { "epoch": 0.7620970285786386, "grad_norm": 0.654498815536499, "learning_rate": 8.130332747157542e-06, "loss": 0.7697, "step": 3020 }, { "epoch": 0.7646205286732698, "grad_norm": 0.8270076513290405, "learning_rate": 7.968399713177366e-06, "loss": 0.825, "step": 3030 }, { "epoch": 0.7671440287679011, "grad_norm": 0.6423079967498779, "learning_rate": 7.807789293828204e-06, "loss": 0.8366, "step": 3040 }, { "epoch": 0.7696675288625323, "grad_norm": 0.662451446056366, "learning_rate": 7.648513961554607e-06, "loss": 0.7695, "step": 3050 }, { "epoch": 0.7721910289571636, "grad_norm": 0.5953843593597412, "learning_rate": 7.4905860851229605e-06, "loss": 0.8296, "step": 3060 }, { "epoch": 0.7747145290517948, "grad_norm": 0.7210749387741089, "learning_rate": 7.334017928660902e-06, "loss": 0.8201, "step": 3070 }, { "epoch": 0.7772380291464261, "grad_norm": 0.6214151382446289, "learning_rate": 7.1788216507049865e-06, "loss": 0.8034, "step": 3080 }, { "epoch": 0.7797615292410573, "grad_norm": 0.6791695356369019, "learning_rate": 7.0250093032564494e-06, "loss": 0.7624, "step": 3090 }, { "epoch": 0.7822850293356886, "grad_norm": 0.6388612985610962, "learning_rate": 6.872592830845339e-06, "loss": 0.8004, "step": 3100 }, { "epoch": 0.7848085294303199, "grad_norm": 0.5958021283149719, "learning_rate": 6.72158406960289e-06, "loss": 0.8275, "step": 3110 }, { "epoch": 0.7873320295249511, "grad_norm": 0.572040855884552, "learning_rate": 6.571994746342439e-06, "loss": 0.8078, "step": 3120 }, { "epoch": 0.7898555296195824, "grad_norm": 0.6328415274620056, "learning_rate": 6.4238364776486785e-06, "loss": 0.7883, "step": 3130 }, { "epoch": 0.7923790297142136, "grad_norm": 0.6552072763442993, "learning_rate": 6.277120768975644e-06, "loss": 0.8398, "step": 3140 }, { "epoch": 0.7949025298088449, "grad_norm": 0.7182049751281738, "learning_rate": 6.131859013753155e-06, "loss": 0.7919, "step": 3150 }, { "epoch": 0.7974260299034761, "grad_norm": 0.7126038074493408, "learning_rate": 5.988062492502117e-06, "loss": 0.7782, "step": 3160 }, { "epoch": 0.7999495299981074, "grad_norm": 0.6005820631980896, "learning_rate": 5.8457423719584435e-06, "loss": 0.7979, "step": 3170 }, { "epoch": 0.8024730300927386, "grad_norm": 0.6624283790588379, "learning_rate": 5.704909704205949e-06, "loss": 0.8297, "step": 3180 }, { "epoch": 0.8049965301873698, "grad_norm": 0.6289507150650024, "learning_rate": 5.565575425818054e-06, "loss": 0.8147, "step": 3190 }, { "epoch": 0.8075200302820011, "grad_norm": 0.6975149512290955, "learning_rate": 5.427750357008468e-06, "loss": 0.7733, "step": 3200 }, { "epoch": 0.8100435303766323, "grad_norm": 0.6802620887756348, "learning_rate": 5.291445200790982e-06, "loss": 0.8226, "step": 3210 }, { "epoch": 0.8125670304712637, "grad_norm": 0.6158818602561951, "learning_rate": 5.156670542148267e-06, "loss": 0.8282, "step": 3220 }, { "epoch": 0.8150905305658949, "grad_norm": 0.7228125333786011, "learning_rate": 5.023436847209887e-06, "loss": 0.816, "step": 3230 }, { "epoch": 0.8176140306605262, "grad_norm": 0.6515725255012512, "learning_rate": 4.891754462439557e-06, "loss": 0.775, "step": 3240 }, { "epoch": 0.8201375307551574, "grad_norm": 0.6829689741134644, "learning_rate": 4.761633613831645e-06, "loss": 0.8156, "step": 3250 }, { "epoch": 0.8226610308497887, "grad_norm": 0.7261675596237183, "learning_rate": 4.6330844061170914e-06, "loss": 0.7862, "step": 3260 }, { "epoch": 0.8251845309444199, "grad_norm": 0.6911167502403259, "learning_rate": 4.506116821978662e-06, "loss": 0.8016, "step": 3270 }, { "epoch": 0.8277080310390512, "grad_norm": 0.5780116319656372, "learning_rate": 4.380740721275786e-06, "loss": 0.824, "step": 3280 }, { "epoch": 0.8302315311336824, "grad_norm": 0.6704926490783691, "learning_rate": 4.25696584027882e-06, "loss": 0.8037, "step": 3290 }, { "epoch": 0.8327550312283136, "grad_norm": 0.7162071466445923, "learning_rate": 4.134801790913006e-06, "loss": 0.7651, "step": 3300 }, { "epoch": 0.8352785313229449, "grad_norm": 0.7350740432739258, "learning_rate": 4.014258060012005e-06, "loss": 0.8278, "step": 3310 }, { "epoch": 0.8378020314175761, "grad_norm": 0.6031658053398132, "learning_rate": 3.895344008581222e-06, "loss": 0.7945, "step": 3320 }, { "epoch": 0.8403255315122075, "grad_norm": 0.6996452212333679, "learning_rate": 3.7780688710708223e-06, "loss": 0.7821, "step": 3330 }, { "epoch": 0.8428490316068387, "grad_norm": 0.6655017733573914, "learning_rate": 3.6624417546586574e-06, "loss": 0.7526, "step": 3340 }, { "epoch": 0.84537253170147, "grad_norm": 0.7387165427207947, "learning_rate": 3.548471638542991e-06, "loss": 0.8259, "step": 3350 }, { "epoch": 0.8478960317961012, "grad_norm": 0.6410266757011414, "learning_rate": 3.436167373245247e-06, "loss": 0.8054, "step": 3360 }, { "epoch": 0.8504195318907325, "grad_norm": 0.6522373557090759, "learning_rate": 3.325537679922672e-06, "loss": 0.8168, "step": 3370 }, { "epoch": 0.8529430319853637, "grad_norm": 0.7458412647247314, "learning_rate": 3.2165911496911173e-06, "loss": 0.7892, "step": 3380 }, { "epoch": 0.8554665320799949, "grad_norm": 0.6441506743431091, "learning_rate": 3.1093362429578414e-06, "loss": 0.8105, "step": 3390 }, { "epoch": 0.8579900321746262, "grad_norm": 0.5970674753189087, "learning_rate": 3.0037812887645483e-06, "loss": 0.8326, "step": 3400 }, { "epoch": 0.8605135322692574, "grad_norm": 0.6173757314682007, "learning_rate": 2.8999344841405373e-06, "loss": 0.7956, "step": 3410 }, { "epoch": 0.8630370323638887, "grad_norm": 0.6268020868301392, "learning_rate": 2.7978038934662024e-06, "loss": 0.7859, "step": 3420 }, { "epoch": 0.8655605324585199, "grad_norm": 0.6534834504127502, "learning_rate": 2.697397447846725e-06, "loss": 0.8041, "step": 3430 }, { "epoch": 0.8680840325531513, "grad_norm": 0.6108519434928894, "learning_rate": 2.5987229444962237e-06, "loss": 0.823, "step": 3440 }, { "epoch": 0.8706075326477825, "grad_norm": 0.6347935795783997, "learning_rate": 2.501788046132203e-06, "loss": 0.831, "step": 3450 }, { "epoch": 0.8731310327424138, "grad_norm": 0.6183903813362122, "learning_rate": 2.4066002803805386e-06, "loss": 0.7974, "step": 3460 }, { "epoch": 0.875654532837045, "grad_norm": 0.6723082065582275, "learning_rate": 2.313167039190861e-06, "loss": 0.8058, "step": 3470 }, { "epoch": 0.8781780329316763, "grad_norm": 0.6427431702613831, "learning_rate": 2.2214955782625752e-06, "loss": 0.805, "step": 3480 }, { "epoch": 0.8807015330263075, "grad_norm": 0.7344009280204773, "learning_rate": 2.1315930164813507e-06, "loss": 0.8366, "step": 3490 }, { "epoch": 0.8832250331209387, "grad_norm": 0.6524431109428406, "learning_rate": 2.0434663353663536e-06, "loss": 0.8022, "step": 3500 }, { "epoch": 0.88574853321557, "grad_norm": 0.6769471168518066, "learning_rate": 1.9571223785280314e-06, "loss": 0.8062, "step": 3510 }, { "epoch": 0.8882720333102012, "grad_norm": 0.6867194771766663, "learning_rate": 1.8725678511367001e-06, "loss": 0.8171, "step": 3520 }, { "epoch": 0.8907955334048325, "grad_norm": 0.6660215854644775, "learning_rate": 1.789809319401825e-06, "loss": 0.8169, "step": 3530 }, { "epoch": 0.8933190334994637, "grad_norm": 0.6402613520622253, "learning_rate": 1.7088532100621224e-06, "loss": 0.7813, "step": 3540 }, { "epoch": 0.895842533594095, "grad_norm": 0.6413708925247192, "learning_rate": 1.629705809886467e-06, "loss": 0.7837, "step": 3550 }, { "epoch": 0.8983660336887263, "grad_norm": 0.6048439741134644, "learning_rate": 1.5523732651857082e-06, "loss": 0.7984, "step": 3560 }, { "epoch": 0.9008895337833576, "grad_norm": 0.6774916052818298, "learning_rate": 1.4768615813353398e-06, "loss": 0.8033, "step": 3570 }, { "epoch": 0.9034130338779888, "grad_norm": 0.6154995560646057, "learning_rate": 1.4031766223091603e-06, "loss": 0.8015, "step": 3580 }, { "epoch": 0.90593653397262, "grad_norm": 0.6018934845924377, "learning_rate": 1.3313241102239054e-06, "loss": 0.7761, "step": 3590 }, { "epoch": 0.9084600340672513, "grad_norm": 0.658366322517395, "learning_rate": 1.261309624894863e-06, "loss": 0.8173, "step": 3600 }, { "epoch": 0.9109835341618825, "grad_norm": 0.6167306900024414, "learning_rate": 1.1931386034025882e-06, "loss": 0.8024, "step": 3610 }, { "epoch": 0.9135070342565138, "grad_norm": 0.5509990453720093, "learning_rate": 1.1268163396706583e-06, "loss": 0.8128, "step": 3620 }, { "epoch": 0.916030534351145, "grad_norm": 0.6154832243919373, "learning_rate": 1.0623479840545874e-06, "loss": 0.7569, "step": 3630 }, { "epoch": 0.9185540344457763, "grad_norm": 0.679389238357544, "learning_rate": 9.997385429418555e-07, "loss": 0.8276, "step": 3640 }, { "epoch": 0.9210775345404075, "grad_norm": 0.662276566028595, "learning_rate": 9.389928783631207e-07, "loss": 0.8304, "step": 3650 }, { "epoch": 0.9236010346350388, "grad_norm": 0.6233845949172974, "learning_rate": 8.801157076146705e-07, "loss": 0.7851, "step": 3660 }, { "epoch": 0.92612453472967, "grad_norm": 0.7036879658699036, "learning_rate": 8.231116028920765e-07, "loss": 0.793, "step": 3670 }, { "epoch": 0.9286480348243014, "grad_norm": 0.6103026270866394, "learning_rate": 7.679849909351472e-07, "loss": 0.7818, "step": 3680 }, { "epoch": 0.9311715349189326, "grad_norm": 0.6900059580802917, "learning_rate": 7.147401526841485e-07, "loss": 0.773, "step": 3690 }, { "epoch": 0.9336950350135638, "grad_norm": 0.681058943271637, "learning_rate": 6.633812229473791e-07, "loss": 0.8357, "step": 3700 }, { "epoch": 0.9362185351081951, "grad_norm": 0.7187952995300293, "learning_rate": 6.139121900800515e-07, "loss": 0.7779, "step": 3710 }, { "epoch": 0.9387420352028263, "grad_norm": 0.6179840564727783, "learning_rate": 5.663368956745963e-07, "loss": 0.7871, "step": 3720 }, { "epoch": 0.9412655352974576, "grad_norm": 0.6663089394569397, "learning_rate": 5.206590342623164e-07, "loss": 0.7901, "step": 3730 }, { "epoch": 0.9437890353920888, "grad_norm": 0.6079100370407104, "learning_rate": 4.768821530264977e-07, "loss": 0.8226, "step": 3740 }, { "epoch": 0.9463125354867201, "grad_norm": 0.68614262342453, "learning_rate": 4.350096515269325e-07, "loss": 0.8185, "step": 3750 }, { "epoch": 0.9488360355813513, "grad_norm": 0.6491347551345825, "learning_rate": 3.950447814359409e-07, "loss": 0.817, "step": 3760 }, { "epoch": 0.9513595356759826, "grad_norm": 0.6513685584068298, "learning_rate": 3.5699064628583745e-07, "loss": 0.7997, "step": 3770 }, { "epoch": 0.9538830357706138, "grad_norm": 0.6080814003944397, "learning_rate": 3.2085020122793186e-07, "loss": 0.7956, "step": 3780 }, { "epoch": 0.956406535865245, "grad_norm": 0.6476254463195801, "learning_rate": 2.8662625280304613e-07, "loss": 0.7888, "step": 3790 }, { "epoch": 0.9589300359598764, "grad_norm": 0.6439909934997559, "learning_rate": 2.5432145872355816e-07, "loss": 0.7847, "step": 3800 }, { "epoch": 0.9614535360545076, "grad_norm": 0.6744981408119202, "learning_rate": 2.2393832766701706e-07, "loss": 0.8093, "step": 3810 }, { "epoch": 0.9639770361491389, "grad_norm": 0.5795860886573792, "learning_rate": 1.9547921908133483e-07, "loss": 0.8082, "step": 3820 }, { "epoch": 0.9665005362437701, "grad_norm": 0.6693094968795776, "learning_rate": 1.689463430015442e-07, "loss": 0.7857, "step": 3830 }, { "epoch": 0.9690240363384014, "grad_norm": 0.645203173160553, "learning_rate": 1.443417598781971e-07, "loss": 0.8056, "step": 3840 }, { "epoch": 0.9715475364330326, "grad_norm": 0.6820341348648071, "learning_rate": 1.2166738041733684e-07, "loss": 0.802, "step": 3850 }, { "epoch": 0.9740710365276639, "grad_norm": 0.6292694807052612, "learning_rate": 1.0092496543212814e-07, "loss": 0.7937, "step": 3860 }, { "epoch": 0.9765945366222951, "grad_norm": 0.6253132224082947, "learning_rate": 8.211612570611926e-08, "loss": 0.7846, "step": 3870 }, { "epoch": 0.9791180367169264, "grad_norm": 0.6571831107139587, "learning_rate": 6.524232186815305e-08, "loss": 0.785, "step": 3880 }, { "epoch": 0.9816415368115576, "grad_norm": 0.6356094479560852, "learning_rate": 5.03048642789411e-08, "loss": 0.7789, "step": 3890 }, { "epoch": 0.9841650369061888, "grad_norm": 0.8404703140258789, "learning_rate": 3.730491292930072e-08, "loss": 0.7954, "step": 3900 }, { "epoch": 0.9866885370008202, "grad_norm": 0.7891058325767517, "learning_rate": 2.624347735007693e-08, "loss": 0.8129, "step": 3910 }, { "epoch": 0.9892120370954514, "grad_norm": 0.6858798265457153, "learning_rate": 1.7121416533749658e-08, "loss": 0.8076, "step": 3920 }, { "epoch": 0.9917355371900827, "grad_norm": 0.6489024758338928, "learning_rate": 9.939438867723194e-09, "loss": 0.8087, "step": 3930 }, { "epoch": 0.9942590372847139, "grad_norm": 0.6204003691673279, "learning_rate": 4.6981020793118725e-09, "loss": 0.8162, "step": 3940 }, { "epoch": 0.9967825373793452, "grad_norm": 0.6356140971183777, "learning_rate": 1.3978131924385906e-09, "loss": 0.7862, "step": 3950 }, { "epoch": 0.9993060374739764, "grad_norm": 0.6472454071044922, "learning_rate": 3.88284960184393e-11, "loss": 0.8188, "step": 3960 } ], "logging_steps": 10, "max_steps": 3962, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 1000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2.0380844918675866e+18, "train_batch_size": 2, "trial_name": null, "trial_params": null }