{ "best_metric": null, "best_model_checkpoint": null, "epoch": 9.992193598750976, "eval_steps": 500, "global_step": 6400, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.00156128024980484, "grad_norm": 15.554603576660156, "learning_rate": 3.125e-07, "loss": 1.2028, "step": 1 }, { "epoch": 0.0078064012490242, "grad_norm": 14.548426628112793, "learning_rate": 1.5625e-06, "loss": 1.23, "step": 5 }, { "epoch": 0.0156128024980484, "grad_norm": 10.735854148864746, "learning_rate": 3.125e-06, "loss": 1.1711, "step": 10 }, { "epoch": 0.0234192037470726, "grad_norm": 9.330694198608398, "learning_rate": 4.6875000000000004e-06, "loss": 1.137, "step": 15 }, { "epoch": 0.0312256049960968, "grad_norm": 5.181936740875244, "learning_rate": 6.25e-06, "loss": 1.0235, "step": 20 }, { "epoch": 0.039032006245121, "grad_norm": 3.862912654876709, "learning_rate": 7.8125e-06, "loss": 0.9359, "step": 25 }, { "epoch": 0.0468384074941452, "grad_norm": 3.142946720123291, "learning_rate": 9.375000000000001e-06, "loss": 0.8746, "step": 30 }, { "epoch": 0.0546448087431694, "grad_norm": 1.8955553770065308, "learning_rate": 1.09375e-05, "loss": 0.8135, "step": 35 }, { "epoch": 0.0624512099921936, "grad_norm": 1.3491796255111694, "learning_rate": 1.25e-05, "loss": 0.7632, "step": 40 }, { "epoch": 0.0702576112412178, "grad_norm": 1.4237151145935059, "learning_rate": 1.4062500000000001e-05, "loss": 0.7152, "step": 45 }, { "epoch": 0.078064012490242, "grad_norm": 1.992169737815857, "learning_rate": 1.5625e-05, "loss": 0.6802, "step": 50 }, { "epoch": 0.0858704137392662, "grad_norm": 1.0017743110656738, "learning_rate": 1.71875e-05, "loss": 0.6617, "step": 55 }, { "epoch": 0.0936768149882904, "grad_norm": 0.9323798418045044, "learning_rate": 1.8750000000000002e-05, "loss": 0.6236, "step": 60 }, { "epoch": 0.1014832162373146, "grad_norm": 0.7461961507797241, "learning_rate": 2.0312500000000002e-05, "loss": 0.6, "step": 65 }, { "epoch": 0.1092896174863388, "grad_norm": 0.7259904742240906, "learning_rate": 2.1875e-05, "loss": 0.5814, "step": 70 }, { "epoch": 0.117096018735363, "grad_norm": 0.5570379495620728, "learning_rate": 2.34375e-05, "loss": 0.5752, "step": 75 }, { "epoch": 0.1249024199843872, "grad_norm": 0.5848241448402405, "learning_rate": 2.5e-05, "loss": 0.5455, "step": 80 }, { "epoch": 0.1327088212334114, "grad_norm": 0.9087640643119812, "learning_rate": 2.6562500000000002e-05, "loss": 0.549, "step": 85 }, { "epoch": 0.1405152224824356, "grad_norm": 0.6453709602355957, "learning_rate": 2.8125000000000003e-05, "loss": 0.5382, "step": 90 }, { "epoch": 0.1483216237314598, "grad_norm": 0.6244491934776306, "learning_rate": 2.96875e-05, "loss": 0.5416, "step": 95 }, { "epoch": 0.156128024980484, "grad_norm": 0.6923064589500427, "learning_rate": 3.125e-05, "loss": 0.5209, "step": 100 }, { "epoch": 0.16393442622950818, "grad_norm": 0.8899492025375366, "learning_rate": 3.2812500000000005e-05, "loss": 0.5193, "step": 105 }, { "epoch": 0.1717408274785324, "grad_norm": 0.6282894611358643, "learning_rate": 3.4375e-05, "loss": 0.5283, "step": 110 }, { "epoch": 0.1795472287275566, "grad_norm": 0.637545645236969, "learning_rate": 3.59375e-05, "loss": 0.5161, "step": 115 }, { "epoch": 0.1873536299765808, "grad_norm": 0.6604511141777039, "learning_rate": 3.7500000000000003e-05, "loss": 0.5024, "step": 120 }, { "epoch": 0.195160031225605, "grad_norm": 0.5799457430839539, "learning_rate": 3.90625e-05, "loss": 0.5113, "step": 125 }, { "epoch": 0.2029664324746292, "grad_norm": 0.745388388633728, "learning_rate": 4.0625000000000005e-05, "loss": 0.509, "step": 130 }, { "epoch": 0.2107728337236534, "grad_norm": 0.8749244809150696, "learning_rate": 4.21875e-05, "loss": 0.5085, "step": 135 }, { "epoch": 0.2185792349726776, "grad_norm": 0.832904040813446, "learning_rate": 4.375e-05, "loss": 0.4885, "step": 140 }, { "epoch": 0.2263856362217018, "grad_norm": 0.7716369032859802, "learning_rate": 4.5312500000000004e-05, "loss": 0.4886, "step": 145 }, { "epoch": 0.234192037470726, "grad_norm": 0.6549221277236938, "learning_rate": 4.6875e-05, "loss": 0.4879, "step": 150 }, { "epoch": 0.2419984387197502, "grad_norm": 0.5921856164932251, "learning_rate": 4.8437500000000005e-05, "loss": 0.4902, "step": 155 }, { "epoch": 0.2498048399687744, "grad_norm": 0.5582632422447205, "learning_rate": 5e-05, "loss": 0.4954, "step": 160 }, { "epoch": 0.2576112412177986, "grad_norm": 0.5605296492576599, "learning_rate": 5.15625e-05, "loss": 0.4882, "step": 165 }, { "epoch": 0.2654176424668228, "grad_norm": 0.7804284691810608, "learning_rate": 5.3125000000000004e-05, "loss": 0.4751, "step": 170 }, { "epoch": 0.273224043715847, "grad_norm": 0.9444372653961182, "learning_rate": 5.46875e-05, "loss": 0.4777, "step": 175 }, { "epoch": 0.2810304449648712, "grad_norm": 0.5533527135848999, "learning_rate": 5.6250000000000005e-05, "loss": 0.4889, "step": 180 }, { "epoch": 0.2888368462138954, "grad_norm": 0.5322621464729309, "learning_rate": 5.78125e-05, "loss": 0.4837, "step": 185 }, { "epoch": 0.2966432474629196, "grad_norm": 0.5449932217597961, "learning_rate": 5.9375e-05, "loss": 0.4695, "step": 190 }, { "epoch": 0.3044496487119438, "grad_norm": 0.6229299306869507, "learning_rate": 6.0937500000000004e-05, "loss": 0.4891, "step": 195 }, { "epoch": 0.312256049960968, "grad_norm": 0.5832580327987671, "learning_rate": 6.25e-05, "loss": 0.4697, "step": 200 }, { "epoch": 0.3200624512099922, "grad_norm": 0.6701719164848328, "learning_rate": 6.40625e-05, "loss": 0.466, "step": 205 }, { "epoch": 0.32786885245901637, "grad_norm": 1.088781476020813, "learning_rate": 6.562500000000001e-05, "loss": 0.4684, "step": 210 }, { "epoch": 0.3356752537080406, "grad_norm": 0.5555649399757385, "learning_rate": 6.71875e-05, "loss": 0.471, "step": 215 }, { "epoch": 0.3434816549570648, "grad_norm": 0.7747379541397095, "learning_rate": 6.875e-05, "loss": 0.4749, "step": 220 }, { "epoch": 0.351288056206089, "grad_norm": 0.5692268013954163, "learning_rate": 7.031250000000001e-05, "loss": 0.4716, "step": 225 }, { "epoch": 0.3590944574551132, "grad_norm": 0.6532680988311768, "learning_rate": 7.1875e-05, "loss": 0.4658, "step": 230 }, { "epoch": 0.3669008587041374, "grad_norm": 0.5181140303611755, "learning_rate": 7.34375e-05, "loss": 0.4631, "step": 235 }, { "epoch": 0.3747072599531616, "grad_norm": 0.5234228372573853, "learning_rate": 7.500000000000001e-05, "loss": 0.4632, "step": 240 }, { "epoch": 0.3825136612021858, "grad_norm": 0.7775989174842834, "learning_rate": 7.65625e-05, "loss": 0.4767, "step": 245 }, { "epoch": 0.39032006245121, "grad_norm": 0.736663281917572, "learning_rate": 7.8125e-05, "loss": 0.4552, "step": 250 }, { "epoch": 0.3981264637002342, "grad_norm": 0.49683699011802673, "learning_rate": 7.96875e-05, "loss": 0.458, "step": 255 }, { "epoch": 0.4059328649492584, "grad_norm": 0.5580967664718628, "learning_rate": 8.125000000000001e-05, "loss": 0.4553, "step": 260 }, { "epoch": 0.4137392661982826, "grad_norm": 0.5070011019706726, "learning_rate": 8.28125e-05, "loss": 0.4667, "step": 265 }, { "epoch": 0.4215456674473068, "grad_norm": 0.5351033806800842, "learning_rate": 8.4375e-05, "loss": 0.4633, "step": 270 }, { "epoch": 0.42935206869633097, "grad_norm": 0.5223821997642517, "learning_rate": 8.593750000000001e-05, "loss": 0.4614, "step": 275 }, { "epoch": 0.4371584699453552, "grad_norm": 0.5035797953605652, "learning_rate": 8.75e-05, "loss": 0.4577, "step": 280 }, { "epoch": 0.4449648711943794, "grad_norm": 0.5388506054878235, "learning_rate": 8.90625e-05, "loss": 0.4647, "step": 285 }, { "epoch": 0.4527712724434036, "grad_norm": 0.5311102867126465, "learning_rate": 9.062500000000001e-05, "loss": 0.4537, "step": 290 }, { "epoch": 0.4605776736924278, "grad_norm": 0.5614621043205261, "learning_rate": 9.21875e-05, "loss": 0.4635, "step": 295 }, { "epoch": 0.468384074941452, "grad_norm": 0.7148458957672119, "learning_rate": 9.375e-05, "loss": 0.457, "step": 300 }, { "epoch": 0.47619047619047616, "grad_norm": 0.7659621834754944, "learning_rate": 9.53125e-05, "loss": 0.4522, "step": 305 }, { "epoch": 0.4839968774395004, "grad_norm": 0.6256824731826782, "learning_rate": 9.687500000000001e-05, "loss": 0.4547, "step": 310 }, { "epoch": 0.4918032786885246, "grad_norm": 0.5664246082305908, "learning_rate": 9.84375e-05, "loss": 0.456, "step": 315 }, { "epoch": 0.4996096799375488, "grad_norm": 0.6836294531822205, "learning_rate": 0.0001, "loss": 0.4525, "step": 320 }, { "epoch": 0.507416081186573, "grad_norm": 0.5858592391014099, "learning_rate": 0.00010156250000000001, "loss": 0.4439, "step": 325 }, { "epoch": 0.5152224824355972, "grad_norm": 0.5709215402603149, "learning_rate": 0.000103125, "loss": 0.4591, "step": 330 }, { "epoch": 0.5230288836846214, "grad_norm": 0.5698142647743225, "learning_rate": 0.0001046875, "loss": 0.4412, "step": 335 }, { "epoch": 0.5308352849336456, "grad_norm": 0.5309858918190002, "learning_rate": 0.00010625000000000001, "loss": 0.4473, "step": 340 }, { "epoch": 0.5386416861826698, "grad_norm": 0.49240779876708984, "learning_rate": 0.00010781250000000001, "loss": 0.4415, "step": 345 }, { "epoch": 0.546448087431694, "grad_norm": 0.4742411971092224, "learning_rate": 0.000109375, "loss": 0.441, "step": 350 }, { "epoch": 0.5542544886807181, "grad_norm": 0.4116191864013672, "learning_rate": 0.0001109375, "loss": 0.4389, "step": 355 }, { "epoch": 0.5620608899297423, "grad_norm": 0.5430789589881897, "learning_rate": 0.00011250000000000001, "loss": 0.4636, "step": 360 }, { "epoch": 0.5698672911787666, "grad_norm": 0.6789748668670654, "learning_rate": 0.0001140625, "loss": 0.4391, "step": 365 }, { "epoch": 0.5776736924277908, "grad_norm": 0.4417556822299957, "learning_rate": 0.000115625, "loss": 0.4383, "step": 370 }, { "epoch": 0.585480093676815, "grad_norm": 0.4192645847797394, "learning_rate": 0.00011718750000000001, "loss": 0.4392, "step": 375 }, { "epoch": 0.5932864949258392, "grad_norm": 0.42620405554771423, "learning_rate": 0.00011875, "loss": 0.4509, "step": 380 }, { "epoch": 0.6010928961748634, "grad_norm": 0.44085854291915894, "learning_rate": 0.0001203125, "loss": 0.4458, "step": 385 }, { "epoch": 0.6088992974238876, "grad_norm": 0.45056504011154175, "learning_rate": 0.00012187500000000001, "loss": 0.4461, "step": 390 }, { "epoch": 0.6167056986729118, "grad_norm": 0.5633181929588318, "learning_rate": 0.0001234375, "loss": 0.4358, "step": 395 }, { "epoch": 0.624512099921936, "grad_norm": 0.5012897253036499, "learning_rate": 0.000125, "loss": 0.4498, "step": 400 }, { "epoch": 0.6323185011709602, "grad_norm": 0.4812583923339844, "learning_rate": 0.0001265625, "loss": 0.4239, "step": 405 }, { "epoch": 0.6401249024199844, "grad_norm": 0.42709508538246155, "learning_rate": 0.000128125, "loss": 0.452, "step": 410 }, { "epoch": 0.6479313036690086, "grad_norm": 0.5168142914772034, "learning_rate": 0.0001296875, "loss": 0.4365, "step": 415 }, { "epoch": 0.6557377049180327, "grad_norm": 0.4785514771938324, "learning_rate": 0.00013125000000000002, "loss": 0.4356, "step": 420 }, { "epoch": 0.663544106167057, "grad_norm": 0.44413048028945923, "learning_rate": 0.0001328125, "loss": 0.4398, "step": 425 }, { "epoch": 0.6713505074160812, "grad_norm": 0.4412420094013214, "learning_rate": 0.000134375, "loss": 0.4363, "step": 430 }, { "epoch": 0.6791569086651054, "grad_norm": 0.4392535090446472, "learning_rate": 0.00013593750000000002, "loss": 0.4339, "step": 435 }, { "epoch": 0.6869633099141296, "grad_norm": 0.42931196093559265, "learning_rate": 0.0001375, "loss": 0.4279, "step": 440 }, { "epoch": 0.6947697111631538, "grad_norm": 0.41424766182899475, "learning_rate": 0.0001390625, "loss": 0.4313, "step": 445 }, { "epoch": 0.702576112412178, "grad_norm": 0.4159262776374817, "learning_rate": 0.00014062500000000002, "loss": 0.4273, "step": 450 }, { "epoch": 0.7103825136612022, "grad_norm": 0.397058367729187, "learning_rate": 0.0001421875, "loss": 0.4416, "step": 455 }, { "epoch": 0.7181889149102264, "grad_norm": 0.42139485478401184, "learning_rate": 0.00014375, "loss": 0.4348, "step": 460 }, { "epoch": 0.7259953161592506, "grad_norm": 0.47808170318603516, "learning_rate": 0.00014531250000000002, "loss": 0.4244, "step": 465 }, { "epoch": 0.7338017174082748, "grad_norm": 0.4168015122413635, "learning_rate": 0.000146875, "loss": 0.4351, "step": 470 }, { "epoch": 0.741608118657299, "grad_norm": 0.3950536251068115, "learning_rate": 0.0001484375, "loss": 0.4338, "step": 475 }, { "epoch": 0.7494145199063232, "grad_norm": 0.5723301768302917, "learning_rate": 0.00015000000000000001, "loss": 0.4348, "step": 480 }, { "epoch": 0.7572209211553473, "grad_norm": 0.3689775764942169, "learning_rate": 0.0001515625, "loss": 0.4355, "step": 485 }, { "epoch": 0.7650273224043715, "grad_norm": 0.4008382260799408, "learning_rate": 0.000153125, "loss": 0.422, "step": 490 }, { "epoch": 0.7728337236533958, "grad_norm": 0.49685850739479065, "learning_rate": 0.0001546875, "loss": 0.4258, "step": 495 }, { "epoch": 0.78064012490242, "grad_norm": 0.39265674352645874, "learning_rate": 0.00015625, "loss": 0.4334, "step": 500 }, { "epoch": 0.7884465261514442, "grad_norm": 0.6721764206886292, "learning_rate": 0.00015781250000000002, "loss": 0.4283, "step": 505 }, { "epoch": 0.7962529274004684, "grad_norm": 0.43760669231414795, "learning_rate": 0.000159375, "loss": 0.4315, "step": 510 }, { "epoch": 0.8040593286494926, "grad_norm": 0.3985048532485962, "learning_rate": 0.0001609375, "loss": 0.4318, "step": 515 }, { "epoch": 0.8118657298985168, "grad_norm": 0.3855305016040802, "learning_rate": 0.00016250000000000002, "loss": 0.4236, "step": 520 }, { "epoch": 0.819672131147541, "grad_norm": 0.4576222002506256, "learning_rate": 0.0001640625, "loss": 0.4246, "step": 525 }, { "epoch": 0.8274785323965652, "grad_norm": 0.3545086979866028, "learning_rate": 0.000165625, "loss": 0.4248, "step": 530 }, { "epoch": 0.8352849336455894, "grad_norm": 0.360779345035553, "learning_rate": 0.00016718750000000002, "loss": 0.4225, "step": 535 }, { "epoch": 0.8430913348946136, "grad_norm": 0.344586044549942, "learning_rate": 0.00016875, "loss": 0.4231, "step": 540 }, { "epoch": 0.8508977361436377, "grad_norm": 0.4501068890094757, "learning_rate": 0.0001703125, "loss": 0.423, "step": 545 }, { "epoch": 0.8587041373926619, "grad_norm": 0.5521011352539062, "learning_rate": 0.00017187500000000002, "loss": 0.424, "step": 550 }, { "epoch": 0.8665105386416861, "grad_norm": 0.4081909954547882, "learning_rate": 0.0001734375, "loss": 0.4126, "step": 555 }, { "epoch": 0.8743169398907104, "grad_norm": 0.36447080969810486, "learning_rate": 0.000175, "loss": 0.4197, "step": 560 }, { "epoch": 0.8821233411397346, "grad_norm": 0.43303537368774414, "learning_rate": 0.00017656250000000002, "loss": 0.4261, "step": 565 }, { "epoch": 0.8899297423887588, "grad_norm": 0.35191860795021057, "learning_rate": 0.000178125, "loss": 0.4233, "step": 570 }, { "epoch": 0.897736143637783, "grad_norm": 0.33722788095474243, "learning_rate": 0.0001796875, "loss": 0.4149, "step": 575 }, { "epoch": 0.9055425448868072, "grad_norm": 0.36060288548469543, "learning_rate": 0.00018125000000000001, "loss": 0.4244, "step": 580 }, { "epoch": 0.9133489461358314, "grad_norm": 0.353076696395874, "learning_rate": 0.0001828125, "loss": 0.4202, "step": 585 }, { "epoch": 0.9211553473848556, "grad_norm": 0.39187219738960266, "learning_rate": 0.000184375, "loss": 0.4133, "step": 590 }, { "epoch": 0.9289617486338798, "grad_norm": 0.5925287008285522, "learning_rate": 0.0001859375, "loss": 0.4231, "step": 595 }, { "epoch": 0.936768149882904, "grad_norm": 0.5176442861557007, "learning_rate": 0.0001875, "loss": 0.4245, "step": 600 }, { "epoch": 0.9445745511319282, "grad_norm": 0.33046722412109375, "learning_rate": 0.00018906250000000002, "loss": 0.4147, "step": 605 }, { "epoch": 0.9523809523809523, "grad_norm": 0.3665611147880554, "learning_rate": 0.000190625, "loss": 0.4275, "step": 610 }, { "epoch": 0.9601873536299765, "grad_norm": 0.3403601050376892, "learning_rate": 0.0001921875, "loss": 0.4158, "step": 615 }, { "epoch": 0.9679937548790007, "grad_norm": 0.3726797103881836, "learning_rate": 0.00019375000000000002, "loss": 0.4197, "step": 620 }, { "epoch": 0.975800156128025, "grad_norm": 0.38149702548980713, "learning_rate": 0.0001953125, "loss": 0.4102, "step": 625 }, { "epoch": 0.9836065573770492, "grad_norm": 0.3473910987377167, "learning_rate": 0.000196875, "loss": 0.4294, "step": 630 }, { "epoch": 0.9914129586260734, "grad_norm": 0.32832297682762146, "learning_rate": 0.00019843750000000002, "loss": 0.4162, "step": 635 }, { "epoch": 0.9992193598750976, "grad_norm": 0.33780425786972046, "learning_rate": 0.0002, "loss": 0.419, "step": 640 }, { "epoch": 0.9992193598750976, "eval_loss": 1.0502797365188599, "eval_runtime": 0.3387, "eval_samples_per_second": 14.76, "eval_steps_per_second": 2.952, "step": 640 }, { "epoch": 1.0070257611241218, "grad_norm": 0.34341296553611755, "learning_rate": 0.00019999962815279418, "loss": 0.4058, "step": 645 }, { "epoch": 1.014832162373146, "grad_norm": 0.3360665440559387, "learning_rate": 0.00019999851261394218, "loss": 0.4051, "step": 650 }, { "epoch": 1.0226385636221702, "grad_norm": 0.33649712800979614, "learning_rate": 0.00019999665339174013, "loss": 0.4055, "step": 655 }, { "epoch": 1.0304449648711944, "grad_norm": 0.33783483505249023, "learning_rate": 0.00019999405050001497, "loss": 0.4124, "step": 660 }, { "epoch": 1.0382513661202186, "grad_norm": 0.35244283080101013, "learning_rate": 0.0001999907039581243, "loss": 0.4148, "step": 665 }, { "epoch": 1.0460577673692428, "grad_norm": 0.3347136974334717, "learning_rate": 0.0001999866137909562, "loss": 0.4032, "step": 670 }, { "epoch": 1.053864168618267, "grad_norm": 0.34604641795158386, "learning_rate": 0.0001999817800289289, "loss": 0.4093, "step": 675 }, { "epoch": 1.0616705698672912, "grad_norm": 0.36726704239845276, "learning_rate": 0.00019997620270799092, "loss": 0.4022, "step": 680 }, { "epoch": 1.0694769711163155, "grad_norm": 0.410674512386322, "learning_rate": 0.00019996988186962041, "loss": 0.4095, "step": 685 }, { "epoch": 1.0772833723653397, "grad_norm": 0.3476228713989258, "learning_rate": 0.00019996281756082517, "loss": 0.4141, "step": 690 }, { "epoch": 1.0850897736143639, "grad_norm": 0.3787543475627899, "learning_rate": 0.00019995500983414202, "loss": 0.3973, "step": 695 }, { "epoch": 1.092896174863388, "grad_norm": 0.35693004727363586, "learning_rate": 0.00019994645874763658, "loss": 0.4155, "step": 700 }, { "epoch": 1.100702576112412, "grad_norm": 0.3264383375644684, "learning_rate": 0.0001999371643649028, "loss": 0.3979, "step": 705 }, { "epoch": 1.1085089773614363, "grad_norm": 0.32450568675994873, "learning_rate": 0.00019992712675506253, "loss": 0.41, "step": 710 }, { "epoch": 1.1163153786104605, "grad_norm": 0.3869469165802002, "learning_rate": 0.00019991634599276486, "loss": 0.4127, "step": 715 }, { "epoch": 1.1241217798594847, "grad_norm": 0.46077024936676025, "learning_rate": 0.0001999048221581858, "loss": 0.4067, "step": 720 }, { "epoch": 1.131928181108509, "grad_norm": 0.3199751675128937, "learning_rate": 0.00019989255533702736, "loss": 0.4092, "step": 725 }, { "epoch": 1.139734582357533, "grad_norm": 0.341310054063797, "learning_rate": 0.00019987954562051725, "loss": 0.4192, "step": 730 }, { "epoch": 1.1475409836065573, "grad_norm": 0.3507387340068817, "learning_rate": 0.000199865793105408, "loss": 0.4063, "step": 735 }, { "epoch": 1.1553473848555815, "grad_norm": 0.33445993065834045, "learning_rate": 0.00019985129789397633, "loss": 0.4178, "step": 740 }, { "epoch": 1.1631537861046057, "grad_norm": 0.3056560158729553, "learning_rate": 0.00019983606009402224, "loss": 0.395, "step": 745 }, { "epoch": 1.17096018735363, "grad_norm": 0.31608232855796814, "learning_rate": 0.00019982007981886847, "loss": 0.4089, "step": 750 }, { "epoch": 1.1787665886026542, "grad_norm": 0.44849371910095215, "learning_rate": 0.00019980335718735944, "loss": 0.4042, "step": 755 }, { "epoch": 1.1865729898516784, "grad_norm": 0.34841999411582947, "learning_rate": 0.00019978589232386035, "loss": 0.4017, "step": 760 }, { "epoch": 1.1943793911007026, "grad_norm": 0.30834051966667175, "learning_rate": 0.0001997676853582565, "loss": 0.4059, "step": 765 }, { "epoch": 1.2021857923497268, "grad_norm": 0.319698303937912, "learning_rate": 0.000199748736425952, "loss": 0.4133, "step": 770 }, { "epoch": 1.209992193598751, "grad_norm": 0.3290104269981384, "learning_rate": 0.00019972904566786903, "loss": 0.4025, "step": 775 }, { "epoch": 1.2177985948477752, "grad_norm": 0.3340936601161957, "learning_rate": 0.00019970861323044667, "loss": 0.4056, "step": 780 }, { "epoch": 1.2256049960967994, "grad_norm": 0.32487761974334717, "learning_rate": 0.00019968743926563983, "loss": 0.4043, "step": 785 }, { "epoch": 1.2334113973458236, "grad_norm": 0.3178825378417969, "learning_rate": 0.00019966552393091804, "loss": 0.4063, "step": 790 }, { "epoch": 1.2412177985948478, "grad_norm": 0.3003855347633362, "learning_rate": 0.00019964286738926448, "loss": 0.3961, "step": 795 }, { "epoch": 1.249024199843872, "grad_norm": 0.32938238978385925, "learning_rate": 0.00019961946980917456, "loss": 0.3991, "step": 800 }, { "epoch": 1.2568306010928962, "grad_norm": 0.3072170615196228, "learning_rate": 0.0001995953313646548, "loss": 0.3943, "step": 805 }, { "epoch": 1.2646370023419204, "grad_norm": 0.36918073892593384, "learning_rate": 0.0001995704522352214, "loss": 0.3998, "step": 810 }, { "epoch": 1.2724434035909447, "grad_norm": 0.3280356228351593, "learning_rate": 0.00019954483260589911, "loss": 0.4035, "step": 815 }, { "epoch": 1.2802498048399689, "grad_norm": 0.29702678322792053, "learning_rate": 0.0001995184726672197, "loss": 0.3901, "step": 820 }, { "epoch": 1.288056206088993, "grad_norm": 0.30307725071907043, "learning_rate": 0.00019949137261522052, "loss": 0.4053, "step": 825 }, { "epoch": 1.2958626073380173, "grad_norm": 0.31273552775382996, "learning_rate": 0.00019946353265144312, "loss": 0.4136, "step": 830 }, { "epoch": 1.3036690085870415, "grad_norm": 0.30774056911468506, "learning_rate": 0.00019943495298293181, "loss": 0.3995, "step": 835 }, { "epoch": 1.3114754098360657, "grad_norm": 0.3068085312843323, "learning_rate": 0.00019940563382223197, "loss": 0.4136, "step": 840 }, { "epoch": 1.31928181108509, "grad_norm": 0.33068013191223145, "learning_rate": 0.00019937557538738854, "loss": 0.3957, "step": 845 }, { "epoch": 1.327088212334114, "grad_norm": 0.36017951369285583, "learning_rate": 0.00019934477790194445, "loss": 0.4094, "step": 850 }, { "epoch": 1.334894613583138, "grad_norm": 0.3393058776855469, "learning_rate": 0.00019931324159493886, "loss": 0.4072, "step": 855 }, { "epoch": 1.3427010148321623, "grad_norm": 0.29926055669784546, "learning_rate": 0.0001992809667009055, "loss": 0.3991, "step": 860 }, { "epoch": 1.3505074160811865, "grad_norm": 0.3126608431339264, "learning_rate": 0.000199247953459871, "loss": 0.3997, "step": 865 }, { "epoch": 1.3583138173302107, "grad_norm": 0.31976112723350525, "learning_rate": 0.00019921420211735299, "loss": 0.4043, "step": 870 }, { "epoch": 1.366120218579235, "grad_norm": 0.37549304962158203, "learning_rate": 0.00019917971292435826, "loss": 0.4112, "step": 875 }, { "epoch": 1.3739266198282591, "grad_norm": 0.3190479278564453, "learning_rate": 0.00019914448613738106, "loss": 0.402, "step": 880 }, { "epoch": 1.3817330210772834, "grad_norm": 0.30075913667678833, "learning_rate": 0.00019910852201840103, "loss": 0.4159, "step": 885 }, { "epoch": 1.3895394223263076, "grad_norm": 0.3074491024017334, "learning_rate": 0.00019907182083488129, "loss": 0.4045, "step": 890 }, { "epoch": 1.3973458235753318, "grad_norm": 0.3855604827404022, "learning_rate": 0.0001990343828597665, "loss": 0.4045, "step": 895 }, { "epoch": 1.405152224824356, "grad_norm": 0.3558739721775055, "learning_rate": 0.00019899620837148077, "loss": 0.3969, "step": 900 }, { "epoch": 1.4129586260733802, "grad_norm": 0.3128485083580017, "learning_rate": 0.00019895729765392572, "loss": 0.403, "step": 905 }, { "epoch": 1.4207650273224044, "grad_norm": 0.30848538875579834, "learning_rate": 0.0001989176509964781, "loss": 0.398, "step": 910 }, { "epoch": 1.4285714285714286, "grad_norm": 0.2908243238925934, "learning_rate": 0.00019887726869398793, "loss": 0.394, "step": 915 }, { "epoch": 1.4363778298204528, "grad_norm": 0.29241564869880676, "learning_rate": 0.00019883615104677608, "loss": 0.4007, "step": 920 }, { "epoch": 1.444184231069477, "grad_norm": 0.29839983582496643, "learning_rate": 0.00019879429836063226, "loss": 0.3961, "step": 925 }, { "epoch": 1.4519906323185012, "grad_norm": 0.298441618680954, "learning_rate": 0.00019875171094681248, "loss": 0.4003, "step": 930 }, { "epoch": 1.4597970335675254, "grad_norm": 0.30495795607566833, "learning_rate": 0.00019870838912203705, "loss": 0.4036, "step": 935 }, { "epoch": 1.4676034348165496, "grad_norm": 0.2891780436038971, "learning_rate": 0.0001986643332084879, "loss": 0.3901, "step": 940 }, { "epoch": 1.4754098360655736, "grad_norm": 0.2930201292037964, "learning_rate": 0.00019861954353380643, "loss": 0.3943, "step": 945 }, { "epoch": 1.4832162373145978, "grad_norm": 0.289411336183548, "learning_rate": 0.0001985740204310909, "loss": 0.4003, "step": 950 }, { "epoch": 1.491022638563622, "grad_norm": 0.3018743693828583, "learning_rate": 0.0001985277642388941, "loss": 0.3949, "step": 955 }, { "epoch": 1.4988290398126463, "grad_norm": 0.3346049189567566, "learning_rate": 0.00019848077530122083, "loss": 0.3949, "step": 960 }, { "epoch": 1.5066354410616705, "grad_norm": 0.35057857632637024, "learning_rate": 0.00019843305396752509, "loss": 0.3916, "step": 965 }, { "epoch": 1.5144418423106947, "grad_norm": 0.33775442838668823, "learning_rate": 0.00019838460059270775, "loss": 0.4003, "step": 970 }, { "epoch": 1.5222482435597189, "grad_norm": 0.3016526997089386, "learning_rate": 0.00019833541553711395, "loss": 0.4032, "step": 975 }, { "epoch": 1.530054644808743, "grad_norm": 0.2809526026248932, "learning_rate": 0.0001982854991665301, "loss": 0.3983, "step": 980 }, { "epoch": 1.5378610460577673, "grad_norm": 0.3081335723400116, "learning_rate": 0.00019823485185218158, "loss": 0.3997, "step": 985 }, { "epoch": 1.5456674473067915, "grad_norm": 0.28975608944892883, "learning_rate": 0.00019818347397072955, "loss": 0.3943, "step": 990 }, { "epoch": 1.5534738485558157, "grad_norm": 0.32257601618766785, "learning_rate": 0.00019813136590426844, "loss": 0.3948, "step": 995 }, { "epoch": 1.56128024980484, "grad_norm": 0.3275687098503113, "learning_rate": 0.00019807852804032305, "loss": 0.4001, "step": 1000 }, { "epoch": 1.5690866510538641, "grad_norm": 0.3463219404220581, "learning_rate": 0.0001980249607718456, "loss": 0.3992, "step": 1005 }, { "epoch": 1.5768930523028883, "grad_norm": 0.3080371022224426, "learning_rate": 0.00019797066449721295, "loss": 0.4023, "step": 1010 }, { "epoch": 1.5846994535519126, "grad_norm": 0.29140976071357727, "learning_rate": 0.00019791563962022338, "loss": 0.4036, "step": 1015 }, { "epoch": 1.5925058548009368, "grad_norm": 0.3033461570739746, "learning_rate": 0.00019785988655009385, "loss": 0.4051, "step": 1020 }, { "epoch": 1.600312256049961, "grad_norm": 0.27476766705513, "learning_rate": 0.0001978034057014568, "loss": 0.3818, "step": 1025 }, { "epoch": 1.6081186572989852, "grad_norm": 0.3026203513145447, "learning_rate": 0.0001977461974943572, "loss": 0.3959, "step": 1030 }, { "epoch": 1.6159250585480094, "grad_norm": 0.28781238198280334, "learning_rate": 0.00019768826235424926, "loss": 0.4067, "step": 1035 }, { "epoch": 1.6237314597970336, "grad_norm": 0.28916335105895996, "learning_rate": 0.00019762960071199333, "loss": 0.3999, "step": 1040 }, { "epoch": 1.6315378610460578, "grad_norm": 0.2844083607196808, "learning_rate": 0.00019757021300385286, "loss": 0.3958, "step": 1045 }, { "epoch": 1.639344262295082, "grad_norm": 0.30581745505332947, "learning_rate": 0.00019751009967149087, "loss": 0.4016, "step": 1050 }, { "epoch": 1.6471506635441062, "grad_norm": 0.27910909056663513, "learning_rate": 0.00019744926116196685, "loss": 0.3903, "step": 1055 }, { "epoch": 1.6549570647931304, "grad_norm": 0.2814294695854187, "learning_rate": 0.00019738769792773336, "loss": 0.3896, "step": 1060 }, { "epoch": 1.6627634660421546, "grad_norm": 0.29259413480758667, "learning_rate": 0.00019732541042663282, "loss": 0.4039, "step": 1065 }, { "epoch": 1.6705698672911788, "grad_norm": 0.2868483364582062, "learning_rate": 0.00019726239912189382, "loss": 0.3893, "step": 1070 }, { "epoch": 1.678376268540203, "grad_norm": 0.30284589529037476, "learning_rate": 0.00019719866448212795, "loss": 0.3973, "step": 1075 }, { "epoch": 1.6861826697892273, "grad_norm": 0.29238229990005493, "learning_rate": 0.00019713420698132614, "loss": 0.404, "step": 1080 }, { "epoch": 1.6939890710382515, "grad_norm": 0.2957955300807953, "learning_rate": 0.00019706902709885524, "loss": 0.3929, "step": 1085 }, { "epoch": 1.7017954722872757, "grad_norm": 0.29362279176712036, "learning_rate": 0.00019700312531945442, "loss": 0.393, "step": 1090 }, { "epoch": 1.7096018735362999, "grad_norm": 0.3154332637786865, "learning_rate": 0.00019693650213323144, "loss": 0.3971, "step": 1095 }, { "epoch": 1.717408274785324, "grad_norm": 0.31909048557281494, "learning_rate": 0.00019686915803565934, "loss": 0.3991, "step": 1100 }, { "epoch": 1.7252146760343483, "grad_norm": 0.31985750794410706, "learning_rate": 0.00019680109352757227, "loss": 0.4076, "step": 1105 }, { "epoch": 1.7330210772833725, "grad_norm": 0.2867513597011566, "learning_rate": 0.00019673230911516226, "loss": 0.3957, "step": 1110 }, { "epoch": 1.7408274785323967, "grad_norm": 0.30128851532936096, "learning_rate": 0.00019666280530997518, "loss": 0.4041, "step": 1115 }, { "epoch": 1.748633879781421, "grad_norm": 0.2900910973548889, "learning_rate": 0.00019659258262890683, "loss": 0.3898, "step": 1120 }, { "epoch": 1.756440281030445, "grad_norm": 0.3005325496196747, "learning_rate": 0.00019652164159419946, "loss": 0.4086, "step": 1125 }, { "epoch": 1.7642466822794691, "grad_norm": 0.2881026268005371, "learning_rate": 0.00019644998273343753, "loss": 0.4005, "step": 1130 }, { "epoch": 1.7720530835284933, "grad_norm": 0.2854783535003662, "learning_rate": 0.000196377606579544, "loss": 0.3942, "step": 1135 }, { "epoch": 1.7798594847775175, "grad_norm": 0.2979474365711212, "learning_rate": 0.00019630451367077628, "loss": 0.3935, "step": 1140 }, { "epoch": 1.7876658860265418, "grad_norm": 0.28160732984542847, "learning_rate": 0.00019623070455072224, "loss": 0.4153, "step": 1145 }, { "epoch": 1.795472287275566, "grad_norm": 0.27550724148750305, "learning_rate": 0.0001961561797682962, "loss": 0.3921, "step": 1150 }, { "epoch": 1.8032786885245902, "grad_norm": 0.27477726340293884, "learning_rate": 0.00019608093987773478, "loss": 0.4017, "step": 1155 }, { "epoch": 1.8110850897736144, "grad_norm": 0.2824558913707733, "learning_rate": 0.0001960049854385929, "loss": 0.3897, "step": 1160 }, { "epoch": 1.8188914910226386, "grad_norm": 0.2761404514312744, "learning_rate": 0.00019592831701573937, "loss": 0.4095, "step": 1165 }, { "epoch": 1.8266978922716628, "grad_norm": 0.28211623430252075, "learning_rate": 0.00019585093517935305, "loss": 0.3912, "step": 1170 }, { "epoch": 1.834504293520687, "grad_norm": 0.27295398712158203, "learning_rate": 0.0001957728405049183, "loss": 0.4125, "step": 1175 }, { "epoch": 1.8423106947697112, "grad_norm": 0.2896745800971985, "learning_rate": 0.0001956940335732209, "loss": 0.3927, "step": 1180 }, { "epoch": 1.8501170960187352, "grad_norm": 0.3007790446281433, "learning_rate": 0.00019561451497034354, "loss": 0.3932, "step": 1185 }, { "epoch": 1.8579234972677594, "grad_norm": 0.29394078254699707, "learning_rate": 0.00019553428528766163, "loss": 0.3905, "step": 1190 }, { "epoch": 1.8657298985167836, "grad_norm": 0.2767265737056732, "learning_rate": 0.00019545334512183885, "loss": 0.395, "step": 1195 }, { "epoch": 1.8735362997658078, "grad_norm": 0.26828575134277344, "learning_rate": 0.0001953716950748227, "loss": 0.3996, "step": 1200 }, { "epoch": 1.881342701014832, "grad_norm": 0.2718822658061981, "learning_rate": 0.00019528933575384, "loss": 0.3803, "step": 1205 }, { "epoch": 1.8891491022638562, "grad_norm": 0.2866761386394501, "learning_rate": 0.00019520626777139245, "loss": 0.4056, "step": 1210 }, { "epoch": 1.8969555035128804, "grad_norm": 0.3028269410133362, "learning_rate": 0.00019512249174525193, "loss": 0.4, "step": 1215 }, { "epoch": 1.9047619047619047, "grad_norm": 0.2755603492259979, "learning_rate": 0.0001950380082984561, "loss": 0.3829, "step": 1220 }, { "epoch": 1.9125683060109289, "grad_norm": 0.28492555022239685, "learning_rate": 0.00019495281805930367, "loss": 0.3839, "step": 1225 }, { "epoch": 1.920374707259953, "grad_norm": 0.2751103639602661, "learning_rate": 0.00019486692166134964, "loss": 0.3907, "step": 1230 }, { "epoch": 1.9281811085089773, "grad_norm": 0.2826032042503357, "learning_rate": 0.00019478031974340073, "loss": 0.3884, "step": 1235 }, { "epoch": 1.9359875097580015, "grad_norm": 0.28445425629615784, "learning_rate": 0.0001946930129495106, "loss": 0.3956, "step": 1240 }, { "epoch": 1.9437939110070257, "grad_norm": 0.2712079584598541, "learning_rate": 0.0001946050019289749, "loss": 0.3893, "step": 1245 }, { "epoch": 1.95160031225605, "grad_norm": 0.27717921137809753, "learning_rate": 0.0001945162873363268, "loss": 0.3912, "step": 1250 }, { "epoch": 1.9594067135050741, "grad_norm": 0.2907007336616516, "learning_rate": 0.00019442686983133168, "loss": 0.4004, "step": 1255 }, { "epoch": 1.9672131147540983, "grad_norm": 0.2781454026699066, "learning_rate": 0.00019433675007898255, "loss": 0.4037, "step": 1260 }, { "epoch": 1.9750195160031225, "grad_norm": 0.27712446451187134, "learning_rate": 0.00019424592874949496, "loss": 0.3902, "step": 1265 }, { "epoch": 1.9828259172521467, "grad_norm": 0.2906958758831024, "learning_rate": 0.00019415440651830208, "loss": 0.3955, "step": 1270 }, { "epoch": 1.990632318501171, "grad_norm": 0.28120070695877075, "learning_rate": 0.00019406218406604965, "loss": 0.3878, "step": 1275 }, { "epoch": 1.9984387197501952, "grad_norm": 0.2804274260997772, "learning_rate": 0.00019396926207859084, "loss": 0.3829, "step": 1280 }, { "epoch": 2.0, "eval_loss": 1.0417580604553223, "eval_runtime": 0.3233, "eval_samples_per_second": 15.465, "eval_steps_per_second": 3.093, "step": 1281 }, { "epoch": 2.0062451209992194, "grad_norm": 0.2731517255306244, "learning_rate": 0.00019387564124698133, "loss": 0.3747, "step": 1285 }, { "epoch": 2.0140515222482436, "grad_norm": 0.2837215065956116, "learning_rate": 0.00019378132226747398, "loss": 0.3651, "step": 1290 }, { "epoch": 2.021857923497268, "grad_norm": 0.29142889380455017, "learning_rate": 0.0001936863058415138, "loss": 0.3736, "step": 1295 }, { "epoch": 2.029664324746292, "grad_norm": 0.29420003294944763, "learning_rate": 0.0001935905926757326, "loss": 0.3764, "step": 1300 }, { "epoch": 2.037470725995316, "grad_norm": 0.2812630832195282, "learning_rate": 0.00019349418348194383, "loss": 0.3674, "step": 1305 }, { "epoch": 2.0452771272443404, "grad_norm": 0.3169541656970978, "learning_rate": 0.00019339707897713738, "loss": 0.3625, "step": 1310 }, { "epoch": 2.0530835284933646, "grad_norm": 0.31742027401924133, "learning_rate": 0.0001932992798834739, "loss": 0.3718, "step": 1315 }, { "epoch": 2.060889929742389, "grad_norm": 0.2889857888221741, "learning_rate": 0.00019320078692827987, "loss": 0.3798, "step": 1320 }, { "epoch": 2.068696330991413, "grad_norm": 0.27867743372917175, "learning_rate": 0.00019310160084404186, "loss": 0.3647, "step": 1325 }, { "epoch": 2.0765027322404372, "grad_norm": 0.29241612553596497, "learning_rate": 0.00019300172236840122, "loss": 0.3761, "step": 1330 }, { "epoch": 2.0843091334894615, "grad_norm": 0.28482168912887573, "learning_rate": 0.00019290115224414862, "loss": 0.3743, "step": 1335 }, { "epoch": 2.0921155347384857, "grad_norm": 0.3034825325012207, "learning_rate": 0.00019279989121921847, "loss": 0.3757, "step": 1340 }, { "epoch": 2.09992193598751, "grad_norm": 0.3016110956668854, "learning_rate": 0.0001926979400466833, "loss": 0.3774, "step": 1345 }, { "epoch": 2.107728337236534, "grad_norm": 0.3048888146877289, "learning_rate": 0.00019259529948474833, "loss": 0.369, "step": 1350 }, { "epoch": 2.1155347384855583, "grad_norm": 0.2792314291000366, "learning_rate": 0.00019249197029674566, "loss": 0.3727, "step": 1355 }, { "epoch": 2.1233411397345825, "grad_norm": 0.30155235528945923, "learning_rate": 0.0001923879532511287, "loss": 0.3691, "step": 1360 }, { "epoch": 2.1311475409836067, "grad_norm": 0.2807096838951111, "learning_rate": 0.0001922832491214664, "loss": 0.3618, "step": 1365 }, { "epoch": 2.138953942232631, "grad_norm": 0.2897432744503021, "learning_rate": 0.00019217785868643747, "loss": 0.3644, "step": 1370 }, { "epoch": 2.146760343481655, "grad_norm": 0.2933688759803772, "learning_rate": 0.0001920717827298248, "loss": 0.3753, "step": 1375 }, { "epoch": 2.1545667447306793, "grad_norm": 0.2933692932128906, "learning_rate": 0.00019196502204050922, "loss": 0.3725, "step": 1380 }, { "epoch": 2.1623731459797035, "grad_norm": 0.28731822967529297, "learning_rate": 0.00019185757741246412, "loss": 0.3733, "step": 1385 }, { "epoch": 2.1701795472287277, "grad_norm": 0.3193456530570984, "learning_rate": 0.00019174944964474912, "loss": 0.3741, "step": 1390 }, { "epoch": 2.177985948477752, "grad_norm": 0.2853163480758667, "learning_rate": 0.00019164063954150447, "loss": 0.3774, "step": 1395 }, { "epoch": 2.185792349726776, "grad_norm": 0.3006320893764496, "learning_rate": 0.00019153114791194473, "loss": 0.3677, "step": 1400 }, { "epoch": 2.1935987509758004, "grad_norm": 0.2752683758735657, "learning_rate": 0.00019142097557035308, "loss": 0.3675, "step": 1405 }, { "epoch": 2.201405152224824, "grad_norm": 0.2900081276893616, "learning_rate": 0.00019131012333607506, "loss": 0.375, "step": 1410 }, { "epoch": 2.209211553473849, "grad_norm": 0.3021766245365143, "learning_rate": 0.00019119859203351255, "loss": 0.3778, "step": 1415 }, { "epoch": 2.2170179547228726, "grad_norm": 0.2842157185077667, "learning_rate": 0.00019108638249211758, "loss": 0.3668, "step": 1420 }, { "epoch": 2.2248243559718968, "grad_norm": 0.27962204813957214, "learning_rate": 0.0001909734955463863, "loss": 0.3731, "step": 1425 }, { "epoch": 2.232630757220921, "grad_norm": 0.29117563366889954, "learning_rate": 0.0001908599320358526, "loss": 0.3726, "step": 1430 }, { "epoch": 2.240437158469945, "grad_norm": 0.307558536529541, "learning_rate": 0.0001907456928050819, "loss": 0.37, "step": 1435 }, { "epoch": 2.2482435597189694, "grad_norm": 0.3160630762577057, "learning_rate": 0.000190630778703665, "loss": 0.3769, "step": 1440 }, { "epoch": 2.2560499609679936, "grad_norm": 0.2901064157485962, "learning_rate": 0.0001905151905862117, "loss": 0.3703, "step": 1445 }, { "epoch": 2.263856362217018, "grad_norm": 0.2909891605377197, "learning_rate": 0.00019039892931234435, "loss": 0.374, "step": 1450 }, { "epoch": 2.271662763466042, "grad_norm": 0.3095717132091522, "learning_rate": 0.00019028199574669153, "loss": 0.3848, "step": 1455 }, { "epoch": 2.279469164715066, "grad_norm": 0.2892928421497345, "learning_rate": 0.00019016439075888162, "loss": 0.364, "step": 1460 }, { "epoch": 2.2872755659640904, "grad_norm": 0.3012056350708008, "learning_rate": 0.00019004611522353636, "loss": 0.3704, "step": 1465 }, { "epoch": 2.2950819672131146, "grad_norm": 0.28810742497444153, "learning_rate": 0.00018992717002026433, "loss": 0.3757, "step": 1470 }, { "epoch": 2.302888368462139, "grad_norm": 0.3010272979736328, "learning_rate": 0.0001898075560336543, "loss": 0.3695, "step": 1475 }, { "epoch": 2.310694769711163, "grad_norm": 0.2966187000274658, "learning_rate": 0.00018968727415326884, "loss": 0.3753, "step": 1480 }, { "epoch": 2.3185011709601873, "grad_norm": 0.303035169839859, "learning_rate": 0.00018956632527363757, "loss": 0.3844, "step": 1485 }, { "epoch": 2.3263075722092115, "grad_norm": 0.2904973030090332, "learning_rate": 0.00018944471029425054, "loss": 0.3733, "step": 1490 }, { "epoch": 2.3341139734582357, "grad_norm": 0.2944584786891937, "learning_rate": 0.00018932243011955154, "loss": 0.3705, "step": 1495 }, { "epoch": 2.34192037470726, "grad_norm": 0.3047090470790863, "learning_rate": 0.00018919948565893142, "loss": 0.3776, "step": 1500 }, { "epoch": 2.349726775956284, "grad_norm": 0.28889063000679016, "learning_rate": 0.00018907587782672128, "loss": 0.3748, "step": 1505 }, { "epoch": 2.3575331772053083, "grad_norm": 0.29488006234169006, "learning_rate": 0.00018895160754218561, "loss": 0.3719, "step": 1510 }, { "epoch": 2.3653395784543325, "grad_norm": 0.28277385234832764, "learning_rate": 0.00018882667572951562, "loss": 0.3737, "step": 1515 }, { "epoch": 2.3731459797033567, "grad_norm": 0.2788912057876587, "learning_rate": 0.00018870108331782217, "loss": 0.37, "step": 1520 }, { "epoch": 2.380952380952381, "grad_norm": 0.2984495759010315, "learning_rate": 0.00018857483124112907, "loss": 0.3816, "step": 1525 }, { "epoch": 2.388758782201405, "grad_norm": 0.28799429535865784, "learning_rate": 0.00018844792043836589, "loss": 0.3849, "step": 1530 }, { "epoch": 2.3965651834504293, "grad_norm": 0.2892215847969055, "learning_rate": 0.00018832035185336124, "loss": 0.372, "step": 1535 }, { "epoch": 2.4043715846994536, "grad_norm": 0.30197983980178833, "learning_rate": 0.0001881921264348355, "loss": 0.3764, "step": 1540 }, { "epoch": 2.4121779859484778, "grad_norm": 0.3232642412185669, "learning_rate": 0.000188063245136394, "loss": 0.3788, "step": 1545 }, { "epoch": 2.419984387197502, "grad_norm": 0.29107898473739624, "learning_rate": 0.00018793370891651972, "loss": 0.3786, "step": 1550 }, { "epoch": 2.427790788446526, "grad_norm": 0.29140210151672363, "learning_rate": 0.00018780351873856627, "loss": 0.3763, "step": 1555 }, { "epoch": 2.4355971896955504, "grad_norm": 0.30477312207221985, "learning_rate": 0.0001876726755707508, "loss": 0.3866, "step": 1560 }, { "epoch": 2.4434035909445746, "grad_norm": 0.3051372170448303, "learning_rate": 0.00018754118038614653, "loss": 0.378, "step": 1565 }, { "epoch": 2.451209992193599, "grad_norm": 0.2856997847557068, "learning_rate": 0.0001874090341626759, "loss": 0.3744, "step": 1570 }, { "epoch": 2.459016393442623, "grad_norm": 0.26825210452079773, "learning_rate": 0.00018727623788310292, "loss": 0.369, "step": 1575 }, { "epoch": 2.4668227946916472, "grad_norm": 0.2855939269065857, "learning_rate": 0.00018714279253502616, "loss": 0.3829, "step": 1580 }, { "epoch": 2.4746291959406714, "grad_norm": 0.28602635860443115, "learning_rate": 0.00018700869911087115, "loss": 0.3657, "step": 1585 }, { "epoch": 2.4824355971896956, "grad_norm": 0.30138617753982544, "learning_rate": 0.00018687395860788323, "loss": 0.3694, "step": 1590 }, { "epoch": 2.49024199843872, "grad_norm": 0.2945271134376526, "learning_rate": 0.00018673857202812, "loss": 0.3663, "step": 1595 }, { "epoch": 2.498048399687744, "grad_norm": 0.28233060240745544, "learning_rate": 0.00018660254037844388, "loss": 0.3839, "step": 1600 }, { "epoch": 2.5058548009367683, "grad_norm": 0.30950790643692017, "learning_rate": 0.00018646586467051463, "loss": 0.3864, "step": 1605 }, { "epoch": 2.5136612021857925, "grad_norm": 0.30938637256622314, "learning_rate": 0.00018632854592078184, "loss": 0.3703, "step": 1610 }, { "epoch": 2.5214676034348167, "grad_norm": 0.2714425325393677, "learning_rate": 0.00018619058515047745, "loss": 0.3748, "step": 1615 }, { "epoch": 2.529274004683841, "grad_norm": 0.30516138672828674, "learning_rate": 0.0001860519833856079, "loss": 0.3868, "step": 1620 }, { "epoch": 2.537080405932865, "grad_norm": 0.30349424481391907, "learning_rate": 0.00018591274165694687, "loss": 0.3682, "step": 1625 }, { "epoch": 2.5448868071818893, "grad_norm": 0.29412826895713806, "learning_rate": 0.00018577286100002723, "loss": 0.3683, "step": 1630 }, { "epoch": 2.552693208430913, "grad_norm": 0.2976747155189514, "learning_rate": 0.00018563234245513364, "loss": 0.3697, "step": 1635 }, { "epoch": 2.5604996096799377, "grad_norm": 0.28876587748527527, "learning_rate": 0.00018549118706729468, "loss": 0.3751, "step": 1640 }, { "epoch": 2.5683060109289615, "grad_norm": 0.2986242175102234, "learning_rate": 0.00018534939588627503, "loss": 0.3772, "step": 1645 }, { "epoch": 2.576112412177986, "grad_norm": 0.29054194688796997, "learning_rate": 0.00018520696996656788, "loss": 0.388, "step": 1650 }, { "epoch": 2.58391881342701, "grad_norm": 0.2982032001018524, "learning_rate": 0.00018506391036738674, "loss": 0.3705, "step": 1655 }, { "epoch": 2.5917252146760346, "grad_norm": 0.2838602364063263, "learning_rate": 0.0001849202181526579, "loss": 0.3705, "step": 1660 }, { "epoch": 2.5995316159250583, "grad_norm": 0.28826209902763367, "learning_rate": 0.00018477589439101232, "loss": 0.3698, "step": 1665 }, { "epoch": 2.607338017174083, "grad_norm": 0.2916529178619385, "learning_rate": 0.00018463094015577775, "loss": 0.3711, "step": 1670 }, { "epoch": 2.6151444184231067, "grad_norm": 0.2822958528995514, "learning_rate": 0.00018448535652497073, "loss": 0.3773, "step": 1675 }, { "epoch": 2.6229508196721314, "grad_norm": 0.2998236119747162, "learning_rate": 0.0001843391445812886, "loss": 0.3769, "step": 1680 }, { "epoch": 2.630757220921155, "grad_norm": 0.2773909866809845, "learning_rate": 0.00018419230541210139, "loss": 0.3656, "step": 1685 }, { "epoch": 2.63856362217018, "grad_norm": 0.2902582585811615, "learning_rate": 0.0001840448401094438, "loss": 0.3811, "step": 1690 }, { "epoch": 2.6463700234192036, "grad_norm": 0.28828758001327515, "learning_rate": 0.00018389674977000708, "loss": 0.3718, "step": 1695 }, { "epoch": 2.654176424668228, "grad_norm": 0.31182461977005005, "learning_rate": 0.0001837480354951308, "loss": 0.3719, "step": 1700 }, { "epoch": 2.661982825917252, "grad_norm": 0.2795233130455017, "learning_rate": 0.0001835986983907947, "loss": 0.3635, "step": 1705 }, { "epoch": 2.669789227166276, "grad_norm": 0.28305554389953613, "learning_rate": 0.00018344873956761045, "loss": 0.3659, "step": 1710 }, { "epoch": 2.6775956284153004, "grad_norm": 0.27487748861312866, "learning_rate": 0.00018329816014081346, "loss": 0.3738, "step": 1715 }, { "epoch": 2.6854020296643246, "grad_norm": 0.2843542695045471, "learning_rate": 0.00018314696123025454, "loss": 0.3729, "step": 1720 }, { "epoch": 2.693208430913349, "grad_norm": 0.28278475999832153, "learning_rate": 0.0001829951439603915, "loss": 0.3675, "step": 1725 }, { "epoch": 2.701014832162373, "grad_norm": 0.28049561381340027, "learning_rate": 0.00018284270946028092, "loss": 0.3722, "step": 1730 }, { "epoch": 2.7088212334113972, "grad_norm": 0.2751220464706421, "learning_rate": 0.00018268965886356964, "loss": 0.3708, "step": 1735 }, { "epoch": 2.7166276346604215, "grad_norm": 0.2924399971961975, "learning_rate": 0.00018253599330848638, "loss": 0.3728, "step": 1740 }, { "epoch": 2.7244340359094457, "grad_norm": 0.27712133526802063, "learning_rate": 0.00018238171393783335, "loss": 0.3716, "step": 1745 }, { "epoch": 2.73224043715847, "grad_norm": 0.28451547026634216, "learning_rate": 0.00018222682189897752, "loss": 0.3826, "step": 1750 }, { "epoch": 2.740046838407494, "grad_norm": 0.275689035654068, "learning_rate": 0.0001820713183438424, "loss": 0.3809, "step": 1755 }, { "epoch": 2.7478532396565183, "grad_norm": 0.2970561683177948, "learning_rate": 0.0001819152044288992, "loss": 0.3807, "step": 1760 }, { "epoch": 2.7556596409055425, "grad_norm": 0.27083873748779297, "learning_rate": 0.00018175848131515837, "loss": 0.3777, "step": 1765 }, { "epoch": 2.7634660421545667, "grad_norm": 0.2864970266819, "learning_rate": 0.000181601150168161, "loss": 0.3697, "step": 1770 }, { "epoch": 2.771272443403591, "grad_norm": 0.2779759466648102, "learning_rate": 0.00018144321215797, "loss": 0.371, "step": 1775 }, { "epoch": 2.779078844652615, "grad_norm": 0.2953694462776184, "learning_rate": 0.00018128466845916154, "loss": 0.3782, "step": 1780 }, { "epoch": 2.7868852459016393, "grad_norm": 0.27182242274284363, "learning_rate": 0.00018112552025081625, "loss": 0.3688, "step": 1785 }, { "epoch": 2.7946916471506635, "grad_norm": 0.28873521089553833, "learning_rate": 0.00018096576871651042, "loss": 0.3703, "step": 1790 }, { "epoch": 2.8024980483996877, "grad_norm": 0.2740427851676941, "learning_rate": 0.00018080541504430733, "loss": 0.3751, "step": 1795 }, { "epoch": 2.810304449648712, "grad_norm": 0.2902701199054718, "learning_rate": 0.00018064446042674828, "loss": 0.3691, "step": 1800 }, { "epoch": 2.818110850897736, "grad_norm": 0.28125742077827454, "learning_rate": 0.00018048290606084371, "loss": 0.3767, "step": 1805 }, { "epoch": 2.8259172521467604, "grad_norm": 0.28101617097854614, "learning_rate": 0.00018032075314806448, "loss": 0.3767, "step": 1810 }, { "epoch": 2.8337236533957846, "grad_norm": 0.31095340847969055, "learning_rate": 0.00018015800289433273, "loss": 0.3841, "step": 1815 }, { "epoch": 2.841530054644809, "grad_norm": 0.26109251379966736, "learning_rate": 0.00017999465651001296, "loss": 0.3821, "step": 1820 }, { "epoch": 2.849336455893833, "grad_norm": 0.293575257062912, "learning_rate": 0.00017983071520990315, "loss": 0.3809, "step": 1825 }, { "epoch": 2.857142857142857, "grad_norm": 0.26493799686431885, "learning_rate": 0.00017966618021322557, "loss": 0.3752, "step": 1830 }, { "epoch": 2.8649492583918814, "grad_norm": 0.2778892517089844, "learning_rate": 0.00017950105274361776, "loss": 0.369, "step": 1835 }, { "epoch": 2.8727556596409056, "grad_norm": 0.28492116928100586, "learning_rate": 0.00017933533402912354, "loss": 0.3714, "step": 1840 }, { "epoch": 2.88056206088993, "grad_norm": 0.281095415353775, "learning_rate": 0.00017916902530218368, "loss": 0.3727, "step": 1845 }, { "epoch": 2.888368462138954, "grad_norm": 0.277915358543396, "learning_rate": 0.0001790021277996269, "loss": 0.3728, "step": 1850 }, { "epoch": 2.8961748633879782, "grad_norm": 0.27478280663490295, "learning_rate": 0.00017883464276266064, "loss": 0.3799, "step": 1855 }, { "epoch": 2.9039812646370025, "grad_norm": 0.28556427359580994, "learning_rate": 0.00017866657143686168, "loss": 0.3762, "step": 1860 }, { "epoch": 2.9117876658860267, "grad_norm": 0.2784564197063446, "learning_rate": 0.0001784979150721672, "loss": 0.3772, "step": 1865 }, { "epoch": 2.919594067135051, "grad_norm": 0.3048378825187683, "learning_rate": 0.00017832867492286505, "loss": 0.3691, "step": 1870 }, { "epoch": 2.927400468384075, "grad_norm": 0.2872634828090668, "learning_rate": 0.0001781588522475848, "loss": 0.3686, "step": 1875 }, { "epoch": 2.9352068696330993, "grad_norm": 0.28422531485557556, "learning_rate": 0.00017798844830928817, "loss": 0.3814, "step": 1880 }, { "epoch": 2.9430132708821235, "grad_norm": 0.30014801025390625, "learning_rate": 0.0001778174643752598, "loss": 0.383, "step": 1885 }, { "epoch": 2.9508196721311473, "grad_norm": 0.3135896921157837, "learning_rate": 0.0001776459017170976, "loss": 0.3665, "step": 1890 }, { "epoch": 2.958626073380172, "grad_norm": 0.28260594606399536, "learning_rate": 0.00017747376161070344, "loss": 0.3856, "step": 1895 }, { "epoch": 2.9664324746291957, "grad_norm": 0.26851585507392883, "learning_rate": 0.0001773010453362737, "loss": 0.3838, "step": 1900 }, { "epoch": 2.9742388758782203, "grad_norm": 0.2853715121746063, "learning_rate": 0.00017712775417828968, "loss": 0.3776, "step": 1905 }, { "epoch": 2.982045277127244, "grad_norm": 0.28013601899147034, "learning_rate": 0.00017695388942550805, "loss": 0.3741, "step": 1910 }, { "epoch": 2.9898516783762688, "grad_norm": 0.270375519990921, "learning_rate": 0.00017677945237095123, "loss": 0.3707, "step": 1915 }, { "epoch": 2.9976580796252925, "grad_norm": 0.29110994935035706, "learning_rate": 0.0001766044443118978, "loss": 0.3774, "step": 1920 }, { "epoch": 2.999219359875098, "eval_loss": 1.0361436605453491, "eval_runtime": 0.337, "eval_samples_per_second": 14.838, "eval_steps_per_second": 2.968, "step": 1921 }, { "epoch": 3.0054644808743167, "grad_norm": 0.27365708351135254, "learning_rate": 0.000176428866549873, "loss": 0.357, "step": 1925 }, { "epoch": 3.013270882123341, "grad_norm": 0.29497936367988586, "learning_rate": 0.00017625272039063883, "loss": 0.3487, "step": 1930 }, { "epoch": 3.021077283372365, "grad_norm": 0.28680676221847534, "learning_rate": 0.00017607600714418436, "loss": 0.3427, "step": 1935 }, { "epoch": 3.0288836846213893, "grad_norm": 0.29143229126930237, "learning_rate": 0.0001758987281247162, "loss": 0.3472, "step": 1940 }, { "epoch": 3.0366900858704136, "grad_norm": 0.28377851843833923, "learning_rate": 0.00017572088465064848, "loss": 0.3468, "step": 1945 }, { "epoch": 3.0444964871194378, "grad_norm": 0.27560412883758545, "learning_rate": 0.00017554247804459316, "loss": 0.3444, "step": 1950 }, { "epoch": 3.052302888368462, "grad_norm": 0.2871435880661011, "learning_rate": 0.00017536350963335022, "loss": 0.355, "step": 1955 }, { "epoch": 3.060109289617486, "grad_norm": 0.29234611988067627, "learning_rate": 0.00017518398074789775, "loss": 0.3405, "step": 1960 }, { "epoch": 3.0679156908665104, "grad_norm": 0.2977675199508667, "learning_rate": 0.000175003892723382, "loss": 0.3522, "step": 1965 }, { "epoch": 3.0757220921155346, "grad_norm": 0.2990856468677521, "learning_rate": 0.00017482324689910758, "loss": 0.352, "step": 1970 }, { "epoch": 3.083528493364559, "grad_norm": 0.28264856338500977, "learning_rate": 0.00017464204461852738, "loss": 0.3497, "step": 1975 }, { "epoch": 3.091334894613583, "grad_norm": 0.3084184229373932, "learning_rate": 0.00017446028722923265, "loss": 0.3586, "step": 1980 }, { "epoch": 3.0991412958626072, "grad_norm": 0.29917633533477783, "learning_rate": 0.00017427797608294288, "loss": 0.3542, "step": 1985 }, { "epoch": 3.1069476971116314, "grad_norm": 0.2877940237522125, "learning_rate": 0.00017409511253549593, "loss": 0.3527, "step": 1990 }, { "epoch": 3.1147540983606556, "grad_norm": 0.3050166070461273, "learning_rate": 0.00017391169794683776, "loss": 0.3481, "step": 1995 }, { "epoch": 3.12256049960968, "grad_norm": 0.29156407713890076, "learning_rate": 0.0001737277336810124, "loss": 0.3614, "step": 2000 }, { "epoch": 3.130366900858704, "grad_norm": 0.3042657971382141, "learning_rate": 0.00017354322110615188, "loss": 0.3608, "step": 2005 }, { "epoch": 3.1381733021077283, "grad_norm": 0.29401668906211853, "learning_rate": 0.00017335816159446584, "loss": 0.3618, "step": 2010 }, { "epoch": 3.1459797033567525, "grad_norm": 0.3223430812358856, "learning_rate": 0.00017317255652223157, "loss": 0.3637, "step": 2015 }, { "epoch": 3.1537861046057767, "grad_norm": 0.3130653500556946, "learning_rate": 0.00017298640726978357, "loss": 0.3561, "step": 2020 }, { "epoch": 3.161592505854801, "grad_norm": 0.3067520260810852, "learning_rate": 0.00017279971522150348, "loss": 0.3536, "step": 2025 }, { "epoch": 3.169398907103825, "grad_norm": 0.30372655391693115, "learning_rate": 0.00017261248176580958, "loss": 0.3511, "step": 2030 }, { "epoch": 3.1772053083528493, "grad_norm": 0.34251293540000916, "learning_rate": 0.00017242470829514672, "loss": 0.3591, "step": 2035 }, { "epoch": 3.1850117096018735, "grad_norm": 0.34350651502609253, "learning_rate": 0.00017223639620597556, "loss": 0.345, "step": 2040 }, { "epoch": 3.1928181108508977, "grad_norm": 0.29779571294784546, "learning_rate": 0.0001720475468987627, "loss": 0.3496, "step": 2045 }, { "epoch": 3.200624512099922, "grad_norm": 0.29767748713493347, "learning_rate": 0.0001718581617779698, "loss": 0.35, "step": 2050 }, { "epoch": 3.208430913348946, "grad_norm": 0.30058038234710693, "learning_rate": 0.00017166824225204352, "loss": 0.3481, "step": 2055 }, { "epoch": 3.2162373145979704, "grad_norm": 0.2958409786224365, "learning_rate": 0.00017147778973340466, "loss": 0.3554, "step": 2060 }, { "epoch": 3.2240437158469946, "grad_norm": 0.29515761137008667, "learning_rate": 0.000171286805638438, "loss": 0.3489, "step": 2065 }, { "epoch": 3.2318501170960188, "grad_norm": 0.30113205313682556, "learning_rate": 0.00017109529138748157, "loss": 0.3437, "step": 2070 }, { "epoch": 3.239656518345043, "grad_norm": 0.2853497862815857, "learning_rate": 0.0001709032484048162, "loss": 0.3515, "step": 2075 }, { "epoch": 3.247462919594067, "grad_norm": 0.3135252892971039, "learning_rate": 0.00017071067811865476, "loss": 0.3502, "step": 2080 }, { "epoch": 3.2552693208430914, "grad_norm": 0.30913615226745605, "learning_rate": 0.00017051758196113173, "loss": 0.3468, "step": 2085 }, { "epoch": 3.2630757220921156, "grad_norm": 0.2965530753135681, "learning_rate": 0.00017032396136829246, "loss": 0.3565, "step": 2090 }, { "epoch": 3.27088212334114, "grad_norm": 0.2974913418292999, "learning_rate": 0.00017012981778008243, "loss": 0.3442, "step": 2095 }, { "epoch": 3.278688524590164, "grad_norm": 0.2917117476463318, "learning_rate": 0.00016993515264033672, "loss": 0.3479, "step": 2100 }, { "epoch": 3.2864949258391882, "grad_norm": 0.29723039269447327, "learning_rate": 0.00016973996739676905, "loss": 0.3523, "step": 2105 }, { "epoch": 3.2943013270882124, "grad_norm": 0.3127197027206421, "learning_rate": 0.00016954426350096116, "loss": 0.3574, "step": 2110 }, { "epoch": 3.3021077283372366, "grad_norm": 0.28733986616134644, "learning_rate": 0.00016934804240835205, "loss": 0.3495, "step": 2115 }, { "epoch": 3.309914129586261, "grad_norm": 0.30774375796318054, "learning_rate": 0.00016915130557822695, "loss": 0.3515, "step": 2120 }, { "epoch": 3.317720530835285, "grad_norm": 0.3072982132434845, "learning_rate": 0.0001689540544737067, "loss": 0.3473, "step": 2125 }, { "epoch": 3.3255269320843093, "grad_norm": 0.30740848183631897, "learning_rate": 0.00016875629056173675, "loss": 0.3631, "step": 2130 }, { "epoch": 3.3333333333333335, "grad_norm": 0.2990294396877289, "learning_rate": 0.00016855801531307624, "loss": 0.3508, "step": 2135 }, { "epoch": 3.3411397345823577, "grad_norm": 0.28540557622909546, "learning_rate": 0.00016835923020228712, "loss": 0.353, "step": 2140 }, { "epoch": 3.348946135831382, "grad_norm": 0.29948052763938904, "learning_rate": 0.0001681599367077232, "loss": 0.3627, "step": 2145 }, { "epoch": 3.356752537080406, "grad_norm": 0.30102795362472534, "learning_rate": 0.00016796013631151897, "loss": 0.3526, "step": 2150 }, { "epoch": 3.3645589383294303, "grad_norm": 0.2947191298007965, "learning_rate": 0.00016775983049957887, "loss": 0.3645, "step": 2155 }, { "epoch": 3.3723653395784545, "grad_norm": 0.3223245143890381, "learning_rate": 0.00016755902076156604, "loss": 0.3623, "step": 2160 }, { "epoch": 3.3801717408274783, "grad_norm": 0.28947457671165466, "learning_rate": 0.00016735770859089123, "loss": 0.3638, "step": 2165 }, { "epoch": 3.387978142076503, "grad_norm": 0.28693681955337524, "learning_rate": 0.00016715589548470185, "loss": 0.3614, "step": 2170 }, { "epoch": 3.3957845433255267, "grad_norm": 0.29449644684791565, "learning_rate": 0.00016695358294387065, "loss": 0.3529, "step": 2175 }, { "epoch": 3.4035909445745514, "grad_norm": 0.289899080991745, "learning_rate": 0.00016675077247298474, "loss": 0.3556, "step": 2180 }, { "epoch": 3.411397345823575, "grad_norm": 0.2969209849834442, "learning_rate": 0.00016654746558033423, "loss": 0.3598, "step": 2185 }, { "epoch": 3.4192037470725998, "grad_norm": 0.3075496554374695, "learning_rate": 0.00016634366377790114, "loss": 0.3509, "step": 2190 }, { "epoch": 3.4270101483216235, "grad_norm": 0.28629180788993835, "learning_rate": 0.00016613936858134802, "loss": 0.3481, "step": 2195 }, { "epoch": 3.4348165495706477, "grad_norm": 0.3028337061405182, "learning_rate": 0.00016593458151000688, "loss": 0.358, "step": 2200 }, { "epoch": 3.442622950819672, "grad_norm": 0.2874985933303833, "learning_rate": 0.00016572930408686776, "loss": 0.3593, "step": 2205 }, { "epoch": 3.450429352068696, "grad_norm": 0.29345887899398804, "learning_rate": 0.00016552353783856732, "loss": 0.3485, "step": 2210 }, { "epoch": 3.4582357533177204, "grad_norm": 0.28079068660736084, "learning_rate": 0.00016531728429537766, "loss": 0.3606, "step": 2215 }, { "epoch": 3.4660421545667446, "grad_norm": 0.2924216389656067, "learning_rate": 0.0001651105449911949, "loss": 0.351, "step": 2220 }, { "epoch": 3.473848555815769, "grad_norm": 0.29718244075775146, "learning_rate": 0.0001649033214635277, "loss": 0.3521, "step": 2225 }, { "epoch": 3.481654957064793, "grad_norm": 0.2965914309024811, "learning_rate": 0.00016469561525348577, "loss": 0.3564, "step": 2230 }, { "epoch": 3.489461358313817, "grad_norm": 0.2874649167060852, "learning_rate": 0.0001644874279057686, "loss": 0.3554, "step": 2235 }, { "epoch": 3.4972677595628414, "grad_norm": 0.2855139970779419, "learning_rate": 0.00016427876096865394, "loss": 0.3619, "step": 2240 }, { "epoch": 3.5050741608118656, "grad_norm": 0.2949666976928711, "learning_rate": 0.0001640696159939861, "loss": 0.3558, "step": 2245 }, { "epoch": 3.51288056206089, "grad_norm": 0.29184606671333313, "learning_rate": 0.00016385999453716454, "loss": 0.3625, "step": 2250 }, { "epoch": 3.520686963309914, "grad_norm": 0.29414233565330505, "learning_rate": 0.00016364989815713233, "loss": 0.3559, "step": 2255 }, { "epoch": 3.5284933645589383, "grad_norm": 0.2970221936702728, "learning_rate": 0.00016343932841636456, "loss": 0.3488, "step": 2260 }, { "epoch": 3.5362997658079625, "grad_norm": 0.305912047624588, "learning_rate": 0.00016322828688085657, "loss": 0.3553, "step": 2265 }, { "epoch": 3.5441061670569867, "grad_norm": 0.3016732931137085, "learning_rate": 0.00016301677512011248, "loss": 0.359, "step": 2270 }, { "epoch": 3.551912568306011, "grad_norm": 0.2802044749259949, "learning_rate": 0.00016280479470713344, "loss": 0.347, "step": 2275 }, { "epoch": 3.559718969555035, "grad_norm": 0.3013835549354553, "learning_rate": 0.00016259234721840591, "loss": 0.3664, "step": 2280 }, { "epoch": 3.5675253708040593, "grad_norm": 0.2899465262889862, "learning_rate": 0.00016237943423389001, "loss": 0.3638, "step": 2285 }, { "epoch": 3.5753317720530835, "grad_norm": 0.2953374981880188, "learning_rate": 0.00016216605733700775, "loss": 0.3621, "step": 2290 }, { "epoch": 3.5831381733021077, "grad_norm": 0.28580641746520996, "learning_rate": 0.00016195221811463113, "loss": 0.3558, "step": 2295 }, { "epoch": 3.590944574551132, "grad_norm": 0.2812565565109253, "learning_rate": 0.00016173791815707051, "loss": 0.357, "step": 2300 }, { "epoch": 3.598750975800156, "grad_norm": 0.3074684739112854, "learning_rate": 0.00016152315905806268, "loss": 0.3593, "step": 2305 }, { "epoch": 3.6065573770491803, "grad_norm": 0.29887428879737854, "learning_rate": 0.0001613079424147591, "loss": 0.3607, "step": 2310 }, { "epoch": 3.6143637782982045, "grad_norm": 0.2820027470588684, "learning_rate": 0.00016109226982771393, "loss": 0.3535, "step": 2315 }, { "epoch": 3.6221701795472288, "grad_norm": 0.28952109813690186, "learning_rate": 0.00016087614290087208, "loss": 0.3563, "step": 2320 }, { "epoch": 3.629976580796253, "grad_norm": 0.29456931352615356, "learning_rate": 0.00016065956324155746, "loss": 0.3601, "step": 2325 }, { "epoch": 3.637782982045277, "grad_norm": 0.29095590114593506, "learning_rate": 0.0001604425324604609, "loss": 0.3633, "step": 2330 }, { "epoch": 3.6455893832943014, "grad_norm": 0.27606719732284546, "learning_rate": 0.00016022505217162822, "loss": 0.3621, "step": 2335 }, { "epoch": 3.6533957845433256, "grad_norm": 0.2846406400203705, "learning_rate": 0.0001600071239924481, "loss": 0.349, "step": 2340 }, { "epoch": 3.66120218579235, "grad_norm": 0.28319764137268066, "learning_rate": 0.00015978874954364033, "loss": 0.3514, "step": 2345 }, { "epoch": 3.669008587041374, "grad_norm": 0.3063547909259796, "learning_rate": 0.00015956993044924334, "loss": 0.3504, "step": 2350 }, { "epoch": 3.676814988290398, "grad_norm": 0.27655765414237976, "learning_rate": 0.0001593506683366026, "loss": 0.3466, "step": 2355 }, { "epoch": 3.6846213895394224, "grad_norm": 0.2881523072719574, "learning_rate": 0.00015913096483635824, "loss": 0.3466, "step": 2360 }, { "epoch": 3.6924277907884466, "grad_norm": 0.2950555682182312, "learning_rate": 0.00015891082158243282, "loss": 0.3512, "step": 2365 }, { "epoch": 3.700234192037471, "grad_norm": 0.2711721360683441, "learning_rate": 0.00015869024021201948, "loss": 0.3545, "step": 2370 }, { "epoch": 3.708040593286495, "grad_norm": 0.31127429008483887, "learning_rate": 0.00015846922236556946, "loss": 0.3616, "step": 2375 }, { "epoch": 3.7158469945355193, "grad_norm": 0.28716492652893066, "learning_rate": 0.00015824776968678024, "loss": 0.3548, "step": 2380 }, { "epoch": 3.7236533957845435, "grad_norm": 0.28540754318237305, "learning_rate": 0.00015802588382258292, "loss": 0.3571, "step": 2385 }, { "epoch": 3.7314597970335677, "grad_norm": 0.30483660101890564, "learning_rate": 0.00015780356642313036, "loss": 0.3475, "step": 2390 }, { "epoch": 3.739266198282592, "grad_norm": 0.28474995493888855, "learning_rate": 0.00015758081914178456, "loss": 0.3475, "step": 2395 }, { "epoch": 3.747072599531616, "grad_norm": 0.28476497530937195, "learning_rate": 0.0001573576436351046, "loss": 0.3525, "step": 2400 }, { "epoch": 3.7548790007806403, "grad_norm": 0.29259222745895386, "learning_rate": 0.00015713404156283435, "loss": 0.3683, "step": 2405 }, { "epoch": 3.762685402029664, "grad_norm": 0.2946162223815918, "learning_rate": 0.00015691001458788983, "loss": 0.3587, "step": 2410 }, { "epoch": 3.7704918032786887, "grad_norm": 0.2999446988105774, "learning_rate": 0.0001566855643763472, "loss": 0.355, "step": 2415 }, { "epoch": 3.7782982045277125, "grad_norm": 0.2700735032558441, "learning_rate": 0.00015646069259743007, "loss": 0.3531, "step": 2420 }, { "epoch": 3.786104605776737, "grad_norm": 0.28989067673683167, "learning_rate": 0.00015623540092349732, "loss": 0.3568, "step": 2425 }, { "epoch": 3.793911007025761, "grad_norm": 0.2979315221309662, "learning_rate": 0.00015600969103003056, "loss": 0.3514, "step": 2430 }, { "epoch": 3.8017174082747855, "grad_norm": 0.27748894691467285, "learning_rate": 0.00015578356459562163, "loss": 0.3474, "step": 2435 }, { "epoch": 3.8095238095238093, "grad_norm": 0.28995972871780396, "learning_rate": 0.00015555702330196023, "loss": 0.352, "step": 2440 }, { "epoch": 3.817330210772834, "grad_norm": 0.2819332480430603, "learning_rate": 0.00015533006883382123, "loss": 0.3482, "step": 2445 }, { "epoch": 3.8251366120218577, "grad_norm": 0.2890748083591461, "learning_rate": 0.0001551027028790524, "loss": 0.3608, "step": 2450 }, { "epoch": 3.8329430132708824, "grad_norm": 0.295684278011322, "learning_rate": 0.0001548749271285616, "loss": 0.3541, "step": 2455 }, { "epoch": 3.840749414519906, "grad_norm": 0.2782337963581085, "learning_rate": 0.00015464674327630434, "loss": 0.3467, "step": 2460 }, { "epoch": 3.848555815768931, "grad_norm": 0.2889052629470825, "learning_rate": 0.00015441815301927122, "loss": 0.3597, "step": 2465 }, { "epoch": 3.8563622170179546, "grad_norm": 0.27629801630973816, "learning_rate": 0.00015418915805747517, "loss": 0.3543, "step": 2470 }, { "epoch": 3.8641686182669788, "grad_norm": 0.2963428497314453, "learning_rate": 0.00015395976009393894, "loss": 0.3637, "step": 2475 }, { "epoch": 3.871975019516003, "grad_norm": 0.29081323742866516, "learning_rate": 0.0001537299608346824, "loss": 0.3563, "step": 2480 }, { "epoch": 3.879781420765027, "grad_norm": 0.29357287287712097, "learning_rate": 0.00015349976198870973, "loss": 0.3633, "step": 2485 }, { "epoch": 3.8875878220140514, "grad_norm": 0.29132893681526184, "learning_rate": 0.00015326916526799692, "loss": 0.3563, "step": 2490 }, { "epoch": 3.8953942232630756, "grad_norm": 0.29773959517478943, "learning_rate": 0.00015303817238747892, "loss": 0.3511, "step": 2495 }, { "epoch": 3.9032006245121, "grad_norm": 0.2842431366443634, "learning_rate": 0.0001528067850650368, "loss": 0.3533, "step": 2500 }, { "epoch": 3.911007025761124, "grad_norm": 0.277873694896698, "learning_rate": 0.00015257500502148522, "loss": 0.3549, "step": 2505 }, { "epoch": 3.9188134270101482, "grad_norm": 0.2919045388698578, "learning_rate": 0.0001523428339805594, "loss": 0.3521, "step": 2510 }, { "epoch": 3.9266198282591724, "grad_norm": 0.289386123418808, "learning_rate": 0.00015211027366890235, "loss": 0.3603, "step": 2515 }, { "epoch": 3.9344262295081966, "grad_norm": 0.27099329233169556, "learning_rate": 0.00015187732581605217, "loss": 0.3553, "step": 2520 }, { "epoch": 3.942232630757221, "grad_norm": 0.2989470660686493, "learning_rate": 0.00015164399215442898, "loss": 0.3562, "step": 2525 }, { "epoch": 3.950039032006245, "grad_norm": 0.2968309223651886, "learning_rate": 0.00015141027441932216, "loss": 0.3581, "step": 2530 }, { "epoch": 3.9578454332552693, "grad_norm": 0.29724830389022827, "learning_rate": 0.00015117617434887753, "loss": 0.3514, "step": 2535 }, { "epoch": 3.9656518345042935, "grad_norm": 0.30498626828193665, "learning_rate": 0.0001509416936840842, "loss": 0.3691, "step": 2540 }, { "epoch": 3.9734582357533177, "grad_norm": 0.28580141067504883, "learning_rate": 0.00015070683416876172, "loss": 0.3509, "step": 2545 }, { "epoch": 3.981264637002342, "grad_norm": 0.28979378938674927, "learning_rate": 0.0001504715975495472, "loss": 0.3598, "step": 2550 }, { "epoch": 3.989071038251366, "grad_norm": 0.30698639154434204, "learning_rate": 0.0001502359855758823, "loss": 0.357, "step": 2555 }, { "epoch": 3.9968774395003903, "grad_norm": 0.2776874303817749, "learning_rate": 0.00015000000000000001, "loss": 0.355, "step": 2560 }, { "epoch": 4.0, "eval_loss": 1.041857361793518, "eval_runtime": 0.3229, "eval_samples_per_second": 15.485, "eval_steps_per_second": 3.097, "step": 2562 }, { "epoch": 4.004683840749414, "grad_norm": 0.2781620919704437, "learning_rate": 0.00014976364257691192, "loss": 0.3413, "step": 2565 }, { "epoch": 4.012490241998439, "grad_norm": 0.3088448643684387, "learning_rate": 0.00014952691506439495, "loss": 0.3303, "step": 2570 }, { "epoch": 4.0202966432474625, "grad_norm": 0.28063511848449707, "learning_rate": 0.00014928981922297842, "loss": 0.3268, "step": 2575 }, { "epoch": 4.028103044496487, "grad_norm": 0.32527846097946167, "learning_rate": 0.0001490523568159308, "loss": 0.3184, "step": 2580 }, { "epoch": 4.035909445745511, "grad_norm": 0.29716962575912476, "learning_rate": 0.00014881452960924677, "loss": 0.3287, "step": 2585 }, { "epoch": 4.043715846994536, "grad_norm": 0.30141621828079224, "learning_rate": 0.00014857633937163402, "loss": 0.3331, "step": 2590 }, { "epoch": 4.051522248243559, "grad_norm": 0.2995307743549347, "learning_rate": 0.00014833778787449998, "loss": 0.3313, "step": 2595 }, { "epoch": 4.059328649492584, "grad_norm": 0.30431655049324036, "learning_rate": 0.00014809887689193877, "loss": 0.334, "step": 2600 }, { "epoch": 4.067135050741608, "grad_norm": 0.3243234157562256, "learning_rate": 0.0001478596082007181, "loss": 0.3339, "step": 2605 }, { "epoch": 4.074941451990632, "grad_norm": 0.3121318817138672, "learning_rate": 0.0001476199835802658, "loss": 0.3306, "step": 2610 }, { "epoch": 4.082747853239656, "grad_norm": 0.31301409006118774, "learning_rate": 0.0001473800048126568, "loss": 0.3336, "step": 2615 }, { "epoch": 4.090554254488681, "grad_norm": 0.29308003187179565, "learning_rate": 0.0001471396736825998, "loss": 0.3239, "step": 2620 }, { "epoch": 4.098360655737705, "grad_norm": 0.30689167976379395, "learning_rate": 0.0001468989919774239, "loss": 0.3342, "step": 2625 }, { "epoch": 4.106167056986729, "grad_norm": 0.32097750902175903, "learning_rate": 0.0001466579614870656, "loss": 0.3386, "step": 2630 }, { "epoch": 4.113973458235753, "grad_norm": 0.3277100622653961, "learning_rate": 0.00014641658400405515, "loss": 0.3434, "step": 2635 }, { "epoch": 4.121779859484778, "grad_norm": 0.30586951971054077, "learning_rate": 0.00014617486132350343, "loss": 0.3328, "step": 2640 }, { "epoch": 4.129586260733801, "grad_norm": 0.30447080731391907, "learning_rate": 0.0001459327952430884, "loss": 0.3343, "step": 2645 }, { "epoch": 4.137392661982826, "grad_norm": 0.31396037340164185, "learning_rate": 0.00014569038756304207, "loss": 0.3331, "step": 2650 }, { "epoch": 4.14519906323185, "grad_norm": 0.30158740282058716, "learning_rate": 0.0001454476400861368, "loss": 0.33, "step": 2655 }, { "epoch": 4.1530054644808745, "grad_norm": 0.30618518590927124, "learning_rate": 0.000145204554617672, "loss": 0.3297, "step": 2660 }, { "epoch": 4.160811865729898, "grad_norm": 0.3204987049102783, "learning_rate": 0.00014496113296546067, "loss": 0.3356, "step": 2665 }, { "epoch": 4.168618266978923, "grad_norm": 0.31640496850013733, "learning_rate": 0.0001447173769398161, "loss": 0.3397, "step": 2670 }, { "epoch": 4.176424668227947, "grad_norm": 0.3094062805175781, "learning_rate": 0.0001444732883535382, "loss": 0.3359, "step": 2675 }, { "epoch": 4.184231069476971, "grad_norm": 0.3020561933517456, "learning_rate": 0.00014422886902190014, "loss": 0.3313, "step": 2680 }, { "epoch": 4.192037470725995, "grad_norm": 0.3145982027053833, "learning_rate": 0.00014398412076263486, "loss": 0.3441, "step": 2685 }, { "epoch": 4.19984387197502, "grad_norm": 0.3120006322860718, "learning_rate": 0.00014373904539592144, "loss": 0.3334, "step": 2690 }, { "epoch": 4.2076502732240435, "grad_norm": 0.319619357585907, "learning_rate": 0.00014349364474437173, "loss": 0.3287, "step": 2695 }, { "epoch": 4.215456674473068, "grad_norm": 0.3287660479545593, "learning_rate": 0.00014324792063301662, "loss": 0.3296, "step": 2700 }, { "epoch": 4.223263075722092, "grad_norm": 0.3166675567626953, "learning_rate": 0.00014300187488929265, "loss": 0.3433, "step": 2705 }, { "epoch": 4.231069476971117, "grad_norm": 0.3454567492008209, "learning_rate": 0.00014275550934302823, "loss": 0.3384, "step": 2710 }, { "epoch": 4.23887587822014, "grad_norm": 0.30334779620170593, "learning_rate": 0.00014250882582643012, "loss": 0.334, "step": 2715 }, { "epoch": 4.246682279469165, "grad_norm": 0.31963109970092773, "learning_rate": 0.00014226182617406996, "loss": 0.3428, "step": 2720 }, { "epoch": 4.254488680718189, "grad_norm": 0.30890125036239624, "learning_rate": 0.00014201451222287025, "loss": 0.3326, "step": 2725 }, { "epoch": 4.262295081967213, "grad_norm": 0.32028844952583313, "learning_rate": 0.0001417668858120911, "loss": 0.333, "step": 2730 }, { "epoch": 4.270101483216237, "grad_norm": 0.3037426471710205, "learning_rate": 0.00014151894878331622, "loss": 0.3302, "step": 2735 }, { "epoch": 4.277907884465262, "grad_norm": 0.305708110332489, "learning_rate": 0.00014127070298043947, "loss": 0.3387, "step": 2740 }, { "epoch": 4.285714285714286, "grad_norm": 0.3226141333580017, "learning_rate": 0.00014102215024965105, "loss": 0.3352, "step": 2745 }, { "epoch": 4.29352068696331, "grad_norm": 0.3047237992286682, "learning_rate": 0.00014077329243942369, "loss": 0.3377, "step": 2750 }, { "epoch": 4.301327088212334, "grad_norm": 0.3010615408420563, "learning_rate": 0.000140524131400499, "loss": 0.3325, "step": 2755 }, { "epoch": 4.309133489461359, "grad_norm": 0.3127407431602478, "learning_rate": 0.00014027466898587374, "loss": 0.3357, "step": 2760 }, { "epoch": 4.316939890710382, "grad_norm": 0.31216832995414734, "learning_rate": 0.00014002490705078592, "loss": 0.346, "step": 2765 }, { "epoch": 4.324746291959407, "grad_norm": 0.31948304176330566, "learning_rate": 0.00013977484745270112, "loss": 0.3292, "step": 2770 }, { "epoch": 4.332552693208431, "grad_norm": 0.3235209584236145, "learning_rate": 0.00013952449205129855, "loss": 0.3415, "step": 2775 }, { "epoch": 4.3403590944574555, "grad_norm": 0.3177550137042999, "learning_rate": 0.0001392738427084574, "loss": 0.3341, "step": 2780 }, { "epoch": 4.348165495706479, "grad_norm": 0.31263765692710876, "learning_rate": 0.00013902290128824283, "loss": 0.3348, "step": 2785 }, { "epoch": 4.355971896955504, "grad_norm": 0.31187278032302856, "learning_rate": 0.00013877166965689205, "loss": 0.3405, "step": 2790 }, { "epoch": 4.363778298204528, "grad_norm": 0.3060557246208191, "learning_rate": 0.00013852014968280082, "loss": 0.3328, "step": 2795 }, { "epoch": 4.371584699453552, "grad_norm": 0.30810457468032837, "learning_rate": 0.000138268343236509, "loss": 0.3303, "step": 2800 }, { "epoch": 4.379391100702576, "grad_norm": 0.309929758310318, "learning_rate": 0.0001380162521906871, "loss": 0.3371, "step": 2805 }, { "epoch": 4.387197501951601, "grad_norm": 0.31695547699928284, "learning_rate": 0.00013776387842012218, "loss": 0.3393, "step": 2810 }, { "epoch": 4.3950039032006245, "grad_norm": 0.31112900376319885, "learning_rate": 0.0001375112238017038, "loss": 0.338, "step": 2815 }, { "epoch": 4.402810304449648, "grad_norm": 0.29906710982322693, "learning_rate": 0.0001372582902144103, "loss": 0.3318, "step": 2820 }, { "epoch": 4.410616705698673, "grad_norm": 0.30970847606658936, "learning_rate": 0.00013700507953929463, "loss": 0.3387, "step": 2825 }, { "epoch": 4.418423106947698, "grad_norm": 0.3163338005542755, "learning_rate": 0.00013675159365947036, "loss": 0.3371, "step": 2830 }, { "epoch": 4.426229508196721, "grad_norm": 0.31127607822418213, "learning_rate": 0.0001364978344600979, "loss": 0.3372, "step": 2835 }, { "epoch": 4.434035909445745, "grad_norm": 0.2967492640018463, "learning_rate": 0.00013624380382837016, "loss": 0.339, "step": 2840 }, { "epoch": 4.44184231069477, "grad_norm": 0.31639572978019714, "learning_rate": 0.00013598950365349883, "loss": 0.3411, "step": 2845 }, { "epoch": 4.4496487119437935, "grad_norm": 0.3143470883369446, "learning_rate": 0.00013573493582670003, "loss": 0.3386, "step": 2850 }, { "epoch": 4.457455113192818, "grad_norm": 0.3157510757446289, "learning_rate": 0.0001354801022411805, "loss": 0.3339, "step": 2855 }, { "epoch": 4.465261514441842, "grad_norm": 0.3069648742675781, "learning_rate": 0.00013522500479212337, "loss": 0.3385, "step": 2860 }, { "epoch": 4.473067915690867, "grad_norm": 0.31706541776657104, "learning_rate": 0.0001349696453766741, "loss": 0.3353, "step": 2865 }, { "epoch": 4.48087431693989, "grad_norm": 0.3200201988220215, "learning_rate": 0.00013471402589392638, "loss": 0.3419, "step": 2870 }, { "epoch": 4.488680718188915, "grad_norm": 0.32029232382774353, "learning_rate": 0.00013445814824490805, "loss": 0.3417, "step": 2875 }, { "epoch": 4.496487119437939, "grad_norm": 0.3023779094219208, "learning_rate": 0.00013420201433256689, "loss": 0.3388, "step": 2880 }, { "epoch": 4.504293520686963, "grad_norm": 0.3167456090450287, "learning_rate": 0.00013394562606175648, "loss": 0.335, "step": 2885 }, { "epoch": 4.512099921935987, "grad_norm": 0.31978872418403625, "learning_rate": 0.000133688985339222, "loss": 0.3425, "step": 2890 }, { "epoch": 4.519906323185012, "grad_norm": 0.32426413893699646, "learning_rate": 0.00013343209407358627, "loss": 0.3325, "step": 2895 }, { "epoch": 4.527712724434036, "grad_norm": 0.28932276368141174, "learning_rate": 0.00013317495417533524, "loss": 0.3365, "step": 2900 }, { "epoch": 4.53551912568306, "grad_norm": 0.31238290667533875, "learning_rate": 0.00013291756755680388, "loss": 0.3353, "step": 2905 }, { "epoch": 4.543325526932084, "grad_norm": 0.3126748204231262, "learning_rate": 0.00013265993613216224, "loss": 0.3363, "step": 2910 }, { "epoch": 4.551131928181109, "grad_norm": 0.331447571516037, "learning_rate": 0.0001324020618174007, "loss": 0.3434, "step": 2915 }, { "epoch": 4.558938329430132, "grad_norm": 0.31981754302978516, "learning_rate": 0.00013214394653031616, "loss": 0.3321, "step": 2920 }, { "epoch": 4.566744730679157, "grad_norm": 0.30969661474227905, "learning_rate": 0.0001318855921904976, "loss": 0.3432, "step": 2925 }, { "epoch": 4.574551131928181, "grad_norm": 0.3188318610191345, "learning_rate": 0.00013162700071931184, "loss": 0.3397, "step": 2930 }, { "epoch": 4.5823575331772055, "grad_norm": 0.3379952907562256, "learning_rate": 0.00013136817403988917, "loss": 0.3389, "step": 2935 }, { "epoch": 4.590163934426229, "grad_norm": 0.2957393229007721, "learning_rate": 0.00013110911407710908, "loss": 0.3294, "step": 2940 }, { "epoch": 4.597970335675254, "grad_norm": 0.30884405970573425, "learning_rate": 0.00013084982275758615, "loss": 0.3367, "step": 2945 }, { "epoch": 4.605776736924278, "grad_norm": 0.31198689341545105, "learning_rate": 0.00013059030200965536, "loss": 0.3301, "step": 2950 }, { "epoch": 4.613583138173302, "grad_norm": 0.3176429271697998, "learning_rate": 0.00013033055376335804, "loss": 0.3356, "step": 2955 }, { "epoch": 4.621389539422326, "grad_norm": 0.3074142038822174, "learning_rate": 0.00013007057995042732, "loss": 0.3315, "step": 2960 }, { "epoch": 4.629195940671351, "grad_norm": 0.30663779377937317, "learning_rate": 0.00012981038250427403, "loss": 0.3376, "step": 2965 }, { "epoch": 4.6370023419203745, "grad_norm": 0.30579113960266113, "learning_rate": 0.0001295499633599719, "loss": 0.3351, "step": 2970 }, { "epoch": 4.644808743169399, "grad_norm": 0.3152572214603424, "learning_rate": 0.00012928932445424365, "loss": 0.3369, "step": 2975 }, { "epoch": 4.652615144418423, "grad_norm": 0.3045886754989624, "learning_rate": 0.00012902846772544624, "loss": 0.3317, "step": 2980 }, { "epoch": 4.660421545667448, "grad_norm": 0.3164273500442505, "learning_rate": 0.00012876739511355657, "loss": 0.3408, "step": 2985 }, { "epoch": 4.668227946916471, "grad_norm": 0.3016981780529022, "learning_rate": 0.0001285061085601571, "loss": 0.3417, "step": 2990 }, { "epoch": 4.676034348165496, "grad_norm": 0.3070276081562042, "learning_rate": 0.00012824461000842126, "loss": 0.3283, "step": 2995 }, { "epoch": 4.68384074941452, "grad_norm": 0.30681294202804565, "learning_rate": 0.00012798290140309923, "loss": 0.3395, "step": 3000 }, { "epoch": 4.691647150663544, "grad_norm": 0.3165208697319031, "learning_rate": 0.0001277209846905032, "loss": 0.3358, "step": 3005 }, { "epoch": 4.699453551912568, "grad_norm": 0.30810585618019104, "learning_rate": 0.00012745886181849325, "loss": 0.3409, "step": 3010 }, { "epoch": 4.707259953161593, "grad_norm": 0.3146737515926361, "learning_rate": 0.00012719653473646243, "loss": 0.3449, "step": 3015 }, { "epoch": 4.715066354410617, "grad_norm": 0.29630905389785767, "learning_rate": 0.00012693400539532263, "loss": 0.3384, "step": 3020 }, { "epoch": 4.722872755659641, "grad_norm": 0.3214413523674011, "learning_rate": 0.00012667127574748986, "loss": 0.3347, "step": 3025 }, { "epoch": 4.730679156908665, "grad_norm": 0.3026057183742523, "learning_rate": 0.00012640834774686985, "loss": 0.3342, "step": 3030 }, { "epoch": 4.73848555815769, "grad_norm": 0.31248968839645386, "learning_rate": 0.00012614522334884344, "loss": 0.3381, "step": 3035 }, { "epoch": 4.7462919594067134, "grad_norm": 0.3201672434806824, "learning_rate": 0.00012588190451025207, "loss": 0.3282, "step": 3040 }, { "epoch": 4.754098360655737, "grad_norm": 0.32542794942855835, "learning_rate": 0.00012561839318938326, "loss": 0.3464, "step": 3045 }, { "epoch": 4.761904761904762, "grad_norm": 0.2962396740913391, "learning_rate": 0.00012535469134595595, "loss": 0.3451, "step": 3050 }, { "epoch": 4.7697111631537865, "grad_norm": 0.3082274794578552, "learning_rate": 0.00012509080094110604, "loss": 0.3325, "step": 3055 }, { "epoch": 4.77751756440281, "grad_norm": 0.31192833185195923, "learning_rate": 0.00012482672393737164, "loss": 0.3348, "step": 3060 }, { "epoch": 4.785323965651834, "grad_norm": 0.30983906984329224, "learning_rate": 0.00012456246229867877, "loss": 0.3418, "step": 3065 }, { "epoch": 4.793130366900859, "grad_norm": 0.34676092863082886, "learning_rate": 0.0001242980179903264, "loss": 0.3435, "step": 3070 }, { "epoch": 4.800936768149883, "grad_norm": 0.3103564977645874, "learning_rate": 0.0001240333929789721, "loss": 0.3392, "step": 3075 }, { "epoch": 4.808743169398907, "grad_norm": 0.32498899102211, "learning_rate": 0.00012376858923261733, "loss": 0.3387, "step": 3080 }, { "epoch": 4.816549570647931, "grad_norm": 0.3146662414073944, "learning_rate": 0.0001235036087205927, "loss": 0.3406, "step": 3085 }, { "epoch": 4.8243559718969555, "grad_norm": 0.32241055369377136, "learning_rate": 0.0001232384534135435, "loss": 0.3421, "step": 3090 }, { "epoch": 4.83216237314598, "grad_norm": 0.29581108689308167, "learning_rate": 0.00012297312528341484, "loss": 0.3309, "step": 3095 }, { "epoch": 4.839968774395004, "grad_norm": 0.3068925440311432, "learning_rate": 0.00012270762630343734, "loss": 0.3419, "step": 3100 }, { "epoch": 4.847775175644028, "grad_norm": 0.3263617753982544, "learning_rate": 0.000122441958448112, "loss": 0.3359, "step": 3105 }, { "epoch": 4.855581576893052, "grad_norm": 0.3101602792739868, "learning_rate": 0.00012217612369319579, "loss": 0.342, "step": 3110 }, { "epoch": 4.863387978142076, "grad_norm": 0.3126956522464752, "learning_rate": 0.00012191012401568698, "loss": 0.3386, "step": 3115 }, { "epoch": 4.871194379391101, "grad_norm": 0.3099190890789032, "learning_rate": 0.00012164396139381029, "loss": 0.3288, "step": 3120 }, { "epoch": 4.8790007806401245, "grad_norm": 0.3113436996936798, "learning_rate": 0.00012137763780700227, "loss": 0.3412, "step": 3125 }, { "epoch": 4.886807181889149, "grad_norm": 0.29880639910697937, "learning_rate": 0.00012111115523589651, "loss": 0.3292, "step": 3130 }, { "epoch": 4.894613583138173, "grad_norm": 0.31602269411087036, "learning_rate": 0.00012084451566230906, "loss": 0.3372, "step": 3135 }, { "epoch": 4.902419984387198, "grad_norm": 0.3172509968280792, "learning_rate": 0.00012057772106922349, "loss": 0.3395, "step": 3140 }, { "epoch": 4.910226385636221, "grad_norm": 0.31268373131752014, "learning_rate": 0.0001203107734407763, "loss": 0.339, "step": 3145 }, { "epoch": 4.918032786885246, "grad_norm": 0.3162590265274048, "learning_rate": 0.00012004367476224206, "loss": 0.3484, "step": 3150 }, { "epoch": 4.92583918813427, "grad_norm": 0.34318986535072327, "learning_rate": 0.00011977642702001875, "loss": 0.3473, "step": 3155 }, { "epoch": 4.9336455893832944, "grad_norm": 0.2913156747817993, "learning_rate": 0.00011950903220161285, "loss": 0.3415, "step": 3160 }, { "epoch": 4.941451990632318, "grad_norm": 0.29538801312446594, "learning_rate": 0.0001192414922956247, "loss": 0.3361, "step": 3165 }, { "epoch": 4.949258391881343, "grad_norm": 0.3098941743373871, "learning_rate": 0.00011897380929173364, "loss": 0.3384, "step": 3170 }, { "epoch": 4.957064793130367, "grad_norm": 0.3089110851287842, "learning_rate": 0.0001187059851806832, "loss": 0.335, "step": 3175 }, { "epoch": 4.964871194379391, "grad_norm": 0.29929882287979126, "learning_rate": 0.00011843802195426635, "loss": 0.3364, "step": 3180 }, { "epoch": 4.972677595628415, "grad_norm": 0.31270211935043335, "learning_rate": 0.00011816992160531062, "loss": 0.3482, "step": 3185 }, { "epoch": 4.98048399687744, "grad_norm": 0.3091076910495758, "learning_rate": 0.0001179016861276633, "loss": 0.3387, "step": 3190 }, { "epoch": 4.9882903981264635, "grad_norm": 0.3052199184894562, "learning_rate": 0.00011763331751617663, "loss": 0.3454, "step": 3195 }, { "epoch": 4.996096799375488, "grad_norm": 0.2918078303337097, "learning_rate": 0.00011736481776669306, "loss": 0.3413, "step": 3200 }, { "epoch": 4.999219359875098, "eval_loss": 1.0547945499420166, "eval_runtime": 0.336, "eval_samples_per_second": 14.88, "eval_steps_per_second": 2.976, "step": 3202 }, { "epoch": 5.003903200624512, "grad_norm": 0.2900161147117615, "learning_rate": 0.00011709618887603014, "loss": 0.3236, "step": 3205 }, { "epoch": 5.0117096018735365, "grad_norm": 0.3155598044395447, "learning_rate": 0.00011682743284196593, "loss": 0.3129, "step": 3210 }, { "epoch": 5.01951600312256, "grad_norm": 0.3263396918773651, "learning_rate": 0.00011655855166322409, "loss": 0.3154, "step": 3215 }, { "epoch": 5.027322404371585, "grad_norm": 0.3178098797798157, "learning_rate": 0.0001162895473394589, "loss": 0.3133, "step": 3220 }, { "epoch": 5.035128805620609, "grad_norm": 0.3300285339355469, "learning_rate": 0.00011602042187124045, "loss": 0.3103, "step": 3225 }, { "epoch": 5.042935206869633, "grad_norm": 0.3271562159061432, "learning_rate": 0.00011575117726003979, "loss": 0.3092, "step": 3230 }, { "epoch": 5.050741608118657, "grad_norm": 0.3253972828388214, "learning_rate": 0.00011548181550821411, "loss": 0.3076, "step": 3235 }, { "epoch": 5.058548009367682, "grad_norm": 0.33159464597702026, "learning_rate": 0.00011521233861899167, "loss": 0.3223, "step": 3240 }, { "epoch": 5.0663544106167056, "grad_norm": 0.3398463726043701, "learning_rate": 0.00011494274859645707, "loss": 0.3163, "step": 3245 }, { "epoch": 5.07416081186573, "grad_norm": 0.332904577255249, "learning_rate": 0.00011467304744553618, "loss": 0.308, "step": 3250 }, { "epoch": 5.081967213114754, "grad_norm": 0.32349440455436707, "learning_rate": 0.00011440323717198144, "loss": 0.3045, "step": 3255 }, { "epoch": 5.089773614363779, "grad_norm": 0.3356363773345947, "learning_rate": 0.00011413331978235678, "loss": 0.3188, "step": 3260 }, { "epoch": 5.097580015612802, "grad_norm": 0.3371128737926483, "learning_rate": 0.00011386329728402268, "loss": 0.3103, "step": 3265 }, { "epoch": 5.105386416861827, "grad_norm": 0.3300149142742157, "learning_rate": 0.00011359317168512142, "loss": 0.3055, "step": 3270 }, { "epoch": 5.113192818110851, "grad_norm": 0.3270377814769745, "learning_rate": 0.000113322944994562, "loss": 0.3115, "step": 3275 }, { "epoch": 5.1209992193598755, "grad_norm": 0.3231879472732544, "learning_rate": 0.00011305261922200519, "loss": 0.3075, "step": 3280 }, { "epoch": 5.128805620608899, "grad_norm": 0.33695080876350403, "learning_rate": 0.00011278219637784861, "loss": 0.316, "step": 3285 }, { "epoch": 5.136612021857924, "grad_norm": 0.34212931990623474, "learning_rate": 0.00011251167847321193, "loss": 0.3158, "step": 3290 }, { "epoch": 5.144418423106948, "grad_norm": 0.3377404808998108, "learning_rate": 0.00011224106751992163, "loss": 0.317, "step": 3295 }, { "epoch": 5.152224824355972, "grad_norm": 0.34452709555625916, "learning_rate": 0.00011197036553049625, "loss": 0.3228, "step": 3300 }, { "epoch": 5.160031225604996, "grad_norm": 0.3356159031391144, "learning_rate": 0.00011169957451813138, "loss": 0.3124, "step": 3305 }, { "epoch": 5.167837626854021, "grad_norm": 0.3585182726383209, "learning_rate": 0.00011142869649668466, "loss": 0.3168, "step": 3310 }, { "epoch": 5.1756440281030445, "grad_norm": 0.3430497348308563, "learning_rate": 0.00011115773348066074, "loss": 0.3178, "step": 3315 }, { "epoch": 5.183450429352069, "grad_norm": 0.3532644808292389, "learning_rate": 0.00011088668748519647, "loss": 0.3192, "step": 3320 }, { "epoch": 5.191256830601093, "grad_norm": 0.34457042813301086, "learning_rate": 0.00011061556052604578, "loss": 0.3181, "step": 3325 }, { "epoch": 5.199063231850117, "grad_norm": 0.33349376916885376, "learning_rate": 0.00011034435461956466, "loss": 0.3121, "step": 3330 }, { "epoch": 5.206869633099141, "grad_norm": 0.3395507335662842, "learning_rate": 0.00011007307178269632, "loss": 0.3183, "step": 3335 }, { "epoch": 5.214676034348165, "grad_norm": 0.3544360399246216, "learning_rate": 0.0001098017140329561, "loss": 0.3243, "step": 3340 }, { "epoch": 5.22248243559719, "grad_norm": 0.3494921326637268, "learning_rate": 0.0001095302833884163, "loss": 0.3146, "step": 3345 }, { "epoch": 5.2302888368462135, "grad_norm": 0.33652257919311523, "learning_rate": 0.00010925878186769158, "loss": 0.3135, "step": 3350 }, { "epoch": 5.238095238095238, "grad_norm": 0.3407340943813324, "learning_rate": 0.00010898721148992351, "loss": 0.3205, "step": 3355 }, { "epoch": 5.245901639344262, "grad_norm": 0.3510075509548187, "learning_rate": 0.00010871557427476583, "loss": 0.3127, "step": 3360 }, { "epoch": 5.2537080405932866, "grad_norm": 0.34021666646003723, "learning_rate": 0.00010844387224236932, "loss": 0.3192, "step": 3365 }, { "epoch": 5.26151444184231, "grad_norm": 0.32510459423065186, "learning_rate": 0.00010817210741336684, "loss": 0.3108, "step": 3370 }, { "epoch": 5.269320843091335, "grad_norm": 0.36883097887039185, "learning_rate": 0.00010790028180885821, "loss": 0.3212, "step": 3375 }, { "epoch": 5.277127244340359, "grad_norm": 0.3306715190410614, "learning_rate": 0.00010762839745039526, "loss": 0.325, "step": 3380 }, { "epoch": 5.284933645589383, "grad_norm": 0.3390478491783142, "learning_rate": 0.00010735645635996676, "loss": 0.3159, "step": 3385 }, { "epoch": 5.292740046838407, "grad_norm": 0.32592055201530457, "learning_rate": 0.00010708446055998343, "loss": 0.3127, "step": 3390 }, { "epoch": 5.300546448087432, "grad_norm": 0.3241756856441498, "learning_rate": 0.00010681241207326282, "loss": 0.3087, "step": 3395 }, { "epoch": 5.308352849336456, "grad_norm": 0.3321205675601959, "learning_rate": 0.00010654031292301432, "loss": 0.3191, "step": 3400 }, { "epoch": 5.31615925058548, "grad_norm": 0.3427116572856903, "learning_rate": 0.00010626816513282412, "loss": 0.3167, "step": 3405 }, { "epoch": 5.323965651834504, "grad_norm": 0.33668723702430725, "learning_rate": 0.00010599597072664012, "loss": 0.3173, "step": 3410 }, { "epoch": 5.331772053083529, "grad_norm": 0.3475385010242462, "learning_rate": 0.0001057237317287569, "loss": 0.3211, "step": 3415 }, { "epoch": 5.339578454332552, "grad_norm": 0.3414543867111206, "learning_rate": 0.00010545145016380065, "loss": 0.3171, "step": 3420 }, { "epoch": 5.347384855581577, "grad_norm": 0.37167057394981384, "learning_rate": 0.00010517912805671419, "loss": 0.3199, "step": 3425 }, { "epoch": 5.355191256830601, "grad_norm": 0.3445757031440735, "learning_rate": 0.00010490676743274181, "loss": 0.3176, "step": 3430 }, { "epoch": 5.3629976580796255, "grad_norm": 0.33130553364753723, "learning_rate": 0.00010463437031741423, "loss": 0.3171, "step": 3435 }, { "epoch": 5.370804059328649, "grad_norm": 0.34514832496643066, "learning_rate": 0.00010436193873653361, "loss": 0.3174, "step": 3440 }, { "epoch": 5.378610460577674, "grad_norm": 0.3419683575630188, "learning_rate": 0.00010408947471615835, "loss": 0.3229, "step": 3445 }, { "epoch": 5.386416861826698, "grad_norm": 0.3344433903694153, "learning_rate": 0.00010381698028258817, "loss": 0.3238, "step": 3450 }, { "epoch": 5.394223263075722, "grad_norm": 0.3199335038661957, "learning_rate": 0.00010354445746234894, "loss": 0.3218, "step": 3455 }, { "epoch": 5.402029664324746, "grad_norm": 0.3573337197303772, "learning_rate": 0.00010327190828217763, "loss": 0.3274, "step": 3460 }, { "epoch": 5.409836065573771, "grad_norm": 0.3186776638031006, "learning_rate": 0.00010299933476900727, "loss": 0.3278, "step": 3465 }, { "epoch": 5.4176424668227945, "grad_norm": 0.3291524648666382, "learning_rate": 0.00010272673894995187, "loss": 0.3207, "step": 3470 }, { "epoch": 5.425448868071819, "grad_norm": 0.3383654057979584, "learning_rate": 0.00010245412285229124, "loss": 0.309, "step": 3475 }, { "epoch": 5.433255269320843, "grad_norm": 0.33380231261253357, "learning_rate": 0.00010218148850345613, "loss": 0.3139, "step": 3480 }, { "epoch": 5.441061670569868, "grad_norm": 0.3124459385871887, "learning_rate": 0.0001019088379310129, "loss": 0.3167, "step": 3485 }, { "epoch": 5.448868071818891, "grad_norm": 0.34139421582221985, "learning_rate": 0.00010163617316264869, "loss": 0.3177, "step": 3490 }, { "epoch": 5.456674473067916, "grad_norm": 0.33428099751472473, "learning_rate": 0.00010136349622615608, "loss": 0.3185, "step": 3495 }, { "epoch": 5.46448087431694, "grad_norm": 0.3460736870765686, "learning_rate": 0.00010109080914941824, "loss": 0.3187, "step": 3500 }, { "epoch": 5.472287275565964, "grad_norm": 0.35471439361572266, "learning_rate": 0.00010081811396039373, "loss": 0.3213, "step": 3505 }, { "epoch": 5.480093676814988, "grad_norm": 0.3312179446220398, "learning_rate": 0.00010054541268710138, "loss": 0.3148, "step": 3510 }, { "epoch": 5.487900078064013, "grad_norm": 0.34200319647789, "learning_rate": 0.00010027270735760541, "loss": 0.3268, "step": 3515 }, { "epoch": 5.495706479313037, "grad_norm": 0.339599072933197, "learning_rate": 0.0001, "loss": 0.3148, "step": 3520 }, { "epoch": 5.503512880562061, "grad_norm": 0.3320743143558502, "learning_rate": 9.972729264239461e-05, "loss": 0.3176, "step": 3525 }, { "epoch": 5.511319281811085, "grad_norm": 0.3287266790866852, "learning_rate": 9.945458731289862e-05, "loss": 0.3235, "step": 3530 }, { "epoch": 5.51912568306011, "grad_norm": 0.3429989814758301, "learning_rate": 9.918188603960632e-05, "loss": 0.3257, "step": 3535 }, { "epoch": 5.526932084309133, "grad_norm": 0.3366888165473938, "learning_rate": 9.890919085058178e-05, "loss": 0.3194, "step": 3540 }, { "epoch": 5.534738485558158, "grad_norm": 0.3251045346260071, "learning_rate": 9.863650377384395e-05, "loss": 0.315, "step": 3545 }, { "epoch": 5.542544886807182, "grad_norm": 0.3391668200492859, "learning_rate": 9.836382683735132e-05, "loss": 0.3225, "step": 3550 }, { "epoch": 5.550351288056206, "grad_norm": 0.3379986584186554, "learning_rate": 9.809116206898712e-05, "loss": 0.3251, "step": 3555 }, { "epoch": 5.55815768930523, "grad_norm": 0.336603581905365, "learning_rate": 9.78185114965439e-05, "loss": 0.3211, "step": 3560 }, { "epoch": 5.565964090554255, "grad_norm": 0.3426755368709564, "learning_rate": 9.754587714770878e-05, "loss": 0.323, "step": 3565 }, { "epoch": 5.573770491803279, "grad_norm": 0.3463711738586426, "learning_rate": 9.727326105004817e-05, "loss": 0.3202, "step": 3570 }, { "epoch": 5.581576893052302, "grad_norm": 0.36325374245643616, "learning_rate": 9.700066523099273e-05, "loss": 0.3143, "step": 3575 }, { "epoch": 5.589383294301327, "grad_norm": 0.32558673620224, "learning_rate": 9.67280917178224e-05, "loss": 0.3286, "step": 3580 }, { "epoch": 5.597189695550352, "grad_norm": 0.33436307311058044, "learning_rate": 9.64555425376511e-05, "loss": 0.3168, "step": 3585 }, { "epoch": 5.6049960967993755, "grad_norm": 0.33413755893707275, "learning_rate": 9.618301971741184e-05, "loss": 0.3135, "step": 3590 }, { "epoch": 5.612802498048399, "grad_norm": 0.3465372920036316, "learning_rate": 9.591052528384167e-05, "loss": 0.3175, "step": 3595 }, { "epoch": 5.620608899297424, "grad_norm": 0.3320391774177551, "learning_rate": 9.563806126346642e-05, "loss": 0.3201, "step": 3600 }, { "epoch": 5.628415300546449, "grad_norm": 0.3287176191806793, "learning_rate": 9.536562968258578e-05, "loss": 0.3224, "step": 3605 }, { "epoch": 5.636221701795472, "grad_norm": 0.3486092686653137, "learning_rate": 9.509323256725821e-05, "loss": 0.3254, "step": 3610 }, { "epoch": 5.644028103044496, "grad_norm": 0.3497357666492462, "learning_rate": 9.482087194328583e-05, "loss": 0.3178, "step": 3615 }, { "epoch": 5.651834504293521, "grad_norm": 0.338351309299469, "learning_rate": 9.454854983619936e-05, "loss": 0.3312, "step": 3620 }, { "epoch": 5.6596409055425445, "grad_norm": 0.3393017649650574, "learning_rate": 9.427626827124317e-05, "loss": 0.3171, "step": 3625 }, { "epoch": 5.667447306791569, "grad_norm": 0.34753820300102234, "learning_rate": 9.400402927335992e-05, "loss": 0.3231, "step": 3630 }, { "epoch": 5.675253708040593, "grad_norm": 0.3426099121570587, "learning_rate": 9.373183486717592e-05, "loss": 0.3195, "step": 3635 }, { "epoch": 5.683060109289618, "grad_norm": 0.3154798150062561, "learning_rate": 9.345968707698569e-05, "loss": 0.3234, "step": 3640 }, { "epoch": 5.690866510538641, "grad_norm": 0.3466772735118866, "learning_rate": 9.318758792673722e-05, "loss": 0.3117, "step": 3645 }, { "epoch": 5.698672911787666, "grad_norm": 0.3402402997016907, "learning_rate": 9.29155394400166e-05, "loss": 0.3197, "step": 3650 }, { "epoch": 5.70647931303669, "grad_norm": 0.34711048007011414, "learning_rate": 9.264354364003327e-05, "loss": 0.3238, "step": 3655 }, { "epoch": 5.714285714285714, "grad_norm": 0.33199217915534973, "learning_rate": 9.237160254960476e-05, "loss": 0.3136, "step": 3660 }, { "epoch": 5.722092115534738, "grad_norm": 0.3648219108581543, "learning_rate": 9.209971819114181e-05, "loss": 0.3257, "step": 3665 }, { "epoch": 5.729898516783763, "grad_norm": 0.3323144316673279, "learning_rate": 9.182789258663321e-05, "loss": 0.3162, "step": 3670 }, { "epoch": 5.737704918032787, "grad_norm": 0.32556962966918945, "learning_rate": 9.155612775763069e-05, "loss": 0.3191, "step": 3675 }, { "epoch": 5.745511319281811, "grad_norm": 0.33691149950027466, "learning_rate": 9.128442572523417e-05, "loss": 0.3221, "step": 3680 }, { "epoch": 5.753317720530835, "grad_norm": 0.33508819341659546, "learning_rate": 9.10127885100765e-05, "loss": 0.3215, "step": 3685 }, { "epoch": 5.76112412177986, "grad_norm": 0.31548982858657837, "learning_rate": 9.074121813230847e-05, "loss": 0.3134, "step": 3690 }, { "epoch": 5.768930523028883, "grad_norm": 0.33069273829460144, "learning_rate": 9.046971661158372e-05, "loss": 0.3138, "step": 3695 }, { "epoch": 5.776736924277908, "grad_norm": 0.3511803150177002, "learning_rate": 9.019828596704394e-05, "loss": 0.322, "step": 3700 }, { "epoch": 5.784543325526932, "grad_norm": 0.34525299072265625, "learning_rate": 8.992692821730368e-05, "loss": 0.3268, "step": 3705 }, { "epoch": 5.7923497267759565, "grad_norm": 0.32578587532043457, "learning_rate": 8.965564538043535e-05, "loss": 0.3254, "step": 3710 }, { "epoch": 5.80015612802498, "grad_norm": 0.33581212162971497, "learning_rate": 8.938443947395427e-05, "loss": 0.3249, "step": 3715 }, { "epoch": 5.807962529274005, "grad_norm": 0.34412479400634766, "learning_rate": 8.911331251480357e-05, "loss": 0.3263, "step": 3720 }, { "epoch": 5.815768930523029, "grad_norm": 0.3513209819793701, "learning_rate": 8.884226651933927e-05, "loss": 0.3199, "step": 3725 }, { "epoch": 5.823575331772053, "grad_norm": 0.3510242700576782, "learning_rate": 8.857130350331535e-05, "loss": 0.3156, "step": 3730 }, { "epoch": 5.831381733021077, "grad_norm": 0.3359852135181427, "learning_rate": 8.830042548186863e-05, "loss": 0.3225, "step": 3735 }, { "epoch": 5.839188134270102, "grad_norm": 0.3355961740016937, "learning_rate": 8.802963446950377e-05, "loss": 0.3229, "step": 3740 }, { "epoch": 5.8469945355191255, "grad_norm": 0.3400302827358246, "learning_rate": 8.775893248007839e-05, "loss": 0.331, "step": 3745 }, { "epoch": 5.85480093676815, "grad_norm": 0.3256356120109558, "learning_rate": 8.74883215267881e-05, "loss": 0.327, "step": 3750 }, { "epoch": 5.862607338017174, "grad_norm": 0.35503053665161133, "learning_rate": 8.721780362215138e-05, "loss": 0.3273, "step": 3755 }, { "epoch": 5.870413739266199, "grad_norm": 0.34520792961120605, "learning_rate": 8.694738077799488e-05, "loss": 0.3198, "step": 3760 }, { "epoch": 5.878220140515222, "grad_norm": 0.32543712854385376, "learning_rate": 8.667705500543803e-05, "loss": 0.3171, "step": 3765 }, { "epoch": 5.886026541764247, "grad_norm": 0.33950576186180115, "learning_rate": 8.64068283148786e-05, "loss": 0.3207, "step": 3770 }, { "epoch": 5.893832943013271, "grad_norm": 0.3403869569301605, "learning_rate": 8.613670271597733e-05, "loss": 0.3243, "step": 3775 }, { "epoch": 5.901639344262295, "grad_norm": 0.34109991788864136, "learning_rate": 8.586668021764329e-05, "loss": 0.3221, "step": 3780 }, { "epoch": 5.909445745511319, "grad_norm": 0.3320345878601074, "learning_rate": 8.559676282801858e-05, "loss": 0.3171, "step": 3785 }, { "epoch": 5.917252146760344, "grad_norm": 0.34376391768455505, "learning_rate": 8.532695255446383e-05, "loss": 0.3191, "step": 3790 }, { "epoch": 5.925058548009368, "grad_norm": 0.33876165747642517, "learning_rate": 8.505725140354294e-05, "loss": 0.3198, "step": 3795 }, { "epoch": 5.932864949258392, "grad_norm": 0.34862759709358215, "learning_rate": 8.478766138100834e-05, "loss": 0.3195, "step": 3800 }, { "epoch": 5.940671350507416, "grad_norm": 0.34864258766174316, "learning_rate": 8.451818449178591e-05, "loss": 0.3271, "step": 3805 }, { "epoch": 5.948477751756441, "grad_norm": 0.35281243920326233, "learning_rate": 8.424882273996024e-05, "loss": 0.3246, "step": 3810 }, { "epoch": 5.956284153005464, "grad_norm": 0.33554619550704956, "learning_rate": 8.397957812875959e-05, "loss": 0.3239, "step": 3815 }, { "epoch": 5.964090554254488, "grad_norm": 0.3569699227809906, "learning_rate": 8.371045266054114e-05, "loss": 0.3182, "step": 3820 }, { "epoch": 5.971896955503513, "grad_norm": 0.3249468505382538, "learning_rate": 8.344144833677594e-05, "loss": 0.3218, "step": 3825 }, { "epoch": 5.9797033567525375, "grad_norm": 0.3645569682121277, "learning_rate": 8.317256715803407e-05, "loss": 0.3158, "step": 3830 }, { "epoch": 5.987509758001561, "grad_norm": 0.32271501421928406, "learning_rate": 8.290381112396987e-05, "loss": 0.3194, "step": 3835 }, { "epoch": 5.995316159250585, "grad_norm": 0.3314792215824127, "learning_rate": 8.263518223330697e-05, "loss": 0.318, "step": 3840 }, { "epoch": 6.0, "eval_loss": 1.0760587453842163, "eval_runtime": 0.3225, "eval_samples_per_second": 15.506, "eval_steps_per_second": 3.101, "step": 3843 }, { "epoch": 6.00312256049961, "grad_norm": 0.31583714485168457, "learning_rate": 8.236668248382337e-05, "loss": 0.3, "step": 3845 }, { "epoch": 6.0109289617486334, "grad_norm": 0.36563432216644287, "learning_rate": 8.209831387233676e-05, "loss": 0.2928, "step": 3850 }, { "epoch": 6.018735362997658, "grad_norm": 0.3586893081665039, "learning_rate": 8.183007839468942e-05, "loss": 0.2947, "step": 3855 }, { "epoch": 6.026541764246682, "grad_norm": 0.351180762052536, "learning_rate": 8.156197804573366e-05, "loss": 0.2974, "step": 3860 }, { "epoch": 6.0343481654957065, "grad_norm": 0.3694261610507965, "learning_rate": 8.12940148193168e-05, "loss": 0.2943, "step": 3865 }, { "epoch": 6.04215456674473, "grad_norm": 0.3579952120780945, "learning_rate": 8.102619070826639e-05, "loss": 0.2993, "step": 3870 }, { "epoch": 6.049960967993755, "grad_norm": 0.3623141348361969, "learning_rate": 8.075850770437534e-05, "loss": 0.3012, "step": 3875 }, { "epoch": 6.057767369242779, "grad_norm": 0.35049453377723694, "learning_rate": 8.049096779838719e-05, "loss": 0.2954, "step": 3880 }, { "epoch": 6.065573770491803, "grad_norm": 0.36742958426475525, "learning_rate": 8.022357297998127e-05, "loss": 0.292, "step": 3885 }, { "epoch": 6.073380171740827, "grad_norm": 0.36501386761665344, "learning_rate": 7.995632523775795e-05, "loss": 0.2938, "step": 3890 }, { "epoch": 6.081186572989852, "grad_norm": 0.35817086696624756, "learning_rate": 7.968922655922374e-05, "loss": 0.302, "step": 3895 }, { "epoch": 6.0889929742388755, "grad_norm": 0.3561493158340454, "learning_rate": 7.942227893077652e-05, "loss": 0.3042, "step": 3900 }, { "epoch": 6.0967993754879, "grad_norm": 0.3571511507034302, "learning_rate": 7.915548433769095e-05, "loss": 0.2971, "step": 3905 }, { "epoch": 6.104605776736924, "grad_norm": 0.34343209862709045, "learning_rate": 7.888884476410348e-05, "loss": 0.2956, "step": 3910 }, { "epoch": 6.112412177985949, "grad_norm": 0.3509383201599121, "learning_rate": 7.862236219299777e-05, "loss": 0.2991, "step": 3915 }, { "epoch": 6.120218579234972, "grad_norm": 0.35262179374694824, "learning_rate": 7.835603860618972e-05, "loss": 0.2984, "step": 3920 }, { "epoch": 6.128024980483997, "grad_norm": 0.35837092995643616, "learning_rate": 7.808987598431303e-05, "loss": 0.2949, "step": 3925 }, { "epoch": 6.135831381733021, "grad_norm": 0.3698333501815796, "learning_rate": 7.782387630680421e-05, "loss": 0.3056, "step": 3930 }, { "epoch": 6.143637782982045, "grad_norm": 0.3528449535369873, "learning_rate": 7.755804155188802e-05, "loss": 0.2861, "step": 3935 }, { "epoch": 6.151444184231069, "grad_norm": 0.3811301290988922, "learning_rate": 7.729237369656269e-05, "loss": 0.3052, "step": 3940 }, { "epoch": 6.159250585480094, "grad_norm": 0.37010228633880615, "learning_rate": 7.702687471658518e-05, "loss": 0.2911, "step": 3945 }, { "epoch": 6.167056986729118, "grad_norm": 0.37986573576927185, "learning_rate": 7.676154658645656e-05, "loss": 0.2997, "step": 3950 }, { "epoch": 6.174863387978142, "grad_norm": 0.34996771812438965, "learning_rate": 7.649639127940735e-05, "loss": 0.3029, "step": 3955 }, { "epoch": 6.182669789227166, "grad_norm": 0.3746223449707031, "learning_rate": 7.623141076738271e-05, "loss": 0.3064, "step": 3960 }, { "epoch": 6.190476190476191, "grad_norm": 0.3640713393688202, "learning_rate": 7.596660702102791e-05, "loss": 0.3013, "step": 3965 }, { "epoch": 6.1982825917252145, "grad_norm": 0.3776625096797943, "learning_rate": 7.570198200967362e-05, "loss": 0.3063, "step": 3970 }, { "epoch": 6.206088992974239, "grad_norm": 0.3669188916683197, "learning_rate": 7.543753770132127e-05, "loss": 0.2982, "step": 3975 }, { "epoch": 6.213895394223263, "grad_norm": 0.37241482734680176, "learning_rate": 7.517327606262836e-05, "loss": 0.3, "step": 3980 }, { "epoch": 6.2217017954722875, "grad_norm": 0.36023199558258057, "learning_rate": 7.490919905889403e-05, "loss": 0.2997, "step": 3985 }, { "epoch": 6.229508196721311, "grad_norm": 0.36012396216392517, "learning_rate": 7.464530865404407e-05, "loss": 0.2926, "step": 3990 }, { "epoch": 6.237314597970336, "grad_norm": 0.3803520202636719, "learning_rate": 7.438160681061676e-05, "loss": 0.3045, "step": 3995 }, { "epoch": 6.24512099921936, "grad_norm": 0.34980347752571106, "learning_rate": 7.411809548974792e-05, "loss": 0.2991, "step": 4000 }, { "epoch": 6.252927400468384, "grad_norm": 0.3561069369316101, "learning_rate": 7.385477665115658e-05, "loss": 0.3094, "step": 4005 }, { "epoch": 6.260733801717408, "grad_norm": 0.36508041620254517, "learning_rate": 7.359165225313019e-05, "loss": 0.2969, "step": 4010 }, { "epoch": 6.268540202966433, "grad_norm": 0.36842647194862366, "learning_rate": 7.332872425251018e-05, "loss": 0.3009, "step": 4015 }, { "epoch": 6.2763466042154565, "grad_norm": 0.3771759569644928, "learning_rate": 7.30659946046774e-05, "loss": 0.3027, "step": 4020 }, { "epoch": 6.284153005464481, "grad_norm": 0.36950060725212097, "learning_rate": 7.280346526353759e-05, "loss": 0.3052, "step": 4025 }, { "epoch": 6.291959406713505, "grad_norm": 0.36776700615882874, "learning_rate": 7.25411381815068e-05, "loss": 0.298, "step": 4030 }, { "epoch": 6.29976580796253, "grad_norm": 0.3699568510055542, "learning_rate": 7.22790153094968e-05, "loss": 0.2961, "step": 4035 }, { "epoch": 6.307572209211553, "grad_norm": 0.36055341362953186, "learning_rate": 7.20170985969008e-05, "loss": 0.297, "step": 4040 }, { "epoch": 6.315378610460578, "grad_norm": 0.37468183040618896, "learning_rate": 7.175538999157876e-05, "loss": 0.2989, "step": 4045 }, { "epoch": 6.323185011709602, "grad_norm": 0.3778867721557617, "learning_rate": 7.149389143984295e-05, "loss": 0.2994, "step": 4050 }, { "epoch": 6.330991412958626, "grad_norm": 0.37846189737319946, "learning_rate": 7.123260488644345e-05, "loss": 0.2949, "step": 4055 }, { "epoch": 6.33879781420765, "grad_norm": 0.3595486581325531, "learning_rate": 7.097153227455379e-05, "loss": 0.3063, "step": 4060 }, { "epoch": 6.346604215456675, "grad_norm": 0.38495534658432007, "learning_rate": 7.071067554575637e-05, "loss": 0.3037, "step": 4065 }, { "epoch": 6.354410616705699, "grad_norm": 0.37678053975105286, "learning_rate": 7.045003664002809e-05, "loss": 0.3111, "step": 4070 }, { "epoch": 6.362217017954723, "grad_norm": 0.353000283241272, "learning_rate": 7.018961749572604e-05, "loss": 0.3003, "step": 4075 }, { "epoch": 6.370023419203747, "grad_norm": 0.381798654794693, "learning_rate": 6.992942004957271e-05, "loss": 0.3037, "step": 4080 }, { "epoch": 6.377829820452771, "grad_norm": 0.34018296003341675, "learning_rate": 6.9669446236642e-05, "loss": 0.2956, "step": 4085 }, { "epoch": 6.3856362217017955, "grad_norm": 0.3533703684806824, "learning_rate": 6.940969799034465e-05, "loss": 0.298, "step": 4090 }, { "epoch": 6.39344262295082, "grad_norm": 0.36671626567840576, "learning_rate": 6.915017724241389e-05, "loss": 0.3004, "step": 4095 }, { "epoch": 6.401249024199844, "grad_norm": 0.3814579248428345, "learning_rate": 6.889088592289093e-05, "loss": 0.2992, "step": 4100 }, { "epoch": 6.409055425448868, "grad_norm": 0.38459497690200806, "learning_rate": 6.863182596011087e-05, "loss": 0.3036, "step": 4105 }, { "epoch": 6.416861826697892, "grad_norm": 0.3617273271083832, "learning_rate": 6.837299928068817e-05, "loss": 0.3037, "step": 4110 }, { "epoch": 6.424668227946917, "grad_norm": 0.3558304011821747, "learning_rate": 6.811440780950237e-05, "loss": 0.2934, "step": 4115 }, { "epoch": 6.432474629195941, "grad_norm": 0.36529627442359924, "learning_rate": 6.785605346968386e-05, "loss": 0.3034, "step": 4120 }, { "epoch": 6.4402810304449645, "grad_norm": 0.3953593373298645, "learning_rate": 6.759793818259933e-05, "loss": 0.3005, "step": 4125 }, { "epoch": 6.448087431693989, "grad_norm": 0.38395586609840393, "learning_rate": 6.73400638678378e-05, "loss": 0.3009, "step": 4130 }, { "epoch": 6.455893832943013, "grad_norm": 0.35242578387260437, "learning_rate": 6.708243244319611e-05, "loss": 0.3053, "step": 4135 }, { "epoch": 6.4637002341920375, "grad_norm": 0.3590448200702667, "learning_rate": 6.682504582466482e-05, "loss": 0.2994, "step": 4140 }, { "epoch": 6.471506635441061, "grad_norm": 0.37192437052726746, "learning_rate": 6.656790592641375e-05, "loss": 0.3056, "step": 4145 }, { "epoch": 6.479313036690086, "grad_norm": 0.35877934098243713, "learning_rate": 6.6311014660778e-05, "loss": 0.2958, "step": 4150 }, { "epoch": 6.48711943793911, "grad_norm": 0.3730364143848419, "learning_rate": 6.605437393824356e-05, "loss": 0.3047, "step": 4155 }, { "epoch": 6.494925839188134, "grad_norm": 0.3790695071220398, "learning_rate": 6.579798566743314e-05, "loss": 0.3027, "step": 4160 }, { "epoch": 6.502732240437158, "grad_norm": 0.36754992604255676, "learning_rate": 6.554185175509197e-05, "loss": 0.2916, "step": 4165 }, { "epoch": 6.510538641686183, "grad_norm": 0.37978771328926086, "learning_rate": 6.528597410607364e-05, "loss": 0.3023, "step": 4170 }, { "epoch": 6.518345042935207, "grad_norm": 0.35783836245536804, "learning_rate": 6.503035462332592e-05, "loss": 0.3031, "step": 4175 }, { "epoch": 6.526151444184231, "grad_norm": 0.3850097060203552, "learning_rate": 6.477499520787665e-05, "loss": 0.3025, "step": 4180 }, { "epoch": 6.533957845433255, "grad_norm": 0.38699930906295776, "learning_rate": 6.451989775881955e-05, "loss": 0.3032, "step": 4185 }, { "epoch": 6.54176424668228, "grad_norm": 0.3620811998844147, "learning_rate": 6.42650641733e-05, "loss": 0.3029, "step": 4190 }, { "epoch": 6.549570647931303, "grad_norm": 0.3833983838558197, "learning_rate": 6.401049634650118e-05, "loss": 0.3018, "step": 4195 }, { "epoch": 6.557377049180328, "grad_norm": 0.3642504811286926, "learning_rate": 6.375619617162985e-05, "loss": 0.3052, "step": 4200 }, { "epoch": 6.565183450429352, "grad_norm": 0.34987977147102356, "learning_rate": 6.350216553990212e-05, "loss": 0.3027, "step": 4205 }, { "epoch": 6.5729898516783765, "grad_norm": 0.3789624571800232, "learning_rate": 6.324840634052967e-05, "loss": 0.2983, "step": 4210 }, { "epoch": 6.5807962529274, "grad_norm": 0.39051589369773865, "learning_rate": 6.29949204607054e-05, "loss": 0.3099, "step": 4215 }, { "epoch": 6.588602654176425, "grad_norm": 0.36654597520828247, "learning_rate": 6.27417097855897e-05, "loss": 0.3067, "step": 4220 }, { "epoch": 6.596409055425449, "grad_norm": 0.3750831186771393, "learning_rate": 6.248877619829619e-05, "loss": 0.304, "step": 4225 }, { "epoch": 6.604215456674473, "grad_norm": 0.3867753744125366, "learning_rate": 6.223612157987786e-05, "loss": 0.3082, "step": 4230 }, { "epoch": 6.612021857923497, "grad_norm": 0.3690248429775238, "learning_rate": 6.198374780931293e-05, "loss": 0.3044, "step": 4235 }, { "epoch": 6.619828259172522, "grad_norm": 0.37476563453674316, "learning_rate": 6.173165676349103e-05, "loss": 0.3027, "step": 4240 }, { "epoch": 6.6276346604215455, "grad_norm": 0.36592650413513184, "learning_rate": 6.14798503171992e-05, "loss": 0.3015, "step": 4245 }, { "epoch": 6.63544106167057, "grad_norm": 0.3869721293449402, "learning_rate": 6.122833034310793e-05, "loss": 0.2986, "step": 4250 }, { "epoch": 6.643247462919594, "grad_norm": 0.3800086975097656, "learning_rate": 6.097709871175723e-05, "loss": 0.3026, "step": 4255 }, { "epoch": 6.6510538641686185, "grad_norm": 0.3717212677001953, "learning_rate": 6.0726157291542605e-05, "loss": 0.303, "step": 4260 }, { "epoch": 6.658860265417642, "grad_norm": 0.35861897468566895, "learning_rate": 6.047550794870145e-05, "loss": 0.3008, "step": 4265 }, { "epoch": 6.666666666666667, "grad_norm": 0.3755732476711273, "learning_rate": 6.02251525472989e-05, "loss": 0.3006, "step": 4270 }, { "epoch": 6.674473067915691, "grad_norm": 0.3563028872013092, "learning_rate": 5.9975092949214116e-05, "loss": 0.3041, "step": 4275 }, { "epoch": 6.682279469164715, "grad_norm": 0.3711301386356354, "learning_rate": 5.9725331014126294e-05, "loss": 0.3047, "step": 4280 }, { "epoch": 6.690085870413739, "grad_norm": 0.3619491457939148, "learning_rate": 5.947586859950103e-05, "loss": 0.3034, "step": 4285 }, { "epoch": 6.697892271662764, "grad_norm": 0.3621855080127716, "learning_rate": 5.922670756057633e-05, "loss": 0.2993, "step": 4290 }, { "epoch": 6.705698672911788, "grad_norm": 0.3833586573600769, "learning_rate": 5.8977849750348944e-05, "loss": 0.2969, "step": 4295 }, { "epoch": 6.713505074160812, "grad_norm": 0.36535149812698364, "learning_rate": 5.872929701956054e-05, "loss": 0.3109, "step": 4300 }, { "epoch": 6.721311475409836, "grad_norm": 0.34852883219718933, "learning_rate": 5.848105121668381e-05, "loss": 0.2996, "step": 4305 }, { "epoch": 6.729117876658861, "grad_norm": 0.35054901242256165, "learning_rate": 5.8233114187908935e-05, "loss": 0.3038, "step": 4310 }, { "epoch": 6.736924277907884, "grad_norm": 0.38017308712005615, "learning_rate": 5.7985487777129765e-05, "loss": 0.3023, "step": 4315 }, { "epoch": 6.744730679156909, "grad_norm": 0.3646107017993927, "learning_rate": 5.773817382593008e-05, "loss": 0.3097, "step": 4320 }, { "epoch": 6.752537080405933, "grad_norm": 0.3594595491886139, "learning_rate": 5.749117417356988e-05, "loss": 0.3088, "step": 4325 }, { "epoch": 6.760343481654957, "grad_norm": 0.39560645818710327, "learning_rate": 5.7244490656971815e-05, "loss": 0.3058, "step": 4330 }, { "epoch": 6.768149882903981, "grad_norm": 0.3735295236110687, "learning_rate": 5.699812511070734e-05, "loss": 0.3109, "step": 4335 }, { "epoch": 6.775956284153006, "grad_norm": 0.3793189823627472, "learning_rate": 5.675207936698337e-05, "loss": 0.2962, "step": 4340 }, { "epoch": 6.78376268540203, "grad_norm": 0.36545222997665405, "learning_rate": 5.6506355255628284e-05, "loss": 0.3034, "step": 4345 }, { "epoch": 6.791569086651053, "grad_norm": 0.3819272816181183, "learning_rate": 5.6260954604078585e-05, "loss": 0.3055, "step": 4350 }, { "epoch": 6.799375487900078, "grad_norm": 0.40963688492774963, "learning_rate": 5.601587923736518e-05, "loss": 0.297, "step": 4355 }, { "epoch": 6.807181889149103, "grad_norm": 0.3582998812198639, "learning_rate": 5.577113097809989e-05, "loss": 0.3024, "step": 4360 }, { "epoch": 6.8149882903981265, "grad_norm": 0.3594261407852173, "learning_rate": 5.5526711646461835e-05, "loss": 0.2976, "step": 4365 }, { "epoch": 6.82279469164715, "grad_norm": 0.36158230900764465, "learning_rate": 5.528262306018395e-05, "loss": 0.3015, "step": 4370 }, { "epoch": 6.830601092896175, "grad_norm": 0.37753212451934814, "learning_rate": 5.503886703453933e-05, "loss": 0.2942, "step": 4375 }, { "epoch": 6.8384074941451995, "grad_norm": 0.37955158948898315, "learning_rate": 5.4795445382328037e-05, "loss": 0.3017, "step": 4380 }, { "epoch": 6.846213895394223, "grad_norm": 0.3600054979324341, "learning_rate": 5.4552359913863214e-05, "loss": 0.3061, "step": 4385 }, { "epoch": 6.854020296643247, "grad_norm": 0.3673628270626068, "learning_rate": 5.4309612436957937e-05, "loss": 0.3049, "step": 4390 }, { "epoch": 6.861826697892272, "grad_norm": 0.36428987979888916, "learning_rate": 5.4067204756911624e-05, "loss": 0.3096, "step": 4395 }, { "epoch": 6.8696330991412955, "grad_norm": 0.3699190616607666, "learning_rate": 5.382513867649663e-05, "loss": 0.3008, "step": 4400 }, { "epoch": 6.87743950039032, "grad_norm": 0.36936306953430176, "learning_rate": 5.358341599594483e-05, "loss": 0.2979, "step": 4405 }, { "epoch": 6.885245901639344, "grad_norm": 0.3822793960571289, "learning_rate": 5.3342038512934424e-05, "loss": 0.3009, "step": 4410 }, { "epoch": 6.893052302888369, "grad_norm": 0.37353184819221497, "learning_rate": 5.3101008022576126e-05, "loss": 0.3061, "step": 4415 }, { "epoch": 6.900858704137392, "grad_norm": 0.37411707639694214, "learning_rate": 5.286032631740023e-05, "loss": 0.3022, "step": 4420 }, { "epoch": 6.908665105386417, "grad_norm": 0.3575087785720825, "learning_rate": 5.261999518734322e-05, "loss": 0.302, "step": 4425 }, { "epoch": 6.916471506635441, "grad_norm": 0.3561764359474182, "learning_rate": 5.238001641973422e-05, "loss": 0.2998, "step": 4430 }, { "epoch": 6.924277907884465, "grad_norm": 0.3756863474845886, "learning_rate": 5.214039179928194e-05, "loss": 0.3009, "step": 4435 }, { "epoch": 6.932084309133489, "grad_norm": 0.36556264758110046, "learning_rate": 5.190112310806126e-05, "loss": 0.296, "step": 4440 }, { "epoch": 6.939890710382514, "grad_norm": 0.37201789021492004, "learning_rate": 5.1662212125500075e-05, "loss": 0.2993, "step": 4445 }, { "epoch": 6.947697111631538, "grad_norm": 0.34447523951530457, "learning_rate": 5.142366062836599e-05, "loss": 0.3002, "step": 4450 }, { "epoch": 6.955503512880562, "grad_norm": 0.3652222156524658, "learning_rate": 5.118547039075326e-05, "loss": 0.3015, "step": 4455 }, { "epoch": 6.963309914129586, "grad_norm": 0.3540942072868347, "learning_rate": 5.09476431840692e-05, "loss": 0.3031, "step": 4460 }, { "epoch": 6.971116315378611, "grad_norm": 0.36961525678634644, "learning_rate": 5.071018077702161e-05, "loss": 0.3134, "step": 4465 }, { "epoch": 6.978922716627634, "grad_norm": 0.3560226559638977, "learning_rate": 5.047308493560506e-05, "loss": 0.2953, "step": 4470 }, { "epoch": 6.986729117876659, "grad_norm": 0.3567667007446289, "learning_rate": 5.023635742308807e-05, "loss": 0.3081, "step": 4475 }, { "epoch": 6.994535519125683, "grad_norm": 0.3771991431713104, "learning_rate": 5.000000000000002e-05, "loss": 0.3056, "step": 4480 }, { "epoch": 6.999219359875098, "eval_loss": 1.1023039817810059, "eval_runtime": 0.3388, "eval_samples_per_second": 14.759, "eval_steps_per_second": 2.952, "step": 4483 }, { "epoch": 7.0023419203747075, "grad_norm": 0.34990745782852173, "learning_rate": 4.976401442411774e-05, "loss": 0.2964, "step": 4485 }, { "epoch": 7.010148321623731, "grad_norm": 0.3766123354434967, "learning_rate": 4.952840245045278e-05, "loss": 0.2868, "step": 4490 }, { "epoch": 7.017954722872756, "grad_norm": 0.3829561769962311, "learning_rate": 4.92931658312383e-05, "loss": 0.2865, "step": 4495 }, { "epoch": 7.02576112412178, "grad_norm": 0.3801458179950714, "learning_rate": 4.9058306315915826e-05, "loss": 0.277, "step": 4500 }, { "epoch": 7.033567525370804, "grad_norm": 0.3700219690799713, "learning_rate": 4.882382565112248e-05, "loss": 0.2803, "step": 4505 }, { "epoch": 7.041373926619828, "grad_norm": 0.3695278763771057, "learning_rate": 4.8589725580677835e-05, "loss": 0.2817, "step": 4510 }, { "epoch": 7.049180327868853, "grad_norm": 0.3825577199459076, "learning_rate": 4.835600784557106e-05, "loss": 0.2791, "step": 4515 }, { "epoch": 7.0569867291178765, "grad_norm": 0.36848512291908264, "learning_rate": 4.8122674183947836e-05, "loss": 0.2808, "step": 4520 }, { "epoch": 7.064793130366901, "grad_norm": 0.3669551610946655, "learning_rate": 4.7889726331097686e-05, "loss": 0.2813, "step": 4525 }, { "epoch": 7.072599531615925, "grad_norm": 0.3677949905395508, "learning_rate": 4.7657166019440614e-05, "loss": 0.2761, "step": 4530 }, { "epoch": 7.08040593286495, "grad_norm": 0.39766794443130493, "learning_rate": 4.742499497851478e-05, "loss": 0.2888, "step": 4535 }, { "epoch": 7.088212334113973, "grad_norm": 0.39841169118881226, "learning_rate": 4.7193214934963206e-05, "loss": 0.2874, "step": 4540 }, { "epoch": 7.096018735362998, "grad_norm": 0.37418073415756226, "learning_rate": 4.696182761252112e-05, "loss": 0.2794, "step": 4545 }, { "epoch": 7.103825136612022, "grad_norm": 0.3763929307460785, "learning_rate": 4.6730834732003104e-05, "loss": 0.2851, "step": 4550 }, { "epoch": 7.111631537861046, "grad_norm": 0.39079606533050537, "learning_rate": 4.6500238011290295e-05, "loss": 0.2804, "step": 4555 }, { "epoch": 7.11943793911007, "grad_norm": 0.3997804522514343, "learning_rate": 4.6270039165317605e-05, "loss": 0.2827, "step": 4560 }, { "epoch": 7.127244340359095, "grad_norm": 0.3801449239253998, "learning_rate": 4.604023990606105e-05, "loss": 0.2845, "step": 4565 }, { "epoch": 7.135050741608119, "grad_norm": 0.3795543611049652, "learning_rate": 4.5810841942524864e-05, "loss": 0.2793, "step": 4570 }, { "epoch": 7.142857142857143, "grad_norm": 0.4058801829814911, "learning_rate": 4.5581846980728794e-05, "loss": 0.2844, "step": 4575 }, { "epoch": 7.150663544106167, "grad_norm": 0.38217005133628845, "learning_rate": 4.535325672369567e-05, "loss": 0.2879, "step": 4580 }, { "epoch": 7.158469945355192, "grad_norm": 0.4147985279560089, "learning_rate": 4.512507287143842e-05, "loss": 0.2829, "step": 4585 }, { "epoch": 7.166276346604215, "grad_norm": 0.3910820484161377, "learning_rate": 4.4897297120947624e-05, "loss": 0.2898, "step": 4590 }, { "epoch": 7.17408274785324, "grad_norm": 0.3985479176044464, "learning_rate": 4.466993116617878e-05, "loss": 0.2851, "step": 4595 }, { "epoch": 7.181889149102264, "grad_norm": 0.3941868245601654, "learning_rate": 4.444297669803981e-05, "loss": 0.2828, "step": 4600 }, { "epoch": 7.1896955503512885, "grad_norm": 0.3968215882778168, "learning_rate": 4.4216435404378355e-05, "loss": 0.2917, "step": 4605 }, { "epoch": 7.197501951600312, "grad_norm": 0.3764592707157135, "learning_rate": 4.399030896996945e-05, "loss": 0.2811, "step": 4610 }, { "epoch": 7.205308352849336, "grad_norm": 0.3845047056674957, "learning_rate": 4.3764599076502696e-05, "loss": 0.2765, "step": 4615 }, { "epoch": 7.213114754098361, "grad_norm": 0.39302965998649597, "learning_rate": 4.353930740256996e-05, "loss": 0.2839, "step": 4620 }, { "epoch": 7.220921155347384, "grad_norm": 0.38965940475463867, "learning_rate": 4.331443562365285e-05, "loss": 0.2834, "step": 4625 }, { "epoch": 7.228727556596409, "grad_norm": 0.36734485626220703, "learning_rate": 4.308998541211015e-05, "loss": 0.2846, "step": 4630 }, { "epoch": 7.236533957845433, "grad_norm": 0.41275784373283386, "learning_rate": 4.286595843716569e-05, "loss": 0.2903, "step": 4635 }, { "epoch": 7.2443403590944575, "grad_norm": 0.38159552216529846, "learning_rate": 4.264235636489542e-05, "loss": 0.288, "step": 4640 }, { "epoch": 7.252146760343481, "grad_norm": 0.38839638233184814, "learning_rate": 4.241918085821547e-05, "loss": 0.2831, "step": 4645 }, { "epoch": 7.259953161592506, "grad_norm": 0.42755383253097534, "learning_rate": 4.219643357686967e-05, "loss": 0.2908, "step": 4650 }, { "epoch": 7.26775956284153, "grad_norm": 0.43094560503959656, "learning_rate": 4.19741161774171e-05, "loss": 0.2841, "step": 4655 }, { "epoch": 7.275565964090554, "grad_norm": 0.3952085077762604, "learning_rate": 4.17522303132198e-05, "loss": 0.2858, "step": 4660 }, { "epoch": 7.283372365339578, "grad_norm": 0.401659220457077, "learning_rate": 4.153077763443055e-05, "loss": 0.2801, "step": 4665 }, { "epoch": 7.291178766588603, "grad_norm": 0.3862559199333191, "learning_rate": 4.1309759787980565e-05, "loss": 0.2818, "step": 4670 }, { "epoch": 7.2989851678376265, "grad_norm": 0.39351990818977356, "learning_rate": 4.1089178417567164e-05, "loss": 0.2916, "step": 4675 }, { "epoch": 7.306791569086651, "grad_norm": 0.39333629608154297, "learning_rate": 4.086903516364179e-05, "loss": 0.2885, "step": 4680 }, { "epoch": 7.314597970335675, "grad_norm": 0.40127068758010864, "learning_rate": 4.064933166339737e-05, "loss": 0.2892, "step": 4685 }, { "epoch": 7.3224043715847, "grad_norm": 0.4141682982444763, "learning_rate": 4.0430069550756665e-05, "loss": 0.2823, "step": 4690 }, { "epoch": 7.330210772833723, "grad_norm": 0.39413705468177795, "learning_rate": 4.021125045635973e-05, "loss": 0.2775, "step": 4695 }, { "epoch": 7.338017174082748, "grad_norm": 0.3831061124801636, "learning_rate": 3.999287600755192e-05, "loss": 0.2867, "step": 4700 }, { "epoch": 7.345823575331772, "grad_norm": 0.3949141204357147, "learning_rate": 3.977494782837182e-05, "loss": 0.2895, "step": 4705 }, { "epoch": 7.353629976580796, "grad_norm": 0.41716331243515015, "learning_rate": 3.9557467539539115e-05, "loss": 0.2857, "step": 4710 }, { "epoch": 7.36143637782982, "grad_norm": 0.385775625705719, "learning_rate": 3.9340436758442536e-05, "loss": 0.2819, "step": 4715 }, { "epoch": 7.369242779078845, "grad_norm": 0.40228816866874695, "learning_rate": 3.9123857099127936e-05, "loss": 0.2891, "step": 4720 }, { "epoch": 7.377049180327869, "grad_norm": 0.38272202014923096, "learning_rate": 3.8907730172286124e-05, "loss": 0.284, "step": 4725 }, { "epoch": 7.384855581576893, "grad_norm": 0.39481300115585327, "learning_rate": 3.8692057585240905e-05, "loss": 0.2822, "step": 4730 }, { "epoch": 7.392661982825917, "grad_norm": 0.3933485150337219, "learning_rate": 3.847684094193733e-05, "loss": 0.294, "step": 4735 }, { "epoch": 7.400468384074942, "grad_norm": 0.542770504951477, "learning_rate": 3.826208184292952e-05, "loss": 0.2887, "step": 4740 }, { "epoch": 7.408274785323965, "grad_norm": 0.4088205397129059, "learning_rate": 3.804778188536887e-05, "loss": 0.2891, "step": 4745 }, { "epoch": 7.41608118657299, "grad_norm": 0.3968678116798401, "learning_rate": 3.783394266299228e-05, "loss": 0.2853, "step": 4750 }, { "epoch": 7.423887587822014, "grad_norm": 0.41700446605682373, "learning_rate": 3.7620565766109975e-05, "loss": 0.2903, "step": 4755 }, { "epoch": 7.4316939890710385, "grad_norm": 0.399271160364151, "learning_rate": 3.7407652781594095e-05, "loss": 0.289, "step": 4760 }, { "epoch": 7.439500390320062, "grad_norm": 0.3900463879108429, "learning_rate": 3.719520529286659e-05, "loss": 0.2846, "step": 4765 }, { "epoch": 7.447306791569087, "grad_norm": 0.43415582180023193, "learning_rate": 3.698322487988755e-05, "loss": 0.2898, "step": 4770 }, { "epoch": 7.455113192818111, "grad_norm": 0.38591185212135315, "learning_rate": 3.677171311914346e-05, "loss": 0.2832, "step": 4775 }, { "epoch": 7.462919594067135, "grad_norm": 0.3995843827724457, "learning_rate": 3.6560671583635467e-05, "loss": 0.2904, "step": 4780 }, { "epoch": 7.470725995316159, "grad_norm": 0.4198484718799591, "learning_rate": 3.6350101842867645e-05, "loss": 0.2769, "step": 4785 }, { "epoch": 7.478532396565184, "grad_norm": 0.39973580837249756, "learning_rate": 3.614000546283547e-05, "loss": 0.2867, "step": 4790 }, { "epoch": 7.4863387978142075, "grad_norm": 0.37141865491867065, "learning_rate": 3.593038400601395e-05, "loss": 0.2834, "step": 4795 }, { "epoch": 7.494145199063232, "grad_norm": 0.4066362679004669, "learning_rate": 3.5721239031346066e-05, "loss": 0.2847, "step": 4800 }, { "epoch": 7.501951600312256, "grad_norm": 0.40516340732574463, "learning_rate": 3.55125720942314e-05, "loss": 0.2854, "step": 4805 }, { "epoch": 7.509758001561281, "grad_norm": 0.428204208612442, "learning_rate": 3.530438474651428e-05, "loss": 0.2816, "step": 4810 }, { "epoch": 7.517564402810304, "grad_norm": 0.4114013612270355, "learning_rate": 3.509667853647235e-05, "loss": 0.2783, "step": 4815 }, { "epoch": 7.525370804059329, "grad_norm": 0.4049510955810547, "learning_rate": 3.4889455008805106e-05, "loss": 0.2904, "step": 4820 }, { "epoch": 7.533177205308353, "grad_norm": 0.4051942825317383, "learning_rate": 3.468271570462235e-05, "loss": 0.2893, "step": 4825 }, { "epoch": 7.540983606557377, "grad_norm": 0.4027831554412842, "learning_rate": 3.447646216143268e-05, "loss": 0.2903, "step": 4830 }, { "epoch": 7.548790007806401, "grad_norm": 0.3996686339378357, "learning_rate": 3.427069591313226e-05, "loss": 0.2823, "step": 4835 }, { "epoch": 7.556596409055425, "grad_norm": 0.40199506282806396, "learning_rate": 3.406541848999312e-05, "loss": 0.2817, "step": 4840 }, { "epoch": 7.56440281030445, "grad_norm": 0.4444729685783386, "learning_rate": 3.3860631418652e-05, "loss": 0.2859, "step": 4845 }, { "epoch": 7.572209211553474, "grad_norm": 0.40995243191719055, "learning_rate": 3.365633622209891e-05, "loss": 0.2869, "step": 4850 }, { "epoch": 7.580015612802498, "grad_norm": 0.4094167649745941, "learning_rate": 3.345253441966579e-05, "loss": 0.2862, "step": 4855 }, { "epoch": 7.587822014051522, "grad_norm": 0.38202327489852905, "learning_rate": 3.324922752701528e-05, "loss": 0.2842, "step": 4860 }, { "epoch": 7.595628415300546, "grad_norm": 0.4048561155796051, "learning_rate": 3.3046417056129366e-05, "loss": 0.2829, "step": 4865 }, { "epoch": 7.603434816549571, "grad_norm": 0.3873648941516876, "learning_rate": 3.2844104515298155e-05, "loss": 0.2837, "step": 4870 }, { "epoch": 7.611241217798595, "grad_norm": 0.39025747776031494, "learning_rate": 3.2642291409108775e-05, "loss": 0.2777, "step": 4875 }, { "epoch": 7.619047619047619, "grad_norm": 0.39962899684906006, "learning_rate": 3.244097923843398e-05, "loss": 0.2897, "step": 4880 }, { "epoch": 7.626854020296643, "grad_norm": 0.39887768030166626, "learning_rate": 3.2240169500421135e-05, "loss": 0.2777, "step": 4885 }, { "epoch": 7.634660421545668, "grad_norm": 0.42155277729034424, "learning_rate": 3.2039863688481055e-05, "loss": 0.2882, "step": 4890 }, { "epoch": 7.642466822794692, "grad_norm": 0.40622079372406006, "learning_rate": 3.184006329227684e-05, "loss": 0.292, "step": 4895 }, { "epoch": 7.6502732240437155, "grad_norm": 0.41190096735954285, "learning_rate": 3.164076979771287e-05, "loss": 0.283, "step": 4900 }, { "epoch": 7.65807962529274, "grad_norm": 0.41508612036705017, "learning_rate": 3.144198468692379e-05, "loss": 0.279, "step": 4905 }, { "epoch": 7.665886026541764, "grad_norm": 0.39837446808815, "learning_rate": 3.1243709438263255e-05, "loss": 0.2843, "step": 4910 }, { "epoch": 7.6736924277907885, "grad_norm": 0.4004950523376465, "learning_rate": 3.104594552629331e-05, "loss": 0.2866, "step": 4915 }, { "epoch": 7.681498829039812, "grad_norm": 0.3863111436367035, "learning_rate": 3.0848694421773075e-05, "loss": 0.2803, "step": 4920 }, { "epoch": 7.689305230288837, "grad_norm": 0.39881014823913574, "learning_rate": 3.065195759164797e-05, "loss": 0.2832, "step": 4925 }, { "epoch": 7.697111631537861, "grad_norm": 0.3900108337402344, "learning_rate": 3.0455736499038845e-05, "loss": 0.2955, "step": 4930 }, { "epoch": 7.704918032786885, "grad_norm": 0.4136905372142792, "learning_rate": 3.0260032603230982e-05, "loss": 0.2938, "step": 4935 }, { "epoch": 7.712724434035909, "grad_norm": 0.4050344228744507, "learning_rate": 3.0064847359663284e-05, "loss": 0.2903, "step": 4940 }, { "epoch": 7.720530835284934, "grad_norm": 0.4190680980682373, "learning_rate": 2.9870182219917564e-05, "loss": 0.2944, "step": 4945 }, { "epoch": 7.7283372365339575, "grad_norm": 0.39599907398223877, "learning_rate": 2.9676038631707593e-05, "loss": 0.2911, "step": 4950 }, { "epoch": 7.736143637782982, "grad_norm": 0.39728233218193054, "learning_rate": 2.9482418038868275e-05, "loss": 0.2861, "step": 4955 }, { "epoch": 7.743950039032006, "grad_norm": 0.39015698432922363, "learning_rate": 2.9289321881345254e-05, "loss": 0.2846, "step": 4960 }, { "epoch": 7.751756440281031, "grad_norm": 0.394378662109375, "learning_rate": 2.9096751595183823e-05, "loss": 0.2879, "step": 4965 }, { "epoch": 7.759562841530054, "grad_norm": 0.39528360962867737, "learning_rate": 2.8904708612518404e-05, "loss": 0.2899, "step": 4970 }, { "epoch": 7.767369242779079, "grad_norm": 0.41304799914360046, "learning_rate": 2.8713194361562036e-05, "loss": 0.2898, "step": 4975 }, { "epoch": 7.775175644028103, "grad_norm": 0.4061037003993988, "learning_rate": 2.8522210266595384e-05, "loss": 0.2901, "step": 4980 }, { "epoch": 7.7829820452771274, "grad_norm": 0.42109227180480957, "learning_rate": 2.833175774795651e-05, "loss": 0.2868, "step": 4985 }, { "epoch": 7.790788446526151, "grad_norm": 0.38924074172973633, "learning_rate": 2.814183822203019e-05, "loss": 0.2887, "step": 4990 }, { "epoch": 7.798594847775176, "grad_norm": 0.4080544114112854, "learning_rate": 2.795245310123732e-05, "loss": 0.2884, "step": 4995 }, { "epoch": 7.8064012490242, "grad_norm": 0.39689868688583374, "learning_rate": 2.776360379402445e-05, "loss": 0.2859, "step": 5000 }, { "epoch": 7.814207650273224, "grad_norm": 0.40437766909599304, "learning_rate": 2.7575291704853323e-05, "loss": 0.287, "step": 5005 }, { "epoch": 7.822014051522248, "grad_norm": 0.4085007905960083, "learning_rate": 2.738751823419041e-05, "loss": 0.2875, "step": 5010 }, { "epoch": 7.829820452771273, "grad_norm": 0.3995596468448639, "learning_rate": 2.720028477849652e-05, "loss": 0.2818, "step": 5015 }, { "epoch": 7.8376268540202965, "grad_norm": 0.3908860683441162, "learning_rate": 2.7013592730216465e-05, "loss": 0.283, "step": 5020 }, { "epoch": 7.845433255269321, "grad_norm": 0.39162319898605347, "learning_rate": 2.6827443477768454e-05, "loss": 0.2833, "step": 5025 }, { "epoch": 7.853239656518345, "grad_norm": 0.3898274600505829, "learning_rate": 2.664183840553417e-05, "loss": 0.2903, "step": 5030 }, { "epoch": 7.8610460577673695, "grad_norm": 0.4035649597644806, "learning_rate": 2.6456778893848144e-05, "loss": 0.2826, "step": 5035 }, { "epoch": 7.868852459016393, "grad_norm": 0.4092506170272827, "learning_rate": 2.6272266318987603e-05, "loss": 0.284, "step": 5040 }, { "epoch": 7.876658860265418, "grad_norm": 0.40052658319473267, "learning_rate": 2.6088302053162272e-05, "loss": 0.2894, "step": 5045 }, { "epoch": 7.884465261514442, "grad_norm": 0.3952447175979614, "learning_rate": 2.5904887464504114e-05, "loss": 0.281, "step": 5050 }, { "epoch": 7.892271662763466, "grad_norm": 0.3892611563205719, "learning_rate": 2.5722023917057125e-05, "loss": 0.2824, "step": 5055 }, { "epoch": 7.90007806401249, "grad_norm": 0.3946845531463623, "learning_rate": 2.5539712770767376e-05, "loss": 0.2865, "step": 5060 }, { "epoch": 7.907884465261515, "grad_norm": 0.39248526096343994, "learning_rate": 2.535795538147262e-05, "loss": 0.2863, "step": 5065 }, { "epoch": 7.9156908665105385, "grad_norm": 0.4364306330680847, "learning_rate": 2.5176753100892426e-05, "loss": 0.2916, "step": 5070 }, { "epoch": 7.923497267759563, "grad_norm": 0.3956867754459381, "learning_rate": 2.4996107276618008e-05, "loss": 0.2819, "step": 5075 }, { "epoch": 7.931303669008587, "grad_norm": 0.3908245265483856, "learning_rate": 2.4816019252102273e-05, "loss": 0.2864, "step": 5080 }, { "epoch": 7.939110070257612, "grad_norm": 0.3847108483314514, "learning_rate": 2.46364903666498e-05, "loss": 0.2937, "step": 5085 }, { "epoch": 7.946916471506635, "grad_norm": 0.412410169839859, "learning_rate": 2.445752195540687e-05, "loss": 0.2795, "step": 5090 }, { "epoch": 7.95472287275566, "grad_norm": 0.39699193835258484, "learning_rate": 2.4279115349351543e-05, "loss": 0.2909, "step": 5095 }, { "epoch": 7.962529274004684, "grad_norm": 0.4221264719963074, "learning_rate": 2.4101271875283817e-05, "loss": 0.2862, "step": 5100 }, { "epoch": 7.970335675253708, "grad_norm": 0.38855311274528503, "learning_rate": 2.3923992855815647e-05, "loss": 0.2847, "step": 5105 }, { "epoch": 7.978142076502732, "grad_norm": 0.4506378173828125, "learning_rate": 2.3747279609361196e-05, "loss": 0.2904, "step": 5110 }, { "epoch": 7.985948477751757, "grad_norm": 0.4097312390804291, "learning_rate": 2.3571133450127002e-05, "loss": 0.2903, "step": 5115 }, { "epoch": 7.993754879000781, "grad_norm": 0.4036017060279846, "learning_rate": 2.339555568810221e-05, "loss": 0.2857, "step": 5120 }, { "epoch": 8.0, "eval_loss": 1.1256442070007324, "eval_runtime": 0.3225, "eval_samples_per_second": 15.502, "eval_steps_per_second": 3.1, "step": 5124 }, { "epoch": 8.001561280249804, "grad_norm": 0.3814958333969116, "learning_rate": 2.3220547629048796e-05, "loss": 0.2894, "step": 5125 }, { "epoch": 8.009367681498828, "grad_norm": 0.3972679376602173, "learning_rate": 2.3046110574491985e-05, "loss": 0.2791, "step": 5130 }, { "epoch": 8.017174082747854, "grad_norm": 0.3934348225593567, "learning_rate": 2.2872245821710346e-05, "loss": 0.2719, "step": 5135 }, { "epoch": 8.024980483996877, "grad_norm": 0.4403572380542755, "learning_rate": 2.26989546637263e-05, "loss": 0.2688, "step": 5140 }, { "epoch": 8.032786885245901, "grad_norm": 0.43504083156585693, "learning_rate": 2.2526238389296585e-05, "loss": 0.2727, "step": 5145 }, { "epoch": 8.040593286494925, "grad_norm": 0.40251481533050537, "learning_rate": 2.2354098282902446e-05, "loss": 0.2706, "step": 5150 }, { "epoch": 8.04839968774395, "grad_norm": 0.3875437080860138, "learning_rate": 2.218253562474023e-05, "loss": 0.2684, "step": 5155 }, { "epoch": 8.056206088992974, "grad_norm": 0.39775198698043823, "learning_rate": 2.201155169071184e-05, "loss": 0.2685, "step": 5160 }, { "epoch": 8.064012490241998, "grad_norm": 0.4047357439994812, "learning_rate": 2.1841147752415235e-05, "loss": 0.2686, "step": 5165 }, { "epoch": 8.071818891491022, "grad_norm": 0.4089979827404022, "learning_rate": 2.1671325077134963e-05, "loss": 0.2695, "step": 5170 }, { "epoch": 8.079625292740047, "grad_norm": 0.40754222869873047, "learning_rate": 2.1502084927832845e-05, "loss": 0.2801, "step": 5175 }, { "epoch": 8.087431693989071, "grad_norm": 0.4052791893482208, "learning_rate": 2.1333428563138303e-05, "loss": 0.2708, "step": 5180 }, { "epoch": 8.095238095238095, "grad_norm": 0.4029444754123688, "learning_rate": 2.116535723733938e-05, "loss": 0.2688, "step": 5185 }, { "epoch": 8.103044496487119, "grad_norm": 0.43849509954452515, "learning_rate": 2.0997872200373116e-05, "loss": 0.2749, "step": 5190 }, { "epoch": 8.110850897736144, "grad_norm": 0.4012914001941681, "learning_rate": 2.083097469781632e-05, "loss": 0.2726, "step": 5195 }, { "epoch": 8.118657298985168, "grad_norm": 0.42504939436912537, "learning_rate": 2.0664665970876496e-05, "loss": 0.2791, "step": 5200 }, { "epoch": 8.126463700234192, "grad_norm": 0.41797903180122375, "learning_rate": 2.0498947256382273e-05, "loss": 0.2728, "step": 5205 }, { "epoch": 8.134270101483215, "grad_norm": 0.3988747000694275, "learning_rate": 2.0333819786774444e-05, "loss": 0.2757, "step": 5210 }, { "epoch": 8.142076502732241, "grad_norm": 0.4232763350009918, "learning_rate": 2.0169284790096853e-05, "loss": 0.2749, "step": 5215 }, { "epoch": 8.149882903981265, "grad_norm": 0.41461604833602905, "learning_rate": 2.000534348998704e-05, "loss": 0.2747, "step": 5220 }, { "epoch": 8.157689305230289, "grad_norm": 0.4166080355644226, "learning_rate": 1.9841997105667275e-05, "loss": 0.2721, "step": 5225 }, { "epoch": 8.165495706479312, "grad_norm": 0.41639989614486694, "learning_rate": 1.967924685193552e-05, "loss": 0.2645, "step": 5230 }, { "epoch": 8.173302107728338, "grad_norm": 0.4174750745296478, "learning_rate": 1.9517093939156295e-05, "loss": 0.2732, "step": 5235 }, { "epoch": 8.181108508977362, "grad_norm": 0.4450904130935669, "learning_rate": 1.9355539573251734e-05, "loss": 0.2777, "step": 5240 }, { "epoch": 8.188914910226385, "grad_norm": 0.4246358573436737, "learning_rate": 1.9194584955692706e-05, "loss": 0.2791, "step": 5245 }, { "epoch": 8.19672131147541, "grad_norm": 0.4177229106426239, "learning_rate": 1.903423128348959e-05, "loss": 0.2743, "step": 5250 }, { "epoch": 8.204527712724435, "grad_norm": 0.42174917459487915, "learning_rate": 1.8874479749183782e-05, "loss": 0.2754, "step": 5255 }, { "epoch": 8.212334113973458, "grad_norm": 0.4163845181465149, "learning_rate": 1.8715331540838487e-05, "loss": 0.2718, "step": 5260 }, { "epoch": 8.220140515222482, "grad_norm": 0.4283103346824646, "learning_rate": 1.855678784203002e-05, "loss": 0.2704, "step": 5265 }, { "epoch": 8.227946916471506, "grad_norm": 0.4160711467266083, "learning_rate": 1.8398849831839014e-05, "loss": 0.2679, "step": 5270 }, { "epoch": 8.235753317720532, "grad_norm": 0.4049237370491028, "learning_rate": 1.824151868484164e-05, "loss": 0.269, "step": 5275 }, { "epoch": 8.243559718969555, "grad_norm": 0.4092932641506195, "learning_rate": 1.808479557110081e-05, "loss": 0.2725, "step": 5280 }, { "epoch": 8.251366120218579, "grad_norm": 0.42792901396751404, "learning_rate": 1.792868165615762e-05, "loss": 0.2778, "step": 5285 }, { "epoch": 8.259172521467603, "grad_norm": 0.4112985134124756, "learning_rate": 1.7773178101022514e-05, "loss": 0.2708, "step": 5290 }, { "epoch": 8.266978922716628, "grad_norm": 0.4147091209888458, "learning_rate": 1.7618286062166677e-05, "loss": 0.2772, "step": 5295 }, { "epoch": 8.274785323965652, "grad_norm": 0.4274720847606659, "learning_rate": 1.7464006691513623e-05, "loss": 0.273, "step": 5300 }, { "epoch": 8.282591725214676, "grad_norm": 0.40364208817481995, "learning_rate": 1.7310341136430385e-05, "loss": 0.2778, "step": 5305 }, { "epoch": 8.2903981264637, "grad_norm": 0.42018428444862366, "learning_rate": 1.7157290539719106e-05, "loss": 0.2743, "step": 5310 }, { "epoch": 8.298204527712725, "grad_norm": 0.4169312119483948, "learning_rate": 1.700485603960853e-05, "loss": 0.275, "step": 5315 }, { "epoch": 8.306010928961749, "grad_norm": 0.40692824125289917, "learning_rate": 1.6853038769745467e-05, "loss": 0.2683, "step": 5320 }, { "epoch": 8.313817330210773, "grad_norm": 0.41851934790611267, "learning_rate": 1.6701839859186542e-05, "loss": 0.2677, "step": 5325 }, { "epoch": 8.321623731459797, "grad_norm": 0.4233751595020294, "learning_rate": 1.655126043238957e-05, "loss": 0.27, "step": 5330 }, { "epoch": 8.329430132708822, "grad_norm": 0.4326801002025604, "learning_rate": 1.6401301609205335e-05, "loss": 0.2774, "step": 5335 }, { "epoch": 8.337236533957846, "grad_norm": 0.42542752623558044, "learning_rate": 1.6251964504869222e-05, "loss": 0.2692, "step": 5340 }, { "epoch": 8.34504293520687, "grad_norm": 0.4217033386230469, "learning_rate": 1.6103250229992927e-05, "loss": 0.2804, "step": 5345 }, { "epoch": 8.352849336455893, "grad_norm": 0.4149416387081146, "learning_rate": 1.595515989055618e-05, "loss": 0.2732, "step": 5350 }, { "epoch": 8.360655737704919, "grad_norm": 0.42739465832710266, "learning_rate": 1.5807694587898648e-05, "loss": 0.2782, "step": 5355 }, { "epoch": 8.368462138953943, "grad_norm": 0.42222627997398376, "learning_rate": 1.566085541871145e-05, "loss": 0.2699, "step": 5360 }, { "epoch": 8.376268540202966, "grad_norm": 0.410786509513855, "learning_rate": 1.551464347502929e-05, "loss": 0.2743, "step": 5365 }, { "epoch": 8.38407494145199, "grad_norm": 0.42929258942604065, "learning_rate": 1.5369059844222278e-05, "loss": 0.2785, "step": 5370 }, { "epoch": 8.391881342701016, "grad_norm": 0.4153624475002289, "learning_rate": 1.5224105608987704e-05, "loss": 0.2687, "step": 5375 }, { "epoch": 8.39968774395004, "grad_norm": 0.3873469829559326, "learning_rate": 1.5079781847342123e-05, "loss": 0.2678, "step": 5380 }, { "epoch": 8.407494145199063, "grad_norm": 0.43077465891838074, "learning_rate": 1.4936089632613281e-05, "loss": 0.2765, "step": 5385 }, { "epoch": 8.415300546448087, "grad_norm": 0.4113737940788269, "learning_rate": 1.4793030033432142e-05, "loss": 0.2688, "step": 5390 }, { "epoch": 8.42310694769711, "grad_norm": 0.40002962946891785, "learning_rate": 1.4650604113724953e-05, "loss": 0.2693, "step": 5395 }, { "epoch": 8.430913348946136, "grad_norm": 0.4060341417789459, "learning_rate": 1.4508812932705363e-05, "loss": 0.2754, "step": 5400 }, { "epoch": 8.43871975019516, "grad_norm": 0.43747276067733765, "learning_rate": 1.4367657544866375e-05, "loss": 0.273, "step": 5405 }, { "epoch": 8.446526151444184, "grad_norm": 0.41913843154907227, "learning_rate": 1.42271389999728e-05, "loss": 0.2756, "step": 5410 }, { "epoch": 8.454332552693208, "grad_norm": 0.39348214864730835, "learning_rate": 1.4087258343053167e-05, "loss": 0.2701, "step": 5415 }, { "epoch": 8.462138953942233, "grad_norm": 0.4029614329338074, "learning_rate": 1.3948016614392112e-05, "loss": 0.269, "step": 5420 }, { "epoch": 8.469945355191257, "grad_norm": 0.45508867502212524, "learning_rate": 1.3809414849522584e-05, "loss": 0.2738, "step": 5425 }, { "epoch": 8.47775175644028, "grad_norm": 0.4153214395046234, "learning_rate": 1.367145407921817e-05, "loss": 0.2713, "step": 5430 }, { "epoch": 8.485558157689304, "grad_norm": 0.4336087703704834, "learning_rate": 1.3534135329485376e-05, "loss": 0.2786, "step": 5435 }, { "epoch": 8.49336455893833, "grad_norm": 0.4236217141151428, "learning_rate": 1.339745962155613e-05, "loss": 0.277, "step": 5440 }, { "epoch": 8.501170960187354, "grad_norm": 0.43844056129455566, "learning_rate": 1.3261427971880025e-05, "loss": 0.2709, "step": 5445 }, { "epoch": 8.508977361436378, "grad_norm": 0.43588918447494507, "learning_rate": 1.3126041392116772e-05, "loss": 0.2759, "step": 5450 }, { "epoch": 8.516783762685401, "grad_norm": 0.4094368517398834, "learning_rate": 1.2991300889128866e-05, "loss": 0.2776, "step": 5455 }, { "epoch": 8.524590163934427, "grad_norm": 0.41443923115730286, "learning_rate": 1.2857207464973875e-05, "loss": 0.2746, "step": 5460 }, { "epoch": 8.53239656518345, "grad_norm": 0.44371187686920166, "learning_rate": 1.2723762116897075e-05, "loss": 0.2821, "step": 5465 }, { "epoch": 8.540202966432474, "grad_norm": 0.3933749198913574, "learning_rate": 1.2590965837324131e-05, "loss": 0.2713, "step": 5470 }, { "epoch": 8.548009367681498, "grad_norm": 0.4023985266685486, "learning_rate": 1.2458819613853468e-05, "loss": 0.2714, "step": 5475 }, { "epoch": 8.555815768930524, "grad_norm": 0.4138771891593933, "learning_rate": 1.2327324429249232e-05, "loss": 0.2749, "step": 5480 }, { "epoch": 8.563622170179547, "grad_norm": 0.4046929180622101, "learning_rate": 1.2196481261433735e-05, "loss": 0.2732, "step": 5485 }, { "epoch": 8.571428571428571, "grad_norm": 0.4175087511539459, "learning_rate": 1.2066291083480296e-05, "loss": 0.2691, "step": 5490 }, { "epoch": 8.579234972677595, "grad_norm": 0.4213341176509857, "learning_rate": 1.1936754863606014e-05, "loss": 0.2769, "step": 5495 }, { "epoch": 8.58704137392662, "grad_norm": 0.43556031584739685, "learning_rate": 1.1807873565164506e-05, "loss": 0.2712, "step": 5500 }, { "epoch": 8.594847775175644, "grad_norm": 0.39848992228507996, "learning_rate": 1.1679648146638766e-05, "loss": 0.2718, "step": 5505 }, { "epoch": 8.602654176424668, "grad_norm": 0.4068874418735504, "learning_rate": 1.155207956163411e-05, "loss": 0.2714, "step": 5510 }, { "epoch": 8.610460577673692, "grad_norm": 0.41433241963386536, "learning_rate": 1.1425168758870964e-05, "loss": 0.275, "step": 5515 }, { "epoch": 8.618266978922717, "grad_norm": 0.42830371856689453, "learning_rate": 1.129891668217783e-05, "loss": 0.2743, "step": 5520 }, { "epoch": 8.626073380171741, "grad_norm": 0.4217815697193146, "learning_rate": 1.1173324270484397e-05, "loss": 0.2756, "step": 5525 }, { "epoch": 8.633879781420765, "grad_norm": 0.3956458270549774, "learning_rate": 1.1048392457814405e-05, "loss": 0.2743, "step": 5530 }, { "epoch": 8.641686182669789, "grad_norm": 0.42598819732666016, "learning_rate": 1.0924122173278751e-05, "loss": 0.2741, "step": 5535 }, { "epoch": 8.649492583918814, "grad_norm": 0.44088292121887207, "learning_rate": 1.080051434106859e-05, "loss": 0.2784, "step": 5540 }, { "epoch": 8.657298985167838, "grad_norm": 0.40059834718704224, "learning_rate": 1.067756988044848e-05, "loss": 0.2736, "step": 5545 }, { "epoch": 8.665105386416862, "grad_norm": 0.41283807158470154, "learning_rate": 1.0555289705749483e-05, "loss": 0.2781, "step": 5550 }, { "epoch": 8.672911787665885, "grad_norm": 0.40279173851013184, "learning_rate": 1.0433674726362442e-05, "loss": 0.2689, "step": 5555 }, { "epoch": 8.680718188914911, "grad_norm": 0.4150051176548004, "learning_rate": 1.0312725846731175e-05, "loss": 0.2733, "step": 5560 }, { "epoch": 8.688524590163935, "grad_norm": 0.4160909950733185, "learning_rate": 1.019244396634571e-05, "loss": 0.272, "step": 5565 }, { "epoch": 8.696330991412959, "grad_norm": 0.4090689718723297, "learning_rate": 1.0072829979735699e-05, "loss": 0.2779, "step": 5570 }, { "epoch": 8.704137392661982, "grad_norm": 0.43035534024238586, "learning_rate": 9.953884776463652e-06, "loss": 0.2799, "step": 5575 }, { "epoch": 8.711943793911008, "grad_norm": 0.4087963402271271, "learning_rate": 9.835609241118404e-06, "loss": 0.2724, "step": 5580 }, { "epoch": 8.719750195160032, "grad_norm": 0.4282284677028656, "learning_rate": 9.71800425330851e-06, "loss": 0.2766, "step": 5585 }, { "epoch": 8.727556596409055, "grad_norm": 0.39750391244888306, "learning_rate": 9.601070687655667e-06, "loss": 0.2751, "step": 5590 }, { "epoch": 8.735362997658079, "grad_norm": 0.43247750401496887, "learning_rate": 9.48480941378831e-06, "loss": 0.2701, "step": 5595 }, { "epoch": 8.743169398907105, "grad_norm": 0.42617279291152954, "learning_rate": 9.369221296335006e-06, "loss": 0.2774, "step": 5600 }, { "epoch": 8.750975800156128, "grad_norm": 0.4098498225212097, "learning_rate": 9.254307194918144e-06, "loss": 0.2779, "step": 5605 }, { "epoch": 8.758782201405152, "grad_norm": 0.4429432153701782, "learning_rate": 9.140067964147447e-06, "loss": 0.2809, "step": 5610 }, { "epoch": 8.766588602654176, "grad_norm": 0.42042461037635803, "learning_rate": 9.026504453613726e-06, "loss": 0.2755, "step": 5615 }, { "epoch": 8.774395003903201, "grad_norm": 0.41905051469802856, "learning_rate": 8.91361750788241e-06, "loss": 0.2679, "step": 5620 }, { "epoch": 8.782201405152225, "grad_norm": 0.4195885956287384, "learning_rate": 8.801407966487486e-06, "loss": 0.2734, "step": 5625 }, { "epoch": 8.790007806401249, "grad_norm": 0.4253963232040405, "learning_rate": 8.689876663924957e-06, "loss": 0.2674, "step": 5630 }, { "epoch": 8.797814207650273, "grad_norm": 0.43018728494644165, "learning_rate": 8.579024429646932e-06, "loss": 0.2762, "step": 5635 }, { "epoch": 8.805620608899297, "grad_norm": 0.41742637753486633, "learning_rate": 8.46885208805529e-06, "loss": 0.2768, "step": 5640 }, { "epoch": 8.813427010148322, "grad_norm": 0.41150280833244324, "learning_rate": 8.359360458495557e-06, "loss": 0.2777, "step": 5645 }, { "epoch": 8.821233411397346, "grad_norm": 0.4265673756599426, "learning_rate": 8.250550355250875e-06, "loss": 0.2743, "step": 5650 }, { "epoch": 8.82903981264637, "grad_norm": 0.4104432463645935, "learning_rate": 8.142422587535903e-06, "loss": 0.279, "step": 5655 }, { "epoch": 8.836846213895395, "grad_norm": 0.4088299572467804, "learning_rate": 8.034977959490775e-06, "loss": 0.2723, "step": 5660 }, { "epoch": 8.844652615144419, "grad_norm": 0.4326065182685852, "learning_rate": 7.92821727017523e-06, "loss": 0.2748, "step": 5665 }, { "epoch": 8.852459016393443, "grad_norm": 0.4133962392807007, "learning_rate": 7.822141313562547e-06, "loss": 0.2699, "step": 5670 }, { "epoch": 8.860265417642466, "grad_norm": 0.42703020572662354, "learning_rate": 7.71675087853364e-06, "loss": 0.2735, "step": 5675 }, { "epoch": 8.86807181889149, "grad_norm": 0.43023622035980225, "learning_rate": 7.612046748871327e-06, "loss": 0.2764, "step": 5680 }, { "epoch": 8.875878220140516, "grad_norm": 0.4223518371582031, "learning_rate": 7.50802970325436e-06, "loss": 0.276, "step": 5685 }, { "epoch": 8.88368462138954, "grad_norm": 0.43024882674217224, "learning_rate": 7.404700515251672e-06, "loss": 0.2745, "step": 5690 }, { "epoch": 8.891491022638563, "grad_norm": 0.4196966886520386, "learning_rate": 7.30205995331672e-06, "loss": 0.2791, "step": 5695 }, { "epoch": 8.899297423887587, "grad_norm": 0.39749372005462646, "learning_rate": 7.200108780781556e-06, "loss": 0.2753, "step": 5700 }, { "epoch": 8.907103825136613, "grad_norm": 0.4266413748264313, "learning_rate": 7.0988477558513785e-06, "loss": 0.2769, "step": 5705 }, { "epoch": 8.914910226385636, "grad_norm": 0.4070897102355957, "learning_rate": 6.998277631598793e-06, "loss": 0.2757, "step": 5710 }, { "epoch": 8.92271662763466, "grad_norm": 0.39892005920410156, "learning_rate": 6.898399155958168e-06, "loss": 0.2741, "step": 5715 }, { "epoch": 8.930523028883684, "grad_norm": 0.42477500438690186, "learning_rate": 6.7992130717201564e-06, "loss": 0.2791, "step": 5720 }, { "epoch": 8.93832943013271, "grad_norm": 0.44194042682647705, "learning_rate": 6.700720116526116e-06, "loss": 0.2786, "step": 5725 }, { "epoch": 8.946135831381733, "grad_norm": 0.4111786484718323, "learning_rate": 6.6029210228626626e-06, "loss": 0.2728, "step": 5730 }, { "epoch": 8.953942232630757, "grad_norm": 0.43993067741394043, "learning_rate": 6.505816518056162e-06, "loss": 0.2799, "step": 5735 }, { "epoch": 8.96174863387978, "grad_norm": 0.413835346698761, "learning_rate": 6.409407324267447e-06, "loss": 0.2761, "step": 5740 }, { "epoch": 8.969555035128806, "grad_norm": 0.4155209958553314, "learning_rate": 6.313694158486228e-06, "loss": 0.2726, "step": 5745 }, { "epoch": 8.97736143637783, "grad_norm": 0.43269336223602295, "learning_rate": 6.218677732526035e-06, "loss": 0.2697, "step": 5750 }, { "epoch": 8.985167837626854, "grad_norm": 0.4364638924598694, "learning_rate": 6.124358753018689e-06, "loss": 0.2788, "step": 5755 }, { "epoch": 8.992974238875878, "grad_norm": 0.40509918332099915, "learning_rate": 6.030737921409169e-06, "loss": 0.2732, "step": 5760 }, { "epoch": 8.999219359875097, "eval_loss": 1.1471812725067139, "eval_runtime": 0.3388, "eval_samples_per_second": 14.758, "eval_steps_per_second": 2.952, "step": 5764 }, { "epoch": 9.000780640124903, "grad_norm": 0.40420717000961304, "learning_rate": 5.937815933950375e-06, "loss": 0.2633, "step": 5765 }, { "epoch": 9.008587041373927, "grad_norm": 0.41395288705825806, "learning_rate": 5.8455934816979305e-06, "loss": 0.2729, "step": 5770 }, { "epoch": 9.01639344262295, "grad_norm": 0.4095096290111542, "learning_rate": 5.7540712505050444e-06, "loss": 0.265, "step": 5775 }, { "epoch": 9.024199843871974, "grad_norm": 0.41725656390190125, "learning_rate": 5.663249921017477e-06, "loss": 0.2632, "step": 5780 }, { "epoch": 9.032006245121, "grad_norm": 0.41932108998298645, "learning_rate": 5.57313016866835e-06, "loss": 0.2709, "step": 5785 }, { "epoch": 9.039812646370024, "grad_norm": 0.41799718141555786, "learning_rate": 5.483712663673224e-06, "loss": 0.2696, "step": 5790 }, { "epoch": 9.047619047619047, "grad_norm": 0.41482260823249817, "learning_rate": 5.394998071025104e-06, "loss": 0.2623, "step": 5795 }, { "epoch": 9.055425448868071, "grad_norm": 0.43754109740257263, "learning_rate": 5.306987050489442e-06, "loss": 0.2673, "step": 5800 }, { "epoch": 9.063231850117097, "grad_norm": 0.41212213039398193, "learning_rate": 5.21968025659928e-06, "loss": 0.2643, "step": 5805 }, { "epoch": 9.07103825136612, "grad_norm": 0.43795979022979736, "learning_rate": 5.1330783386503765e-06, "loss": 0.2672, "step": 5810 }, { "epoch": 9.078844652615144, "grad_norm": 0.416830837726593, "learning_rate": 5.047181940696333e-06, "loss": 0.2637, "step": 5815 }, { "epoch": 9.086651053864168, "grad_norm": 0.4341103136539459, "learning_rate": 4.961991701543889e-06, "loss": 0.2712, "step": 5820 }, { "epoch": 9.094457455113194, "grad_norm": 0.4144740402698517, "learning_rate": 4.877508254748076e-06, "loss": 0.2665, "step": 5825 }, { "epoch": 9.102263856362217, "grad_norm": 0.42530107498168945, "learning_rate": 4.7937322286075725e-06, "loss": 0.2678, "step": 5830 }, { "epoch": 9.110070257611241, "grad_norm": 0.4266839623451233, "learning_rate": 4.710664246160013e-06, "loss": 0.2709, "step": 5835 }, { "epoch": 9.117876658860265, "grad_norm": 0.42478978633880615, "learning_rate": 4.628304925177318e-06, "loss": 0.2692, "step": 5840 }, { "epoch": 9.12568306010929, "grad_norm": 0.42860227823257446, "learning_rate": 4.54665487816115e-06, "loss": 0.2694, "step": 5845 }, { "epoch": 9.133489461358314, "grad_norm": 0.43037670850753784, "learning_rate": 4.465714712338398e-06, "loss": 0.2686, "step": 5850 }, { "epoch": 9.141295862607338, "grad_norm": 0.4481579661369324, "learning_rate": 4.385485029656489e-06, "loss": 0.2672, "step": 5855 }, { "epoch": 9.149102263856362, "grad_norm": 0.406190425157547, "learning_rate": 4.305966426779118e-06, "loss": 0.2697, "step": 5860 }, { "epoch": 9.156908665105387, "grad_norm": 0.41195255517959595, "learning_rate": 4.22715949508169e-06, "loss": 0.268, "step": 5865 }, { "epoch": 9.164715066354411, "grad_norm": 0.41223201155662537, "learning_rate": 4.149064820646953e-06, "loss": 0.2625, "step": 5870 }, { "epoch": 9.172521467603435, "grad_norm": 0.40198594331741333, "learning_rate": 4.071682984260638e-06, "loss": 0.2654, "step": 5875 }, { "epoch": 9.180327868852459, "grad_norm": 0.4083414673805237, "learning_rate": 3.99501456140714e-06, "loss": 0.2666, "step": 5880 }, { "epoch": 9.188134270101482, "grad_norm": 0.43248122930526733, "learning_rate": 3.919060122265228e-06, "loss": 0.2712, "step": 5885 }, { "epoch": 9.195940671350508, "grad_norm": 0.4074991047382355, "learning_rate": 3.8438202317037986e-06, "loss": 0.2636, "step": 5890 }, { "epoch": 9.203747072599532, "grad_norm": 0.410256028175354, "learning_rate": 3.7692954492777764e-06, "loss": 0.2663, "step": 5895 }, { "epoch": 9.211553473848555, "grad_norm": 0.41726231575012207, "learning_rate": 3.6954863292237297e-06, "loss": 0.2649, "step": 5900 }, { "epoch": 9.21935987509758, "grad_norm": 0.40415236353874207, "learning_rate": 3.622393420456016e-06, "loss": 0.2691, "step": 5905 }, { "epoch": 9.227166276346605, "grad_norm": 0.40107595920562744, "learning_rate": 3.550017266562489e-06, "loss": 0.2636, "step": 5910 }, { "epoch": 9.234972677595628, "grad_norm": 0.4478661119937897, "learning_rate": 3.4783584058005527e-06, "loss": 0.2711, "step": 5915 }, { "epoch": 9.242779078844652, "grad_norm": 0.40624120831489563, "learning_rate": 3.40741737109318e-06, "loss": 0.2652, "step": 5920 }, { "epoch": 9.250585480093676, "grad_norm": 0.42705732583999634, "learning_rate": 3.3371946900248473e-06, "loss": 0.2675, "step": 5925 }, { "epoch": 9.258391881342702, "grad_norm": 0.4034976661205292, "learning_rate": 3.267690884837726e-06, "loss": 0.2631, "step": 5930 }, { "epoch": 9.266198282591725, "grad_norm": 0.4358353912830353, "learning_rate": 3.198906472427732e-06, "loss": 0.2702, "step": 5935 }, { "epoch": 9.274004683840749, "grad_norm": 0.42044761776924133, "learning_rate": 3.130841964340692e-06, "loss": 0.269, "step": 5940 }, { "epoch": 9.281811085089773, "grad_norm": 0.4694254994392395, "learning_rate": 3.06349786676855e-06, "loss": 0.27, "step": 5945 }, { "epoch": 9.289617486338798, "grad_norm": 0.4230342507362366, "learning_rate": 2.996874680545603e-06, "loss": 0.2645, "step": 5950 }, { "epoch": 9.297423887587822, "grad_norm": 0.436117947101593, "learning_rate": 2.9309729011447573e-06, "loss": 0.2713, "step": 5955 }, { "epoch": 9.305230288836846, "grad_norm": 0.4295603632926941, "learning_rate": 2.8657930186738567e-06, "loss": 0.2666, "step": 5960 }, { "epoch": 9.31303669008587, "grad_norm": 0.42392897605895996, "learning_rate": 2.8013355178720612e-06, "loss": 0.2682, "step": 5965 }, { "epoch": 9.320843091334895, "grad_norm": 0.40090325474739075, "learning_rate": 2.7376008781061835e-06, "loss": 0.2624, "step": 5970 }, { "epoch": 9.328649492583919, "grad_norm": 0.4344378113746643, "learning_rate": 2.674589573367192e-06, "loss": 0.274, "step": 5975 }, { "epoch": 9.336455893832943, "grad_norm": 0.4555279016494751, "learning_rate": 2.612302072266637e-06, "loss": 0.2696, "step": 5980 }, { "epoch": 9.344262295081966, "grad_norm": 0.4352642297744751, "learning_rate": 2.5507388380331843e-06, "loss": 0.2677, "step": 5985 }, { "epoch": 9.352068696330992, "grad_norm": 0.4107036590576172, "learning_rate": 2.4899003285091536e-06, "loss": 0.2692, "step": 5990 }, { "epoch": 9.359875097580016, "grad_norm": 0.4284498691558838, "learning_rate": 2.429786996147154e-06, "loss": 0.2623, "step": 5995 }, { "epoch": 9.36768149882904, "grad_norm": 0.42979347705841064, "learning_rate": 2.3703992880066638e-06, "loss": 0.2705, "step": 6000 }, { "epoch": 9.375487900078063, "grad_norm": 0.4255581796169281, "learning_rate": 2.3117376457507667e-06, "loss": 0.2678, "step": 6005 }, { "epoch": 9.383294301327089, "grad_norm": 0.4213410019874573, "learning_rate": 2.2538025056428214e-06, "loss": 0.2679, "step": 6010 }, { "epoch": 9.391100702576113, "grad_norm": 0.42378342151641846, "learning_rate": 2.1965942985431977e-06, "loss": 0.2603, "step": 6015 }, { "epoch": 9.398907103825136, "grad_norm": 0.43251511454582214, "learning_rate": 2.140113449906167e-06, "loss": 0.2678, "step": 6020 }, { "epoch": 9.40671350507416, "grad_norm": 0.428505539894104, "learning_rate": 2.0843603797766287e-06, "loss": 0.2716, "step": 6025 }, { "epoch": 9.414519906323186, "grad_norm": 0.40141308307647705, "learning_rate": 2.0293355027870554e-06, "loss": 0.264, "step": 6030 }, { "epoch": 9.42232630757221, "grad_norm": 0.42421501874923706, "learning_rate": 1.975039228154385e-06, "loss": 0.2675, "step": 6035 }, { "epoch": 9.430132708821233, "grad_norm": 0.41700848937034607, "learning_rate": 1.921471959676957e-06, "loss": 0.2663, "step": 6040 }, { "epoch": 9.437939110070257, "grad_norm": 0.4247893989086151, "learning_rate": 1.8686340957315695e-06, "loss": 0.2671, "step": 6045 }, { "epoch": 9.445745511319283, "grad_norm": 0.4123760163784027, "learning_rate": 1.8165260292704711e-06, "loss": 0.264, "step": 6050 }, { "epoch": 9.453551912568306, "grad_norm": 0.4423754811286926, "learning_rate": 1.7651481478184296e-06, "loss": 0.2617, "step": 6055 }, { "epoch": 9.46135831381733, "grad_norm": 0.4133644998073578, "learning_rate": 1.7145008334698898e-06, "loss": 0.2624, "step": 6060 }, { "epoch": 9.469164715066354, "grad_norm": 0.4270353317260742, "learning_rate": 1.6645844628860764e-06, "loss": 0.2674, "step": 6065 }, { "epoch": 9.47697111631538, "grad_norm": 0.41615259647369385, "learning_rate": 1.6153994072922506e-06, "loss": 0.2708, "step": 6070 }, { "epoch": 9.484777517564403, "grad_norm": 0.4244523346424103, "learning_rate": 1.5669460324749586e-06, "loss": 0.2698, "step": 6075 }, { "epoch": 9.492583918813427, "grad_norm": 0.4362230598926544, "learning_rate": 1.5192246987791981e-06, "loss": 0.2617, "step": 6080 }, { "epoch": 9.50039032006245, "grad_norm": 0.4365394413471222, "learning_rate": 1.472235761105878e-06, "loss": 0.2669, "step": 6085 }, { "epoch": 9.508196721311476, "grad_norm": 0.4158194363117218, "learning_rate": 1.4259795689090972e-06, "loss": 0.2694, "step": 6090 }, { "epoch": 9.5160031225605, "grad_norm": 0.41145652532577515, "learning_rate": 1.3804564661935915e-06, "loss": 0.2629, "step": 6095 }, { "epoch": 9.523809523809524, "grad_norm": 0.39693784713745117, "learning_rate": 1.3356667915121025e-06, "loss": 0.2683, "step": 6100 }, { "epoch": 9.531615925058547, "grad_norm": 0.4337891638278961, "learning_rate": 1.2916108779629566e-06, "loss": 0.2691, "step": 6105 }, { "epoch": 9.539422326307573, "grad_norm": 0.431761234998703, "learning_rate": 1.2482890531875125e-06, "loss": 0.2704, "step": 6110 }, { "epoch": 9.547228727556597, "grad_norm": 0.42788565158843994, "learning_rate": 1.2057016393677623e-06, "loss": 0.2721, "step": 6115 }, { "epoch": 9.55503512880562, "grad_norm": 0.4381605088710785, "learning_rate": 1.163848953223934e-06, "loss": 0.2731, "step": 6120 }, { "epoch": 9.562841530054644, "grad_norm": 0.42890533804893494, "learning_rate": 1.1227313060120926e-06, "loss": 0.2683, "step": 6125 }, { "epoch": 9.570647931303668, "grad_norm": 0.41614386439323425, "learning_rate": 1.0823490035218987e-06, "loss": 0.2634, "step": 6130 }, { "epoch": 9.578454332552694, "grad_norm": 0.41682758927345276, "learning_rate": 1.042702346074287e-06, "loss": 0.2621, "step": 6135 }, { "epoch": 9.586260733801717, "grad_norm": 0.4195772707462311, "learning_rate": 1.003791628519213e-06, "loss": 0.2677, "step": 6140 }, { "epoch": 9.594067135050741, "grad_norm": 0.40635907649993896, "learning_rate": 9.656171402335213e-07, "loss": 0.2668, "step": 6145 }, { "epoch": 9.601873536299767, "grad_norm": 0.40993356704711914, "learning_rate": 9.281791651187366e-07, "loss": 0.2661, "step": 6150 }, { "epoch": 9.60967993754879, "grad_norm": 0.4103589355945587, "learning_rate": 8.914779815989982e-07, "loss": 0.2691, "step": 6155 }, { "epoch": 9.617486338797814, "grad_norm": 0.44015854597091675, "learning_rate": 8.555138626189618e-07, "loss": 0.2649, "step": 6160 }, { "epoch": 9.625292740046838, "grad_norm": 0.414385586977005, "learning_rate": 8.202870756417569e-07, "loss": 0.2616, "step": 6165 }, { "epoch": 9.633099141295862, "grad_norm": 0.4170667827129364, "learning_rate": 7.857978826470325e-07, "loss": 0.2658, "step": 6170 }, { "epoch": 9.640905542544887, "grad_norm": 0.42618632316589355, "learning_rate": 7.520465401290033e-07, "loss": 0.2691, "step": 6175 }, { "epoch": 9.648711943793911, "grad_norm": 0.41920483112335205, "learning_rate": 7.19033299094496e-07, "loss": 0.2639, "step": 6180 }, { "epoch": 9.656518345042935, "grad_norm": 0.4324471056461334, "learning_rate": 6.867584050611498e-07, "loss": 0.2715, "step": 6185 }, { "epoch": 9.664324746291959, "grad_norm": 0.4518769681453705, "learning_rate": 6.552220980555635e-07, "loss": 0.2749, "step": 6190 }, { "epoch": 9.672131147540984, "grad_norm": 0.4381062984466553, "learning_rate": 6.244246126114627e-07, "loss": 0.2734, "step": 6195 }, { "epoch": 9.679937548790008, "grad_norm": 0.4453407824039459, "learning_rate": 5.943661777680354e-07, "loss": 0.2692, "step": 6200 }, { "epoch": 9.687743950039032, "grad_norm": 0.41984906792640686, "learning_rate": 5.650470170681876e-07, "loss": 0.2644, "step": 6205 }, { "epoch": 9.695550351288055, "grad_norm": 0.4198761284351349, "learning_rate": 5.364673485568794e-07, "loss": 0.2676, "step": 6210 }, { "epoch": 9.703356752537081, "grad_norm": 0.42462319135665894, "learning_rate": 5.086273847795031e-07, "loss": 0.2671, "step": 6215 }, { "epoch": 9.711163153786105, "grad_norm": 0.4393913447856903, "learning_rate": 4.815273327803182e-07, "loss": 0.2689, "step": 6220 }, { "epoch": 9.718969555035128, "grad_norm": 0.4282847046852112, "learning_rate": 4.5516739410087494e-07, "loss": 0.2696, "step": 6225 }, { "epoch": 9.726775956284152, "grad_norm": 0.41957032680511475, "learning_rate": 4.2954776477860393e-07, "loss": 0.2748, "step": 6230 }, { "epoch": 9.734582357533178, "grad_norm": 0.4238908886909485, "learning_rate": 4.0466863534522893e-07, "loss": 0.2741, "step": 6235 }, { "epoch": 9.742388758782202, "grad_norm": 0.4146591126918793, "learning_rate": 3.805301908254455e-07, "loss": 0.2653, "step": 6240 }, { "epoch": 9.750195160031225, "grad_norm": 0.4316459000110626, "learning_rate": 3.571326107355333e-07, "loss": 0.2687, "step": 6245 }, { "epoch": 9.758001561280249, "grad_norm": 0.4224886894226074, "learning_rate": 3.3447606908196817e-07, "loss": 0.2646, "step": 6250 }, { "epoch": 9.765807962529275, "grad_norm": 0.439411461353302, "learning_rate": 3.125607343602011e-07, "loss": 0.2657, "step": 6255 }, { "epoch": 9.773614363778298, "grad_norm": 0.4261423647403717, "learning_rate": 2.9138676955333676e-07, "loss": 0.2606, "step": 6260 }, { "epoch": 9.781420765027322, "grad_norm": 0.41440731287002563, "learning_rate": 2.709543321309793e-07, "loss": 0.2673, "step": 6265 }, { "epoch": 9.789227166276346, "grad_norm": 0.42373111844062805, "learning_rate": 2.5126357404802183e-07, "loss": 0.2751, "step": 6270 }, { "epoch": 9.797033567525371, "grad_norm": 0.4395394027233124, "learning_rate": 2.3231464174352512e-07, "loss": 0.2747, "step": 6275 }, { "epoch": 9.804839968774395, "grad_norm": 0.4383135735988617, "learning_rate": 2.141076761396521e-07, "loss": 0.2708, "step": 6280 }, { "epoch": 9.812646370023419, "grad_norm": 0.4313544034957886, "learning_rate": 1.966428126405795e-07, "loss": 0.2686, "step": 6285 }, { "epoch": 9.820452771272443, "grad_norm": 0.39951956272125244, "learning_rate": 1.7992018113152098e-07, "loss": 0.2591, "step": 6290 }, { "epoch": 9.828259172521468, "grad_norm": 0.4347188174724579, "learning_rate": 1.6393990597775022e-07, "loss": 0.2709, "step": 6295 }, { "epoch": 9.836065573770492, "grad_norm": 0.4294663965702057, "learning_rate": 1.487021060236904e-07, "loss": 0.2644, "step": 6300 }, { "epoch": 9.843871975019516, "grad_norm": 0.437920480966568, "learning_rate": 1.3420689459200386e-07, "loss": 0.2693, "step": 6305 }, { "epoch": 9.85167837626854, "grad_norm": 0.4170249104499817, "learning_rate": 1.204543794827595e-07, "loss": 0.2748, "step": 6310 }, { "epoch": 9.859484777517565, "grad_norm": 0.40519675612449646, "learning_rate": 1.0744466297265554e-07, "loss": 0.267, "step": 6315 }, { "epoch": 9.867291178766589, "grad_norm": 0.40046894550323486, "learning_rate": 9.517784181422019e-08, "loss": 0.2614, "step": 6320 }, { "epoch": 9.875097580015613, "grad_norm": 0.4310961067676544, "learning_rate": 8.365400723512328e-08, "loss": 0.2617, "step": 6325 }, { "epoch": 9.882903981264636, "grad_norm": 0.43310293555259705, "learning_rate": 7.287324493747693e-08, "loss": 0.2685, "step": 6330 }, { "epoch": 9.890710382513662, "grad_norm": 0.43305081129074097, "learning_rate": 6.283563509719148e-08, "loss": 0.2607, "step": 6335 }, { "epoch": 9.898516783762686, "grad_norm": 0.42313021421432495, "learning_rate": 5.354125236343155e-08, "loss": 0.2714, "step": 6340 }, { "epoch": 9.90632318501171, "grad_norm": 0.43003058433532715, "learning_rate": 4.499016585799431e-08, "loss": 0.2685, "step": 6345 }, { "epoch": 9.914129586260733, "grad_norm": 0.4060039818286896, "learning_rate": 3.7182439174832106e-08, "loss": 0.2728, "step": 6350 }, { "epoch": 9.921935987509759, "grad_norm": 0.4126955270767212, "learning_rate": 3.0118130379575005e-08, "loss": 0.2675, "step": 6355 }, { "epoch": 9.929742388758783, "grad_norm": 0.43146437406539917, "learning_rate": 2.379729200908676e-08, "loss": 0.2695, "step": 6360 }, { "epoch": 9.937548790007806, "grad_norm": 0.4640655219554901, "learning_rate": 1.8219971071098408e-08, "loss": 0.2713, "step": 6365 }, { "epoch": 9.94535519125683, "grad_norm": 0.41949042677879333, "learning_rate": 1.3386209043819708e-08, "loss": 0.2707, "step": 6370 }, { "epoch": 9.953161592505854, "grad_norm": 0.42296043038368225, "learning_rate": 9.296041875683781e-09, "loss": 0.2636, "step": 6375 }, { "epoch": 9.96096799375488, "grad_norm": 0.45912352204322815, "learning_rate": 5.949499985025142e-09, "loss": 0.2702, "step": 6380 }, { "epoch": 9.968774395003903, "grad_norm": 0.4141922891139984, "learning_rate": 3.346608259890971e-09, "loss": 0.2663, "step": 6385 }, { "epoch": 9.976580796252927, "grad_norm": 0.42737066745758057, "learning_rate": 1.4873860578412668e-09, "loss": 0.2692, "step": 6390 }, { "epoch": 9.984387197501952, "grad_norm": 0.441078782081604, "learning_rate": 3.7184720581562306e-10, "loss": 0.2707, "step": 6395 }, { "epoch": 9.992193598750976, "grad_norm": 0.45411616563796997, "learning_rate": 0.0, "loss": 0.2714, "step": 6400 }, { "epoch": 9.992193598750976, "eval_loss": 1.1572602987289429, "eval_runtime": 0.3457, "eval_samples_per_second": 14.464, "eval_steps_per_second": 2.893, "step": 6400 }, { "epoch": 9.992193598750976, "step": 6400, "total_flos": 8.968401612833817e+18, "train_loss": 0.34100163986906407, "train_runtime": 19497.7947, "train_samples_per_second": 10.509, "train_steps_per_second": 0.328 } ], "logging_steps": 5, "max_steps": 6400, "num_input_tokens_seen": 0, "num_train_epochs": 10, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 8.968401612833817e+18, "train_batch_size": 4, "trial_name": null, "trial_params": null }