{ "best_metric": 0.44520583748817444, "best_model_checkpoint": "saves/Orca/lora/checkpoint-3000", "epoch": 0.9998413957176844, "eval_steps": 500, "global_step": 3152, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0031720856463124504, "grad_norm": 0.8383375406265259, "learning_rate": 5.000000000000001e-07, "loss": 2.0081, "step": 10 }, { "epoch": 0.006344171292624901, "grad_norm": 0.9924185276031494, "learning_rate": 1.0000000000000002e-06, "loss": 2.0141, "step": 20 }, { "epoch": 0.00951625693893735, "grad_norm": 1.7123312950134277, "learning_rate": 1.5e-06, "loss": 2.0134, "step": 30 }, { "epoch": 0.012688342585249802, "grad_norm": 1.6002644300460815, "learning_rate": 2.0000000000000003e-06, "loss": 1.9886, "step": 40 }, { "epoch": 0.01586042823156225, "grad_norm": 1.4284940958023071, "learning_rate": 2.5e-06, "loss": 1.9963, "step": 50 }, { "epoch": 0.0190325138778747, "grad_norm": 1.9293030500411987, "learning_rate": 3e-06, "loss": 1.9858, "step": 60 }, { "epoch": 0.022204599524187154, "grad_norm": 1.8281705379486084, "learning_rate": 3.5000000000000004e-06, "loss": 1.9608, "step": 70 }, { "epoch": 0.025376685170499604, "grad_norm": 1.8404144048690796, "learning_rate": 4.000000000000001e-06, "loss": 1.9317, "step": 80 }, { "epoch": 0.028548770816812053, "grad_norm": 1.8080644607543945, "learning_rate": 4.5e-06, "loss": 1.8928, "step": 90 }, { "epoch": 0.0317208564631245, "grad_norm": 1.4789849519729614, "learning_rate": 5e-06, "loss": 1.8453, "step": 100 }, { "epoch": 0.034892942109436956, "grad_norm": 2.3675060272216797, "learning_rate": 5.500000000000001e-06, "loss": 1.7683, "step": 110 }, { "epoch": 0.0380650277557494, "grad_norm": 6.9262309074401855, "learning_rate": 6e-06, "loss": 1.6648, "step": 120 }, { "epoch": 0.041237113402061855, "grad_norm": 1.5342968702316284, "learning_rate": 6.5000000000000004e-06, "loss": 1.5507, "step": 130 }, { "epoch": 0.04440919904837431, "grad_norm": 1.7138350009918213, "learning_rate": 7.000000000000001e-06, "loss": 1.4388, "step": 140 }, { "epoch": 0.047581284694686754, "grad_norm": 0.9868146181106567, "learning_rate": 7.5e-06, "loss": 1.3484, "step": 150 }, { "epoch": 0.05075337034099921, "grad_norm": 0.649817705154419, "learning_rate": 8.000000000000001e-06, "loss": 1.2671, "step": 160 }, { "epoch": 0.05392545598731166, "grad_norm": 0.5127522945404053, "learning_rate": 8.500000000000002e-06, "loss": 1.19, "step": 170 }, { "epoch": 0.057097541633624106, "grad_norm": 0.43639031052589417, "learning_rate": 9e-06, "loss": 1.1479, "step": 180 }, { "epoch": 0.06026962727993656, "grad_norm": 0.31427648663520813, "learning_rate": 9.5e-06, "loss": 1.0992, "step": 190 }, { "epoch": 0.063441712926249, "grad_norm": 0.2897365689277649, "learning_rate": 1e-05, "loss": 1.0673, "step": 200 }, { "epoch": 0.06661379857256146, "grad_norm": 0.43495845794677734, "learning_rate": 1.05e-05, "loss": 1.0399, "step": 210 }, { "epoch": 0.06978588421887391, "grad_norm": 0.2536582946777344, "learning_rate": 1.1000000000000001e-05, "loss": 1.0043, "step": 220 }, { "epoch": 0.07295796986518636, "grad_norm": 0.3480685353279114, "learning_rate": 1.1500000000000002e-05, "loss": 0.9747, "step": 230 }, { "epoch": 0.0761300555114988, "grad_norm": 0.25048044323921204, "learning_rate": 1.2e-05, "loss": 0.9477, "step": 240 }, { "epoch": 0.07930214115781126, "grad_norm": 0.31881338357925415, "learning_rate": 1.25e-05, "loss": 0.9099, "step": 250 }, { "epoch": 0.08247422680412371, "grad_norm": 0.25372961163520813, "learning_rate": 1.3000000000000001e-05, "loss": 0.8661, "step": 260 }, { "epoch": 0.08564631245043616, "grad_norm": 0.2892467975616455, "learning_rate": 1.3500000000000001e-05, "loss": 0.8306, "step": 270 }, { "epoch": 0.08881839809674862, "grad_norm": 0.265289306640625, "learning_rate": 1.4000000000000001e-05, "loss": 0.79, "step": 280 }, { "epoch": 0.09199048374306107, "grad_norm": 0.2384192794561386, "learning_rate": 1.45e-05, "loss": 0.7414, "step": 290 }, { "epoch": 0.09516256938937351, "grad_norm": 0.2318211942911148, "learning_rate": 1.5e-05, "loss": 0.6994, "step": 300 }, { "epoch": 0.09833465503568596, "grad_norm": 0.22584764659404755, "learning_rate": 1.55e-05, "loss": 0.6663, "step": 310 }, { "epoch": 0.10150674068199841, "grad_norm": 0.2281956821680069, "learning_rate": 1.6000000000000003e-05, "loss": 0.6366, "step": 320 }, { "epoch": 0.10467882632831087, "grad_norm": 0.23853954672813416, "learning_rate": 1.65e-05, "loss": 0.6105, "step": 330 }, { "epoch": 0.10785091197462332, "grad_norm": 0.2249988317489624, "learning_rate": 1.7000000000000003e-05, "loss": 0.5923, "step": 340 }, { "epoch": 0.11102299762093576, "grad_norm": 0.22325117886066437, "learning_rate": 1.75e-05, "loss": 0.5782, "step": 350 }, { "epoch": 0.11419508326724821, "grad_norm": 0.26995840668678284, "learning_rate": 1.8e-05, "loss": 0.5588, "step": 360 }, { "epoch": 0.11736716891356067, "grad_norm": 0.3128603994846344, "learning_rate": 1.85e-05, "loss": 0.5539, "step": 370 }, { "epoch": 0.12053925455987312, "grad_norm": 0.2401484102010727, "learning_rate": 1.9e-05, "loss": 0.5463, "step": 380 }, { "epoch": 0.12371134020618557, "grad_norm": 0.252087265253067, "learning_rate": 1.9500000000000003e-05, "loss": 0.5396, "step": 390 }, { "epoch": 0.126883425852498, "grad_norm": 0.3551071286201477, "learning_rate": 2e-05, "loss": 0.5293, "step": 400 }, { "epoch": 0.13005551149881048, "grad_norm": 0.2852734625339508, "learning_rate": 2.05e-05, "loss": 0.5327, "step": 410 }, { "epoch": 0.13322759714512292, "grad_norm": 0.25599339604377747, "learning_rate": 2.1e-05, "loss": 0.5215, "step": 420 }, { "epoch": 0.13639968279143536, "grad_norm": 0.2308124303817749, "learning_rate": 2.15e-05, "loss": 0.5179, "step": 430 }, { "epoch": 0.13957176843774782, "grad_norm": 0.28417715430259705, "learning_rate": 2.2000000000000003e-05, "loss": 0.5122, "step": 440 }, { "epoch": 0.14274385408406026, "grad_norm": 0.32094892859458923, "learning_rate": 2.25e-05, "loss": 0.5078, "step": 450 }, { "epoch": 0.14591593973037273, "grad_norm": 0.28217512369155884, "learning_rate": 2.3000000000000003e-05, "loss": 0.5064, "step": 460 }, { "epoch": 0.14908802537668517, "grad_norm": 0.3319534659385681, "learning_rate": 2.35e-05, "loss": 0.5056, "step": 470 }, { "epoch": 0.1522601110229976, "grad_norm": 0.35069146752357483, "learning_rate": 2.4e-05, "loss": 0.5001, "step": 480 }, { "epoch": 0.15543219666931007, "grad_norm": 0.2778172492980957, "learning_rate": 2.45e-05, "loss": 0.4994, "step": 490 }, { "epoch": 0.1586042823156225, "grad_norm": 0.2950364053249359, "learning_rate": 2.5e-05, "loss": 0.5009, "step": 500 }, { "epoch": 0.1586042823156225, "eval_loss": 0.49637770652770996, "eval_runtime": 23627.0877, "eval_samples_per_second": 0.044, "eval_steps_per_second": 0.022, "step": 500 }, { "epoch": 0.16177636796193498, "grad_norm": 0.26569539308547974, "learning_rate": 2.5500000000000003e-05, "loss": 0.4978, "step": 510 }, { "epoch": 0.16494845360824742, "grad_norm": 0.26948851346969604, "learning_rate": 2.6000000000000002e-05, "loss": 0.4926, "step": 520 }, { "epoch": 0.1681205392545599, "grad_norm": 0.29083552956581116, "learning_rate": 2.6500000000000004e-05, "loss": 0.4893, "step": 530 }, { "epoch": 0.17129262490087233, "grad_norm": 0.2939367890357971, "learning_rate": 2.7000000000000002e-05, "loss": 0.491, "step": 540 }, { "epoch": 0.17446471054718476, "grad_norm": 0.28195658326148987, "learning_rate": 2.7500000000000004e-05, "loss": 0.4865, "step": 550 }, { "epoch": 0.17763679619349723, "grad_norm": 0.2827642858028412, "learning_rate": 2.8000000000000003e-05, "loss": 0.4851, "step": 560 }, { "epoch": 0.18080888183980967, "grad_norm": 0.2715272605419159, "learning_rate": 2.8499999999999998e-05, "loss": 0.4833, "step": 570 }, { "epoch": 0.18398096748612214, "grad_norm": 0.30428314208984375, "learning_rate": 2.9e-05, "loss": 0.4834, "step": 580 }, { "epoch": 0.18715305313243458, "grad_norm": 0.2857949435710907, "learning_rate": 2.95e-05, "loss": 0.4796, "step": 590 }, { "epoch": 0.19032513877874702, "grad_norm": 0.302212655544281, "learning_rate": 3e-05, "loss": 0.4826, "step": 600 }, { "epoch": 0.19349722442505948, "grad_norm": 0.34079357981681824, "learning_rate": 3.05e-05, "loss": 0.4826, "step": 610 }, { "epoch": 0.19666931007137192, "grad_norm": 0.3239024877548218, "learning_rate": 3.1e-05, "loss": 0.4804, "step": 620 }, { "epoch": 0.1998413957176844, "grad_norm": 0.36531686782836914, "learning_rate": 3.15e-05, "loss": 0.4782, "step": 630 }, { "epoch": 0.20301348136399683, "grad_norm": 0.3582271337509155, "learning_rate": 3.2000000000000005e-05, "loss": 0.4784, "step": 640 }, { "epoch": 0.20618556701030927, "grad_norm": 0.3250937759876251, "learning_rate": 3.2500000000000004e-05, "loss": 0.4716, "step": 650 }, { "epoch": 0.20935765265662173, "grad_norm": 0.28494277596473694, "learning_rate": 3.3e-05, "loss": 0.4764, "step": 660 }, { "epoch": 0.21252973830293417, "grad_norm": 0.3156846761703491, "learning_rate": 3.35e-05, "loss": 0.4742, "step": 670 }, { "epoch": 0.21570182394924664, "grad_norm": 0.27053794264793396, "learning_rate": 3.4000000000000007e-05, "loss": 0.4765, "step": 680 }, { "epoch": 0.21887390959555908, "grad_norm": 0.31487756967544556, "learning_rate": 3.45e-05, "loss": 0.4749, "step": 690 }, { "epoch": 0.22204599524187152, "grad_norm": 0.2739400267601013, "learning_rate": 3.5e-05, "loss": 0.4693, "step": 700 }, { "epoch": 0.22521808088818399, "grad_norm": 0.2924807369709015, "learning_rate": 3.55e-05, "loss": 0.4697, "step": 710 }, { "epoch": 0.22839016653449642, "grad_norm": 0.27241086959838867, "learning_rate": 3.6e-05, "loss": 0.4711, "step": 720 }, { "epoch": 0.2315622521808089, "grad_norm": 0.249298095703125, "learning_rate": 3.65e-05, "loss": 0.4748, "step": 730 }, { "epoch": 0.23473433782712133, "grad_norm": 0.33761996030807495, "learning_rate": 3.7e-05, "loss": 0.4692, "step": 740 }, { "epoch": 0.23790642347343377, "grad_norm": 0.3454744517803192, "learning_rate": 3.7500000000000003e-05, "loss": 0.4691, "step": 750 }, { "epoch": 0.24107850911974624, "grad_norm": 0.3940749168395996, "learning_rate": 3.8e-05, "loss": 0.4694, "step": 760 }, { "epoch": 0.24425059476605868, "grad_norm": 0.2833056151866913, "learning_rate": 3.85e-05, "loss": 0.4651, "step": 770 }, { "epoch": 0.24742268041237114, "grad_norm": 0.3439841866493225, "learning_rate": 3.9000000000000006e-05, "loss": 0.4718, "step": 780 }, { "epoch": 0.2505947660586836, "grad_norm": 0.28585824370384216, "learning_rate": 3.9500000000000005e-05, "loss": 0.4678, "step": 790 }, { "epoch": 0.253766851704996, "grad_norm": 0.2761208415031433, "learning_rate": 4e-05, "loss": 0.4672, "step": 800 }, { "epoch": 0.25693893735130846, "grad_norm": 0.287431538105011, "learning_rate": 4.05e-05, "loss": 0.4654, "step": 810 }, { "epoch": 0.26011102299762096, "grad_norm": 0.2529178261756897, "learning_rate": 4.1e-05, "loss": 0.4658, "step": 820 }, { "epoch": 0.2632831086439334, "grad_norm": 0.22234879434108734, "learning_rate": 4.15e-05, "loss": 0.467, "step": 830 }, { "epoch": 0.26645519429024583, "grad_norm": 0.2847014367580414, "learning_rate": 4.2e-05, "loss": 0.4643, "step": 840 }, { "epoch": 0.2696272799365583, "grad_norm": 0.27712419629096985, "learning_rate": 4.25e-05, "loss": 0.4604, "step": 850 }, { "epoch": 0.2727993655828707, "grad_norm": 0.24886064231395721, "learning_rate": 4.3e-05, "loss": 0.4657, "step": 860 }, { "epoch": 0.2759714512291832, "grad_norm": 0.3196752369403839, "learning_rate": 4.35e-05, "loss": 0.4652, "step": 870 }, { "epoch": 0.27914353687549565, "grad_norm": 0.27527421712875366, "learning_rate": 4.4000000000000006e-05, "loss": 0.4634, "step": 880 }, { "epoch": 0.2823156225218081, "grad_norm": 0.26819467544555664, "learning_rate": 4.4500000000000004e-05, "loss": 0.4633, "step": 890 }, { "epoch": 0.2854877081681205, "grad_norm": 0.26196032762527466, "learning_rate": 4.5e-05, "loss": 0.4656, "step": 900 }, { "epoch": 0.28865979381443296, "grad_norm": 0.23337453603744507, "learning_rate": 4.55e-05, "loss": 0.4622, "step": 910 }, { "epoch": 0.29183187946074546, "grad_norm": 0.2958989143371582, "learning_rate": 4.600000000000001e-05, "loss": 0.4627, "step": 920 }, { "epoch": 0.2950039651070579, "grad_norm": 0.23167885839939117, "learning_rate": 4.6500000000000005e-05, "loss": 0.4636, "step": 930 }, { "epoch": 0.29817605075337034, "grad_norm": 0.2608003318309784, "learning_rate": 4.7e-05, "loss": 0.4628, "step": 940 }, { "epoch": 0.3013481363996828, "grad_norm": 0.2806340456008911, "learning_rate": 4.75e-05, "loss": 0.4624, "step": 950 }, { "epoch": 0.3045202220459952, "grad_norm": 0.2206449806690216, "learning_rate": 4.8e-05, "loss": 0.4594, "step": 960 }, { "epoch": 0.3076923076923077, "grad_norm": 0.2883213460445404, "learning_rate": 4.85e-05, "loss": 0.4612, "step": 970 }, { "epoch": 0.31086439333862015, "grad_norm": 0.23365403711795807, "learning_rate": 4.9e-05, "loss": 0.463, "step": 980 }, { "epoch": 0.3140364789849326, "grad_norm": 0.28844788670539856, "learning_rate": 4.9500000000000004e-05, "loss": 0.4612, "step": 990 }, { "epoch": 0.317208564631245, "grad_norm": 0.26974356174468994, "learning_rate": 5e-05, "loss": 0.4641, "step": 1000 }, { "epoch": 0.317208564631245, "eval_loss": 0.45910927653312683, "eval_runtime": 419.1696, "eval_samples_per_second": 2.457, "eval_steps_per_second": 1.229, "step": 1000 }, { "epoch": 0.3203806502775575, "grad_norm": 0.23146626353263855, "learning_rate": 4.9997336102492574e-05, "loss": 0.4583, "step": 1010 }, { "epoch": 0.32355273592386996, "grad_norm": 0.24651704728603363, "learning_rate": 4.9989344977678285e-05, "loss": 0.461, "step": 1020 }, { "epoch": 0.3267248215701824, "grad_norm": 0.2995850741863251, "learning_rate": 4.997602832856013e-05, "loss": 0.4583, "step": 1030 }, { "epoch": 0.32989690721649484, "grad_norm": 0.2886098325252533, "learning_rate": 4.995738899307319e-05, "loss": 0.4565, "step": 1040 }, { "epoch": 0.3330689928628073, "grad_norm": 0.2367076575756073, "learning_rate": 4.99334309434798e-05, "loss": 0.4598, "step": 1050 }, { "epoch": 0.3362410785091198, "grad_norm": 0.29130271077156067, "learning_rate": 4.990415928552305e-05, "loss": 0.4564, "step": 1060 }, { "epoch": 0.3394131641554322, "grad_norm": 0.23702391982078552, "learning_rate": 4.9869580257338685e-05, "loss": 0.4567, "step": 1070 }, { "epoch": 0.34258524980174465, "grad_norm": 0.24812078475952148, "learning_rate": 4.9829701228125655e-05, "loss": 0.4587, "step": 1080 }, { "epoch": 0.3457573354480571, "grad_norm": 0.2195880264043808, "learning_rate": 4.9784530696575684e-05, "loss": 0.4578, "step": 1090 }, { "epoch": 0.34892942109436953, "grad_norm": 0.2531008720397949, "learning_rate": 4.973407828906208e-05, "loss": 0.4551, "step": 1100 }, { "epoch": 0.352101506740682, "grad_norm": 0.2672252655029297, "learning_rate": 4.967835475758825e-05, "loss": 0.4596, "step": 1110 }, { "epoch": 0.35527359238699446, "grad_norm": 0.2659941613674164, "learning_rate": 4.961737197749633e-05, "loss": 0.4536, "step": 1120 }, { "epoch": 0.3584456780333069, "grad_norm": 0.2517182230949402, "learning_rate": 4.955114294493639e-05, "loss": 0.4579, "step": 1130 }, { "epoch": 0.36161776367961934, "grad_norm": 0.2654310166835785, "learning_rate": 4.947968177409681e-05, "loss": 0.4548, "step": 1140 }, { "epoch": 0.3647898493259318, "grad_norm": 0.36614441871643066, "learning_rate": 4.940300369419637e-05, "loss": 0.4575, "step": 1150 }, { "epoch": 0.3679619349722443, "grad_norm": 0.2583513855934143, "learning_rate": 4.9321125046238756e-05, "loss": 0.4598, "step": 1160 }, { "epoch": 0.3711340206185567, "grad_norm": 0.23981873691082, "learning_rate": 4.923406327953008e-05, "loss": 0.4546, "step": 1170 }, { "epoch": 0.37430610626486915, "grad_norm": 0.23798957467079163, "learning_rate": 4.9141836947960165e-05, "loss": 0.458, "step": 1180 }, { "epoch": 0.3774781919111816, "grad_norm": 0.2931083142757416, "learning_rate": 4.904446570604863e-05, "loss": 0.4543, "step": 1190 }, { "epoch": 0.38065027755749403, "grad_norm": 0.27312615513801575, "learning_rate": 4.8941970304756144e-05, "loss": 0.4552, "step": 1200 }, { "epoch": 0.3838223632038065, "grad_norm": 0.2659391164779663, "learning_rate": 4.883437258706224e-05, "loss": 0.4583, "step": 1210 }, { "epoch": 0.38699444885011897, "grad_norm": 0.23010990023612976, "learning_rate": 4.8721695483310275e-05, "loss": 0.4564, "step": 1220 }, { "epoch": 0.3901665344964314, "grad_norm": 0.2815942168235779, "learning_rate": 4.860396300632072e-05, "loss": 0.4548, "step": 1230 }, { "epoch": 0.39333862014274384, "grad_norm": 0.36561158299446106, "learning_rate": 4.848120024627372e-05, "loss": 0.4528, "step": 1240 }, { "epoch": 0.3965107057890563, "grad_norm": 0.25015613436698914, "learning_rate": 4.835343336536212e-05, "loss": 0.4567, "step": 1250 }, { "epoch": 0.3996827914353688, "grad_norm": 0.2373395711183548, "learning_rate": 4.822068959221598e-05, "loss": 0.4544, "step": 1260 }, { "epoch": 0.4028548770816812, "grad_norm": 0.26331648230552673, "learning_rate": 4.80829972160998e-05, "loss": 0.4532, "step": 1270 }, { "epoch": 0.40602696272799366, "grad_norm": 0.2622736692428589, "learning_rate": 4.794038558088378e-05, "loss": 0.4535, "step": 1280 }, { "epoch": 0.4091990483743061, "grad_norm": 0.1999824345111847, "learning_rate": 4.779288507879031e-05, "loss": 0.4568, "step": 1290 }, { "epoch": 0.41237113402061853, "grad_norm": 0.24161754548549652, "learning_rate": 4.764052714391695e-05, "loss": 0.4534, "step": 1300 }, { "epoch": 0.41554321966693103, "grad_norm": 0.2368980497121811, "learning_rate": 4.7483344245537545e-05, "loss": 0.4558, "step": 1310 }, { "epoch": 0.41871530531324347, "grad_norm": 0.1948440819978714, "learning_rate": 4.7321369881182584e-05, "loss": 0.4564, "step": 1320 }, { "epoch": 0.4218873909595559, "grad_norm": 0.2509821355342865, "learning_rate": 4.715463856950053e-05, "loss": 0.4558, "step": 1330 }, { "epoch": 0.42505947660586835, "grad_norm": 0.23150520026683807, "learning_rate": 4.698318584290141e-05, "loss": 0.4538, "step": 1340 }, { "epoch": 0.4282315622521808, "grad_norm": 0.22577635943889618, "learning_rate": 4.680704823998452e-05, "loss": 0.4571, "step": 1350 }, { "epoch": 0.4314036478984933, "grad_norm": 0.22882512211799622, "learning_rate": 4.6626263297751546e-05, "loss": 0.453, "step": 1360 }, { "epoch": 0.4345757335448057, "grad_norm": 0.1979285478591919, "learning_rate": 4.644086954360708e-05, "loss": 0.4542, "step": 1370 }, { "epoch": 0.43774781919111816, "grad_norm": 0.2554062008857727, "learning_rate": 4.625090648714786e-05, "loss": 0.4561, "step": 1380 }, { "epoch": 0.4409199048374306, "grad_norm": 0.2330465316772461, "learning_rate": 4.60564146117429e-05, "loss": 0.4525, "step": 1390 }, { "epoch": 0.44409199048374304, "grad_norm": 0.22269243001937866, "learning_rate": 4.585743536590599e-05, "loss": 0.4544, "step": 1400 }, { "epoch": 0.44726407613005553, "grad_norm": 0.2291986495256424, "learning_rate": 4.565401115446246e-05, "loss": 0.4548, "step": 1410 }, { "epoch": 0.45043616177636797, "grad_norm": 0.21675272285938263, "learning_rate": 4.5446185329512314e-05, "loss": 0.4518, "step": 1420 }, { "epoch": 0.4536082474226804, "grad_norm": 0.21207275986671448, "learning_rate": 4.5234002181191304e-05, "loss": 0.4511, "step": 1430 }, { "epoch": 0.45678033306899285, "grad_norm": 0.22509776055812836, "learning_rate": 4.5017506928232247e-05, "loss": 0.4518, "step": 1440 }, { "epoch": 0.4599524187153053, "grad_norm": 0.23429358005523682, "learning_rate": 4.4796745708328294e-05, "loss": 0.4556, "step": 1450 }, { "epoch": 0.4631245043616178, "grad_norm": 0.18599826097488403, "learning_rate": 4.457176556830054e-05, "loss": 0.4538, "step": 1460 }, { "epoch": 0.4662965900079302, "grad_norm": 0.2500051259994507, "learning_rate": 4.434261445407172e-05, "loss": 0.4538, "step": 1470 }, { "epoch": 0.46946867565424266, "grad_norm": 0.213278666138649, "learning_rate": 4.410934120044838e-05, "loss": 0.4545, "step": 1480 }, { "epoch": 0.4726407613005551, "grad_norm": 0.2551578879356384, "learning_rate": 4.387199552071366e-05, "loss": 0.4512, "step": 1490 }, { "epoch": 0.47581284694686754, "grad_norm": 0.23356303572654724, "learning_rate": 4.3630627996032706e-05, "loss": 0.4514, "step": 1500 }, { "epoch": 0.47581284694686754, "eval_loss": 0.4515945613384247, "eval_runtime": 293.5383, "eval_samples_per_second": 3.509, "eval_steps_per_second": 1.754, "step": 1500 }, { "epoch": 0.47898493259318004, "grad_norm": 0.22167843580245972, "learning_rate": 4.3385290064673316e-05, "loss": 0.4529, "step": 1510 }, { "epoch": 0.4821570182394925, "grad_norm": 0.2573811709880829, "learning_rate": 4.31360340110438e-05, "loss": 0.4517, "step": 1520 }, { "epoch": 0.4853291038858049, "grad_norm": 0.20118848979473114, "learning_rate": 4.288291295455055e-05, "loss": 0.4538, "step": 1530 }, { "epoch": 0.48850118953211735, "grad_norm": 0.24297507107257843, "learning_rate": 4.262598083827769e-05, "loss": 0.4519, "step": 1540 }, { "epoch": 0.4916732751784298, "grad_norm": 0.18671482801437378, "learning_rate": 4.236529241749114e-05, "loss": 0.4515, "step": 1550 }, { "epoch": 0.4948453608247423, "grad_norm": 0.20851927995681763, "learning_rate": 4.2100903247969644e-05, "loss": 0.4526, "step": 1560 }, { "epoch": 0.4980174464710547, "grad_norm": 0.23694059252738953, "learning_rate": 4.1832869674165204e-05, "loss": 0.4519, "step": 1570 }, { "epoch": 0.5011895321173672, "grad_norm": 0.3004760146141052, "learning_rate": 4.156124881719533e-05, "loss": 0.4493, "step": 1580 }, { "epoch": 0.5043616177636796, "grad_norm": 0.20880846679210663, "learning_rate": 4.1286098562669925e-05, "loss": 0.4508, "step": 1590 }, { "epoch": 0.507533703409992, "grad_norm": 0.2033112347126007, "learning_rate": 4.1007477548355185e-05, "loss": 0.4522, "step": 1600 }, { "epoch": 0.5107057890563045, "grad_norm": 0.24385987222194672, "learning_rate": 4.072544515167714e-05, "loss": 0.45, "step": 1610 }, { "epoch": 0.5138778747026169, "grad_norm": 0.19505774974822998, "learning_rate": 4.044006147706768e-05, "loss": 0.4528, "step": 1620 }, { "epoch": 0.5170499603489295, "grad_norm": 0.20980720221996307, "learning_rate": 4.0151387343155545e-05, "loss": 0.4512, "step": 1630 }, { "epoch": 0.5202220459952419, "grad_norm": 0.2582622766494751, "learning_rate": 3.985948426980521e-05, "loss": 0.452, "step": 1640 }, { "epoch": 0.5233941316415543, "grad_norm": 0.18902088701725006, "learning_rate": 3.956441446500624e-05, "loss": 0.4498, "step": 1650 }, { "epoch": 0.5265662172878668, "grad_norm": 0.21405866742134094, "learning_rate": 3.926624081161604e-05, "loss": 0.4462, "step": 1660 }, { "epoch": 0.5297383029341792, "grad_norm": 0.21340113878250122, "learning_rate": 3.896502685395876e-05, "loss": 0.449, "step": 1670 }, { "epoch": 0.5329103885804917, "grad_norm": 0.2296970933675766, "learning_rate": 3.8660836784283275e-05, "loss": 0.4513, "step": 1680 }, { "epoch": 0.5360824742268041, "grad_norm": 0.2143164724111557, "learning_rate": 3.835373542908308e-05, "loss": 0.4488, "step": 1690 }, { "epoch": 0.5392545598731165, "grad_norm": 0.1945250779390335, "learning_rate": 3.804378823528093e-05, "loss": 0.4529, "step": 1700 }, { "epoch": 0.542426645519429, "grad_norm": 0.20191755890846252, "learning_rate": 3.7731061256281394e-05, "loss": 0.4507, "step": 1710 }, { "epoch": 0.5455987311657414, "grad_norm": 0.21345514059066772, "learning_rate": 3.7415621137894056e-05, "loss": 0.4487, "step": 1720 }, { "epoch": 0.548770816812054, "grad_norm": 0.2062884271144867, "learning_rate": 3.709753510413052e-05, "loss": 0.4504, "step": 1730 }, { "epoch": 0.5519429024583664, "grad_norm": 0.20377105474472046, "learning_rate": 3.6776870942878196e-05, "loss": 0.4489, "step": 1740 }, { "epoch": 0.5551149881046789, "grad_norm": 0.19985787570476532, "learning_rate": 3.645369699145387e-05, "loss": 0.4497, "step": 1750 }, { "epoch": 0.5582870737509913, "grad_norm": 0.2308078110218048, "learning_rate": 3.6128082122040224e-05, "loss": 0.453, "step": 1760 }, { "epoch": 0.5614591593973037, "grad_norm": 0.19868697226047516, "learning_rate": 3.5800095727008395e-05, "loss": 0.4512, "step": 1770 }, { "epoch": 0.5646312450436162, "grad_norm": 0.20231026411056519, "learning_rate": 3.54698077041296e-05, "loss": 0.4482, "step": 1780 }, { "epoch": 0.5678033306899286, "grad_norm": 0.21103453636169434, "learning_rate": 3.51372884416791e-05, "loss": 0.452, "step": 1790 }, { "epoch": 0.570975416336241, "grad_norm": 0.21350406110286713, "learning_rate": 3.4802608803435646e-05, "loss": 0.4518, "step": 1800 }, { "epoch": 0.5741475019825535, "grad_norm": 0.2247006744146347, "learning_rate": 3.446584011357957e-05, "loss": 0.451, "step": 1810 }, { "epoch": 0.5773195876288659, "grad_norm": 0.20522333681583405, "learning_rate": 3.412705414149276e-05, "loss": 0.4491, "step": 1820 }, { "epoch": 0.5804916732751785, "grad_norm": 0.20249220728874207, "learning_rate": 3.3786323086463736e-05, "loss": 0.4508, "step": 1830 }, { "epoch": 0.5836637589214909, "grad_norm": 0.17756901681423187, "learning_rate": 3.3443719562301147e-05, "loss": 0.4493, "step": 1840 }, { "epoch": 0.5868358445678034, "grad_norm": 0.19246827065944672, "learning_rate": 3.309931658185892e-05, "loss": 0.4501, "step": 1850 }, { "epoch": 0.5900079302141158, "grad_norm": 0.20991584658622742, "learning_rate": 3.275318754147636e-05, "loss": 0.4478, "step": 1860 }, { "epoch": 0.5931800158604282, "grad_norm": 0.21498160064220428, "learning_rate": 3.240540620533649e-05, "loss": 0.4454, "step": 1870 }, { "epoch": 0.5963521015067407, "grad_norm": 0.199667289853096, "learning_rate": 3.205604668974607e-05, "loss": 0.4469, "step": 1880 }, { "epoch": 0.5995241871530531, "grad_norm": 0.21041718125343323, "learning_rate": 3.170518344734051e-05, "loss": 0.4517, "step": 1890 }, { "epoch": 0.6026962727993656, "grad_norm": 0.2041397988796234, "learning_rate": 3.135289125121718e-05, "loss": 0.4518, "step": 1900 }, { "epoch": 0.605868358445678, "grad_norm": 0.18860045075416565, "learning_rate": 3.0999245179000325e-05, "loss": 0.4511, "step": 1910 }, { "epoch": 0.6090404440919904, "grad_norm": 0.20879918336868286, "learning_rate": 3.064432059684117e-05, "loss": 0.4504, "step": 1920 }, { "epoch": 0.612212529738303, "grad_norm": 0.18158119916915894, "learning_rate": 3.0288193143356484e-05, "loss": 0.4501, "step": 1930 }, { "epoch": 0.6153846153846154, "grad_norm": 0.20405800640583038, "learning_rate": 2.9930938713509125e-05, "loss": 0.4478, "step": 1940 }, { "epoch": 0.6185567010309279, "grad_norm": 0.19368775188922882, "learning_rate": 2.9572633442433917e-05, "loss": 0.449, "step": 1950 }, { "epoch": 0.6217287866772403, "grad_norm": 0.19516830146312714, "learning_rate": 2.9213353689212337e-05, "loss": 0.4509, "step": 1960 }, { "epoch": 0.6249008723235527, "grad_norm": 0.2207994908094406, "learning_rate": 2.8853176020599504e-05, "loss": 0.4524, "step": 1970 }, { "epoch": 0.6280729579698652, "grad_norm": 0.20896515250205994, "learning_rate": 2.849217719470691e-05, "loss": 0.447, "step": 1980 }, { "epoch": 0.6312450436161776, "grad_norm": 0.19302915036678314, "learning_rate": 2.8130434144644364e-05, "loss": 0.4462, "step": 1990 }, { "epoch": 0.63441712926249, "grad_norm": 0.21044501662254333, "learning_rate": 2.776802396212461e-05, "loss": 0.4522, "step": 2000 }, { "epoch": 0.63441712926249, "eval_loss": 0.44816499948501587, "eval_runtime": 293.5621, "eval_samples_per_second": 3.509, "eval_steps_per_second": 1.754, "step": 2000 }, { "epoch": 0.6375892149088025, "grad_norm": 0.2075473517179489, "learning_rate": 2.7405023881034204e-05, "loss": 0.4487, "step": 2010 }, { "epoch": 0.640761300555115, "grad_norm": 0.1936168372631073, "learning_rate": 2.7041511260974028e-05, "loss": 0.4485, "step": 2020 }, { "epoch": 0.6439333862014275, "grad_norm": 0.2282039374113083, "learning_rate": 2.66775635707731e-05, "loss": 0.4478, "step": 2030 }, { "epoch": 0.6471054718477399, "grad_norm": 0.18652349710464478, "learning_rate": 2.6313258371978994e-05, "loss": 0.449, "step": 2040 }, { "epoch": 0.6502775574940524, "grad_norm": 0.20407763123512268, "learning_rate": 2.5948673302328587e-05, "loss": 0.4486, "step": 2050 }, { "epoch": 0.6534496431403648, "grad_norm": 0.22227470576763153, "learning_rate": 2.558388605920255e-05, "loss": 0.4506, "step": 2060 }, { "epoch": 0.6566217287866772, "grad_norm": 0.20888665318489075, "learning_rate": 2.5218974383067085e-05, "loss": 0.4487, "step": 2070 }, { "epoch": 0.6597938144329897, "grad_norm": 0.19797295331954956, "learning_rate": 2.4854016040906572e-05, "loss": 0.4521, "step": 2080 }, { "epoch": 0.6629659000793021, "grad_norm": 0.22293215990066528, "learning_rate": 2.4489088809650425e-05, "loss": 0.449, "step": 2090 }, { "epoch": 0.6661379857256146, "grad_norm": 0.2310066670179367, "learning_rate": 2.4124270459598007e-05, "loss": 0.4479, "step": 2100 }, { "epoch": 0.669310071371927, "grad_norm": 0.19995397329330444, "learning_rate": 2.375963873784478e-05, "loss": 0.4504, "step": 2110 }, { "epoch": 0.6724821570182395, "grad_norm": 0.18815317749977112, "learning_rate": 2.3395271351713515e-05, "loss": 0.4501, "step": 2120 }, { "epoch": 0.675654242664552, "grad_norm": 0.19861653447151184, "learning_rate": 2.303124595219395e-05, "loss": 0.4506, "step": 2130 }, { "epoch": 0.6788263283108644, "grad_norm": 0.17454959452152252, "learning_rate": 2.2667640117394442e-05, "loss": 0.4453, "step": 2140 }, { "epoch": 0.6819984139571769, "grad_norm": 0.19067569077014923, "learning_rate": 2.230453133600915e-05, "loss": 0.4466, "step": 2150 }, { "epoch": 0.6851704996034893, "grad_norm": 0.1979096531867981, "learning_rate": 2.1941996990804288e-05, "loss": 0.4479, "step": 2160 }, { "epoch": 0.6883425852498017, "grad_norm": 0.18306080996990204, "learning_rate": 2.1580114342126933e-05, "loss": 0.4482, "step": 2170 }, { "epoch": 0.6915146708961142, "grad_norm": 0.18473172187805176, "learning_rate": 2.1218960511439952e-05, "loss": 0.4468, "step": 2180 }, { "epoch": 0.6946867565424266, "grad_norm": 0.17530548572540283, "learning_rate": 2.0858612464886505e-05, "loss": 0.449, "step": 2190 }, { "epoch": 0.6978588421887391, "grad_norm": 0.1818351149559021, "learning_rate": 2.0499146996887618e-05, "loss": 0.4458, "step": 2200 }, { "epoch": 0.7010309278350515, "grad_norm": 0.19750525057315826, "learning_rate": 2.0140640713776438e-05, "loss": 0.4489, "step": 2210 }, { "epoch": 0.704203013481364, "grad_norm": 0.20060710608959198, "learning_rate": 1.978317001747248e-05, "loss": 0.4482, "step": 2220 }, { "epoch": 0.7073750991276765, "grad_norm": 0.19399641454219818, "learning_rate": 1.942681108919949e-05, "loss": 0.4494, "step": 2230 }, { "epoch": 0.7105471847739889, "grad_norm": 0.2124367654323578, "learning_rate": 1.9071639873250334e-05, "loss": 0.4491, "step": 2240 }, { "epoch": 0.7137192704203014, "grad_norm": 0.2064894288778305, "learning_rate": 1.871773206080236e-05, "loss": 0.4433, "step": 2250 }, { "epoch": 0.7168913560666138, "grad_norm": 0.20361967384815216, "learning_rate": 1.836516307378671e-05, "loss": 0.4482, "step": 2260 }, { "epoch": 0.7200634417129262, "grad_norm": 0.22753477096557617, "learning_rate": 1.801400804881507e-05, "loss": 0.4434, "step": 2270 }, { "epoch": 0.7232355273592387, "grad_norm": 0.2242439240217209, "learning_rate": 1.766434182116708e-05, "loss": 0.4487, "step": 2280 }, { "epoch": 0.7264076130055511, "grad_norm": 0.1731296181678772, "learning_rate": 1.7316238908842127e-05, "loss": 0.4456, "step": 2290 }, { "epoch": 0.7295796986518636, "grad_norm": 0.19372820854187012, "learning_rate": 1.6969773496678647e-05, "loss": 0.446, "step": 2300 }, { "epoch": 0.732751784298176, "grad_norm": 0.21722117066383362, "learning_rate": 1.662501942054447e-05, "loss": 0.4465, "step": 2310 }, { "epoch": 0.7359238699444886, "grad_norm": 0.24181324243545532, "learning_rate": 1.6282050151601518e-05, "loss": 0.4496, "step": 2320 }, { "epoch": 0.739095955590801, "grad_norm": 0.2091435343027115, "learning_rate": 1.594093878064825e-05, "loss": 0.4469, "step": 2330 }, { "epoch": 0.7422680412371134, "grad_norm": 0.19474650919437408, "learning_rate": 1.5601758002543137e-05, "loss": 0.4465, "step": 2340 }, { "epoch": 0.7454401268834259, "grad_norm": 0.21081089973449707, "learning_rate": 1.5264580100712507e-05, "loss": 0.4449, "step": 2350 }, { "epoch": 0.7486122125297383, "grad_norm": 0.21220536530017853, "learning_rate": 1.4929476931746167e-05, "loss": 0.4461, "step": 2360 }, { "epoch": 0.7517842981760507, "grad_norm": 0.19944216310977936, "learning_rate": 1.4596519910083825e-05, "loss": 0.4476, "step": 2370 }, { "epoch": 0.7549563838223632, "grad_norm": 0.19958476722240448, "learning_rate": 1.4265779992795893e-05, "loss": 0.4467, "step": 2380 }, { "epoch": 0.7581284694686756, "grad_norm": 0.22247722744941711, "learning_rate": 1.3937327664461672e-05, "loss": 0.4427, "step": 2390 }, { "epoch": 0.7613005551149881, "grad_norm": 0.19884702563285828, "learning_rate": 1.361123292214826e-05, "loss": 0.4498, "step": 2400 }, { "epoch": 0.7644726407613005, "grad_norm": 0.17867441475391388, "learning_rate": 1.3287565260493357e-05, "loss": 0.4484, "step": 2410 }, { "epoch": 0.767644726407613, "grad_norm": 0.17917264997959137, "learning_rate": 1.2966393656895134e-05, "loss": 0.4439, "step": 2420 }, { "epoch": 0.7708168120539255, "grad_norm": 0.24791833758354187, "learning_rate": 1.2647786556812332e-05, "loss": 0.4463, "step": 2430 }, { "epoch": 0.7739888977002379, "grad_norm": 0.1971379816532135, "learning_rate": 1.2331811859177722e-05, "loss": 0.4475, "step": 2440 }, { "epoch": 0.7771609833465504, "grad_norm": 0.19460198283195496, "learning_rate": 1.2018536901928079e-05, "loss": 0.4468, "step": 2450 }, { "epoch": 0.7803330689928628, "grad_norm": 0.21018314361572266, "learning_rate": 1.1708028447653613e-05, "loss": 0.4489, "step": 2460 }, { "epoch": 0.7835051546391752, "grad_norm": 0.18419425189495087, "learning_rate": 1.1400352669370115e-05, "loss": 0.4469, "step": 2470 }, { "epoch": 0.7866772402854877, "grad_norm": 0.192842036485672, "learning_rate": 1.1095575136416695e-05, "loss": 0.4457, "step": 2480 }, { "epoch": 0.7898493259318001, "grad_norm": 0.21617653965950012, "learning_rate": 1.0793760800482179e-05, "loss": 0.4511, "step": 2490 }, { "epoch": 0.7930214115781126, "grad_norm": 0.1891343593597412, "learning_rate": 1.0494973981763146e-05, "loss": 0.4436, "step": 2500 }, { "epoch": 0.7930214115781126, "eval_loss": 0.44588717818260193, "eval_runtime": 293.564, "eval_samples_per_second": 3.509, "eval_steps_per_second": 1.754, "step": 2500 }, { "epoch": 0.796193497224425, "grad_norm": 0.193357452750206, "learning_rate": 1.0199278355256522e-05, "loss": 0.4436, "step": 2510 }, { "epoch": 0.7993655828707376, "grad_norm": 0.19498059153556824, "learning_rate": 9.906736937189697e-06, "loss": 0.4457, "step": 2520 }, { "epoch": 0.80253766851705, "grad_norm": 0.2086293250322342, "learning_rate": 9.61741207159104e-06, "loss": 0.4488, "step": 2530 }, { "epoch": 0.8057097541633624, "grad_norm": 0.1872304081916809, "learning_rate": 9.331365417003601e-06, "loss": 0.4436, "step": 2540 }, { "epoch": 0.8088818398096749, "grad_norm": 0.196652352809906, "learning_rate": 9.048657933345e-06, "loss": 0.4448, "step": 2550 }, { "epoch": 0.8120539254559873, "grad_norm": 0.21112217009067535, "learning_rate": 8.769349868916118e-06, "loss": 0.4481, "step": 2560 }, { "epoch": 0.8152260111022998, "grad_norm": 0.19768798351287842, "learning_rate": 8.493500747561488e-06, "loss": 0.4493, "step": 2570 }, { "epoch": 0.8183980967486122, "grad_norm": 0.19622166454792023, "learning_rate": 8.221169355984052e-06, "loss": 0.4448, "step": 2580 }, { "epoch": 0.8215701823949246, "grad_norm": 0.2044159471988678, "learning_rate": 7.952413731217025e-06, "loss": 0.4472, "step": 2590 }, { "epoch": 0.8247422680412371, "grad_norm": 0.206452876329422, "learning_rate": 7.687291148255527e-06, "loss": 0.4452, "step": 2600 }, { "epoch": 0.8279143536875495, "grad_norm": 0.20858535170555115, "learning_rate": 7.425858107850578e-06, "loss": 0.4419, "step": 2610 }, { "epoch": 0.8310864393338621, "grad_norm": 0.18986332416534424, "learning_rate": 7.168170324468171e-06, "loss": 0.449, "step": 2620 }, { "epoch": 0.8342585249801745, "grad_norm": 0.21051757037639618, "learning_rate": 6.9142827144158066e-06, "loss": 0.4468, "step": 2630 }, { "epoch": 0.8374306106264869, "grad_norm": 0.20453451573848724, "learning_rate": 6.66424938413921e-06, "loss": 0.4471, "step": 2640 }, { "epoch": 0.8406026962727994, "grad_norm": 0.18574683368206024, "learning_rate": 6.418123618691607e-06, "loss": 0.4473, "step": 2650 }, { "epoch": 0.8437747819191118, "grad_norm": 0.2309403419494629, "learning_rate": 6.175957870378043e-06, "loss": 0.4477, "step": 2660 }, { "epoch": 0.8469468675654243, "grad_norm": 0.19485293328762054, "learning_rate": 5.937803747577186e-06, "loss": 0.4443, "step": 2670 }, { "epoch": 0.8501189532117367, "grad_norm": 0.17677082121372223, "learning_rate": 5.7037120037429645e-06, "loss": 0.4501, "step": 2680 }, { "epoch": 0.8532910388580491, "grad_norm": 0.18835794925689697, "learning_rate": 5.473732526588407e-06, "loss": 0.4451, "step": 2690 }, { "epoch": 0.8564631245043616, "grad_norm": 0.2117832899093628, "learning_rate": 5.2479143274539954e-06, "loss": 0.4464, "step": 2700 }, { "epoch": 0.8596352101506741, "grad_norm": 0.18720006942749023, "learning_rate": 5.026305530862749e-06, "loss": 0.4484, "step": 2710 }, { "epoch": 0.8628072957969866, "grad_norm": 0.17910288274288177, "learning_rate": 4.80895336426434e-06, "loss": 0.4457, "step": 2720 }, { "epoch": 0.865979381443299, "grad_norm": 0.1965160220861435, "learning_rate": 4.595904147970356e-06, "loss": 0.4455, "step": 2730 }, { "epoch": 0.8691514670896114, "grad_norm": 0.16718564927577972, "learning_rate": 4.3872032852828955e-06, "loss": 0.4455, "step": 2740 }, { "epoch": 0.8723235527359239, "grad_norm": 0.19451048970222473, "learning_rate": 4.182895252818589e-06, "loss": 0.4454, "step": 2750 }, { "epoch": 0.8754956383822363, "grad_norm": 0.2160135954618454, "learning_rate": 3.983023591030113e-06, "loss": 0.4462, "step": 2760 }, { "epoch": 0.8786677240285488, "grad_norm": 0.1994054913520813, "learning_rate": 3.7876308949271995e-06, "loss": 0.4455, "step": 2770 }, { "epoch": 0.8818398096748612, "grad_norm": 0.20801877975463867, "learning_rate": 3.5967588049991313e-06, "loss": 0.4465, "step": 2780 }, { "epoch": 0.8850118953211736, "grad_norm": 0.20264749228954315, "learning_rate": 3.410447998340688e-06, "loss": 0.4451, "step": 2790 }, { "epoch": 0.8881839809674861, "grad_norm": 0.173824280500412, "learning_rate": 3.2287381799833428e-06, "loss": 0.4453, "step": 2800 }, { "epoch": 0.8913560666137986, "grad_norm": 0.18437042832374573, "learning_rate": 3.051668074433667e-06, "loss": 0.449, "step": 2810 }, { "epoch": 0.8945281522601111, "grad_norm": 0.17420920729637146, "learning_rate": 2.8792754174206905e-06, "loss": 0.4453, "step": 2820 }, { "epoch": 0.8977002379064235, "grad_norm": 0.18584947288036346, "learning_rate": 2.7115969478539562e-06, "loss": 0.4446, "step": 2830 }, { "epoch": 0.9008723235527359, "grad_norm": 0.18006384372711182, "learning_rate": 2.5486683999940335e-06, "loss": 0.4449, "step": 2840 }, { "epoch": 0.9040444091990484, "grad_norm": 0.19664862751960754, "learning_rate": 2.3905244958371263e-06, "loss": 0.4463, "step": 2850 }, { "epoch": 0.9072164948453608, "grad_norm": 0.20159141719341278, "learning_rate": 2.2371989377154013e-06, "loss": 0.444, "step": 2860 }, { "epoch": 0.9103885804916733, "grad_norm": 0.23178640007972717, "learning_rate": 2.088724401114625e-06, "loss": 0.4472, "step": 2870 }, { "epoch": 0.9135606661379857, "grad_norm": 0.18047955632209778, "learning_rate": 1.9451325277106413e-06, "loss": 0.4489, "step": 2880 }, { "epoch": 0.9167327517842981, "grad_norm": 0.18607738614082336, "learning_rate": 1.8064539186261387e-06, "loss": 0.4489, "step": 2890 }, { "epoch": 0.9199048374306106, "grad_norm": 0.19285809993743896, "learning_rate": 1.6727181279092036e-06, "loss": 0.4448, "step": 2900 }, { "epoch": 0.9230769230769231, "grad_norm": 0.1946035474538803, "learning_rate": 1.5439536562349976e-06, "loss": 0.4474, "step": 2910 }, { "epoch": 0.9262490087232356, "grad_norm": 0.1947924941778183, "learning_rate": 1.4201879448319355e-06, "loss": 0.4465, "step": 2920 }, { "epoch": 0.929421094369548, "grad_norm": 0.17740875482559204, "learning_rate": 1.301447369633621e-06, "loss": 0.4475, "step": 2930 }, { "epoch": 0.9325931800158604, "grad_norm": 0.18701794743537903, "learning_rate": 1.187757235657841e-06, "loss": 0.4488, "step": 2940 }, { "epoch": 0.9357652656621729, "grad_norm": 0.19531536102294922, "learning_rate": 1.0791417716137565e-06, "loss": 0.4458, "step": 2950 }, { "epoch": 0.9389373513084853, "grad_norm": 0.17441792786121368, "learning_rate": 9.756241247384807e-07, "loss": 0.4426, "step": 2960 }, { "epoch": 0.9421094369547978, "grad_norm": 0.20463122427463531, "learning_rate": 8.772263558641386e-07, "loss": 0.4485, "step": 2970 }, { "epoch": 0.9452815226011102, "grad_norm": 0.18601520359516144, "learning_rate": 7.839694347164223e-07, "loss": 0.4421, "step": 2980 }, { "epoch": 0.9484536082474226, "grad_norm": 0.18209311366081238, "learning_rate": 6.958732354457209e-07, "loss": 0.4499, "step": 2990 }, { "epoch": 0.9516256938937351, "grad_norm": 0.1675465852022171, "learning_rate": 6.129565323916813e-07, "loss": 0.4463, "step": 3000 }, { "epoch": 0.9516256938937351, "eval_loss": 0.44520583748817444, "eval_runtime": 293.5596, "eval_samples_per_second": 3.509, "eval_steps_per_second": 1.754, "step": 3000 }, { "epoch": 0.9547977795400476, "grad_norm": 0.21267291903495789, "learning_rate": 5.352369960821946e-07, "loss": 0.4428, "step": 3010 }, { "epoch": 0.9579698651863601, "grad_norm": 0.19138512015342712, "learning_rate": 4.627311894675856e-07, "loss": 0.446, "step": 3020 }, { "epoch": 0.9611419508326725, "grad_norm": 0.19043755531311035, "learning_rate": 3.954545643908514e-07, "loss": 0.4446, "step": 3030 }, { "epoch": 0.964314036478985, "grad_norm": 0.1850534826517105, "learning_rate": 3.3342145829469983e-07, "loss": 0.4435, "step": 3040 }, { "epoch": 0.9674861221252974, "grad_norm": 0.21564887464046478, "learning_rate": 2.7664509116607506e-07, "loss": 0.4451, "step": 3050 }, { "epoch": 0.9706582077716098, "grad_norm": 0.1970747411251068, "learning_rate": 2.251375627187996e-07, "loss": 0.4465, "step": 3060 }, { "epoch": 0.9738302934179223, "grad_norm": 0.1968993842601776, "learning_rate": 1.789098498150066e-07, "loss": 0.4479, "step": 3070 }, { "epoch": 0.9770023790642347, "grad_norm": 0.18383654952049255, "learning_rate": 1.3797180412583322e-07, "loss": 0.4472, "step": 3080 }, { "epoch": 0.9801744647105471, "grad_norm": 0.18183210492134094, "learning_rate": 1.0233215003190577e-07, "loss": 0.4484, "step": 3090 }, { "epoch": 0.9833465503568596, "grad_norm": 0.18428853154182434, "learning_rate": 7.199848276408238e-08, "loss": 0.4445, "step": 3100 }, { "epoch": 0.9865186360031721, "grad_norm": 0.21674080193042755, "learning_rate": 4.6977266784811736e-08, "loss": 0.4465, "step": 3110 }, { "epoch": 0.9896907216494846, "grad_norm": 0.2023499608039856, "learning_rate": 2.7273834410485034e-08, "loss": 0.4448, "step": 3120 }, { "epoch": 0.992862807295797, "grad_norm": 0.19151511788368225, "learning_rate": 1.2892384675056158e-08, "loss": 0.4461, "step": 3130 }, { "epoch": 0.9960348929421095, "grad_norm": 0.16607537865638733, "learning_rate": 3.8359824351685836e-09, "loss": 0.4453, "step": 3140 }, { "epoch": 0.9992069785884219, "grad_norm": 0.21372340619564056, "learning_rate": 1.0655771701395534e-10, "loss": 0.448, "step": 3150 }, { "epoch": 0.9998413957176844, "step": 3152, "total_flos": 2.0499162760961065e+18, "train_loss": 0.5496831483344742, "train_runtime": 71646.9274, "train_samples_per_second": 0.704, "train_steps_per_second": 0.044 } ], "logging_steps": 10, "max_steps": 3152, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 500, "total_flos": 2.0499162760961065e+18, "train_batch_size": 4, "trial_name": null, "trial_params": null }