{ "best_metric": 0.018277771770954132, "best_model_checkpoint": "runs/deepseek_lora_20240424-122712/checkpoint-2000", "epoch": 0.9399232396020991, "eval_steps": 500, "global_step": 3000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0, "grad_norm": 2.6060585975646973, "learning_rate": 4.0000000000000003e-07, "loss": 1.6667, "step": 10 }, { "epoch": 0.01, "grad_norm": 5.726912021636963, "learning_rate": 8.000000000000001e-07, "loss": 1.6606, "step": 20 }, { "epoch": 0.01, "grad_norm": 6.3224616050720215, "learning_rate": 1.2000000000000002e-06, "loss": 1.6875, "step": 30 }, { "epoch": 0.01, "grad_norm": 4.4585185050964355, "learning_rate": 1.6000000000000001e-06, "loss": 1.5512, "step": 40 }, { "epoch": 0.02, "grad_norm": 2.849210023880005, "learning_rate": 2.0000000000000003e-06, "loss": 1.598, "step": 50 }, { "epoch": 0.02, "grad_norm": 1.7917169332504272, "learning_rate": 2.4000000000000003e-06, "loss": 1.3203, "step": 60 }, { "epoch": 0.02, "grad_norm": 10.056238174438477, "learning_rate": 2.8000000000000003e-06, "loss": 1.4914, "step": 70 }, { "epoch": 0.03, "grad_norm": 5.542996406555176, "learning_rate": 3.2000000000000003e-06, "loss": 1.3369, "step": 80 }, { "epoch": 0.03, "grad_norm": 1.5463522672653198, "learning_rate": 3.6000000000000003e-06, "loss": 1.3185, "step": 90 }, { "epoch": 0.03, "grad_norm": 9.401649475097656, "learning_rate": 4.000000000000001e-06, "loss": 1.2827, "step": 100 }, { "epoch": 0.03, "grad_norm": 1.238974928855896, "learning_rate": 4.4e-06, "loss": 1.0366, "step": 110 }, { "epoch": 0.04, "grad_norm": 1.2282520532608032, "learning_rate": 4.800000000000001e-06, "loss": 1.0715, "step": 120 }, { "epoch": 0.04, "grad_norm": 1.4583073854446411, "learning_rate": 5.2e-06, "loss": 0.868, "step": 130 }, { "epoch": 0.04, "grad_norm": 4.3194146156311035, "learning_rate": 5.600000000000001e-06, "loss": 0.8281, "step": 140 }, { "epoch": 0.05, "grad_norm": 4.280028820037842, "learning_rate": 6e-06, "loss": 0.8666, "step": 150 }, { "epoch": 0.05, "grad_norm": 0.9424476027488708, "learning_rate": 6.4000000000000006e-06, "loss": 0.6691, "step": 160 }, { "epoch": 0.05, "grad_norm": 0.9174453616142273, "learning_rate": 6.800000000000001e-06, "loss": 0.5405, "step": 170 }, { "epoch": 0.06, "grad_norm": 1.0179359912872314, "learning_rate": 7.2000000000000005e-06, "loss": 0.5103, "step": 180 }, { "epoch": 0.06, "grad_norm": 0.4516351521015167, "learning_rate": 7.600000000000001e-06, "loss": 0.2679, "step": 190 }, { "epoch": 0.06, "grad_norm": 6.054721355438232, "learning_rate": 8.000000000000001e-06, "loss": 0.2634, "step": 200 }, { "epoch": 0.07, "grad_norm": 0.40070173144340515, "learning_rate": 8.400000000000001e-06, "loss": 0.2254, "step": 210 }, { "epoch": 0.07, "grad_norm": 0.6031996011734009, "learning_rate": 8.8e-06, "loss": 0.1281, "step": 220 }, { "epoch": 0.07, "grad_norm": 3.0129740238189697, "learning_rate": 9.200000000000002e-06, "loss": 0.2365, "step": 230 }, { "epoch": 0.08, "grad_norm": 0.6611989140510559, "learning_rate": 9.600000000000001e-06, "loss": 0.3255, "step": 240 }, { "epoch": 0.08, "grad_norm": 0.3195549547672272, "learning_rate": 1e-05, "loss": 0.3287, "step": 250 }, { "epoch": 0.08, "grad_norm": 0.27022993564605713, "learning_rate": 1.04e-05, "loss": 0.1439, "step": 260 }, { "epoch": 0.08, "grad_norm": 4.671290874481201, "learning_rate": 1.0800000000000002e-05, "loss": 0.0676, "step": 270 }, { "epoch": 0.09, "grad_norm": 0.2435157746076584, "learning_rate": 1.1200000000000001e-05, "loss": 0.2338, "step": 280 }, { "epoch": 0.09, "grad_norm": 0.40632137656211853, "learning_rate": 1.16e-05, "loss": 0.1661, "step": 290 }, { "epoch": 0.09, "grad_norm": 0.2910804748535156, "learning_rate": 1.2e-05, "loss": 0.1781, "step": 300 }, { "epoch": 0.1, "grad_norm": 7.506131649017334, "learning_rate": 1.2400000000000002e-05, "loss": 0.1283, "step": 310 }, { "epoch": 0.1, "grad_norm": 0.04508267343044281, "learning_rate": 1.2800000000000001e-05, "loss": 0.1022, "step": 320 }, { "epoch": 0.1, "grad_norm": 0.23127306997776031, "learning_rate": 1.3200000000000002e-05, "loss": 0.1243, "step": 330 }, { "epoch": 0.11, "grad_norm": 6.1613383293151855, "learning_rate": 1.3600000000000002e-05, "loss": 0.1485, "step": 340 }, { "epoch": 0.11, "grad_norm": 0.023777758702635765, "learning_rate": 1.4e-05, "loss": 0.0538, "step": 350 }, { "epoch": 0.11, "grad_norm": 0.014769719913601875, "learning_rate": 1.4400000000000001e-05, "loss": 0.0285, "step": 360 }, { "epoch": 0.12, "grad_norm": 5.690178871154785, "learning_rate": 1.48e-05, "loss": 0.1325, "step": 370 }, { "epoch": 0.12, "grad_norm": 0.02117346040904522, "learning_rate": 1.5200000000000002e-05, "loss": 0.0507, "step": 380 }, { "epoch": 0.12, "grad_norm": 0.06630611419677734, "learning_rate": 1.5600000000000003e-05, "loss": 0.199, "step": 390 }, { "epoch": 0.13, "grad_norm": 0.16532257199287415, "learning_rate": 1.6000000000000003e-05, "loss": 0.0652, "step": 400 }, { "epoch": 0.13, "grad_norm": 0.009799620136618614, "learning_rate": 1.64e-05, "loss": 0.1011, "step": 410 }, { "epoch": 0.13, "grad_norm": 0.12324853241443634, "learning_rate": 1.6800000000000002e-05, "loss": 0.0364, "step": 420 }, { "epoch": 0.13, "grad_norm": 0.010026533156633377, "learning_rate": 1.72e-05, "loss": 0.1973, "step": 430 }, { "epoch": 0.14, "grad_norm": 0.0071570691652596, "learning_rate": 1.76e-05, "loss": 0.0819, "step": 440 }, { "epoch": 0.14, "grad_norm": 0.04149964451789856, "learning_rate": 1.8e-05, "loss": 0.1046, "step": 450 }, { "epoch": 0.14, "grad_norm": 3.45990252494812, "learning_rate": 1.8400000000000003e-05, "loss": 0.1011, "step": 460 }, { "epoch": 0.15, "grad_norm": 0.00734300771728158, "learning_rate": 1.88e-05, "loss": 0.223, "step": 470 }, { "epoch": 0.15, "grad_norm": 0.01419814396649599, "learning_rate": 1.9200000000000003e-05, "loss": 0.1178, "step": 480 }, { "epoch": 0.15, "grad_norm": 0.10111651569604874, "learning_rate": 1.9600000000000002e-05, "loss": 0.1416, "step": 490 }, { "epoch": 0.16, "grad_norm": 3.1214537620544434, "learning_rate": 2e-05, "loss": 0.1258, "step": 500 }, { "epoch": 0.16, "eval_loss": 0.05367860943078995, "eval_runtime": 62.0108, "eval_samples_per_second": 16.126, "eval_steps_per_second": 16.126, "step": 500 }, { "epoch": 0.16, "grad_norm": 0.013164438307285309, "learning_rate": 1.9955555555555557e-05, "loss": 0.0268, "step": 510 }, { "epoch": 0.16, "grad_norm": 0.15392057597637177, "learning_rate": 1.9911111111111112e-05, "loss": 0.0843, "step": 520 }, { "epoch": 0.17, "grad_norm": 0.006180985830724239, "learning_rate": 1.9866666666666667e-05, "loss": 0.1384, "step": 530 }, { "epoch": 0.17, "grad_norm": 0.2702454924583435, "learning_rate": 1.9822222222222226e-05, "loss": 0.0956, "step": 540 }, { "epoch": 0.17, "grad_norm": 1.5135071277618408, "learning_rate": 1.977777777777778e-05, "loss": 0.1615, "step": 550 }, { "epoch": 0.18, "grad_norm": 0.005121675785630941, "learning_rate": 1.9733333333333336e-05, "loss": 0.1201, "step": 560 }, { "epoch": 0.18, "grad_norm": 0.09728775173425674, "learning_rate": 1.968888888888889e-05, "loss": 0.1112, "step": 570 }, { "epoch": 0.18, "grad_norm": 4.736963748931885, "learning_rate": 1.9644444444444447e-05, "loss": 0.1546, "step": 580 }, { "epoch": 0.18, "grad_norm": 0.006898015271872282, "learning_rate": 1.9600000000000002e-05, "loss": 0.1116, "step": 590 }, { "epoch": 0.19, "grad_norm": 0.005388608202338219, "learning_rate": 1.9555555555555557e-05, "loss": 0.1597, "step": 600 }, { "epoch": 0.19, "grad_norm": 0.0133949751034379, "learning_rate": 1.9511111111111113e-05, "loss": 0.0799, "step": 610 }, { "epoch": 0.19, "grad_norm": 0.11226101964712143, "learning_rate": 1.9466666666666668e-05, "loss": 0.0448, "step": 620 }, { "epoch": 0.2, "grad_norm": 0.011205198243260384, "learning_rate": 1.9422222222222223e-05, "loss": 0.0739, "step": 630 }, { "epoch": 0.2, "grad_norm": 0.007312687113881111, "learning_rate": 1.9377777777777778e-05, "loss": 0.0779, "step": 640 }, { "epoch": 0.2, "grad_norm": 0.006550287362188101, "learning_rate": 1.9333333333333333e-05, "loss": 0.2039, "step": 650 }, { "epoch": 0.21, "grad_norm": 1.961844563484192, "learning_rate": 1.928888888888889e-05, "loss": 0.0618, "step": 660 }, { "epoch": 0.21, "grad_norm": 0.012318034656345844, "learning_rate": 1.9244444444444444e-05, "loss": 0.1166, "step": 670 }, { "epoch": 0.21, "grad_norm": 0.08778905868530273, "learning_rate": 1.9200000000000003e-05, "loss": 0.0822, "step": 680 }, { "epoch": 0.22, "grad_norm": 0.010623461566865444, "learning_rate": 1.9155555555555558e-05, "loss": 0.095, "step": 690 }, { "epoch": 0.22, "grad_norm": 0.0858495905995369, "learning_rate": 1.9111111111111113e-05, "loss": 0.0468, "step": 700 }, { "epoch": 0.22, "grad_norm": 1.7685880661010742, "learning_rate": 1.9066666666666668e-05, "loss": 0.0288, "step": 710 }, { "epoch": 0.23, "grad_norm": 0.032682083547115326, "learning_rate": 1.9022222222222223e-05, "loss": 0.1024, "step": 720 }, { "epoch": 0.23, "grad_norm": 0.010010900907218456, "learning_rate": 1.897777777777778e-05, "loss": 0.0129, "step": 730 }, { "epoch": 0.23, "grad_norm": 0.037693481892347336, "learning_rate": 1.8933333333333334e-05, "loss": 0.1217, "step": 740 }, { "epoch": 0.23, "grad_norm": 2.2111008167266846, "learning_rate": 1.888888888888889e-05, "loss": 0.1206, "step": 750 }, { "epoch": 0.24, "grad_norm": 1.4000599384307861, "learning_rate": 1.8844444444444444e-05, "loss": 0.1214, "step": 760 }, { "epoch": 0.24, "grad_norm": 4.3877034187316895, "learning_rate": 1.88e-05, "loss": 0.0408, "step": 770 }, { "epoch": 0.24, "grad_norm": 4.164296627044678, "learning_rate": 1.8755555555555558e-05, "loss": 0.1315, "step": 780 }, { "epoch": 0.25, "grad_norm": 0.07969211786985397, "learning_rate": 1.8711111111111113e-05, "loss": 0.0149, "step": 790 }, { "epoch": 0.25, "grad_norm": 3.9340951442718506, "learning_rate": 1.866666666666667e-05, "loss": 0.1169, "step": 800 }, { "epoch": 0.25, "grad_norm": 0.006517359986901283, "learning_rate": 1.8622222222222224e-05, "loss": 0.0517, "step": 810 }, { "epoch": 0.26, "grad_norm": 0.008951540105044842, "learning_rate": 1.857777777777778e-05, "loss": 0.0935, "step": 820 }, { "epoch": 0.26, "grad_norm": 0.03655437007546425, "learning_rate": 1.8533333333333334e-05, "loss": 0.0883, "step": 830 }, { "epoch": 0.26, "grad_norm": 0.0346522182226181, "learning_rate": 1.848888888888889e-05, "loss": 0.0479, "step": 840 }, { "epoch": 0.27, "grad_norm": 0.006223162170499563, "learning_rate": 1.8444444444444448e-05, "loss": 0.0075, "step": 850 }, { "epoch": 0.27, "grad_norm": 0.004037676844745874, "learning_rate": 1.8400000000000003e-05, "loss": 0.0333, "step": 860 }, { "epoch": 0.27, "grad_norm": 0.004569903016090393, "learning_rate": 1.835555555555556e-05, "loss": 0.0536, "step": 870 }, { "epoch": 0.28, "grad_norm": 0.005509174894541502, "learning_rate": 1.8311111111111114e-05, "loss": 0.1528, "step": 880 }, { "epoch": 0.28, "grad_norm": 0.007597978692501783, "learning_rate": 1.826666666666667e-05, "loss": 0.0052, "step": 890 }, { "epoch": 0.28, "grad_norm": 4.2217936515808105, "learning_rate": 1.8222222222222224e-05, "loss": 0.0776, "step": 900 }, { "epoch": 0.29, "grad_norm": 0.004342419560998678, "learning_rate": 1.817777777777778e-05, "loss": 0.0552, "step": 910 }, { "epoch": 0.29, "grad_norm": 0.0696156695485115, "learning_rate": 1.8133333333333335e-05, "loss": 0.0412, "step": 920 }, { "epoch": 0.29, "grad_norm": 0.022448547184467316, "learning_rate": 1.808888888888889e-05, "loss": 0.0109, "step": 930 }, { "epoch": 0.29, "grad_norm": 4.305741310119629, "learning_rate": 1.8044444444444445e-05, "loss": 0.0428, "step": 940 }, { "epoch": 0.3, "grad_norm": 3.344078779220581, "learning_rate": 1.8e-05, "loss": 0.0651, "step": 950 }, { "epoch": 0.3, "grad_norm": 2.585327386856079, "learning_rate": 1.7955555555555556e-05, "loss": 0.1474, "step": 960 }, { "epoch": 0.3, "grad_norm": 2.2476861476898193, "learning_rate": 1.791111111111111e-05, "loss": 0.0892, "step": 970 }, { "epoch": 0.31, "grad_norm": 0.06212488189339638, "learning_rate": 1.7866666666666666e-05, "loss": 0.0659, "step": 980 }, { "epoch": 0.31, "grad_norm": 0.012895594350993633, "learning_rate": 1.782222222222222e-05, "loss": 0.0915, "step": 990 }, { "epoch": 0.31, "grad_norm": 0.1314150094985962, "learning_rate": 1.7777777777777777e-05, "loss": 0.0456, "step": 1000 }, { "epoch": 0.31, "eval_loss": 0.02997952327132225, "eval_runtime": 62.0027, "eval_samples_per_second": 16.128, "eval_steps_per_second": 16.128, "step": 1000 }, { "epoch": 0.32, "grad_norm": 0.01634177938103676, "learning_rate": 1.7733333333333335e-05, "loss": 0.0424, "step": 1010 }, { "epoch": 0.32, "grad_norm": 0.009437276981770992, "learning_rate": 1.768888888888889e-05, "loss": 0.0663, "step": 1020 }, { "epoch": 0.32, "grad_norm": 0.04742557182908058, "learning_rate": 1.7644444444444446e-05, "loss": 0.1235, "step": 1030 }, { "epoch": 0.33, "grad_norm": 1.101121187210083, "learning_rate": 1.76e-05, "loss": 0.0574, "step": 1040 }, { "epoch": 0.33, "grad_norm": 3.8775153160095215, "learning_rate": 1.7555555555555556e-05, "loss": 0.0916, "step": 1050 }, { "epoch": 0.33, "grad_norm": 0.0064536286517977715, "learning_rate": 1.751111111111111e-05, "loss": 0.0363, "step": 1060 }, { "epoch": 0.34, "grad_norm": 0.004769014660269022, "learning_rate": 1.7466666666666667e-05, "loss": 0.013, "step": 1070 }, { "epoch": 0.34, "grad_norm": 1.7518908977508545, "learning_rate": 1.7422222222222222e-05, "loss": 0.082, "step": 1080 }, { "epoch": 0.34, "grad_norm": 0.07153653353452682, "learning_rate": 1.737777777777778e-05, "loss": 0.0388, "step": 1090 }, { "epoch": 0.34, "grad_norm": 0.015164585784077644, "learning_rate": 1.7333333333333336e-05, "loss": 0.0613, "step": 1100 }, { "epoch": 0.35, "grad_norm": 3.7508952617645264, "learning_rate": 1.728888888888889e-05, "loss": 0.0989, "step": 1110 }, { "epoch": 0.35, "grad_norm": 0.025457441806793213, "learning_rate": 1.7244444444444446e-05, "loss": 0.0268, "step": 1120 }, { "epoch": 0.35, "grad_norm": 4.794126033782959, "learning_rate": 1.72e-05, "loss": 0.0741, "step": 1130 }, { "epoch": 0.36, "grad_norm": 4.877151012420654, "learning_rate": 1.7155555555555557e-05, "loss": 0.0653, "step": 1140 }, { "epoch": 0.36, "grad_norm": 0.009955652058124542, "learning_rate": 1.7111111111111112e-05, "loss": 0.1041, "step": 1150 }, { "epoch": 0.36, "grad_norm": 0.006371030583977699, "learning_rate": 1.706666666666667e-05, "loss": 0.0365, "step": 1160 }, { "epoch": 0.37, "grad_norm": 0.005713317077606916, "learning_rate": 1.7022222222222226e-05, "loss": 0.0402, "step": 1170 }, { "epoch": 0.37, "grad_norm": 0.0035403750371187925, "learning_rate": 1.697777777777778e-05, "loss": 0.0208, "step": 1180 }, { "epoch": 0.37, "grad_norm": 1.2340508699417114, "learning_rate": 1.6933333333333336e-05, "loss": 0.0507, "step": 1190 }, { "epoch": 0.38, "grad_norm": 0.01031398307532072, "learning_rate": 1.688888888888889e-05, "loss": 0.1464, "step": 1200 }, { "epoch": 0.38, "grad_norm": 1.4290913343429565, "learning_rate": 1.6844444444444447e-05, "loss": 0.01, "step": 1210 }, { "epoch": 0.38, "grad_norm": 1.9291528463363647, "learning_rate": 1.6800000000000002e-05, "loss": 0.0077, "step": 1220 }, { "epoch": 0.39, "grad_norm": 0.049461908638477325, "learning_rate": 1.6755555555555557e-05, "loss": 0.034, "step": 1230 }, { "epoch": 0.39, "grad_norm": 0.004318055231124163, "learning_rate": 1.6711111111111112e-05, "loss": 0.0612, "step": 1240 }, { "epoch": 0.39, "grad_norm": 0.003021540120244026, "learning_rate": 1.6666666666666667e-05, "loss": 0.0135, "step": 1250 }, { "epoch": 0.39, "grad_norm": 0.04815623164176941, "learning_rate": 1.6622222222222223e-05, "loss": 0.0884, "step": 1260 }, { "epoch": 0.4, "grad_norm": 2.4603257179260254, "learning_rate": 1.6577777777777778e-05, "loss": 0.0149, "step": 1270 }, { "epoch": 0.4, "grad_norm": 0.002652758965268731, "learning_rate": 1.6533333333333333e-05, "loss": 0.0239, "step": 1280 }, { "epoch": 0.4, "grad_norm": 0.053305696696043015, "learning_rate": 1.648888888888889e-05, "loss": 0.0615, "step": 1290 }, { "epoch": 0.41, "grad_norm": 0.0942620038986206, "learning_rate": 1.6444444444444444e-05, "loss": 0.0789, "step": 1300 }, { "epoch": 0.41, "grad_norm": 2.3062126636505127, "learning_rate": 1.64e-05, "loss": 0.0526, "step": 1310 }, { "epoch": 0.41, "grad_norm": 0.004340393468737602, "learning_rate": 1.6355555555555557e-05, "loss": 0.0638, "step": 1320 }, { "epoch": 0.42, "grad_norm": 2.277503252029419, "learning_rate": 1.6311111111111113e-05, "loss": 0.0512, "step": 1330 }, { "epoch": 0.42, "grad_norm": 5.073038578033447, "learning_rate": 1.6266666666666668e-05, "loss": 0.1041, "step": 1340 }, { "epoch": 0.42, "grad_norm": 0.0030644198413938284, "learning_rate": 1.6222222222222223e-05, "loss": 0.0016, "step": 1350 }, { "epoch": 0.43, "grad_norm": 0.4875153601169586, "learning_rate": 1.617777777777778e-05, "loss": 0.0946, "step": 1360 }, { "epoch": 0.43, "grad_norm": 0.003334318520501256, "learning_rate": 1.6133333333333334e-05, "loss": 0.0548, "step": 1370 }, { "epoch": 0.43, "grad_norm": 2.697582960128784, "learning_rate": 1.608888888888889e-05, "loss": 0.0628, "step": 1380 }, { "epoch": 0.44, "grad_norm": 0.021118061617016792, "learning_rate": 1.6044444444444444e-05, "loss": 0.0537, "step": 1390 }, { "epoch": 0.44, "grad_norm": 0.007942981086671352, "learning_rate": 1.6000000000000003e-05, "loss": 0.057, "step": 1400 }, { "epoch": 0.44, "grad_norm": 0.003399541135877371, "learning_rate": 1.5955555555555558e-05, "loss": 0.0089, "step": 1410 }, { "epoch": 0.44, "grad_norm": 0.4597236216068268, "learning_rate": 1.5911111111111113e-05, "loss": 0.079, "step": 1420 }, { "epoch": 0.45, "grad_norm": 1.0684938430786133, "learning_rate": 1.586666666666667e-05, "loss": 0.0149, "step": 1430 }, { "epoch": 0.45, "grad_norm": 3.3166937828063965, "learning_rate": 1.5822222222222224e-05, "loss": 0.1189, "step": 1440 }, { "epoch": 0.45, "grad_norm": 5.2161736488342285, "learning_rate": 1.577777777777778e-05, "loss": 0.1408, "step": 1450 }, { "epoch": 0.46, "grad_norm": 0.3570309281349182, "learning_rate": 1.5733333333333334e-05, "loss": 0.0247, "step": 1460 }, { "epoch": 0.46, "grad_norm": 0.005081634968519211, "learning_rate": 1.5688888888888893e-05, "loss": 0.0231, "step": 1470 }, { "epoch": 0.46, "grad_norm": 4.267640590667725, "learning_rate": 1.5644444444444448e-05, "loss": 0.0928, "step": 1480 }, { "epoch": 0.47, "grad_norm": 0.014545031823217869, "learning_rate": 1.5600000000000003e-05, "loss": 0.0561, "step": 1490 }, { "epoch": 0.47, "grad_norm": 2.007899522781372, "learning_rate": 1.555555555555556e-05, "loss": 0.058, "step": 1500 }, { "epoch": 0.47, "eval_loss": 0.021991439163684845, "eval_runtime": 61.9363, "eval_samples_per_second": 16.146, "eval_steps_per_second": 16.146, "step": 1500 }, { "epoch": 0.47, "grad_norm": 0.18892046809196472, "learning_rate": 1.5511111111111114e-05, "loss": 0.0541, "step": 1510 }, { "epoch": 0.48, "grad_norm": 0.9835280179977417, "learning_rate": 1.546666666666667e-05, "loss": 0.0349, "step": 1520 }, { "epoch": 0.48, "grad_norm": 0.004944021347910166, "learning_rate": 1.5422222222222224e-05, "loss": 0.0688, "step": 1530 }, { "epoch": 0.48, "grad_norm": 0.0025414193514734507, "learning_rate": 1.537777777777778e-05, "loss": 0.0374, "step": 1540 }, { "epoch": 0.49, "grad_norm": 0.047879841178655624, "learning_rate": 1.5333333333333334e-05, "loss": 0.0867, "step": 1550 }, { "epoch": 0.49, "grad_norm": 0.12313953042030334, "learning_rate": 1.528888888888889e-05, "loss": 0.0142, "step": 1560 }, { "epoch": 0.49, "grad_norm": 0.0055120717734098434, "learning_rate": 1.5244444444444447e-05, "loss": 0.0332, "step": 1570 }, { "epoch": 0.5, "grad_norm": 0.0030517149716615677, "learning_rate": 1.5200000000000002e-05, "loss": 0.0739, "step": 1580 }, { "epoch": 0.5, "grad_norm": 0.04954369366168976, "learning_rate": 1.5155555555555557e-05, "loss": 0.0609, "step": 1590 }, { "epoch": 0.5, "grad_norm": 2.16231369972229, "learning_rate": 1.5111111111111112e-05, "loss": 0.084, "step": 1600 }, { "epoch": 0.5, "grad_norm": 0.03328130766749382, "learning_rate": 1.5066666666666668e-05, "loss": 0.0561, "step": 1610 }, { "epoch": 0.51, "grad_norm": 0.3883110582828522, "learning_rate": 1.5022222222222223e-05, "loss": 0.065, "step": 1620 }, { "epoch": 0.51, "grad_norm": 0.007501136511564255, "learning_rate": 1.497777777777778e-05, "loss": 0.0645, "step": 1630 }, { "epoch": 0.51, "grad_norm": 0.0025652372278273106, "learning_rate": 1.4933333333333335e-05, "loss": 0.1242, "step": 1640 }, { "epoch": 0.52, "grad_norm": 2.322174549102783, "learning_rate": 1.488888888888889e-05, "loss": 0.0639, "step": 1650 }, { "epoch": 0.52, "grad_norm": 3.1910171508789062, "learning_rate": 1.4844444444444445e-05, "loss": 0.1048, "step": 1660 }, { "epoch": 0.52, "grad_norm": 0.00425474438816309, "learning_rate": 1.48e-05, "loss": 0.0244, "step": 1670 }, { "epoch": 0.53, "grad_norm": 4.180226802825928, "learning_rate": 1.4755555555555556e-05, "loss": 0.0874, "step": 1680 }, { "epoch": 0.53, "grad_norm": 1.1955620050430298, "learning_rate": 1.4711111111111111e-05, "loss": 0.0371, "step": 1690 }, { "epoch": 0.53, "grad_norm": 0.010523403063416481, "learning_rate": 1.4666666666666666e-05, "loss": 0.0441, "step": 1700 }, { "epoch": 0.54, "grad_norm": 0.3403013348579407, "learning_rate": 1.4622222222222225e-05, "loss": 0.0758, "step": 1710 }, { "epoch": 0.54, "grad_norm": 0.046740252524614334, "learning_rate": 1.457777777777778e-05, "loss": 0.1069, "step": 1720 }, { "epoch": 0.54, "grad_norm": 0.1384706348180771, "learning_rate": 1.4533333333333335e-05, "loss": 0.0656, "step": 1730 }, { "epoch": 0.55, "grad_norm": 1.4094263315200806, "learning_rate": 1.448888888888889e-05, "loss": 0.0255, "step": 1740 }, { "epoch": 0.55, "grad_norm": 0.055873990058898926, "learning_rate": 1.4444444444444446e-05, "loss": 0.0181, "step": 1750 }, { "epoch": 0.55, "grad_norm": 0.002195443492382765, "learning_rate": 1.4400000000000001e-05, "loss": 0.0004, "step": 1760 }, { "epoch": 0.55, "grad_norm": 0.004009276628494263, "learning_rate": 1.4355555555555556e-05, "loss": 0.0724, "step": 1770 }, { "epoch": 0.56, "grad_norm": 2.4936776161193848, "learning_rate": 1.4311111111111111e-05, "loss": 0.0822, "step": 1780 }, { "epoch": 0.56, "grad_norm": 1.6697956323623657, "learning_rate": 1.4266666666666668e-05, "loss": 0.0041, "step": 1790 }, { "epoch": 0.56, "grad_norm": 2.3623549938201904, "learning_rate": 1.4222222222222224e-05, "loss": 0.0805, "step": 1800 }, { "epoch": 0.57, "grad_norm": 3.6882166862487793, "learning_rate": 1.4177777777777779e-05, "loss": 0.0353, "step": 1810 }, { "epoch": 0.57, "grad_norm": 2.8029565811157227, "learning_rate": 1.4133333333333334e-05, "loss": 0.0666, "step": 1820 }, { "epoch": 0.57, "grad_norm": 0.0027399081736803055, "learning_rate": 1.408888888888889e-05, "loss": 0.0129, "step": 1830 }, { "epoch": 0.58, "grad_norm": 0.002227638615295291, "learning_rate": 1.4044444444444445e-05, "loss": 0.0644, "step": 1840 }, { "epoch": 0.58, "grad_norm": 0.0038109265733510256, "learning_rate": 1.4e-05, "loss": 0.0554, "step": 1850 }, { "epoch": 0.58, "grad_norm": 0.23026524484157562, "learning_rate": 1.3955555555555558e-05, "loss": 0.0513, "step": 1860 }, { "epoch": 0.59, "grad_norm": 0.003622630378231406, "learning_rate": 1.3911111111111114e-05, "loss": 0.0595, "step": 1870 }, { "epoch": 0.59, "grad_norm": 0.004487840924412012, "learning_rate": 1.3866666666666669e-05, "loss": 0.0507, "step": 1880 }, { "epoch": 0.59, "grad_norm": 0.0032237153500318527, "learning_rate": 1.3822222222222224e-05, "loss": 0.0246, "step": 1890 }, { "epoch": 0.6, "grad_norm": 2.904846429824829, "learning_rate": 1.377777777777778e-05, "loss": 0.041, "step": 1900 }, { "epoch": 0.6, "grad_norm": 2.9861977100372314, "learning_rate": 1.3733333333333335e-05, "loss": 0.0529, "step": 1910 }, { "epoch": 0.6, "grad_norm": 0.002036773832514882, "learning_rate": 1.368888888888889e-05, "loss": 0.0515, "step": 1920 }, { "epoch": 0.6, "grad_norm": 2.882114887237549, "learning_rate": 1.3644444444444445e-05, "loss": 0.0437, "step": 1930 }, { "epoch": 0.61, "grad_norm": 3.0705764293670654, "learning_rate": 1.3600000000000002e-05, "loss": 0.0203, "step": 1940 }, { "epoch": 0.61, "grad_norm": 0.003252115799114108, "learning_rate": 1.3555555555555557e-05, "loss": 0.0288, "step": 1950 }, { "epoch": 0.61, "grad_norm": 0.002082700841128826, "learning_rate": 1.3511111111111112e-05, "loss": 0.0737, "step": 1960 }, { "epoch": 0.62, "grad_norm": 0.004269629716873169, "learning_rate": 1.3466666666666668e-05, "loss": 0.0283, "step": 1970 }, { "epoch": 0.62, "grad_norm": 0.004804402124136686, "learning_rate": 1.3422222222222223e-05, "loss": 0.0769, "step": 1980 }, { "epoch": 0.62, "grad_norm": 0.006747941020876169, "learning_rate": 1.3377777777777778e-05, "loss": 0.0169, "step": 1990 }, { "epoch": 0.63, "grad_norm": 0.00186560966540128, "learning_rate": 1.3333333333333333e-05, "loss": 0.0346, "step": 2000 }, { "epoch": 0.63, "eval_loss": 0.018277771770954132, "eval_runtime": 61.9406, "eval_samples_per_second": 16.144, "eval_steps_per_second": 16.144, "step": 2000 }, { "epoch": 0.63, "grad_norm": 0.0024101065937429667, "learning_rate": 1.3288888888888889e-05, "loss": 0.047, "step": 2010 }, { "epoch": 0.63, "grad_norm": 0.013266036286950111, "learning_rate": 1.3244444444444447e-05, "loss": 0.0304, "step": 2020 }, { "epoch": 0.64, "grad_norm": 0.500153660774231, "learning_rate": 1.3200000000000002e-05, "loss": 0.0443, "step": 2030 }, { "epoch": 0.64, "grad_norm": 0.0013359179720282555, "learning_rate": 1.3155555555555558e-05, "loss": 0.0002, "step": 2040 }, { "epoch": 0.64, "grad_norm": 1.1259698867797852, "learning_rate": 1.3111111111111113e-05, "loss": 0.0382, "step": 2050 }, { "epoch": 0.65, "grad_norm": 0.009357116185128689, "learning_rate": 1.3066666666666668e-05, "loss": 0.0624, "step": 2060 }, { "epoch": 0.65, "grad_norm": 0.1253252774477005, "learning_rate": 1.3022222222222223e-05, "loss": 0.0734, "step": 2070 }, { "epoch": 0.65, "grad_norm": 0.0024754456244409084, "learning_rate": 1.2977777777777779e-05, "loss": 0.0683, "step": 2080 }, { "epoch": 0.65, "grad_norm": 1.6212913990020752, "learning_rate": 1.2933333333333334e-05, "loss": 0.0467, "step": 2090 }, { "epoch": 0.66, "grad_norm": 1.4906156063079834, "learning_rate": 1.288888888888889e-05, "loss": 0.1055, "step": 2100 }, { "epoch": 0.66, "grad_norm": 0.005489406641572714, "learning_rate": 1.2844444444444446e-05, "loss": 0.0425, "step": 2110 }, { "epoch": 0.66, "grad_norm": 0.018557880073785782, "learning_rate": 1.2800000000000001e-05, "loss": 0.0497, "step": 2120 }, { "epoch": 0.67, "grad_norm": 0.00175235525239259, "learning_rate": 1.2755555555555556e-05, "loss": 0.0517, "step": 2130 }, { "epoch": 0.67, "grad_norm": 0.002169775078073144, "learning_rate": 1.2711111111111112e-05, "loss": 0.038, "step": 2140 }, { "epoch": 0.67, "grad_norm": 0.004229371901601553, "learning_rate": 1.2666666666666667e-05, "loss": 0.0897, "step": 2150 }, { "epoch": 0.68, "grad_norm": 0.0044867489486932755, "learning_rate": 1.2622222222222222e-05, "loss": 0.092, "step": 2160 }, { "epoch": 0.68, "grad_norm": 0.00232448847964406, "learning_rate": 1.257777777777778e-05, "loss": 0.0474, "step": 2170 }, { "epoch": 0.68, "grad_norm": 0.10797467082738876, "learning_rate": 1.2533333333333336e-05, "loss": 0.0065, "step": 2180 }, { "epoch": 0.69, "grad_norm": 0.001698866835795343, "learning_rate": 1.2488888888888891e-05, "loss": 0.1059, "step": 2190 }, { "epoch": 0.69, "grad_norm": 0.0014356509782373905, "learning_rate": 1.2444444444444446e-05, "loss": 0.0352, "step": 2200 }, { "epoch": 0.69, "grad_norm": 0.001666803378611803, "learning_rate": 1.2400000000000002e-05, "loss": 0.0723, "step": 2210 }, { "epoch": 0.7, "grad_norm": 1.3424423933029175, "learning_rate": 1.2355555555555557e-05, "loss": 0.029, "step": 2220 }, { "epoch": 0.7, "grad_norm": 0.0017589009366929531, "learning_rate": 1.2311111111111112e-05, "loss": 0.0261, "step": 2230 }, { "epoch": 0.7, "grad_norm": 2.2947793006896973, "learning_rate": 1.2266666666666667e-05, "loss": 0.1309, "step": 2240 }, { "epoch": 0.7, "grad_norm": 2.9271962642669678, "learning_rate": 1.2222222222222224e-05, "loss": 0.062, "step": 2250 }, { "epoch": 0.71, "grad_norm": 0.006457278039306402, "learning_rate": 1.217777777777778e-05, "loss": 0.0099, "step": 2260 }, { "epoch": 0.71, "grad_norm": 1.2980073690414429, "learning_rate": 1.2133333333333335e-05, "loss": 0.0846, "step": 2270 }, { "epoch": 0.71, "grad_norm": 0.0023306766524910927, "learning_rate": 1.208888888888889e-05, "loss": 0.028, "step": 2280 }, { "epoch": 0.72, "grad_norm": 0.0022570204455405474, "learning_rate": 1.2044444444444445e-05, "loss": 0.0283, "step": 2290 }, { "epoch": 0.72, "grad_norm": 0.0026164520531892776, "learning_rate": 1.2e-05, "loss": 0.0502, "step": 2300 }, { "epoch": 0.72, "grad_norm": 0.0019713249057531357, "learning_rate": 1.1955555555555556e-05, "loss": 0.0354, "step": 2310 }, { "epoch": 0.73, "grad_norm": 0.00487458985298872, "learning_rate": 1.191111111111111e-05, "loss": 0.0579, "step": 2320 }, { "epoch": 0.73, "grad_norm": 0.002555105835199356, "learning_rate": 1.186666666666667e-05, "loss": 0.0566, "step": 2330 }, { "epoch": 0.73, "grad_norm": 0.28949007391929626, "learning_rate": 1.1822222222222225e-05, "loss": 0.0149, "step": 2340 }, { "epoch": 0.74, "grad_norm": 0.5910075902938843, "learning_rate": 1.177777777777778e-05, "loss": 0.075, "step": 2350 }, { "epoch": 0.74, "grad_norm": 0.001737405196763575, "learning_rate": 1.1733333333333335e-05, "loss": 0.0469, "step": 2360 }, { "epoch": 0.74, "grad_norm": 5.248105049133301, "learning_rate": 1.168888888888889e-05, "loss": 0.0715, "step": 2370 }, { "epoch": 0.75, "grad_norm": 0.0018054692773148417, "learning_rate": 1.1644444444444446e-05, "loss": 0.0354, "step": 2380 }, { "epoch": 0.75, "grad_norm": 0.001644105650484562, "learning_rate": 1.16e-05, "loss": 0.0321, "step": 2390 }, { "epoch": 0.75, "grad_norm": 0.001677204272709787, "learning_rate": 1.1555555555555556e-05, "loss": 0.0183, "step": 2400 }, { "epoch": 0.76, "grad_norm": 3.1084885597229004, "learning_rate": 1.1511111111111113e-05, "loss": 0.029, "step": 2410 }, { "epoch": 0.76, "grad_norm": 0.007074132561683655, "learning_rate": 1.1466666666666668e-05, "loss": 0.0525, "step": 2420 }, { "epoch": 0.76, "grad_norm": 0.0032486789859831333, "learning_rate": 1.1422222222222223e-05, "loss": 0.0338, "step": 2430 }, { "epoch": 0.76, "grad_norm": 2.797699213027954, "learning_rate": 1.1377777777777779e-05, "loss": 0.1017, "step": 2440 }, { "epoch": 0.77, "grad_norm": 1.977303385734558, "learning_rate": 1.1333333333333334e-05, "loss": 0.061, "step": 2450 }, { "epoch": 0.77, "grad_norm": 4.702297210693359, "learning_rate": 1.1288888888888889e-05, "loss": 0.0382, "step": 2460 }, { "epoch": 0.77, "grad_norm": 0.05130983144044876, "learning_rate": 1.1244444444444444e-05, "loss": 0.0273, "step": 2470 }, { "epoch": 0.78, "grad_norm": 0.001360241905786097, "learning_rate": 1.1200000000000001e-05, "loss": 0.0615, "step": 2480 }, { "epoch": 0.78, "grad_norm": 0.65842205286026, "learning_rate": 1.1155555555555556e-05, "loss": 0.0349, "step": 2490 }, { "epoch": 0.78, "grad_norm": 0.0062817600555717945, "learning_rate": 1.1111111111111113e-05, "loss": 0.055, "step": 2500 }, { "epoch": 0.78, "eval_loss": 0.020550861954689026, "eval_runtime": 61.9123, "eval_samples_per_second": 16.152, "eval_steps_per_second": 16.152, "step": 2500 }, { "epoch": 0.79, "grad_norm": 0.002629642840474844, "learning_rate": 1.1066666666666669e-05, "loss": 0.0775, "step": 2510 }, { "epoch": 0.79, "grad_norm": 0.003445986658334732, "learning_rate": 1.1022222222222224e-05, "loss": 0.0285, "step": 2520 }, { "epoch": 0.79, "grad_norm": 2.4178996086120605, "learning_rate": 1.0977777777777779e-05, "loss": 0.1755, "step": 2530 }, { "epoch": 0.8, "grad_norm": 0.0058800880797207355, "learning_rate": 1.0933333333333334e-05, "loss": 0.0055, "step": 2540 }, { "epoch": 0.8, "grad_norm": 0.013078362680971622, "learning_rate": 1.088888888888889e-05, "loss": 0.0088, "step": 2550 }, { "epoch": 0.8, "grad_norm": 0.0052205640822649, "learning_rate": 1.0844444444444446e-05, "loss": 0.023, "step": 2560 }, { "epoch": 0.81, "grad_norm": 1.2594823837280273, "learning_rate": 1.0800000000000002e-05, "loss": 0.0222, "step": 2570 }, { "epoch": 0.81, "grad_norm": 3.3159544467926025, "learning_rate": 1.0755555555555557e-05, "loss": 0.0426, "step": 2580 }, { "epoch": 0.81, "grad_norm": 2.103391647338867, "learning_rate": 1.0711111111111112e-05, "loss": 0.0457, "step": 2590 }, { "epoch": 0.81, "grad_norm": 0.000918123172596097, "learning_rate": 1.0666666666666667e-05, "loss": 0.0391, "step": 2600 }, { "epoch": 0.82, "grad_norm": 0.0012873125961050391, "learning_rate": 1.0622222222222223e-05, "loss": 0.0217, "step": 2610 }, { "epoch": 0.82, "grad_norm": 0.002058672485873103, "learning_rate": 1.0577777777777778e-05, "loss": 0.0463, "step": 2620 }, { "epoch": 0.82, "grad_norm": 0.027771245688199997, "learning_rate": 1.0533333333333333e-05, "loss": 0.0805, "step": 2630 }, { "epoch": 0.83, "grad_norm": 0.0015083320904523134, "learning_rate": 1.048888888888889e-05, "loss": 0.05, "step": 2640 }, { "epoch": 0.83, "grad_norm": 2.7888853549957275, "learning_rate": 1.0444444444444445e-05, "loss": 0.0752, "step": 2650 }, { "epoch": 0.83, "grad_norm": 3.569105863571167, "learning_rate": 1.04e-05, "loss": 0.0585, "step": 2660 }, { "epoch": 0.84, "grad_norm": 0.034804560244083405, "learning_rate": 1.0355555555555557e-05, "loss": 0.0113, "step": 2670 }, { "epoch": 0.84, "grad_norm": 0.0017612408846616745, "learning_rate": 1.0311111111111113e-05, "loss": 0.0535, "step": 2680 }, { "epoch": 0.84, "grad_norm": 0.0011774456361308694, "learning_rate": 1.0266666666666668e-05, "loss": 0.0622, "step": 2690 }, { "epoch": 0.85, "grad_norm": 2.2724449634552, "learning_rate": 1.0222222222222223e-05, "loss": 0.0059, "step": 2700 }, { "epoch": 0.85, "grad_norm": 0.8430375456809998, "learning_rate": 1.0177777777777778e-05, "loss": 0.043, "step": 2710 }, { "epoch": 0.85, "grad_norm": 2.0912680625915527, "learning_rate": 1.0133333333333335e-05, "loss": 0.0356, "step": 2720 }, { "epoch": 0.86, "grad_norm": 0.043006353080272675, "learning_rate": 1.008888888888889e-05, "loss": 0.0247, "step": 2730 }, { "epoch": 0.86, "grad_norm": 0.0013607463333755732, "learning_rate": 1.0044444444444446e-05, "loss": 0.0115, "step": 2740 }, { "epoch": 0.86, "grad_norm": 0.003235200187191367, "learning_rate": 1e-05, "loss": 0.0447, "step": 2750 }, { "epoch": 0.86, "grad_norm": 0.006067329086363316, "learning_rate": 9.955555555555556e-06, "loss": 0.049, "step": 2760 }, { "epoch": 0.87, "grad_norm": 0.001446689828298986, "learning_rate": 9.911111111111113e-06, "loss": 0.0502, "step": 2770 }, { "epoch": 0.87, "grad_norm": 0.0018873319495469332, "learning_rate": 9.866666666666668e-06, "loss": 0.0854, "step": 2780 }, { "epoch": 0.87, "grad_norm": 0.002605983056128025, "learning_rate": 9.822222222222223e-06, "loss": 0.0284, "step": 2790 }, { "epoch": 0.88, "grad_norm": 0.002705740975216031, "learning_rate": 9.777777777777779e-06, "loss": 0.0639, "step": 2800 }, { "epoch": 0.88, "grad_norm": 0.004726971033960581, "learning_rate": 9.733333333333334e-06, "loss": 0.1064, "step": 2810 }, { "epoch": 0.88, "grad_norm": 0.01625528745353222, "learning_rate": 9.688888888888889e-06, "loss": 0.0679, "step": 2820 }, { "epoch": 0.89, "grad_norm": 1.4221800565719604, "learning_rate": 9.644444444444444e-06, "loss": 0.0504, "step": 2830 }, { "epoch": 0.89, "grad_norm": 1.1785073280334473, "learning_rate": 9.600000000000001e-06, "loss": 0.0574, "step": 2840 }, { "epoch": 0.89, "grad_norm": 0.0024502715095877647, "learning_rate": 9.555555555555556e-06, "loss": 0.0211, "step": 2850 }, { "epoch": 0.9, "grad_norm": 0.0017725643701851368, "learning_rate": 9.511111111111112e-06, "loss": 0.057, "step": 2860 }, { "epoch": 0.9, "grad_norm": 0.0027383696287870407, "learning_rate": 9.466666666666667e-06, "loss": 0.1225, "step": 2870 }, { "epoch": 0.9, "grad_norm": 0.00296800397336483, "learning_rate": 9.422222222222222e-06, "loss": 0.0167, "step": 2880 }, { "epoch": 0.91, "grad_norm": 0.003540828125551343, "learning_rate": 9.377777777777779e-06, "loss": 0.0484, "step": 2890 }, { "epoch": 0.91, "grad_norm": 0.03577937185764313, "learning_rate": 9.333333333333334e-06, "loss": 0.0334, "step": 2900 }, { "epoch": 0.91, "grad_norm": 0.0013176521752029657, "learning_rate": 9.28888888888889e-06, "loss": 0.0706, "step": 2910 }, { "epoch": 0.91, "grad_norm": 0.0014053646009415388, "learning_rate": 9.244444444444445e-06, "loss": 0.0206, "step": 2920 }, { "epoch": 0.92, "grad_norm": 0.002312118886038661, "learning_rate": 9.200000000000002e-06, "loss": 0.0398, "step": 2930 }, { "epoch": 0.92, "grad_norm": 0.002495428314432502, "learning_rate": 9.155555555555557e-06, "loss": 0.0151, "step": 2940 }, { "epoch": 0.92, "grad_norm": 0.0020435641054064035, "learning_rate": 9.111111111111112e-06, "loss": 0.034, "step": 2950 }, { "epoch": 0.93, "grad_norm": 0.0015432636719197035, "learning_rate": 9.066666666666667e-06, "loss": 0.0485, "step": 2960 }, { "epoch": 0.93, "grad_norm": 1.6057056188583374, "learning_rate": 9.022222222222223e-06, "loss": 0.0566, "step": 2970 }, { "epoch": 0.93, "grad_norm": 0.0020012864843010902, "learning_rate": 8.977777777777778e-06, "loss": 0.0012, "step": 2980 }, { "epoch": 0.94, "grad_norm": 0.0008306769304908812, "learning_rate": 8.933333333333333e-06, "loss": 0.037, "step": 2990 }, { "epoch": 0.94, "grad_norm": 0.0008226165664382279, "learning_rate": 8.888888888888888e-06, "loss": 0.0365, "step": 3000 }, { "epoch": 0.94, "eval_loss": 0.021811991930007935, "eval_runtime": 61.9881, "eval_samples_per_second": 16.132, "eval_steps_per_second": 16.132, "step": 3000 } ], "logging_steps": 10, "max_steps": 5000, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 1000, "total_flos": 4.8306377981952e+16, "train_batch_size": 1, "trial_name": null, "trial_params": null }