{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.3, "eval_steps": 500, "global_step": 1875, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0008, "grad_norm": 1.0, "learning_rate": 0.0001, "loss": 2.6853, "step": 5 }, { "epoch": 0.0016, "grad_norm": 0.65625, "learning_rate": 0.0001, "loss": 1.2371, "step": 10 }, { "epoch": 0.0024, "grad_norm": 0.3828125, "learning_rate": 0.0001, "loss": 0.8883, "step": 15 }, { "epoch": 0.0032, "grad_norm": 0.34765625, "learning_rate": 0.0001, "loss": 0.6803, "step": 20 }, { "epoch": 0.004, "grad_norm": 0.388671875, "learning_rate": 0.0001, "loss": 0.5467, "step": 25 }, { "epoch": 0.0048, "grad_norm": 0.32421875, "learning_rate": 0.0001, "loss": 0.4222, "step": 30 }, { "epoch": 0.0056, "grad_norm": 0.31640625, "learning_rate": 0.0001, "loss": 0.4292, "step": 35 }, { "epoch": 0.0064, "grad_norm": 0.34765625, "learning_rate": 0.0001, "loss": 0.356, "step": 40 }, { "epoch": 0.0072, "grad_norm": 0.31640625, "learning_rate": 0.0001, "loss": 0.2943, "step": 45 }, { "epoch": 0.008, "grad_norm": 0.1943359375, "learning_rate": 0.0001, "loss": 0.2583, "step": 50 }, { "epoch": 0.0088, "grad_norm": 0.19921875, "learning_rate": 0.0001, "loss": 0.2666, "step": 55 }, { "epoch": 0.0096, "grad_norm": 0.244140625, "learning_rate": 0.0001, "loss": 0.2787, "step": 60 }, { "epoch": 0.0104, "grad_norm": 0.275390625, "learning_rate": 0.0001, "loss": 0.3379, "step": 65 }, { "epoch": 0.0112, "grad_norm": 0.33203125, "learning_rate": 0.0001, "loss": 0.3266, "step": 70 }, { "epoch": 0.012, "grad_norm": 0.14453125, "learning_rate": 0.0001, "loss": 0.2819, "step": 75 }, { "epoch": 0.0128, "grad_norm": 0.1787109375, "learning_rate": 0.0001, "loss": 0.2406, "step": 80 }, { "epoch": 0.0136, "grad_norm": 0.203125, "learning_rate": 0.0001, "loss": 0.3264, "step": 85 }, { "epoch": 0.0144, "grad_norm": 0.1953125, "learning_rate": 0.0001, "loss": 0.3031, "step": 90 }, { "epoch": 0.0152, "grad_norm": 0.265625, "learning_rate": 0.0001, "loss": 0.3208, "step": 95 }, { "epoch": 0.016, "grad_norm": 0.2333984375, "learning_rate": 0.0001, "loss": 0.2912, "step": 100 }, { "epoch": 0.0168, "grad_norm": 0.2314453125, "learning_rate": 0.0001, "loss": 0.193, "step": 105 }, { "epoch": 0.0176, "grad_norm": 0.208984375, "learning_rate": 0.0001, "loss": 0.2738, "step": 110 }, { "epoch": 0.0184, "grad_norm": 0.1806640625, "learning_rate": 0.0001, "loss": 0.2672, "step": 115 }, { "epoch": 0.0192, "grad_norm": 0.2080078125, "learning_rate": 0.0001, "loss": 0.316, "step": 120 }, { "epoch": 0.02, "grad_norm": 0.1884765625, "learning_rate": 0.0001, "loss": 0.2612, "step": 125 }, { "epoch": 0.0208, "grad_norm": 0.220703125, "learning_rate": 0.0001, "loss": 0.2953, "step": 130 }, { "epoch": 0.0216, "grad_norm": 0.1982421875, "learning_rate": 0.0001, "loss": 0.2785, "step": 135 }, { "epoch": 0.0224, "grad_norm": 0.1611328125, "learning_rate": 0.0001, "loss": 0.319, "step": 140 }, { "epoch": 0.0232, "grad_norm": 0.12890625, "learning_rate": 0.0001, "loss": 0.2766, "step": 145 }, { "epoch": 0.024, "grad_norm": 0.216796875, "learning_rate": 0.0001, "loss": 0.1815, "step": 150 }, { "epoch": 0.0248, "grad_norm": 0.11572265625, "learning_rate": 0.0001, "loss": 0.1364, "step": 155 }, { "epoch": 0.0256, "grad_norm": 0.16015625, "learning_rate": 0.0001, "loss": 0.2344, "step": 160 }, { "epoch": 0.0264, "grad_norm": 0.1748046875, "learning_rate": 0.0001, "loss": 0.2177, "step": 165 }, { "epoch": 0.0272, "grad_norm": 0.3671875, "learning_rate": 0.0001, "loss": 0.2535, "step": 170 }, { "epoch": 0.028, "grad_norm": 0.1630859375, "learning_rate": 0.0001, "loss": 0.2629, "step": 175 }, { "epoch": 0.0288, "grad_norm": 0.2255859375, "learning_rate": 0.0001, "loss": 0.2015, "step": 180 }, { "epoch": 0.0296, "grad_norm": 0.1630859375, "learning_rate": 0.0001, "loss": 0.2467, "step": 185 }, { "epoch": 0.0304, "grad_norm": 0.2080078125, "learning_rate": 0.0001, "loss": 0.2361, "step": 190 }, { "epoch": 0.0312, "grad_norm": 0.1484375, "learning_rate": 0.0001, "loss": 0.2341, "step": 195 }, { "epoch": 0.032, "grad_norm": 0.154296875, "learning_rate": 0.0001, "loss": 0.199, "step": 200 }, { "epoch": 0.0328, "grad_norm": 0.162109375, "learning_rate": 0.0001, "loss": 0.1996, "step": 205 }, { "epoch": 0.0336, "grad_norm": 0.1982421875, "learning_rate": 0.0001, "loss": 0.2472, "step": 210 }, { "epoch": 0.0344, "grad_norm": 0.1171875, "learning_rate": 0.0001, "loss": 0.2503, "step": 215 }, { "epoch": 0.0352, "grad_norm": 0.2158203125, "learning_rate": 0.0001, "loss": 0.2538, "step": 220 }, { "epoch": 0.036, "grad_norm": 0.1513671875, "learning_rate": 0.0001, "loss": 0.244, "step": 225 }, { "epoch": 0.0368, "grad_norm": 1.1875, "learning_rate": 0.0001, "loss": 0.2082, "step": 230 }, { "epoch": 0.0376, "grad_norm": 0.1435546875, "learning_rate": 0.0001, "loss": 0.2536, "step": 235 }, { "epoch": 0.0384, "grad_norm": 0.1337890625, "learning_rate": 0.0001, "loss": 0.2415, "step": 240 }, { "epoch": 0.0392, "grad_norm": 0.1591796875, "learning_rate": 0.0001, "loss": 0.3025, "step": 245 }, { "epoch": 0.04, "grad_norm": 0.11962890625, "learning_rate": 0.0001, "loss": 0.191, "step": 250 }, { "epoch": 0.0408, "grad_norm": 0.2001953125, "learning_rate": 0.0001, "loss": 0.17, "step": 255 }, { "epoch": 0.0416, "grad_norm": 0.1455078125, "learning_rate": 0.0001, "loss": 0.227, "step": 260 }, { "epoch": 0.0424, "grad_norm": 0.173828125, "learning_rate": 0.0001, "loss": 0.2285, "step": 265 }, { "epoch": 0.0432, "grad_norm": 0.1708984375, "learning_rate": 0.0001, "loss": 0.2092, "step": 270 }, { "epoch": 0.044, "grad_norm": 0.1533203125, "learning_rate": 0.0001, "loss": 0.2688, "step": 275 }, { "epoch": 0.0448, "grad_norm": 0.1767578125, "learning_rate": 0.0001, "loss": 0.2806, "step": 280 }, { "epoch": 0.0456, "grad_norm": 0.154296875, "learning_rate": 0.0001, "loss": 0.253, "step": 285 }, { "epoch": 0.0464, "grad_norm": 0.23828125, "learning_rate": 0.0001, "loss": 0.2322, "step": 290 }, { "epoch": 0.0472, "grad_norm": 0.18359375, "learning_rate": 0.0001, "loss": 0.2227, "step": 295 }, { "epoch": 0.048, "grad_norm": 0.2119140625, "learning_rate": 0.0001, "loss": 0.2486, "step": 300 }, { "epoch": 0.0488, "grad_norm": 0.2431640625, "learning_rate": 0.0001, "loss": 0.161, "step": 305 }, { "epoch": 0.0496, "grad_norm": 0.193359375, "learning_rate": 0.0001, "loss": 0.2632, "step": 310 }, { "epoch": 0.0504, "grad_norm": 0.2099609375, "learning_rate": 0.0001, "loss": 0.2556, "step": 315 }, { "epoch": 0.0512, "grad_norm": 0.10302734375, "learning_rate": 0.0001, "loss": 0.2304, "step": 320 }, { "epoch": 0.052, "grad_norm": 0.1259765625, "learning_rate": 0.0001, "loss": 0.2208, "step": 325 }, { "epoch": 0.0528, "grad_norm": 0.1552734375, "learning_rate": 0.0001, "loss": 0.2628, "step": 330 }, { "epoch": 0.0536, "grad_norm": 0.1787109375, "learning_rate": 0.0001, "loss": 0.2933, "step": 335 }, { "epoch": 0.0544, "grad_norm": 0.166015625, "learning_rate": 0.0001, "loss": 0.2582, "step": 340 }, { "epoch": 0.0552, "grad_norm": 0.140625, "learning_rate": 0.0001, "loss": 0.2362, "step": 345 }, { "epoch": 0.056, "grad_norm": 0.1328125, "learning_rate": 0.0001, "loss": 0.1985, "step": 350 }, { "epoch": 0.0568, "grad_norm": 0.19140625, "learning_rate": 0.0001, "loss": 0.149, "step": 355 }, { "epoch": 0.0576, "grad_norm": 0.115234375, "learning_rate": 0.0001, "loss": 0.1826, "step": 360 }, { "epoch": 0.0584, "grad_norm": 0.1748046875, "learning_rate": 0.0001, "loss": 0.2389, "step": 365 }, { "epoch": 0.0592, "grad_norm": 0.185546875, "learning_rate": 0.0001, "loss": 0.3073, "step": 370 }, { "epoch": 0.06, "grad_norm": 0.16015625, "learning_rate": 0.0001, "loss": 0.2194, "step": 375 }, { "epoch": 0.0608, "grad_norm": 0.1708984375, "learning_rate": 0.0001, "loss": 0.2179, "step": 380 }, { "epoch": 0.0616, "grad_norm": 0.185546875, "learning_rate": 0.0001, "loss": 0.2711, "step": 385 }, { "epoch": 0.0624, "grad_norm": 0.12060546875, "learning_rate": 0.0001, "loss": 0.2003, "step": 390 }, { "epoch": 0.0632, "grad_norm": 0.1650390625, "learning_rate": 0.0001, "loss": 0.2109, "step": 395 }, { "epoch": 0.064, "grad_norm": 0.1484375, "learning_rate": 0.0001, "loss": 0.1646, "step": 400 }, { "epoch": 0.0648, "grad_norm": 0.380859375, "learning_rate": 0.0001, "loss": 0.1521, "step": 405 }, { "epoch": 0.0656, "grad_norm": 0.142578125, "learning_rate": 0.0001, "loss": 0.2176, "step": 410 }, { "epoch": 0.0664, "grad_norm": 0.193359375, "learning_rate": 0.0001, "loss": 0.2385, "step": 415 }, { "epoch": 0.0672, "grad_norm": 0.1494140625, "learning_rate": 0.0001, "loss": 0.1947, "step": 420 }, { "epoch": 0.068, "grad_norm": 0.123046875, "learning_rate": 0.0001, "loss": 0.2462, "step": 425 }, { "epoch": 0.0688, "grad_norm": 0.1162109375, "learning_rate": 0.0001, "loss": 0.2514, "step": 430 }, { "epoch": 0.0696, "grad_norm": 0.12451171875, "learning_rate": 0.0001, "loss": 0.1767, "step": 435 }, { "epoch": 0.0704, "grad_norm": 0.17578125, "learning_rate": 0.0001, "loss": 0.2061, "step": 440 }, { "epoch": 0.0712, "grad_norm": 0.169921875, "learning_rate": 0.0001, "loss": 0.2258, "step": 445 }, { "epoch": 0.072, "grad_norm": 0.19140625, "learning_rate": 0.0001, "loss": 0.2221, "step": 450 }, { "epoch": 0.0728, "grad_norm": 0.09228515625, "learning_rate": 0.0001, "loss": 0.1451, "step": 455 }, { "epoch": 0.0736, "grad_norm": 0.1904296875, "learning_rate": 0.0001, "loss": 0.2632, "step": 460 }, { "epoch": 0.0744, "grad_norm": 0.171875, "learning_rate": 0.0001, "loss": 0.2082, "step": 465 }, { "epoch": 0.0752, "grad_norm": 0.1708984375, "learning_rate": 0.0001, "loss": 0.2701, "step": 470 }, { "epoch": 0.076, "grad_norm": 0.154296875, "learning_rate": 0.0001, "loss": 0.2151, "step": 475 }, { "epoch": 0.0768, "grad_norm": 0.15234375, "learning_rate": 0.0001, "loss": 0.1857, "step": 480 }, { "epoch": 0.0776, "grad_norm": 0.1416015625, "learning_rate": 0.0001, "loss": 0.2237, "step": 485 }, { "epoch": 0.0784, "grad_norm": 0.125, "learning_rate": 0.0001, "loss": 0.2201, "step": 490 }, { "epoch": 0.0792, "grad_norm": 0.1875, "learning_rate": 0.0001, "loss": 0.2417, "step": 495 }, { "epoch": 0.08, "grad_norm": 0.11328125, "learning_rate": 0.0001, "loss": 0.1779, "step": 500 }, { "epoch": 0.0808, "grad_norm": 0.087890625, "learning_rate": 0.0001, "loss": 0.1163, "step": 505 }, { "epoch": 0.0816, "grad_norm": 0.11474609375, "learning_rate": 0.0001, "loss": 0.2181, "step": 510 }, { "epoch": 0.0824, "grad_norm": 0.1513671875, "learning_rate": 0.0001, "loss": 0.2088, "step": 515 }, { "epoch": 0.0832, "grad_norm": 0.1787109375, "learning_rate": 0.0001, "loss": 0.2407, "step": 520 }, { "epoch": 0.084, "grad_norm": 0.150390625, "learning_rate": 0.0001, "loss": 0.2139, "step": 525 }, { "epoch": 0.0848, "grad_norm": 0.138671875, "learning_rate": 0.0001, "loss": 0.2365, "step": 530 }, { "epoch": 0.0856, "grad_norm": 0.1826171875, "learning_rate": 0.0001, "loss": 0.2131, "step": 535 }, { "epoch": 0.0864, "grad_norm": 0.1259765625, "learning_rate": 0.0001, "loss": 0.1849, "step": 540 }, { "epoch": 0.0872, "grad_norm": 0.1796875, "learning_rate": 0.0001, "loss": 0.1907, "step": 545 }, { "epoch": 0.088, "grad_norm": 0.095703125, "learning_rate": 0.0001, "loss": 0.1759, "step": 550 }, { "epoch": 0.0888, "grad_norm": 0.119140625, "learning_rate": 0.0001, "loss": 0.114, "step": 555 }, { "epoch": 0.0896, "grad_norm": 0.1708984375, "learning_rate": 0.0001, "loss": 0.2083, "step": 560 }, { "epoch": 0.0904, "grad_norm": 0.107421875, "learning_rate": 0.0001, "loss": 0.2194, "step": 565 }, { "epoch": 0.0912, "grad_norm": 0.1259765625, "learning_rate": 0.0001, "loss": 0.2528, "step": 570 }, { "epoch": 0.092, "grad_norm": 0.1796875, "learning_rate": 0.0001, "loss": 0.2222, "step": 575 }, { "epoch": 0.0928, "grad_norm": 0.07568359375, "learning_rate": 0.0001, "loss": 0.2424, "step": 580 }, { "epoch": 0.0936, "grad_norm": 0.1142578125, "learning_rate": 0.0001, "loss": 0.2629, "step": 585 }, { "epoch": 0.0944, "grad_norm": 0.11474609375, "learning_rate": 0.0001, "loss": 0.2342, "step": 590 }, { "epoch": 0.0952, "grad_norm": 0.181640625, "learning_rate": 0.0001, "loss": 0.2613, "step": 595 }, { "epoch": 0.096, "grad_norm": 0.1484375, "learning_rate": 0.0001, "loss": 0.1535, "step": 600 }, { "epoch": 0.0968, "grad_norm": 0.2421875, "learning_rate": 0.0001, "loss": 0.1473, "step": 605 }, { "epoch": 0.0976, "grad_norm": 0.0908203125, "learning_rate": 0.0001, "loss": 0.1887, "step": 610 }, { "epoch": 0.0984, "grad_norm": 0.1259765625, "learning_rate": 0.0001, "loss": 0.2135, "step": 615 }, { "epoch": 0.0992, "grad_norm": 0.158203125, "learning_rate": 0.0001, "loss": 0.2272, "step": 620 }, { "epoch": 0.1, "grad_norm": 0.10986328125, "learning_rate": 0.0001, "loss": 0.2209, "step": 625 }, { "epoch": 0.1008, "grad_norm": 0.1474609375, "learning_rate": 0.0001, "loss": 0.2423, "step": 630 }, { "epoch": 0.1016, "grad_norm": 0.103515625, "learning_rate": 0.0001, "loss": 0.2474, "step": 635 }, { "epoch": 0.1024, "grad_norm": 0.091796875, "learning_rate": 0.0001, "loss": 0.2031, "step": 640 }, { "epoch": 0.1032, "grad_norm": 0.1455078125, "learning_rate": 0.0001, "loss": 0.1811, "step": 645 }, { "epoch": 0.104, "grad_norm": 0.173828125, "learning_rate": 0.0001, "loss": 0.1811, "step": 650 }, { "epoch": 0.1048, "grad_norm": 0.248046875, "learning_rate": 0.0001, "loss": 0.1475, "step": 655 }, { "epoch": 0.1056, "grad_norm": 0.169921875, "learning_rate": 0.0001, "loss": 0.2052, "step": 660 }, { "epoch": 0.1064, "grad_norm": 0.14453125, "learning_rate": 0.0001, "loss": 0.201, "step": 665 }, { "epoch": 0.1072, "grad_norm": 0.173828125, "learning_rate": 0.0001, "loss": 0.2274, "step": 670 }, { "epoch": 0.108, "grad_norm": 0.2138671875, "learning_rate": 0.0001, "loss": 0.2101, "step": 675 }, { "epoch": 0.1088, "grad_norm": 0.1328125, "learning_rate": 0.0001, "loss": 0.1785, "step": 680 }, { "epoch": 0.1096, "grad_norm": 0.1494140625, "learning_rate": 0.0001, "loss": 0.2615, "step": 685 }, { "epoch": 0.1104, "grad_norm": 0.13671875, "learning_rate": 0.0001, "loss": 0.2318, "step": 690 }, { "epoch": 0.1112, "grad_norm": 0.1806640625, "learning_rate": 0.0001, "loss": 0.2392, "step": 695 }, { "epoch": 0.112, "grad_norm": 0.1767578125, "learning_rate": 0.0001, "loss": 0.2709, "step": 700 }, { "epoch": 0.1128, "grad_norm": 0.130859375, "learning_rate": 0.0001, "loss": 0.1266, "step": 705 }, { "epoch": 0.1136, "grad_norm": 0.146484375, "learning_rate": 0.0001, "loss": 0.1663, "step": 710 }, { "epoch": 0.1144, "grad_norm": 0.138671875, "learning_rate": 0.0001, "loss": 0.2361, "step": 715 }, { "epoch": 0.1152, "grad_norm": 0.2099609375, "learning_rate": 0.0001, "loss": 0.2468, "step": 720 }, { "epoch": 0.116, "grad_norm": 0.142578125, "learning_rate": 0.0001, "loss": 0.2149, "step": 725 }, { "epoch": 0.1168, "grad_norm": 0.11083984375, "learning_rate": 0.0001, "loss": 0.1436, "step": 730 }, { "epoch": 0.1176, "grad_norm": 0.1171875, "learning_rate": 0.0001, "loss": 0.1838, "step": 735 }, { "epoch": 0.1184, "grad_norm": 0.126953125, "learning_rate": 0.0001, "loss": 0.2348, "step": 740 }, { "epoch": 0.1192, "grad_norm": 0.09716796875, "learning_rate": 0.0001, "loss": 0.2123, "step": 745 }, { "epoch": 0.12, "grad_norm": 0.150390625, "learning_rate": 0.0001, "loss": 0.199, "step": 750 }, { "epoch": 0.1208, "grad_norm": 0.1455078125, "learning_rate": 0.0001, "loss": 0.1164, "step": 755 }, { "epoch": 0.1216, "grad_norm": 0.115234375, "learning_rate": 0.0001, "loss": 0.2354, "step": 760 }, { "epoch": 0.1224, "grad_norm": 0.1220703125, "learning_rate": 0.0001, "loss": 0.1993, "step": 765 }, { "epoch": 0.1232, "grad_norm": 0.08984375, "learning_rate": 0.0001, "loss": 0.1933, "step": 770 }, { "epoch": 0.124, "grad_norm": 0.13671875, "learning_rate": 0.0001, "loss": 0.2397, "step": 775 }, { "epoch": 0.1248, "grad_norm": 0.1318359375, "learning_rate": 0.0001, "loss": 0.2339, "step": 780 }, { "epoch": 0.1256, "grad_norm": 0.205078125, "learning_rate": 0.0001, "loss": 0.1864, "step": 785 }, { "epoch": 0.1264, "grad_norm": 0.1103515625, "learning_rate": 0.0001, "loss": 0.225, "step": 790 }, { "epoch": 0.1272, "grad_norm": 0.203125, "learning_rate": 0.0001, "loss": 0.2022, "step": 795 }, { "epoch": 0.128, "grad_norm": 0.130859375, "learning_rate": 0.0001, "loss": 0.153, "step": 800 }, { "epoch": 0.1288, "grad_norm": 0.09033203125, "learning_rate": 0.0001, "loss": 0.0892, "step": 805 }, { "epoch": 0.1296, "grad_norm": 0.189453125, "learning_rate": 0.0001, "loss": 0.1925, "step": 810 }, { "epoch": 0.1304, "grad_norm": 0.15625, "learning_rate": 0.0001, "loss": 0.2262, "step": 815 }, { "epoch": 0.1312, "grad_norm": 0.21484375, "learning_rate": 0.0001, "loss": 0.2531, "step": 820 }, { "epoch": 0.132, "grad_norm": 0.12109375, "learning_rate": 0.0001, "loss": 0.1882, "step": 825 }, { "epoch": 0.1328, "grad_norm": 0.11279296875, "learning_rate": 0.0001, "loss": 0.1733, "step": 830 }, { "epoch": 0.1336, "grad_norm": 0.123046875, "learning_rate": 0.0001, "loss": 0.2105, "step": 835 }, { "epoch": 0.1344, "grad_norm": 0.1796875, "learning_rate": 0.0001, "loss": 0.1886, "step": 840 }, { "epoch": 0.1352, "grad_norm": 0.15625, "learning_rate": 0.0001, "loss": 0.2178, "step": 845 }, { "epoch": 0.136, "grad_norm": 0.1552734375, "learning_rate": 0.0001, "loss": 0.22, "step": 850 }, { "epoch": 0.1368, "grad_norm": 0.09765625, "learning_rate": 0.0001, "loss": 0.12, "step": 855 }, { "epoch": 0.1376, "grad_norm": 0.11767578125, "learning_rate": 0.0001, "loss": 0.2326, "step": 860 }, { "epoch": 0.1384, "grad_norm": 0.12890625, "learning_rate": 0.0001, "loss": 0.2202, "step": 865 }, { "epoch": 0.1392, "grad_norm": 0.11181640625, "learning_rate": 0.0001, "loss": 0.1767, "step": 870 }, { "epoch": 0.14, "grad_norm": 0.115234375, "learning_rate": 0.0001, "loss": 0.2105, "step": 875 }, { "epoch": 0.1408, "grad_norm": 0.138671875, "learning_rate": 0.0001, "loss": 0.2157, "step": 880 }, { "epoch": 0.1416, "grad_norm": 0.1298828125, "learning_rate": 0.0001, "loss": 0.2028, "step": 885 }, { "epoch": 0.1424, "grad_norm": 0.15625, "learning_rate": 0.0001, "loss": 0.2519, "step": 890 }, { "epoch": 0.1432, "grad_norm": 0.10693359375, "learning_rate": 0.0001, "loss": 0.1949, "step": 895 }, { "epoch": 0.144, "grad_norm": 0.1259765625, "learning_rate": 0.0001, "loss": 0.2124, "step": 900 }, { "epoch": 0.1448, "grad_norm": 0.09375, "learning_rate": 0.0001, "loss": 0.1201, "step": 905 }, { "epoch": 0.1456, "grad_norm": 0.10791015625, "learning_rate": 0.0001, "loss": 0.2231, "step": 910 }, { "epoch": 0.1464, "grad_norm": 0.0986328125, "learning_rate": 0.0001, "loss": 0.2124, "step": 915 }, { "epoch": 0.1472, "grad_norm": 0.1474609375, "learning_rate": 0.0001, "loss": 0.1918, "step": 920 }, { "epoch": 0.148, "grad_norm": 0.1376953125, "learning_rate": 0.0001, "loss": 0.2144, "step": 925 }, { "epoch": 0.1488, "grad_norm": 0.2001953125, "learning_rate": 0.0001, "loss": 0.2364, "step": 930 }, { "epoch": 0.1496, "grad_norm": 0.140625, "learning_rate": 0.0001, "loss": 0.197, "step": 935 }, { "epoch": 0.1504, "grad_norm": 0.091796875, "learning_rate": 0.0001, "loss": 0.2397, "step": 940 }, { "epoch": 0.1512, "grad_norm": 0.14453125, "learning_rate": 0.0001, "loss": 0.1847, "step": 945 }, { "epoch": 0.152, "grad_norm": 0.091796875, "learning_rate": 0.0001, "loss": 0.1096, "step": 950 }, { "epoch": 0.1528, "grad_norm": 0.142578125, "learning_rate": 0.0001, "loss": 0.1233, "step": 955 }, { "epoch": 0.1536, "grad_norm": 0.1298828125, "learning_rate": 0.0001, "loss": 0.1697, "step": 960 }, { "epoch": 0.1544, "grad_norm": 0.13671875, "learning_rate": 0.0001, "loss": 0.2191, "step": 965 }, { "epoch": 0.1552, "grad_norm": 0.1552734375, "learning_rate": 0.0001, "loss": 0.1959, "step": 970 }, { "epoch": 0.156, "grad_norm": 0.1328125, "learning_rate": 0.0001, "loss": 0.1908, "step": 975 }, { "epoch": 0.1568, "grad_norm": 0.15234375, "learning_rate": 0.0001, "loss": 0.1795, "step": 980 }, { "epoch": 0.1576, "grad_norm": 0.126953125, "learning_rate": 0.0001, "loss": 0.2114, "step": 985 }, { "epoch": 0.1584, "grad_norm": 0.10400390625, "learning_rate": 0.0001, "loss": 0.1933, "step": 990 }, { "epoch": 0.1592, "grad_norm": 0.09033203125, "learning_rate": 0.0001, "loss": 0.1937, "step": 995 }, { "epoch": 0.16, "grad_norm": 0.09619140625, "learning_rate": 0.0001, "loss": 0.1617, "step": 1000 }, { "epoch": 0.1608, "grad_norm": 0.10400390625, "learning_rate": 0.0001, "loss": 0.1408, "step": 1005 }, { "epoch": 0.1616, "grad_norm": 0.07373046875, "learning_rate": 0.0001, "loss": 0.1704, "step": 1010 }, { "epoch": 0.1624, "grad_norm": 0.1787109375, "learning_rate": 0.0001, "loss": 0.2203, "step": 1015 }, { "epoch": 0.1632, "grad_norm": 0.10986328125, "learning_rate": 0.0001, "loss": 0.197, "step": 1020 }, { "epoch": 0.164, "grad_norm": 0.11572265625, "learning_rate": 0.0001, "loss": 0.223, "step": 1025 }, { "epoch": 0.1648, "grad_norm": 0.1142578125, "learning_rate": 0.0001, "loss": 0.2103, "step": 1030 }, { "epoch": 0.1656, "grad_norm": 0.2158203125, "learning_rate": 0.0001, "loss": 0.2367, "step": 1035 }, { "epoch": 0.1664, "grad_norm": 0.1142578125, "learning_rate": 0.0001, "loss": 0.2092, "step": 1040 }, { "epoch": 0.1672, "grad_norm": 0.1513671875, "learning_rate": 0.0001, "loss": 0.2141, "step": 1045 }, { "epoch": 0.168, "grad_norm": 0.15625, "learning_rate": 0.0001, "loss": 0.2024, "step": 1050 }, { "epoch": 0.1688, "grad_norm": 0.1435546875, "learning_rate": 0.0001, "loss": 0.1401, "step": 1055 }, { "epoch": 0.1696, "grad_norm": 0.1337890625, "learning_rate": 0.0001, "loss": 0.1902, "step": 1060 }, { "epoch": 0.1704, "grad_norm": 0.119140625, "learning_rate": 0.0001, "loss": 0.1907, "step": 1065 }, { "epoch": 0.1712, "grad_norm": 0.1259765625, "learning_rate": 0.0001, "loss": 0.2219, "step": 1070 }, { "epoch": 0.172, "grad_norm": 0.115234375, "learning_rate": 0.0001, "loss": 0.2513, "step": 1075 }, { "epoch": 0.1728, "grad_norm": 0.08642578125, "learning_rate": 0.0001, "loss": 0.2536, "step": 1080 }, { "epoch": 0.1736, "grad_norm": 0.12158203125, "learning_rate": 0.0001, "loss": 0.172, "step": 1085 }, { "epoch": 0.1744, "grad_norm": 0.177734375, "learning_rate": 0.0001, "loss": 0.2361, "step": 1090 }, { "epoch": 0.1752, "grad_norm": 0.125, "learning_rate": 0.0001, "loss": 0.2239, "step": 1095 }, { "epoch": 0.176, "grad_norm": 0.09619140625, "learning_rate": 0.0001, "loss": 0.166, "step": 1100 }, { "epoch": 0.1768, "grad_norm": 0.1259765625, "learning_rate": 0.0001, "loss": 0.1204, "step": 1105 }, { "epoch": 0.1776, "grad_norm": 0.109375, "learning_rate": 0.0001, "loss": 0.1967, "step": 1110 }, { "epoch": 0.1784, "grad_norm": 0.1396484375, "learning_rate": 0.0001, "loss": 0.1697, "step": 1115 }, { "epoch": 0.1792, "grad_norm": 0.10888671875, "learning_rate": 0.0001, "loss": 0.1977, "step": 1120 }, { "epoch": 0.18, "grad_norm": 0.0771484375, "learning_rate": 0.0001, "loss": 0.1822, "step": 1125 }, { "epoch": 0.1808, "grad_norm": 0.1357421875, "learning_rate": 0.0001, "loss": 0.1855, "step": 1130 }, { "epoch": 0.1816, "grad_norm": 0.12060546875, "learning_rate": 0.0001, "loss": 0.1619, "step": 1135 }, { "epoch": 0.1824, "grad_norm": 0.25, "learning_rate": 0.0001, "loss": 0.2243, "step": 1140 }, { "epoch": 0.1832, "grad_norm": 0.16796875, "learning_rate": 0.0001, "loss": 0.2044, "step": 1145 }, { "epoch": 0.184, "grad_norm": 0.13671875, "learning_rate": 0.0001, "loss": 0.1434, "step": 1150 }, { "epoch": 0.1848, "grad_norm": 0.21875, "learning_rate": 0.0001, "loss": 0.1165, "step": 1155 }, { "epoch": 0.1856, "grad_norm": 0.1298828125, "learning_rate": 0.0001, "loss": 0.1748, "step": 1160 }, { "epoch": 0.1864, "grad_norm": 0.09326171875, "learning_rate": 0.0001, "loss": 0.1974, "step": 1165 }, { "epoch": 0.1872, "grad_norm": 0.12451171875, "learning_rate": 0.0001, "loss": 0.1945, "step": 1170 }, { "epoch": 0.188, "grad_norm": 0.0927734375, "learning_rate": 0.0001, "loss": 0.2474, "step": 1175 }, { "epoch": 0.1888, "grad_norm": 0.1552734375, "learning_rate": 0.0001, "loss": 0.1762, "step": 1180 }, { "epoch": 0.1896, "grad_norm": 0.1103515625, "learning_rate": 0.0001, "loss": 0.2316, "step": 1185 }, { "epoch": 0.1904, "grad_norm": 0.09228515625, "learning_rate": 0.0001, "loss": 0.2118, "step": 1190 }, { "epoch": 0.1912, "grad_norm": 0.140625, "learning_rate": 0.0001, "loss": 0.1775, "step": 1195 }, { "epoch": 0.192, "grad_norm": 0.1376953125, "learning_rate": 0.0001, "loss": 0.178, "step": 1200 }, { "epoch": 0.1928, "grad_norm": 0.1083984375, "learning_rate": 0.0001, "loss": 0.1185, "step": 1205 }, { "epoch": 0.1936, "grad_norm": 0.1357421875, "learning_rate": 0.0001, "loss": 0.2021, "step": 1210 }, { "epoch": 0.1944, "grad_norm": 0.146484375, "learning_rate": 0.0001, "loss": 0.1596, "step": 1215 }, { "epoch": 0.1952, "grad_norm": 0.10400390625, "learning_rate": 0.0001, "loss": 0.1944, "step": 1220 }, { "epoch": 0.196, "grad_norm": 0.125, "learning_rate": 0.0001, "loss": 0.196, "step": 1225 }, { "epoch": 0.1968, "grad_norm": 0.1474609375, "learning_rate": 0.0001, "loss": 0.2183, "step": 1230 }, { "epoch": 0.1976, "grad_norm": 0.125, "learning_rate": 0.0001, "loss": 0.1977, "step": 1235 }, { "epoch": 0.1984, "grad_norm": 0.0849609375, "learning_rate": 0.0001, "loss": 0.1946, "step": 1240 }, { "epoch": 0.1992, "grad_norm": 0.1328125, "learning_rate": 0.0001, "loss": 0.2813, "step": 1245 }, { "epoch": 0.2, "grad_norm": 0.244140625, "learning_rate": 0.0001, "loss": 0.1522, "step": 1250 }, { "epoch": 0.2008, "grad_norm": 0.12255859375, "learning_rate": 0.0001, "loss": 0.098, "step": 1255 }, { "epoch": 0.2016, "grad_norm": 0.138671875, "learning_rate": 0.0001, "loss": 0.2329, "step": 1260 }, { "epoch": 0.2024, "grad_norm": 0.11669921875, "learning_rate": 0.0001, "loss": 0.1867, "step": 1265 }, { "epoch": 0.2032, "grad_norm": 0.166015625, "learning_rate": 0.0001, "loss": 0.228, "step": 1270 }, { "epoch": 0.204, "grad_norm": 0.1005859375, "learning_rate": 0.0001, "loss": 0.2163, "step": 1275 }, { "epoch": 0.2048, "grad_norm": 0.103515625, "learning_rate": 0.0001, "loss": 0.2058, "step": 1280 }, { "epoch": 0.2056, "grad_norm": 0.1552734375, "learning_rate": 0.0001, "loss": 0.1649, "step": 1285 }, { "epoch": 0.2064, "grad_norm": 0.12353515625, "learning_rate": 0.0001, "loss": 0.2473, "step": 1290 }, { "epoch": 0.2072, "grad_norm": 0.11279296875, "learning_rate": 0.0001, "loss": 0.2023, "step": 1295 }, { "epoch": 0.208, "grad_norm": 0.11376953125, "learning_rate": 0.0001, "loss": 0.1992, "step": 1300 }, { "epoch": 0.2088, "grad_norm": 0.10791015625, "learning_rate": 0.0001, "loss": 0.1262, "step": 1305 }, { "epoch": 0.2096, "grad_norm": 0.103515625, "learning_rate": 0.0001, "loss": 0.1956, "step": 1310 }, { "epoch": 0.2104, "grad_norm": 0.095703125, "learning_rate": 0.0001, "loss": 0.2246, "step": 1315 }, { "epoch": 0.2112, "grad_norm": 0.11083984375, "learning_rate": 0.0001, "loss": 0.1975, "step": 1320 }, { "epoch": 0.212, "grad_norm": 0.0986328125, "learning_rate": 0.0001, "loss": 0.1886, "step": 1325 }, { "epoch": 0.2128, "grad_norm": 0.146484375, "learning_rate": 0.0001, "loss": 0.2105, "step": 1330 }, { "epoch": 0.2136, "grad_norm": 0.10009765625, "learning_rate": 0.0001, "loss": 0.1991, "step": 1335 }, { "epoch": 0.2144, "grad_norm": 0.1416015625, "learning_rate": 0.0001, "loss": 0.2136, "step": 1340 }, { "epoch": 0.2152, "grad_norm": 0.1123046875, "learning_rate": 0.0001, "loss": 0.2297, "step": 1345 }, { "epoch": 0.216, "grad_norm": 0.15234375, "learning_rate": 0.0001, "loss": 0.1733, "step": 1350 }, { "epoch": 0.2168, "grad_norm": 0.1455078125, "learning_rate": 0.0001, "loss": 0.1281, "step": 1355 }, { "epoch": 0.2176, "grad_norm": 0.08203125, "learning_rate": 0.0001, "loss": 0.1533, "step": 1360 }, { "epoch": 0.2184, "grad_norm": 0.1376953125, "learning_rate": 0.0001, "loss": 0.2096, "step": 1365 }, { "epoch": 0.2192, "grad_norm": 0.150390625, "learning_rate": 0.0001, "loss": 0.2105, "step": 1370 }, { "epoch": 0.22, "grad_norm": 0.10693359375, "learning_rate": 0.0001, "loss": 0.2055, "step": 1375 }, { "epoch": 0.2208, "grad_norm": 0.126953125, "learning_rate": 0.0001, "loss": 0.2199, "step": 1380 }, { "epoch": 0.2216, "grad_norm": 0.08251953125, "learning_rate": 0.0001, "loss": 0.1908, "step": 1385 }, { "epoch": 0.2224, "grad_norm": 0.10400390625, "learning_rate": 0.0001, "loss": 0.1514, "step": 1390 }, { "epoch": 0.2232, "grad_norm": 0.11376953125, "learning_rate": 0.0001, "loss": 0.179, "step": 1395 }, { "epoch": 0.224, "grad_norm": 0.1318359375, "learning_rate": 0.0001, "loss": 0.1964, "step": 1400 }, { "epoch": 0.2248, "grad_norm": 0.1005859375, "learning_rate": 0.0001, "loss": 0.0723, "step": 1405 }, { "epoch": 0.2256, "grad_norm": 0.11962890625, "learning_rate": 0.0001, "loss": 0.1509, "step": 1410 }, { "epoch": 0.2264, "grad_norm": 0.12890625, "learning_rate": 0.0001, "loss": 0.2137, "step": 1415 }, { "epoch": 0.2272, "grad_norm": 0.11376953125, "learning_rate": 0.0001, "loss": 0.2211, "step": 1420 }, { "epoch": 0.228, "grad_norm": 0.11669921875, "learning_rate": 0.0001, "loss": 0.2314, "step": 1425 }, { "epoch": 0.2288, "grad_norm": 0.17578125, "learning_rate": 0.0001, "loss": 0.2205, "step": 1430 }, { "epoch": 0.2296, "grad_norm": 0.091796875, "learning_rate": 0.0001, "loss": 0.1623, "step": 1435 }, { "epoch": 0.2304, "grad_norm": 0.10986328125, "learning_rate": 0.0001, "loss": 0.2121, "step": 1440 }, { "epoch": 0.2312, "grad_norm": 0.11328125, "learning_rate": 0.0001, "loss": 0.2019, "step": 1445 }, { "epoch": 0.232, "grad_norm": 0.0888671875, "learning_rate": 0.0001, "loss": 0.1582, "step": 1450 }, { "epoch": 0.2328, "grad_norm": 0.1025390625, "learning_rate": 0.0001, "loss": 0.1088, "step": 1455 }, { "epoch": 0.2336, "grad_norm": 0.11962890625, "learning_rate": 0.0001, "loss": 0.1655, "step": 1460 }, { "epoch": 0.2344, "grad_norm": 0.130859375, "learning_rate": 0.0001, "loss": 0.2243, "step": 1465 }, { "epoch": 0.2352, "grad_norm": 0.1982421875, "learning_rate": 0.0001, "loss": 0.2082, "step": 1470 }, { "epoch": 0.236, "grad_norm": 0.1259765625, "learning_rate": 0.0001, "loss": 0.2045, "step": 1475 }, { "epoch": 0.2368, "grad_norm": 0.09619140625, "learning_rate": 0.0001, "loss": 0.1976, "step": 1480 }, { "epoch": 0.2376, "grad_norm": 0.142578125, "learning_rate": 0.0001, "loss": 0.1891, "step": 1485 }, { "epoch": 0.2384, "grad_norm": 0.2255859375, "learning_rate": 0.0001, "loss": 0.1925, "step": 1490 }, { "epoch": 0.2392, "grad_norm": 0.2275390625, "learning_rate": 0.0001, "loss": 0.2374, "step": 1495 }, { "epoch": 0.24, "grad_norm": 0.1318359375, "learning_rate": 0.0001, "loss": 0.1455, "step": 1500 }, { "epoch": 0.2408, "grad_norm": 0.146484375, "learning_rate": 0.0001, "loss": 0.096, "step": 1505 }, { "epoch": 0.2416, "grad_norm": 0.12109375, "learning_rate": 0.0001, "loss": 0.1953, "step": 1510 }, { "epoch": 0.2424, "grad_norm": 0.1630859375, "learning_rate": 0.0001, "loss": 0.22, "step": 1515 }, { "epoch": 0.2432, "grad_norm": 0.20703125, "learning_rate": 0.0001, "loss": 0.2312, "step": 1520 }, { "epoch": 0.244, "grad_norm": 0.1484375, "learning_rate": 0.0001, "loss": 0.1818, "step": 1525 }, { "epoch": 0.2448, "grad_norm": 0.134765625, "learning_rate": 0.0001, "loss": 0.1695, "step": 1530 }, { "epoch": 0.2456, "grad_norm": 0.0849609375, "learning_rate": 0.0001, "loss": 0.2185, "step": 1535 }, { "epoch": 0.2464, "grad_norm": 0.1708984375, "learning_rate": 0.0001, "loss": 0.1862, "step": 1540 }, { "epoch": 0.2472, "grad_norm": 0.1259765625, "learning_rate": 0.0001, "loss": 0.1921, "step": 1545 }, { "epoch": 0.248, "grad_norm": 0.158203125, "learning_rate": 0.0001, "loss": 0.1692, "step": 1550 }, { "epoch": 0.2488, "grad_norm": 0.111328125, "learning_rate": 0.0001, "loss": 0.1138, "step": 1555 }, { "epoch": 0.2496, "grad_norm": 0.1748046875, "learning_rate": 0.0001, "loss": 0.234, "step": 1560 }, { "epoch": 0.2504, "grad_norm": 0.11767578125, "learning_rate": 0.0001, "loss": 0.1303, "step": 1565 }, { "epoch": 0.2512, "grad_norm": 0.09765625, "learning_rate": 0.0001, "loss": 0.1755, "step": 1570 }, { "epoch": 0.252, "grad_norm": 0.1533203125, "learning_rate": 0.0001, "loss": 0.2128, "step": 1575 }, { "epoch": 0.2528, "grad_norm": 0.06201171875, "learning_rate": 0.0001, "loss": 0.2095, "step": 1580 }, { "epoch": 0.2536, "grad_norm": 0.130859375, "learning_rate": 0.0001, "loss": 0.2124, "step": 1585 }, { "epoch": 0.2544, "grad_norm": 0.1083984375, "learning_rate": 0.0001, "loss": 0.1692, "step": 1590 }, { "epoch": 0.2552, "grad_norm": 0.1552734375, "learning_rate": 0.0001, "loss": 0.1912, "step": 1595 }, { "epoch": 0.256, "grad_norm": 0.1484375, "learning_rate": 0.0001, "loss": 0.1554, "step": 1600 }, { "epoch": 0.2568, "grad_norm": 0.11669921875, "learning_rate": 0.0001, "loss": 0.1131, "step": 1605 }, { "epoch": 0.2576, "grad_norm": 0.12109375, "learning_rate": 0.0001, "loss": 0.2317, "step": 1610 }, { "epoch": 0.2584, "grad_norm": 0.1064453125, "learning_rate": 0.0001, "loss": 0.1861, "step": 1615 }, { "epoch": 0.2592, "grad_norm": 0.12255859375, "learning_rate": 0.0001, "loss": 0.1415, "step": 1620 }, { "epoch": 0.26, "grad_norm": 0.15625, "learning_rate": 0.0001, "loss": 0.1972, "step": 1625 }, { "epoch": 0.2608, "grad_norm": 0.10009765625, "learning_rate": 0.0001, "loss": 0.1924, "step": 1630 }, { "epoch": 0.2616, "grad_norm": 0.1259765625, "learning_rate": 0.0001, "loss": 0.2503, "step": 1635 }, { "epoch": 0.2624, "grad_norm": 0.169921875, "learning_rate": 0.0001, "loss": 0.2585, "step": 1640 }, { "epoch": 0.2632, "grad_norm": 0.171875, "learning_rate": 0.0001, "loss": 0.2544, "step": 1645 }, { "epoch": 0.264, "grad_norm": 0.123046875, "learning_rate": 0.0001, "loss": 0.1862, "step": 1650 }, { "epoch": 0.2648, "grad_norm": 0.1298828125, "learning_rate": 0.0001, "loss": 0.095, "step": 1655 }, { "epoch": 0.2656, "grad_norm": 0.0966796875, "learning_rate": 0.0001, "loss": 0.1524, "step": 1660 }, { "epoch": 0.2664, "grad_norm": 0.1455078125, "learning_rate": 0.0001, "loss": 0.2259, "step": 1665 }, { "epoch": 0.2672, "grad_norm": 0.1533203125, "learning_rate": 0.0001, "loss": 0.2124, "step": 1670 }, { "epoch": 0.268, "grad_norm": 0.0927734375, "learning_rate": 0.0001, "loss": 0.2102, "step": 1675 }, { "epoch": 0.2688, "grad_norm": 0.1103515625, "learning_rate": 0.0001, "loss": 0.2028, "step": 1680 }, { "epoch": 0.2696, "grad_norm": 0.0927734375, "learning_rate": 0.0001, "loss": 0.1989, "step": 1685 }, { "epoch": 0.2704, "grad_norm": 0.09716796875, "learning_rate": 0.0001, "loss": 0.2101, "step": 1690 }, { "epoch": 0.2712, "grad_norm": 0.09716796875, "learning_rate": 0.0001, "loss": 0.2219, "step": 1695 }, { "epoch": 0.272, "grad_norm": 0.05517578125, "learning_rate": 0.0001, "loss": 0.1282, "step": 1700 }, { "epoch": 0.2728, "grad_norm": 0.1328125, "learning_rate": 0.0001, "loss": 0.147, "step": 1705 }, { "epoch": 0.2736, "grad_norm": 0.130859375, "learning_rate": 0.0001, "loss": 0.1769, "step": 1710 }, { "epoch": 0.2744, "grad_norm": 0.154296875, "learning_rate": 0.0001, "loss": 0.1632, "step": 1715 }, { "epoch": 0.2752, "grad_norm": 0.1357421875, "learning_rate": 0.0001, "loss": 0.156, "step": 1720 }, { "epoch": 0.276, "grad_norm": 0.14453125, "learning_rate": 0.0001, "loss": 0.2096, "step": 1725 }, { "epoch": 0.2768, "grad_norm": 0.1064453125, "learning_rate": 0.0001, "loss": 0.1689, "step": 1730 }, { "epoch": 0.2776, "grad_norm": 0.1005859375, "learning_rate": 0.0001, "loss": 0.2363, "step": 1735 }, { "epoch": 0.2784, "grad_norm": 0.1435546875, "learning_rate": 0.0001, "loss": 0.2207, "step": 1740 }, { "epoch": 0.2792, "grad_norm": 0.1640625, "learning_rate": 0.0001, "loss": 0.2129, "step": 1745 }, { "epoch": 0.28, "grad_norm": 0.123046875, "learning_rate": 0.0001, "loss": 0.1312, "step": 1750 }, { "epoch": 0.2808, "grad_norm": 0.1376953125, "learning_rate": 0.0001, "loss": 0.1228, "step": 1755 }, { "epoch": 0.2816, "grad_norm": 0.1064453125, "learning_rate": 0.0001, "loss": 0.1979, "step": 1760 }, { "epoch": 0.2824, "grad_norm": 0.1474609375, "learning_rate": 0.0001, "loss": 0.1808, "step": 1765 }, { "epoch": 0.2832, "grad_norm": 0.1484375, "learning_rate": 0.0001, "loss": 0.1939, "step": 1770 }, { "epoch": 0.284, "grad_norm": 0.1923828125, "learning_rate": 0.0001, "loss": 0.2308, "step": 1775 }, { "epoch": 0.2848, "grad_norm": 0.1376953125, "learning_rate": 0.0001, "loss": 0.1864, "step": 1780 }, { "epoch": 0.2856, "grad_norm": 0.13671875, "learning_rate": 0.0001, "loss": 0.2134, "step": 1785 }, { "epoch": 0.2864, "grad_norm": 0.1416015625, "learning_rate": 0.0001, "loss": 0.2068, "step": 1790 }, { "epoch": 0.2872, "grad_norm": 0.1015625, "learning_rate": 0.0001, "loss": 0.2407, "step": 1795 }, { "epoch": 0.288, "grad_norm": 0.07763671875, "learning_rate": 0.0001, "loss": 0.1336, "step": 1800 }, { "epoch": 0.2888, "grad_norm": 0.11474609375, "learning_rate": 0.0001, "loss": 0.155, "step": 1805 }, { "epoch": 0.2896, "grad_norm": 0.1376953125, "learning_rate": 0.0001, "loss": 0.1847, "step": 1810 }, { "epoch": 0.2904, "grad_norm": 0.1259765625, "learning_rate": 0.0001, "loss": 0.2186, "step": 1815 }, { "epoch": 0.2912, "grad_norm": 0.12255859375, "learning_rate": 0.0001, "loss": 0.2134, "step": 1820 }, { "epoch": 0.292, "grad_norm": 0.158203125, "learning_rate": 0.0001, "loss": 0.2259, "step": 1825 }, { "epoch": 0.2928, "grad_norm": 0.11328125, "learning_rate": 0.0001, "loss": 0.218, "step": 1830 }, { "epoch": 0.2936, "grad_norm": 0.13671875, "learning_rate": 0.0001, "loss": 0.1995, "step": 1835 }, { "epoch": 0.2944, "grad_norm": 0.10205078125, "learning_rate": 0.0001, "loss": 0.167, "step": 1840 }, { "epoch": 0.2952, "grad_norm": 0.1162109375, "learning_rate": 0.0001, "loss": 0.1578, "step": 1845 }, { "epoch": 0.296, "grad_norm": 0.06298828125, "learning_rate": 0.0001, "loss": 0.1664, "step": 1850 }, { "epoch": 0.2968, "grad_norm": 0.1494140625, "learning_rate": 0.0001, "loss": 0.128, "step": 1855 }, { "epoch": 0.2976, "grad_norm": 0.095703125, "learning_rate": 0.0001, "loss": 0.1704, "step": 1860 }, { "epoch": 0.2984, "grad_norm": 0.11474609375, "learning_rate": 0.0001, "loss": 0.2083, "step": 1865 }, { "epoch": 0.2992, "grad_norm": 0.1376953125, "learning_rate": 0.0001, "loss": 0.1926, "step": 1870 }, { "epoch": 0.3, "grad_norm": 0.123046875, "learning_rate": 0.0001, "loss": 0.1636, "step": 1875 }, { "epoch": 0.3, "step": 1875, "total_flos": 7.012531902575002e+17, "train_loss": 0.2200873402118683, "train_runtime": 60240.5637, "train_samples_per_second": 0.498, "train_steps_per_second": 0.031 } ], "logging_steps": 5, "max_steps": 1875, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 90, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 7.012531902575002e+17, "train_batch_size": 1, "trial_name": null, "trial_params": null }