{ "best_metric": null, "best_model_checkpoint": null, "epoch": 2.2776, "eval_steps": 1000, "global_step": 5000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.005, "grad_norm": 88.54288482666016, "learning_rate": 4.2000000000000006e-07, "loss": 2.5968, "step": 25 }, { "epoch": 0.01, "grad_norm": 71.25262451171875, "learning_rate": 9.200000000000001e-07, "loss": 2.3534, "step": 50 }, { "epoch": 0.015, "grad_norm": 80.52667999267578, "learning_rate": 1.42e-06, "loss": 1.9642, "step": 75 }, { "epoch": 0.02, "grad_norm": 57.064151763916016, "learning_rate": 1.9200000000000003e-06, "loss": 1.6321, "step": 100 }, { "epoch": 0.025, "grad_norm": 50.983245849609375, "learning_rate": 2.42e-06, "loss": 1.5842, "step": 125 }, { "epoch": 0.03, "grad_norm": 52.22635269165039, "learning_rate": 2.92e-06, "loss": 1.4106, "step": 150 }, { "epoch": 0.035, "grad_norm": 53.9218864440918, "learning_rate": 3.4200000000000007e-06, "loss": 1.417, "step": 175 }, { "epoch": 0.04, "grad_norm": 48.82598876953125, "learning_rate": 3.920000000000001e-06, "loss": 1.4496, "step": 200 }, { "epoch": 0.045, "grad_norm": 63.08296585083008, "learning_rate": 4.42e-06, "loss": 1.3191, "step": 225 }, { "epoch": 0.05, "grad_norm": 41.98011016845703, "learning_rate": 4.92e-06, "loss": 1.0756, "step": 250 }, { "epoch": 0.055, "grad_norm": 53.777217864990234, "learning_rate": 5.420000000000001e-06, "loss": 1.2538, "step": 275 }, { "epoch": 0.06, "grad_norm": 57.49655532836914, "learning_rate": 5.92e-06, "loss": 1.2225, "step": 300 }, { "epoch": 0.065, "grad_norm": 52.660003662109375, "learning_rate": 6.4000000000000006e-06, "loss": 1.2031, "step": 325 }, { "epoch": 0.07, "grad_norm": 45.03782272338867, "learning_rate": 6.9e-06, "loss": 1.2953, "step": 350 }, { "epoch": 0.075, "grad_norm": 49.226261138916016, "learning_rate": 7.4e-06, "loss": 1.0642, "step": 375 }, { "epoch": 0.08, "grad_norm": 33.70058822631836, "learning_rate": 7.9e-06, "loss": 1.0962, "step": 400 }, { "epoch": 0.085, "grad_norm": 47.71772766113281, "learning_rate": 8.400000000000001e-06, "loss": 1.3042, "step": 425 }, { "epoch": 0.09, "grad_norm": 55.22367477416992, "learning_rate": 8.900000000000001e-06, "loss": 1.3794, "step": 450 }, { "epoch": 0.095, "grad_norm": 62.34690475463867, "learning_rate": 9.4e-06, "loss": 1.1516, "step": 475 }, { "epoch": 0.1, "grad_norm": 84.5876693725586, "learning_rate": 9.9e-06, "loss": 1.2527, "step": 500 }, { "epoch": 0.105, "grad_norm": 53.2620735168457, "learning_rate": 9.955555555555556e-06, "loss": 1.1349, "step": 525 }, { "epoch": 0.11, "grad_norm": 54.91654586791992, "learning_rate": 9.9e-06, "loss": 1.2107, "step": 550 }, { "epoch": 0.115, "grad_norm": 43.851173400878906, "learning_rate": 9.844444444444446e-06, "loss": 1.4065, "step": 575 }, { "epoch": 0.12, "grad_norm": 48.8751335144043, "learning_rate": 9.78888888888889e-06, "loss": 1.2255, "step": 600 }, { "epoch": 0.125, "grad_norm": 58.15187454223633, "learning_rate": 9.733333333333334e-06, "loss": 1.1948, "step": 625 }, { "epoch": 0.13, "grad_norm": 53.069602966308594, "learning_rate": 9.677777777777778e-06, "loss": 1.2878, "step": 650 }, { "epoch": 0.135, "grad_norm": 45.524715423583984, "learning_rate": 9.622222222222222e-06, "loss": 1.1325, "step": 675 }, { "epoch": 0.14, "grad_norm": 26.86398696899414, "learning_rate": 9.566666666666668e-06, "loss": 1.0762, "step": 700 }, { "epoch": 0.145, "grad_norm": 48.63008499145508, "learning_rate": 9.511111111111112e-06, "loss": 1.279, "step": 725 }, { "epoch": 0.15, "grad_norm": 50.58299255371094, "learning_rate": 9.455555555555557e-06, "loss": 1.0627, "step": 750 }, { "epoch": 0.155, "grad_norm": 32.71992492675781, "learning_rate": 9.4e-06, "loss": 1.0438, "step": 775 }, { "epoch": 0.16, "grad_norm": 42.12641143798828, "learning_rate": 9.344444444444446e-06, "loss": 1.1204, "step": 800 }, { "epoch": 0.165, "grad_norm": 44.01963806152344, "learning_rate": 9.28888888888889e-06, "loss": 0.908, "step": 825 }, { "epoch": 0.17, "grad_norm": 39.93177795410156, "learning_rate": 9.233333333333334e-06, "loss": 1.158, "step": 850 }, { "epoch": 0.175, "grad_norm": 38.166259765625, "learning_rate": 9.17777777777778e-06, "loss": 1.2021, "step": 875 }, { "epoch": 0.18, "grad_norm": 40.696678161621094, "learning_rate": 9.122222222222223e-06, "loss": 1.0054, "step": 900 }, { "epoch": 0.185, "grad_norm": 48.79116439819336, "learning_rate": 9.066666666666667e-06, "loss": 0.9587, "step": 925 }, { "epoch": 0.19, "grad_norm": 39.57270431518555, "learning_rate": 9.011111111111111e-06, "loss": 1.1084, "step": 950 }, { "epoch": 0.195, "grad_norm": 54.647216796875, "learning_rate": 8.955555555555555e-06, "loss": 1.0416, "step": 975 }, { "epoch": 0.2, "grad_norm": 42.69215774536133, "learning_rate": 8.900000000000001e-06, "loss": 0.97, "step": 1000 }, { "epoch": 0.2, "eval_cer": 38.476190476190474, "eval_loss": 0.7355929613113403, "eval_model_preparation_time": 0.0121, "eval_runtime": 116.5648, "eval_samples_per_second": 2.574, "eval_steps_per_second": 1.287, "eval_wer": 38.17307692307692, "step": 1000 }, { "epoch": 0.205, "grad_norm": 53.78874206542969, "learning_rate": 8.844444444444445e-06, "loss": 1.2455, "step": 1025 }, { "epoch": 0.21, "grad_norm": 34.4659423828125, "learning_rate": 8.788888888888891e-06, "loss": 1.0515, "step": 1050 }, { "epoch": 0.215, "grad_norm": 51.5352668762207, "learning_rate": 8.733333333333333e-06, "loss": 0.9412, "step": 1075 }, { "epoch": 0.22, "grad_norm": 41.004310607910156, "learning_rate": 8.677777777777779e-06, "loss": 1.0831, "step": 1100 }, { "epoch": 0.225, "grad_norm": 43.79686737060547, "learning_rate": 8.622222222222223e-06, "loss": 0.8895, "step": 1125 }, { "epoch": 0.23, "grad_norm": 32.330204010009766, "learning_rate": 8.566666666666667e-06, "loss": 0.8984, "step": 1150 }, { "epoch": 0.235, "grad_norm": 43.38914108276367, "learning_rate": 8.511111111111113e-06, "loss": 1.1215, "step": 1175 }, { "epoch": 0.24, "grad_norm": 62.52465057373047, "learning_rate": 8.455555555555555e-06, "loss": 0.9912, "step": 1200 }, { "epoch": 0.245, "grad_norm": 39.1291618347168, "learning_rate": 8.400000000000001e-06, "loss": 1.0422, "step": 1225 }, { "epoch": 0.25, "grad_norm": 53.39470291137695, "learning_rate": 8.344444444444445e-06, "loss": 1.1502, "step": 1250 }, { "epoch": 0.255, "grad_norm": 42.28481674194336, "learning_rate": 8.288888888888889e-06, "loss": 1.112, "step": 1275 }, { "epoch": 0.26, "grad_norm": 37.374542236328125, "learning_rate": 8.233333333333335e-06, "loss": 1.0152, "step": 1300 }, { "epoch": 0.265, "grad_norm": 52.94354248046875, "learning_rate": 8.177777777777779e-06, "loss": 0.9147, "step": 1325 }, { "epoch": 0.27, "grad_norm": 46.23354721069336, "learning_rate": 8.122222222222223e-06, "loss": 0.8429, "step": 1350 }, { "epoch": 0.275, "grad_norm": 40.089210510253906, "learning_rate": 8.066666666666667e-06, "loss": 0.8026, "step": 1375 }, { "epoch": 0.28, "grad_norm": 42.95212173461914, "learning_rate": 8.011111111111113e-06, "loss": 0.7804, "step": 1400 }, { "epoch": 0.285, "grad_norm": 56.216678619384766, "learning_rate": 7.955555555555557e-06, "loss": 0.8045, "step": 1425 }, { "epoch": 0.29, "grad_norm": 45.01803207397461, "learning_rate": 7.902222222222223e-06, "loss": 0.9807, "step": 1450 }, { "epoch": 0.295, "grad_norm": 63.81135940551758, "learning_rate": 7.846666666666667e-06, "loss": 0.9447, "step": 1475 }, { "epoch": 0.3, "grad_norm": 38.958457946777344, "learning_rate": 7.791111111111111e-06, "loss": 1.0144, "step": 1500 }, { "epoch": 0.305, "grad_norm": 48.87187957763672, "learning_rate": 7.735555555555557e-06, "loss": 1.0261, "step": 1525 }, { "epoch": 0.31, "grad_norm": 34.80329513549805, "learning_rate": 7.680000000000001e-06, "loss": 0.8152, "step": 1550 }, { "epoch": 0.315, "grad_norm": 45.12413024902344, "learning_rate": 7.624444444444445e-06, "loss": 0.8943, "step": 1575 }, { "epoch": 0.32, "grad_norm": 38.103729248046875, "learning_rate": 7.56888888888889e-06, "loss": 1.085, "step": 1600 }, { "epoch": 0.325, "grad_norm": 38.715492248535156, "learning_rate": 7.513333333333334e-06, "loss": 0.9853, "step": 1625 }, { "epoch": 0.33, "grad_norm": 42.01364517211914, "learning_rate": 7.457777777777778e-06, "loss": 0.8214, "step": 1650 }, { "epoch": 0.335, "grad_norm": 37.475799560546875, "learning_rate": 7.402222222222223e-06, "loss": 0.7706, "step": 1675 }, { "epoch": 0.34, "grad_norm": 40.387210845947266, "learning_rate": 7.346666666666668e-06, "loss": 1.081, "step": 1700 }, { "epoch": 0.345, "grad_norm": 21.215946197509766, "learning_rate": 7.291111111111112e-06, "loss": 0.8047, "step": 1725 }, { "epoch": 0.35, "grad_norm": 21.454147338867188, "learning_rate": 7.235555555555556e-06, "loss": 0.836, "step": 1750 }, { "epoch": 0.355, "grad_norm": 46.05310821533203, "learning_rate": 7.180000000000001e-06, "loss": 0.8883, "step": 1775 }, { "epoch": 0.36, "grad_norm": 61.988643646240234, "learning_rate": 7.124444444444445e-06, "loss": 0.8254, "step": 1800 }, { "epoch": 1.0038, "grad_norm": 34.00222396850586, "learning_rate": 7.06888888888889e-06, "loss": 0.5202, "step": 1825 }, { "epoch": 1.0088, "grad_norm": 23.966150283813477, "learning_rate": 7.0133333333333345e-06, "loss": 0.362, "step": 1850 }, { "epoch": 1.0138, "grad_norm": 19.050518035888672, "learning_rate": 6.9577777777777785e-06, "loss": 0.2891, "step": 1875 }, { "epoch": 1.0188, "grad_norm": 34.59785079956055, "learning_rate": 6.902222222222223e-06, "loss": 0.3438, "step": 1900 }, { "epoch": 1.0238, "grad_norm": 29.40850257873535, "learning_rate": 6.846666666666667e-06, "loss": 0.2822, "step": 1925 }, { "epoch": 1.0288, "grad_norm": 15.476716041564941, "learning_rate": 6.7911111111111115e-06, "loss": 0.4843, "step": 1950 }, { "epoch": 1.0338, "grad_norm": 33.912174224853516, "learning_rate": 6.735555555555556e-06, "loss": 0.3741, "step": 1975 }, { "epoch": 1.0388, "grad_norm": 30.411020278930664, "learning_rate": 6.680000000000001e-06, "loss": 0.3044, "step": 2000 }, { "epoch": 1.0388, "eval_cer": 23.904761904761905, "eval_loss": 0.309874564409256, "eval_model_preparation_time": 0.0121, "eval_runtime": 119.6464, "eval_samples_per_second": 2.507, "eval_steps_per_second": 1.254, "eval_wer": 23.46153846153846, "step": 2000 }, { "epoch": 1.0438, "grad_norm": 19.959854125976562, "learning_rate": 6.6244444444444445e-06, "loss": 0.2255, "step": 2025 }, { "epoch": 1.0488, "grad_norm": 20.751602172851562, "learning_rate": 6.568888888888889e-06, "loss": 0.3341, "step": 2050 }, { "epoch": 1.0538, "grad_norm": 24.52460479736328, "learning_rate": 6.513333333333333e-06, "loss": 0.4861, "step": 2075 }, { "epoch": 1.0588, "grad_norm": 5.740904331207275, "learning_rate": 6.457777777777778e-06, "loss": 0.4165, "step": 2100 }, { "epoch": 1.0638, "grad_norm": 24.452116012573242, "learning_rate": 6.402222222222223e-06, "loss": 0.4478, "step": 2125 }, { "epoch": 1.0688, "grad_norm": 29.230716705322266, "learning_rate": 6.346666666666668e-06, "loss": 0.4387, "step": 2150 }, { "epoch": 1.0738, "grad_norm": 41.07571792602539, "learning_rate": 6.291111111111111e-06, "loss": 0.2466, "step": 2175 }, { "epoch": 1.0788, "grad_norm": 19.89525032043457, "learning_rate": 6.235555555555556e-06, "loss": 0.3156, "step": 2200 }, { "epoch": 1.0838, "grad_norm": 33.628971099853516, "learning_rate": 6.18e-06, "loss": 0.3624, "step": 2225 }, { "epoch": 1.0888, "grad_norm": 23.34870147705078, "learning_rate": 6.124444444444445e-06, "loss": 0.4263, "step": 2250 }, { "epoch": 1.0937999999999999, "grad_norm": 30.75408172607422, "learning_rate": 6.06888888888889e-06, "loss": 0.4153, "step": 2275 }, { "epoch": 1.0988, "grad_norm": 7.229944705963135, "learning_rate": 6.013333333333335e-06, "loss": 0.3068, "step": 2300 }, { "epoch": 1.1038000000000001, "grad_norm": 37.87436294555664, "learning_rate": 5.957777777777778e-06, "loss": 0.3049, "step": 2325 }, { "epoch": 1.1088, "grad_norm": 33.34481430053711, "learning_rate": 5.902222222222223e-06, "loss": 0.2759, "step": 2350 }, { "epoch": 1.1138, "grad_norm": 24.211904525756836, "learning_rate": 5.846666666666667e-06, "loss": 0.35, "step": 2375 }, { "epoch": 1.1188, "grad_norm": 41.01383590698242, "learning_rate": 5.791111111111112e-06, "loss": 0.3992, "step": 2400 }, { "epoch": 1.1238, "grad_norm": 6.543262004852295, "learning_rate": 5.735555555555557e-06, "loss": 0.3356, "step": 2425 }, { "epoch": 1.1288, "grad_norm": 34.344913482666016, "learning_rate": 5.68e-06, "loss": 0.3364, "step": 2450 }, { "epoch": 1.1338, "grad_norm": 9.35561752319336, "learning_rate": 5.624444444444445e-06, "loss": 0.3606, "step": 2475 }, { "epoch": 1.1388, "grad_norm": 21.762096405029297, "learning_rate": 5.56888888888889e-06, "loss": 0.3437, "step": 2500 }, { "epoch": 1.1438, "grad_norm": 29.119796752929688, "learning_rate": 5.513333333333334e-06, "loss": 0.428, "step": 2525 }, { "epoch": 1.1488, "grad_norm": 46.66371536254883, "learning_rate": 5.4577777777777785e-06, "loss": 0.4566, "step": 2550 }, { "epoch": 1.1538, "grad_norm": 15.0108642578125, "learning_rate": 5.402222222222223e-06, "loss": 0.3162, "step": 2575 }, { "epoch": 1.1588, "grad_norm": 60.40862274169922, "learning_rate": 5.346666666666667e-06, "loss": 0.3696, "step": 2600 }, { "epoch": 1.1638, "grad_norm": 26.4654598236084, "learning_rate": 5.2911111111111115e-06, "loss": 0.2831, "step": 2625 }, { "epoch": 1.1688, "grad_norm": 23.651691436767578, "learning_rate": 5.235555555555556e-06, "loss": 0.4484, "step": 2650 }, { "epoch": 1.1738, "grad_norm": 33.675167083740234, "learning_rate": 5.18e-06, "loss": 0.3062, "step": 2675 }, { "epoch": 1.1788, "grad_norm": 27.336896896362305, "learning_rate": 5.124444444444445e-06, "loss": 0.3738, "step": 2700 }, { "epoch": 1.1838, "grad_norm": 17.323768615722656, "learning_rate": 5.06888888888889e-06, "loss": 0.3182, "step": 2725 }, { "epoch": 1.1888, "grad_norm": 18.657014846801758, "learning_rate": 5.013333333333333e-06, "loss": 0.3589, "step": 2750 }, { "epoch": 1.1938, "grad_norm": 39.54949188232422, "learning_rate": 4.957777777777778e-06, "loss": 0.4187, "step": 2775 }, { "epoch": 1.1988, "grad_norm": 57.646602630615234, "learning_rate": 4.902222222222222e-06, "loss": 0.3374, "step": 2800 }, { "epoch": 1.2038, "grad_norm": 24.029294967651367, "learning_rate": 4.846666666666667e-06, "loss": 0.2373, "step": 2825 }, { "epoch": 1.2088, "grad_norm": 34.673709869384766, "learning_rate": 4.791111111111111e-06, "loss": 0.3274, "step": 2850 }, { "epoch": 1.2138, "grad_norm": 40.28670120239258, "learning_rate": 4.735555555555556e-06, "loss": 0.3105, "step": 2875 }, { "epoch": 1.2187999999999999, "grad_norm": 22.120824813842773, "learning_rate": 4.680000000000001e-06, "loss": 0.2678, "step": 2900 }, { "epoch": 1.2238, "grad_norm": 21.942718505859375, "learning_rate": 4.624444444444445e-06, "loss": 0.5025, "step": 2925 }, { "epoch": 1.2288000000000001, "grad_norm": 31.645906448364258, "learning_rate": 4.568888888888889e-06, "loss": 0.3753, "step": 2950 }, { "epoch": 1.2338, "grad_norm": 19.00498390197754, "learning_rate": 4.513333333333333e-06, "loss": 0.3421, "step": 2975 }, { "epoch": 1.2388, "grad_norm": 42.73635482788086, "learning_rate": 4.457777777777778e-06, "loss": 0.3108, "step": 3000 }, { "epoch": 1.2388, "eval_cer": 7.7142857142857135, "eval_loss": 0.11533673852682114, "eval_model_preparation_time": 0.0121, "eval_runtime": 120.7728, "eval_samples_per_second": 2.484, "eval_steps_per_second": 1.242, "eval_wer": 7.5, "step": 3000 }, { "epoch": 1.2438, "grad_norm": 20.023395538330078, "learning_rate": 4.402222222222223e-06, "loss": 0.3144, "step": 3025 }, { "epoch": 1.2488, "grad_norm": 32.98371505737305, "learning_rate": 4.346666666666667e-06, "loss": 0.2618, "step": 3050 }, { "epoch": 1.2538, "grad_norm": 42.568119049072266, "learning_rate": 4.291111111111112e-06, "loss": 0.331, "step": 3075 }, { "epoch": 1.2588, "grad_norm": 0.5769469738006592, "learning_rate": 4.235555555555556e-06, "loss": 0.3391, "step": 3100 }, { "epoch": 1.2638, "grad_norm": 20.518579483032227, "learning_rate": 4.18e-06, "loss": 0.2234, "step": 3125 }, { "epoch": 1.2688, "grad_norm": 27.69402313232422, "learning_rate": 4.124444444444445e-06, "loss": 0.295, "step": 3150 }, { "epoch": 1.2738, "grad_norm": 31.121999740600586, "learning_rate": 4.0688888888888896e-06, "loss": 0.3753, "step": 3175 }, { "epoch": 1.2788, "grad_norm": 14.77844524383545, "learning_rate": 4.013333333333334e-06, "loss": 0.3706, "step": 3200 }, { "epoch": 1.2838, "grad_norm": 27.002138137817383, "learning_rate": 3.9577777777777785e-06, "loss": 0.2957, "step": 3225 }, { "epoch": 1.2888, "grad_norm": 13.426039695739746, "learning_rate": 3.9022222222222225e-06, "loss": 0.2902, "step": 3250 }, { "epoch": 1.2938, "grad_norm": 41.75296401977539, "learning_rate": 3.8466666666666665e-06, "loss": 0.2924, "step": 3275 }, { "epoch": 1.2988, "grad_norm": 52.886409759521484, "learning_rate": 3.7911111111111114e-06, "loss": 0.2519, "step": 3300 }, { "epoch": 1.3038, "grad_norm": 33.73487854003906, "learning_rate": 3.7355555555555555e-06, "loss": 0.2966, "step": 3325 }, { "epoch": 1.3088, "grad_norm": 27.994157791137695, "learning_rate": 3.6800000000000003e-06, "loss": 0.2887, "step": 3350 }, { "epoch": 1.3138, "grad_norm": 32.397579193115234, "learning_rate": 3.624444444444445e-06, "loss": 0.2368, "step": 3375 }, { "epoch": 1.3188, "grad_norm": 25.96181869506836, "learning_rate": 3.568888888888889e-06, "loss": 0.2911, "step": 3400 }, { "epoch": 1.3237999999999999, "grad_norm": 7.705018520355225, "learning_rate": 3.5133333333333337e-06, "loss": 0.2647, "step": 3425 }, { "epoch": 1.3288, "grad_norm": 28.499221801757812, "learning_rate": 3.457777777777778e-06, "loss": 0.4149, "step": 3450 }, { "epoch": 1.3338, "grad_norm": 40.06334686279297, "learning_rate": 3.4022222222222222e-06, "loss": 0.2952, "step": 3475 }, { "epoch": 1.3388, "grad_norm": 1.474075198173523, "learning_rate": 3.346666666666667e-06, "loss": 0.207, "step": 3500 }, { "epoch": 1.3437999999999999, "grad_norm": 27.121145248413086, "learning_rate": 3.2911111111111116e-06, "loss": 0.3432, "step": 3525 }, { "epoch": 1.3488, "grad_norm": 22.38213539123535, "learning_rate": 3.2355555555555556e-06, "loss": 0.3522, "step": 3550 }, { "epoch": 1.3538000000000001, "grad_norm": 42.13764190673828, "learning_rate": 3.1800000000000005e-06, "loss": 0.3729, "step": 3575 }, { "epoch": 1.3588, "grad_norm": 27.436649322509766, "learning_rate": 3.124444444444445e-06, "loss": 0.1482, "step": 3600 }, { "epoch": 2.0026, "grad_norm": 16.43301010131836, "learning_rate": 3.068888888888889e-06, "loss": 0.1802, "step": 3625 }, { "epoch": 2.0076, "grad_norm": 11.994711875915527, "learning_rate": 3.013333333333334e-06, "loss": 0.0922, "step": 3650 }, { "epoch": 2.0126, "grad_norm": 7.826560020446777, "learning_rate": 2.957777777777778e-06, "loss": 0.0312, "step": 3675 }, { "epoch": 2.0176, "grad_norm": 6.457350730895996, "learning_rate": 2.9022222222222223e-06, "loss": 0.0343, "step": 3700 }, { "epoch": 2.0226, "grad_norm": 1.624599575996399, "learning_rate": 2.8466666666666672e-06, "loss": 0.0751, "step": 3725 }, { "epoch": 2.0276, "grad_norm": 4.553808212280273, "learning_rate": 2.7911111111111113e-06, "loss": 0.0905, "step": 3750 }, { "epoch": 2.0326, "grad_norm": 4.33929967880249, "learning_rate": 2.7355555555555557e-06, "loss": 0.079, "step": 3775 }, { "epoch": 2.0376, "grad_norm": 11.390565872192383, "learning_rate": 2.68e-06, "loss": 0.064, "step": 3800 }, { "epoch": 2.0426, "grad_norm": 0.5454270243644714, "learning_rate": 2.6244444444444446e-06, "loss": 0.0351, "step": 3825 }, { "epoch": 2.0476, "grad_norm": 25.50840950012207, "learning_rate": 2.568888888888889e-06, "loss": 0.0726, "step": 3850 }, { "epoch": 2.0526, "grad_norm": 5.380075931549072, "learning_rate": 2.5133333333333336e-06, "loss": 0.0405, "step": 3875 }, { "epoch": 2.0576, "grad_norm": 1.0126858949661255, "learning_rate": 2.457777777777778e-06, "loss": 0.0858, "step": 3900 }, { "epoch": 2.0626, "grad_norm": 1.7211323976516724, "learning_rate": 2.4022222222222225e-06, "loss": 0.0431, "step": 3925 }, { "epoch": 2.0676, "grad_norm": 2.049405574798584, "learning_rate": 2.346666666666667e-06, "loss": 0.0127, "step": 3950 }, { "epoch": 2.0726, "grad_norm": 1.535748839378357, "learning_rate": 2.2911111111111114e-06, "loss": 0.0559, "step": 3975 }, { "epoch": 2.0776, "grad_norm": 19.566328048706055, "learning_rate": 2.235555555555556e-06, "loss": 0.0544, "step": 4000 }, { "epoch": 2.0776, "eval_cer": 2.2857142857142856, "eval_loss": 0.029534637928009033, "eval_model_preparation_time": 0.0121, "eval_runtime": 119.7419, "eval_samples_per_second": 2.505, "eval_steps_per_second": 1.253, "eval_wer": 2.307692307692308, "step": 4000 }, { "epoch": 2.0826000000000002, "grad_norm": 0.4915294945240021, "learning_rate": 2.1800000000000003e-06, "loss": 0.0556, "step": 4025 }, { "epoch": 2.0876, "grad_norm": 13.67375659942627, "learning_rate": 2.1244444444444443e-06, "loss": 0.0377, "step": 4050 }, { "epoch": 2.0926, "grad_norm": 0.06471225619316101, "learning_rate": 2.0688888888888892e-06, "loss": 0.0535, "step": 4075 }, { "epoch": 2.0976, "grad_norm": 0.1591552495956421, "learning_rate": 2.0133333333333337e-06, "loss": 0.037, "step": 4100 }, { "epoch": 2.1026, "grad_norm": 1.6957018375396729, "learning_rate": 1.9577777777777777e-06, "loss": 0.0548, "step": 4125 }, { "epoch": 2.1076, "grad_norm": 7.906589508056641, "learning_rate": 1.9022222222222222e-06, "loss": 0.066, "step": 4150 }, { "epoch": 2.1126, "grad_norm": 0.4227987825870514, "learning_rate": 1.8466666666666668e-06, "loss": 0.0687, "step": 4175 }, { "epoch": 2.1176, "grad_norm": 0.3206275999546051, "learning_rate": 1.7911111111111113e-06, "loss": 0.0263, "step": 4200 }, { "epoch": 2.1226, "grad_norm": 0.16232123970985413, "learning_rate": 1.7355555555555555e-06, "loss": 0.0257, "step": 4225 }, { "epoch": 2.1276, "grad_norm": 21.80668830871582, "learning_rate": 1.6800000000000002e-06, "loss": 0.0346, "step": 4250 }, { "epoch": 2.1326, "grad_norm": 3.021885871887207, "learning_rate": 1.6244444444444447e-06, "loss": 0.0715, "step": 4275 }, { "epoch": 2.1376, "grad_norm": 9.51052188873291, "learning_rate": 1.568888888888889e-06, "loss": 0.0686, "step": 4300 }, { "epoch": 2.1426, "grad_norm": 7.770538806915283, "learning_rate": 1.5133333333333334e-06, "loss": 0.0681, "step": 4325 }, { "epoch": 2.1476, "grad_norm": 24.541362762451172, "learning_rate": 1.457777777777778e-06, "loss": 0.0589, "step": 4350 }, { "epoch": 2.1526, "grad_norm": 0.2108163684606552, "learning_rate": 1.4022222222222223e-06, "loss": 0.0407, "step": 4375 }, { "epoch": 2.1576, "grad_norm": 39.73221206665039, "learning_rate": 1.3466666666666668e-06, "loss": 0.0895, "step": 4400 }, { "epoch": 2.1626, "grad_norm": 4.3332438468933105, "learning_rate": 1.2911111111111112e-06, "loss": 0.0645, "step": 4425 }, { "epoch": 2.1676, "grad_norm": 1.142255425453186, "learning_rate": 1.2355555555555557e-06, "loss": 0.0592, "step": 4450 }, { "epoch": 2.1726, "grad_norm": 0.046843066811561584, "learning_rate": 1.1800000000000001e-06, "loss": 0.034, "step": 4475 }, { "epoch": 2.1776, "grad_norm": 14.69985580444336, "learning_rate": 1.1244444444444446e-06, "loss": 0.0704, "step": 4500 }, { "epoch": 2.1826, "grad_norm": 11.02341079711914, "learning_rate": 1.068888888888889e-06, "loss": 0.0621, "step": 4525 }, { "epoch": 2.1875999999999998, "grad_norm": 0.05967080965638161, "learning_rate": 1.0133333333333333e-06, "loss": 0.0331, "step": 4550 }, { "epoch": 2.1926, "grad_norm": 9.536999702453613, "learning_rate": 9.57777777777778e-07, "loss": 0.0762, "step": 4575 }, { "epoch": 2.1976, "grad_norm": 3.740366220474243, "learning_rate": 9.022222222222222e-07, "loss": 0.0346, "step": 4600 }, { "epoch": 2.2026, "grad_norm": 1.1851881742477417, "learning_rate": 8.466666666666668e-07, "loss": 0.0502, "step": 4625 }, { "epoch": 2.2076000000000002, "grad_norm": 23.711048126220703, "learning_rate": 7.911111111111111e-07, "loss": 0.0661, "step": 4650 }, { "epoch": 2.2126, "grad_norm": 1.0020112991333008, "learning_rate": 7.355555555555556e-07, "loss": 0.0252, "step": 4675 }, { "epoch": 2.2176, "grad_norm": 21.021774291992188, "learning_rate": 6.800000000000001e-07, "loss": 0.0415, "step": 4700 }, { "epoch": 2.2226, "grad_norm": 0.8725367784500122, "learning_rate": 6.244444444444445e-07, "loss": 0.0293, "step": 4725 }, { "epoch": 2.2276, "grad_norm": 0.3742901086807251, "learning_rate": 5.68888888888889e-07, "loss": 0.0146, "step": 4750 }, { "epoch": 2.2326, "grad_norm": 12.225566864013672, "learning_rate": 5.133333333333334e-07, "loss": 0.0775, "step": 4775 }, { "epoch": 2.2376, "grad_norm": 0.5857837200164795, "learning_rate": 4.5777777777777784e-07, "loss": 0.0335, "step": 4800 }, { "epoch": 2.2426, "grad_norm": 2.0213797092437744, "learning_rate": 4.0222222222222224e-07, "loss": 0.06, "step": 4825 }, { "epoch": 2.2476, "grad_norm": 0.22115977108478546, "learning_rate": 3.466666666666667e-07, "loss": 0.0437, "step": 4850 }, { "epoch": 2.2526, "grad_norm": 11.783333778381348, "learning_rate": 2.9111111111111116e-07, "loss": 0.0454, "step": 4875 }, { "epoch": 2.2576, "grad_norm": 2.67048978805542, "learning_rate": 2.3555555555555556e-07, "loss": 0.093, "step": 4900 }, { "epoch": 2.2626, "grad_norm": 0.16820040345191956, "learning_rate": 1.8e-07, "loss": 0.1193, "step": 4925 }, { "epoch": 2.2676, "grad_norm": 1.518930435180664, "learning_rate": 1.2444444444444446e-07, "loss": 0.0493, "step": 4950 }, { "epoch": 2.2726, "grad_norm": 0.10118613392114639, "learning_rate": 6.888888888888889e-08, "loss": 0.0538, "step": 4975 }, { "epoch": 2.2776, "grad_norm": 0.5119066834449768, "learning_rate": 1.3333333333333334e-08, "loss": 0.0678, "step": 5000 }, { "epoch": 2.2776, "eval_cer": 0.9523809523809524, "eval_loss": 0.014132725074887276, "eval_model_preparation_time": 0.0121, "eval_runtime": 119.856, "eval_samples_per_second": 2.503, "eval_steps_per_second": 1.252, "eval_wer": 0.9615384615384616, "step": 5000 }, { "epoch": 2.2776, "step": 5000, "total_flos": 1.02060490752e+19, "train_loss": 0.5477467903137208, "train_runtime": 5409.143, "train_samples_per_second": 1.849, "train_steps_per_second": 0.924 } ], "logging_steps": 25, "max_steps": 5000, "num_input_tokens_seen": 0, "num_train_epochs": 9223372036854775807, "save_steps": 1000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.02060490752e+19, "train_batch_size": 2, "trial_name": null, "trial_params": null }