|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 4.6690790241624835, |
|
"eval_steps": 1000, |
|
"global_step": 5000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 38.9212532043457, |
|
"learning_rate": 4.6e-09, |
|
"loss": 1.2718, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 38.4885368347168, |
|
"learning_rate": 9.599999999999998e-09, |
|
"loss": 1.2556, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 41.782493591308594, |
|
"learning_rate": 1.46e-08, |
|
"loss": 1.2688, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 37.6844367980957, |
|
"learning_rate": 1.9599999999999997e-08, |
|
"loss": 1.2761, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 33.974815368652344, |
|
"learning_rate": 2.46e-08, |
|
"loss": 1.2864, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 35.566436767578125, |
|
"learning_rate": 2.96e-08, |
|
"loss": 1.2951, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 36.60067367553711, |
|
"learning_rate": 3.46e-08, |
|
"loss": 1.3226, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 34.72969055175781, |
|
"learning_rate": 3.9600000000000004e-08, |
|
"loss": 1.2698, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 34.141136169433594, |
|
"learning_rate": 4.4599999999999996e-08, |
|
"loss": 1.2497, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 32.15596389770508, |
|
"learning_rate": 4.9599999999999994e-08, |
|
"loss": 1.2155, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 31.284732818603516, |
|
"learning_rate": 5.46e-08, |
|
"loss": 1.2187, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 29.053382873535156, |
|
"learning_rate": 5.96e-08, |
|
"loss": 1.1431, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 28.53986167907715, |
|
"learning_rate": 6.46e-08, |
|
"loss": 1.1342, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 26.937896728515625, |
|
"learning_rate": 6.959999999999999e-08, |
|
"loss": 1.0694, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 28.588802337646484, |
|
"learning_rate": 7.459999999999999e-08, |
|
"loss": 0.9657, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 28.448694229125977, |
|
"learning_rate": 7.96e-08, |
|
"loss": 0.9004, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 27.52383804321289, |
|
"learning_rate": 8.459999999999999e-08, |
|
"loss": 0.8309, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 30.11163902282715, |
|
"learning_rate": 8.96e-08, |
|
"loss": 0.6978, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 23.116390228271484, |
|
"learning_rate": 9.46e-08, |
|
"loss": 0.6042, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"grad_norm": 22.033464431762695, |
|
"learning_rate": 9.959999999999999e-08, |
|
"loss": 0.5676, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"grad_norm": 21.632957458496094, |
|
"learning_rate": 1.046e-07, |
|
"loss": 0.4607, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"grad_norm": 17.29313087463379, |
|
"learning_rate": 1.096e-07, |
|
"loss": 0.4817, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 19.6722412109375, |
|
"learning_rate": 1.146e-07, |
|
"loss": 0.4447, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"grad_norm": 16.24555206298828, |
|
"learning_rate": 1.1959999999999999e-07, |
|
"loss": 0.4028, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"grad_norm": 15.58441162109375, |
|
"learning_rate": 1.246e-07, |
|
"loss": 0.3872, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 14.177542686462402, |
|
"learning_rate": 1.296e-07, |
|
"loss": 0.3831, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 14.774901390075684, |
|
"learning_rate": 1.346e-07, |
|
"loss": 0.3505, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 13.562920570373535, |
|
"learning_rate": 1.396e-07, |
|
"loss": 0.3468, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 12.933964729309082, |
|
"learning_rate": 1.446e-07, |
|
"loss": 0.3402, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 12.395262718200684, |
|
"learning_rate": 1.4960000000000002e-07, |
|
"loss": 0.3317, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 14.720085144042969, |
|
"learning_rate": 1.5459999999999997e-07, |
|
"loss": 0.2923, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"grad_norm": 14.832243919372559, |
|
"learning_rate": 1.5959999999999997e-07, |
|
"loss": 0.3061, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"grad_norm": 11.47397518157959, |
|
"learning_rate": 1.6459999999999998e-07, |
|
"loss": 0.2994, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 17.4206485748291, |
|
"learning_rate": 1.6959999999999998e-07, |
|
"loss": 0.2873, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"grad_norm": 12.757941246032715, |
|
"learning_rate": 1.746e-07, |
|
"loss": 0.263, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"grad_norm": 15.563201904296875, |
|
"learning_rate": 1.796e-07, |
|
"loss": 0.2659, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"grad_norm": 12.566412925720215, |
|
"learning_rate": 1.8459999999999997e-07, |
|
"loss": 0.269, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"grad_norm": 11.327591896057129, |
|
"learning_rate": 1.8959999999999998e-07, |
|
"loss": 0.2678, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"grad_norm": 11.70463752746582, |
|
"learning_rate": 1.9459999999999998e-07, |
|
"loss": 0.2567, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"grad_norm": 11.723058700561523, |
|
"learning_rate": 1.996e-07, |
|
"loss": 0.2552, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"eval_loss": 0.21999725699424744, |
|
"eval_runtime": 7789.979, |
|
"eval_samples_per_second": 1.208, |
|
"eval_steps_per_second": 0.604, |
|
"eval_wer": 0.12200871952204101, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 15.0565824508667, |
|
"learning_rate": 2.046e-07, |
|
"loss": 0.2629, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 12.149203300476074, |
|
"learning_rate": 2.096e-07, |
|
"loss": 0.2535, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 4.3349385261535645, |
|
"learning_rate": 2.146e-07, |
|
"loss": 0.2429, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"grad_norm": 4.134186744689941, |
|
"learning_rate": 2.1959999999999998e-07, |
|
"loss": 0.2214, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"grad_norm": 3.6862359046936035, |
|
"learning_rate": 2.2459999999999999e-07, |
|
"loss": 0.2307, |
|
"step": 1125 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"grad_norm": 4.983838081359863, |
|
"learning_rate": 2.296e-07, |
|
"loss": 0.2327, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"grad_norm": 5.041070461273193, |
|
"learning_rate": 2.346e-07, |
|
"loss": 0.224, |
|
"step": 1175 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"grad_norm": 4.438192844390869, |
|
"learning_rate": 2.396e-07, |
|
"loss": 0.2376, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"grad_norm": 5.2538161277771, |
|
"learning_rate": 2.446e-07, |
|
"loss": 0.2234, |
|
"step": 1225 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"grad_norm": 4.371279239654541, |
|
"learning_rate": 2.4959999999999996e-07, |
|
"loss": 0.2104, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"grad_norm": 3.8350584506988525, |
|
"learning_rate": 2.5459999999999996e-07, |
|
"loss": 0.2221, |
|
"step": 1275 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"grad_norm": 4.0391845703125, |
|
"learning_rate": 2.5959999999999997e-07, |
|
"loss": 0.2147, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"grad_norm": 4.656424045562744, |
|
"learning_rate": 2.646e-07, |
|
"loss": 0.2259, |
|
"step": 1325 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"grad_norm": 4.003029823303223, |
|
"learning_rate": 2.696e-07, |
|
"loss": 0.2073, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"grad_norm": 3.689852476119995, |
|
"learning_rate": 2.746e-07, |
|
"loss": 0.1964, |
|
"step": 1375 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"grad_norm": 2.934403419494629, |
|
"learning_rate": 2.796e-07, |
|
"loss": 0.1997, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"grad_norm": 4.054893493652344, |
|
"learning_rate": 2.846e-07, |
|
"loss": 0.1945, |
|
"step": 1425 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"grad_norm": 4.135422229766846, |
|
"learning_rate": 2.896e-07, |
|
"loss": 0.2014, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"grad_norm": 4.4028706550598145, |
|
"learning_rate": 2.9459999999999995e-07, |
|
"loss": 0.2013, |
|
"step": 1475 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"grad_norm": 4.187674045562744, |
|
"learning_rate": 2.9959999999999996e-07, |
|
"loss": 0.2108, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"grad_norm": 5.227431297302246, |
|
"learning_rate": 3.0459999999999996e-07, |
|
"loss": 0.21, |
|
"step": 1525 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"grad_norm": 3.673109769821167, |
|
"learning_rate": 3.0959999999999997e-07, |
|
"loss": 0.1994, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"grad_norm": 4.424607753753662, |
|
"learning_rate": 3.1459999999999997e-07, |
|
"loss": 0.2143, |
|
"step": 1575 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"grad_norm": 4.143054485321045, |
|
"learning_rate": 3.196e-07, |
|
"loss": 0.1961, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"grad_norm": 5.010011672973633, |
|
"learning_rate": 3.246e-07, |
|
"loss": 0.1935, |
|
"step": 1625 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"grad_norm": 4.047982215881348, |
|
"learning_rate": 3.296e-07, |
|
"loss": 0.1925, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"grad_norm": 3.367611885070801, |
|
"learning_rate": 3.346e-07, |
|
"loss": 0.2028, |
|
"step": 1675 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"grad_norm": 3.4524459838867188, |
|
"learning_rate": 3.396e-07, |
|
"loss": 0.2027, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"grad_norm": 4.717390537261963, |
|
"learning_rate": 3.446e-07, |
|
"loss": 0.2, |
|
"step": 1725 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"grad_norm": 4.299264430999756, |
|
"learning_rate": 3.496e-07, |
|
"loss": 0.2117, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"grad_norm": 3.89416241645813, |
|
"learning_rate": 3.546e-07, |
|
"loss": 0.2055, |
|
"step": 1775 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"grad_norm": 4.462187767028809, |
|
"learning_rate": 3.5959999999999996e-07, |
|
"loss": 0.2061, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"grad_norm": 4.449103355407715, |
|
"learning_rate": 3.6459999999999997e-07, |
|
"loss": 0.1999, |
|
"step": 1825 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"grad_norm": 2.920868158340454, |
|
"learning_rate": 3.696e-07, |
|
"loss": 0.2063, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"grad_norm": 3.8898050785064697, |
|
"learning_rate": 3.746e-07, |
|
"loss": 0.1862, |
|
"step": 1875 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"grad_norm": 3.56378173828125, |
|
"learning_rate": 3.796e-07, |
|
"loss": 0.1974, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"grad_norm": 3.6201984882354736, |
|
"learning_rate": 3.846e-07, |
|
"loss": 0.2001, |
|
"step": 1925 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"grad_norm": 4.066067218780518, |
|
"learning_rate": 3.896e-07, |
|
"loss": 0.1954, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"grad_norm": 3.4573721885681152, |
|
"learning_rate": 3.946e-07, |
|
"loss": 0.1885, |
|
"step": 1975 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"grad_norm": 3.976249933242798, |
|
"learning_rate": 3.996e-07, |
|
"loss": 0.1928, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"eval_loss": 0.1644940823316574, |
|
"eval_runtime": 8054.0522, |
|
"eval_samples_per_second": 1.169, |
|
"eval_steps_per_second": 0.584, |
|
"eval_wer": 0.10624899079606007, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"grad_norm": 3.10294246673584, |
|
"learning_rate": 4.0439999999999994e-07, |
|
"loss": 0.1868, |
|
"step": 2025 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"grad_norm": 3.2514965534210205, |
|
"learning_rate": 4.0939999999999995e-07, |
|
"loss": 0.194, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"grad_norm": 4.614846229553223, |
|
"learning_rate": 4.1439999999999995e-07, |
|
"loss": 0.1998, |
|
"step": 2075 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"grad_norm": 3.231562852859497, |
|
"learning_rate": 4.1939999999999996e-07, |
|
"loss": 0.1841, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"grad_norm": 4.466641902923584, |
|
"learning_rate": 4.2439999999999996e-07, |
|
"loss": 0.175, |
|
"step": 2125 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"grad_norm": 3.2693140506744385, |
|
"learning_rate": 4.2939999999999997e-07, |
|
"loss": 0.1824, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"grad_norm": 3.7038068771362305, |
|
"learning_rate": 4.3439999999999997e-07, |
|
"loss": 0.1779, |
|
"step": 2175 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"grad_norm": 4.4231181144714355, |
|
"learning_rate": 4.394e-07, |
|
"loss": 0.1704, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"grad_norm": 3.4143128395080566, |
|
"learning_rate": 4.444e-07, |
|
"loss": 0.1559, |
|
"step": 2225 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"grad_norm": 3.6965668201446533, |
|
"learning_rate": 4.494e-07, |
|
"loss": 0.1708, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"grad_norm": 4.041949272155762, |
|
"learning_rate": 4.544e-07, |
|
"loss": 0.1625, |
|
"step": 2275 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"grad_norm": 3.2054636478424072, |
|
"learning_rate": 4.5939999999999994e-07, |
|
"loss": 0.1631, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"grad_norm": 4.389201641082764, |
|
"learning_rate": 4.6439999999999995e-07, |
|
"loss": 0.1833, |
|
"step": 2325 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"grad_norm": 3.396207094192505, |
|
"learning_rate": 4.6939999999999995e-07, |
|
"loss": 0.1629, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"grad_norm": 3.7798011302948, |
|
"learning_rate": 4.7439999999999996e-07, |
|
"loss": 0.158, |
|
"step": 2375 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"grad_norm": 3.4497921466827393, |
|
"learning_rate": 4.794e-07, |
|
"loss": 0.1659, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"grad_norm": 4.368722915649414, |
|
"learning_rate": 4.844e-07, |
|
"loss": 0.1624, |
|
"step": 2425 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"grad_norm": 3.625420570373535, |
|
"learning_rate": 4.894e-07, |
|
"loss": 0.1729, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"grad_norm": 4.549854278564453, |
|
"learning_rate": 4.944e-07, |
|
"loss": 0.1647, |
|
"step": 2475 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"grad_norm": 3.4779980182647705, |
|
"learning_rate": 4.994e-07, |
|
"loss": 0.1638, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"grad_norm": 6.931514739990234, |
|
"learning_rate": 5.043999999999999e-07, |
|
"loss": 0.1617, |
|
"step": 2525 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"grad_norm": 5.141421318054199, |
|
"learning_rate": 5.093999999999999e-07, |
|
"loss": 0.1682, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"grad_norm": 3.208487033843994, |
|
"learning_rate": 5.143999999999999e-07, |
|
"loss": 0.1478, |
|
"step": 2575 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"grad_norm": 4.365899562835693, |
|
"learning_rate": 5.194e-07, |
|
"loss": 0.1542, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"grad_norm": 3.7716119289398193, |
|
"learning_rate": 5.243999999999999e-07, |
|
"loss": 0.1668, |
|
"step": 2625 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"grad_norm": 4.948827743530273, |
|
"learning_rate": 5.294e-07, |
|
"loss": 0.1583, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"grad_norm": 4.176829814910889, |
|
"learning_rate": 5.343999999999999e-07, |
|
"loss": 0.1607, |
|
"step": 2675 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"grad_norm": 4.278001308441162, |
|
"learning_rate": 5.394e-07, |
|
"loss": 0.1558, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"grad_norm": 3.226017951965332, |
|
"learning_rate": 5.443999999999999e-07, |
|
"loss": 0.1495, |
|
"step": 2725 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"grad_norm": 3.3415420055389404, |
|
"learning_rate": 5.494e-07, |
|
"loss": 0.1762, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"grad_norm": 3.7578542232513428, |
|
"learning_rate": 5.543999999999999e-07, |
|
"loss": 0.1684, |
|
"step": 2775 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"grad_norm": 4.540279388427734, |
|
"learning_rate": 5.594e-07, |
|
"loss": 0.1629, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"grad_norm": 2.953626871109009, |
|
"learning_rate": 5.643999999999999e-07, |
|
"loss": 0.1539, |
|
"step": 2825 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"grad_norm": 4.061501502990723, |
|
"learning_rate": 5.694e-07, |
|
"loss": 0.1428, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"grad_norm": 3.305874824523926, |
|
"learning_rate": 5.744e-07, |
|
"loss": 0.1713, |
|
"step": 2875 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"grad_norm": 3.4820492267608643, |
|
"learning_rate": 5.794e-07, |
|
"loss": 0.1712, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"grad_norm": 4.517523765563965, |
|
"learning_rate": 5.844e-07, |
|
"loss": 0.1626, |
|
"step": 2925 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"grad_norm": 5.121035099029541, |
|
"learning_rate": 5.894e-07, |
|
"loss": 0.1591, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"grad_norm": 3.2896835803985596, |
|
"learning_rate": 5.944e-07, |
|
"loss": 0.1652, |
|
"step": 2975 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"grad_norm": 3.9433748722076416, |
|
"learning_rate": 5.994e-07, |
|
"loss": 0.1646, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"eval_loss": 0.15075725317001343, |
|
"eval_runtime": 7921.0228, |
|
"eval_samples_per_second": 1.188, |
|
"eval_steps_per_second": 0.594, |
|
"eval_wer": 0.10159857904085258, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"grad_norm": 6.672876834869385, |
|
"learning_rate": 6.044e-07, |
|
"loss": 0.1694, |
|
"step": 3025 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"grad_norm": 3.1442010402679443, |
|
"learning_rate": 6.094e-07, |
|
"loss": 0.1655, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"grad_norm": 4.091799736022949, |
|
"learning_rate": 6.143999999999999e-07, |
|
"loss": 0.1568, |
|
"step": 3075 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"grad_norm": 3.820805311203003, |
|
"learning_rate": 6.194e-07, |
|
"loss": 0.1635, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"grad_norm": 3.704472303390503, |
|
"learning_rate": 6.242e-07, |
|
"loss": 0.1587, |
|
"step": 3125 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"grad_norm": 3.817859649658203, |
|
"learning_rate": 6.291999999999999e-07, |
|
"loss": 0.1443, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"grad_norm": 4.0161452293396, |
|
"learning_rate": 6.342e-07, |
|
"loss": 0.147, |
|
"step": 3175 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"grad_norm": 5.137659072875977, |
|
"learning_rate": 6.392e-07, |
|
"loss": 0.1574, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 3.01, |
|
"grad_norm": 3.7066938877105713, |
|
"learning_rate": 6.442e-07, |
|
"loss": 0.142, |
|
"step": 3225 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"grad_norm": 4.4287919998168945, |
|
"learning_rate": 6.492e-07, |
|
"loss": 0.1367, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 3.06, |
|
"grad_norm": 4.2995381355285645, |
|
"learning_rate": 6.542e-07, |
|
"loss": 0.1226, |
|
"step": 3275 |
|
}, |
|
{ |
|
"epoch": 3.08, |
|
"grad_norm": 3.6767024993896484, |
|
"learning_rate": 6.592e-07, |
|
"loss": 0.1533, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 3.1, |
|
"grad_norm": 4.893792152404785, |
|
"learning_rate": 6.642e-07, |
|
"loss": 0.1304, |
|
"step": 3325 |
|
}, |
|
{ |
|
"epoch": 3.13, |
|
"grad_norm": 1.7213168144226074, |
|
"learning_rate": 6.692e-07, |
|
"loss": 0.143, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 3.15, |
|
"grad_norm": 4.113219738006592, |
|
"learning_rate": 6.742e-07, |
|
"loss": 0.1311, |
|
"step": 3375 |
|
}, |
|
{ |
|
"epoch": 3.17, |
|
"grad_norm": 4.480813980102539, |
|
"learning_rate": 6.792e-07, |
|
"loss": 0.1429, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 3.2, |
|
"grad_norm": 4.102989196777344, |
|
"learning_rate": 6.842e-07, |
|
"loss": 0.1361, |
|
"step": 3425 |
|
}, |
|
{ |
|
"epoch": 3.22, |
|
"grad_norm": 4.409080982208252, |
|
"learning_rate": 6.892e-07, |
|
"loss": 0.1271, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 3.25, |
|
"grad_norm": 3.579719305038452, |
|
"learning_rate": 6.942e-07, |
|
"loss": 0.1303, |
|
"step": 3475 |
|
}, |
|
{ |
|
"epoch": 3.27, |
|
"grad_norm": 5.013184547424316, |
|
"learning_rate": 6.992e-07, |
|
"loss": 0.1359, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 3.29, |
|
"grad_norm": 4.512373447418213, |
|
"learning_rate": 7.042e-07, |
|
"loss": 0.1359, |
|
"step": 3525 |
|
}, |
|
{ |
|
"epoch": 3.32, |
|
"grad_norm": 4.140831470489502, |
|
"learning_rate": 7.092e-07, |
|
"loss": 0.1239, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 3.34, |
|
"grad_norm": 3.4033215045928955, |
|
"learning_rate": 7.141999999999999e-07, |
|
"loss": 0.133, |
|
"step": 3575 |
|
}, |
|
{ |
|
"epoch": 3.36, |
|
"grad_norm": 3.3019678592681885, |
|
"learning_rate": 7.191999999999999e-07, |
|
"loss": 0.1302, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 3.39, |
|
"grad_norm": 3.88134765625, |
|
"learning_rate": 7.241999999999999e-07, |
|
"loss": 0.1437, |
|
"step": 3625 |
|
}, |
|
{ |
|
"epoch": 3.41, |
|
"grad_norm": 4.493922710418701, |
|
"learning_rate": 7.291999999999999e-07, |
|
"loss": 0.1346, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 3.43, |
|
"grad_norm": 4.8288421630859375, |
|
"learning_rate": 7.341999999999999e-07, |
|
"loss": 0.1331, |
|
"step": 3675 |
|
}, |
|
{ |
|
"epoch": 3.46, |
|
"grad_norm": 4.28986120223999, |
|
"learning_rate": 7.392e-07, |
|
"loss": 0.1309, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 3.48, |
|
"grad_norm": 3.411266326904297, |
|
"learning_rate": 7.441999999999999e-07, |
|
"loss": 0.1549, |
|
"step": 3725 |
|
}, |
|
{ |
|
"epoch": 3.5, |
|
"grad_norm": 3.317307949066162, |
|
"learning_rate": 7.492e-07, |
|
"loss": 0.1265, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 3.53, |
|
"grad_norm": 4.695957183837891, |
|
"learning_rate": 7.541999999999999e-07, |
|
"loss": 0.1422, |
|
"step": 3775 |
|
}, |
|
{ |
|
"epoch": 3.55, |
|
"grad_norm": 3.6588873863220215, |
|
"learning_rate": 7.592e-07, |
|
"loss": 0.1437, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 3.57, |
|
"grad_norm": 5.599457740783691, |
|
"learning_rate": 7.641999999999999e-07, |
|
"loss": 0.1436, |
|
"step": 3825 |
|
}, |
|
{ |
|
"epoch": 3.6, |
|
"grad_norm": 4.065281391143799, |
|
"learning_rate": 7.692e-07, |
|
"loss": 0.1281, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 3.62, |
|
"grad_norm": 4.9737935066223145, |
|
"learning_rate": 7.741999999999999e-07, |
|
"loss": 0.1293, |
|
"step": 3875 |
|
}, |
|
{ |
|
"epoch": 3.64, |
|
"grad_norm": 3.3383023738861084, |
|
"learning_rate": 7.792e-07, |
|
"loss": 0.1329, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 3.67, |
|
"grad_norm": 3.6315879821777344, |
|
"learning_rate": 7.841999999999999e-07, |
|
"loss": 0.1273, |
|
"step": 3925 |
|
}, |
|
{ |
|
"epoch": 3.69, |
|
"grad_norm": 4.887773036956787, |
|
"learning_rate": 7.892e-07, |
|
"loss": 0.136, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 3.71, |
|
"grad_norm": 4.8274455070495605, |
|
"learning_rate": 7.942e-07, |
|
"loss": 0.1438, |
|
"step": 3975 |
|
}, |
|
{ |
|
"epoch": 3.74, |
|
"grad_norm": 4.641810417175293, |
|
"learning_rate": 7.992e-07, |
|
"loss": 0.1333, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 3.74, |
|
"eval_loss": 0.14375422894954681, |
|
"eval_runtime": 7811.8504, |
|
"eval_samples_per_second": 1.205, |
|
"eval_steps_per_second": 0.603, |
|
"eval_wer": 0.09701275633780074, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 3.76, |
|
"grad_norm": 5.432655334472656, |
|
"learning_rate": 8.042e-07, |
|
"loss": 0.1238, |
|
"step": 4025 |
|
}, |
|
{ |
|
"epoch": 3.78, |
|
"grad_norm": 4.146833419799805, |
|
"learning_rate": 8.092e-07, |
|
"loss": 0.1234, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 3.81, |
|
"grad_norm": 4.496214389801025, |
|
"learning_rate": 8.142e-07, |
|
"loss": 0.1385, |
|
"step": 4075 |
|
}, |
|
{ |
|
"epoch": 3.83, |
|
"grad_norm": 4.550703525543213, |
|
"learning_rate": 8.192e-07, |
|
"loss": 0.1327, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 3.85, |
|
"grad_norm": 4.395373821258545, |
|
"learning_rate": 8.242e-07, |
|
"loss": 0.1336, |
|
"step": 4125 |
|
}, |
|
{ |
|
"epoch": 3.88, |
|
"grad_norm": 3.292907238006592, |
|
"learning_rate": 8.292e-07, |
|
"loss": 0.1243, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 3.9, |
|
"grad_norm": 3.963186502456665, |
|
"learning_rate": 8.342e-07, |
|
"loss": 0.1287, |
|
"step": 4175 |
|
}, |
|
{ |
|
"epoch": 3.92, |
|
"grad_norm": 3.482177972793579, |
|
"learning_rate": 8.391999999999999e-07, |
|
"loss": 0.1305, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 3.95, |
|
"grad_norm": 3.331468105316162, |
|
"learning_rate": 8.441999999999999e-07, |
|
"loss": 0.1308, |
|
"step": 4225 |
|
}, |
|
{ |
|
"epoch": 3.97, |
|
"grad_norm": 5.990370750427246, |
|
"learning_rate": 8.492e-07, |
|
"loss": 0.1411, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 3.99, |
|
"grad_norm": 3.512251615524292, |
|
"learning_rate": 8.541999999999999e-07, |
|
"loss": 0.132, |
|
"step": 4275 |
|
}, |
|
{ |
|
"epoch": 4.02, |
|
"grad_norm": 3.779296398162842, |
|
"learning_rate": 8.592e-07, |
|
"loss": 0.1277, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 4.04, |
|
"grad_norm": 4.156597137451172, |
|
"learning_rate": 8.641999999999999e-07, |
|
"loss": 0.1082, |
|
"step": 4325 |
|
}, |
|
{ |
|
"epoch": 4.06, |
|
"grad_norm": 3.5820066928863525, |
|
"learning_rate": 8.692e-07, |
|
"loss": 0.117, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 4.09, |
|
"grad_norm": 3.3737404346466064, |
|
"learning_rate": 8.741999999999999e-07, |
|
"loss": 0.1068, |
|
"step": 4375 |
|
}, |
|
{ |
|
"epoch": 4.11, |
|
"grad_norm": 4.2788166999816895, |
|
"learning_rate": 8.792e-07, |
|
"loss": 0.1129, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 4.13, |
|
"grad_norm": 3.872753381729126, |
|
"learning_rate": 8.841999999999999e-07, |
|
"loss": 0.1124, |
|
"step": 4425 |
|
}, |
|
{ |
|
"epoch": 4.16, |
|
"grad_norm": 5.087253570556641, |
|
"learning_rate": 8.892e-07, |
|
"loss": 0.1149, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 4.18, |
|
"grad_norm": 4.363770484924316, |
|
"learning_rate": 8.941999999999999e-07, |
|
"loss": 0.1105, |
|
"step": 4475 |
|
}, |
|
{ |
|
"epoch": 4.2, |
|
"grad_norm": 3.211000680923462, |
|
"learning_rate": 8.992e-07, |
|
"loss": 0.1162, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 4.23, |
|
"grad_norm": 4.609809875488281, |
|
"learning_rate": 9.042e-07, |
|
"loss": 0.1165, |
|
"step": 4525 |
|
}, |
|
{ |
|
"epoch": 4.25, |
|
"grad_norm": 3.679553508758545, |
|
"learning_rate": 9.092e-07, |
|
"loss": 0.111, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 4.27, |
|
"grad_norm": 3.6184656620025635, |
|
"learning_rate": 9.142e-07, |
|
"loss": 0.1243, |
|
"step": 4575 |
|
}, |
|
{ |
|
"epoch": 4.3, |
|
"grad_norm": 4.844442367553711, |
|
"learning_rate": 9.192e-07, |
|
"loss": 0.1112, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 4.32, |
|
"grad_norm": 5.363363265991211, |
|
"learning_rate": 9.242e-07, |
|
"loss": 0.1193, |
|
"step": 4625 |
|
}, |
|
{ |
|
"epoch": 4.34, |
|
"grad_norm": 2.0149080753326416, |
|
"learning_rate": 9.292e-07, |
|
"loss": 0.1143, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 4.37, |
|
"grad_norm": 5.45695161819458, |
|
"learning_rate": 9.342e-07, |
|
"loss": 0.1056, |
|
"step": 4675 |
|
}, |
|
{ |
|
"epoch": 4.39, |
|
"grad_norm": 5.325240612030029, |
|
"learning_rate": 9.391999999999999e-07, |
|
"loss": 0.114, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 4.41, |
|
"grad_norm": 3.116534471511841, |
|
"learning_rate": 9.442e-07, |
|
"loss": 0.0995, |
|
"step": 4725 |
|
}, |
|
{ |
|
"epoch": 4.44, |
|
"grad_norm": 2.405822515487671, |
|
"learning_rate": 9.492e-07, |
|
"loss": 0.1165, |
|
"step": 4750 |
|
}, |
|
{ |
|
"epoch": 4.46, |
|
"grad_norm": 4.517841339111328, |
|
"learning_rate": 9.542e-07, |
|
"loss": 0.1052, |
|
"step": 4775 |
|
}, |
|
{ |
|
"epoch": 4.48, |
|
"grad_norm": 2.7329490184783936, |
|
"learning_rate": 9.592e-07, |
|
"loss": 0.102, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 4.51, |
|
"grad_norm": 3.871314287185669, |
|
"learning_rate": 9.641999999999998e-07, |
|
"loss": 0.1081, |
|
"step": 4825 |
|
}, |
|
{ |
|
"epoch": 4.53, |
|
"grad_norm": 3.3892979621887207, |
|
"learning_rate": 9.692e-07, |
|
"loss": 0.1138, |
|
"step": 4850 |
|
}, |
|
{ |
|
"epoch": 4.55, |
|
"grad_norm": 4.42065954208374, |
|
"learning_rate": 9.742e-07, |
|
"loss": 0.106, |
|
"step": 4875 |
|
}, |
|
{ |
|
"epoch": 4.58, |
|
"grad_norm": 4.373811721801758, |
|
"learning_rate": 9.791999999999999e-07, |
|
"loss": 0.1043, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 4.6, |
|
"grad_norm": 5.577483177185059, |
|
"learning_rate": 9.841999999999998e-07, |
|
"loss": 0.0962, |
|
"step": 4925 |
|
}, |
|
{ |
|
"epoch": 4.62, |
|
"grad_norm": 3.65592360496521, |
|
"learning_rate": 9.892e-07, |
|
"loss": 0.1161, |
|
"step": 4950 |
|
}, |
|
{ |
|
"epoch": 4.65, |
|
"grad_norm": 4.68386173248291, |
|
"learning_rate": 9.942e-07, |
|
"loss": 0.1071, |
|
"step": 4975 |
|
}, |
|
{ |
|
"epoch": 4.67, |
|
"grad_norm": 3.9556305408477783, |
|
"learning_rate": 9.992e-07, |
|
"loss": 0.1027, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 4.67, |
|
"eval_loss": 0.14086556434631348, |
|
"eval_runtime": 7847.8302, |
|
"eval_samples_per_second": 1.2, |
|
"eval_steps_per_second": 0.6, |
|
"eval_wer": 0.09421927983206846, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 4.67, |
|
"step": 5000, |
|
"total_flos": 1.632967868976988e+20, |
|
"train_loss": 0.2694411336898804, |
|
"train_runtime": 106369.5753, |
|
"train_samples_per_second": 1.504, |
|
"train_steps_per_second": 0.047 |
|
} |
|
], |
|
"logging_steps": 25, |
|
"max_steps": 5000, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 5, |
|
"save_steps": 1000, |
|
"total_flos": 1.632967868976988e+20, |
|
"train_batch_size": 1, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|