|
{ |
|
"best_metric": 24.71689240755159, |
|
"best_model_checkpoint": "output_dir_run_2/checkpoint-42000", |
|
"epoch": 1.834423353006859, |
|
"global_step": 46000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 4.4e-07, |
|
"loss": 3.8256, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 9.200000000000001e-07, |
|
"loss": 3.5128, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 1.42e-06, |
|
"loss": 2.8953, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 1.9200000000000003e-06, |
|
"loss": 2.3779, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 2.42e-06, |
|
"loss": 2.065, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 2.92e-06, |
|
"loss": 1.8663, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 3.4200000000000007e-06, |
|
"loss": 1.6813, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 3.920000000000001e-06, |
|
"loss": 1.6136, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 4.42e-06, |
|
"loss": 1.5066, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 4.92e-06, |
|
"loss": 1.3971, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 5.420000000000001e-06, |
|
"loss": 1.3437, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 5.92e-06, |
|
"loss": 1.2227, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 6.42e-06, |
|
"loss": 1.2204, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 6.92e-06, |
|
"loss": 1.1546, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 7.420000000000001e-06, |
|
"loss": 1.1521, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 7.92e-06, |
|
"loss": 1.0848, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 8.42e-06, |
|
"loss": 1.0626, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 8.920000000000001e-06, |
|
"loss": 1.0112, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.42e-06, |
|
"loss": 0.988, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.920000000000002e-06, |
|
"loss": 0.9656, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.99957957957958e-06, |
|
"loss": 0.9418, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.99907907907908e-06, |
|
"loss": 0.9225, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.99857857857858e-06, |
|
"loss": 0.9479, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.998078078078079e-06, |
|
"loss": 0.8927, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 9.997577577577578e-06, |
|
"loss": 0.8688, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.997077077077078e-06, |
|
"loss": 0.8846, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.996576576576577e-06, |
|
"loss": 0.8514, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.996076076076076e-06, |
|
"loss": 0.8254, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.995575575575577e-06, |
|
"loss": 0.8167, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.995075075075076e-06, |
|
"loss": 0.7695, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.994574574574576e-06, |
|
"loss": 0.8109, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.994074074074075e-06, |
|
"loss": 0.7718, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.993573573573574e-06, |
|
"loss": 0.7937, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.993073073073074e-06, |
|
"loss": 0.7566, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 9.992572572572573e-06, |
|
"loss": 0.7627, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.992072072072074e-06, |
|
"loss": 0.7564, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.991571571571572e-06, |
|
"loss": 0.766, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.991071071071073e-06, |
|
"loss": 0.7255, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.990570570570572e-06, |
|
"loss": 0.7385, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.990070070070071e-06, |
|
"loss": 0.7455, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"eval_loss": 0.9031199216842651, |
|
"eval_runtime": 2867.5101, |
|
"eval_samples_per_second": 5.654, |
|
"eval_steps_per_second": 0.142, |
|
"eval_wer": 58.282005226601704, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.98956956956957e-06, |
|
"loss": 0.7141, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.98906906906907e-06, |
|
"loss": 0.7036, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.988568568568569e-06, |
|
"loss": 0.7157, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.988068068068068e-06, |
|
"loss": 0.7362, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 9.98756756756757e-06, |
|
"loss": 0.7134, |
|
"step": 1125 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.987067067067067e-06, |
|
"loss": 0.6704, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.986566566566568e-06, |
|
"loss": 0.6907, |
|
"step": 1175 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.986066066066067e-06, |
|
"loss": 0.6851, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.985565565565567e-06, |
|
"loss": 0.7214, |
|
"step": 1225 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.985065065065066e-06, |
|
"loss": 0.6876, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.984564564564565e-06, |
|
"loss": 0.6627, |
|
"step": 1275 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.984064064064064e-06, |
|
"loss": 0.6739, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.983563563563564e-06, |
|
"loss": 0.6629, |
|
"step": 1325 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.983063063063065e-06, |
|
"loss": 0.6493, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9.982562562562564e-06, |
|
"loss": 0.6512, |
|
"step": 1375 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.982062062062063e-06, |
|
"loss": 0.6481, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.981561561561563e-06, |
|
"loss": 0.6459, |
|
"step": 1425 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.981061061061062e-06, |
|
"loss": 0.643, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.980560560560561e-06, |
|
"loss": 0.6344, |
|
"step": 1475 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.98006006006006e-06, |
|
"loss": 0.6312, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.979559559559561e-06, |
|
"loss": 0.6587, |
|
"step": 1525 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.979059059059059e-06, |
|
"loss": 0.6156, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.97855855855856e-06, |
|
"loss": 0.64, |
|
"step": 1575 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.97805805805806e-06, |
|
"loss": 0.5947, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.977557557557559e-06, |
|
"loss": 0.6496, |
|
"step": 1625 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.977057057057058e-06, |
|
"loss": 0.6434, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.976556556556557e-06, |
|
"loss": 0.5917, |
|
"step": 1675 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.976056056056057e-06, |
|
"loss": 0.6057, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.975555555555556e-06, |
|
"loss": 0.6132, |
|
"step": 1725 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.975055055055057e-06, |
|
"loss": 0.5929, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.974554554554554e-06, |
|
"loss": 0.6166, |
|
"step": 1775 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.974054054054055e-06, |
|
"loss": 0.5866, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.973553553553555e-06, |
|
"loss": 0.6123, |
|
"step": 1825 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.973053053053054e-06, |
|
"loss": 0.594, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.972552552552553e-06, |
|
"loss": 0.5949, |
|
"step": 1875 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.972052052052053e-06, |
|
"loss": 0.5862, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.971551551551552e-06, |
|
"loss": 0.609, |
|
"step": 1925 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.971051051051051e-06, |
|
"loss": 0.5688, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.970550550550552e-06, |
|
"loss": 0.592, |
|
"step": 1975 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.970050050050052e-06, |
|
"loss": 0.5906, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"eval_loss": 0.7455159425735474, |
|
"eval_runtime": 3054.18, |
|
"eval_samples_per_second": 5.308, |
|
"eval_steps_per_second": 0.133, |
|
"eval_wer": 50.6001341340261, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.96954954954955e-06, |
|
"loss": 0.5755, |
|
"step": 2025 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.96904904904905e-06, |
|
"loss": 0.5761, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.96854854854855e-06, |
|
"loss": 0.5832, |
|
"step": 2075 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.968048048048049e-06, |
|
"loss": 0.5661, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.967547547547548e-06, |
|
"loss": 0.5906, |
|
"step": 2125 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.967047047047049e-06, |
|
"loss": 0.5769, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.966546546546547e-06, |
|
"loss": 0.5491, |
|
"step": 2175 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.966046046046048e-06, |
|
"loss": 0.5794, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.965545545545547e-06, |
|
"loss": 0.5319, |
|
"step": 2225 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.965045045045044e-06, |
|
"loss": 0.5656, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.964544544544545e-06, |
|
"loss": 0.5714, |
|
"step": 2275 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.964044044044045e-06, |
|
"loss": 0.5667, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.963543543543544e-06, |
|
"loss": 0.5628, |
|
"step": 2325 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.963043043043043e-06, |
|
"loss": 0.5816, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.962542542542544e-06, |
|
"loss": 0.5713, |
|
"step": 2375 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.962042042042042e-06, |
|
"loss": 0.5513, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.961541541541543e-06, |
|
"loss": 0.5416, |
|
"step": 2425 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.961041041041042e-06, |
|
"loss": 0.5273, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.960540540540542e-06, |
|
"loss": 0.5414, |
|
"step": 2475 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.960040040040041e-06, |
|
"loss": 0.5454, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.95953953953954e-06, |
|
"loss": 0.5335, |
|
"step": 2525 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.95903903903904e-06, |
|
"loss": 0.538, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.958538538538539e-06, |
|
"loss": 0.5275, |
|
"step": 2575 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.95803803803804e-06, |
|
"loss": 0.541, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.957537537537539e-06, |
|
"loss": 0.5653, |
|
"step": 2625 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.957037037037038e-06, |
|
"loss": 0.5407, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.956536536536538e-06, |
|
"loss": 0.5508, |
|
"step": 2675 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.956036036036037e-06, |
|
"loss": 0.507, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.955535535535536e-06, |
|
"loss": 0.5289, |
|
"step": 2725 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.955035035035036e-06, |
|
"loss": 0.543, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.954534534534537e-06, |
|
"loss": 0.5489, |
|
"step": 2775 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.954034034034034e-06, |
|
"loss": 0.5433, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.953533533533535e-06, |
|
"loss": 0.5372, |
|
"step": 2825 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.953033033033034e-06, |
|
"loss": 0.5397, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.952532532532532e-06, |
|
"loss": 0.5429, |
|
"step": 2875 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.952032032032033e-06, |
|
"loss": 0.5178, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.951531531531532e-06, |
|
"loss": 0.5426, |
|
"step": 2925 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.951031031031032e-06, |
|
"loss": 0.5301, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.950530530530531e-06, |
|
"loss": 0.5113, |
|
"step": 2975 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.950030030030032e-06, |
|
"loss": 0.5015, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"eval_loss": 0.6700246334075928, |
|
"eval_runtime": 2829.2547, |
|
"eval_samples_per_second": 5.73, |
|
"eval_steps_per_second": 0.144, |
|
"eval_wer": 45.80060283223225, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.94952952952953e-06, |
|
"loss": 0.5075, |
|
"step": 3025 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.94902902902903e-06, |
|
"loss": 0.548, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.94852852852853e-06, |
|
"loss": 0.5221, |
|
"step": 3075 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.948028028028029e-06, |
|
"loss": 0.5108, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.947527527527528e-06, |
|
"loss": 0.4792, |
|
"step": 3125 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.947027027027028e-06, |
|
"loss": 0.4993, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.946526526526527e-06, |
|
"loss": 0.5159, |
|
"step": 3175 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.946026026026026e-06, |
|
"loss": 0.4948, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.945525525525527e-06, |
|
"loss": 0.527, |
|
"step": 3225 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.945025025025027e-06, |
|
"loss": 0.4831, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.944524524524526e-06, |
|
"loss": 0.5106, |
|
"step": 3275 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.944024024024025e-06, |
|
"loss": 0.5323, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.943523523523524e-06, |
|
"loss": 0.502, |
|
"step": 3325 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.943023023023024e-06, |
|
"loss": 0.495, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.942522522522523e-06, |
|
"loss": 0.519, |
|
"step": 3375 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.942022022022022e-06, |
|
"loss": 0.5351, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.941521521521522e-06, |
|
"loss": 0.4833, |
|
"step": 3425 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.941021021021023e-06, |
|
"loss": 0.5227, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.940520520520522e-06, |
|
"loss": 0.4969, |
|
"step": 3475 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.94002002002002e-06, |
|
"loss": 0.5103, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.93951951951952e-06, |
|
"loss": 0.5109, |
|
"step": 3525 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.93901901901902e-06, |
|
"loss": 0.5228, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.938518518518519e-06, |
|
"loss": 0.4983, |
|
"step": 3575 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.938018018018018e-06, |
|
"loss": 0.4873, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.93751751751752e-06, |
|
"loss": 0.4955, |
|
"step": 3625 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.937017017017017e-06, |
|
"loss": 0.5064, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.936516516516518e-06, |
|
"loss": 0.506, |
|
"step": 3675 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.936016016016017e-06, |
|
"loss": 0.483, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.935515515515517e-06, |
|
"loss": 0.4931, |
|
"step": 3725 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.935015015015016e-06, |
|
"loss": 0.4854, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.934514514514515e-06, |
|
"loss": 0.4697, |
|
"step": 3775 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.934014014014014e-06, |
|
"loss": 0.5078, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.933513513513514e-06, |
|
"loss": 0.4759, |
|
"step": 3825 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.933013013013015e-06, |
|
"loss": 0.5202, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.932512512512514e-06, |
|
"loss": 0.4633, |
|
"step": 3875 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.932012012012013e-06, |
|
"loss": 0.4947, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.931511511511513e-06, |
|
"loss": 0.4952, |
|
"step": 3925 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.931011011011012e-06, |
|
"loss": 0.4768, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.930510510510511e-06, |
|
"loss": 0.485, |
|
"step": 3975 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.93001001001001e-06, |
|
"loss": 0.4855, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"eval_loss": 0.6402216553688049, |
|
"eval_runtime": 2763.3301, |
|
"eval_samples_per_second": 5.867, |
|
"eval_steps_per_second": 0.147, |
|
"eval_wer": 41.90377810840188, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.92950950950951e-06, |
|
"loss": 0.4653, |
|
"step": 4025 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.929009009009009e-06, |
|
"loss": 0.4994, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.92850850850851e-06, |
|
"loss": 0.4641, |
|
"step": 4075 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.92800800800801e-06, |
|
"loss": 0.4851, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.927507507507507e-06, |
|
"loss": 0.489, |
|
"step": 4125 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.927007007007008e-06, |
|
"loss": 0.4881, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.926506506506507e-06, |
|
"loss": 0.4975, |
|
"step": 4175 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.926006006006007e-06, |
|
"loss": 0.4785, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.925505505505506e-06, |
|
"loss": 0.4689, |
|
"step": 4225 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.925005005005007e-06, |
|
"loss": 0.4688, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.924504504504505e-06, |
|
"loss": 0.4842, |
|
"step": 4275 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.924004004004006e-06, |
|
"loss": 0.462, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.923503503503505e-06, |
|
"loss": 0.4674, |
|
"step": 4325 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.923003003003004e-06, |
|
"loss": 0.4632, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.922502502502503e-06, |
|
"loss": 0.4618, |
|
"step": 4375 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.922002002002003e-06, |
|
"loss": 0.4655, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.921501501501502e-06, |
|
"loss": 0.4557, |
|
"step": 4425 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.921001001001001e-06, |
|
"loss": 0.4549, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.920500500500502e-06, |
|
"loss": 0.4757, |
|
"step": 4475 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.920000000000002e-06, |
|
"loss": 0.4501, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.919499499499501e-06, |
|
"loss": 0.4789, |
|
"step": 4525 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.918998998999e-06, |
|
"loss": 0.4768, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.9184984984985e-06, |
|
"loss": 0.4799, |
|
"step": 4575 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.917997997997999e-06, |
|
"loss": 0.4648, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.917497497497498e-06, |
|
"loss": 0.4671, |
|
"step": 4625 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.916996996996997e-06, |
|
"loss": 0.4672, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.916496496496497e-06, |
|
"loss": 0.4664, |
|
"step": 4675 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.915995995995998e-06, |
|
"loss": 0.4769, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.915495495495497e-06, |
|
"loss": 0.4828, |
|
"step": 4725 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.914994994994995e-06, |
|
"loss": 0.4648, |
|
"step": 4750 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.914494494494496e-06, |
|
"loss": 0.4501, |
|
"step": 4775 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.913993993993995e-06, |
|
"loss": 0.4638, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.913493493493494e-06, |
|
"loss": 0.463, |
|
"step": 4825 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.912992992992993e-06, |
|
"loss": 0.4759, |
|
"step": 4850 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.912492492492494e-06, |
|
"loss": 0.4967, |
|
"step": 4875 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.911991991991992e-06, |
|
"loss": 0.4483, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.911491491491493e-06, |
|
"loss": 0.4514, |
|
"step": 4925 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.910990990990992e-06, |
|
"loss": 0.482, |
|
"step": 4950 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.910490490490492e-06, |
|
"loss": 0.4439, |
|
"step": 4975 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.909989989989991e-06, |
|
"loss": 0.4379, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"eval_loss": 0.6058462858200073, |
|
"eval_runtime": 2619.0085, |
|
"eval_samples_per_second": 6.191, |
|
"eval_steps_per_second": 0.155, |
|
"eval_wer": 38.43171113389505, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.90948948948949e-06, |
|
"loss": 0.4567, |
|
"step": 5025 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.90898898898899e-06, |
|
"loss": 0.4506, |
|
"step": 5050 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.908488488488489e-06, |
|
"loss": 0.4717, |
|
"step": 5075 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.90798798798799e-06, |
|
"loss": 0.4775, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.907487487487487e-06, |
|
"loss": 0.4344, |
|
"step": 5125 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.906986986986987e-06, |
|
"loss": 0.4462, |
|
"step": 5150 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.906486486486488e-06, |
|
"loss": 0.4492, |
|
"step": 5175 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.905985985985987e-06, |
|
"loss": 0.4532, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.905485485485486e-06, |
|
"loss": 0.4595, |
|
"step": 5225 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.904984984984986e-06, |
|
"loss": 0.4369, |
|
"step": 5250 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.904484484484485e-06, |
|
"loss": 0.4668, |
|
"step": 5275 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.903983983983984e-06, |
|
"loss": 0.4507, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.903483483483485e-06, |
|
"loss": 0.4438, |
|
"step": 5325 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.902982982982984e-06, |
|
"loss": 0.4679, |
|
"step": 5350 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.902482482482482e-06, |
|
"loss": 0.468, |
|
"step": 5375 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.901981981981983e-06, |
|
"loss": 0.4502, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.901481481481482e-06, |
|
"loss": 0.4319, |
|
"step": 5425 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.900980980980982e-06, |
|
"loss": 0.4799, |
|
"step": 5450 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.900480480480481e-06, |
|
"loss": 0.4553, |
|
"step": 5475 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.899979979979982e-06, |
|
"loss": 0.4309, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.89947947947948e-06, |
|
"loss": 0.4385, |
|
"step": 5525 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.89897897897898e-06, |
|
"loss": 0.431, |
|
"step": 5550 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.89847847847848e-06, |
|
"loss": 0.4478, |
|
"step": 5575 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.897977977977979e-06, |
|
"loss": 0.4379, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.897477477477478e-06, |
|
"loss": 0.4334, |
|
"step": 5625 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.896976976976978e-06, |
|
"loss": 0.4542, |
|
"step": 5650 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.896476476476477e-06, |
|
"loss": 0.4454, |
|
"step": 5675 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.895975975975976e-06, |
|
"loss": 0.4519, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.895475475475477e-06, |
|
"loss": 0.4625, |
|
"step": 5725 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.894974974974975e-06, |
|
"loss": 0.4361, |
|
"step": 5750 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.894474474474474e-06, |
|
"loss": 0.4305, |
|
"step": 5775 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.893973973973975e-06, |
|
"loss": 0.4558, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.893473473473475e-06, |
|
"loss": 0.4554, |
|
"step": 5825 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.892972972972974e-06, |
|
"loss": 0.4562, |
|
"step": 5850 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.892472472472473e-06, |
|
"loss": 0.4378, |
|
"step": 5875 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.891971971971972e-06, |
|
"loss": 0.464, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.891471471471472e-06, |
|
"loss": 0.4279, |
|
"step": 5925 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.890970970970973e-06, |
|
"loss": 0.4414, |
|
"step": 5950 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.890470470470472e-06, |
|
"loss": 0.4537, |
|
"step": 5975 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.88996996996997e-06, |
|
"loss": 0.4399, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"eval_loss": 0.5941766500473022, |
|
"eval_runtime": 2644.5549, |
|
"eval_samples_per_second": 6.131, |
|
"eval_steps_per_second": 0.154, |
|
"eval_wer": 37.588362716907824, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.88946946946947e-06, |
|
"loss": 0.4419, |
|
"step": 6025 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.88896896896897e-06, |
|
"loss": 0.4219, |
|
"step": 6050 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.88846846846847e-06, |
|
"loss": 0.4271, |
|
"step": 6075 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.887967967967968e-06, |
|
"loss": 0.4565, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.88746746746747e-06, |
|
"loss": 0.4588, |
|
"step": 6125 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.886966966966967e-06, |
|
"loss": 0.4384, |
|
"step": 6150 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.886466466466468e-06, |
|
"loss": 0.4341, |
|
"step": 6175 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.885965965965967e-06, |
|
"loss": 0.4302, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.885465465465467e-06, |
|
"loss": 0.4307, |
|
"step": 6225 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.884964964964966e-06, |
|
"loss": 0.4634, |
|
"step": 6250 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.884464464464465e-06, |
|
"loss": 0.4374, |
|
"step": 6275 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.883963963963965e-06, |
|
"loss": 0.4352, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.883463463463464e-06, |
|
"loss": 0.4376, |
|
"step": 6325 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.882962962962965e-06, |
|
"loss": 0.4284, |
|
"step": 6350 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.882462462462462e-06, |
|
"loss": 0.4143, |
|
"step": 6375 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.881961961961962e-06, |
|
"loss": 0.4286, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.881461461461463e-06, |
|
"loss": 0.4358, |
|
"step": 6425 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.880960960960962e-06, |
|
"loss": 0.4306, |
|
"step": 6450 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.880460460460461e-06, |
|
"loss": 0.4357, |
|
"step": 6475 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.87995995995996e-06, |
|
"loss": 0.4344, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.87945945945946e-06, |
|
"loss": 0.4508, |
|
"step": 6525 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.87895895895896e-06, |
|
"loss": 0.441, |
|
"step": 6550 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.87845845845846e-06, |
|
"loss": 0.419, |
|
"step": 6575 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.87795795795796e-06, |
|
"loss": 0.4221, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.877457457457457e-06, |
|
"loss": 0.4288, |
|
"step": 6625 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.876956956956958e-06, |
|
"loss": 0.4288, |
|
"step": 6650 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.876456456456457e-06, |
|
"loss": 0.4314, |
|
"step": 6675 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.875955955955957e-06, |
|
"loss": 0.4138, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.875455455455456e-06, |
|
"loss": 0.4291, |
|
"step": 6725 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.874954954954957e-06, |
|
"loss": 0.4446, |
|
"step": 6750 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.874454454454455e-06, |
|
"loss": 0.4153, |
|
"step": 6775 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.873953953953956e-06, |
|
"loss": 0.4325, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.873453453453455e-06, |
|
"loss": 0.4158, |
|
"step": 6825 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.872952952952952e-06, |
|
"loss": 0.4248, |
|
"step": 6850 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.872452452452453e-06, |
|
"loss": 0.4221, |
|
"step": 6875 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.871951951951953e-06, |
|
"loss": 0.4206, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.871451451451452e-06, |
|
"loss": 0.4173, |
|
"step": 6925 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.870950950950951e-06, |
|
"loss": 0.4285, |
|
"step": 6950 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.870450450450452e-06, |
|
"loss": 0.4102, |
|
"step": 6975 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.86994994994995e-06, |
|
"loss": 0.4335, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"eval_loss": 0.5607422590255737, |
|
"eval_runtime": 2701.8903, |
|
"eval_samples_per_second": 6.001, |
|
"eval_steps_per_second": 0.15, |
|
"eval_wer": 36.83443698398872, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.86944944944945e-06, |
|
"loss": 0.4267, |
|
"step": 7025 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.86894894894895e-06, |
|
"loss": 0.439, |
|
"step": 7050 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.86844844844845e-06, |
|
"loss": 0.4139, |
|
"step": 7075 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.867947947947949e-06, |
|
"loss": 0.4248, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.867447447447448e-06, |
|
"loss": 0.4279, |
|
"step": 7125 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.866946946946947e-06, |
|
"loss": 0.4172, |
|
"step": 7150 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.866446446446447e-06, |
|
"loss": 0.4382, |
|
"step": 7175 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.865945945945948e-06, |
|
"loss": 0.4367, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.865445445445447e-06, |
|
"loss": 0.4247, |
|
"step": 7225 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.864944944944945e-06, |
|
"loss": 0.4223, |
|
"step": 7250 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.864444444444446e-06, |
|
"loss": 0.4207, |
|
"step": 7275 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.863943943943945e-06, |
|
"loss": 0.4166, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.863443443443444e-06, |
|
"loss": 0.4155, |
|
"step": 7325 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.862942942942944e-06, |
|
"loss": 0.4131, |
|
"step": 7350 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.862442442442443e-06, |
|
"loss": 0.4, |
|
"step": 7375 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.861941941941942e-06, |
|
"loss": 0.4051, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.861441441441443e-06, |
|
"loss": 0.3976, |
|
"step": 7425 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.860940940940942e-06, |
|
"loss": 0.4396, |
|
"step": 7450 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.86044044044044e-06, |
|
"loss": 0.3999, |
|
"step": 7475 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.859939939939941e-06, |
|
"loss": 0.4167, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.85943943943944e-06, |
|
"loss": 0.4324, |
|
"step": 7525 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.85893893893894e-06, |
|
"loss": 0.4185, |
|
"step": 7550 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.858438438438439e-06, |
|
"loss": 0.4205, |
|
"step": 7575 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.85793793793794e-06, |
|
"loss": 0.4011, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.857437437437437e-06, |
|
"loss": 0.4032, |
|
"step": 7625 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.856936936936937e-06, |
|
"loss": 0.4223, |
|
"step": 7650 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.856436436436438e-06, |
|
"loss": 0.4024, |
|
"step": 7675 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.855935935935937e-06, |
|
"loss": 0.4071, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.855435435435436e-06, |
|
"loss": 0.4024, |
|
"step": 7725 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.854934934934936e-06, |
|
"loss": 0.3904, |
|
"step": 7750 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.854434434434435e-06, |
|
"loss": 0.4041, |
|
"step": 7775 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.853933933933934e-06, |
|
"loss": 0.4201, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.853433433433435e-06, |
|
"loss": 0.393, |
|
"step": 7825 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.852932932932935e-06, |
|
"loss": 0.4405, |
|
"step": 7850 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.852432432432432e-06, |
|
"loss": 0.402, |
|
"step": 7875 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.851931931931933e-06, |
|
"loss": 0.4064, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.851431431431432e-06, |
|
"loss": 0.3908, |
|
"step": 7925 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.850930930930932e-06, |
|
"loss": 0.4232, |
|
"step": 7950 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.850430430430431e-06, |
|
"loss": 0.4311, |
|
"step": 7975 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.84992992992993e-06, |
|
"loss": 0.4266, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"eval_loss": 0.5630306601524353, |
|
"eval_runtime": 2643.0014, |
|
"eval_samples_per_second": 6.134, |
|
"eval_steps_per_second": 0.154, |
|
"eval_wer": 34.73531656401045, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.84942942942943e-06, |
|
"loss": 0.4117, |
|
"step": 8025 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.84894894894895e-06, |
|
"loss": 0.3983, |
|
"step": 8050 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.848448448448448e-06, |
|
"loss": 0.4025, |
|
"step": 8075 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.847947947947949e-06, |
|
"loss": 0.4119, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.847447447447448e-06, |
|
"loss": 0.4135, |
|
"step": 8125 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.846946946946948e-06, |
|
"loss": 0.4048, |
|
"step": 8150 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.846446446446447e-06, |
|
"loss": 0.4371, |
|
"step": 8175 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.845945945945948e-06, |
|
"loss": 0.3954, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.845445445445446e-06, |
|
"loss": 0.3933, |
|
"step": 8225 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.844944944944945e-06, |
|
"loss": 0.3956, |
|
"step": 8250 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.844444444444446e-06, |
|
"loss": 0.4109, |
|
"step": 8275 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.843943943943945e-06, |
|
"loss": 0.4087, |
|
"step": 8300 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.843443443443444e-06, |
|
"loss": 0.3904, |
|
"step": 8325 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.842942942942944e-06, |
|
"loss": 0.4309, |
|
"step": 8350 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.842442442442443e-06, |
|
"loss": 0.411, |
|
"step": 8375 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.841941941941942e-06, |
|
"loss": 0.384, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.841441441441443e-06, |
|
"loss": 0.3993, |
|
"step": 8425 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.840940940940943e-06, |
|
"loss": 0.3931, |
|
"step": 8450 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.84044044044044e-06, |
|
"loss": 0.4044, |
|
"step": 8475 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.839939939939941e-06, |
|
"loss": 0.3985, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.83943943943944e-06, |
|
"loss": 0.4317, |
|
"step": 8525 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.83893893893894e-06, |
|
"loss": 0.3932, |
|
"step": 8550 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.838438438438439e-06, |
|
"loss": 0.3925, |
|
"step": 8575 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.837937937937938e-06, |
|
"loss": 0.3799, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.837437437437438e-06, |
|
"loss": 0.3975, |
|
"step": 8625 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.836936936936937e-06, |
|
"loss": 0.4047, |
|
"step": 8650 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.836436436436438e-06, |
|
"loss": 0.4092, |
|
"step": 8675 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.835935935935936e-06, |
|
"loss": 0.3945, |
|
"step": 8700 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.835435435435437e-06, |
|
"loss": 0.4068, |
|
"step": 8725 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.834934934934936e-06, |
|
"loss": 0.3877, |
|
"step": 8750 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.834434434434435e-06, |
|
"loss": 0.3949, |
|
"step": 8775 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.833933933933934e-06, |
|
"loss": 0.3841, |
|
"step": 8800 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.833433433433435e-06, |
|
"loss": 0.4074, |
|
"step": 8825 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.832932932932933e-06, |
|
"loss": 0.4171, |
|
"step": 8850 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.832452452452454e-06, |
|
"loss": 0.3953, |
|
"step": 8875 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.831951951951953e-06, |
|
"loss": 0.3761, |
|
"step": 8900 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 9.831451451451452e-06, |
|
"loss": 0.4115, |
|
"step": 8925 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 9.830950950950952e-06, |
|
"loss": 0.3979, |
|
"step": 8950 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 9.830450450450451e-06, |
|
"loss": 0.405, |
|
"step": 8975 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 9.82994994994995e-06, |
|
"loss": 0.4057, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"eval_loss": 0.5405702590942383, |
|
"eval_runtime": 2698.8083, |
|
"eval_samples_per_second": 6.007, |
|
"eval_steps_per_second": 0.15, |
|
"eval_wer": 34.710648237370975, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 9.829449449449451e-06, |
|
"loss": 0.3833, |
|
"step": 9025 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 9.82894894894895e-06, |
|
"loss": 0.4058, |
|
"step": 9050 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 9.828448448448448e-06, |
|
"loss": 0.3861, |
|
"step": 9075 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 9.82794794794795e-06, |
|
"loss": 0.4031, |
|
"step": 9100 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 9.827447447447449e-06, |
|
"loss": 0.3894, |
|
"step": 9125 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 9.826946946946948e-06, |
|
"loss": 0.3892, |
|
"step": 9150 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 9.826446446446447e-06, |
|
"loss": 0.3991, |
|
"step": 9175 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 9.825945945945946e-06, |
|
"loss": 0.3803, |
|
"step": 9200 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 9.825445445445446e-06, |
|
"loss": 0.3927, |
|
"step": 9225 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 9.824944944944945e-06, |
|
"loss": 0.4084, |
|
"step": 9250 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 9.824444444444446e-06, |
|
"loss": 0.3916, |
|
"step": 9275 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 9.823943943943944e-06, |
|
"loss": 0.3674, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 9.823443443443445e-06, |
|
"loss": 0.3683, |
|
"step": 9325 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 9.822942942942944e-06, |
|
"loss": 0.4016, |
|
"step": 9350 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 9.822442442442443e-06, |
|
"loss": 0.395, |
|
"step": 9375 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 9.821941941941942e-06, |
|
"loss": 0.3906, |
|
"step": 9400 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 9.821441441441443e-06, |
|
"loss": 0.3843, |
|
"step": 9425 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 9.820940940940941e-06, |
|
"loss": 0.3732, |
|
"step": 9450 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 9.82044044044044e-06, |
|
"loss": 0.3927, |
|
"step": 9475 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 9.819939939939941e-06, |
|
"loss": 0.4151, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 9.81943943943944e-06, |
|
"loss": 0.3933, |
|
"step": 9525 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 9.81893893893894e-06, |
|
"loss": 0.3856, |
|
"step": 9550 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 9.81843843843844e-06, |
|
"loss": 0.3818, |
|
"step": 9575 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 9.817937937937939e-06, |
|
"loss": 0.3932, |
|
"step": 9600 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 9.817437437437438e-06, |
|
"loss": 0.4085, |
|
"step": 9625 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 9.816936936936939e-06, |
|
"loss": 0.4023, |
|
"step": 9650 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 9.816436436436436e-06, |
|
"loss": 0.3918, |
|
"step": 9675 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 9.815935935935936e-06, |
|
"loss": 0.3965, |
|
"step": 9700 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 9.815435435435437e-06, |
|
"loss": 0.3934, |
|
"step": 9725 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 9.814934934934936e-06, |
|
"loss": 0.3807, |
|
"step": 9750 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 9.814434434434435e-06, |
|
"loss": 0.3809, |
|
"step": 9775 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 9.813933933933935e-06, |
|
"loss": 0.3804, |
|
"step": 9800 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 9.813433433433434e-06, |
|
"loss": 0.4112, |
|
"step": 9825 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 9.812932932932933e-06, |
|
"loss": 0.3853, |
|
"step": 9850 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 9.812432432432433e-06, |
|
"loss": 0.3967, |
|
"step": 9875 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 9.811931931931934e-06, |
|
"loss": 0.3934, |
|
"step": 9900 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 9.811431431431431e-06, |
|
"loss": 0.4013, |
|
"step": 9925 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 9.810930930930932e-06, |
|
"loss": 0.4004, |
|
"step": 9950 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 9.810430430430431e-06, |
|
"loss": 0.3825, |
|
"step": 9975 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 9.80992992992993e-06, |
|
"loss": 0.3798, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"eval_loss": 0.5242385864257812, |
|
"eval_runtime": 2613.1802, |
|
"eval_samples_per_second": 6.204, |
|
"eval_steps_per_second": 0.155, |
|
"eval_wer": 32.9584261607604, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 9.80942942942943e-06, |
|
"loss": 0.379, |
|
"step": 10025 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 9.808928928928931e-06, |
|
"loss": 0.397, |
|
"step": 10050 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 9.808428428428429e-06, |
|
"loss": 0.3818, |
|
"step": 10075 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 9.807927927927928e-06, |
|
"loss": 0.373, |
|
"step": 10100 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 9.807427427427429e-06, |
|
"loss": 0.3592, |
|
"step": 10125 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 9.806926926926928e-06, |
|
"loss": 0.3824, |
|
"step": 10150 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 9.806426426426427e-06, |
|
"loss": 0.3907, |
|
"step": 10175 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 9.805925925925927e-06, |
|
"loss": 0.3787, |
|
"step": 10200 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 9.805425425425426e-06, |
|
"loss": 0.3571, |
|
"step": 10225 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 9.804924924924925e-06, |
|
"loss": 0.3785, |
|
"step": 10250 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 9.804424424424426e-06, |
|
"loss": 0.41, |
|
"step": 10275 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 9.803923923923924e-06, |
|
"loss": 0.3703, |
|
"step": 10300 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 9.803423423423423e-06, |
|
"loss": 0.38, |
|
"step": 10325 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 9.802922922922924e-06, |
|
"loss": 0.3787, |
|
"step": 10350 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 9.802422422422424e-06, |
|
"loss": 0.3948, |
|
"step": 10375 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 9.801921921921923e-06, |
|
"loss": 0.3775, |
|
"step": 10400 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 9.801421421421422e-06, |
|
"loss": 0.3823, |
|
"step": 10425 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 9.800920920920921e-06, |
|
"loss": 0.3868, |
|
"step": 10450 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 9.80042042042042e-06, |
|
"loss": 0.378, |
|
"step": 10475 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 9.79991991991992e-06, |
|
"loss": 0.3911, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 9.799419419419421e-06, |
|
"loss": 0.3845, |
|
"step": 10525 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 9.798918918918919e-06, |
|
"loss": 0.3899, |
|
"step": 10550 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 9.79841841841842e-06, |
|
"loss": 0.3919, |
|
"step": 10575 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 9.797917917917919e-06, |
|
"loss": 0.4163, |
|
"step": 10600 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 9.797417417417418e-06, |
|
"loss": 0.382, |
|
"step": 10625 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 9.796916916916918e-06, |
|
"loss": 0.3796, |
|
"step": 10650 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 9.796416416416419e-06, |
|
"loss": 0.3751, |
|
"step": 10675 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 9.795915915915916e-06, |
|
"loss": 0.3768, |
|
"step": 10700 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 9.795415415415415e-06, |
|
"loss": 0.3878, |
|
"step": 10725 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 9.794914914914916e-06, |
|
"loss": 0.3821, |
|
"step": 10750 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 9.794414414414416e-06, |
|
"loss": 0.3879, |
|
"step": 10775 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 9.793913913913915e-06, |
|
"loss": 0.3685, |
|
"step": 10800 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 9.793413413413414e-06, |
|
"loss": 0.3673, |
|
"step": 10825 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 9.792912912912914e-06, |
|
"loss": 0.3955, |
|
"step": 10850 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 9.792412412412413e-06, |
|
"loss": 0.3616, |
|
"step": 10875 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 9.791911911911914e-06, |
|
"loss": 0.3692, |
|
"step": 10900 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 9.791411411411411e-06, |
|
"loss": 0.3766, |
|
"step": 10925 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 9.79091091091091e-06, |
|
"loss": 0.3873, |
|
"step": 10950 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 9.790410410410412e-06, |
|
"loss": 0.3792, |
|
"step": 10975 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 9.789909909909911e-06, |
|
"loss": 0.3781, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"eval_loss": 0.5272489190101624, |
|
"eval_runtime": 2804.5214, |
|
"eval_samples_per_second": 5.781, |
|
"eval_steps_per_second": 0.145, |
|
"eval_wer": 34.24349180163582, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 9.78940940940941e-06, |
|
"loss": 0.3926, |
|
"step": 11025 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 9.78890890890891e-06, |
|
"loss": 0.3849, |
|
"step": 11050 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 9.788408408408409e-06, |
|
"loss": 0.3726, |
|
"step": 11075 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 9.787907907907908e-06, |
|
"loss": 0.3798, |
|
"step": 11100 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 9.787407407407408e-06, |
|
"loss": 0.3877, |
|
"step": 11125 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 9.786906906906909e-06, |
|
"loss": 0.3873, |
|
"step": 11150 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 9.786426426426428e-06, |
|
"loss": 0.381, |
|
"step": 11175 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 9.785925925925927e-06, |
|
"loss": 0.3769, |
|
"step": 11200 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 9.785425425425426e-06, |
|
"loss": 0.3903, |
|
"step": 11225 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 9.784924924924926e-06, |
|
"loss": 0.3778, |
|
"step": 11250 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 9.784424424424427e-06, |
|
"loss": 0.3966, |
|
"step": 11275 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 9.783923923923924e-06, |
|
"loss": 0.3851, |
|
"step": 11300 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 9.783423423423423e-06, |
|
"loss": 0.3737, |
|
"step": 11325 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 9.782922922922924e-06, |
|
"loss": 0.3602, |
|
"step": 11350 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 9.782422422422424e-06, |
|
"loss": 0.3722, |
|
"step": 11375 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 9.781921921921923e-06, |
|
"loss": 0.3874, |
|
"step": 11400 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 9.781421421421422e-06, |
|
"loss": 0.3895, |
|
"step": 11425 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 9.780920920920922e-06, |
|
"loss": 0.375, |
|
"step": 11450 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 9.780420420420421e-06, |
|
"loss": 0.3858, |
|
"step": 11475 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 9.77991991991992e-06, |
|
"loss": 0.3877, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 9.77941941941942e-06, |
|
"loss": 0.3932, |
|
"step": 11525 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 9.778918918918919e-06, |
|
"loss": 0.3792, |
|
"step": 11550 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 9.77841841841842e-06, |
|
"loss": 0.3709, |
|
"step": 11575 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 9.777917917917919e-06, |
|
"loss": 0.3815, |
|
"step": 11600 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 9.777417417417418e-06, |
|
"loss": 0.3834, |
|
"step": 11625 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 9.776916916916918e-06, |
|
"loss": 0.3598, |
|
"step": 11650 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 9.776416416416417e-06, |
|
"loss": 0.3648, |
|
"step": 11675 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 9.775915915915916e-06, |
|
"loss": 0.3582, |
|
"step": 11700 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 9.775415415415416e-06, |
|
"loss": 0.3753, |
|
"step": 11725 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 9.774914914914917e-06, |
|
"loss": 0.3689, |
|
"step": 11750 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 9.774414414414414e-06, |
|
"loss": 0.3774, |
|
"step": 11775 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 9.773913913913915e-06, |
|
"loss": 0.3825, |
|
"step": 11800 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 9.773413413413414e-06, |
|
"loss": 0.36, |
|
"step": 11825 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 9.772912912912914e-06, |
|
"loss": 0.3672, |
|
"step": 11850 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 9.772412412412413e-06, |
|
"loss": 0.3848, |
|
"step": 11875 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 9.771911911911914e-06, |
|
"loss": 0.357, |
|
"step": 11900 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 9.771411411411412e-06, |
|
"loss": 0.3658, |
|
"step": 11925 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 9.770910910910911e-06, |
|
"loss": 0.3733, |
|
"step": 11950 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 9.770410410410412e-06, |
|
"loss": 0.3731, |
|
"step": 11975 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 9.76990990990991e-06, |
|
"loss": 0.3674, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"eval_loss": 0.5061578750610352, |
|
"eval_runtime": 2740.8866, |
|
"eval_samples_per_second": 5.915, |
|
"eval_steps_per_second": 0.148, |
|
"eval_wer": 33.014700780906715, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 9.76940940940941e-06, |
|
"loss": 0.3981, |
|
"step": 12025 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 9.76890890890891e-06, |
|
"loss": 0.3759, |
|
"step": 12050 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 9.76840840840841e-06, |
|
"loss": 0.3647, |
|
"step": 12075 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 9.767907907907908e-06, |
|
"loss": 0.3756, |
|
"step": 12100 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 9.767407407407408e-06, |
|
"loss": 0.3444, |
|
"step": 12125 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 9.766906906906907e-06, |
|
"loss": 0.3536, |
|
"step": 12150 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 9.766406406406406e-06, |
|
"loss": 0.3584, |
|
"step": 12175 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 9.765905905905907e-06, |
|
"loss": 0.3879, |
|
"step": 12200 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 9.765405405405407e-06, |
|
"loss": 0.3403, |
|
"step": 12225 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 9.764904904904906e-06, |
|
"loss": 0.3646, |
|
"step": 12250 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 9.764404404404405e-06, |
|
"loss": 0.3697, |
|
"step": 12275 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 9.763903903903905e-06, |
|
"loss": 0.3607, |
|
"step": 12300 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 9.763403403403404e-06, |
|
"loss": 0.3588, |
|
"step": 12325 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 9.762902902902903e-06, |
|
"loss": 0.3668, |
|
"step": 12350 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 9.762402402402404e-06, |
|
"loss": 0.3933, |
|
"step": 12375 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 9.761901901901902e-06, |
|
"loss": 0.3744, |
|
"step": 12400 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 9.761401401401403e-06, |
|
"loss": 0.3905, |
|
"step": 12425 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 9.760900900900902e-06, |
|
"loss": 0.3641, |
|
"step": 12450 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 9.760400400400401e-06, |
|
"loss": 0.3647, |
|
"step": 12475 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 9.7598998998999e-06, |
|
"loss": 0.35, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 9.7593993993994e-06, |
|
"loss": 0.3545, |
|
"step": 12525 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 9.7588988988989e-06, |
|
"loss": 0.3645, |
|
"step": 12550 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 9.758398398398398e-06, |
|
"loss": 0.3625, |
|
"step": 12575 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 9.7578978978979e-06, |
|
"loss": 0.3801, |
|
"step": 12600 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 9.757397397397397e-06, |
|
"loss": 0.3562, |
|
"step": 12625 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 9.756896896896898e-06, |
|
"loss": 0.3594, |
|
"step": 12650 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 9.756396396396397e-06, |
|
"loss": 0.3644, |
|
"step": 12675 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 9.755895895895897e-06, |
|
"loss": 0.356, |
|
"step": 12700 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 9.755395395395396e-06, |
|
"loss": 0.3676, |
|
"step": 12725 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 9.754894894894895e-06, |
|
"loss": 0.3722, |
|
"step": 12750 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 9.754394394394395e-06, |
|
"loss": 0.396, |
|
"step": 12775 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 9.753893893893894e-06, |
|
"loss": 0.3613, |
|
"step": 12800 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 9.753393393393395e-06, |
|
"loss": 0.3626, |
|
"step": 12825 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 9.752892892892894e-06, |
|
"loss": 0.3642, |
|
"step": 12850 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 9.752392392392393e-06, |
|
"loss": 0.3674, |
|
"step": 12875 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 9.751891891891893e-06, |
|
"loss": 0.3526, |
|
"step": 12900 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.751391391391392e-06, |
|
"loss": 0.3767, |
|
"step": 12925 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.750890890890891e-06, |
|
"loss": 0.3616, |
|
"step": 12950 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.75039039039039e-06, |
|
"loss": 0.3475, |
|
"step": 12975 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.749889889889892e-06, |
|
"loss": 0.3507, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"eval_loss": 0.49944427609443665, |
|
"eval_runtime": 2897.6163, |
|
"eval_samples_per_second": 5.595, |
|
"eval_steps_per_second": 0.14, |
|
"eval_wer": 33.72160251616932, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.74938938938939e-06, |
|
"loss": 0.3743, |
|
"step": 13025 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.74888888888889e-06, |
|
"loss": 0.3597, |
|
"step": 13050 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.74838838838839e-06, |
|
"loss": 0.3862, |
|
"step": 13075 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.747887887887889e-06, |
|
"loss": 0.3606, |
|
"step": 13100 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.747387387387388e-06, |
|
"loss": 0.4099, |
|
"step": 13125 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.746886886886887e-06, |
|
"loss": 0.3583, |
|
"step": 13150 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.746386386386387e-06, |
|
"loss": 0.3673, |
|
"step": 13175 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.745885885885886e-06, |
|
"loss": 0.3597, |
|
"step": 13200 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.745385385385387e-06, |
|
"loss": 0.3715, |
|
"step": 13225 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.744884884884885e-06, |
|
"loss": 0.3681, |
|
"step": 13250 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.744384384384386e-06, |
|
"loss": 0.3793, |
|
"step": 13275 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.743883883883885e-06, |
|
"loss": 0.3574, |
|
"step": 13300 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.743383383383384e-06, |
|
"loss": 0.3602, |
|
"step": 13325 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.742882882882883e-06, |
|
"loss": 0.3663, |
|
"step": 13350 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.742382382382383e-06, |
|
"loss": 0.3947, |
|
"step": 13375 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.741881881881882e-06, |
|
"loss": 0.3709, |
|
"step": 13400 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.741381381381381e-06, |
|
"loss": 0.3625, |
|
"step": 13425 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.740880880880882e-06, |
|
"loss": 0.3574, |
|
"step": 13450 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.740380380380382e-06, |
|
"loss": 0.3642, |
|
"step": 13475 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.739879879879881e-06, |
|
"loss": 0.3639, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.73937937937938e-06, |
|
"loss": 0.3669, |
|
"step": 13525 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.73887887887888e-06, |
|
"loss": 0.3762, |
|
"step": 13550 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.738378378378379e-06, |
|
"loss": 0.3564, |
|
"step": 13575 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.737877877877878e-06, |
|
"loss": 0.3602, |
|
"step": 13600 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.737377377377379e-06, |
|
"loss": 0.3464, |
|
"step": 13625 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.736876876876877e-06, |
|
"loss": 0.3618, |
|
"step": 13650 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 9.736376376376378e-06, |
|
"loss": 0.3761, |
|
"step": 13675 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 9.735875875875877e-06, |
|
"loss": 0.3604, |
|
"step": 13700 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 9.735375375375376e-06, |
|
"loss": 0.3714, |
|
"step": 13725 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 9.734874874874876e-06, |
|
"loss": 0.3699, |
|
"step": 13750 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 9.734374374374375e-06, |
|
"loss": 0.3728, |
|
"step": 13775 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 9.733873873873874e-06, |
|
"loss": 0.3878, |
|
"step": 13800 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 9.733373373373374e-06, |
|
"loss": 0.3461, |
|
"step": 13825 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 9.732872872872875e-06, |
|
"loss": 0.3415, |
|
"step": 13850 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 9.732372372372372e-06, |
|
"loss": 0.3548, |
|
"step": 13875 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 9.731871871871873e-06, |
|
"loss": 0.3729, |
|
"step": 13900 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 9.731371371371372e-06, |
|
"loss": 0.3705, |
|
"step": 13925 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 9.730870870870872e-06, |
|
"loss": 0.3605, |
|
"step": 13950 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 9.730370370370371e-06, |
|
"loss": 0.3631, |
|
"step": 13975 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 9.72986986986987e-06, |
|
"loss": 0.375, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"eval_loss": 0.5072089433670044, |
|
"eval_runtime": 2512.5863, |
|
"eval_samples_per_second": 6.453, |
|
"eval_steps_per_second": 0.162, |
|
"eval_wer": 30.171676135706633, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 9.72936936936937e-06, |
|
"loss": 0.3642, |
|
"step": 14025 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 9.728868868868869e-06, |
|
"loss": 0.348, |
|
"step": 14050 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 9.72836836836837e-06, |
|
"loss": 0.3413, |
|
"step": 14075 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 9.72786786786787e-06, |
|
"loss": 0.3528, |
|
"step": 14100 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 9.727367367367368e-06, |
|
"loss": 0.3501, |
|
"step": 14125 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 9.726866866866868e-06, |
|
"loss": 0.3318, |
|
"step": 14150 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 9.726366366366367e-06, |
|
"loss": 0.3655, |
|
"step": 14175 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 9.725865865865866e-06, |
|
"loss": 0.3496, |
|
"step": 14200 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 9.725365365365366e-06, |
|
"loss": 0.3379, |
|
"step": 14225 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 9.724884884884885e-06, |
|
"loss": 0.3893, |
|
"step": 14250 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 9.724384384384386e-06, |
|
"loss": 0.3703, |
|
"step": 14275 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 9.723883883883885e-06, |
|
"loss": 0.3531, |
|
"step": 14300 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 9.723383383383384e-06, |
|
"loss": 0.3604, |
|
"step": 14325 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 9.722882882882884e-06, |
|
"loss": 0.3527, |
|
"step": 14350 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 9.722382382382383e-06, |
|
"loss": 0.3556, |
|
"step": 14375 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 9.721881881881882e-06, |
|
"loss": 0.3664, |
|
"step": 14400 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 9.721381381381382e-06, |
|
"loss": 0.3641, |
|
"step": 14425 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 9.720880880880883e-06, |
|
"loss": 0.3553, |
|
"step": 14450 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 9.72038038038038e-06, |
|
"loss": 0.3605, |
|
"step": 14475 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 9.719879879879881e-06, |
|
"loss": 0.3412, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 9.71937937937938e-06, |
|
"loss": 0.3588, |
|
"step": 14525 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 9.71887887887888e-06, |
|
"loss": 0.3334, |
|
"step": 14550 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 9.718378378378379e-06, |
|
"loss": 0.3624, |
|
"step": 14575 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 9.717877877877878e-06, |
|
"loss": 0.3609, |
|
"step": 14600 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 9.717377377377378e-06, |
|
"loss": 0.3339, |
|
"step": 14625 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 9.716876876876877e-06, |
|
"loss": 0.353, |
|
"step": 14650 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 9.716376376376378e-06, |
|
"loss": 0.3517, |
|
"step": 14675 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 9.715875875875877e-06, |
|
"loss": 0.3626, |
|
"step": 14700 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 9.715375375375377e-06, |
|
"loss": 0.3268, |
|
"step": 14725 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 9.714874874874876e-06, |
|
"loss": 0.3648, |
|
"step": 14750 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 9.714374374374375e-06, |
|
"loss": 0.3579, |
|
"step": 14775 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 9.713873873873874e-06, |
|
"loss": 0.3403, |
|
"step": 14800 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 9.713373373373374e-06, |
|
"loss": 0.3597, |
|
"step": 14825 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 9.712872872872873e-06, |
|
"loss": 0.3323, |
|
"step": 14850 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 9.712372372372372e-06, |
|
"loss": 0.3412, |
|
"step": 14875 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 9.711871871871873e-06, |
|
"loss": 0.3574, |
|
"step": 14900 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 9.711371371371373e-06, |
|
"loss": 0.348, |
|
"step": 14925 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 9.710870870870872e-06, |
|
"loss": 0.3588, |
|
"step": 14950 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 9.710370370370371e-06, |
|
"loss": 0.3445, |
|
"step": 14975 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 9.70986986986987e-06, |
|
"loss": 0.3545, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"eval_loss": 0.5023101568222046, |
|
"eval_runtime": 2571.7112, |
|
"eval_samples_per_second": 6.304, |
|
"eval_steps_per_second": 0.158, |
|
"eval_wer": 30.215616592533205, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 9.70936936936937e-06, |
|
"loss": 0.3526, |
|
"step": 15025 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 9.708868868868869e-06, |
|
"loss": 0.361, |
|
"step": 15050 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 9.70836836836837e-06, |
|
"loss": 0.3731, |
|
"step": 15075 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 9.707867867867868e-06, |
|
"loss": 0.3499, |
|
"step": 15100 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 9.707367367367369e-06, |
|
"loss": 0.3608, |
|
"step": 15125 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 9.706866866866868e-06, |
|
"loss": 0.3465, |
|
"step": 15150 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 9.706366366366367e-06, |
|
"loss": 0.362, |
|
"step": 15175 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 9.705865865865867e-06, |
|
"loss": 0.348, |
|
"step": 15200 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 9.705365365365366e-06, |
|
"loss": 0.3532, |
|
"step": 15225 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 9.704864864864865e-06, |
|
"loss": 0.3352, |
|
"step": 15250 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 9.704364364364364e-06, |
|
"loss": 0.3424, |
|
"step": 15275 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 9.703863863863865e-06, |
|
"loss": 0.3569, |
|
"step": 15300 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 9.703363363363365e-06, |
|
"loss": 0.329, |
|
"step": 15325 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 9.702862862862864e-06, |
|
"loss": 0.3551, |
|
"step": 15350 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 9.702362362362363e-06, |
|
"loss": 0.3592, |
|
"step": 15375 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 9.701861861861863e-06, |
|
"loss": 0.359, |
|
"step": 15400 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 9.701361361361362e-06, |
|
"loss": 0.3498, |
|
"step": 15425 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 9.700860860860861e-06, |
|
"loss": 0.352, |
|
"step": 15450 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 9.70036036036036e-06, |
|
"loss": 0.3451, |
|
"step": 15475 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 9.69985985985986e-06, |
|
"loss": 0.3582, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 9.69935935935936e-06, |
|
"loss": 0.3495, |
|
"step": 15525 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 9.69885885885886e-06, |
|
"loss": 0.3603, |
|
"step": 15550 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 9.69835835835836e-06, |
|
"loss": 0.3398, |
|
"step": 15575 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 9.697857857857859e-06, |
|
"loss": 0.3532, |
|
"step": 15600 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 9.697357357357358e-06, |
|
"loss": 0.3848, |
|
"step": 15625 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 9.696856856856857e-06, |
|
"loss": 0.3341, |
|
"step": 15650 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 9.696356356356357e-06, |
|
"loss": 0.37, |
|
"step": 15675 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 9.695855855855858e-06, |
|
"loss": 0.3635, |
|
"step": 15700 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 9.695355355355355e-06, |
|
"loss": 0.3614, |
|
"step": 15725 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 9.694854854854856e-06, |
|
"loss": 0.3651, |
|
"step": 15750 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 9.694354354354355e-06, |
|
"loss": 0.3733, |
|
"step": 15775 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 9.693853853853855e-06, |
|
"loss": 0.3455, |
|
"step": 15800 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 9.693353353353354e-06, |
|
"loss": 0.3551, |
|
"step": 15825 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 9.692852852852853e-06, |
|
"loss": 0.3517, |
|
"step": 15850 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 9.692352352352353e-06, |
|
"loss": 0.3595, |
|
"step": 15875 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 9.691851851851852e-06, |
|
"loss": 0.3694, |
|
"step": 15900 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 9.691351351351353e-06, |
|
"loss": 0.3433, |
|
"step": 15925 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 9.69085085085085e-06, |
|
"loss": 0.3443, |
|
"step": 15950 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 9.690350350350352e-06, |
|
"loss": 0.3754, |
|
"step": 15975 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 9.689849849849851e-06, |
|
"loss": 0.369, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"eval_loss": 0.4858362674713135, |
|
"eval_runtime": 2697.5515, |
|
"eval_samples_per_second": 6.01, |
|
"eval_steps_per_second": 0.151, |
|
"eval_wer": 31.24320657410905, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 9.68934934934935e-06, |
|
"loss": 0.3536, |
|
"step": 16025 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 9.68884884884885e-06, |
|
"loss": 0.339, |
|
"step": 16050 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 9.688348348348349e-06, |
|
"loss": 0.3319, |
|
"step": 16075 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 9.687847847847848e-06, |
|
"loss": 0.3742, |
|
"step": 16100 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 9.687347347347347e-06, |
|
"loss": 0.3299, |
|
"step": 16125 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 9.686846846846848e-06, |
|
"loss": 0.3526, |
|
"step": 16150 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 9.686346346346348e-06, |
|
"loss": 0.3479, |
|
"step": 16175 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 9.685845845845847e-06, |
|
"loss": 0.3631, |
|
"step": 16200 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 9.685345345345346e-06, |
|
"loss": 0.3545, |
|
"step": 16225 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 9.684844844844846e-06, |
|
"loss": 0.3651, |
|
"step": 16250 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 9.684344344344345e-06, |
|
"loss": 0.3438, |
|
"step": 16275 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 9.683843843843844e-06, |
|
"loss": 0.3502, |
|
"step": 16300 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 9.683343343343345e-06, |
|
"loss": 0.357, |
|
"step": 16325 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 9.682842842842843e-06, |
|
"loss": 0.349, |
|
"step": 16350 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 9.682342342342344e-06, |
|
"loss": 0.3369, |
|
"step": 16375 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 9.681841841841843e-06, |
|
"loss": 0.361, |
|
"step": 16400 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 9.681341341341342e-06, |
|
"loss": 0.3435, |
|
"step": 16425 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 9.680840840840842e-06, |
|
"loss": 0.3185, |
|
"step": 16450 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 9.680340340340341e-06, |
|
"loss": 0.3341, |
|
"step": 16475 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 9.67983983983984e-06, |
|
"loss": 0.3349, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 9.67933933933934e-06, |
|
"loss": 0.3453, |
|
"step": 16525 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 9.67883883883884e-06, |
|
"loss": 0.3519, |
|
"step": 16550 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 9.678338338338338e-06, |
|
"loss": 0.3529, |
|
"step": 16575 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 9.677857857857859e-06, |
|
"loss": 0.3486, |
|
"step": 16600 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 9.677357357357358e-06, |
|
"loss": 0.3424, |
|
"step": 16625 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 9.676856856856858e-06, |
|
"loss": 0.3455, |
|
"step": 16650 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 9.676356356356357e-06, |
|
"loss": 0.3431, |
|
"step": 16675 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 9.675855855855856e-06, |
|
"loss": 0.3379, |
|
"step": 16700 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 9.675355355355355e-06, |
|
"loss": 0.3539, |
|
"step": 16725 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 9.674854854854856e-06, |
|
"loss": 0.327, |
|
"step": 16750 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 9.674354354354356e-06, |
|
"loss": 0.3275, |
|
"step": 16775 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 9.673853853853855e-06, |
|
"loss": 0.3396, |
|
"step": 16800 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 9.673353353353354e-06, |
|
"loss": 0.3476, |
|
"step": 16825 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 9.672852852852854e-06, |
|
"loss": 0.3675, |
|
"step": 16850 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 9.672352352352353e-06, |
|
"loss": 0.3571, |
|
"step": 16875 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 9.671851851851852e-06, |
|
"loss": 0.3463, |
|
"step": 16900 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 9.671351351351353e-06, |
|
"loss": 0.3381, |
|
"step": 16925 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 9.67085085085085e-06, |
|
"loss": 0.3582, |
|
"step": 16950 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 9.670350350350352e-06, |
|
"loss": 0.3633, |
|
"step": 16975 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 9.669849849849851e-06, |
|
"loss": 0.3585, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"eval_loss": 0.4826071560382843, |
|
"eval_runtime": 2556.6287, |
|
"eval_samples_per_second": 6.342, |
|
"eval_steps_per_second": 0.159, |
|
"eval_wer": 29.16952536597775, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 9.66934934934935e-06, |
|
"loss": 0.3504, |
|
"step": 17025 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 9.66884884884885e-06, |
|
"loss": 0.3523, |
|
"step": 17050 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 9.668348348348349e-06, |
|
"loss": 0.3572, |
|
"step": 17075 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 9.667847847847848e-06, |
|
"loss": 0.364, |
|
"step": 17100 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 9.667347347347348e-06, |
|
"loss": 0.3251, |
|
"step": 17125 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 9.666846846846849e-06, |
|
"loss": 0.3318, |
|
"step": 17150 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 9.666346346346346e-06, |
|
"loss": 0.3456, |
|
"step": 17175 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 9.665845845845847e-06, |
|
"loss": 0.345, |
|
"step": 17200 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 9.665345345345346e-06, |
|
"loss": 0.3224, |
|
"step": 17225 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 9.664844844844846e-06, |
|
"loss": 0.3397, |
|
"step": 17250 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 9.664344344344345e-06, |
|
"loss": 0.3484, |
|
"step": 17275 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 9.663843843843844e-06, |
|
"loss": 0.3267, |
|
"step": 17300 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 9.663343343343344e-06, |
|
"loss": 0.3642, |
|
"step": 17325 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 9.662842842842843e-06, |
|
"loss": 0.3447, |
|
"step": 17350 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 9.662342342342344e-06, |
|
"loss": 0.3429, |
|
"step": 17375 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 9.661841841841843e-06, |
|
"loss": 0.3258, |
|
"step": 17400 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 9.661341341341343e-06, |
|
"loss": 0.3305, |
|
"step": 17425 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 9.660840840840842e-06, |
|
"loss": 0.3376, |
|
"step": 17450 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 9.660340340340341e-06, |
|
"loss": 0.3417, |
|
"step": 17475 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 9.65983983983984e-06, |
|
"loss": 0.3326, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 9.65933933933934e-06, |
|
"loss": 0.334, |
|
"step": 17525 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 9.65883883883884e-06, |
|
"loss": 0.3461, |
|
"step": 17550 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 9.658338338338338e-06, |
|
"loss": 0.3512, |
|
"step": 17575 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 9.65783783783784e-06, |
|
"loss": 0.3609, |
|
"step": 17600 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 9.657337337337339e-06, |
|
"loss": 0.3502, |
|
"step": 17625 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 9.656836836836838e-06, |
|
"loss": 0.3508, |
|
"step": 17650 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 9.656336336336337e-06, |
|
"loss": 0.3436, |
|
"step": 17675 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 9.655835835835836e-06, |
|
"loss": 0.3525, |
|
"step": 17700 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 9.655335335335336e-06, |
|
"loss": 0.339, |
|
"step": 17725 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 9.654834834834835e-06, |
|
"loss": 0.3507, |
|
"step": 17750 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 9.654334334334336e-06, |
|
"loss": 0.3277, |
|
"step": 17775 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 9.653833833833834e-06, |
|
"loss": 0.3486, |
|
"step": 17800 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 9.653353353353354e-06, |
|
"loss": 0.3529, |
|
"step": 17825 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 9.652852852852854e-06, |
|
"loss": 0.3228, |
|
"step": 17850 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 9.652352352352353e-06, |
|
"loss": 0.335, |
|
"step": 17875 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 9.651851851851852e-06, |
|
"loss": 0.3641, |
|
"step": 17900 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 9.651351351351352e-06, |
|
"loss": 0.3248, |
|
"step": 17925 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 9.650850850850851e-06, |
|
"loss": 0.3308, |
|
"step": 17950 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 9.650350350350352e-06, |
|
"loss": 0.333, |
|
"step": 17975 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 9.649849849849851e-06, |
|
"loss": 0.3302, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"eval_loss": 0.4842660427093506, |
|
"eval_runtime": 2644.843, |
|
"eval_samples_per_second": 6.13, |
|
"eval_steps_per_second": 0.154, |
|
"eval_wer": 29.56498947741692, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 9.64934934934935e-06, |
|
"loss": 0.3624, |
|
"step": 18025 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 9.64884884884885e-06, |
|
"loss": 0.3645, |
|
"step": 18050 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 9.648348348348349e-06, |
|
"loss": 0.3563, |
|
"step": 18075 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 9.647847847847848e-06, |
|
"loss": 0.3354, |
|
"step": 18100 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 9.647347347347348e-06, |
|
"loss": 0.3352, |
|
"step": 18125 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 9.646846846846849e-06, |
|
"loss": 0.3426, |
|
"step": 18150 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 9.646346346346346e-06, |
|
"loss": 0.3534, |
|
"step": 18175 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 9.645845845845847e-06, |
|
"loss": 0.3316, |
|
"step": 18200 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 9.645345345345347e-06, |
|
"loss": 0.3387, |
|
"step": 18225 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 9.644844844844846e-06, |
|
"loss": 0.3489, |
|
"step": 18250 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 9.644344344344345e-06, |
|
"loss": 0.3339, |
|
"step": 18275 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 9.643843843843845e-06, |
|
"loss": 0.3416, |
|
"step": 18300 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 9.643343343343344e-06, |
|
"loss": 0.3259, |
|
"step": 18325 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 9.642842842842843e-06, |
|
"loss": 0.3476, |
|
"step": 18350 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 9.642342342342344e-06, |
|
"loss": 0.3499, |
|
"step": 18375 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 9.641841841841842e-06, |
|
"loss": 0.3438, |
|
"step": 18400 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 9.641341341341343e-06, |
|
"loss": 0.3316, |
|
"step": 18425 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 9.640840840840842e-06, |
|
"loss": 0.3397, |
|
"step": 18450 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 9.640340340340341e-06, |
|
"loss": 0.3281, |
|
"step": 18475 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 9.63983983983984e-06, |
|
"loss": 0.3316, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 9.63933933933934e-06, |
|
"loss": 0.3487, |
|
"step": 18525 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 9.63883883883884e-06, |
|
"loss": 0.3273, |
|
"step": 18550 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 9.638338338338338e-06, |
|
"loss": 0.3522, |
|
"step": 18575 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 9.63783783783784e-06, |
|
"loss": 0.3261, |
|
"step": 18600 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 9.637337337337339e-06, |
|
"loss": 0.3551, |
|
"step": 18625 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 9.636836836836838e-06, |
|
"loss": 0.343, |
|
"step": 18650 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 9.636336336336337e-06, |
|
"loss": 0.3302, |
|
"step": 18675 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 9.635835835835837e-06, |
|
"loss": 0.3287, |
|
"step": 18700 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 9.635335335335336e-06, |
|
"loss": 0.3317, |
|
"step": 18725 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 9.634834834834835e-06, |
|
"loss": 0.3346, |
|
"step": 18750 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 9.634334334334336e-06, |
|
"loss": 0.3531, |
|
"step": 18775 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 9.633833833833834e-06, |
|
"loss": 0.35, |
|
"step": 18800 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 9.633333333333335e-06, |
|
"loss": 0.3407, |
|
"step": 18825 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 9.632832832832834e-06, |
|
"loss": 0.3193, |
|
"step": 18850 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 9.632332332332333e-06, |
|
"loss": 0.351, |
|
"step": 18875 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 9.631831831831833e-06, |
|
"loss": 0.3554, |
|
"step": 18900 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 9.631331331331332e-06, |
|
"loss": 0.3389, |
|
"step": 18925 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 9.630830830830831e-06, |
|
"loss": 0.3196, |
|
"step": 18950 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 9.63033033033033e-06, |
|
"loss": 0.3271, |
|
"step": 18975 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 9.629829829829832e-06, |
|
"loss": 0.3252, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"eval_loss": 0.4709344506263733, |
|
"eval_runtime": 2536.9152, |
|
"eval_samples_per_second": 6.391, |
|
"eval_steps_per_second": 0.16, |
|
"eval_wer": 28.698514504205182, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 9.62932932932933e-06, |
|
"loss": 0.3142, |
|
"step": 19025 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 9.62882882882883e-06, |
|
"loss": 0.3346, |
|
"step": 19050 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 9.62832832832833e-06, |
|
"loss": 0.3582, |
|
"step": 19075 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 9.627827827827829e-06, |
|
"loss": 0.327, |
|
"step": 19100 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 9.627327327327328e-06, |
|
"loss": 0.3347, |
|
"step": 19125 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 9.626826826826827e-06, |
|
"loss": 0.3509, |
|
"step": 19150 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 9.626326326326327e-06, |
|
"loss": 0.3349, |
|
"step": 19175 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 9.625825825825826e-06, |
|
"loss": 0.335, |
|
"step": 19200 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 9.625325325325327e-06, |
|
"loss": 0.3371, |
|
"step": 19225 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 9.624824824824826e-06, |
|
"loss": 0.3475, |
|
"step": 19250 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 9.624324324324326e-06, |
|
"loss": 0.3428, |
|
"step": 19275 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 9.623823823823825e-06, |
|
"loss": 0.3375, |
|
"step": 19300 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 9.623323323323324e-06, |
|
"loss": 0.335, |
|
"step": 19325 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 9.622822822822823e-06, |
|
"loss": 0.3488, |
|
"step": 19350 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 9.622322322322323e-06, |
|
"loss": 0.3255, |
|
"step": 19375 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 9.621821821821822e-06, |
|
"loss": 0.3419, |
|
"step": 19400 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 9.621321321321321e-06, |
|
"loss": 0.3257, |
|
"step": 19425 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 9.620820820820822e-06, |
|
"loss": 0.337, |
|
"step": 19450 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 9.620320320320322e-06, |
|
"loss": 0.3159, |
|
"step": 19475 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 9.619819819819821e-06, |
|
"loss": 0.336, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 9.61931931931932e-06, |
|
"loss": 0.3218, |
|
"step": 19525 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 9.61881881881882e-06, |
|
"loss": 0.3447, |
|
"step": 19550 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 9.618318318318319e-06, |
|
"loss": 0.3395, |
|
"step": 19575 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 9.617817817817818e-06, |
|
"loss": 0.3237, |
|
"step": 19600 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 9.617317317317319e-06, |
|
"loss": 0.331, |
|
"step": 19625 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 9.616816816816817e-06, |
|
"loss": 0.3152, |
|
"step": 19650 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 9.616316316316318e-06, |
|
"loss": 0.3391, |
|
"step": 19675 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 9.615815815815817e-06, |
|
"loss": 0.3293, |
|
"step": 19700 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 9.615315315315316e-06, |
|
"loss": 0.3322, |
|
"step": 19725 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 9.614814814814816e-06, |
|
"loss": 0.3294, |
|
"step": 19750 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 9.614314314314315e-06, |
|
"loss": 0.334, |
|
"step": 19775 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 9.613813813813814e-06, |
|
"loss": 0.3418, |
|
"step": 19800 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 9.613313313313314e-06, |
|
"loss": 0.3456, |
|
"step": 19825 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 9.612812812812815e-06, |
|
"loss": 0.3513, |
|
"step": 19850 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 9.612312312312314e-06, |
|
"loss": 0.3395, |
|
"step": 19875 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 9.611811811811813e-06, |
|
"loss": 0.3478, |
|
"step": 19900 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 9.611311311311312e-06, |
|
"loss": 0.3278, |
|
"step": 19925 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 9.610810810810812e-06, |
|
"loss": 0.3012, |
|
"step": 19950 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 9.610310310310311e-06, |
|
"loss": 0.3453, |
|
"step": 19975 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 9.60980980980981e-06, |
|
"loss": 0.3453, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"eval_loss": 0.471629798412323, |
|
"eval_runtime": 2478.9552, |
|
"eval_samples_per_second": 6.54, |
|
"eval_steps_per_second": 0.164, |
|
"eval_wer": 27.84051926827576, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 9.60930930930931e-06, |
|
"loss": 0.3442, |
|
"step": 20025 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 9.608808808808809e-06, |
|
"loss": 0.3431, |
|
"step": 20050 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 9.60830830830831e-06, |
|
"loss": 0.3146, |
|
"step": 20075 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 9.60780780780781e-06, |
|
"loss": 0.3474, |
|
"step": 20100 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 9.607307307307307e-06, |
|
"loss": 0.3335, |
|
"step": 20125 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 9.606806806806808e-06, |
|
"loss": 0.3374, |
|
"step": 20150 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 9.606306306306307e-06, |
|
"loss": 0.3235, |
|
"step": 20175 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 9.605805805805806e-06, |
|
"loss": 0.339, |
|
"step": 20200 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 9.605305305305306e-06, |
|
"loss": 0.3468, |
|
"step": 20225 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 9.604804804804807e-06, |
|
"loss": 0.3587, |
|
"step": 20250 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 9.604304304304304e-06, |
|
"loss": 0.3113, |
|
"step": 20275 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 9.603803803803805e-06, |
|
"loss": 0.3224, |
|
"step": 20300 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 9.603303303303305e-06, |
|
"loss": 0.3298, |
|
"step": 20325 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 9.602802802802804e-06, |
|
"loss": 0.3489, |
|
"step": 20350 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 9.602302302302303e-06, |
|
"loss": 0.2943, |
|
"step": 20375 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 9.601801801801802e-06, |
|
"loss": 0.3283, |
|
"step": 20400 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 9.601301301301302e-06, |
|
"loss": 0.3167, |
|
"step": 20425 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 9.600800800800801e-06, |
|
"loss": 0.3427, |
|
"step": 20450 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 9.600300300300302e-06, |
|
"loss": 0.3538, |
|
"step": 20475 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 9.599799799799801e-06, |
|
"loss": 0.3353, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 9.5992992992993e-06, |
|
"loss": 0.3427, |
|
"step": 20525 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 9.5987987987988e-06, |
|
"loss": 0.3349, |
|
"step": 20550 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 9.5982982982983e-06, |
|
"loss": 0.334, |
|
"step": 20575 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 9.597797797797799e-06, |
|
"loss": 0.3281, |
|
"step": 20600 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 9.597297297297298e-06, |
|
"loss": 0.3231, |
|
"step": 20625 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 9.596796796796797e-06, |
|
"loss": 0.336, |
|
"step": 20650 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 9.596296296296296e-06, |
|
"loss": 0.3208, |
|
"step": 20675 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 9.595795795795797e-06, |
|
"loss": 0.3261, |
|
"step": 20700 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 9.595295295295297e-06, |
|
"loss": 0.3354, |
|
"step": 20725 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 9.594794794794794e-06, |
|
"loss": 0.3311, |
|
"step": 20750 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 9.594294294294295e-06, |
|
"loss": 0.3372, |
|
"step": 20775 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 9.593793793793795e-06, |
|
"loss": 0.3263, |
|
"step": 20800 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 9.593293293293294e-06, |
|
"loss": 0.3187, |
|
"step": 20825 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 9.592792792792793e-06, |
|
"loss": 0.317, |
|
"step": 20850 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 9.592292292292294e-06, |
|
"loss": 0.3428, |
|
"step": 20875 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 9.591791791791792e-06, |
|
"loss": 0.3049, |
|
"step": 20900 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 9.591291291291293e-06, |
|
"loss": 0.3125, |
|
"step": 20925 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 9.590790790790792e-06, |
|
"loss": 0.3206, |
|
"step": 20950 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 9.590290290290291e-06, |
|
"loss": 0.3515, |
|
"step": 20975 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 9.58978978978979e-06, |
|
"loss": 0.3412, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"eval_loss": 0.4714227020740509, |
|
"eval_runtime": 2637.8451, |
|
"eval_samples_per_second": 6.146, |
|
"eval_steps_per_second": 0.154, |
|
"eval_wer": 29.17954687367504, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 9.58928928928929e-06, |
|
"loss": 0.3176, |
|
"step": 21025 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 9.58878878878879e-06, |
|
"loss": 0.3355, |
|
"step": 21050 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 9.588288288288289e-06, |
|
"loss": 0.3306, |
|
"step": 21075 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 9.58778778778779e-06, |
|
"loss": 0.3532, |
|
"step": 21100 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 9.587287287287287e-06, |
|
"loss": 0.3311, |
|
"step": 21125 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 9.586786786786788e-06, |
|
"loss": 0.3377, |
|
"step": 21150 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 9.586286286286287e-06, |
|
"loss": 0.3171, |
|
"step": 21175 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 9.585785785785787e-06, |
|
"loss": 0.3454, |
|
"step": 21200 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 9.585285285285286e-06, |
|
"loss": 0.311, |
|
"step": 21225 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 9.584784784784785e-06, |
|
"loss": 0.3303, |
|
"step": 21250 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 9.584284284284285e-06, |
|
"loss": 0.3204, |
|
"step": 21275 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 9.583783783783784e-06, |
|
"loss": 0.3104, |
|
"step": 21300 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 9.583283283283285e-06, |
|
"loss": 0.3244, |
|
"step": 21325 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 9.582782782782784e-06, |
|
"loss": 0.3192, |
|
"step": 21350 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 9.582282282282282e-06, |
|
"loss": 0.3238, |
|
"step": 21375 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 9.581781781781783e-06, |
|
"loss": 0.3298, |
|
"step": 21400 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 9.581281281281282e-06, |
|
"loss": 0.3223, |
|
"step": 21425 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 9.580780780780781e-06, |
|
"loss": 0.3576, |
|
"step": 21450 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 9.58028028028028e-06, |
|
"loss": 0.305, |
|
"step": 21475 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 9.579779779779782e-06, |
|
"loss": 0.3287, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 9.57927927927928e-06, |
|
"loss": 0.3372, |
|
"step": 21525 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 9.57877877877878e-06, |
|
"loss": 0.3142, |
|
"step": 21550 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 9.57827827827828e-06, |
|
"loss": 0.3222, |
|
"step": 21575 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 9.577777777777779e-06, |
|
"loss": 0.3413, |
|
"step": 21600 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 9.577277277277278e-06, |
|
"loss": 0.3194, |
|
"step": 21625 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 9.576776776776777e-06, |
|
"loss": 0.3226, |
|
"step": 21650 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 9.576276276276277e-06, |
|
"loss": 0.3162, |
|
"step": 21675 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 9.575775775775776e-06, |
|
"loss": 0.3097, |
|
"step": 21700 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 9.575275275275277e-06, |
|
"loss": 0.3366, |
|
"step": 21725 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 9.574774774774775e-06, |
|
"loss": 0.322, |
|
"step": 21750 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 9.574274274274276e-06, |
|
"loss": 0.3407, |
|
"step": 21775 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 9.573773773773775e-06, |
|
"loss": 0.325, |
|
"step": 21800 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 9.573273273273274e-06, |
|
"loss": 0.319, |
|
"step": 21825 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 9.572772772772774e-06, |
|
"loss": 0.318, |
|
"step": 21850 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 9.572272272272273e-06, |
|
"loss": 0.3029, |
|
"step": 21875 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 9.571771771771772e-06, |
|
"loss": 0.3398, |
|
"step": 21900 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 9.571271271271271e-06, |
|
"loss": 0.345, |
|
"step": 21925 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 9.570770770770772e-06, |
|
"loss": 0.3366, |
|
"step": 21950 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 9.570270270270272e-06, |
|
"loss": 0.3192, |
|
"step": 21975 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 9.56976976976977e-06, |
|
"loss": 0.3221, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"eval_loss": 0.4648185670375824, |
|
"eval_runtime": 2509.1979, |
|
"eval_samples_per_second": 6.461, |
|
"eval_steps_per_second": 0.162, |
|
"eval_wer": 27.938421689626196, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 9.56926926926927e-06, |
|
"loss": 0.3333, |
|
"step": 22025 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 9.56876876876877e-06, |
|
"loss": 0.3309, |
|
"step": 22050 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 9.568268268268269e-06, |
|
"loss": 0.3252, |
|
"step": 22075 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 9.567767767767768e-06, |
|
"loss": 0.339, |
|
"step": 22100 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 9.56726726726727e-06, |
|
"loss": 0.3185, |
|
"step": 22125 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 9.566766766766767e-06, |
|
"loss": 0.3312, |
|
"step": 22150 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 9.566266266266268e-06, |
|
"loss": 0.3279, |
|
"step": 22175 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 9.565765765765767e-06, |
|
"loss": 0.3111, |
|
"step": 22200 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 9.565265265265266e-06, |
|
"loss": 0.3294, |
|
"step": 22225 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 9.564764764764766e-06, |
|
"loss": 0.3359, |
|
"step": 22250 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 9.564264264264265e-06, |
|
"loss": 0.3324, |
|
"step": 22275 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 9.563763763763764e-06, |
|
"loss": 0.3331, |
|
"step": 22300 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 9.563263263263264e-06, |
|
"loss": 0.3331, |
|
"step": 22325 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 9.562762762762765e-06, |
|
"loss": 0.307, |
|
"step": 22350 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 9.562262262262262e-06, |
|
"loss": 0.3384, |
|
"step": 22375 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 9.561761761761763e-06, |
|
"loss": 0.3291, |
|
"step": 22400 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 9.561261261261262e-06, |
|
"loss": 0.3411, |
|
"step": 22425 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 9.560760760760762e-06, |
|
"loss": 0.3141, |
|
"step": 22450 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 9.560260260260261e-06, |
|
"loss": 0.3289, |
|
"step": 22475 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 9.55975975975976e-06, |
|
"loss": 0.3289, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 9.55925925925926e-06, |
|
"loss": 0.2942, |
|
"step": 22525 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 9.558758758758759e-06, |
|
"loss": 0.3307, |
|
"step": 22550 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 9.55825825825826e-06, |
|
"loss": 0.3565, |
|
"step": 22575 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 9.55775775775776e-06, |
|
"loss": 0.3065, |
|
"step": 22600 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 9.557257257257257e-06, |
|
"loss": 0.3128, |
|
"step": 22625 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 9.556756756756758e-06, |
|
"loss": 0.3181, |
|
"step": 22650 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 9.556256256256257e-06, |
|
"loss": 0.321, |
|
"step": 22675 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 9.555755755755756e-06, |
|
"loss": 0.3305, |
|
"step": 22700 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 9.555255255255256e-06, |
|
"loss": 0.3454, |
|
"step": 22725 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 9.554754754754757e-06, |
|
"loss": 0.3236, |
|
"step": 22750 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 9.554254254254254e-06, |
|
"loss": 0.3345, |
|
"step": 22775 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 9.553753753753755e-06, |
|
"loss": 0.3221, |
|
"step": 22800 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 9.553253253253255e-06, |
|
"loss": 0.3333, |
|
"step": 22825 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 9.552752752752752e-06, |
|
"loss": 0.3176, |
|
"step": 22850 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 9.552252252252253e-06, |
|
"loss": 0.3286, |
|
"step": 22875 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 9.551751751751753e-06, |
|
"loss": 0.3165, |
|
"step": 22900 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 9.551251251251252e-06, |
|
"loss": 0.3312, |
|
"step": 22925 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 9.550750750750751e-06, |
|
"loss": 0.3334, |
|
"step": 22950 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 9.550250250250252e-06, |
|
"loss": 0.3092, |
|
"step": 22975 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 9.54974974974975e-06, |
|
"loss": 0.322, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"eval_loss": 0.46254613995552063, |
|
"eval_runtime": 2464.8795, |
|
"eval_samples_per_second": 6.578, |
|
"eval_steps_per_second": 0.165, |
|
"eval_wer": 27.390322307105247, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 9.54924924924925e-06, |
|
"loss": 0.3245, |
|
"step": 23025 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 9.54874874874875e-06, |
|
"loss": 0.3263, |
|
"step": 23050 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 9.54824824824825e-06, |
|
"loss": 0.3246, |
|
"step": 23075 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 9.547747747747749e-06, |
|
"loss": 0.3407, |
|
"step": 23100 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 9.547247247247248e-06, |
|
"loss": 0.3116, |
|
"step": 23125 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 9.546746746746747e-06, |
|
"loss": 0.3278, |
|
"step": 23150 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 9.546246246246246e-06, |
|
"loss": 0.3369, |
|
"step": 23175 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 9.545745745745747e-06, |
|
"loss": 0.3133, |
|
"step": 23200 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 9.545245245245247e-06, |
|
"loss": 0.3112, |
|
"step": 23225 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 9.544744744744744e-06, |
|
"loss": 0.3316, |
|
"step": 23250 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 9.544244244244245e-06, |
|
"loss": 0.3266, |
|
"step": 23275 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 9.543743743743745e-06, |
|
"loss": 0.3237, |
|
"step": 23300 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 9.543243243243244e-06, |
|
"loss": 0.3073, |
|
"step": 23325 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 9.542742742742743e-06, |
|
"loss": 0.3326, |
|
"step": 23350 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 9.542242242242243e-06, |
|
"loss": 0.3063, |
|
"step": 23375 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 9.541741741741742e-06, |
|
"loss": 0.3145, |
|
"step": 23400 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 9.541241241241243e-06, |
|
"loss": 0.3329, |
|
"step": 23425 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 9.540740740740742e-06, |
|
"loss": 0.3285, |
|
"step": 23450 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 9.54024024024024e-06, |
|
"loss": 0.3312, |
|
"step": 23475 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 9.53973973973974e-06, |
|
"loss": 0.3203, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 9.53923923923924e-06, |
|
"loss": 0.327, |
|
"step": 23525 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 9.53873873873874e-06, |
|
"loss": 0.3136, |
|
"step": 23550 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 9.538238238238239e-06, |
|
"loss": 0.314, |
|
"step": 23575 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 9.53773773773774e-06, |
|
"loss": 0.3294, |
|
"step": 23600 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 9.537237237237237e-06, |
|
"loss": 0.3331, |
|
"step": 23625 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 9.536736736736738e-06, |
|
"loss": 0.3112, |
|
"step": 23650 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 9.536236236236237e-06, |
|
"loss": 0.3047, |
|
"step": 23675 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 9.535735735735737e-06, |
|
"loss": 0.3153, |
|
"step": 23700 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 9.535235235235236e-06, |
|
"loss": 0.3195, |
|
"step": 23725 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 9.534734734734735e-06, |
|
"loss": 0.3187, |
|
"step": 23750 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 9.534234234234235e-06, |
|
"loss": 0.3351, |
|
"step": 23775 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 9.533733733733734e-06, |
|
"loss": 0.3172, |
|
"step": 23800 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 9.533233233233235e-06, |
|
"loss": 0.3206, |
|
"step": 23825 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 9.532752752752752e-06, |
|
"loss": 0.3261, |
|
"step": 23850 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 9.532252252252253e-06, |
|
"loss": 0.3204, |
|
"step": 23875 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 9.531751751751753e-06, |
|
"loss": 0.3154, |
|
"step": 23900 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 9.531251251251252e-06, |
|
"loss": 0.3159, |
|
"step": 23925 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 9.530750750750751e-06, |
|
"loss": 0.326, |
|
"step": 23950 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 9.53025025025025e-06, |
|
"loss": 0.3014, |
|
"step": 23975 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 9.52974974974975e-06, |
|
"loss": 0.312, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"eval_loss": 0.46919143199920654, |
|
"eval_runtime": 2551.3752, |
|
"eval_samples_per_second": 6.355, |
|
"eval_steps_per_second": 0.159, |
|
"eval_wer": 28.381680683929357, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 9.529249249249251e-06, |
|
"loss": 0.317, |
|
"step": 24025 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 9.52874874874875e-06, |
|
"loss": 0.3128, |
|
"step": 24050 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 9.528248248248248e-06, |
|
"loss": 0.324, |
|
"step": 24075 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 9.527747747747749e-06, |
|
"loss": 0.3367, |
|
"step": 24100 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 9.527247247247248e-06, |
|
"loss": 0.3159, |
|
"step": 24125 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 9.526746746746747e-06, |
|
"loss": 0.3288, |
|
"step": 24150 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 9.526246246246247e-06, |
|
"loss": 0.3379, |
|
"step": 24175 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 9.525745745745748e-06, |
|
"loss": 0.3359, |
|
"step": 24200 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 9.525245245245245e-06, |
|
"loss": 0.3146, |
|
"step": 24225 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 9.524744744744745e-06, |
|
"loss": 0.3054, |
|
"step": 24250 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 9.524244244244246e-06, |
|
"loss": 0.3329, |
|
"step": 24275 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 9.523743743743745e-06, |
|
"loss": 0.3157, |
|
"step": 24300 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 9.523243243243244e-06, |
|
"loss": 0.3252, |
|
"step": 24325 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 9.522742742742743e-06, |
|
"loss": 0.3199, |
|
"step": 24350 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 9.522242242242243e-06, |
|
"loss": 0.2962, |
|
"step": 24375 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 9.521741741741742e-06, |
|
"loss": 0.3197, |
|
"step": 24400 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 9.521241241241243e-06, |
|
"loss": 0.3105, |
|
"step": 24425 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 9.520740740740742e-06, |
|
"loss": 0.3158, |
|
"step": 24450 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 9.52024024024024e-06, |
|
"loss": 0.3084, |
|
"step": 24475 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 9.519739739739741e-06, |
|
"loss": 0.3258, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 9.51923923923924e-06, |
|
"loss": 0.3087, |
|
"step": 24525 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 9.51873873873874e-06, |
|
"loss": 0.323, |
|
"step": 24550 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 9.518238238238239e-06, |
|
"loss": 0.3153, |
|
"step": 24575 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 9.517737737737738e-06, |
|
"loss": 0.3256, |
|
"step": 24600 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 9.517237237237237e-06, |
|
"loss": 0.3461, |
|
"step": 24625 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 9.516736736736738e-06, |
|
"loss": 0.319, |
|
"step": 24650 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 9.516236236236238e-06, |
|
"loss": 0.3091, |
|
"step": 24675 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 9.515735735735735e-06, |
|
"loss": 0.3109, |
|
"step": 24700 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 9.515235235235236e-06, |
|
"loss": 0.3062, |
|
"step": 24725 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 9.514734734734736e-06, |
|
"loss": 0.3162, |
|
"step": 24750 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 9.514234234234235e-06, |
|
"loss": 0.3184, |
|
"step": 24775 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 9.513733733733734e-06, |
|
"loss": 0.3168, |
|
"step": 24800 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 9.513233233233235e-06, |
|
"loss": 0.315, |
|
"step": 24825 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 9.512732732732733e-06, |
|
"loss": 0.3023, |
|
"step": 24850 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 9.512232232232232e-06, |
|
"loss": 0.3242, |
|
"step": 24875 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 9.511731731731733e-06, |
|
"loss": 0.3337, |
|
"step": 24900 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 9.511231231231232e-06, |
|
"loss": 0.3149, |
|
"step": 24925 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 9.510730730730732e-06, |
|
"loss": 0.3085, |
|
"step": 24950 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 9.510230230230231e-06, |
|
"loss": 0.3263, |
|
"step": 24975 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 9.50972972972973e-06, |
|
"loss": 0.3089, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_loss": 0.457874059677124, |
|
"eval_runtime": 2679.5488, |
|
"eval_samples_per_second": 6.051, |
|
"eval_steps_per_second": 0.152, |
|
"eval_wer": 27.82818510495602, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 9.50922922922923e-06, |
|
"loss": 0.3246, |
|
"step": 25025 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 9.50872872872873e-06, |
|
"loss": 0.3371, |
|
"step": 25050 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 9.50822822822823e-06, |
|
"loss": 0.2992, |
|
"step": 25075 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 9.507727727727727e-06, |
|
"loss": 0.2763, |
|
"step": 25100 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 9.507227227227228e-06, |
|
"loss": 0.277, |
|
"step": 25125 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 9.506726726726728e-06, |
|
"loss": 0.2702, |
|
"step": 25150 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 9.506226226226227e-06, |
|
"loss": 0.2656, |
|
"step": 25175 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 9.505725725725726e-06, |
|
"loss": 0.2767, |
|
"step": 25200 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 9.505225225225226e-06, |
|
"loss": 0.2745, |
|
"step": 25225 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 9.504724724724725e-06, |
|
"loss": 0.2637, |
|
"step": 25250 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 9.504224224224226e-06, |
|
"loss": 0.2704, |
|
"step": 25275 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 9.503723723723725e-06, |
|
"loss": 0.2554, |
|
"step": 25300 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 9.503223223223223e-06, |
|
"loss": 0.271, |
|
"step": 25325 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 9.502722722722724e-06, |
|
"loss": 0.2757, |
|
"step": 25350 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 9.502222222222223e-06, |
|
"loss": 0.2606, |
|
"step": 25375 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 9.501721721721722e-06, |
|
"loss": 0.2664, |
|
"step": 25400 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 9.501221221221222e-06, |
|
"loss": 0.2702, |
|
"step": 25425 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 9.500720720720723e-06, |
|
"loss": 0.2633, |
|
"step": 25450 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 9.50022022022022e-06, |
|
"loss": 0.2816, |
|
"step": 25475 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 9.49971971971972e-06, |
|
"loss": 0.2848, |
|
"step": 25500 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 9.49921921921922e-06, |
|
"loss": 0.2655, |
|
"step": 25525 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 9.49871871871872e-06, |
|
"loss": 0.2652, |
|
"step": 25550 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 9.49821821821822e-06, |
|
"loss": 0.2704, |
|
"step": 25575 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 9.497717717717718e-06, |
|
"loss": 0.2605, |
|
"step": 25600 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 9.497217217217218e-06, |
|
"loss": 0.2674, |
|
"step": 25625 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 9.496716716716717e-06, |
|
"loss": 0.2677, |
|
"step": 25650 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 9.496216216216218e-06, |
|
"loss": 0.2761, |
|
"step": 25675 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 9.495715715715716e-06, |
|
"loss": 0.2626, |
|
"step": 25700 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 9.495215215215215e-06, |
|
"loss": 0.2637, |
|
"step": 25725 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 9.494714714714716e-06, |
|
"loss": 0.2704, |
|
"step": 25750 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 9.494214214214215e-06, |
|
"loss": 0.2756, |
|
"step": 25775 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 9.493713713713715e-06, |
|
"loss": 0.2807, |
|
"step": 25800 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 9.493213213213214e-06, |
|
"loss": 0.2798, |
|
"step": 25825 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 9.492712712712713e-06, |
|
"loss": 0.2728, |
|
"step": 25850 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 9.492212212212212e-06, |
|
"loss": 0.2664, |
|
"step": 25875 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 9.491731731731733e-06, |
|
"loss": 0.2811, |
|
"step": 25900 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 9.491231231231231e-06, |
|
"loss": 0.2582, |
|
"step": 25925 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 9.490730730730732e-06, |
|
"loss": 0.2739, |
|
"step": 25950 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 9.490230230230231e-06, |
|
"loss": 0.264, |
|
"step": 25975 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 9.48972972972973e-06, |
|
"loss": 0.2625, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"eval_loss": 0.45843926072120667, |
|
"eval_runtime": 2831.694, |
|
"eval_samples_per_second": 5.726, |
|
"eval_steps_per_second": 0.143, |
|
"eval_wer": 27.744929502547777, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 9.48922922922923e-06, |
|
"loss": 0.2841, |
|
"step": 26025 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 9.48872872872873e-06, |
|
"loss": 0.2881, |
|
"step": 26050 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 9.488228228228228e-06, |
|
"loss": 0.2915, |
|
"step": 26075 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 9.487727727727728e-06, |
|
"loss": 0.2687, |
|
"step": 26100 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 9.487227227227229e-06, |
|
"loss": 0.2678, |
|
"step": 26125 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 9.486726726726728e-06, |
|
"loss": 0.2724, |
|
"step": 26150 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 9.486226226226227e-06, |
|
"loss": 0.2792, |
|
"step": 26175 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 9.485725725725727e-06, |
|
"loss": 0.2657, |
|
"step": 26200 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 9.485225225225226e-06, |
|
"loss": 0.2722, |
|
"step": 26225 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 9.484724724724725e-06, |
|
"loss": 0.2704, |
|
"step": 26250 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 9.484224224224226e-06, |
|
"loss": 0.2595, |
|
"step": 26275 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 9.483723723723724e-06, |
|
"loss": 0.2699, |
|
"step": 26300 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 9.483223223223223e-06, |
|
"loss": 0.2857, |
|
"step": 26325 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 9.482722722722724e-06, |
|
"loss": 0.2664, |
|
"step": 26350 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 9.482222222222223e-06, |
|
"loss": 0.2606, |
|
"step": 26375 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 9.481721721721723e-06, |
|
"loss": 0.2607, |
|
"step": 26400 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 9.481221221221222e-06, |
|
"loss": 0.2487, |
|
"step": 26425 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 9.480720720720721e-06, |
|
"loss": 0.2603, |
|
"step": 26450 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 9.48022022022022e-06, |
|
"loss": 0.2819, |
|
"step": 26475 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 9.479719719719721e-06, |
|
"loss": 0.27, |
|
"step": 26500 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 9.47921921921922e-06, |
|
"loss": 0.2577, |
|
"step": 26525 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 9.478718718718718e-06, |
|
"loss": 0.2743, |
|
"step": 26550 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 9.47821821821822e-06, |
|
"loss": 0.2806, |
|
"step": 26575 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 9.477717717717719e-06, |
|
"loss": 0.2736, |
|
"step": 26600 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 9.477217217217218e-06, |
|
"loss": 0.2699, |
|
"step": 26625 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 9.476716716716717e-06, |
|
"loss": 0.2743, |
|
"step": 26650 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 9.476216216216218e-06, |
|
"loss": 0.2706, |
|
"step": 26675 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 9.475715715715716e-06, |
|
"loss": 0.2658, |
|
"step": 26700 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 9.475215215215215e-06, |
|
"loss": 0.263, |
|
"step": 26725 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 9.474714714714716e-06, |
|
"loss": 0.2653, |
|
"step": 26750 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 9.474214214214215e-06, |
|
"loss": 0.2622, |
|
"step": 26775 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 9.473713713713715e-06, |
|
"loss": 0.2743, |
|
"step": 26800 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 9.473213213213214e-06, |
|
"loss": 0.2531, |
|
"step": 26825 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 9.472712712712713e-06, |
|
"loss": 0.2686, |
|
"step": 26850 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 9.472212212212213e-06, |
|
"loss": 0.2847, |
|
"step": 26875 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 9.471711711711714e-06, |
|
"loss": 0.2556, |
|
"step": 26900 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 9.471211211211211e-06, |
|
"loss": 0.2642, |
|
"step": 26925 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 9.47071071071071e-06, |
|
"loss": 0.2719, |
|
"step": 26950 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 9.470210210210212e-06, |
|
"loss": 0.2632, |
|
"step": 26975 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 9.46970970970971e-06, |
|
"loss": 0.2758, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"eval_loss": 0.45501989126205444, |
|
"eval_runtime": 2547.6663, |
|
"eval_samples_per_second": 6.364, |
|
"eval_steps_per_second": 0.159, |
|
"eval_wer": 26.85301531748907, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 9.46920920920921e-06, |
|
"loss": 0.2766, |
|
"step": 27025 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 9.46870870870871e-06, |
|
"loss": 0.2724, |
|
"step": 27050 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 9.468208208208209e-06, |
|
"loss": 0.265, |
|
"step": 27075 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 9.467707707707708e-06, |
|
"loss": 0.2627, |
|
"step": 27100 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 9.467207207207207e-06, |
|
"loss": 0.2648, |
|
"step": 27125 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 9.466706706706708e-06, |
|
"loss": 0.2676, |
|
"step": 27150 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 9.466206206206206e-06, |
|
"loss": 0.2634, |
|
"step": 27175 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 9.465705705705707e-06, |
|
"loss": 0.2631, |
|
"step": 27200 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 9.465205205205206e-06, |
|
"loss": 0.2729, |
|
"step": 27225 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 9.464704704704705e-06, |
|
"loss": 0.2772, |
|
"step": 27250 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 9.464204204204205e-06, |
|
"loss": 0.2605, |
|
"step": 27275 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 9.463703703703706e-06, |
|
"loss": 0.2753, |
|
"step": 27300 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 9.463203203203203e-06, |
|
"loss": 0.2719, |
|
"step": 27325 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 9.462702702702703e-06, |
|
"loss": 0.2802, |
|
"step": 27350 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 9.462202202202204e-06, |
|
"loss": 0.2579, |
|
"step": 27375 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 9.461701701701701e-06, |
|
"loss": 0.2781, |
|
"step": 27400 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 9.461201201201202e-06, |
|
"loss": 0.292, |
|
"step": 27425 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 9.460700700700702e-06, |
|
"loss": 0.25, |
|
"step": 27450 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 9.460200200200201e-06, |
|
"loss": 0.2518, |
|
"step": 27475 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 9.4596996996997e-06, |
|
"loss": 0.264, |
|
"step": 27500 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 9.459199199199201e-06, |
|
"loss": 0.2603, |
|
"step": 27525 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 9.458698698698699e-06, |
|
"loss": 0.2601, |
|
"step": 27550 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 9.458198198198198e-06, |
|
"loss": 0.275, |
|
"step": 27575 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 9.457697697697699e-06, |
|
"loss": 0.2681, |
|
"step": 27600 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 9.457197197197198e-06, |
|
"loss": 0.2595, |
|
"step": 27625 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 9.456696696696698e-06, |
|
"loss": 0.2797, |
|
"step": 27650 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 9.456196196196197e-06, |
|
"loss": 0.2564, |
|
"step": 27675 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 9.455695695695696e-06, |
|
"loss": 0.2565, |
|
"step": 27700 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 9.455195195195196e-06, |
|
"loss": 0.2833, |
|
"step": 27725 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 9.454694694694695e-06, |
|
"loss": 0.2648, |
|
"step": 27750 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 9.454194194194196e-06, |
|
"loss": 0.2607, |
|
"step": 27775 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 9.453693693693693e-06, |
|
"loss": 0.2697, |
|
"step": 27800 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 9.453193193193194e-06, |
|
"loss": 0.2778, |
|
"step": 27825 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 9.452692692692694e-06, |
|
"loss": 0.2616, |
|
"step": 27850 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 9.452192192192193e-06, |
|
"loss": 0.27, |
|
"step": 27875 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 9.451691691691692e-06, |
|
"loss": 0.2726, |
|
"step": 27900 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 9.451191191191193e-06, |
|
"loss": 0.2444, |
|
"step": 27925 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 9.450690690690691e-06, |
|
"loss": 0.258, |
|
"step": 27950 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 9.45019019019019e-06, |
|
"loss": 0.2758, |
|
"step": 27975 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 9.449689689689691e-06, |
|
"loss": 0.265, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"eval_loss": 0.45008718967437744, |
|
"eval_runtime": 2652.1712, |
|
"eval_samples_per_second": 6.113, |
|
"eval_steps_per_second": 0.153, |
|
"eval_wer": 28.59213234557242, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 9.44920920920921e-06, |
|
"loss": 0.2703, |
|
"step": 28025 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 9.44870870870871e-06, |
|
"loss": 0.2573, |
|
"step": 28050 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 9.448208208208209e-06, |
|
"loss": 0.2824, |
|
"step": 28075 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 9.447707707707708e-06, |
|
"loss": 0.2697, |
|
"step": 28100 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 9.44720720720721e-06, |
|
"loss": 0.2586, |
|
"step": 28125 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 9.446706706706707e-06, |
|
"loss": 0.2882, |
|
"step": 28150 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 9.446206206206206e-06, |
|
"loss": 0.2649, |
|
"step": 28175 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 9.445705705705707e-06, |
|
"loss": 0.2542, |
|
"step": 28200 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 9.445205205205206e-06, |
|
"loss": 0.263, |
|
"step": 28225 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 9.444704704704706e-06, |
|
"loss": 0.2729, |
|
"step": 28250 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 9.444204204204205e-06, |
|
"loss": 0.2861, |
|
"step": 28275 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 9.443703703703704e-06, |
|
"loss": 0.2692, |
|
"step": 28300 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 9.443203203203204e-06, |
|
"loss": 0.2826, |
|
"step": 28325 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 9.442702702702703e-06, |
|
"loss": 0.2795, |
|
"step": 28350 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 9.442202202202204e-06, |
|
"loss": 0.2655, |
|
"step": 28375 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 9.441701701701701e-06, |
|
"loss": 0.2605, |
|
"step": 28400 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 9.441201201201202e-06, |
|
"loss": 0.2626, |
|
"step": 28425 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 9.440700700700702e-06, |
|
"loss": 0.2867, |
|
"step": 28450 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 9.440200200200201e-06, |
|
"loss": 0.272, |
|
"step": 28475 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 9.4396996996997e-06, |
|
"loss": 0.257, |
|
"step": 28500 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 9.4391991991992e-06, |
|
"loss": 0.2783, |
|
"step": 28525 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 9.438698698698699e-06, |
|
"loss": 0.2602, |
|
"step": 28550 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 9.438198198198198e-06, |
|
"loss": 0.2644, |
|
"step": 28575 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 9.4376976976977e-06, |
|
"loss": 0.2554, |
|
"step": 28600 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 9.437197197197197e-06, |
|
"loss": 0.2743, |
|
"step": 28625 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 9.436696696696698e-06, |
|
"loss": 0.2724, |
|
"step": 28650 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 9.436196196196197e-06, |
|
"loss": 0.2607, |
|
"step": 28675 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 9.435695695695696e-06, |
|
"loss": 0.2626, |
|
"step": 28700 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 9.435195195195196e-06, |
|
"loss": 0.2616, |
|
"step": 28725 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 9.434694694694697e-06, |
|
"loss": 0.2602, |
|
"step": 28750 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 9.434194194194194e-06, |
|
"loss": 0.2639, |
|
"step": 28775 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 9.433693693693694e-06, |
|
"loss": 0.2564, |
|
"step": 28800 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 9.433193193193195e-06, |
|
"loss": 0.2544, |
|
"step": 28825 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 9.432692692692694e-06, |
|
"loss": 0.2874, |
|
"step": 28850 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 9.432192192192193e-06, |
|
"loss": 0.2663, |
|
"step": 28875 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 9.431691691691692e-06, |
|
"loss": 0.2658, |
|
"step": 28900 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 9.431191191191192e-06, |
|
"loss": 0.286, |
|
"step": 28925 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 9.430690690690691e-06, |
|
"loss": 0.2451, |
|
"step": 28950 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 9.43019019019019e-06, |
|
"loss": 0.2811, |
|
"step": 28975 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 9.429689689689691e-06, |
|
"loss": 0.258, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"eval_loss": 0.44391068816185, |
|
"eval_runtime": 2504.9501, |
|
"eval_samples_per_second": 6.472, |
|
"eval_steps_per_second": 0.162, |
|
"eval_wer": 26.30877036100554, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 9.429189189189189e-06, |
|
"loss": 0.2654, |
|
"step": 29025 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 9.42868868868869e-06, |
|
"loss": 0.2587, |
|
"step": 29050 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 9.42818818818819e-06, |
|
"loss": 0.2571, |
|
"step": 29075 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 9.427687687687689e-06, |
|
"loss": 0.2734, |
|
"step": 29100 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 9.427187187187188e-06, |
|
"loss": 0.2898, |
|
"step": 29125 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 9.426686686686687e-06, |
|
"loss": 0.2514, |
|
"step": 29150 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 9.426186186186186e-06, |
|
"loss": 0.2804, |
|
"step": 29175 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 9.425685685685686e-06, |
|
"loss": 0.2812, |
|
"step": 29200 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 9.425185185185187e-06, |
|
"loss": 0.265, |
|
"step": 29225 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 9.424684684684684e-06, |
|
"loss": 0.2546, |
|
"step": 29250 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 9.424184184184185e-06, |
|
"loss": 0.2818, |
|
"step": 29275 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 9.423683683683685e-06, |
|
"loss": 0.2448, |
|
"step": 29300 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 9.423183183183184e-06, |
|
"loss": 0.2791, |
|
"step": 29325 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 9.422682682682683e-06, |
|
"loss": 0.2583, |
|
"step": 29350 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 9.422182182182184e-06, |
|
"loss": 0.2737, |
|
"step": 29375 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 9.421701701701702e-06, |
|
"loss": 0.266, |
|
"step": 29400 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 9.421201201201203e-06, |
|
"loss": 0.265, |
|
"step": 29425 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 9.420700700700702e-06, |
|
"loss": 0.2771, |
|
"step": 29450 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 9.420200200200201e-06, |
|
"loss": 0.2569, |
|
"step": 29475 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 9.4196996996997e-06, |
|
"loss": 0.2621, |
|
"step": 29500 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 9.4191991991992e-06, |
|
"loss": 0.2703, |
|
"step": 29525 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 9.418698698698699e-06, |
|
"loss": 0.2786, |
|
"step": 29550 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 9.418198198198198e-06, |
|
"loss": 0.2628, |
|
"step": 29575 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 9.4176976976977e-06, |
|
"loss": 0.295, |
|
"step": 29600 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 9.417197197197197e-06, |
|
"loss": 0.2829, |
|
"step": 29625 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 9.416696696696698e-06, |
|
"loss": 0.2721, |
|
"step": 29650 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 9.416196196196197e-06, |
|
"loss": 0.2721, |
|
"step": 29675 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 9.415695695695697e-06, |
|
"loss": 0.2683, |
|
"step": 29700 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 9.415195195195196e-06, |
|
"loss": 0.2655, |
|
"step": 29725 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 9.414694694694695e-06, |
|
"loss": 0.2567, |
|
"step": 29750 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 9.414194194194194e-06, |
|
"loss": 0.2542, |
|
"step": 29775 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 9.413693693693694e-06, |
|
"loss": 0.2722, |
|
"step": 29800 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 9.413193193193195e-06, |
|
"loss": 0.2707, |
|
"step": 29825 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 9.412692692692692e-06, |
|
"loss": 0.2718, |
|
"step": 29850 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 9.412192192192193e-06, |
|
"loss": 0.2634, |
|
"step": 29875 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 9.411691691691693e-06, |
|
"loss": 0.2643, |
|
"step": 29900 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 9.411191191191192e-06, |
|
"loss": 0.2618, |
|
"step": 29925 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 9.410690690690691e-06, |
|
"loss": 0.2684, |
|
"step": 29950 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 9.41019019019019e-06, |
|
"loss": 0.282, |
|
"step": 29975 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 9.40968968968969e-06, |
|
"loss": 0.2837, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"eval_loss": 0.4546422064304352, |
|
"eval_runtime": 2695.5326, |
|
"eval_samples_per_second": 6.015, |
|
"eval_steps_per_second": 0.151, |
|
"eval_wer": 27.167536482142445, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 9.40918918918919e-06, |
|
"loss": 0.2669, |
|
"step": 30025 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 9.40868868868869e-06, |
|
"loss": 0.2881, |
|
"step": 30050 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 9.40818818818819e-06, |
|
"loss": 0.2759, |
|
"step": 30075 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 9.407687687687689e-06, |
|
"loss": 0.2568, |
|
"step": 30100 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 9.407187187187188e-06, |
|
"loss": 0.2779, |
|
"step": 30125 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 9.406686686686687e-06, |
|
"loss": 0.2668, |
|
"step": 30150 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 9.406186186186187e-06, |
|
"loss": 0.272, |
|
"step": 30175 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 9.405685685685686e-06, |
|
"loss": 0.2638, |
|
"step": 30200 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 9.405185185185187e-06, |
|
"loss": 0.27, |
|
"step": 30225 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 9.404684684684685e-06, |
|
"loss": 0.2757, |
|
"step": 30250 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 9.404184184184186e-06, |
|
"loss": 0.2499, |
|
"step": 30275 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 9.403683683683685e-06, |
|
"loss": 0.2886, |
|
"step": 30300 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 9.403183183183184e-06, |
|
"loss": 0.2756, |
|
"step": 30325 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 9.402682682682683e-06, |
|
"loss": 0.261, |
|
"step": 30350 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 9.402182182182183e-06, |
|
"loss": 0.2748, |
|
"step": 30375 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 9.401681681681682e-06, |
|
"loss": 0.261, |
|
"step": 30400 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 9.401181181181181e-06, |
|
"loss": 0.2839, |
|
"step": 30425 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 9.400680680680682e-06, |
|
"loss": 0.2735, |
|
"step": 30450 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 9.40018018018018e-06, |
|
"loss": 0.2659, |
|
"step": 30475 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 9.3996996996997e-06, |
|
"loss": 0.2621, |
|
"step": 30500 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 9.3991991991992e-06, |
|
"loss": 0.2744, |
|
"step": 30525 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 9.3986986986987e-06, |
|
"loss": 0.2632, |
|
"step": 30550 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 9.398198198198199e-06, |
|
"loss": 0.2589, |
|
"step": 30575 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 9.397697697697698e-06, |
|
"loss": 0.2528, |
|
"step": 30600 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 9.397197197197197e-06, |
|
"loss": 0.265, |
|
"step": 30625 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 9.396696696696698e-06, |
|
"loss": 0.2569, |
|
"step": 30650 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 9.396196196196197e-06, |
|
"loss": 0.2646, |
|
"step": 30675 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 9.395695695695697e-06, |
|
"loss": 0.2737, |
|
"step": 30700 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 9.395195195195196e-06, |
|
"loss": 0.2555, |
|
"step": 30725 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 9.394694694694695e-06, |
|
"loss": 0.2671, |
|
"step": 30750 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 9.394194194194195e-06, |
|
"loss": 0.2693, |
|
"step": 30775 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 9.393693693693694e-06, |
|
"loss": 0.2795, |
|
"step": 30800 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 9.393193193193193e-06, |
|
"loss": 0.2641, |
|
"step": 30825 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 9.392692692692693e-06, |
|
"loss": 0.2789, |
|
"step": 30850 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 9.392192192192194e-06, |
|
"loss": 0.2617, |
|
"step": 30875 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 9.391691691691693e-06, |
|
"loss": 0.2536, |
|
"step": 30900 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 9.391191191191192e-06, |
|
"loss": 0.2636, |
|
"step": 30925 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 9.390690690690691e-06, |
|
"loss": 0.2713, |
|
"step": 30950 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 9.39019019019019e-06, |
|
"loss": 0.2741, |
|
"step": 30975 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 9.38968968968969e-06, |
|
"loss": 0.2804, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"eval_loss": 0.44737279415130615, |
|
"eval_runtime": 2512.7508, |
|
"eval_samples_per_second": 6.452, |
|
"eval_steps_per_second": 0.162, |
|
"eval_wer": 26.34577285096476, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 9.38918918918919e-06, |
|
"loss": 0.277, |
|
"step": 31025 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 9.38868868868869e-06, |
|
"loss": 0.2563, |
|
"step": 31050 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 9.388188188188188e-06, |
|
"loss": 0.2455, |
|
"step": 31075 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 9.387687687687689e-06, |
|
"loss": 0.2606, |
|
"step": 31100 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 9.387187187187188e-06, |
|
"loss": 0.2602, |
|
"step": 31125 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 9.386686686686688e-06, |
|
"loss": 0.2576, |
|
"step": 31150 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 9.386186186186187e-06, |
|
"loss": 0.27, |
|
"step": 31175 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 9.385685685685686e-06, |
|
"loss": 0.2806, |
|
"step": 31200 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 9.385185185185185e-06, |
|
"loss": 0.2673, |
|
"step": 31225 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 9.384684684684685e-06, |
|
"loss": 0.2766, |
|
"step": 31250 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 9.384184184184186e-06, |
|
"loss": 0.2575, |
|
"step": 31275 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 9.383683683683685e-06, |
|
"loss": 0.2526, |
|
"step": 31300 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 9.383183183183184e-06, |
|
"loss": 0.2678, |
|
"step": 31325 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 9.382682682682684e-06, |
|
"loss": 0.2711, |
|
"step": 31350 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 9.382182182182183e-06, |
|
"loss": 0.2889, |
|
"step": 31375 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 9.381681681681682e-06, |
|
"loss": 0.2709, |
|
"step": 31400 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 9.381181181181182e-06, |
|
"loss": 0.2736, |
|
"step": 31425 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 9.38068068068068e-06, |
|
"loss": 0.2699, |
|
"step": 31450 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 9.38018018018018e-06, |
|
"loss": 0.2801, |
|
"step": 31475 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 9.379679679679681e-06, |
|
"loss": 0.2774, |
|
"step": 31500 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 9.37917917917918e-06, |
|
"loss": 0.2658, |
|
"step": 31525 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 9.37867867867868e-06, |
|
"loss": 0.2663, |
|
"step": 31550 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 9.378178178178179e-06, |
|
"loss": 0.2673, |
|
"step": 31575 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 9.377677677677678e-06, |
|
"loss": 0.2722, |
|
"step": 31600 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 9.377177177177178e-06, |
|
"loss": 0.2818, |
|
"step": 31625 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 9.376676676676677e-06, |
|
"loss": 0.2689, |
|
"step": 31650 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 9.376176176176178e-06, |
|
"loss": 0.2755, |
|
"step": 31675 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 9.375675675675675e-06, |
|
"loss": 0.2563, |
|
"step": 31700 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 9.375175175175176e-06, |
|
"loss": 0.2691, |
|
"step": 31725 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 9.374674674674676e-06, |
|
"loss": 0.2775, |
|
"step": 31750 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 9.374174174174175e-06, |
|
"loss": 0.2625, |
|
"step": 31775 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 9.373673673673674e-06, |
|
"loss": 0.2526, |
|
"step": 31800 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 9.373173173173174e-06, |
|
"loss": 0.2624, |
|
"step": 31825 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 9.372672672672673e-06, |
|
"loss": 0.2823, |
|
"step": 31850 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 9.372172172172172e-06, |
|
"loss": 0.2637, |
|
"step": 31875 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 9.371671671671673e-06, |
|
"loss": 0.2747, |
|
"step": 31900 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 9.371171171171173e-06, |
|
"loss": 0.2795, |
|
"step": 31925 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 9.370670670670672e-06, |
|
"loss": 0.2643, |
|
"step": 31950 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 9.370170170170171e-06, |
|
"loss": 0.2799, |
|
"step": 31975 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 9.36966966966967e-06, |
|
"loss": 0.2612, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"eval_loss": 0.4532637298107147, |
|
"eval_runtime": 2555.128, |
|
"eval_samples_per_second": 6.345, |
|
"eval_steps_per_second": 0.159, |
|
"eval_wer": 25.984998573862367, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 9.36916916916917e-06, |
|
"loss": 0.2583, |
|
"step": 32025 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 9.368668668668669e-06, |
|
"loss": 0.257, |
|
"step": 32050 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 9.368168168168168e-06, |
|
"loss": 0.2726, |
|
"step": 32075 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 9.367667667667668e-06, |
|
"loss": 0.2675, |
|
"step": 32100 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 9.367167167167169e-06, |
|
"loss": 0.2606, |
|
"step": 32125 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 9.366666666666668e-06, |
|
"loss": 0.2672, |
|
"step": 32150 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 9.366166166166167e-06, |
|
"loss": 0.2622, |
|
"step": 32175 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 9.365665665665666e-06, |
|
"loss": 0.2592, |
|
"step": 32200 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 9.365165165165166e-06, |
|
"loss": 0.2618, |
|
"step": 32225 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 9.364664664664665e-06, |
|
"loss": 0.2669, |
|
"step": 32250 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 9.364164164164164e-06, |
|
"loss": 0.2725, |
|
"step": 32275 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 9.363663663663665e-06, |
|
"loss": 0.2695, |
|
"step": 32300 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 9.363163163163163e-06, |
|
"loss": 0.2727, |
|
"step": 32325 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 9.362662662662664e-06, |
|
"loss": 0.28, |
|
"step": 32350 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 9.362162162162163e-06, |
|
"loss": 0.2826, |
|
"step": 32375 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 9.361661661661663e-06, |
|
"loss": 0.2728, |
|
"step": 32400 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 9.361161161161162e-06, |
|
"loss": 0.2653, |
|
"step": 32425 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 9.360660660660661e-06, |
|
"loss": 0.2655, |
|
"step": 32450 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 9.36016016016016e-06, |
|
"loss": 0.2787, |
|
"step": 32475 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 9.35965965965966e-06, |
|
"loss": 0.2677, |
|
"step": 32500 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 9.35915915915916e-06, |
|
"loss": 0.2587, |
|
"step": 32525 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 9.358658658658658e-06, |
|
"loss": 0.2685, |
|
"step": 32550 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 9.35815815815816e-06, |
|
"loss": 0.2583, |
|
"step": 32575 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 9.357657657657659e-06, |
|
"loss": 0.2673, |
|
"step": 32600 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 9.357157157157158e-06, |
|
"loss": 0.2828, |
|
"step": 32625 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 9.356656656656657e-06, |
|
"loss": 0.2753, |
|
"step": 32650 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 9.356156156156157e-06, |
|
"loss": 0.2847, |
|
"step": 32675 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 9.355655655655656e-06, |
|
"loss": 0.2671, |
|
"step": 32700 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 9.355155155155155e-06, |
|
"loss": 0.2655, |
|
"step": 32725 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 9.354654654654656e-06, |
|
"loss": 0.2602, |
|
"step": 32750 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 9.354154154154155e-06, |
|
"loss": 0.2707, |
|
"step": 32775 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 9.353653653653655e-06, |
|
"loss": 0.2577, |
|
"step": 32800 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 9.353153153153154e-06, |
|
"loss": 0.2786, |
|
"step": 32825 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 9.352652652652653e-06, |
|
"loss": 0.2529, |
|
"step": 32850 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 9.352152152152153e-06, |
|
"loss": 0.2544, |
|
"step": 32875 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 9.351651651651652e-06, |
|
"loss": 0.2748, |
|
"step": 32900 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 9.351151151151153e-06, |
|
"loss": 0.2696, |
|
"step": 32925 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 9.35065065065065e-06, |
|
"loss": 0.2646, |
|
"step": 32950 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 9.350150150150151e-06, |
|
"loss": 0.2598, |
|
"step": 32975 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 9.34964964964965e-06, |
|
"loss": 0.2551, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"eval_loss": 0.4496789276599884, |
|
"eval_runtime": 2555.1297, |
|
"eval_samples_per_second": 6.345, |
|
"eval_steps_per_second": 0.159, |
|
"eval_wer": 26.183886957393177, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 9.34914914914915e-06, |
|
"loss": 0.2877, |
|
"step": 33025 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 9.34864864864865e-06, |
|
"loss": 0.2536, |
|
"step": 33050 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 9.348148148148149e-06, |
|
"loss": 0.2592, |
|
"step": 33075 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 9.347647647647648e-06, |
|
"loss": 0.2691, |
|
"step": 33100 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 9.347147147147147e-06, |
|
"loss": 0.269, |
|
"step": 33125 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 9.346646646646648e-06, |
|
"loss": 0.2585, |
|
"step": 33150 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 9.346146146146146e-06, |
|
"loss": 0.2547, |
|
"step": 33175 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 9.345645645645647e-06, |
|
"loss": 0.278, |
|
"step": 33200 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 9.345145145145146e-06, |
|
"loss": 0.283, |
|
"step": 33225 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 9.344644644644645e-06, |
|
"loss": 0.2708, |
|
"step": 33250 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 9.344144144144145e-06, |
|
"loss": 0.2549, |
|
"step": 33275 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 9.343643643643644e-06, |
|
"loss": 0.2601, |
|
"step": 33300 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 9.343143143143143e-06, |
|
"loss": 0.2585, |
|
"step": 33325 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 9.342642642642643e-06, |
|
"loss": 0.2794, |
|
"step": 33350 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 9.342142142142144e-06, |
|
"loss": 0.2721, |
|
"step": 33375 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 9.341641641641643e-06, |
|
"loss": 0.278, |
|
"step": 33400 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 9.341141141141142e-06, |
|
"loss": 0.2702, |
|
"step": 33425 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 9.340640640640642e-06, |
|
"loss": 0.2577, |
|
"step": 33450 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 9.34014014014014e-06, |
|
"loss": 0.2706, |
|
"step": 33475 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 9.33963963963964e-06, |
|
"loss": 0.25, |
|
"step": 33500 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 9.33913913913914e-06, |
|
"loss": 0.2774, |
|
"step": 33525 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 9.33863863863864e-06, |
|
"loss": 0.2676, |
|
"step": 33550 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 9.338138138138138e-06, |
|
"loss": 0.2776, |
|
"step": 33575 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 9.337637637637639e-06, |
|
"loss": 0.2542, |
|
"step": 33600 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 9.337137137137138e-06, |
|
"loss": 0.273, |
|
"step": 33625 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 9.336636636636638e-06, |
|
"loss": 0.2498, |
|
"step": 33650 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 9.336136136136137e-06, |
|
"loss": 0.2624, |
|
"step": 33675 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 9.335635635635636e-06, |
|
"loss": 0.255, |
|
"step": 33700 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 9.335135135135135e-06, |
|
"loss": 0.2639, |
|
"step": 33725 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 9.334634634634635e-06, |
|
"loss": 0.281, |
|
"step": 33750 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 9.334134134134136e-06, |
|
"loss": 0.2733, |
|
"step": 33775 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 9.333633633633633e-06, |
|
"loss": 0.2613, |
|
"step": 33800 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 9.333133133133134e-06, |
|
"loss": 0.273, |
|
"step": 33825 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 9.332632632632634e-06, |
|
"loss": 0.264, |
|
"step": 33850 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 9.332132132132133e-06, |
|
"loss": 0.2665, |
|
"step": 33875 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 9.331631631631632e-06, |
|
"loss": 0.264, |
|
"step": 33900 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 9.331131131131132e-06, |
|
"loss": 0.2686, |
|
"step": 33925 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 9.330630630630631e-06, |
|
"loss": 0.2709, |
|
"step": 33950 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 9.33013013013013e-06, |
|
"loss": 0.2731, |
|
"step": 33975 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 9.329629629629631e-06, |
|
"loss": 0.28, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"eval_loss": 0.44749680161476135, |
|
"eval_runtime": 2532.2254, |
|
"eval_samples_per_second": 6.403, |
|
"eval_steps_per_second": 0.16, |
|
"eval_wer": 26.161531286376142, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 9.32912912912913e-06, |
|
"loss": 0.2572, |
|
"step": 34025 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 9.32862862862863e-06, |
|
"loss": 0.2828, |
|
"step": 34050 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 9.328128128128129e-06, |
|
"loss": 0.268, |
|
"step": 34075 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 9.327627627627628e-06, |
|
"loss": 0.2602, |
|
"step": 34100 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 9.327127127127128e-06, |
|
"loss": 0.2651, |
|
"step": 34125 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 9.326626626626627e-06, |
|
"loss": 0.2796, |
|
"step": 34150 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 9.326126126126128e-06, |
|
"loss": 0.2649, |
|
"step": 34175 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 9.325625625625626e-06, |
|
"loss": 0.2729, |
|
"step": 34200 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 9.325125125125127e-06, |
|
"loss": 0.2492, |
|
"step": 34225 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 9.324624624624626e-06, |
|
"loss": 0.2389, |
|
"step": 34250 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 9.324124124124125e-06, |
|
"loss": 0.2584, |
|
"step": 34275 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 9.323623623623624e-06, |
|
"loss": 0.2635, |
|
"step": 34300 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 9.323123123123124e-06, |
|
"loss": 0.2556, |
|
"step": 34325 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 9.322622622622623e-06, |
|
"loss": 0.2661, |
|
"step": 34350 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 9.322122122122122e-06, |
|
"loss": 0.2698, |
|
"step": 34375 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 9.321621621621623e-06, |
|
"loss": 0.2622, |
|
"step": 34400 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 9.321121121121121e-06, |
|
"loss": 0.2606, |
|
"step": 34425 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 9.320620620620622e-06, |
|
"loss": 0.2561, |
|
"step": 34450 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 9.320120120120121e-06, |
|
"loss": 0.2681, |
|
"step": 34475 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 9.31961961961962e-06, |
|
"loss": 0.2686, |
|
"step": 34500 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 9.31911911911912e-06, |
|
"loss": 0.2691, |
|
"step": 34525 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 9.318618618618619e-06, |
|
"loss": 0.2481, |
|
"step": 34550 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 9.318118118118118e-06, |
|
"loss": 0.2579, |
|
"step": 34575 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 9.317617617617618e-06, |
|
"loss": 0.2442, |
|
"step": 34600 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 9.317117117117119e-06, |
|
"loss": 0.2605, |
|
"step": 34625 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 9.316616616616618e-06, |
|
"loss": 0.2717, |
|
"step": 34650 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 9.316116116116117e-06, |
|
"loss": 0.2749, |
|
"step": 34675 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 9.315615615615617e-06, |
|
"loss": 0.2525, |
|
"step": 34700 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 9.315115115115116e-06, |
|
"loss": 0.2586, |
|
"step": 34725 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 9.314614614614615e-06, |
|
"loss": 0.2683, |
|
"step": 34750 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 9.314114114114114e-06, |
|
"loss": 0.2538, |
|
"step": 34775 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 9.313613613613614e-06, |
|
"loss": 0.2354, |
|
"step": 34800 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 9.313113113113113e-06, |
|
"loss": 0.2591, |
|
"step": 34825 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 9.312612612612614e-06, |
|
"loss": 0.28, |
|
"step": 34850 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 9.312112112112113e-06, |
|
"loss": 0.2789, |
|
"step": 34875 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 9.311611611611613e-06, |
|
"loss": 0.2741, |
|
"step": 34900 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 9.311111111111112e-06, |
|
"loss": 0.2803, |
|
"step": 34925 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 9.310610610610611e-06, |
|
"loss": 0.2699, |
|
"step": 34950 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 9.31011011011011e-06, |
|
"loss": 0.2618, |
|
"step": 34975 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 9.30960960960961e-06, |
|
"loss": 0.2775, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"eval_loss": 0.44763508439064026, |
|
"eval_runtime": 2480.936, |
|
"eval_samples_per_second": 6.535, |
|
"eval_steps_per_second": 0.164, |
|
"eval_wer": 25.96958086971269, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 9.30910910910911e-06, |
|
"loss": 0.2801, |
|
"step": 35025 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 9.308608608608608e-06, |
|
"loss": 0.2666, |
|
"step": 35050 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 9.30810810810811e-06, |
|
"loss": 0.2817, |
|
"step": 35075 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 9.307607607607609e-06, |
|
"loss": 0.2618, |
|
"step": 35100 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 9.307107107107108e-06, |
|
"loss": 0.2637, |
|
"step": 35125 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 9.306606606606607e-06, |
|
"loss": 0.2679, |
|
"step": 35150 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 9.306106106106107e-06, |
|
"loss": 0.2729, |
|
"step": 35175 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 9.305605605605606e-06, |
|
"loss": 0.2659, |
|
"step": 35200 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 9.305105105105105e-06, |
|
"loss": 0.2642, |
|
"step": 35225 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 9.304604604604606e-06, |
|
"loss": 0.2607, |
|
"step": 35250 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 9.304104104104105e-06, |
|
"loss": 0.2831, |
|
"step": 35275 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 9.303603603603605e-06, |
|
"loss": 0.2681, |
|
"step": 35300 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 9.303103103103104e-06, |
|
"loss": 0.2611, |
|
"step": 35325 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 9.302602602602603e-06, |
|
"loss": 0.2525, |
|
"step": 35350 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 9.302102102102103e-06, |
|
"loss": 0.2629, |
|
"step": 35375 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 9.301601601601602e-06, |
|
"loss": 0.2576, |
|
"step": 35400 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 9.301101101101101e-06, |
|
"loss": 0.2576, |
|
"step": 35425 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 9.3006006006006e-06, |
|
"loss": 0.2558, |
|
"step": 35450 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 9.300100100100102e-06, |
|
"loss": 0.2591, |
|
"step": 35475 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 9.299599599599601e-06, |
|
"loss": 0.2646, |
|
"step": 35500 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 9.2990990990991e-06, |
|
"loss": 0.2515, |
|
"step": 35525 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 9.2985985985986e-06, |
|
"loss": 0.274, |
|
"step": 35550 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 9.298098098098099e-06, |
|
"loss": 0.2696, |
|
"step": 35575 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 9.297597597597598e-06, |
|
"loss": 0.2637, |
|
"step": 35600 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 9.297097097097097e-06, |
|
"loss": 0.2644, |
|
"step": 35625 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 9.296596596596598e-06, |
|
"loss": 0.262, |
|
"step": 35650 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 9.296096096096096e-06, |
|
"loss": 0.2709, |
|
"step": 35675 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 9.295595595595597e-06, |
|
"loss": 0.263, |
|
"step": 35700 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 9.295095095095096e-06, |
|
"loss": 0.2569, |
|
"step": 35725 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 9.294594594594596e-06, |
|
"loss": 0.2525, |
|
"step": 35750 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 9.294094094094095e-06, |
|
"loss": 0.2584, |
|
"step": 35775 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 9.293593593593594e-06, |
|
"loss": 0.2665, |
|
"step": 35800 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 9.293093093093093e-06, |
|
"loss": 0.2693, |
|
"step": 35825 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 9.292592592592593e-06, |
|
"loss": 0.257, |
|
"step": 35850 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 9.292092092092094e-06, |
|
"loss": 0.2616, |
|
"step": 35875 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 9.291591591591593e-06, |
|
"loss": 0.2711, |
|
"step": 35900 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 9.291091091091092e-06, |
|
"loss": 0.2573, |
|
"step": 35925 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 9.290590590590592e-06, |
|
"loss": 0.2708, |
|
"step": 35950 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 9.290090090090091e-06, |
|
"loss": 0.2423, |
|
"step": 35975 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 9.28958958958959e-06, |
|
"loss": 0.2609, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"eval_loss": 0.4396921396255493, |
|
"eval_runtime": 2445.1333, |
|
"eval_samples_per_second": 6.631, |
|
"eval_steps_per_second": 0.166, |
|
"eval_wer": 25.296598083579376, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 9.28908908908909e-06, |
|
"loss": 0.2572, |
|
"step": 36025 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 9.288588588588589e-06, |
|
"loss": 0.2676, |
|
"step": 36050 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 9.288088088088088e-06, |
|
"loss": 0.2508, |
|
"step": 36075 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 9.287587587587589e-06, |
|
"loss": 0.2714, |
|
"step": 36100 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 9.287087087087088e-06, |
|
"loss": 0.26, |
|
"step": 36125 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 9.286586586586588e-06, |
|
"loss": 0.2648, |
|
"step": 36150 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 9.286086086086087e-06, |
|
"loss": 0.2748, |
|
"step": 36175 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 9.285585585585586e-06, |
|
"loss": 0.2602, |
|
"step": 36200 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 9.285085085085086e-06, |
|
"loss": 0.2658, |
|
"step": 36225 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 9.284584584584585e-06, |
|
"loss": 0.2623, |
|
"step": 36250 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 9.284084084084086e-06, |
|
"loss": 0.2634, |
|
"step": 36275 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 9.283583583583583e-06, |
|
"loss": 0.2832, |
|
"step": 36300 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 9.283083083083084e-06, |
|
"loss": 0.2656, |
|
"step": 36325 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 9.282582582582584e-06, |
|
"loss": 0.2565, |
|
"step": 36350 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 9.282082082082083e-06, |
|
"loss": 0.2542, |
|
"step": 36375 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 9.281581581581582e-06, |
|
"loss": 0.2594, |
|
"step": 36400 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 9.281081081081082e-06, |
|
"loss": 0.2654, |
|
"step": 36425 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 9.280580580580581e-06, |
|
"loss": 0.292, |
|
"step": 36450 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 9.28008008008008e-06, |
|
"loss": 0.282, |
|
"step": 36475 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 9.279579579579581e-06, |
|
"loss": 0.2795, |
|
"step": 36500 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 9.27907907907908e-06, |
|
"loss": 0.263, |
|
"step": 36525 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 9.27857857857858e-06, |
|
"loss": 0.272, |
|
"step": 36550 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 9.278098098098099e-06, |
|
"loss": 0.2498, |
|
"step": 36575 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 9.277597597597598e-06, |
|
"loss": 0.2626, |
|
"step": 36600 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 9.277097097097098e-06, |
|
"loss": 0.2671, |
|
"step": 36625 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 9.276596596596597e-06, |
|
"loss": 0.2622, |
|
"step": 36650 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 9.276096096096096e-06, |
|
"loss": 0.2507, |
|
"step": 36675 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 9.275595595595597e-06, |
|
"loss": 0.2606, |
|
"step": 36700 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 9.275095095095096e-06, |
|
"loss": 0.2716, |
|
"step": 36725 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 9.274594594594596e-06, |
|
"loss": 0.2699, |
|
"step": 36750 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 9.274094094094095e-06, |
|
"loss": 0.2819, |
|
"step": 36775 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 9.273593593593594e-06, |
|
"loss": 0.2681, |
|
"step": 36800 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 9.273093093093094e-06, |
|
"loss": 0.2603, |
|
"step": 36825 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 9.272592592592593e-06, |
|
"loss": 0.2556, |
|
"step": 36850 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 9.272092092092094e-06, |
|
"loss": 0.2798, |
|
"step": 36875 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 9.271591591591592e-06, |
|
"loss": 0.2623, |
|
"step": 36900 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 9.271091091091092e-06, |
|
"loss": 0.2502, |
|
"step": 36925 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 9.270590590590592e-06, |
|
"loss": 0.262, |
|
"step": 36950 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 9.270090090090091e-06, |
|
"loss": 0.2504, |
|
"step": 36975 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 9.26958958958959e-06, |
|
"loss": 0.2681, |
|
"step": 37000 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"eval_loss": 0.4397401511669159, |
|
"eval_runtime": 2578.9596, |
|
"eval_samples_per_second": 6.287, |
|
"eval_steps_per_second": 0.157, |
|
"eval_wer": 26.684962342257613, |
|
"step": 37000 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 9.26908908908909e-06, |
|
"loss": 0.2618, |
|
"step": 37025 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 9.268588588588589e-06, |
|
"loss": 0.2717, |
|
"step": 37050 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 9.268088088088088e-06, |
|
"loss": 0.2628, |
|
"step": 37075 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 9.26758758758759e-06, |
|
"loss": 0.262, |
|
"step": 37100 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 9.267087087087087e-06, |
|
"loss": 0.2531, |
|
"step": 37125 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 9.266586586586588e-06, |
|
"loss": 0.2407, |
|
"step": 37150 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 9.266086086086087e-06, |
|
"loss": 0.2717, |
|
"step": 37175 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 9.265585585585586e-06, |
|
"loss": 0.2696, |
|
"step": 37200 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 9.265085085085086e-06, |
|
"loss": 0.2718, |
|
"step": 37225 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 9.264584584584585e-06, |
|
"loss": 0.2523, |
|
"step": 37250 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 9.264084084084084e-06, |
|
"loss": 0.2637, |
|
"step": 37275 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 9.263583583583584e-06, |
|
"loss": 0.2818, |
|
"step": 37300 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 9.263083083083085e-06, |
|
"loss": 0.2812, |
|
"step": 37325 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 9.262582582582584e-06, |
|
"loss": 0.2623, |
|
"step": 37350 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 9.262082082082083e-06, |
|
"loss": 0.2615, |
|
"step": 37375 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 9.261581581581583e-06, |
|
"loss": 0.2783, |
|
"step": 37400 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 9.261081081081082e-06, |
|
"loss": 0.2577, |
|
"step": 37425 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 9.260580580580581e-06, |
|
"loss": 0.2718, |
|
"step": 37450 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 9.26008008008008e-06, |
|
"loss": 0.2569, |
|
"step": 37475 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 9.259579579579581e-06, |
|
"loss": 0.2604, |
|
"step": 37500 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 9.259079079079079e-06, |
|
"loss": 0.265, |
|
"step": 37525 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 9.25857857857858e-06, |
|
"loss": 0.2688, |
|
"step": 37550 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 9.25807807807808e-06, |
|
"loss": 0.2567, |
|
"step": 37575 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 9.257577577577579e-06, |
|
"loss": 0.2747, |
|
"step": 37600 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 9.257077077077078e-06, |
|
"loss": 0.27, |
|
"step": 37625 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 9.256576576576577e-06, |
|
"loss": 0.2699, |
|
"step": 37650 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 9.256076076076076e-06, |
|
"loss": 0.2645, |
|
"step": 37675 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 9.255575575575576e-06, |
|
"loss": 0.2723, |
|
"step": 37700 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 9.255075075075077e-06, |
|
"loss": 0.2693, |
|
"step": 37725 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 9.254574574574574e-06, |
|
"loss": 0.2509, |
|
"step": 37750 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 9.254074074074075e-06, |
|
"loss": 0.2698, |
|
"step": 37775 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 9.253573573573575e-06, |
|
"loss": 0.2359, |
|
"step": 37800 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 9.253073073073074e-06, |
|
"loss": 0.2707, |
|
"step": 37825 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 9.252572572572573e-06, |
|
"loss": 0.2534, |
|
"step": 37850 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 9.252072072072073e-06, |
|
"loss": 0.2554, |
|
"step": 37875 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 9.251571571571572e-06, |
|
"loss": 0.2753, |
|
"step": 37900 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 9.251071071071071e-06, |
|
"loss": 0.2551, |
|
"step": 37925 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 9.250570570570572e-06, |
|
"loss": 0.2576, |
|
"step": 37950 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 9.250070070070071e-06, |
|
"loss": 0.2579, |
|
"step": 37975 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 9.24956956956957e-06, |
|
"loss": 0.2661, |
|
"step": 38000 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"eval_loss": 0.4426652491092682, |
|
"eval_runtime": 2522.4342, |
|
"eval_samples_per_second": 6.428, |
|
"eval_steps_per_second": 0.161, |
|
"eval_wer": 25.423023257606708, |
|
"step": 38000 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 9.24906906906907e-06, |
|
"loss": 0.2819, |
|
"step": 38025 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 9.24856856856857e-06, |
|
"loss": 0.2773, |
|
"step": 38050 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 9.248068068068069e-06, |
|
"loss": 0.2604, |
|
"step": 38075 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 9.247567567567568e-06, |
|
"loss": 0.2645, |
|
"step": 38100 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 9.247067067067069e-06, |
|
"loss": 0.2504, |
|
"step": 38125 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 9.246566566566567e-06, |
|
"loss": 0.2731, |
|
"step": 38150 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 9.246066066066068e-06, |
|
"loss": 0.2543, |
|
"step": 38175 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 9.245565565565567e-06, |
|
"loss": 0.2552, |
|
"step": 38200 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 9.245065065065066e-06, |
|
"loss": 0.2789, |
|
"step": 38225 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 9.244564564564565e-06, |
|
"loss": 0.2668, |
|
"step": 38250 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 9.244064064064065e-06, |
|
"loss": 0.2635, |
|
"step": 38275 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 9.243563563563564e-06, |
|
"loss": 0.248, |
|
"step": 38300 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 9.243063063063063e-06, |
|
"loss": 0.2675, |
|
"step": 38325 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 9.242562562562564e-06, |
|
"loss": 0.265, |
|
"step": 38350 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 9.242062062062062e-06, |
|
"loss": 0.2509, |
|
"step": 38375 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 9.241561561561563e-06, |
|
"loss": 0.2535, |
|
"step": 38400 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 9.241061061061062e-06, |
|
"loss": 0.2716, |
|
"step": 38425 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 9.240560560560561e-06, |
|
"loss": 0.2578, |
|
"step": 38450 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 9.24006006006006e-06, |
|
"loss": 0.2743, |
|
"step": 38475 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 9.23955955955956e-06, |
|
"loss": 0.2701, |
|
"step": 38500 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 9.23905905905906e-06, |
|
"loss": 0.2716, |
|
"step": 38525 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 9.238558558558559e-06, |
|
"loss": 0.2696, |
|
"step": 38550 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 9.23805805805806e-06, |
|
"loss": 0.2701, |
|
"step": 38575 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 9.237577577577577e-06, |
|
"loss": 0.2573, |
|
"step": 38600 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 9.237077077077078e-06, |
|
"loss": 0.2641, |
|
"step": 38625 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 9.236576576576577e-06, |
|
"loss": 0.277, |
|
"step": 38650 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 9.236076076076077e-06, |
|
"loss": 0.2562, |
|
"step": 38675 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 9.235575575575576e-06, |
|
"loss": 0.2584, |
|
"step": 38700 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 9.235075075075077e-06, |
|
"loss": 0.2571, |
|
"step": 38725 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 9.234574574574575e-06, |
|
"loss": 0.2652, |
|
"step": 38750 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 9.234074074074076e-06, |
|
"loss": 0.26, |
|
"step": 38775 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 9.233573573573575e-06, |
|
"loss": 0.2475, |
|
"step": 38800 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 9.233073073073074e-06, |
|
"loss": 0.2728, |
|
"step": 38825 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 9.232572572572573e-06, |
|
"loss": 0.2678, |
|
"step": 38850 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 9.232072072072073e-06, |
|
"loss": 0.2597, |
|
"step": 38875 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 9.231571571571572e-06, |
|
"loss": 0.2532, |
|
"step": 38900 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 9.231071071071071e-06, |
|
"loss": 0.2552, |
|
"step": 38925 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 9.230570570570572e-06, |
|
"loss": 0.2528, |
|
"step": 38950 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 9.23007007007007e-06, |
|
"loss": 0.2356, |
|
"step": 38975 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 9.229569569569571e-06, |
|
"loss": 0.2512, |
|
"step": 39000 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"eval_loss": 0.4447844326496124, |
|
"eval_runtime": 2535.3087, |
|
"eval_samples_per_second": 6.395, |
|
"eval_steps_per_second": 0.16, |
|
"eval_wer": 26.18311607218569, |
|
"step": 39000 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 9.22906906906907e-06, |
|
"loss": 0.28, |
|
"step": 39025 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 9.22856856856857e-06, |
|
"loss": 0.2598, |
|
"step": 39050 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 9.228068068068069e-06, |
|
"loss": 0.2805, |
|
"step": 39075 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 9.227567567567568e-06, |
|
"loss": 0.2677, |
|
"step": 39100 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 9.227067067067067e-06, |
|
"loss": 0.2749, |
|
"step": 39125 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 9.226566566566567e-06, |
|
"loss": 0.2652, |
|
"step": 39150 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 9.226066066066068e-06, |
|
"loss": 0.2609, |
|
"step": 39175 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 9.225565565565567e-06, |
|
"loss": 0.2656, |
|
"step": 39200 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 9.225065065065065e-06, |
|
"loss": 0.2648, |
|
"step": 39225 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 9.224564564564566e-06, |
|
"loss": 0.2567, |
|
"step": 39250 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 9.224064064064065e-06, |
|
"loss": 0.2433, |
|
"step": 39275 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 9.223563563563564e-06, |
|
"loss": 0.276, |
|
"step": 39300 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 9.223063063063064e-06, |
|
"loss": 0.2651, |
|
"step": 39325 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 9.222562562562564e-06, |
|
"loss": 0.2591, |
|
"step": 39350 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 9.222062062062062e-06, |
|
"loss": 0.2616, |
|
"step": 39375 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 9.221561561561563e-06, |
|
"loss": 0.2563, |
|
"step": 39400 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 9.221061061061062e-06, |
|
"loss": 0.2595, |
|
"step": 39425 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 9.22056056056056e-06, |
|
"loss": 0.2646, |
|
"step": 39450 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 9.220060060060061e-06, |
|
"loss": 0.2676, |
|
"step": 39475 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 9.21955955955956e-06, |
|
"loss": 0.2585, |
|
"step": 39500 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 9.21905905905906e-06, |
|
"loss": 0.2599, |
|
"step": 39525 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 9.218558558558559e-06, |
|
"loss": 0.2477, |
|
"step": 39550 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 9.21805805805806e-06, |
|
"loss": 0.2376, |
|
"step": 39575 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 9.217557557557557e-06, |
|
"loss": 0.2491, |
|
"step": 39600 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 9.217057057057058e-06, |
|
"loss": 0.2533, |
|
"step": 39625 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 9.216556556556558e-06, |
|
"loss": 0.2646, |
|
"step": 39650 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 9.216056056056057e-06, |
|
"loss": 0.2824, |
|
"step": 39675 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 9.215555555555556e-06, |
|
"loss": 0.2854, |
|
"step": 39700 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 9.215055055055056e-06, |
|
"loss": 0.2553, |
|
"step": 39725 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 9.214554554554555e-06, |
|
"loss": 0.2484, |
|
"step": 39750 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 9.214054054054054e-06, |
|
"loss": 0.2362, |
|
"step": 39775 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 9.213553553553555e-06, |
|
"loss": 0.2634, |
|
"step": 39800 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 9.213053053053055e-06, |
|
"loss": 0.2554, |
|
"step": 39825 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 9.212552552552552e-06, |
|
"loss": 0.2502, |
|
"step": 39850 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 9.212052052052053e-06, |
|
"loss": 0.2546, |
|
"step": 39875 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 9.211551551551552e-06, |
|
"loss": 0.2723, |
|
"step": 39900 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 9.211051051051052e-06, |
|
"loss": 0.2674, |
|
"step": 39925 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 9.210550550550551e-06, |
|
"loss": 0.2688, |
|
"step": 39950 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 9.21005005005005e-06, |
|
"loss": 0.2512, |
|
"step": 39975 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 9.20954954954955e-06, |
|
"loss": 0.2541, |
|
"step": 40000 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"eval_loss": 0.4458891749382019, |
|
"eval_runtime": 2617.6449, |
|
"eval_samples_per_second": 6.194, |
|
"eval_steps_per_second": 0.155, |
|
"eval_wer": 26.314937442665414, |
|
"step": 40000 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 9.20904904904905e-06, |
|
"loss": 0.2547, |
|
"step": 40025 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 9.20854854854855e-06, |
|
"loss": 0.2557, |
|
"step": 40050 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 9.208048048048048e-06, |
|
"loss": 0.2526, |
|
"step": 40075 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 9.207547547547549e-06, |
|
"loss": 0.2709, |
|
"step": 40100 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 9.207047047047048e-06, |
|
"loss": 0.2643, |
|
"step": 40125 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 9.206546546546547e-06, |
|
"loss": 0.2716, |
|
"step": 40150 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 9.206046046046046e-06, |
|
"loss": 0.2725, |
|
"step": 40175 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 9.205545545545547e-06, |
|
"loss": 0.2532, |
|
"step": 40200 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 9.205045045045045e-06, |
|
"loss": 0.2595, |
|
"step": 40225 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 9.204544544544546e-06, |
|
"loss": 0.2619, |
|
"step": 40250 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 9.204044044044045e-06, |
|
"loss": 0.2627, |
|
"step": 40275 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 9.203543543543545e-06, |
|
"loss": 0.2581, |
|
"step": 40300 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 9.203043043043044e-06, |
|
"loss": 0.2634, |
|
"step": 40325 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 9.202542542542543e-06, |
|
"loss": 0.2392, |
|
"step": 40350 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 9.202042042042042e-06, |
|
"loss": 0.2694, |
|
"step": 40375 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 9.201541541541542e-06, |
|
"loss": 0.2596, |
|
"step": 40400 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 9.201041041041043e-06, |
|
"loss": 0.2844, |
|
"step": 40425 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 9.200540540540542e-06, |
|
"loss": 0.2679, |
|
"step": 40450 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 9.20004004004004e-06, |
|
"loss": 0.2592, |
|
"step": 40475 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 9.19953953953954e-06, |
|
"loss": 0.2523, |
|
"step": 40500 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 9.19903903903904e-06, |
|
"loss": 0.27, |
|
"step": 40525 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 9.19853853853854e-06, |
|
"loss": 0.272, |
|
"step": 40550 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 9.198038038038039e-06, |
|
"loss": 0.2621, |
|
"step": 40575 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 9.197537537537538e-06, |
|
"loss": 0.262, |
|
"step": 40600 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 9.197057057057059e-06, |
|
"loss": 0.2642, |
|
"step": 40625 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 9.196556556556558e-06, |
|
"loss": 0.278, |
|
"step": 40650 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 9.196056056056056e-06, |
|
"loss": 0.2711, |
|
"step": 40675 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 9.195555555555557e-06, |
|
"loss": 0.2728, |
|
"step": 40700 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 9.195055055055056e-06, |
|
"loss": 0.2844, |
|
"step": 40725 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 9.194554554554555e-06, |
|
"loss": 0.2657, |
|
"step": 40750 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 9.194054054054054e-06, |
|
"loss": 0.2559, |
|
"step": 40775 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 9.193553553553555e-06, |
|
"loss": 0.2671, |
|
"step": 40800 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 9.193053053053053e-06, |
|
"loss": 0.2559, |
|
"step": 40825 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 9.192552552552554e-06, |
|
"loss": 0.2546, |
|
"step": 40850 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 9.192052052052053e-06, |
|
"loss": 0.2607, |
|
"step": 40875 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 9.191551551551553e-06, |
|
"loss": 0.2718, |
|
"step": 40900 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 9.191051051051052e-06, |
|
"loss": 0.2672, |
|
"step": 40925 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 9.190550550550551e-06, |
|
"loss": 0.2682, |
|
"step": 40950 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 9.19005005005005e-06, |
|
"loss": 0.2646, |
|
"step": 40975 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 9.18954954954955e-06, |
|
"loss": 0.2535, |
|
"step": 41000 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"eval_loss": 0.44114458560943604, |
|
"eval_runtime": 2473.7687, |
|
"eval_samples_per_second": 6.554, |
|
"eval_steps_per_second": 0.164, |
|
"eval_wer": 25.31355755814402, |
|
"step": 41000 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 9.18904904904905e-06, |
|
"loss": 0.2634, |
|
"step": 41025 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 9.18854854854855e-06, |
|
"loss": 0.2616, |
|
"step": 41050 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 9.188048048048048e-06, |
|
"loss": 0.266, |
|
"step": 41075 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 9.187547547547549e-06, |
|
"loss": 0.264, |
|
"step": 41100 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 9.187047047047048e-06, |
|
"loss": 0.2604, |
|
"step": 41125 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 9.186546546546547e-06, |
|
"loss": 0.2552, |
|
"step": 41150 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 9.186046046046047e-06, |
|
"loss": 0.2539, |
|
"step": 41175 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 9.185545545545546e-06, |
|
"loss": 0.2854, |
|
"step": 41200 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 9.185045045045045e-06, |
|
"loss": 0.2629, |
|
"step": 41225 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 9.184544544544546e-06, |
|
"loss": 0.2613, |
|
"step": 41250 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 9.184044044044045e-06, |
|
"loss": 0.2561, |
|
"step": 41275 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 9.183543543543543e-06, |
|
"loss": 0.2617, |
|
"step": 41300 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 9.183043043043044e-06, |
|
"loss": 0.2573, |
|
"step": 41325 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 9.182542542542543e-06, |
|
"loss": 0.2622, |
|
"step": 41350 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 9.182042042042043e-06, |
|
"loss": 0.2472, |
|
"step": 41375 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 9.181541541541542e-06, |
|
"loss": 0.262, |
|
"step": 41400 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 9.181041041041043e-06, |
|
"loss": 0.2558, |
|
"step": 41425 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 9.18054054054054e-06, |
|
"loss": 0.2572, |
|
"step": 41450 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 9.180040040040042e-06, |
|
"loss": 0.2618, |
|
"step": 41475 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 9.17953953953954e-06, |
|
"loss": 0.2532, |
|
"step": 41500 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 9.17903903903904e-06, |
|
"loss": 0.2522, |
|
"step": 41525 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 9.17853853853854e-06, |
|
"loss": 0.275, |
|
"step": 41550 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 9.178038038038039e-06, |
|
"loss": 0.2754, |
|
"step": 41575 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 9.177537537537538e-06, |
|
"loss": 0.2599, |
|
"step": 41600 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 9.177037037037037e-06, |
|
"loss": 0.2509, |
|
"step": 41625 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 9.176536536536538e-06, |
|
"loss": 0.2554, |
|
"step": 41650 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 9.176036036036036e-06, |
|
"loss": 0.2426, |
|
"step": 41675 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 9.175535535535535e-06, |
|
"loss": 0.2675, |
|
"step": 41700 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 9.175035035035036e-06, |
|
"loss": 0.2559, |
|
"step": 41725 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 9.174534534534536e-06, |
|
"loss": 0.2716, |
|
"step": 41750 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 9.174034034034035e-06, |
|
"loss": 0.2686, |
|
"step": 41775 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 9.173533533533534e-06, |
|
"loss": 0.259, |
|
"step": 41800 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 9.173033033033033e-06, |
|
"loss": 0.2747, |
|
"step": 41825 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 9.172532532532533e-06, |
|
"loss": 0.2507, |
|
"step": 41850 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 9.172032032032034e-06, |
|
"loss": 0.2573, |
|
"step": 41875 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 9.171531531531533e-06, |
|
"loss": 0.2542, |
|
"step": 41900 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 9.17103103103103e-06, |
|
"loss": 0.2584, |
|
"step": 41925 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 9.170530530530532e-06, |
|
"loss": 0.2459, |
|
"step": 41950 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 9.170030030030031e-06, |
|
"loss": 0.2533, |
|
"step": 41975 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 9.16952952952953e-06, |
|
"loss": 0.2518, |
|
"step": 42000 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"eval_loss": 0.42200759053230286, |
|
"eval_runtime": 2416.12, |
|
"eval_samples_per_second": 6.71, |
|
"eval_steps_per_second": 0.168, |
|
"eval_wer": 24.71689240755159, |
|
"step": 42000 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 9.16902902902903e-06, |
|
"loss": 0.2759, |
|
"step": 42025 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 9.16852852852853e-06, |
|
"loss": 0.258, |
|
"step": 42050 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 9.168028028028028e-06, |
|
"loss": 0.2424, |
|
"step": 42075 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 9.167527527527527e-06, |
|
"loss": 0.258, |
|
"step": 42100 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 9.167027027027028e-06, |
|
"loss": 0.274, |
|
"step": 42125 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 9.166526526526528e-06, |
|
"loss": 0.2479, |
|
"step": 42150 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 9.166026026026027e-06, |
|
"loss": 0.2697, |
|
"step": 42175 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 9.165525525525526e-06, |
|
"loss": 0.2621, |
|
"step": 42200 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 9.165025025025026e-06, |
|
"loss": 0.2645, |
|
"step": 42225 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 9.164524524524525e-06, |
|
"loss": 0.2624, |
|
"step": 42250 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 9.164024024024026e-06, |
|
"loss": 0.2592, |
|
"step": 42275 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 9.163523523523523e-06, |
|
"loss": 0.255, |
|
"step": 42300 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 9.163023023023023e-06, |
|
"loss": 0.262, |
|
"step": 42325 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 9.162522522522524e-06, |
|
"loss": 0.2687, |
|
"step": 42350 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 9.162022022022023e-06, |
|
"loss": 0.2566, |
|
"step": 42375 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 9.161521521521522e-06, |
|
"loss": 0.2563, |
|
"step": 42400 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 9.161021021021022e-06, |
|
"loss": 0.2559, |
|
"step": 42425 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 9.160520520520521e-06, |
|
"loss": 0.2641, |
|
"step": 42450 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 9.16002002002002e-06, |
|
"loss": 0.2729, |
|
"step": 42475 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 9.159519519519521e-06, |
|
"loss": 0.2531, |
|
"step": 42500 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 9.15901901901902e-06, |
|
"loss": 0.2819, |
|
"step": 42525 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 9.158518518518518e-06, |
|
"loss": 0.2594, |
|
"step": 42550 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 9.158018018018019e-06, |
|
"loss": 0.2649, |
|
"step": 42575 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 9.157517517517518e-06, |
|
"loss": 0.2593, |
|
"step": 42600 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 9.157017017017018e-06, |
|
"loss": 0.2636, |
|
"step": 42625 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 9.156536536536539e-06, |
|
"loss": 0.2698, |
|
"step": 42650 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 9.156036036036036e-06, |
|
"loss": 0.254, |
|
"step": 42675 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 9.155535535535535e-06, |
|
"loss": 0.2356, |
|
"step": 42700 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 9.155035035035036e-06, |
|
"loss": 0.2506, |
|
"step": 42725 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 9.154534534534536e-06, |
|
"loss": 0.2669, |
|
"step": 42750 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 9.154034034034035e-06, |
|
"loss": 0.2576, |
|
"step": 42775 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 9.153533533533534e-06, |
|
"loss": 0.2536, |
|
"step": 42800 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 9.153033033033034e-06, |
|
"loss": 0.2686, |
|
"step": 42825 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 9.152532532532533e-06, |
|
"loss": 0.2511, |
|
"step": 42850 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 9.152032032032034e-06, |
|
"loss": 0.2667, |
|
"step": 42875 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 9.151531531531531e-06, |
|
"loss": 0.2593, |
|
"step": 42900 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 9.15103103103103e-06, |
|
"loss": 0.2656, |
|
"step": 42925 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 9.150530530530532e-06, |
|
"loss": 0.2574, |
|
"step": 42950 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 9.150030030030031e-06, |
|
"loss": 0.2557, |
|
"step": 42975 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 9.14952952952953e-06, |
|
"loss": 0.255, |
|
"step": 43000 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"eval_loss": 0.43135425448417664, |
|
"eval_runtime": 2587.1846, |
|
"eval_samples_per_second": 6.267, |
|
"eval_steps_per_second": 0.157, |
|
"eval_wer": 25.259595593620155, |
|
"step": 43000 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 9.14902902902903e-06, |
|
"loss": 0.2639, |
|
"step": 43025 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 9.148528528528529e-06, |
|
"loss": 0.256, |
|
"step": 43050 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 9.148028028028028e-06, |
|
"loss": 0.2588, |
|
"step": 43075 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 9.14752752752753e-06, |
|
"loss": 0.2768, |
|
"step": 43100 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 9.147027027027029e-06, |
|
"loss": 0.2557, |
|
"step": 43125 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 9.146526526526526e-06, |
|
"loss": 0.2506, |
|
"step": 43150 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 9.146026026026027e-06, |
|
"loss": 0.2495, |
|
"step": 43175 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 9.145525525525526e-06, |
|
"loss": 0.2652, |
|
"step": 43200 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 9.145025025025026e-06, |
|
"loss": 0.2525, |
|
"step": 43225 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 9.144524524524525e-06, |
|
"loss": 0.2694, |
|
"step": 43250 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 9.144024024024026e-06, |
|
"loss": 0.2489, |
|
"step": 43275 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 9.143523523523524e-06, |
|
"loss": 0.2531, |
|
"step": 43300 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 9.143023023023023e-06, |
|
"loss": 0.2549, |
|
"step": 43325 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 9.142522522522524e-06, |
|
"loss": 0.2526, |
|
"step": 43350 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 9.142022022022023e-06, |
|
"loss": 0.2604, |
|
"step": 43375 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 9.141521521521523e-06, |
|
"loss": 0.2683, |
|
"step": 43400 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 9.141021021021022e-06, |
|
"loss": 0.2659, |
|
"step": 43425 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 9.140520520520521e-06, |
|
"loss": 0.2569, |
|
"step": 43450 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 9.14002002002002e-06, |
|
"loss": 0.2556, |
|
"step": 43475 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 9.139519519519521e-06, |
|
"loss": 0.2752, |
|
"step": 43500 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 9.139019019019019e-06, |
|
"loss": 0.2562, |
|
"step": 43525 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 9.138518518518518e-06, |
|
"loss": 0.2848, |
|
"step": 43550 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 9.13801801801802e-06, |
|
"loss": 0.2598, |
|
"step": 43575 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 9.137517517517519e-06, |
|
"loss": 0.2663, |
|
"step": 43600 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 9.137017017017018e-06, |
|
"loss": 0.2582, |
|
"step": 43625 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 9.136516516516517e-06, |
|
"loss": 0.2646, |
|
"step": 43650 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 9.136016016016016e-06, |
|
"loss": 0.2623, |
|
"step": 43675 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 9.135515515515516e-06, |
|
"loss": 0.2569, |
|
"step": 43700 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 9.135015015015017e-06, |
|
"loss": 0.2504, |
|
"step": 43725 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 9.134514514514516e-06, |
|
"loss": 0.2726, |
|
"step": 43750 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 9.134014014014014e-06, |
|
"loss": 0.2593, |
|
"step": 43775 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 9.133513513513515e-06, |
|
"loss": 0.2478, |
|
"step": 43800 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 9.133013013013014e-06, |
|
"loss": 0.2582, |
|
"step": 43825 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 9.132512512512513e-06, |
|
"loss": 0.2646, |
|
"step": 43850 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 9.132012012012013e-06, |
|
"loss": 0.2822, |
|
"step": 43875 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 9.131511511511514e-06, |
|
"loss": 0.2379, |
|
"step": 43900 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 9.131011011011011e-06, |
|
"loss": 0.2728, |
|
"step": 43925 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 9.13051051051051e-06, |
|
"loss": 0.274, |
|
"step": 43950 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 9.130010010010011e-06, |
|
"loss": 0.2838, |
|
"step": 43975 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 9.129509509509509e-06, |
|
"loss": 0.2639, |
|
"step": 44000 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"eval_loss": 0.43265146017074585, |
|
"eval_runtime": 2614.8693, |
|
"eval_samples_per_second": 6.2, |
|
"eval_steps_per_second": 0.155, |
|
"eval_wer": 25.521696564164632, |
|
"step": 44000 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 9.12900900900901e-06, |
|
"loss": 0.2588, |
|
"step": 44025 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 9.12850850850851e-06, |
|
"loss": 0.2764, |
|
"step": 44050 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 9.128008008008009e-06, |
|
"loss": 0.263, |
|
"step": 44075 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 9.127507507507508e-06, |
|
"loss": 0.2644, |
|
"step": 44100 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 9.127007007007009e-06, |
|
"loss": 0.2476, |
|
"step": 44125 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 9.126506506506507e-06, |
|
"loss": 0.2604, |
|
"step": 44150 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 9.126006006006006e-06, |
|
"loss": 0.2747, |
|
"step": 44175 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 9.125505505505507e-06, |
|
"loss": 0.265, |
|
"step": 44200 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 9.125005005005006e-06, |
|
"loss": 0.2568, |
|
"step": 44225 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 9.124504504504505e-06, |
|
"loss": 0.2485, |
|
"step": 44250 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 9.124004004004005e-06, |
|
"loss": 0.2742, |
|
"step": 44275 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 9.123503503503504e-06, |
|
"loss": 0.2634, |
|
"step": 44300 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 9.123003003003003e-06, |
|
"loss": 0.2805, |
|
"step": 44325 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 9.122502502502504e-06, |
|
"loss": 0.2543, |
|
"step": 44350 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 9.122002002002004e-06, |
|
"loss": 0.2753, |
|
"step": 44375 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 9.121501501501501e-06, |
|
"loss": 0.2607, |
|
"step": 44400 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 9.121001001001002e-06, |
|
"loss": 0.242, |
|
"step": 44425 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 9.120500500500501e-06, |
|
"loss": 0.2633, |
|
"step": 44450 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 9.12e-06, |
|
"loss": 0.2686, |
|
"step": 44475 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 9.1194994994995e-06, |
|
"loss": 0.2608, |
|
"step": 44500 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 9.118998998999e-06, |
|
"loss": 0.2779, |
|
"step": 44525 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 9.118498498498499e-06, |
|
"loss": 0.2648, |
|
"step": 44550 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 9.117997997997998e-06, |
|
"loss": 0.2701, |
|
"step": 44575 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 9.117497497497499e-06, |
|
"loss": 0.245, |
|
"step": 44600 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 9.116996996996997e-06, |
|
"loss": 0.2607, |
|
"step": 44625 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 9.116516516516517e-06, |
|
"loss": 0.2622, |
|
"step": 44650 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 9.116016016016017e-06, |
|
"loss": 0.2466, |
|
"step": 44675 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 9.115515515515516e-06, |
|
"loss": 0.2627, |
|
"step": 44700 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 9.115015015015017e-06, |
|
"loss": 0.26, |
|
"step": 44725 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 9.114514514514515e-06, |
|
"loss": 0.2837, |
|
"step": 44750 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 9.114014014014014e-06, |
|
"loss": 0.2417, |
|
"step": 44775 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 9.113513513513515e-06, |
|
"loss": 0.2663, |
|
"step": 44800 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 9.113013013013014e-06, |
|
"loss": 0.2606, |
|
"step": 44825 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 9.112512512512513e-06, |
|
"loss": 0.2416, |
|
"step": 44850 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 9.112012012012013e-06, |
|
"loss": 0.2537, |
|
"step": 44875 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 9.111511511511512e-06, |
|
"loss": 0.2587, |
|
"step": 44900 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 9.111011011011011e-06, |
|
"loss": 0.2658, |
|
"step": 44925 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 9.11051051051051e-06, |
|
"loss": 0.2653, |
|
"step": 44950 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 9.110010010010012e-06, |
|
"loss": 0.2643, |
|
"step": 44975 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 9.10950950950951e-06, |
|
"loss": 0.2638, |
|
"step": 45000 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"eval_loss": 0.432108074426651, |
|
"eval_runtime": 2486.7031, |
|
"eval_samples_per_second": 6.52, |
|
"eval_steps_per_second": 0.163, |
|
"eval_wer": 24.733851882116234, |
|
"step": 45000 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 9.10900900900901e-06, |
|
"loss": 0.268, |
|
"step": 45025 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 9.10850850850851e-06, |
|
"loss": 0.2566, |
|
"step": 45050 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 9.108008008008009e-06, |
|
"loss": 0.2603, |
|
"step": 45075 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 9.107507507507508e-06, |
|
"loss": 0.2463, |
|
"step": 45100 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 9.107007007007007e-06, |
|
"loss": 0.2597, |
|
"step": 45125 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 9.106506506506507e-06, |
|
"loss": 0.2531, |
|
"step": 45150 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 9.106006006006006e-06, |
|
"loss": 0.2675, |
|
"step": 45175 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 9.105505505505507e-06, |
|
"loss": 0.2445, |
|
"step": 45200 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 9.105005005005005e-06, |
|
"loss": 0.2576, |
|
"step": 45225 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 9.104504504504506e-06, |
|
"loss": 0.2473, |
|
"step": 45250 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 9.104004004004005e-06, |
|
"loss": 0.2551, |
|
"step": 45275 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 9.103503503503504e-06, |
|
"loss": 0.2658, |
|
"step": 45300 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 9.103003003003003e-06, |
|
"loss": 0.2603, |
|
"step": 45325 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 9.102502502502504e-06, |
|
"loss": 0.2618, |
|
"step": 45350 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 9.102002002002002e-06, |
|
"loss": 0.2405, |
|
"step": 45375 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 9.101501501501501e-06, |
|
"loss": 0.263, |
|
"step": 45400 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 9.101001001001002e-06, |
|
"loss": 0.2673, |
|
"step": 45425 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 9.100500500500502e-06, |
|
"loss": 0.2638, |
|
"step": 45450 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 9.100000000000001e-06, |
|
"loss": 0.2583, |
|
"step": 45475 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 9.0994994994995e-06, |
|
"loss": 0.2695, |
|
"step": 45500 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 9.098998998999e-06, |
|
"loss": 0.2544, |
|
"step": 45525 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 9.098498498498499e-06, |
|
"loss": 0.2634, |
|
"step": 45550 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 9.097997997997998e-06, |
|
"loss": 0.2532, |
|
"step": 45575 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 9.097497497497499e-06, |
|
"loss": 0.2559, |
|
"step": 45600 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 9.096996996996997e-06, |
|
"loss": 0.2459, |
|
"step": 45625 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 9.096496496496498e-06, |
|
"loss": 0.2781, |
|
"step": 45650 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 9.095995995995997e-06, |
|
"loss": 0.2712, |
|
"step": 45675 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 9.095495495495496e-06, |
|
"loss": 0.2597, |
|
"step": 45700 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 9.094994994994996e-06, |
|
"loss": 0.275, |
|
"step": 45725 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 9.094494494494495e-06, |
|
"loss": 0.2458, |
|
"step": 45750 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 9.093993993993994e-06, |
|
"loss": 0.2752, |
|
"step": 45775 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 9.093493493493494e-06, |
|
"loss": 0.261, |
|
"step": 45800 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 9.092992992992995e-06, |
|
"loss": 0.2532, |
|
"step": 45825 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 9.092492492492492e-06, |
|
"loss": 0.2847, |
|
"step": 45850 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 9.091991991991993e-06, |
|
"loss": 0.2607, |
|
"step": 45875 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 9.091491491491492e-06, |
|
"loss": 0.2678, |
|
"step": 45900 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 9.090990990990992e-06, |
|
"loss": 0.2488, |
|
"step": 45925 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 9.090490490490491e-06, |
|
"loss": 0.2628, |
|
"step": 45950 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 9.089989989989992e-06, |
|
"loss": 0.2625, |
|
"step": 45975 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 9.08948948948949e-06, |
|
"loss": 0.2647, |
|
"step": 46000 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"eval_loss": 0.44105586409568787, |
|
"eval_runtime": 2496.4644, |
|
"eval_samples_per_second": 6.494, |
|
"eval_steps_per_second": 0.163, |
|
"eval_wer": 24.86027705614357, |
|
"step": 46000 |
|
} |
|
], |
|
"max_steps": 500000, |
|
"num_train_epochs": 20, |
|
"total_flos": 5.309913924895223e+20, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|