|
{ |
|
"best_metric": 20.044683596085466, |
|
"best_model_checkpoint": "all_lang_models/marathi_models/whisper-medium-mr_alldata_multigpu/checkpoint-8200", |
|
"epoch": 2.8472222222222223, |
|
"global_step": 8200, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 2.445014948199563e-06, |
|
"loss": 2.2897, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 4.654519848766012e-06, |
|
"loss": 1.1234, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 5.605150925598839e-06, |
|
"loss": 0.7141, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 6.219461347633451e-06, |
|
"loss": 0.5167, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 6.6741959581207635e-06, |
|
"loss": 0.4023, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 7.035392983998264e-06, |
|
"loss": 0.3122, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 7.3350448445986885e-06, |
|
"loss": 0.2807, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 7.5911031495554354e-06, |
|
"loss": 0.2622, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 7.814655826165288e-06, |
|
"loss": 0.2502, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 8.013033860827016e-06, |
|
"loss": 0.2301, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"eval_loss": 0.304443359375, |
|
"eval_runtime": 284.7229, |
|
"eval_samples_per_second": 10.006, |
|
"eval_steps_per_second": 0.081, |
|
"eval_wer": 45.1272853142012, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 8.19133763518252e-06, |
|
"loss": 0.2152, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 8.353259112018159e-06, |
|
"loss": 0.209, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 8.50155696580508e-06, |
|
"loss": 0.2003, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 8.6383468479903e-06, |
|
"loss": 0.1952, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 8.765286902998114e-06, |
|
"loss": 0.1897, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 8.883700858687214e-06, |
|
"loss": 0.1869, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 8.994662316723439e-06, |
|
"loss": 0.1825, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.099054052977428e-06, |
|
"loss": 0.1779, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.19761072313247e-06, |
|
"loss": 0.1767, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.290950252080677e-06, |
|
"loss": 0.1658, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"eval_loss": 0.2354736328125, |
|
"eval_runtime": 1180.4962, |
|
"eval_samples_per_second": 2.413, |
|
"eval_steps_per_second": 0.019, |
|
"eval_wer": 37.77966581704899, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.379597325032726e-06, |
|
"loss": 0.1638, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.464001251372113e-06, |
|
"loss": 0.1654, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.544549745165139e-06, |
|
"loss": 0.1603, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.621579693581943e-06, |
|
"loss": 0.1589, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.695385670292832e-06, |
|
"loss": 0.1557, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.766226737805446e-06, |
|
"loss": 0.1523, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.83433193552004e-06, |
|
"loss": 0.1576, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.899904746919066e-06, |
|
"loss": 0.1504, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.96312676561764e-06, |
|
"loss": 0.1468, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.998955067920586e-06, |
|
"loss": 0.1492, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"eval_loss": 0.2061767578125, |
|
"eval_runtime": 514.1736, |
|
"eval_samples_per_second": 5.541, |
|
"eval_steps_per_second": 0.045, |
|
"eval_wer": 33.78646275842537, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.995471960989204e-06, |
|
"loss": 0.1447, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.99198885405782e-06, |
|
"loss": 0.1476, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.988505747126437e-06, |
|
"loss": 0.142, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.985022640195054e-06, |
|
"loss": 0.1433, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.981539533263672e-06, |
|
"loss": 0.1379, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.978056426332289e-06, |
|
"loss": 0.1418, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.974573319400907e-06, |
|
"loss": 0.141, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.971090212469524e-06, |
|
"loss": 0.1389, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.96760710553814e-06, |
|
"loss": 0.1352, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.964123998606759e-06, |
|
"loss": 0.139, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"eval_loss": 0.1884765625, |
|
"eval_runtime": 739.6853, |
|
"eval_samples_per_second": 3.852, |
|
"eval_steps_per_second": 0.031, |
|
"eval_wer": 32.59070455332137, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.960640891675375e-06, |
|
"loss": 0.132, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.957157784743993e-06, |
|
"loss": 0.1298, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.95367467781261e-06, |
|
"loss": 0.1269, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.950191570881227e-06, |
|
"loss": 0.1242, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.946708463949845e-06, |
|
"loss": 0.1283, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.943225357018461e-06, |
|
"loss": 0.131, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.939742250087078e-06, |
|
"loss": 0.1232, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.936259143155695e-06, |
|
"loss": 0.1228, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.932776036224313e-06, |
|
"loss": 0.1249, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.92929292929293e-06, |
|
"loss": 0.1238, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"eval_loss": 0.177001953125, |
|
"eval_runtime": 759.5608, |
|
"eval_samples_per_second": 3.751, |
|
"eval_steps_per_second": 0.03, |
|
"eval_wer": 30.28729664243683, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.925809822361548e-06, |
|
"loss": 0.1245, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.922326715430164e-06, |
|
"loss": 0.1251, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.918843608498781e-06, |
|
"loss": 0.1222, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.915360501567399e-06, |
|
"loss": 0.1218, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.911877394636016e-06, |
|
"loss": 0.1235, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.908394287704634e-06, |
|
"loss": 0.1222, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.90491118077325e-06, |
|
"loss": 0.1217, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.901428073841867e-06, |
|
"loss": 0.121, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.897944966910485e-06, |
|
"loss": 0.1178, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.894461859979102e-06, |
|
"loss": 0.1214, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"eval_loss": 0.1671142578125, |
|
"eval_runtime": 424.7805, |
|
"eval_samples_per_second": 6.707, |
|
"eval_steps_per_second": 0.054, |
|
"eval_wer": 29.554108058780955, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.89097875304772e-06, |
|
"loss": 0.1203, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.887495646116337e-06, |
|
"loss": 0.1154, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.884012539184953e-06, |
|
"loss": 0.1181, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.88052943225357e-06, |
|
"loss": 0.1124, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.877046325322188e-06, |
|
"loss": 0.1183, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.873563218390807e-06, |
|
"loss": 0.1114, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.870080111459423e-06, |
|
"loss": 0.1125, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.86659700452804e-06, |
|
"loss": 0.1138, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.863113897596656e-06, |
|
"loss": 0.1127, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.859630790665275e-06, |
|
"loss": 0.11, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"eval_loss": 0.159423828125, |
|
"eval_runtime": 741.6949, |
|
"eval_samples_per_second": 3.841, |
|
"eval_steps_per_second": 0.031, |
|
"eval_wer": 28.60064822681645, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.856147683733891e-06, |
|
"loss": 0.1093, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.85266457680251e-06, |
|
"loss": 0.1184, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.849181469871126e-06, |
|
"loss": 0.1098, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.845698362939743e-06, |
|
"loss": 0.1111, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.842215256008361e-06, |
|
"loss": 0.1048, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.838732149076977e-06, |
|
"loss": 0.1092, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.835249042145594e-06, |
|
"loss": 0.107, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.83176593521421e-06, |
|
"loss": 0.1136, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.828282828282829e-06, |
|
"loss": 0.1092, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.824799721351447e-06, |
|
"loss": 0.1116, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"eval_loss": 0.1539306640625, |
|
"eval_runtime": 290.1579, |
|
"eval_samples_per_second": 9.819, |
|
"eval_steps_per_second": 0.079, |
|
"eval_wer": 28.02794298121401, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.821316614420064e-06, |
|
"loss": 0.1067, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.81783350748868e-06, |
|
"loss": 0.1072, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.814350400557297e-06, |
|
"loss": 0.1105, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.810867293625915e-06, |
|
"loss": 0.108, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.807384186694532e-06, |
|
"loss": 0.1042, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.80390107976315e-06, |
|
"loss": 0.1092, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.800417972831767e-06, |
|
"loss": 0.105, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.796934865900383e-06, |
|
"loss": 0.1067, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.793451758969001e-06, |
|
"loss": 0.1066, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.789968652037618e-06, |
|
"loss": 0.104, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"eval_loss": 0.150634765625, |
|
"eval_runtime": 1070.3526, |
|
"eval_samples_per_second": 2.662, |
|
"eval_steps_per_second": 0.021, |
|
"eval_wer": 27.301047861795524, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.786485545106236e-06, |
|
"loss": 0.1086, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.783002438174853e-06, |
|
"loss": 0.1033, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.77951933124347e-06, |
|
"loss": 0.1054, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.776036224312088e-06, |
|
"loss": 0.1089, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.772553117380704e-06, |
|
"loss": 0.1043, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.769070010449323e-06, |
|
"loss": 0.1046, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.76558690351794e-06, |
|
"loss": 0.1, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.762103796586556e-06, |
|
"loss": 0.102, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.758620689655172e-06, |
|
"loss": 0.1049, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.75513758272379e-06, |
|
"loss": 0.1023, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"eval_loss": 0.146728515625, |
|
"eval_runtime": 692.4983, |
|
"eval_samples_per_second": 4.114, |
|
"eval_steps_per_second": 0.033, |
|
"eval_wer": 26.86679882941565, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.751654475792409e-06, |
|
"loss": 0.1022, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.748171368861024e-06, |
|
"loss": 0.0972, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 9.744688261929642e-06, |
|
"loss": 0.1008, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 9.741205154998259e-06, |
|
"loss": 0.1059, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 9.737722048066877e-06, |
|
"loss": 0.1021, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 9.734238941135494e-06, |
|
"loss": 0.1034, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 9.73075583420411e-06, |
|
"loss": 0.1015, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 9.727272727272728e-06, |
|
"loss": 0.1003, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 9.723789620341345e-06, |
|
"loss": 0.1025, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 9.720306513409963e-06, |
|
"loss": 0.0972, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"eval_loss": 0.1441650390625, |
|
"eval_runtime": 527.1217, |
|
"eval_samples_per_second": 5.405, |
|
"eval_steps_per_second": 0.044, |
|
"eval_wer": 25.756002391516407, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 9.71682340647858e-06, |
|
"loss": 0.1017, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 9.713340299547196e-06, |
|
"loss": 0.1021, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 9.709857192615813e-06, |
|
"loss": 0.1011, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 9.706374085684431e-06, |
|
"loss": 0.0979, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 9.70289097875305e-06, |
|
"loss": 0.0998, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 9.699407871821666e-06, |
|
"loss": 0.0966, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 9.695924764890283e-06, |
|
"loss": 0.099, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 9.6924416579589e-06, |
|
"loss": 0.0984, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 9.688958551027518e-06, |
|
"loss": 0.0986, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 9.685475444096134e-06, |
|
"loss": 0.0972, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"eval_loss": 0.1383056640625, |
|
"eval_runtime": 777.6072, |
|
"eval_samples_per_second": 3.664, |
|
"eval_steps_per_second": 0.03, |
|
"eval_wer": 25.350073948204788, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 9.681992337164752e-06, |
|
"loss": 0.0975, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 9.678509230233369e-06, |
|
"loss": 0.0978, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 9.675026123301986e-06, |
|
"loss": 0.0996, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 9.671543016370604e-06, |
|
"loss": 0.0978, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 9.66805990943922e-06, |
|
"loss": 0.1, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 9.664576802507839e-06, |
|
"loss": 0.0924, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 9.661093695576455e-06, |
|
"loss": 0.0998, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 9.657610588645072e-06, |
|
"loss": 0.0961, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 9.65412748171369e-06, |
|
"loss": 0.0979, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 9.650644374782307e-06, |
|
"loss": 0.0977, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"eval_loss": 0.13623046875, |
|
"eval_runtime": 782.2381, |
|
"eval_samples_per_second": 3.642, |
|
"eval_steps_per_second": 0.029, |
|
"eval_wer": 25.43188898329085, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 9.647161267850923e-06, |
|
"loss": 0.0979, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 9.64367816091954e-06, |
|
"loss": 0.0976, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 9.640195053988158e-06, |
|
"loss": 0.0981, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 9.636711947056775e-06, |
|
"loss": 0.0955, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 9.633228840125393e-06, |
|
"loss": 0.096, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 9.62974573319401e-06, |
|
"loss": 0.0944, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 9.626262626262626e-06, |
|
"loss": 0.0968, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 9.622779519331244e-06, |
|
"loss": 0.0917, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 9.619296412399861e-06, |
|
"loss": 0.0977, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 9.61581330546848e-06, |
|
"loss": 0.0934, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"eval_loss": 0.1353759765625, |
|
"eval_runtime": 476.9686, |
|
"eval_samples_per_second": 5.973, |
|
"eval_steps_per_second": 0.048, |
|
"eval_wer": 25.249378520406555, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 9.612330198537096e-06, |
|
"loss": 0.0918, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 9.608847091605712e-06, |
|
"loss": 0.097, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 9.60536398467433e-06, |
|
"loss": 0.0933, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 9.601880877742947e-06, |
|
"loss": 0.0944, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 9.598397770811566e-06, |
|
"loss": 0.0957, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 9.594914663880182e-06, |
|
"loss": 0.0948, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 9.591431556948799e-06, |
|
"loss": 0.0955, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 9.587948450017415e-06, |
|
"loss": 0.0967, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.584465343086034e-06, |
|
"loss": 0.0935, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.580982236154652e-06, |
|
"loss": 0.0965, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"eval_loss": 0.1326904296875, |
|
"eval_runtime": 295.4901, |
|
"eval_samples_per_second": 9.642, |
|
"eval_steps_per_second": 0.078, |
|
"eval_wer": 25.243085056169168, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.577499129223268e-06, |
|
"loss": 0.097, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.574016022291885e-06, |
|
"loss": 0.0891, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.570532915360502e-06, |
|
"loss": 0.0953, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.56704980842912e-06, |
|
"loss": 0.0948, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.563566701497736e-06, |
|
"loss": 0.0949, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.560083594566355e-06, |
|
"loss": 0.0911, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 9.556600487634971e-06, |
|
"loss": 0.0901, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 9.553117380703588e-06, |
|
"loss": 0.0953, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 9.549634273772206e-06, |
|
"loss": 0.0932, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 9.546151166840823e-06, |
|
"loss": 0.0937, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"eval_loss": 0.130126953125, |
|
"eval_runtime": 803.927, |
|
"eval_samples_per_second": 3.544, |
|
"eval_steps_per_second": 0.029, |
|
"eval_wer": 24.390320652002895, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 9.54266805990944e-06, |
|
"loss": 0.0912, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 9.539184952978056e-06, |
|
"loss": 0.092, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 9.535701846046674e-06, |
|
"loss": 0.0923, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 9.532218739115292e-06, |
|
"loss": 0.0886, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 9.528735632183909e-06, |
|
"loss": 0.0891, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 9.525252525252526e-06, |
|
"loss": 0.0864, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 9.521769418321142e-06, |
|
"loss": 0.0898, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 9.51828631138976e-06, |
|
"loss": 0.0913, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 9.514803204458377e-06, |
|
"loss": 0.089, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 9.511320097526995e-06, |
|
"loss": 0.0926, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"eval_loss": 0.12744140625, |
|
"eval_runtime": 779.0613, |
|
"eval_samples_per_second": 3.657, |
|
"eval_steps_per_second": 0.03, |
|
"eval_wer": 24.30221215267944, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 9.507836990595612e-06, |
|
"loss": 0.0934, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 9.504353883664229e-06, |
|
"loss": 0.0866, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 9.500870776732847e-06, |
|
"loss": 0.0901, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 9.497387669801463e-06, |
|
"loss": 0.0879, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 9.493904562870082e-06, |
|
"loss": 0.0925, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 9.490421455938698e-06, |
|
"loss": 0.0883, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 9.486938349007315e-06, |
|
"loss": 0.0904, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 9.483455242075933e-06, |
|
"loss": 0.0895, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 9.47997213514455e-06, |
|
"loss": 0.0944, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 9.476489028213168e-06, |
|
"loss": 0.0951, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"eval_loss": 0.1268310546875, |
|
"eval_runtime": 625.0326, |
|
"eval_samples_per_second": 4.558, |
|
"eval_steps_per_second": 0.037, |
|
"eval_wer": 24.51933666886938, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 9.473005921281785e-06, |
|
"loss": 0.0887, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 9.469522814350401e-06, |
|
"loss": 0.0914, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 9.466039707419018e-06, |
|
"loss": 0.0906, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 9.462556600487636e-06, |
|
"loss": 0.0908, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 9.459073493556253e-06, |
|
"loss": 0.0901, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 9.455590386624869e-06, |
|
"loss": 0.088, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 9.452107279693487e-06, |
|
"loss": 0.0825, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 9.448624172762104e-06, |
|
"loss": 0.0909, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 9.445141065830722e-06, |
|
"loss": 0.0896, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 9.441657958899339e-06, |
|
"loss": 0.0922, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"eval_loss": 0.1253662109375, |
|
"eval_runtime": 548.6186, |
|
"eval_samples_per_second": 5.193, |
|
"eval_steps_per_second": 0.042, |
|
"eval_wer": 23.770414424620032, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 9.438174851967955e-06, |
|
"loss": 0.0845, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 9.434691745036574e-06, |
|
"loss": 0.0843, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 9.43120863810519e-06, |
|
"loss": 0.0826, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 9.427725531173809e-06, |
|
"loss": 0.0879, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 9.424242424242425e-06, |
|
"loss": 0.0879, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 9.420759317311042e-06, |
|
"loss": 0.0885, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 9.417276210379658e-06, |
|
"loss": 0.09, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 9.413793103448277e-06, |
|
"loss": 0.0898, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 9.410309996516895e-06, |
|
"loss": 0.0884, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 9.406826889585511e-06, |
|
"loss": 0.0881, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"eval_loss": 0.1258544921875, |
|
"eval_runtime": 767.5109, |
|
"eval_samples_per_second": 3.712, |
|
"eval_steps_per_second": 0.03, |
|
"eval_wer": 23.386513106139276, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 9.403343782654128e-06, |
|
"loss": 0.0888, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 9.399860675722745e-06, |
|
"loss": 0.0862, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 9.396377568791363e-06, |
|
"loss": 0.0918, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 9.39289446185998e-06, |
|
"loss": 0.0899, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 9.389411354928598e-06, |
|
"loss": 0.0854, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 9.385928247997214e-06, |
|
"loss": 0.0858, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 9.382445141065831e-06, |
|
"loss": 0.0898, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 9.378962034134449e-06, |
|
"loss": 0.0873, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 9.375478927203066e-06, |
|
"loss": 0.087, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 9.371995820271684e-06, |
|
"loss": 0.0867, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"eval_loss": 0.12060546875, |
|
"eval_runtime": 595.8484, |
|
"eval_samples_per_second": 4.781, |
|
"eval_steps_per_second": 0.039, |
|
"eval_wer": 23.399100034614055, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 9.3685127133403e-06, |
|
"loss": 0.0865, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 9.365029606408917e-06, |
|
"loss": 0.0868, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 9.361546499477535e-06, |
|
"loss": 0.0851, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 9.358063392546152e-06, |
|
"loss": 0.0825, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 9.354580285614769e-06, |
|
"loss": 0.0871, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 9.351097178683385e-06, |
|
"loss": 0.0876, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 9.347614071752003e-06, |
|
"loss": 0.0842, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 9.34413096482062e-06, |
|
"loss": 0.0854, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 9.340647857889238e-06, |
|
"loss": 0.0883, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 9.337164750957855e-06, |
|
"loss": 0.0818, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"eval_loss": 0.1219482421875, |
|
"eval_runtime": 360.525, |
|
"eval_samples_per_second": 7.902, |
|
"eval_steps_per_second": 0.064, |
|
"eval_wer": 23.2512036250354, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 9.333681644026471e-06, |
|
"loss": 0.0826, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 9.33019853709509e-06, |
|
"loss": 0.0823, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 9.326715430163706e-06, |
|
"loss": 0.0845, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 9.323232323232325e-06, |
|
"loss": 0.0814, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 9.319749216300941e-06, |
|
"loss": 0.088, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 9.316266109369558e-06, |
|
"loss": 0.0844, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 9.312783002438176e-06, |
|
"loss": 0.0867, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 9.309299895506793e-06, |
|
"loss": 0.083, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 9.305816788575411e-06, |
|
"loss": 0.0814, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 9.302333681644027e-06, |
|
"loss": 0.0857, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"eval_loss": 0.12091064453125, |
|
"eval_runtime": 790.1779, |
|
"eval_samples_per_second": 3.606, |
|
"eval_steps_per_second": 0.029, |
|
"eval_wer": 23.18512225054281, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 9.298850574712644e-06, |
|
"loss": 0.0849, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 9.295367467781262e-06, |
|
"loss": 0.0821, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 9.291884360849879e-06, |
|
"loss": 0.0845, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 9.288401253918497e-06, |
|
"loss": 0.0839, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 9.284918146987114e-06, |
|
"loss": 0.0839, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 9.28143504005573e-06, |
|
"loss": 0.0827, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 9.277951933124347e-06, |
|
"loss": 0.0835, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 9.274468826192965e-06, |
|
"loss": 0.0851, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 9.270985719261583e-06, |
|
"loss": 0.0833, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 9.267502612330198e-06, |
|
"loss": 0.0856, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"eval_loss": 0.1209716796875, |
|
"eval_runtime": 462.8014, |
|
"eval_samples_per_second": 6.156, |
|
"eval_steps_per_second": 0.05, |
|
"eval_wer": 23.074986626388498, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 9.264019505398817e-06, |
|
"loss": 0.0829, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 9.260536398467433e-06, |
|
"loss": 0.0809, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 9.257053291536052e-06, |
|
"loss": 0.0838, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 9.253570184604668e-06, |
|
"loss": 0.0845, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 9.250087077673285e-06, |
|
"loss": 0.0813, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 9.246603970741903e-06, |
|
"loss": 0.0866, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 9.24312086381052e-06, |
|
"loss": 0.0851, |
|
"step": 2470 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 9.239637756879138e-06, |
|
"loss": 0.0822, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 9.236154649947754e-06, |
|
"loss": 0.0789, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 9.232671543016371e-06, |
|
"loss": 0.0812, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"eval_loss": 0.1201171875, |
|
"eval_runtime": 682.3065, |
|
"eval_samples_per_second": 4.176, |
|
"eval_steps_per_second": 0.034, |
|
"eval_wer": 23.134774536643697, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 9.229188436084988e-06, |
|
"loss": 0.0859, |
|
"step": 2510 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 9.225705329153606e-06, |
|
"loss": 0.0801, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 9.222222222222224e-06, |
|
"loss": 0.0819, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 9.21873911529084e-06, |
|
"loss": 0.0807, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 9.215256008359457e-06, |
|
"loss": 0.085, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 9.211772901428074e-06, |
|
"loss": 0.0822, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 9.208289794496692e-06, |
|
"loss": 0.0837, |
|
"step": 2570 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 9.204806687565309e-06, |
|
"loss": 0.0865, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 9.201323580633927e-06, |
|
"loss": 0.0851, |
|
"step": 2590 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 9.197840473702544e-06, |
|
"loss": 0.0825, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"eval_loss": 0.11773681640625, |
|
"eval_runtime": 780.5221, |
|
"eval_samples_per_second": 3.65, |
|
"eval_steps_per_second": 0.029, |
|
"eval_wer": 22.73828628968816, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 9.19435736677116e-06, |
|
"loss": 0.0857, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 9.190874259839778e-06, |
|
"loss": 0.0839, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 9.187391152908395e-06, |
|
"loss": 0.0814, |
|
"step": 2630 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 9.183908045977013e-06, |
|
"loss": 0.0841, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 9.18042493904563e-06, |
|
"loss": 0.08, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 9.176941832114246e-06, |
|
"loss": 0.0842, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 9.173458725182865e-06, |
|
"loss": 0.0826, |
|
"step": 2670 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 9.169975618251481e-06, |
|
"loss": 0.082, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 9.166492511320098e-06, |
|
"loss": 0.0838, |
|
"step": 2690 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 9.163009404388714e-06, |
|
"loss": 0.0824, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"eval_loss": 0.11480712890625, |
|
"eval_runtime": 515.5719, |
|
"eval_samples_per_second": 5.526, |
|
"eval_steps_per_second": 0.045, |
|
"eval_wer": 22.555775826803863, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 9.159526297457333e-06, |
|
"loss": 0.0819, |
|
"step": 2710 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 9.15604319052595e-06, |
|
"loss": 0.08, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 9.152560083594568e-06, |
|
"loss": 0.0785, |
|
"step": 2730 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 9.149076976663184e-06, |
|
"loss": 0.081, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 9.1455938697318e-06, |
|
"loss": 0.0728, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 9.142110762800419e-06, |
|
"loss": 0.0836, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 9.138627655869036e-06, |
|
"loss": 0.0779, |
|
"step": 2770 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 9.135144548937654e-06, |
|
"loss": 0.0773, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 9.13166144200627e-06, |
|
"loss": 0.0816, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 9.128178335074887e-06, |
|
"loss": 0.0801, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"eval_loss": 0.1163330078125, |
|
"eval_runtime": 770.6819, |
|
"eval_samples_per_second": 3.697, |
|
"eval_steps_per_second": 0.03, |
|
"eval_wer": 22.228515686459613, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 9.124695228143505e-06, |
|
"loss": 0.0821, |
|
"step": 2810 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 9.121212121212122e-06, |
|
"loss": 0.0825, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 9.11772901428074e-06, |
|
"loss": 0.0816, |
|
"step": 2830 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 9.114245907349357e-06, |
|
"loss": 0.0789, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 9.110762800417973e-06, |
|
"loss": 0.0835, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 9.10727969348659e-06, |
|
"loss": 0.0807, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 9.103796586555208e-06, |
|
"loss": 0.083, |
|
"step": 2870 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 9.100313479623826e-06, |
|
"loss": 0.0773, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 9.096830372692443e-06, |
|
"loss": 0.0709, |
|
"step": 2890 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 9.09334726576106e-06, |
|
"loss": 0.0753, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"eval_loss": 0.1170654296875, |
|
"eval_runtime": 291.6827, |
|
"eval_samples_per_second": 9.767, |
|
"eval_steps_per_second": 0.079, |
|
"eval_wer": 22.228515686459613, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 9.089864158829676e-06, |
|
"loss": 0.0767, |
|
"step": 2910 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 9.086381051898294e-06, |
|
"loss": 0.0717, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 9.082897944966911e-06, |
|
"loss": 0.0736, |
|
"step": 2930 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 9.07941483803553e-06, |
|
"loss": 0.0724, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 9.075931731104146e-06, |
|
"loss": 0.0726, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 9.072448624172762e-06, |
|
"loss": 0.0723, |
|
"step": 2960 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 9.06896551724138e-06, |
|
"loss": 0.0735, |
|
"step": 2970 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 9.065482410309997e-06, |
|
"loss": 0.0724, |
|
"step": 2980 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 9.061999303378614e-06, |
|
"loss": 0.07, |
|
"step": 2990 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 9.05851619644723e-06, |
|
"loss": 0.0698, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"eval_loss": 0.1173095703125, |
|
"eval_runtime": 667.2017, |
|
"eval_samples_per_second": 4.27, |
|
"eval_steps_per_second": 0.034, |
|
"eval_wer": 22.59983007646559, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 9.055033089515849e-06, |
|
"loss": 0.0736, |
|
"step": 3010 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 9.051549982584467e-06, |
|
"loss": 0.0697, |
|
"step": 3020 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 9.048066875653084e-06, |
|
"loss": 0.0716, |
|
"step": 3030 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 9.0445837687217e-06, |
|
"loss": 0.0685, |
|
"step": 3040 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 9.041100661790317e-06, |
|
"loss": 0.0744, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 9.037617554858935e-06, |
|
"loss": 0.0712, |
|
"step": 3060 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 9.034134447927552e-06, |
|
"loss": 0.0708, |
|
"step": 3070 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 9.03065134099617e-06, |
|
"loss": 0.0739, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 9.027168234064786e-06, |
|
"loss": 0.0712, |
|
"step": 3090 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 9.023685127133403e-06, |
|
"loss": 0.0738, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"eval_loss": 0.11474609375, |
|
"eval_runtime": 772.0192, |
|
"eval_samples_per_second": 3.69, |
|
"eval_steps_per_second": 0.03, |
|
"eval_wer": 22.162434311967022, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 9.020202020202021e-06, |
|
"loss": 0.0725, |
|
"step": 3110 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 9.016718913270638e-06, |
|
"loss": 0.0753, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 9.013235806339256e-06, |
|
"loss": 0.0705, |
|
"step": 3130 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 9.009752699407873e-06, |
|
"loss": 0.0715, |
|
"step": 3140 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 9.00626959247649e-06, |
|
"loss": 0.0774, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 9.002786485545108e-06, |
|
"loss": 0.0714, |
|
"step": 3160 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 8.999303378613724e-06, |
|
"loss": 0.0699, |
|
"step": 3170 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 8.995820271682342e-06, |
|
"loss": 0.0756, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 8.992337164750959e-06, |
|
"loss": 0.0715, |
|
"step": 3190 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 8.988854057819576e-06, |
|
"loss": 0.0704, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"eval_loss": 0.11749267578125, |
|
"eval_runtime": 486.4812, |
|
"eval_samples_per_second": 5.856, |
|
"eval_steps_per_second": 0.047, |
|
"eval_wer": 22.017684634507066, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 8.985370950888192e-06, |
|
"loss": 0.0723, |
|
"step": 3210 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 8.98188784395681e-06, |
|
"loss": 0.0726, |
|
"step": 3220 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 8.978404737025427e-06, |
|
"loss": 0.0711, |
|
"step": 3230 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 8.974921630094044e-06, |
|
"loss": 0.0704, |
|
"step": 3240 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 8.971438523162662e-06, |
|
"loss": 0.0716, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 8.967955416231279e-06, |
|
"loss": 0.0723, |
|
"step": 3260 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 8.964472309299897e-06, |
|
"loss": 0.0751, |
|
"step": 3270 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 8.960989202368513e-06, |
|
"loss": 0.072, |
|
"step": 3280 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 8.95750609543713e-06, |
|
"loss": 0.0707, |
|
"step": 3290 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 8.954022988505748e-06, |
|
"loss": 0.0671, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"eval_loss": 0.11810302734375, |
|
"eval_runtime": 700.0763, |
|
"eval_samples_per_second": 4.07, |
|
"eval_steps_per_second": 0.033, |
|
"eval_wer": 22.03341829510054, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 8.950539881574365e-06, |
|
"loss": 0.0725, |
|
"step": 3310 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 8.947056774642983e-06, |
|
"loss": 0.0762, |
|
"step": 3320 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 8.9435736677116e-06, |
|
"loss": 0.0706, |
|
"step": 3330 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 8.940090560780216e-06, |
|
"loss": 0.0692, |
|
"step": 3340 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 8.936607453848833e-06, |
|
"loss": 0.0739, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 8.933124346917451e-06, |
|
"loss": 0.0751, |
|
"step": 3360 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 8.92964123998607e-06, |
|
"loss": 0.0733, |
|
"step": 3370 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 8.926158133054686e-06, |
|
"loss": 0.0692, |
|
"step": 3380 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 8.922675026123303e-06, |
|
"loss": 0.0713, |
|
"step": 3390 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 8.919191919191919e-06, |
|
"loss": 0.0699, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"eval_loss": 0.11517333984375, |
|
"eval_runtime": 774.6303, |
|
"eval_samples_per_second": 3.678, |
|
"eval_steps_per_second": 0.03, |
|
"eval_wer": 22.04285849145662, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 8.915708812260537e-06, |
|
"loss": 0.0737, |
|
"step": 3410 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 8.912225705329154e-06, |
|
"loss": 0.0709, |
|
"step": 3420 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 8.908742598397772e-06, |
|
"loss": 0.0702, |
|
"step": 3430 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 8.905259491466389e-06, |
|
"loss": 0.0685, |
|
"step": 3440 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 8.901776384535005e-06, |
|
"loss": 0.0739, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 8.898293277603624e-06, |
|
"loss": 0.0718, |
|
"step": 3460 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 8.89481017067224e-06, |
|
"loss": 0.0717, |
|
"step": 3470 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 8.891327063740859e-06, |
|
"loss": 0.0744, |
|
"step": 3480 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 8.887843956809475e-06, |
|
"loss": 0.0717, |
|
"step": 3490 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 8.884360849878092e-06, |
|
"loss": 0.0686, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"eval_loss": 0.11749267578125, |
|
"eval_runtime": 492.6786, |
|
"eval_samples_per_second": 5.783, |
|
"eval_steps_per_second": 0.047, |
|
"eval_wer": 22.231662418578306, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 8.88087774294671e-06, |
|
"loss": 0.0727, |
|
"step": 3510 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 8.877394636015327e-06, |
|
"loss": 0.0704, |
|
"step": 3520 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 8.873911529083943e-06, |
|
"loss": 0.0737, |
|
"step": 3530 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 8.87042842215256e-06, |
|
"loss": 0.0728, |
|
"step": 3540 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 8.866945315221178e-06, |
|
"loss": 0.0764, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 8.863462208289795e-06, |
|
"loss": 0.0745, |
|
"step": 3560 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 8.859979101358413e-06, |
|
"loss": 0.0688, |
|
"step": 3570 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 8.85649599442703e-06, |
|
"loss": 0.0715, |
|
"step": 3580 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 8.853012887495646e-06, |
|
"loss": 0.0704, |
|
"step": 3590 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 8.849529780564264e-06, |
|
"loss": 0.0742, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"eval_loss": 0.113525390625, |
|
"eval_runtime": 293.9147, |
|
"eval_samples_per_second": 9.693, |
|
"eval_steps_per_second": 0.078, |
|
"eval_wer": 22.181314704679192, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 8.846046673632881e-06, |
|
"loss": 0.0726, |
|
"step": 3610 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 8.8425635667015e-06, |
|
"loss": 0.0731, |
|
"step": 3620 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 8.839080459770116e-06, |
|
"loss": 0.0712, |
|
"step": 3630 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 8.835597352838732e-06, |
|
"loss": 0.0672, |
|
"step": 3640 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 8.83211424590735e-06, |
|
"loss": 0.0704, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 8.828631138975967e-06, |
|
"loss": 0.0721, |
|
"step": 3660 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 8.825148032044585e-06, |
|
"loss": 0.0705, |
|
"step": 3670 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 8.821664925113202e-06, |
|
"loss": 0.0722, |
|
"step": 3680 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 8.818181818181819e-06, |
|
"loss": 0.0749, |
|
"step": 3690 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 8.814698711250435e-06, |
|
"loss": 0.0698, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"eval_loss": 0.113037109375, |
|
"eval_runtime": 720.2659, |
|
"eval_samples_per_second": 3.955, |
|
"eval_steps_per_second": 0.032, |
|
"eval_wer": 22.03027156298184, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 8.811215604319053e-06, |
|
"loss": 0.0668, |
|
"step": 3710 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 8.807732497387672e-06, |
|
"loss": 0.0712, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 8.804249390456288e-06, |
|
"loss": 0.068, |
|
"step": 3730 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 8.800766283524905e-06, |
|
"loss": 0.07, |
|
"step": 3740 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 8.797283176593521e-06, |
|
"loss": 0.0704, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 8.79380006966214e-06, |
|
"loss": 0.0703, |
|
"step": 3760 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 8.790316962730756e-06, |
|
"loss": 0.0739, |
|
"step": 3770 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 8.786833855799373e-06, |
|
"loss": 0.0736, |
|
"step": 3780 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 8.783350748867991e-06, |
|
"loss": 0.0684, |
|
"step": 3790 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 8.779867641936608e-06, |
|
"loss": 0.0712, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"eval_loss": 0.11163330078125, |
|
"eval_runtime": 782.2761, |
|
"eval_samples_per_second": 3.642, |
|
"eval_steps_per_second": 0.029, |
|
"eval_wer": 21.920135938827528, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 8.776384535005226e-06, |
|
"loss": 0.0717, |
|
"step": 3810 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 8.772901428073843e-06, |
|
"loss": 0.0703, |
|
"step": 3820 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 8.76941832114246e-06, |
|
"loss": 0.073, |
|
"step": 3830 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 8.765935214211076e-06, |
|
"loss": 0.0728, |
|
"step": 3840 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 8.762452107279694e-06, |
|
"loss": 0.0713, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 8.758969000348312e-06, |
|
"loss": 0.0701, |
|
"step": 3860 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 8.755485893416929e-06, |
|
"loss": 0.0732, |
|
"step": 3870 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 8.752002786485546e-06, |
|
"loss": 0.0734, |
|
"step": 3880 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 8.748519679554162e-06, |
|
"loss": 0.07, |
|
"step": 3890 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 8.74503657262278e-06, |
|
"loss": 0.0705, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"eval_loss": 0.114990234375, |
|
"eval_runtime": 518.7116, |
|
"eval_samples_per_second": 5.492, |
|
"eval_steps_per_second": 0.044, |
|
"eval_wer": 21.680984297806727, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 8.741553465691397e-06, |
|
"loss": 0.0702, |
|
"step": 3910 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 8.738070358760015e-06, |
|
"loss": 0.0704, |
|
"step": 3920 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 8.734587251828632e-06, |
|
"loss": 0.0689, |
|
"step": 3930 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 8.731104144897248e-06, |
|
"loss": 0.0709, |
|
"step": 3940 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 8.727621037965867e-06, |
|
"loss": 0.07, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 8.724137931034483e-06, |
|
"loss": 0.0722, |
|
"step": 3960 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 8.720654824103102e-06, |
|
"loss": 0.0717, |
|
"step": 3970 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 8.717171717171718e-06, |
|
"loss": 0.0716, |
|
"step": 3980 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 8.713688610240335e-06, |
|
"loss": 0.068, |
|
"step": 3990 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 8.710205503308953e-06, |
|
"loss": 0.0701, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"eval_loss": 0.11181640625, |
|
"eval_runtime": 775.9038, |
|
"eval_samples_per_second": 3.672, |
|
"eval_steps_per_second": 0.03, |
|
"eval_wer": 21.636930048145, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 8.70672239637757e-06, |
|
"loss": 0.0732, |
|
"step": 4010 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 8.703239289446188e-06, |
|
"loss": 0.0666, |
|
"step": 4020 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 8.699756182514804e-06, |
|
"loss": 0.0701, |
|
"step": 4030 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 8.696273075583421e-06, |
|
"loss": 0.0726, |
|
"step": 4040 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 8.692789968652038e-06, |
|
"loss": 0.0724, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 8.689306861720656e-06, |
|
"loss": 0.0701, |
|
"step": 4060 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 8.685823754789272e-06, |
|
"loss": 0.0713, |
|
"step": 4070 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 8.682340647857889e-06, |
|
"loss": 0.0689, |
|
"step": 4080 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 8.678857540926507e-06, |
|
"loss": 0.0707, |
|
"step": 4090 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 8.675374433995124e-06, |
|
"loss": 0.0675, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"eval_loss": 0.1124267578125, |
|
"eval_runtime": 663.0481, |
|
"eval_samples_per_second": 4.297, |
|
"eval_steps_per_second": 0.035, |
|
"eval_wer": 21.68413102992542, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 8.671891327063742e-06, |
|
"loss": 0.0713, |
|
"step": 4110 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 8.668408220132359e-06, |
|
"loss": 0.0678, |
|
"step": 4120 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 8.664925113200975e-06, |
|
"loss": 0.0729, |
|
"step": 4130 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 8.661442006269594e-06, |
|
"loss": 0.0703, |
|
"step": 4140 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 8.65795889933821e-06, |
|
"loss": 0.0749, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 8.654475792406828e-06, |
|
"loss": 0.0687, |
|
"step": 4160 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 8.650992685475445e-06, |
|
"loss": 0.0697, |
|
"step": 4170 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 8.647509578544062e-06, |
|
"loss": 0.0698, |
|
"step": 4180 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 8.644026471612678e-06, |
|
"loss": 0.0722, |
|
"step": 4190 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 8.640543364681296e-06, |
|
"loss": 0.0671, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"eval_loss": 0.109130859375, |
|
"eval_runtime": 486.9376, |
|
"eval_samples_per_second": 5.851, |
|
"eval_steps_per_second": 0.047, |
|
"eval_wer": 21.253028729664244, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 8.637060257749915e-06, |
|
"loss": 0.0704, |
|
"step": 4210 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 8.633577150818531e-06, |
|
"loss": 0.0704, |
|
"step": 4220 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 8.630094043887148e-06, |
|
"loss": 0.073, |
|
"step": 4230 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 8.626610936955764e-06, |
|
"loss": 0.0737, |
|
"step": 4240 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 8.623127830024383e-06, |
|
"loss": 0.0699, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 8.619644723093e-06, |
|
"loss": 0.0694, |
|
"step": 4260 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 8.616161616161618e-06, |
|
"loss": 0.069, |
|
"step": 4270 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 8.612678509230234e-06, |
|
"loss": 0.0711, |
|
"step": 4280 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 8.60919540229885e-06, |
|
"loss": 0.0695, |
|
"step": 4290 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 8.605712295367469e-06, |
|
"loss": 0.0673, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"eval_loss": 0.1119384765625, |
|
"eval_runtime": 294.9453, |
|
"eval_samples_per_second": 9.659, |
|
"eval_steps_per_second": 0.078, |
|
"eval_wer": 21.341137228987698, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 8.602229188436086e-06, |
|
"loss": 0.0731, |
|
"step": 4310 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 8.598746081504704e-06, |
|
"loss": 0.066, |
|
"step": 4320 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 8.59526297457332e-06, |
|
"loss": 0.0702, |
|
"step": 4330 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 8.591779867641937e-06, |
|
"loss": 0.0704, |
|
"step": 4340 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 8.588296760710555e-06, |
|
"loss": 0.0701, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 8.584813653779172e-06, |
|
"loss": 0.0736, |
|
"step": 4360 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 8.581330546847788e-06, |
|
"loss": 0.07, |
|
"step": 4370 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 8.577847439916405e-06, |
|
"loss": 0.0715, |
|
"step": 4380 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 8.574364332985023e-06, |
|
"loss": 0.0687, |
|
"step": 4390 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 8.57088122605364e-06, |
|
"loss": 0.0719, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"eval_loss": 0.111572265625, |
|
"eval_runtime": 650.2805, |
|
"eval_samples_per_second": 4.381, |
|
"eval_steps_per_second": 0.035, |
|
"eval_wer": 21.385191478649425, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 8.567398119122258e-06, |
|
"loss": 0.0691, |
|
"step": 4410 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 8.563915012190875e-06, |
|
"loss": 0.0664, |
|
"step": 4420 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 8.560431905259491e-06, |
|
"loss": 0.0696, |
|
"step": 4430 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 8.55694879832811e-06, |
|
"loss": 0.0721, |
|
"step": 4440 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 8.553465691396726e-06, |
|
"loss": 0.0717, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 8.549982584465344e-06, |
|
"loss": 0.0684, |
|
"step": 4460 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 8.546499477533961e-06, |
|
"loss": 0.0702, |
|
"step": 4470 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 8.543016370602578e-06, |
|
"loss": 0.0689, |
|
"step": 4480 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 8.539533263671196e-06, |
|
"loss": 0.0721, |
|
"step": 4490 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 8.536050156739812e-06, |
|
"loss": 0.0711, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"eval_loss": 0.110595703125, |
|
"eval_runtime": 785.4826, |
|
"eval_samples_per_second": 3.627, |
|
"eval_steps_per_second": 0.029, |
|
"eval_wer": 21.407218603480285, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 8.53256704980843e-06, |
|
"loss": 0.0718, |
|
"step": 4510 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 8.529083942877047e-06, |
|
"loss": 0.0684, |
|
"step": 4520 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 8.525600835945664e-06, |
|
"loss": 0.0664, |
|
"step": 4530 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 8.52211772901428e-06, |
|
"loss": 0.0626, |
|
"step": 4540 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 8.518634622082899e-06, |
|
"loss": 0.0687, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 8.515151515151517e-06, |
|
"loss": 0.066, |
|
"step": 4560 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 8.511668408220134e-06, |
|
"loss": 0.0724, |
|
"step": 4570 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 8.50818530128875e-06, |
|
"loss": 0.0699, |
|
"step": 4580 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 8.504702194357367e-06, |
|
"loss": 0.0677, |
|
"step": 4590 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 8.501219087425985e-06, |
|
"loss": 0.071, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"eval_loss": 0.10986328125, |
|
"eval_runtime": 486.6542, |
|
"eval_samples_per_second": 5.854, |
|
"eval_steps_per_second": 0.047, |
|
"eval_wer": 21.347430693225085, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 8.497735980494602e-06, |
|
"loss": 0.0706, |
|
"step": 4610 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 8.494252873563218e-06, |
|
"loss": 0.0663, |
|
"step": 4620 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 8.490769766631836e-06, |
|
"loss": 0.0686, |
|
"step": 4630 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 8.487286659700453e-06, |
|
"loss": 0.0689, |
|
"step": 4640 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 8.483803552769071e-06, |
|
"loss": 0.0702, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 8.480320445837688e-06, |
|
"loss": 0.0679, |
|
"step": 4660 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 8.476837338906305e-06, |
|
"loss": 0.069, |
|
"step": 4670 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 8.473354231974921e-06, |
|
"loss": 0.0677, |
|
"step": 4680 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 8.46987112504354e-06, |
|
"loss": 0.0749, |
|
"step": 4690 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 8.466388018112158e-06, |
|
"loss": 0.0702, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"eval_loss": 0.1082763671875, |
|
"eval_runtime": 726.1025, |
|
"eval_samples_per_second": 3.924, |
|
"eval_steps_per_second": 0.032, |
|
"eval_wer": 21.14603983762862, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 8.462904911180774e-06, |
|
"loss": 0.069, |
|
"step": 4710 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 8.45942180424939e-06, |
|
"loss": 0.0668, |
|
"step": 4720 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 8.455938697318007e-06, |
|
"loss": 0.0685, |
|
"step": 4730 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 8.452455590386626e-06, |
|
"loss": 0.0706, |
|
"step": 4740 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 8.448972483455242e-06, |
|
"loss": 0.0693, |
|
"step": 4750 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 8.44548937652386e-06, |
|
"loss": 0.0677, |
|
"step": 4760 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 8.442006269592477e-06, |
|
"loss": 0.0719, |
|
"step": 4770 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 8.438523162661094e-06, |
|
"loss": 0.0692, |
|
"step": 4780 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 8.435040055729712e-06, |
|
"loss": 0.0671, |
|
"step": 4790 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 8.431556948798329e-06, |
|
"loss": 0.0686, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"eval_loss": 0.1087646484375, |
|
"eval_runtime": 787.116, |
|
"eval_samples_per_second": 3.62, |
|
"eval_steps_per_second": 0.029, |
|
"eval_wer": 21.725038547468454, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 8.428073841866947e-06, |
|
"loss": 0.066, |
|
"step": 4810 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 8.424590734935563e-06, |
|
"loss": 0.0661, |
|
"step": 4820 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 8.42110762800418e-06, |
|
"loss": 0.0661, |
|
"step": 4830 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 8.417624521072798e-06, |
|
"loss": 0.0677, |
|
"step": 4840 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 8.414141414141415e-06, |
|
"loss": 0.0678, |
|
"step": 4850 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 8.410658307210033e-06, |
|
"loss": 0.0699, |
|
"step": 4860 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 8.40717520027865e-06, |
|
"loss": 0.0696, |
|
"step": 4870 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 8.403692093347266e-06, |
|
"loss": 0.073, |
|
"step": 4880 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 8.400208986415883e-06, |
|
"loss": 0.0693, |
|
"step": 4890 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 8.396725879484501e-06, |
|
"loss": 0.0643, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"eval_loss": 0.10882568359375, |
|
"eval_runtime": 484.8642, |
|
"eval_samples_per_second": 5.876, |
|
"eval_steps_per_second": 0.047, |
|
"eval_wer": 21.312816639919443, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 8.393242772553118e-06, |
|
"loss": 0.0684, |
|
"step": 4910 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 8.389759665621734e-06, |
|
"loss": 0.0676, |
|
"step": 4920 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 8.386276558690353e-06, |
|
"loss": 0.0673, |
|
"step": 4930 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 8.382793451758969e-06, |
|
"loss": 0.0712, |
|
"step": 4940 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 8.379310344827587e-06, |
|
"loss": 0.0666, |
|
"step": 4950 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 8.375827237896204e-06, |
|
"loss": 0.0678, |
|
"step": 4960 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 8.37234413096482e-06, |
|
"loss": 0.0686, |
|
"step": 4970 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 8.368861024033439e-06, |
|
"loss": 0.0696, |
|
"step": 4980 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 8.365377917102055e-06, |
|
"loss": 0.0691, |
|
"step": 4990 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 8.361894810170674e-06, |
|
"loss": 0.0669, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"eval_loss": 0.1083984375, |
|
"eval_runtime": 302.723, |
|
"eval_samples_per_second": 9.411, |
|
"eval_steps_per_second": 0.076, |
|
"eval_wer": 21.199534283646436, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 8.359108324625566e-06, |
|
"loss": 0.0652, |
|
"step": 5010 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 8.355625217694185e-06, |
|
"loss": 0.0708, |
|
"step": 5020 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 8.352142110762801e-06, |
|
"loss": 0.0678, |
|
"step": 5030 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 8.34865900383142e-06, |
|
"loss": 0.0698, |
|
"step": 5040 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 8.345175896900036e-06, |
|
"loss": 0.0677, |
|
"step": 5050 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 8.341692789968653e-06, |
|
"loss": 0.0715, |
|
"step": 5060 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 8.33820968303727e-06, |
|
"loss": 0.0699, |
|
"step": 5070 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 8.334726576105887e-06, |
|
"loss": 0.0701, |
|
"step": 5080 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 8.331243469174504e-06, |
|
"loss": 0.07, |
|
"step": 5090 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 8.32776036224312e-06, |
|
"loss": 0.0712, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"eval_loss": 0.106201171875, |
|
"eval_runtime": 311.8176, |
|
"eval_samples_per_second": 9.137, |
|
"eval_steps_per_second": 0.074, |
|
"eval_wer": 21.13030617703515, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 8.324277255311739e-06, |
|
"loss": 0.0678, |
|
"step": 5110 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 8.320794148380355e-06, |
|
"loss": 0.0679, |
|
"step": 5120 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 8.317311041448974e-06, |
|
"loss": 0.0691, |
|
"step": 5130 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 8.31382793451759e-06, |
|
"loss": 0.0693, |
|
"step": 5140 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 8.310344827586207e-06, |
|
"loss": 0.0668, |
|
"step": 5150 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 8.306861720654825e-06, |
|
"loss": 0.0666, |
|
"step": 5160 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 8.303378613723442e-06, |
|
"loss": 0.0663, |
|
"step": 5170 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 8.29989550679206e-06, |
|
"loss": 0.0683, |
|
"step": 5180 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 8.296412399860677e-06, |
|
"loss": 0.0663, |
|
"step": 5190 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 8.292929292929293e-06, |
|
"loss": 0.066, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"eval_loss": 0.10882568359375, |
|
"eval_runtime": 328.1314, |
|
"eval_samples_per_second": 8.682, |
|
"eval_steps_per_second": 0.07, |
|
"eval_wer": 21.34428396110639, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 8.28944618599791e-06, |
|
"loss": 0.0708, |
|
"step": 5210 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 8.285963079066528e-06, |
|
"loss": 0.0695, |
|
"step": 5220 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 8.282479972135146e-06, |
|
"loss": 0.0647, |
|
"step": 5230 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 8.278996865203763e-06, |
|
"loss": 0.0673, |
|
"step": 5240 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 8.27551375827238e-06, |
|
"loss": 0.0675, |
|
"step": 5250 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 8.272030651340996e-06, |
|
"loss": 0.0672, |
|
"step": 5260 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 8.268547544409614e-06, |
|
"loss": 0.0693, |
|
"step": 5270 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 8.265064437478231e-06, |
|
"loss": 0.0662, |
|
"step": 5280 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 8.26158133054685e-06, |
|
"loss": 0.0683, |
|
"step": 5290 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 8.258098223615466e-06, |
|
"loss": 0.0688, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"eval_loss": 0.10693359375, |
|
"eval_runtime": 683.9313, |
|
"eval_samples_per_second": 4.166, |
|
"eval_steps_per_second": 0.034, |
|
"eval_wer": 20.994996695931274, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 8.254615116684082e-06, |
|
"loss": 0.0683, |
|
"step": 5310 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 8.2511320097527e-06, |
|
"loss": 0.066, |
|
"step": 5320 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 8.247648902821317e-06, |
|
"loss": 0.066, |
|
"step": 5330 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 8.244165795889934e-06, |
|
"loss": 0.073, |
|
"step": 5340 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 8.24068268895855e-06, |
|
"loss": 0.0679, |
|
"step": 5350 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 8.237199582027169e-06, |
|
"loss": 0.0707, |
|
"step": 5360 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 8.233716475095787e-06, |
|
"loss": 0.0693, |
|
"step": 5370 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 8.230233368164403e-06, |
|
"loss": 0.065, |
|
"step": 5380 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 8.22675026123302e-06, |
|
"loss": 0.0672, |
|
"step": 5390 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 8.223267154301637e-06, |
|
"loss": 0.0709, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"eval_loss": 0.103515625, |
|
"eval_runtime": 310.8453, |
|
"eval_samples_per_second": 9.165, |
|
"eval_steps_per_second": 0.074, |
|
"eval_wer": 20.686616948299193, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 8.219784047370255e-06, |
|
"loss": 0.0686, |
|
"step": 5410 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 8.216300940438872e-06, |
|
"loss": 0.0679, |
|
"step": 5420 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 8.21281783350749e-06, |
|
"loss": 0.0677, |
|
"step": 5430 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 8.209334726576106e-06, |
|
"loss": 0.0658, |
|
"step": 5440 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 8.205851619644723e-06, |
|
"loss": 0.0667, |
|
"step": 5450 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 8.202368512713341e-06, |
|
"loss": 0.0644, |
|
"step": 5460 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 8.198885405781958e-06, |
|
"loss": 0.0699, |
|
"step": 5470 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 8.195402298850576e-06, |
|
"loss": 0.0694, |
|
"step": 5480 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 8.191919191919193e-06, |
|
"loss": 0.0675, |
|
"step": 5490 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 8.18843608498781e-06, |
|
"loss": 0.0688, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"eval_loss": 0.10687255859375, |
|
"eval_runtime": 300.4898, |
|
"eval_samples_per_second": 9.481, |
|
"eval_steps_per_second": 0.077, |
|
"eval_wer": 20.699203876773968, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 8.184952978056428e-06, |
|
"loss": 0.0667, |
|
"step": 5510 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 8.181469871125044e-06, |
|
"loss": 0.0683, |
|
"step": 5520 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 8.177986764193662e-06, |
|
"loss": 0.0655, |
|
"step": 5530 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 8.174503657262279e-06, |
|
"loss": 0.0682, |
|
"step": 5540 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 8.171020550330896e-06, |
|
"loss": 0.0662, |
|
"step": 5550 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 8.167537443399512e-06, |
|
"loss": 0.0689, |
|
"step": 5560 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 8.16405433646813e-06, |
|
"loss": 0.0665, |
|
"step": 5570 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 8.160571229536749e-06, |
|
"loss": 0.0675, |
|
"step": 5580 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 8.157088122605365e-06, |
|
"loss": 0.0681, |
|
"step": 5590 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 8.153605015673982e-06, |
|
"loss": 0.0711, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"eval_loss": 0.10504150390625, |
|
"eval_runtime": 679.7242, |
|
"eval_samples_per_second": 4.191, |
|
"eval_steps_per_second": 0.034, |
|
"eval_wer": 21.001290160168665, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 8.150121908742598e-06, |
|
"loss": 0.0661, |
|
"step": 5610 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 8.146638801811217e-06, |
|
"loss": 0.0658, |
|
"step": 5620 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 8.143155694879833e-06, |
|
"loss": 0.0657, |
|
"step": 5630 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 8.13967258794845e-06, |
|
"loss": 0.0681, |
|
"step": 5640 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 8.136189481017068e-06, |
|
"loss": 0.0671, |
|
"step": 5650 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 8.132706374085685e-06, |
|
"loss": 0.0719, |
|
"step": 5660 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 8.129223267154303e-06, |
|
"loss": 0.0695, |
|
"step": 5670 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 8.12574016022292e-06, |
|
"loss": 0.0654, |
|
"step": 5680 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 8.122257053291536e-06, |
|
"loss": 0.0673, |
|
"step": 5690 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 8.118773946360153e-06, |
|
"loss": 0.0676, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"eval_loss": 0.111572265625, |
|
"eval_runtime": 263.5896, |
|
"eval_samples_per_second": 10.808, |
|
"eval_steps_per_second": 0.087, |
|
"eval_wer": 20.837660089996536, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 8.115290839428771e-06, |
|
"loss": 0.0675, |
|
"step": 5710 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 8.11180773249739e-06, |
|
"loss": 0.0694, |
|
"step": 5720 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 8.108324625566006e-06, |
|
"loss": 0.0713, |
|
"step": 5730 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 8.104841518634622e-06, |
|
"loss": 0.0681, |
|
"step": 5740 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 8.101358411703239e-06, |
|
"loss": 0.067, |
|
"step": 5750 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 8.097875304771857e-06, |
|
"loss": 0.0719, |
|
"step": 5760 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 8.094392197840476e-06, |
|
"loss": 0.0601, |
|
"step": 5770 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 8.090909090909092e-06, |
|
"loss": 0.0592, |
|
"step": 5780 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 8.087425983977709e-06, |
|
"loss": 0.0606, |
|
"step": 5790 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 8.083942877046325e-06, |
|
"loss": 0.0591, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"eval_loss": 0.11163330078125, |
|
"eval_runtime": 266.5978, |
|
"eval_samples_per_second": 10.687, |
|
"eval_steps_per_second": 0.086, |
|
"eval_wer": 20.740111394317, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 8.080459770114944e-06, |
|
"loss": 0.0613, |
|
"step": 5810 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 8.07697666318356e-06, |
|
"loss": 0.0589, |
|
"step": 5820 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 8.073493556252178e-06, |
|
"loss": 0.0585, |
|
"step": 5830 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 8.070010449320795e-06, |
|
"loss": 0.0579, |
|
"step": 5840 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 8.066527342389412e-06, |
|
"loss": 0.0576, |
|
"step": 5850 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 8.06304423545803e-06, |
|
"loss": 0.0569, |
|
"step": 5860 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 8.059561128526646e-06, |
|
"loss": 0.0608, |
|
"step": 5870 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 8.056078021595263e-06, |
|
"loss": 0.0568, |
|
"step": 5880 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 8.05259491466388e-06, |
|
"loss": 0.0599, |
|
"step": 5890 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 8.049111807732498e-06, |
|
"loss": 0.062, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"eval_loss": 0.11236572265625, |
|
"eval_runtime": 270.2208, |
|
"eval_samples_per_second": 10.543, |
|
"eval_steps_per_second": 0.085, |
|
"eval_wer": 20.607948645331824, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 8.045628700801116e-06, |
|
"loss": 0.0565, |
|
"step": 5910 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 8.042145593869733e-06, |
|
"loss": 0.0577, |
|
"step": 5920 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 8.03866248693835e-06, |
|
"loss": 0.0571, |
|
"step": 5930 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 8.035179380006966e-06, |
|
"loss": 0.0601, |
|
"step": 5940 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 8.031696273075584e-06, |
|
"loss": 0.0578, |
|
"step": 5950 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 8.0282131661442e-06, |
|
"loss": 0.0577, |
|
"step": 5960 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 8.024730059212819e-06, |
|
"loss": 0.0581, |
|
"step": 5970 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 8.021246952281436e-06, |
|
"loss": 0.0589, |
|
"step": 5980 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 8.017763845350052e-06, |
|
"loss": 0.0601, |
|
"step": 5990 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 8.01428073841867e-06, |
|
"loss": 0.0601, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"eval_loss": 0.11273193359375, |
|
"eval_runtime": 269.3473, |
|
"eval_samples_per_second": 10.577, |
|
"eval_steps_per_second": 0.085, |
|
"eval_wer": 20.504106485414898, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 8.011494252873565e-06, |
|
"loss": 0.0576, |
|
"step": 6010 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 8.008011145942181e-06, |
|
"loss": 0.0588, |
|
"step": 6020 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 8.004528039010798e-06, |
|
"loss": 0.0611, |
|
"step": 6030 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 8.001044932079416e-06, |
|
"loss": 0.057, |
|
"step": 6040 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 7.997561825148033e-06, |
|
"loss": 0.0577, |
|
"step": 6050 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 7.99407871821665e-06, |
|
"loss": 0.0618, |
|
"step": 6060 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 7.990595611285266e-06, |
|
"loss": 0.0614, |
|
"step": 6070 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 7.987112504353884e-06, |
|
"loss": 0.0615, |
|
"step": 6080 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 7.9836293974225e-06, |
|
"loss": 0.0569, |
|
"step": 6090 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 7.980146290491119e-06, |
|
"loss": 0.059, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"eval_loss": 0.11077880859375, |
|
"eval_runtime": 272.3283, |
|
"eval_samples_per_second": 10.462, |
|
"eval_steps_per_second": 0.084, |
|
"eval_wer": 20.57333459202618, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 7.976663183559736e-06, |
|
"loss": 0.057, |
|
"step": 6110 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 7.973180076628352e-06, |
|
"loss": 0.0565, |
|
"step": 6120 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 7.96969696969697e-06, |
|
"loss": 0.0599, |
|
"step": 6130 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 7.966213862765587e-06, |
|
"loss": 0.0564, |
|
"step": 6140 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 7.962730755834205e-06, |
|
"loss": 0.0618, |
|
"step": 6150 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 7.959247648902822e-06, |
|
"loss": 0.0581, |
|
"step": 6160 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 7.955764541971439e-06, |
|
"loss": 0.0564, |
|
"step": 6170 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 7.952281435040057e-06, |
|
"loss": 0.0576, |
|
"step": 6180 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 7.948798328108673e-06, |
|
"loss": 0.0605, |
|
"step": 6190 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 7.945315221177292e-06, |
|
"loss": 0.0594, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"eval_loss": 0.11126708984375, |
|
"eval_runtime": 271.5968, |
|
"eval_samples_per_second": 10.49, |
|
"eval_steps_per_second": 0.085, |
|
"eval_wer": 20.595361716857045, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 7.941832114245908e-06, |
|
"loss": 0.0595, |
|
"step": 6210 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 7.938349007314525e-06, |
|
"loss": 0.0579, |
|
"step": 6220 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 7.934865900383141e-06, |
|
"loss": 0.0574, |
|
"step": 6230 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 7.93138279345176e-06, |
|
"loss": 0.0549, |
|
"step": 6240 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 7.927899686520378e-06, |
|
"loss": 0.0579, |
|
"step": 6250 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 7.924416579588995e-06, |
|
"loss": 0.0588, |
|
"step": 6260 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 7.920933472657611e-06, |
|
"loss": 0.057, |
|
"step": 6270 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 7.917450365726228e-06, |
|
"loss": 0.0593, |
|
"step": 6280 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 7.913967258794846e-06, |
|
"loss": 0.0602, |
|
"step": 6290 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 7.910484151863463e-06, |
|
"loss": 0.0587, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"eval_loss": 0.11077880859375, |
|
"eval_runtime": 295.0822, |
|
"eval_samples_per_second": 9.655, |
|
"eval_steps_per_second": 0.078, |
|
"eval_wer": 20.532427074483152, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 7.90700104493208e-06, |
|
"loss": 0.0593, |
|
"step": 6310 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 7.903517938000697e-06, |
|
"loss": 0.0583, |
|
"step": 6320 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 7.900034831069314e-06, |
|
"loss": 0.0585, |
|
"step": 6330 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 7.896551724137932e-06, |
|
"loss": 0.0583, |
|
"step": 6340 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 7.893068617206549e-06, |
|
"loss": 0.0567, |
|
"step": 6350 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 7.889585510275165e-06, |
|
"loss": 0.0584, |
|
"step": 6360 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 7.886102403343784e-06, |
|
"loss": 0.0588, |
|
"step": 6370 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 7.8826192964124e-06, |
|
"loss": 0.058, |
|
"step": 6380 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 7.879136189481019e-06, |
|
"loss": 0.0577, |
|
"step": 6390 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 7.875653082549635e-06, |
|
"loss": 0.0581, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"eval_loss": 0.112060546875, |
|
"eval_runtime": 273.5545, |
|
"eval_samples_per_second": 10.415, |
|
"eval_steps_per_second": 0.084, |
|
"eval_wer": 20.58592152050096, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 7.872169975618252e-06, |
|
"loss": 0.0556, |
|
"step": 6410 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 7.868686868686868e-06, |
|
"loss": 0.0597, |
|
"step": 6420 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 7.865203761755487e-06, |
|
"loss": 0.0618, |
|
"step": 6430 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 7.861720654824105e-06, |
|
"loss": 0.0567, |
|
"step": 6440 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 7.858237547892721e-06, |
|
"loss": 0.0611, |
|
"step": 6450 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 7.854754440961338e-06, |
|
"loss": 0.062, |
|
"step": 6460 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 7.851271334029955e-06, |
|
"loss": 0.0563, |
|
"step": 6470 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 7.847788227098573e-06, |
|
"loss": 0.0594, |
|
"step": 6480 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 7.84430512016719e-06, |
|
"loss": 0.06, |
|
"step": 6490 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 7.840822013235808e-06, |
|
"loss": 0.0596, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"eval_loss": 0.1087646484375, |
|
"eval_runtime": 268.8291, |
|
"eval_samples_per_second": 10.598, |
|
"eval_steps_per_second": 0.086, |
|
"eval_wer": 20.63941596651877, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 7.837338906304424e-06, |
|
"loss": 0.061, |
|
"step": 6510 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 7.833855799373041e-06, |
|
"loss": 0.0571, |
|
"step": 6520 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 7.830372692441659e-06, |
|
"loss": 0.0588, |
|
"step": 6530 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 7.826889585510276e-06, |
|
"loss": 0.0571, |
|
"step": 6540 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 7.823406478578894e-06, |
|
"loss": 0.0591, |
|
"step": 6550 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 7.81992337164751e-06, |
|
"loss": 0.06, |
|
"step": 6560 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 7.816440264716127e-06, |
|
"loss": 0.0566, |
|
"step": 6570 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 7.812957157784745e-06, |
|
"loss": 0.0595, |
|
"step": 6580 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 7.809474050853362e-06, |
|
"loss": 0.0616, |
|
"step": 6590 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 7.805990943921979e-06, |
|
"loss": 0.0617, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"eval_loss": 0.1103515625, |
|
"eval_runtime": 274.7709, |
|
"eval_samples_per_second": 10.369, |
|
"eval_steps_per_second": 0.084, |
|
"eval_wer": 20.81877969728437, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 7.802507836990595e-06, |
|
"loss": 0.058, |
|
"step": 6610 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 7.799024730059213e-06, |
|
"loss": 0.0592, |
|
"step": 6620 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 7.79554162312783e-06, |
|
"loss": 0.0588, |
|
"step": 6630 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 7.792058516196448e-06, |
|
"loss": 0.0548, |
|
"step": 6640 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 7.788575409265065e-06, |
|
"loss": 0.0567, |
|
"step": 6650 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 7.785092302333681e-06, |
|
"loss": 0.0577, |
|
"step": 6660 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 7.7816091954023e-06, |
|
"loss": 0.058, |
|
"step": 6670 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 7.778126088470916e-06, |
|
"loss": 0.0584, |
|
"step": 6680 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 7.774642981539535e-06, |
|
"loss": 0.0559, |
|
"step": 6690 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 7.771159874608151e-06, |
|
"loss": 0.0557, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"eval_loss": 0.11236572265625, |
|
"eval_runtime": 271.2745, |
|
"eval_samples_per_second": 10.502, |
|
"eval_steps_per_second": 0.085, |
|
"eval_wer": 20.557600931432706, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 7.767676767676768e-06, |
|
"loss": 0.0587, |
|
"step": 6710 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 7.764193660745386e-06, |
|
"loss": 0.0599, |
|
"step": 6720 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 7.760710553814003e-06, |
|
"loss": 0.0596, |
|
"step": 6730 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 7.757227446882621e-06, |
|
"loss": 0.0556, |
|
"step": 6740 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 7.753744339951237e-06, |
|
"loss": 0.06, |
|
"step": 6750 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 7.750261233019854e-06, |
|
"loss": 0.0574, |
|
"step": 6760 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 7.74677812608847e-06, |
|
"loss": 0.0567, |
|
"step": 6770 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 7.743295019157089e-06, |
|
"loss": 0.0599, |
|
"step": 6780 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 7.739811912225707e-06, |
|
"loss": 0.0562, |
|
"step": 6790 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 7.736328805294324e-06, |
|
"loss": 0.0594, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"eval_loss": 0.10858154296875, |
|
"eval_runtime": 272.0479, |
|
"eval_samples_per_second": 10.472, |
|
"eval_steps_per_second": 0.085, |
|
"eval_wer": 20.607948645331824, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 7.73284569836294e-06, |
|
"loss": 0.0571, |
|
"step": 6810 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 7.729362591431557e-06, |
|
"loss": 0.059, |
|
"step": 6820 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 7.725879484500175e-06, |
|
"loss": 0.0576, |
|
"step": 6830 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 7.722396377568792e-06, |
|
"loss": 0.0599, |
|
"step": 6840 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 7.71891327063741e-06, |
|
"loss": 0.056, |
|
"step": 6850 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 7.715430163706027e-06, |
|
"loss": 0.0595, |
|
"step": 6860 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 7.711947056774643e-06, |
|
"loss": 0.0567, |
|
"step": 6870 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 7.708463949843261e-06, |
|
"loss": 0.0581, |
|
"step": 6880 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 7.704980842911878e-06, |
|
"loss": 0.059, |
|
"step": 6890 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 7.701497735980495e-06, |
|
"loss": 0.0611, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"eval_loss": 0.11187744140625, |
|
"eval_runtime": 270.2, |
|
"eval_samples_per_second": 10.544, |
|
"eval_steps_per_second": 0.085, |
|
"eval_wer": 20.488372824821425, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 7.698014629049111e-06, |
|
"loss": 0.0633, |
|
"step": 6910 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 7.69453152211773e-06, |
|
"loss": 0.0577, |
|
"step": 6920 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 7.691048415186348e-06, |
|
"loss": 0.0599, |
|
"step": 6930 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 7.687565308254964e-06, |
|
"loss": 0.059, |
|
"step": 6940 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 7.684082201323581e-06, |
|
"loss": 0.0576, |
|
"step": 6950 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 7.680599094392198e-06, |
|
"loss": 0.0571, |
|
"step": 6960 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 7.677115987460816e-06, |
|
"loss": 0.0606, |
|
"step": 6970 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 7.673632880529432e-06, |
|
"loss": 0.061, |
|
"step": 6980 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 7.67014977359805e-06, |
|
"loss": 0.0603, |
|
"step": 6990 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 7.666666666666667e-06, |
|
"loss": 0.0574, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"eval_loss": 0.112060546875, |
|
"eval_runtime": 270.472, |
|
"eval_samples_per_second": 10.533, |
|
"eval_steps_per_second": 0.085, |
|
"eval_wer": 20.516693413889676, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 7.663183559735284e-06, |
|
"loss": 0.0595, |
|
"step": 7010 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 7.660397074190178e-06, |
|
"loss": 0.0587, |
|
"step": 7020 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 7.656913967258796e-06, |
|
"loss": 0.0575, |
|
"step": 7030 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 7.653430860327413e-06, |
|
"loss": 0.0591, |
|
"step": 7040 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 7.64994775339603e-06, |
|
"loss": 0.0569, |
|
"step": 7050 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 7.646464646464648e-06, |
|
"loss": 0.0564, |
|
"step": 7060 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 7.642981539533264e-06, |
|
"loss": 0.0626, |
|
"step": 7070 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 7.639498432601881e-06, |
|
"loss": 0.0585, |
|
"step": 7080 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 7.636015325670498e-06, |
|
"loss": 0.061, |
|
"step": 7090 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 7.632532218739116e-06, |
|
"loss": 0.0593, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"eval_loss": 0.1129150390625, |
|
"eval_runtime": 324.392, |
|
"eval_samples_per_second": 8.783, |
|
"eval_steps_per_second": 0.071, |
|
"eval_wer": 20.45375877151578, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 7.629049111807733e-06, |
|
"loss": 0.0569, |
|
"step": 7110 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 7.62556600487635e-06, |
|
"loss": 0.0579, |
|
"step": 7120 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 7.622082897944968e-06, |
|
"loss": 0.0567, |
|
"step": 7130 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 7.618599791013585e-06, |
|
"loss": 0.0627, |
|
"step": 7140 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 7.615116684082202e-06, |
|
"loss": 0.0583, |
|
"step": 7150 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 7.611633577150819e-06, |
|
"loss": 0.0567, |
|
"step": 7160 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 7.608150470219436e-06, |
|
"loss": 0.0617, |
|
"step": 7170 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 7.604667363288054e-06, |
|
"loss": 0.0585, |
|
"step": 7180 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 7.601184256356671e-06, |
|
"loss": 0.0588, |
|
"step": 7190 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 7.597701149425288e-06, |
|
"loss": 0.0613, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"eval_loss": 0.1126708984375, |
|
"eval_runtime": 273.1528, |
|
"eval_samples_per_second": 10.43, |
|
"eval_steps_per_second": 0.084, |
|
"eval_wer": 20.582774788382267, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 7.594218042493905e-06, |
|
"loss": 0.0577, |
|
"step": 7210 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 7.590734935562522e-06, |
|
"loss": 0.056, |
|
"step": 7220 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 7.587251828631139e-06, |
|
"loss": 0.0582, |
|
"step": 7230 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 7.583768721699757e-06, |
|
"loss": 0.0578, |
|
"step": 7240 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 7.580285614768375e-06, |
|
"loss": 0.06, |
|
"step": 7250 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 7.576802507836991e-06, |
|
"loss": 0.059, |
|
"step": 7260 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 7.573319400905609e-06, |
|
"loss": 0.0586, |
|
"step": 7270 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 7.569836293974225e-06, |
|
"loss": 0.0555, |
|
"step": 7280 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 7.566353187042843e-06, |
|
"loss": 0.0572, |
|
"step": 7290 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 7.562870080111459e-06, |
|
"loss": 0.0594, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"eval_loss": 0.11083984375, |
|
"eval_runtime": 271.7249, |
|
"eval_samples_per_second": 10.485, |
|
"eval_steps_per_second": 0.085, |
|
"eval_wer": 20.488372824821425, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 7.559386973180078e-06, |
|
"loss": 0.059, |
|
"step": 7310 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 7.555903866248695e-06, |
|
"loss": 0.0571, |
|
"step": 7320 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 7.552420759317312e-06, |
|
"loss": 0.0573, |
|
"step": 7330 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 7.548937652385929e-06, |
|
"loss": 0.0599, |
|
"step": 7340 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 7.545454545454546e-06, |
|
"loss": 0.0536, |
|
"step": 7350 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 7.541971438523164e-06, |
|
"loss": 0.0589, |
|
"step": 7360 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 7.5384883315917805e-06, |
|
"loss": 0.0579, |
|
"step": 7370 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 7.535005224660398e-06, |
|
"loss": 0.0585, |
|
"step": 7380 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 7.531522117729015e-06, |
|
"loss": 0.0595, |
|
"step": 7390 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 7.528039010797632e-06, |
|
"loss": 0.0603, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"eval_loss": 0.11322021484375, |
|
"eval_runtime": 268.4354, |
|
"eval_samples_per_second": 10.613, |
|
"eval_steps_per_second": 0.086, |
|
"eval_wer": 20.66458982346833, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 7.524555903866249e-06, |
|
"loss": 0.0592, |
|
"step": 7410 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 7.521072796934866e-06, |
|
"loss": 0.0595, |
|
"step": 7420 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 7.517589690003484e-06, |
|
"loss": 0.0585, |
|
"step": 7430 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 7.514106583072101e-06, |
|
"loss": 0.0576, |
|
"step": 7440 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 7.510623476140718e-06, |
|
"loss": 0.0582, |
|
"step": 7450 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 7.507140369209336e-06, |
|
"loss": 0.0614, |
|
"step": 7460 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 7.503657262277952e-06, |
|
"loss": 0.0565, |
|
"step": 7470 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 7.5001741553465705e-06, |
|
"loss": 0.0626, |
|
"step": 7480 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 7.496691048415187e-06, |
|
"loss": 0.0587, |
|
"step": 7490 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 7.4932079414838045e-06, |
|
"loss": 0.0602, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"eval_loss": 0.10906982421875, |
|
"eval_runtime": 271.3251, |
|
"eval_samples_per_second": 10.5, |
|
"eval_steps_per_second": 0.085, |
|
"eval_wer": 20.27124830863149, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 7.489724834552421e-06, |
|
"loss": 0.0595, |
|
"step": 7510 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 7.4862417276210385e-06, |
|
"loss": 0.0566, |
|
"step": 7520 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 7.482758620689656e-06, |
|
"loss": 0.0584, |
|
"step": 7530 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 7.4792755137582725e-06, |
|
"loss": 0.0549, |
|
"step": 7540 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 7.475792406826891e-06, |
|
"loss": 0.0598, |
|
"step": 7550 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 7.472309299895507e-06, |
|
"loss": 0.0571, |
|
"step": 7560 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 7.468826192964125e-06, |
|
"loss": 0.0588, |
|
"step": 7570 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 7.465343086032741e-06, |
|
"loss": 0.0558, |
|
"step": 7580 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 7.461859979101359e-06, |
|
"loss": 0.059, |
|
"step": 7590 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 7.458376872169977e-06, |
|
"loss": 0.0607, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"eval_loss": 0.10968017578125, |
|
"eval_runtime": 271.2796, |
|
"eval_samples_per_second": 10.502, |
|
"eval_steps_per_second": 0.085, |
|
"eval_wer": 20.35306334371755, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 7.454893765238594e-06, |
|
"loss": 0.0585, |
|
"step": 7610 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 7.451410658307211e-06, |
|
"loss": 0.0593, |
|
"step": 7620 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 7.447927551375828e-06, |
|
"loss": 0.0596, |
|
"step": 7630 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 7.444444444444445e-06, |
|
"loss": 0.0575, |
|
"step": 7640 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 7.440961337513062e-06, |
|
"loss": 0.0605, |
|
"step": 7650 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 7.437478230581679e-06, |
|
"loss": 0.0611, |
|
"step": 7660 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 7.433995123650297e-06, |
|
"loss": 0.0554, |
|
"step": 7670 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 7.430512016718914e-06, |
|
"loss": 0.0548, |
|
"step": 7680 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 7.427028909787531e-06, |
|
"loss": 0.0589, |
|
"step": 7690 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 7.423545802856148e-06, |
|
"loss": 0.0567, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"eval_loss": 0.10833740234375, |
|
"eval_runtime": 276.9395, |
|
"eval_samples_per_second": 10.287, |
|
"eval_steps_per_second": 0.083, |
|
"eval_wer": 20.321596022530603, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 7.420062695924765e-06, |
|
"loss": 0.0582, |
|
"step": 7710 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 7.416579588993382e-06, |
|
"loss": 0.0593, |
|
"step": 7720 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 7.413096482062e-06, |
|
"loss": 0.0569, |
|
"step": 7730 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 7.409613375130618e-06, |
|
"loss": 0.0613, |
|
"step": 7740 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 7.406130268199234e-06, |
|
"loss": 0.0568, |
|
"step": 7750 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 7.402647161267852e-06, |
|
"loss": 0.0568, |
|
"step": 7760 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 7.399164054336468e-06, |
|
"loss": 0.0591, |
|
"step": 7770 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 7.3956809474050865e-06, |
|
"loss": 0.0606, |
|
"step": 7780 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 7.392197840473703e-06, |
|
"loss": 0.0609, |
|
"step": 7790 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 7.3887147335423205e-06, |
|
"loss": 0.0572, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"eval_loss": 0.10943603515625, |
|
"eval_runtime": 273.1967, |
|
"eval_samples_per_second": 10.428, |
|
"eval_steps_per_second": 0.084, |
|
"eval_wer": 20.548160735076625, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 7.385231626610938e-06, |
|
"loss": 0.0596, |
|
"step": 7810 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 7.3817485196795545e-06, |
|
"loss": 0.0582, |
|
"step": 7820 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 7.378265412748172e-06, |
|
"loss": 0.0581, |
|
"step": 7830 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 7.3747823058167885e-06, |
|
"loss": 0.059, |
|
"step": 7840 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 7.371299198885407e-06, |
|
"loss": 0.0579, |
|
"step": 7850 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 7.367816091954023e-06, |
|
"loss": 0.0595, |
|
"step": 7860 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 7.364332985022641e-06, |
|
"loss": 0.0604, |
|
"step": 7870 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 7.360849878091258e-06, |
|
"loss": 0.0601, |
|
"step": 7880 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 7.357366771159875e-06, |
|
"loss": 0.0635, |
|
"step": 7890 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 7.353883664228493e-06, |
|
"loss": 0.0563, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"eval_loss": 0.10894775390625, |
|
"eval_runtime": 275.1597, |
|
"eval_samples_per_second": 10.354, |
|
"eval_steps_per_second": 0.084, |
|
"eval_wer": 20.541867270839234, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 7.35040055729711e-06, |
|
"loss": 0.0593, |
|
"step": 7910 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 7.346917450365727e-06, |
|
"loss": 0.0577, |
|
"step": 7920 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 7.343434343434344e-06, |
|
"loss": 0.0576, |
|
"step": 7930 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 7.339951236502961e-06, |
|
"loss": 0.0615, |
|
"step": 7940 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 7.3364681295715786e-06, |
|
"loss": 0.0583, |
|
"step": 7950 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 7.332985022640195e-06, |
|
"loss": 0.059, |
|
"step": 7960 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 7.329501915708813e-06, |
|
"loss": 0.0591, |
|
"step": 7970 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 7.32601880877743e-06, |
|
"loss": 0.0573, |
|
"step": 7980 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 7.322535701846047e-06, |
|
"loss": 0.0613, |
|
"step": 7990 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 7.319052594914664e-06, |
|
"loss": 0.0581, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"eval_loss": 0.10784912109375, |
|
"eval_runtime": 272.0815, |
|
"eval_samples_per_second": 10.471, |
|
"eval_steps_per_second": 0.085, |
|
"eval_wer": 20.16111268447717, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 7.3155694879832814e-06, |
|
"loss": 0.0567, |
|
"step": 8010 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 7.312783002438175e-06, |
|
"loss": 0.061, |
|
"step": 8020 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 7.309299895506793e-06, |
|
"loss": 0.0588, |
|
"step": 8030 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 7.30581678857541e-06, |
|
"loss": 0.0577, |
|
"step": 8040 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 7.302333681644027e-06, |
|
"loss": 0.0586, |
|
"step": 8050 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 7.298850574712645e-06, |
|
"loss": 0.0586, |
|
"step": 8060 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 7.295367467781261e-06, |
|
"loss": 0.0598, |
|
"step": 8070 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 7.2918843608498794e-06, |
|
"loss": 0.0558, |
|
"step": 8080 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 7.288401253918496e-06, |
|
"loss": 0.057, |
|
"step": 8090 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 7.2849181469871134e-06, |
|
"loss": 0.0582, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"eval_loss": 0.11016845703125, |
|
"eval_runtime": 270.4011, |
|
"eval_samples_per_second": 10.536, |
|
"eval_steps_per_second": 0.085, |
|
"eval_wer": 20.321596022530603, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 7.28143504005573e-06, |
|
"loss": 0.057, |
|
"step": 8110 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 7.2779519331243474e-06, |
|
"loss": 0.0579, |
|
"step": 8120 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 7.274468826192965e-06, |
|
"loss": 0.0567, |
|
"step": 8130 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 7.2709857192615815e-06, |
|
"loss": 0.057, |
|
"step": 8140 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 7.2675026123302e-06, |
|
"loss": 0.0628, |
|
"step": 8150 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 7.264019505398816e-06, |
|
"loss": 0.0574, |
|
"step": 8160 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 7.260536398467434e-06, |
|
"loss": 0.0575, |
|
"step": 8170 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 7.25705329153605e-06, |
|
"loss": 0.0574, |
|
"step": 8180 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 7.253570184604668e-06, |
|
"loss": 0.0585, |
|
"step": 8190 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 7.250087077673286e-06, |
|
"loss": 0.0571, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"eval_loss": 0.11126708984375, |
|
"eval_runtime": 271.8988, |
|
"eval_samples_per_second": 10.478, |
|
"eval_steps_per_second": 0.085, |
|
"eval_wer": 20.044683596085466, |
|
"step": 8200 |
|
} |
|
], |
|
"max_steps": 29000, |
|
"num_train_epochs": 11, |
|
"total_flos": 2.142217063719011e+21, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|