|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 12.0, |
|
"eval_steps": 2500, |
|
"global_step": 29076, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 3.232462173314994e-07, |
|
"loss": 13.7898, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 6.602475928473178e-07, |
|
"loss": 13.7695, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.972489683631363e-07, |
|
"loss": 13.6652, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.3411279229711144e-06, |
|
"loss": 13.4432, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.6781292984869326e-06, |
|
"loss": 13.2341, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 2.0151306740027512e-06, |
|
"loss": 12.9315, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 2.3452544704264104e-06, |
|
"loss": 12.5548, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 2.682255845942229e-06, |
|
"loss": 11.6273, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 3.019257221458047e-06, |
|
"loss": 10.5914, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 3.3631361760660253e-06, |
|
"loss": 9.7521, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 3.707015130674003e-06, |
|
"loss": 8.8445, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 4.050894085281981e-06, |
|
"loss": 8.3614, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 4.394773039889959e-06, |
|
"loss": 7.5809, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 4.738651994497937e-06, |
|
"loss": 7.0097, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 5.082530949105915e-06, |
|
"loss": 6.7184, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 5.4195323246217335e-06, |
|
"loss": 6.3055, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 5.749656121045393e-06, |
|
"loss": 5.932, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 6.0866574965612115e-06, |
|
"loss": 5.6367, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 6.43053645116919e-06, |
|
"loss": 5.3637, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 6.767537826685008e-06, |
|
"loss": 5.2347, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 7.111416781292986e-06, |
|
"loss": 4.9859, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 7.4552957359009634e-06, |
|
"loss": 4.7688, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 7.799174690508942e-06, |
|
"loss": 4.5529, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 8.1292984869326e-06, |
|
"loss": 4.4404, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 8.46629986244842e-06, |
|
"loss": 4.2761, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 8.810178817056398e-06, |
|
"loss": 4.1596, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 9.154057771664376e-06, |
|
"loss": 4.0284, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 9.497936726272354e-06, |
|
"loss": 3.9322, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 9.814305364511693e-06, |
|
"loss": 3.8223, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 1.0144429160935352e-05, |
|
"loss": 3.7458, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 1.046767537826685e-05, |
|
"loss": 3.7244, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 1.0804676753782671e-05, |
|
"loss": 3.6959, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 1.1134800550206328e-05, |
|
"loss": 3.6831, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 1.1464924346629987e-05, |
|
"loss": 3.7152, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 1.1808803301237965e-05, |
|
"loss": 3.6774, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 1.2145804676753783e-05, |
|
"loss": 3.6656, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 1.2489683631361762e-05, |
|
"loss": 3.6759, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 1.282668500687758e-05, |
|
"loss": 3.6629, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 1.31636863823934e-05, |
|
"loss": 3.6847, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.3507565337001376e-05, |
|
"loss": 3.6718, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.3844566712517195e-05, |
|
"loss": 3.6818, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 1.4181568088033012e-05, |
|
"loss": 3.6721, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 1.4525447042640992e-05, |
|
"loss": 3.6782, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.4869325997248969e-05, |
|
"loss": 3.6706, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.5206327372764788e-05, |
|
"loss": 3.6728, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.5550206327372764e-05, |
|
"loss": 3.6824, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 1.5887207702888585e-05, |
|
"loss": 3.6702, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 1.6224209078404403e-05, |
|
"loss": 3.6709, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 1.6561210453920224e-05, |
|
"loss": 3.6721, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 1.689821182943604e-05, |
|
"loss": 3.6685, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"eval_loss": 4.088207721710205, |
|
"eval_runtime": 6.8428, |
|
"eval_samples_per_second": 207.517, |
|
"eval_steps_per_second": 13.006, |
|
"eval_wer": 1.0, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 1.723521320495186e-05, |
|
"loss": 3.6783, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 1.755845942228336e-05, |
|
"loss": 3.6567, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 1.7888583218707018e-05, |
|
"loss": 3.6711, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 1.8232462173314994e-05, |
|
"loss": 3.6702, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 1.856946354883081e-05, |
|
"loss": 3.6692, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 1.889958734525447e-05, |
|
"loss": 3.6652, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 1.924346629986245e-05, |
|
"loss": 3.6677, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 1.9580467675378268e-05, |
|
"loss": 3.6741, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 1.991746905089409e-05, |
|
"loss": 3.6756, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 1.9970956893916236e-05, |
|
"loss": 3.6782, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 1.9934270865178847e-05, |
|
"loss": 3.6903, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 1.9896820544176094e-05, |
|
"loss": 3.6683, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 1.9858605930907983e-05, |
|
"loss": 3.6756, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 1.9820391317639865e-05, |
|
"loss": 3.6806, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 1.9783705288902476e-05, |
|
"loss": 3.6758, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 1.9746254967899726e-05, |
|
"loss": 3.6761, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 1.9708040354631612e-05, |
|
"loss": 3.6788, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 1.96698257413635e-05, |
|
"loss": 3.6683, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 1.9632375420360748e-05, |
|
"loss": 3.674, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 1.9594925099357994e-05, |
|
"loss": 3.676, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 1.9556710486089883e-05, |
|
"loss": 3.6638, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 1.951926016508713e-05, |
|
"loss": 3.6637, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 1.948104555181902e-05, |
|
"loss": 3.6676, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 1.94428309385509e-05, |
|
"loss": 3.6742, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 1.940538061754815e-05, |
|
"loss": 3.6812, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 1.9367166004280037e-05, |
|
"loss": 3.6827, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 1.9330479975542648e-05, |
|
"loss": 3.6729, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 1.9292265362274537e-05, |
|
"loss": 3.6829, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 1.9254815041271784e-05, |
|
"loss": 3.689, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 1.921660042800367e-05, |
|
"loss": 3.6704, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 1.917915010700092e-05, |
|
"loss": 3.6789, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 1.9140935493732805e-05, |
|
"loss": 3.6798, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 1.910272088046469e-05, |
|
"loss": 3.6657, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 1.9065270559461938e-05, |
|
"loss": 3.6711, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 1.9027820238459188e-05, |
|
"loss": 3.6687, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 1.8990369917456438e-05, |
|
"loss": 3.6858, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 1.8952155304188323e-05, |
|
"loss": 3.6726, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 1.8915469275450934e-05, |
|
"loss": 3.6687, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 1.887725466218282e-05, |
|
"loss": 3.689, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 1.8839804341180066e-05, |
|
"loss": 3.6766, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 1.8802354020177316e-05, |
|
"loss": 3.6642, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 1.8764903699174566e-05, |
|
"loss": 3.6611, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 1.8726689085906452e-05, |
|
"loss": 3.6737, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 1.8689238764903702e-05, |
|
"loss": 3.6685, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 1.865178844390095e-05, |
|
"loss": 3.6684, |
|
"step": 4750 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 1.861510241516356e-05, |
|
"loss": 3.6653, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 1.8576887801895445e-05, |
|
"loss": 3.6628, |
|
"step": 4850 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 1.853867318862733e-05, |
|
"loss": 3.6679, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 1.850045857535922e-05, |
|
"loss": 3.6783, |
|
"step": 4950 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 1.8462243962091103e-05, |
|
"loss": 3.6718, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"eval_loss": 4.088207721710205, |
|
"eval_runtime": 6.279, |
|
"eval_samples_per_second": 226.15, |
|
"eval_steps_per_second": 14.174, |
|
"eval_wer": 1.0, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 1.8424029348822992e-05, |
|
"loss": 3.6698, |
|
"step": 5050 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 1.8385814735554878e-05, |
|
"loss": 3.6662, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 1.8347600122286764e-05, |
|
"loss": 3.6735, |
|
"step": 5150 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 1.8310914093549374e-05, |
|
"loss": 3.6721, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 1.8272699480281263e-05, |
|
"loss": 3.681, |
|
"step": 5250 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 1.823677774380923e-05, |
|
"loss": 3.6774, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 1.819856313054112e-05, |
|
"loss": 3.6913, |
|
"step": 5350 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 1.8161112809538367e-05, |
|
"loss": 3.6684, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 1.8123662488535617e-05, |
|
"loss": 3.6759, |
|
"step": 5450 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 1.8085447875267503e-05, |
|
"loss": 3.6831, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 1.8048761846530114e-05, |
|
"loss": 3.6778, |
|
"step": 5550 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 1.8011311525527364e-05, |
|
"loss": 3.6893, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 1.797309691225925e-05, |
|
"loss": 3.6675, |
|
"step": 5650 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 1.7934882298991135e-05, |
|
"loss": 3.6688, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 1.7897431977988385e-05, |
|
"loss": 3.6726, |
|
"step": 5750 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 1.7859981656985632e-05, |
|
"loss": 3.6644, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 1.782253133598288e-05, |
|
"loss": 3.6727, |
|
"step": 5850 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 1.778508101498013e-05, |
|
"loss": 3.6683, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 1.7747630693977378e-05, |
|
"loss": 3.6652, |
|
"step": 5950 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 1.7710180372974628e-05, |
|
"loss": 3.6618, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 1.7672730051971875e-05, |
|
"loss": 3.6846, |
|
"step": 6050 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 1.7635279730969125e-05, |
|
"loss": 3.6714, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 1.7599357994497096e-05, |
|
"loss": 3.7024, |
|
"step": 6150 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 1.756343625802507e-05, |
|
"loss": 3.6734, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 1.7525985937022317e-05, |
|
"loss": 3.6812, |
|
"step": 6250 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 1.7487771323754207e-05, |
|
"loss": 3.6653, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 1.7449556710486092e-05, |
|
"loss": 3.6763, |
|
"step": 6350 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 1.7412106389483342e-05, |
|
"loss": 3.688, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 1.737465606848059e-05, |
|
"loss": 3.6817, |
|
"step": 6450 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 1.73379700397432e-05, |
|
"loss": 3.665, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 1.7300519718740446e-05, |
|
"loss": 3.6705, |
|
"step": 6550 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 1.7263069397737696e-05, |
|
"loss": 3.6856, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 1.7224854784469582e-05, |
|
"loss": 3.6755, |
|
"step": 6650 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 1.718664017120147e-05, |
|
"loss": 3.6635, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 1.7148425557933354e-05, |
|
"loss": 3.6868, |
|
"step": 6750 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 1.7110210944665243e-05, |
|
"loss": 3.6703, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 1.707276062366249e-05, |
|
"loss": 3.6816, |
|
"step": 6850 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 1.703531030265974e-05, |
|
"loss": 3.6733, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 1.6997095689391625e-05, |
|
"loss": 3.67, |
|
"step": 6950 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 1.695964536838887e-05, |
|
"loss": 3.6895, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 1.692219504738612e-05, |
|
"loss": 3.6769, |
|
"step": 7050 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 1.688474472638337e-05, |
|
"loss": 3.6758, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 1.6847294405380618e-05, |
|
"loss": 3.6722, |
|
"step": 7150 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 1.6809079792112507e-05, |
|
"loss": 3.6679, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 1.6772393763375118e-05, |
|
"loss": 3.6766, |
|
"step": 7250 |
|
}, |
|
{ |
|
"epoch": 3.01, |
|
"learning_rate": 1.6734943442372364e-05, |
|
"loss": 3.6953, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"learning_rate": 1.669672882910425e-05, |
|
"loss": 3.6688, |
|
"step": 7350 |
|
}, |
|
{ |
|
"epoch": 3.05, |
|
"learning_rate": 1.66592785081015e-05, |
|
"loss": 3.6755, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 3.07, |
|
"learning_rate": 1.6621063894833386e-05, |
|
"loss": 3.6672, |
|
"step": 7450 |
|
}, |
|
{ |
|
"epoch": 3.1, |
|
"learning_rate": 1.6584377866095997e-05, |
|
"loss": 3.6684, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 3.1, |
|
"eval_loss": 4.088207721710205, |
|
"eval_runtime": 8.4451, |
|
"eval_samples_per_second": 168.145, |
|
"eval_steps_per_second": 10.539, |
|
"eval_wer": 1.0, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 3.12, |
|
"learning_rate": 1.6546163252827883e-05, |
|
"loss": 3.6898, |
|
"step": 7550 |
|
}, |
|
{ |
|
"epoch": 3.14, |
|
"learning_rate": 1.650794863955977e-05, |
|
"loss": 3.676, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 3.16, |
|
"learning_rate": 1.6469734026291654e-05, |
|
"loss": 3.68, |
|
"step": 7650 |
|
}, |
|
{ |
|
"epoch": 3.18, |
|
"learning_rate": 1.6431519413023543e-05, |
|
"loss": 3.6825, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 3.2, |
|
"learning_rate": 1.639406909202079e-05, |
|
"loss": 3.6714, |
|
"step": 7750 |
|
}, |
|
{ |
|
"epoch": 3.22, |
|
"learning_rate": 1.6355854478752676e-05, |
|
"loss": 3.67, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 3.24, |
|
"learning_rate": 1.6318404157749926e-05, |
|
"loss": 3.6576, |
|
"step": 7850 |
|
}, |
|
{ |
|
"epoch": 3.26, |
|
"learning_rate": 1.6281718129012536e-05, |
|
"loss": 3.6685, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 3.28, |
|
"learning_rate": 1.6243503515744422e-05, |
|
"loss": 3.6725, |
|
"step": 7950 |
|
}, |
|
{ |
|
"epoch": 3.3, |
|
"learning_rate": 1.6206053194741672e-05, |
|
"loss": 3.6761, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 3.32, |
|
"learning_rate": 1.6167838581473555e-05, |
|
"loss": 3.662, |
|
"step": 8050 |
|
}, |
|
{ |
|
"epoch": 3.34, |
|
"learning_rate": 1.6130388260470805e-05, |
|
"loss": 3.6737, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 3.36, |
|
"learning_rate": 1.6092937939468055e-05, |
|
"loss": 3.6758, |
|
"step": 8150 |
|
}, |
|
{ |
|
"epoch": 3.38, |
|
"learning_rate": 1.6056251910730665e-05, |
|
"loss": 3.6768, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 3.4, |
|
"learning_rate": 1.601803729746255e-05, |
|
"loss": 3.6886, |
|
"step": 8250 |
|
}, |
|
{ |
|
"epoch": 3.43, |
|
"learning_rate": 1.5979822684194437e-05, |
|
"loss": 3.6682, |
|
"step": 8300 |
|
}, |
|
{ |
|
"epoch": 3.45, |
|
"learning_rate": 1.5941608070926323e-05, |
|
"loss": 3.6857, |
|
"step": 8350 |
|
}, |
|
{ |
|
"epoch": 3.47, |
|
"learning_rate": 1.590339345765821e-05, |
|
"loss": 3.6767, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 3.49, |
|
"learning_rate": 1.586670742892082e-05, |
|
"loss": 3.6783, |
|
"step": 8450 |
|
}, |
|
{ |
|
"epoch": 3.51, |
|
"learning_rate": 1.583002140018343e-05, |
|
"loss": 3.673, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 3.53, |
|
"learning_rate": 1.579180678691532e-05, |
|
"loss": 3.6855, |
|
"step": 8550 |
|
}, |
|
{ |
|
"epoch": 3.55, |
|
"learning_rate": 1.575512075817793e-05, |
|
"loss": 3.6773, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 3.57, |
|
"learning_rate": 1.57191990217059e-05, |
|
"loss": 3.6653, |
|
"step": 8650 |
|
}, |
|
{ |
|
"epoch": 3.59, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 3.573, |
|
"step": 8700 |
|
}, |
|
{ |
|
"epoch": 3.61, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 8750 |
|
}, |
|
{ |
|
"epoch": 3.63, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 8800 |
|
}, |
|
{ |
|
"epoch": 3.65, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 8850 |
|
}, |
|
{ |
|
"epoch": 3.67, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 8900 |
|
}, |
|
{ |
|
"epoch": 3.69, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 8950 |
|
}, |
|
{ |
|
"epoch": 3.71, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 3.74, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 9050 |
|
}, |
|
{ |
|
"epoch": 3.76, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 9100 |
|
}, |
|
{ |
|
"epoch": 3.78, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 9150 |
|
}, |
|
{ |
|
"epoch": 3.8, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 9200 |
|
}, |
|
{ |
|
"epoch": 3.82, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 9250 |
|
}, |
|
{ |
|
"epoch": 3.84, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 3.86, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 9350 |
|
}, |
|
{ |
|
"epoch": 3.88, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 9400 |
|
}, |
|
{ |
|
"epoch": 3.9, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 9450 |
|
}, |
|
{ |
|
"epoch": 3.92, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 3.94, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 9550 |
|
}, |
|
{ |
|
"epoch": 3.96, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 9600 |
|
}, |
|
{ |
|
"epoch": 3.98, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 9650 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 9700 |
|
}, |
|
{ |
|
"epoch": 4.02, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 9750 |
|
}, |
|
{ |
|
"epoch": 4.04, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 9800 |
|
}, |
|
{ |
|
"epoch": 4.07, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 9850 |
|
}, |
|
{ |
|
"epoch": 4.09, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 9900 |
|
}, |
|
{ |
|
"epoch": 4.11, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 9950 |
|
}, |
|
{ |
|
"epoch": 4.13, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 4.13, |
|
"eval_loss": NaN, |
|
"eval_runtime": 7.5132, |
|
"eval_samples_per_second": 189.002, |
|
"eval_steps_per_second": 11.846, |
|
"eval_wer": 1.0, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 4.15, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 10050 |
|
}, |
|
{ |
|
"epoch": 4.17, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 10100 |
|
}, |
|
{ |
|
"epoch": 4.19, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 10150 |
|
}, |
|
{ |
|
"epoch": 4.21, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 10200 |
|
}, |
|
{ |
|
"epoch": 4.23, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 10250 |
|
}, |
|
{ |
|
"epoch": 4.25, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 10300 |
|
}, |
|
{ |
|
"epoch": 4.27, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 10350 |
|
}, |
|
{ |
|
"epoch": 4.29, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 10400 |
|
}, |
|
{ |
|
"epoch": 4.31, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 10450 |
|
}, |
|
{ |
|
"epoch": 4.33, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 4.35, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 10550 |
|
}, |
|
{ |
|
"epoch": 4.37, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 10600 |
|
}, |
|
{ |
|
"epoch": 4.4, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 10650 |
|
}, |
|
{ |
|
"epoch": 4.42, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 10700 |
|
}, |
|
{ |
|
"epoch": 4.44, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 10750 |
|
}, |
|
{ |
|
"epoch": 4.46, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 10800 |
|
}, |
|
{ |
|
"epoch": 4.48, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 10850 |
|
}, |
|
{ |
|
"epoch": 4.5, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 10900 |
|
}, |
|
{ |
|
"epoch": 4.52, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 10950 |
|
}, |
|
{ |
|
"epoch": 4.54, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 4.56, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 11050 |
|
}, |
|
{ |
|
"epoch": 4.58, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 11100 |
|
}, |
|
{ |
|
"epoch": 4.6, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 11150 |
|
}, |
|
{ |
|
"epoch": 4.62, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 11200 |
|
}, |
|
{ |
|
"epoch": 4.64, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 11250 |
|
}, |
|
{ |
|
"epoch": 4.66, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 11300 |
|
}, |
|
{ |
|
"epoch": 4.68, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 11350 |
|
}, |
|
{ |
|
"epoch": 4.7, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 11400 |
|
}, |
|
{ |
|
"epoch": 4.73, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 11450 |
|
}, |
|
{ |
|
"epoch": 4.75, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 4.77, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 11550 |
|
}, |
|
{ |
|
"epoch": 4.79, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 11600 |
|
}, |
|
{ |
|
"epoch": 4.81, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 11650 |
|
}, |
|
{ |
|
"epoch": 4.83, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 11700 |
|
}, |
|
{ |
|
"epoch": 4.85, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 11750 |
|
}, |
|
{ |
|
"epoch": 4.87, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 11800 |
|
}, |
|
{ |
|
"epoch": 4.89, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 11850 |
|
}, |
|
{ |
|
"epoch": 4.91, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 11900 |
|
}, |
|
{ |
|
"epoch": 4.93, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 11950 |
|
}, |
|
{ |
|
"epoch": 4.95, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 4.97, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 12050 |
|
}, |
|
{ |
|
"epoch": 4.99, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 12100 |
|
}, |
|
{ |
|
"epoch": 5.01, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 12150 |
|
}, |
|
{ |
|
"epoch": 5.04, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 12200 |
|
}, |
|
{ |
|
"epoch": 5.06, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 12250 |
|
}, |
|
{ |
|
"epoch": 5.08, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 12300 |
|
}, |
|
{ |
|
"epoch": 5.1, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 12350 |
|
}, |
|
{ |
|
"epoch": 5.12, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 12400 |
|
}, |
|
{ |
|
"epoch": 5.14, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 12450 |
|
}, |
|
{ |
|
"epoch": 5.16, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 5.16, |
|
"eval_loss": NaN, |
|
"eval_runtime": 6.0384, |
|
"eval_samples_per_second": 235.162, |
|
"eval_steps_per_second": 14.739, |
|
"eval_wer": 1.0, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 5.18, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 12550 |
|
}, |
|
{ |
|
"epoch": 5.2, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 12600 |
|
}, |
|
{ |
|
"epoch": 5.22, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 12650 |
|
}, |
|
{ |
|
"epoch": 5.24, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 12700 |
|
}, |
|
{ |
|
"epoch": 5.26, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 12750 |
|
}, |
|
{ |
|
"epoch": 5.28, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 12800 |
|
}, |
|
{ |
|
"epoch": 5.3, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 12850 |
|
}, |
|
{ |
|
"epoch": 5.32, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 12900 |
|
}, |
|
{ |
|
"epoch": 5.34, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 12950 |
|
}, |
|
{ |
|
"epoch": 5.37, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 5.39, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 13050 |
|
}, |
|
{ |
|
"epoch": 5.41, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 13100 |
|
}, |
|
{ |
|
"epoch": 5.43, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 13150 |
|
}, |
|
{ |
|
"epoch": 5.45, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 13200 |
|
}, |
|
{ |
|
"epoch": 5.47, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 13250 |
|
}, |
|
{ |
|
"epoch": 5.49, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 13300 |
|
}, |
|
{ |
|
"epoch": 5.51, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 13350 |
|
}, |
|
{ |
|
"epoch": 5.53, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 13400 |
|
}, |
|
{ |
|
"epoch": 5.55, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 13450 |
|
}, |
|
{ |
|
"epoch": 5.57, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 5.59, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 13550 |
|
}, |
|
{ |
|
"epoch": 5.61, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 13600 |
|
}, |
|
{ |
|
"epoch": 5.63, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 13650 |
|
}, |
|
{ |
|
"epoch": 5.65, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 13700 |
|
}, |
|
{ |
|
"epoch": 5.67, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 13750 |
|
}, |
|
{ |
|
"epoch": 5.7, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 13800 |
|
}, |
|
{ |
|
"epoch": 5.72, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 13850 |
|
}, |
|
{ |
|
"epoch": 5.74, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 13900 |
|
}, |
|
{ |
|
"epoch": 5.76, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 13950 |
|
}, |
|
{ |
|
"epoch": 5.78, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 5.8, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 14050 |
|
}, |
|
{ |
|
"epoch": 5.82, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 14100 |
|
}, |
|
{ |
|
"epoch": 5.84, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 14150 |
|
}, |
|
{ |
|
"epoch": 5.86, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 14200 |
|
}, |
|
{ |
|
"epoch": 5.88, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 14250 |
|
}, |
|
{ |
|
"epoch": 5.9, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 14300 |
|
}, |
|
{ |
|
"epoch": 5.92, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 14350 |
|
}, |
|
{ |
|
"epoch": 5.94, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 14400 |
|
}, |
|
{ |
|
"epoch": 5.96, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 14450 |
|
}, |
|
{ |
|
"epoch": 5.98, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 14550 |
|
}, |
|
{ |
|
"epoch": 6.03, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 14600 |
|
}, |
|
{ |
|
"epoch": 6.05, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 14650 |
|
}, |
|
{ |
|
"epoch": 6.07, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 14700 |
|
}, |
|
{ |
|
"epoch": 6.09, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 14750 |
|
}, |
|
{ |
|
"epoch": 6.11, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 14800 |
|
}, |
|
{ |
|
"epoch": 6.13, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 14850 |
|
}, |
|
{ |
|
"epoch": 6.15, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 14900 |
|
}, |
|
{ |
|
"epoch": 6.17, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 14950 |
|
}, |
|
{ |
|
"epoch": 6.19, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 6.19, |
|
"eval_loss": NaN, |
|
"eval_runtime": 7.3329, |
|
"eval_samples_per_second": 193.649, |
|
"eval_steps_per_second": 12.137, |
|
"eval_wer": 1.0, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 6.21, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 15050 |
|
}, |
|
{ |
|
"epoch": 6.23, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 15100 |
|
}, |
|
{ |
|
"epoch": 6.25, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 15150 |
|
}, |
|
{ |
|
"epoch": 6.27, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 15200 |
|
}, |
|
{ |
|
"epoch": 6.29, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 15250 |
|
}, |
|
{ |
|
"epoch": 6.31, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 15300 |
|
}, |
|
{ |
|
"epoch": 6.34, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 15350 |
|
}, |
|
{ |
|
"epoch": 6.36, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 15400 |
|
}, |
|
{ |
|
"epoch": 6.38, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 15450 |
|
}, |
|
{ |
|
"epoch": 6.4, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 6.42, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 15550 |
|
}, |
|
{ |
|
"epoch": 6.44, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 15600 |
|
}, |
|
{ |
|
"epoch": 6.46, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 15650 |
|
}, |
|
{ |
|
"epoch": 6.48, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 15700 |
|
}, |
|
{ |
|
"epoch": 6.5, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 15750 |
|
}, |
|
{ |
|
"epoch": 6.52, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 15800 |
|
}, |
|
{ |
|
"epoch": 6.54, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 15850 |
|
}, |
|
{ |
|
"epoch": 6.56, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 15900 |
|
}, |
|
{ |
|
"epoch": 6.58, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 15950 |
|
}, |
|
{ |
|
"epoch": 6.6, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 6.62, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 16050 |
|
}, |
|
{ |
|
"epoch": 6.64, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 16100 |
|
}, |
|
{ |
|
"epoch": 6.67, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 16150 |
|
}, |
|
{ |
|
"epoch": 6.69, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 16200 |
|
}, |
|
{ |
|
"epoch": 6.71, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 16250 |
|
}, |
|
{ |
|
"epoch": 6.73, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 16300 |
|
}, |
|
{ |
|
"epoch": 6.75, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 16350 |
|
}, |
|
{ |
|
"epoch": 6.77, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 16400 |
|
}, |
|
{ |
|
"epoch": 6.79, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 16450 |
|
}, |
|
{ |
|
"epoch": 6.81, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 6.83, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 16550 |
|
}, |
|
{ |
|
"epoch": 6.85, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 16600 |
|
}, |
|
{ |
|
"epoch": 6.87, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 16650 |
|
}, |
|
{ |
|
"epoch": 6.89, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 16700 |
|
}, |
|
{ |
|
"epoch": 6.91, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 16750 |
|
}, |
|
{ |
|
"epoch": 6.93, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 16800 |
|
}, |
|
{ |
|
"epoch": 6.95, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 16850 |
|
}, |
|
{ |
|
"epoch": 6.97, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 16900 |
|
}, |
|
{ |
|
"epoch": 7.0, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 16950 |
|
}, |
|
{ |
|
"epoch": 7.02, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 7.04, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 17050 |
|
}, |
|
{ |
|
"epoch": 7.06, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 17100 |
|
}, |
|
{ |
|
"epoch": 7.08, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 17150 |
|
}, |
|
{ |
|
"epoch": 7.1, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 17200 |
|
}, |
|
{ |
|
"epoch": 7.12, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 17250 |
|
}, |
|
{ |
|
"epoch": 7.14, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 17300 |
|
}, |
|
{ |
|
"epoch": 7.16, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 17350 |
|
}, |
|
{ |
|
"epoch": 7.18, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 17400 |
|
}, |
|
{ |
|
"epoch": 7.2, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 17450 |
|
}, |
|
{ |
|
"epoch": 7.22, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 7.22, |
|
"eval_loss": NaN, |
|
"eval_runtime": 7.6499, |
|
"eval_samples_per_second": 185.624, |
|
"eval_steps_per_second": 11.634, |
|
"eval_wer": 1.0, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 7.24, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 17550 |
|
}, |
|
{ |
|
"epoch": 7.26, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 17600 |
|
}, |
|
{ |
|
"epoch": 7.28, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 17650 |
|
}, |
|
{ |
|
"epoch": 7.3, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 17700 |
|
}, |
|
{ |
|
"epoch": 7.33, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 17750 |
|
}, |
|
{ |
|
"epoch": 7.35, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 17800 |
|
}, |
|
{ |
|
"epoch": 7.37, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 17850 |
|
}, |
|
{ |
|
"epoch": 7.39, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 17900 |
|
}, |
|
{ |
|
"epoch": 7.41, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 17950 |
|
}, |
|
{ |
|
"epoch": 7.43, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 7.45, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 18050 |
|
}, |
|
{ |
|
"epoch": 7.47, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 18100 |
|
}, |
|
{ |
|
"epoch": 7.49, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 18150 |
|
}, |
|
{ |
|
"epoch": 7.51, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 18200 |
|
}, |
|
{ |
|
"epoch": 7.53, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 18250 |
|
}, |
|
{ |
|
"epoch": 7.55, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 18300 |
|
}, |
|
{ |
|
"epoch": 7.57, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 18350 |
|
}, |
|
{ |
|
"epoch": 7.59, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 18400 |
|
}, |
|
{ |
|
"epoch": 7.61, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 18450 |
|
}, |
|
{ |
|
"epoch": 7.64, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 7.66, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 18550 |
|
}, |
|
{ |
|
"epoch": 7.68, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 18600 |
|
}, |
|
{ |
|
"epoch": 7.7, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 18650 |
|
}, |
|
{ |
|
"epoch": 7.72, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 18700 |
|
}, |
|
{ |
|
"epoch": 7.74, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 18750 |
|
}, |
|
{ |
|
"epoch": 7.76, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 18800 |
|
}, |
|
{ |
|
"epoch": 7.78, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 18850 |
|
}, |
|
{ |
|
"epoch": 7.8, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 18900 |
|
}, |
|
{ |
|
"epoch": 7.82, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 18950 |
|
}, |
|
{ |
|
"epoch": 7.84, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 7.86, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 19050 |
|
}, |
|
{ |
|
"epoch": 7.88, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 19100 |
|
}, |
|
{ |
|
"epoch": 7.9, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 19150 |
|
}, |
|
{ |
|
"epoch": 7.92, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 19200 |
|
}, |
|
{ |
|
"epoch": 7.94, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 19250 |
|
}, |
|
{ |
|
"epoch": 7.97, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 19300 |
|
}, |
|
{ |
|
"epoch": 7.99, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 19350 |
|
}, |
|
{ |
|
"epoch": 8.01, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 19400 |
|
}, |
|
{ |
|
"epoch": 8.03, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 19450 |
|
}, |
|
{ |
|
"epoch": 8.05, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 8.07, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 19550 |
|
}, |
|
{ |
|
"epoch": 8.09, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 19600 |
|
}, |
|
{ |
|
"epoch": 8.11, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 19650 |
|
}, |
|
{ |
|
"epoch": 8.13, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 19700 |
|
}, |
|
{ |
|
"epoch": 8.15, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 19750 |
|
}, |
|
{ |
|
"epoch": 8.17, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 19800 |
|
}, |
|
{ |
|
"epoch": 8.19, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 19850 |
|
}, |
|
{ |
|
"epoch": 8.21, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 19900 |
|
}, |
|
{ |
|
"epoch": 8.23, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 19950 |
|
}, |
|
{ |
|
"epoch": 8.25, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 8.25, |
|
"eval_loss": NaN, |
|
"eval_runtime": 7.5114, |
|
"eval_samples_per_second": 189.047, |
|
"eval_steps_per_second": 11.849, |
|
"eval_wer": 1.0, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 8.27, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 20050 |
|
}, |
|
{ |
|
"epoch": 8.3, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 20100 |
|
}, |
|
{ |
|
"epoch": 8.32, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 20150 |
|
}, |
|
{ |
|
"epoch": 8.34, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 20200 |
|
}, |
|
{ |
|
"epoch": 8.36, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 20250 |
|
}, |
|
{ |
|
"epoch": 8.38, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 20300 |
|
}, |
|
{ |
|
"epoch": 8.4, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 20350 |
|
}, |
|
{ |
|
"epoch": 8.42, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 20400 |
|
}, |
|
{ |
|
"epoch": 8.44, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 20450 |
|
}, |
|
{ |
|
"epoch": 8.46, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 8.48, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 20550 |
|
}, |
|
{ |
|
"epoch": 8.5, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 20600 |
|
}, |
|
{ |
|
"epoch": 8.52, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 20650 |
|
}, |
|
{ |
|
"epoch": 8.54, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 20700 |
|
}, |
|
{ |
|
"epoch": 8.56, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 20750 |
|
}, |
|
{ |
|
"epoch": 8.58, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 20800 |
|
}, |
|
{ |
|
"epoch": 8.61, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 20850 |
|
}, |
|
{ |
|
"epoch": 8.63, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 20900 |
|
}, |
|
{ |
|
"epoch": 8.65, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 20950 |
|
}, |
|
{ |
|
"epoch": 8.67, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 8.69, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 21050 |
|
}, |
|
{ |
|
"epoch": 8.71, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 21100 |
|
}, |
|
{ |
|
"epoch": 8.73, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 21150 |
|
}, |
|
{ |
|
"epoch": 8.75, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 21200 |
|
}, |
|
{ |
|
"epoch": 8.77, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 21250 |
|
}, |
|
{ |
|
"epoch": 8.79, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 21300 |
|
}, |
|
{ |
|
"epoch": 8.81, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 21350 |
|
}, |
|
{ |
|
"epoch": 8.83, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 21400 |
|
}, |
|
{ |
|
"epoch": 8.85, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 21450 |
|
}, |
|
{ |
|
"epoch": 8.87, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 8.89, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 21550 |
|
}, |
|
{ |
|
"epoch": 8.91, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 21600 |
|
}, |
|
{ |
|
"epoch": 8.94, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 21650 |
|
}, |
|
{ |
|
"epoch": 8.96, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 21700 |
|
}, |
|
{ |
|
"epoch": 8.98, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 21750 |
|
}, |
|
{ |
|
"epoch": 9.0, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 21800 |
|
}, |
|
{ |
|
"epoch": 9.02, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 21850 |
|
}, |
|
{ |
|
"epoch": 9.04, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 21900 |
|
}, |
|
{ |
|
"epoch": 9.06, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 21950 |
|
}, |
|
{ |
|
"epoch": 9.08, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 9.1, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 22050 |
|
}, |
|
{ |
|
"epoch": 9.12, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 22100 |
|
}, |
|
{ |
|
"epoch": 9.14, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 22150 |
|
}, |
|
{ |
|
"epoch": 9.16, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 22200 |
|
}, |
|
{ |
|
"epoch": 9.18, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 22250 |
|
}, |
|
{ |
|
"epoch": 9.2, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 22300 |
|
}, |
|
{ |
|
"epoch": 9.22, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 22350 |
|
}, |
|
{ |
|
"epoch": 9.24, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 22400 |
|
}, |
|
{ |
|
"epoch": 9.27, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 22450 |
|
}, |
|
{ |
|
"epoch": 9.29, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 9.29, |
|
"eval_loss": NaN, |
|
"eval_runtime": 5.6959, |
|
"eval_samples_per_second": 249.304, |
|
"eval_steps_per_second": 15.625, |
|
"eval_wer": 1.0, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 9.31, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 22550 |
|
}, |
|
{ |
|
"epoch": 9.33, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 22600 |
|
}, |
|
{ |
|
"epoch": 9.35, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 22650 |
|
}, |
|
{ |
|
"epoch": 9.37, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 22700 |
|
}, |
|
{ |
|
"epoch": 9.39, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 22750 |
|
}, |
|
{ |
|
"epoch": 9.41, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 22800 |
|
}, |
|
{ |
|
"epoch": 9.43, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 22850 |
|
}, |
|
{ |
|
"epoch": 9.45, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 22900 |
|
}, |
|
{ |
|
"epoch": 9.47, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 22950 |
|
}, |
|
{ |
|
"epoch": 9.49, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 9.51, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 23050 |
|
}, |
|
{ |
|
"epoch": 9.53, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 23100 |
|
}, |
|
{ |
|
"epoch": 9.55, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 23150 |
|
}, |
|
{ |
|
"epoch": 9.57, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 23200 |
|
}, |
|
{ |
|
"epoch": 9.6, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 23250 |
|
}, |
|
{ |
|
"epoch": 9.62, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 23300 |
|
}, |
|
{ |
|
"epoch": 9.64, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 23350 |
|
}, |
|
{ |
|
"epoch": 9.66, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 23400 |
|
}, |
|
{ |
|
"epoch": 9.68, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 23450 |
|
}, |
|
{ |
|
"epoch": 9.7, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 9.72, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 23550 |
|
}, |
|
{ |
|
"epoch": 9.74, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 23600 |
|
}, |
|
{ |
|
"epoch": 9.76, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 23650 |
|
}, |
|
{ |
|
"epoch": 9.78, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 23700 |
|
}, |
|
{ |
|
"epoch": 9.8, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 23750 |
|
}, |
|
{ |
|
"epoch": 9.82, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 23800 |
|
}, |
|
{ |
|
"epoch": 9.84, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 23850 |
|
}, |
|
{ |
|
"epoch": 9.86, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 23900 |
|
}, |
|
{ |
|
"epoch": 9.88, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 23950 |
|
}, |
|
{ |
|
"epoch": 9.91, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 9.93, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 24050 |
|
}, |
|
{ |
|
"epoch": 9.95, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 24100 |
|
}, |
|
{ |
|
"epoch": 9.97, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 24150 |
|
}, |
|
{ |
|
"epoch": 9.99, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 24200 |
|
}, |
|
{ |
|
"epoch": 10.01, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 24250 |
|
}, |
|
{ |
|
"epoch": 10.03, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 24300 |
|
}, |
|
{ |
|
"epoch": 10.05, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 24350 |
|
}, |
|
{ |
|
"epoch": 10.07, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 24400 |
|
}, |
|
{ |
|
"epoch": 10.09, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 24450 |
|
}, |
|
{ |
|
"epoch": 10.11, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 10.13, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 24550 |
|
}, |
|
{ |
|
"epoch": 10.15, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 24600 |
|
}, |
|
{ |
|
"epoch": 10.17, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 24650 |
|
}, |
|
{ |
|
"epoch": 10.19, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 24700 |
|
}, |
|
{ |
|
"epoch": 10.21, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 24750 |
|
}, |
|
{ |
|
"epoch": 10.24, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 24800 |
|
}, |
|
{ |
|
"epoch": 10.26, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 24850 |
|
}, |
|
{ |
|
"epoch": 10.28, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 24900 |
|
}, |
|
{ |
|
"epoch": 10.3, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 24950 |
|
}, |
|
{ |
|
"epoch": 10.32, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 10.32, |
|
"eval_loss": NaN, |
|
"eval_runtime": 7.3757, |
|
"eval_samples_per_second": 192.524, |
|
"eval_steps_per_second": 12.067, |
|
"eval_wer": 1.0, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 10.34, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 25050 |
|
}, |
|
{ |
|
"epoch": 10.36, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 25100 |
|
}, |
|
{ |
|
"epoch": 10.38, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 25150 |
|
}, |
|
{ |
|
"epoch": 10.4, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 25200 |
|
}, |
|
{ |
|
"epoch": 10.42, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 25250 |
|
}, |
|
{ |
|
"epoch": 10.44, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 25300 |
|
}, |
|
{ |
|
"epoch": 10.46, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 25350 |
|
}, |
|
{ |
|
"epoch": 10.48, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 25400 |
|
}, |
|
{ |
|
"epoch": 10.5, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 25450 |
|
}, |
|
{ |
|
"epoch": 10.52, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 25500 |
|
}, |
|
{ |
|
"epoch": 10.54, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 25550 |
|
}, |
|
{ |
|
"epoch": 10.57, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 25600 |
|
}, |
|
{ |
|
"epoch": 10.59, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 25650 |
|
}, |
|
{ |
|
"epoch": 10.61, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 25700 |
|
}, |
|
{ |
|
"epoch": 10.63, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 25750 |
|
}, |
|
{ |
|
"epoch": 10.65, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 25800 |
|
}, |
|
{ |
|
"epoch": 10.67, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 25850 |
|
}, |
|
{ |
|
"epoch": 10.69, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 25900 |
|
}, |
|
{ |
|
"epoch": 10.71, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 25950 |
|
}, |
|
{ |
|
"epoch": 10.73, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 10.75, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 26050 |
|
}, |
|
{ |
|
"epoch": 10.77, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 26100 |
|
}, |
|
{ |
|
"epoch": 10.79, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 26150 |
|
}, |
|
{ |
|
"epoch": 10.81, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 26200 |
|
}, |
|
{ |
|
"epoch": 10.83, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 26250 |
|
}, |
|
{ |
|
"epoch": 10.85, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 26300 |
|
}, |
|
{ |
|
"epoch": 10.87, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 26350 |
|
}, |
|
{ |
|
"epoch": 10.9, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 26400 |
|
}, |
|
{ |
|
"epoch": 10.92, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 26450 |
|
}, |
|
{ |
|
"epoch": 10.94, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 26500 |
|
}, |
|
{ |
|
"epoch": 10.96, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 26550 |
|
}, |
|
{ |
|
"epoch": 10.98, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 26600 |
|
}, |
|
{ |
|
"epoch": 11.0, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 26650 |
|
}, |
|
{ |
|
"epoch": 11.02, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 26700 |
|
}, |
|
{ |
|
"epoch": 11.04, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 26750 |
|
}, |
|
{ |
|
"epoch": 11.06, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 26800 |
|
}, |
|
{ |
|
"epoch": 11.08, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 26850 |
|
}, |
|
{ |
|
"epoch": 11.1, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 26900 |
|
}, |
|
{ |
|
"epoch": 11.12, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 26950 |
|
}, |
|
{ |
|
"epoch": 11.14, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 11.16, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 27050 |
|
}, |
|
{ |
|
"epoch": 11.18, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 27100 |
|
}, |
|
{ |
|
"epoch": 11.21, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 27150 |
|
}, |
|
{ |
|
"epoch": 11.23, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 27200 |
|
}, |
|
{ |
|
"epoch": 11.25, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 27250 |
|
}, |
|
{ |
|
"epoch": 11.27, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 27300 |
|
}, |
|
{ |
|
"epoch": 11.29, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 27350 |
|
}, |
|
{ |
|
"epoch": 11.31, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 27400 |
|
}, |
|
{ |
|
"epoch": 11.33, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 27450 |
|
}, |
|
{ |
|
"epoch": 11.35, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 27500 |
|
}, |
|
{ |
|
"epoch": 11.35, |
|
"eval_loss": NaN, |
|
"eval_runtime": 7.2191, |
|
"eval_samples_per_second": 196.702, |
|
"eval_steps_per_second": 12.328, |
|
"eval_wer": 1.0, |
|
"step": 27500 |
|
}, |
|
{ |
|
"epoch": 11.37, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 27550 |
|
}, |
|
{ |
|
"epoch": 11.39, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 27600 |
|
}, |
|
{ |
|
"epoch": 11.41, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 27650 |
|
}, |
|
{ |
|
"epoch": 11.43, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 27700 |
|
}, |
|
{ |
|
"epoch": 11.45, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 27750 |
|
}, |
|
{ |
|
"epoch": 11.47, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 27800 |
|
}, |
|
{ |
|
"epoch": 11.49, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 27850 |
|
}, |
|
{ |
|
"epoch": 11.51, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 27900 |
|
}, |
|
{ |
|
"epoch": 11.54, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 27950 |
|
}, |
|
{ |
|
"epoch": 11.56, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 11.58, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 28050 |
|
}, |
|
{ |
|
"epoch": 11.6, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 28100 |
|
}, |
|
{ |
|
"epoch": 11.62, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 28150 |
|
}, |
|
{ |
|
"epoch": 11.64, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 28200 |
|
}, |
|
{ |
|
"epoch": 11.66, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 28250 |
|
}, |
|
{ |
|
"epoch": 11.68, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 28300 |
|
}, |
|
{ |
|
"epoch": 11.7, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 28350 |
|
}, |
|
{ |
|
"epoch": 11.72, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 28400 |
|
}, |
|
{ |
|
"epoch": 11.74, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 28450 |
|
}, |
|
{ |
|
"epoch": 11.76, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 28500 |
|
}, |
|
{ |
|
"epoch": 11.78, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 28550 |
|
}, |
|
{ |
|
"epoch": 11.8, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 28600 |
|
}, |
|
{ |
|
"epoch": 11.82, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 28650 |
|
}, |
|
{ |
|
"epoch": 11.84, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 28700 |
|
}, |
|
{ |
|
"epoch": 11.87, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 28750 |
|
}, |
|
{ |
|
"epoch": 11.89, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 28800 |
|
}, |
|
{ |
|
"epoch": 11.91, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 28850 |
|
}, |
|
{ |
|
"epoch": 11.93, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 28900 |
|
}, |
|
{ |
|
"epoch": 11.95, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 28950 |
|
}, |
|
{ |
|
"epoch": 11.97, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 11.99, |
|
"learning_rate": 1.5703913176398656e-05, |
|
"loss": 0.0, |
|
"step": 29050 |
|
}, |
|
{ |
|
"epoch": 12.0, |
|
"step": 29076, |
|
"total_flos": 1.3204107032278545e+20, |
|
"train_loss": 1.314115549744299, |
|
"train_runtime": 6951.6256, |
|
"train_samples_per_second": 167.289, |
|
"train_steps_per_second": 4.183 |
|
} |
|
], |
|
"logging_steps": 50, |
|
"max_steps": 29076, |
|
"num_train_epochs": 12, |
|
"save_steps": 500, |
|
"total_flos": 1.3204107032278545e+20, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|