whisper-large-v3-ft-cv-cy-en / trainer_state.json
DewiBrynJones's picture
End of training
6466ac2 verified
raw
history blame
36.6 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 3.537318712415989,
"eval_steps": 1000,
"global_step": 5000,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.017686593562079942,
"grad_norm": 11.730201721191406,
"learning_rate": 4.800000000000001e-07,
"loss": 1.906,
"step": 25
},
{
"epoch": 0.035373187124159884,
"grad_norm": 9.43655014038086,
"learning_rate": 9.800000000000001e-07,
"loss": 1.5891,
"step": 50
},
{
"epoch": 0.05305978068623983,
"grad_norm": 8.554606437683105,
"learning_rate": 1.48e-06,
"loss": 0.9582,
"step": 75
},
{
"epoch": 0.07074637424831977,
"grad_norm": 7.075723171234131,
"learning_rate": 1.98e-06,
"loss": 0.8164,
"step": 100
},
{
"epoch": 0.08843296781039972,
"grad_norm": 7.806982517242432,
"learning_rate": 2.46e-06,
"loss": 0.7283,
"step": 125
},
{
"epoch": 0.10611956137247966,
"grad_norm": 6.871402740478516,
"learning_rate": 2.96e-06,
"loss": 0.721,
"step": 150
},
{
"epoch": 0.1238061549345596,
"grad_norm": 7.813017845153809,
"learning_rate": 3.46e-06,
"loss": 0.68,
"step": 175
},
{
"epoch": 0.14149274849663954,
"grad_norm": 7.54969596862793,
"learning_rate": 3.96e-06,
"loss": 0.7248,
"step": 200
},
{
"epoch": 0.1591793420587195,
"grad_norm": 7.196774482727051,
"learning_rate": 4.4600000000000005e-06,
"loss": 0.6417,
"step": 225
},
{
"epoch": 0.17686593562079944,
"grad_norm": 8.30762004852295,
"learning_rate": 4.960000000000001e-06,
"loss": 0.6594,
"step": 250
},
{
"epoch": 0.19455252918287938,
"grad_norm": 6.641123294830322,
"learning_rate": 5.460000000000001e-06,
"loss": 0.6359,
"step": 275
},
{
"epoch": 0.21223912274495932,
"grad_norm": 8.47127914428711,
"learning_rate": 5.9600000000000005e-06,
"loss": 0.6455,
"step": 300
},
{
"epoch": 0.22992571630703926,
"grad_norm": 8.078540802001953,
"learning_rate": 6.460000000000001e-06,
"loss": 0.6563,
"step": 325
},
{
"epoch": 0.2476123098691192,
"grad_norm": 7.481531620025635,
"learning_rate": 6.96e-06,
"loss": 0.6099,
"step": 350
},
{
"epoch": 0.26529890343119916,
"grad_norm": 8.171069145202637,
"learning_rate": 7.4600000000000006e-06,
"loss": 0.5922,
"step": 375
},
{
"epoch": 0.2829854969932791,
"grad_norm": 7.401958465576172,
"learning_rate": 7.960000000000002e-06,
"loss": 0.6232,
"step": 400
},
{
"epoch": 0.30067209055535904,
"grad_norm": 6.448988437652588,
"learning_rate": 8.46e-06,
"loss": 0.6328,
"step": 425
},
{
"epoch": 0.318358684117439,
"grad_norm": 6.956504821777344,
"learning_rate": 8.96e-06,
"loss": 0.6026,
"step": 450
},
{
"epoch": 0.3360452776795189,
"grad_norm": 8.774309158325195,
"learning_rate": 9.460000000000001e-06,
"loss": 0.6318,
"step": 475
},
{
"epoch": 0.3537318712415989,
"grad_norm": 6.220521450042725,
"learning_rate": 9.960000000000001e-06,
"loss": 0.6186,
"step": 500
},
{
"epoch": 0.3714184648036788,
"grad_norm": 6.165685176849365,
"learning_rate": 9.94888888888889e-06,
"loss": 0.5509,
"step": 525
},
{
"epoch": 0.38910505836575876,
"grad_norm": 6.3863348960876465,
"learning_rate": 9.893333333333334e-06,
"loss": 0.63,
"step": 550
},
{
"epoch": 0.40679165192783867,
"grad_norm": 6.65989351272583,
"learning_rate": 9.837777777777778e-06,
"loss": 0.5981,
"step": 575
},
{
"epoch": 0.42447824548991864,
"grad_norm": 5.605571746826172,
"learning_rate": 9.782222222222222e-06,
"loss": 0.5611,
"step": 600
},
{
"epoch": 0.4421648390519986,
"grad_norm": 7.080026149749756,
"learning_rate": 9.726666666666668e-06,
"loss": 0.5498,
"step": 625
},
{
"epoch": 0.4598514326140785,
"grad_norm": 5.857707977294922,
"learning_rate": 9.671111111111112e-06,
"loss": 0.5942,
"step": 650
},
{
"epoch": 0.4775380261761585,
"grad_norm": 6.002816200256348,
"learning_rate": 9.615555555555558e-06,
"loss": 0.5488,
"step": 675
},
{
"epoch": 0.4952246197382384,
"grad_norm": 6.393026828765869,
"learning_rate": 9.56e-06,
"loss": 0.5464,
"step": 700
},
{
"epoch": 0.5129112133003184,
"grad_norm": 6.212553024291992,
"learning_rate": 9.504444444444446e-06,
"loss": 0.5416,
"step": 725
},
{
"epoch": 0.5305978068623983,
"grad_norm": 6.353979110717773,
"learning_rate": 9.44888888888889e-06,
"loss": 0.5493,
"step": 750
},
{
"epoch": 0.5482844004244782,
"grad_norm": 5.198184013366699,
"learning_rate": 9.393333333333334e-06,
"loss": 0.562,
"step": 775
},
{
"epoch": 0.5659709939865581,
"grad_norm": 5.739233493804932,
"learning_rate": 9.33777777777778e-06,
"loss": 0.5245,
"step": 800
},
{
"epoch": 0.5836575875486382,
"grad_norm": 5.637094974517822,
"learning_rate": 9.282222222222222e-06,
"loss": 0.5056,
"step": 825
},
{
"epoch": 0.6013441811107181,
"grad_norm": 5.2869181632995605,
"learning_rate": 9.226666666666668e-06,
"loss": 0.495,
"step": 850
},
{
"epoch": 0.619030774672798,
"grad_norm": 6.895484447479248,
"learning_rate": 9.171111111111112e-06,
"loss": 0.5311,
"step": 875
},
{
"epoch": 0.636717368234878,
"grad_norm": 5.642014026641846,
"learning_rate": 9.115555555555556e-06,
"loss": 0.4611,
"step": 900
},
{
"epoch": 0.6544039617969579,
"grad_norm": 6.39215087890625,
"learning_rate": 9.060000000000001e-06,
"loss": 0.5128,
"step": 925
},
{
"epoch": 0.6720905553590378,
"grad_norm": 4.879695892333984,
"learning_rate": 9.004444444444445e-06,
"loss": 0.502,
"step": 950
},
{
"epoch": 0.6897771489211177,
"grad_norm": 6.497096538543701,
"learning_rate": 8.94888888888889e-06,
"loss": 0.4927,
"step": 975
},
{
"epoch": 0.7074637424831978,
"grad_norm": 6.05295991897583,
"learning_rate": 8.893333333333333e-06,
"loss": 0.4825,
"step": 1000
},
{
"epoch": 0.7074637424831978,
"eval_loss": 0.2707957625389099,
"eval_runtime": 4792.033,
"eval_samples_per_second": 2.246,
"eval_steps_per_second": 0.14,
"eval_wer": 0.18103144801689744,
"step": 1000
},
{
"epoch": 0.7251503360452777,
"grad_norm": 6.068356990814209,
"learning_rate": 8.83777777777778e-06,
"loss": 0.4787,
"step": 1025
},
{
"epoch": 0.7428369296073576,
"grad_norm": 6.832317352294922,
"learning_rate": 8.782222222222223e-06,
"loss": 0.4583,
"step": 1050
},
{
"epoch": 0.7605235231694376,
"grad_norm": 5.528912544250488,
"learning_rate": 8.726666666666667e-06,
"loss": 0.4733,
"step": 1075
},
{
"epoch": 0.7782101167315175,
"grad_norm": 6.220973968505859,
"learning_rate": 8.671111111111113e-06,
"loss": 0.4994,
"step": 1100
},
{
"epoch": 0.7958967102935974,
"grad_norm": 6.543335914611816,
"learning_rate": 8.615555555555555e-06,
"loss": 0.4694,
"step": 1125
},
{
"epoch": 0.8135833038556773,
"grad_norm": 4.369144439697266,
"learning_rate": 8.560000000000001e-06,
"loss": 0.4525,
"step": 1150
},
{
"epoch": 0.8312698974177574,
"grad_norm": 4.928637981414795,
"learning_rate": 8.504444444444445e-06,
"loss": 0.4802,
"step": 1175
},
{
"epoch": 0.8489564909798373,
"grad_norm": 5.016990661621094,
"learning_rate": 8.448888888888889e-06,
"loss": 0.462,
"step": 1200
},
{
"epoch": 0.8666430845419172,
"grad_norm": 5.185001373291016,
"learning_rate": 8.393333333333335e-06,
"loss": 0.4189,
"step": 1225
},
{
"epoch": 0.8843296781039972,
"grad_norm": 5.9500627517700195,
"learning_rate": 8.337777777777777e-06,
"loss": 0.4862,
"step": 1250
},
{
"epoch": 0.9020162716660771,
"grad_norm": 4.7840166091918945,
"learning_rate": 8.282222222222223e-06,
"loss": 0.4201,
"step": 1275
},
{
"epoch": 0.919702865228157,
"grad_norm": 5.590398788452148,
"learning_rate": 8.226666666666667e-06,
"loss": 0.438,
"step": 1300
},
{
"epoch": 0.937389458790237,
"grad_norm": 5.356804370880127,
"learning_rate": 8.171111111111113e-06,
"loss": 0.4343,
"step": 1325
},
{
"epoch": 0.955076052352317,
"grad_norm": 5.994735240936279,
"learning_rate": 8.115555555555557e-06,
"loss": 0.4281,
"step": 1350
},
{
"epoch": 0.9727626459143969,
"grad_norm": 4.92245626449585,
"learning_rate": 8.06e-06,
"loss": 0.4711,
"step": 1375
},
{
"epoch": 0.9904492394764768,
"grad_norm": 6.673391342163086,
"learning_rate": 8.004444444444445e-06,
"loss": 0.4396,
"step": 1400
},
{
"epoch": 1.0081358330385568,
"grad_norm": 4.157800197601318,
"learning_rate": 7.948888888888889e-06,
"loss": 0.3655,
"step": 1425
},
{
"epoch": 1.0258224266006368,
"grad_norm": 4.4156060218811035,
"learning_rate": 7.893333333333335e-06,
"loss": 0.3032,
"step": 1450
},
{
"epoch": 1.0435090201627166,
"grad_norm": 3.762485980987549,
"learning_rate": 7.837777777777779e-06,
"loss": 0.2597,
"step": 1475
},
{
"epoch": 1.0611956137247966,
"grad_norm": 3.0248515605926514,
"learning_rate": 7.782222222222223e-06,
"loss": 0.2829,
"step": 1500
},
{
"epoch": 1.0788822072868764,
"grad_norm": 5.492848873138428,
"learning_rate": 7.726666666666667e-06,
"loss": 0.229,
"step": 1525
},
{
"epoch": 1.0965688008489565,
"grad_norm": 5.225216388702393,
"learning_rate": 7.67111111111111e-06,
"loss": 0.2677,
"step": 1550
},
{
"epoch": 1.1142553944110365,
"grad_norm": 3.8239035606384277,
"learning_rate": 7.6155555555555564e-06,
"loss": 0.2517,
"step": 1575
},
{
"epoch": 1.1319419879731163,
"grad_norm": 5.077475070953369,
"learning_rate": 7.5600000000000005e-06,
"loss": 0.2489,
"step": 1600
},
{
"epoch": 1.1496285815351963,
"grad_norm": 4.364721298217773,
"learning_rate": 7.504444444444445e-06,
"loss": 0.2598,
"step": 1625
},
{
"epoch": 1.1673151750972763,
"grad_norm": 3.3700525760650635,
"learning_rate": 7.44888888888889e-06,
"loss": 0.2472,
"step": 1650
},
{
"epoch": 1.1850017686593561,
"grad_norm": 4.548206329345703,
"learning_rate": 7.393333333333333e-06,
"loss": 0.2635,
"step": 1675
},
{
"epoch": 1.2026883622214362,
"grad_norm": 4.475757122039795,
"learning_rate": 7.337777777777778e-06,
"loss": 0.2345,
"step": 1700
},
{
"epoch": 1.2203749557835162,
"grad_norm": 4.205091953277588,
"learning_rate": 7.282222222222222e-06,
"loss": 0.2464,
"step": 1725
},
{
"epoch": 1.238061549345596,
"grad_norm": 4.778344631195068,
"learning_rate": 7.226666666666667e-06,
"loss": 0.238,
"step": 1750
},
{
"epoch": 1.255748142907676,
"grad_norm": 3.172917604446411,
"learning_rate": 7.171111111111112e-06,
"loss": 0.2364,
"step": 1775
},
{
"epoch": 1.2734347364697558,
"grad_norm": 4.275852203369141,
"learning_rate": 7.115555555555557e-06,
"loss": 0.2573,
"step": 1800
},
{
"epoch": 1.2911213300318358,
"grad_norm": 4.308451175689697,
"learning_rate": 7.06e-06,
"loss": 0.2569,
"step": 1825
},
{
"epoch": 1.3088079235939158,
"grad_norm": 4.079460144042969,
"learning_rate": 7.004444444444445e-06,
"loss": 0.2261,
"step": 1850
},
{
"epoch": 1.3264945171559956,
"grad_norm": 6.641946792602539,
"learning_rate": 6.948888888888889e-06,
"loss": 0.2334,
"step": 1875
},
{
"epoch": 1.3441811107180757,
"grad_norm": 4.23126745223999,
"learning_rate": 6.893333333333334e-06,
"loss": 0.2307,
"step": 1900
},
{
"epoch": 1.3618677042801557,
"grad_norm": 3.1530086994171143,
"learning_rate": 6.837777777777779e-06,
"loss": 0.2154,
"step": 1925
},
{
"epoch": 1.3795542978422355,
"grad_norm": 3.8190791606903076,
"learning_rate": 6.782222222222222e-06,
"loss": 0.2274,
"step": 1950
},
{
"epoch": 1.3972408914043155,
"grad_norm": 3.953749179840088,
"learning_rate": 6.726666666666667e-06,
"loss": 0.2426,
"step": 1975
},
{
"epoch": 1.4149274849663955,
"grad_norm": 3.527085304260254,
"learning_rate": 6.671111111111112e-06,
"loss": 0.2262,
"step": 2000
},
{
"epoch": 1.4149274849663955,
"eval_loss": 0.2485629916191101,
"eval_runtime": 4723.0929,
"eval_samples_per_second": 2.279,
"eval_steps_per_second": 0.142,
"eval_wer": 0.15938160056324807,
"step": 2000
},
{
"epoch": 1.4326140785284753,
"grad_norm": 4.03377628326416,
"learning_rate": 6.615555555555556e-06,
"loss": 0.2367,
"step": 2025
},
{
"epoch": 1.4503006720905554,
"grad_norm": 3.128779888153076,
"learning_rate": 6.560000000000001e-06,
"loss": 0.2372,
"step": 2050
},
{
"epoch": 1.4679872656526354,
"grad_norm": 3.7657673358917236,
"learning_rate": 6.504444444444446e-06,
"loss": 0.2394,
"step": 2075
},
{
"epoch": 1.4856738592147152,
"grad_norm": 3.009976387023926,
"learning_rate": 6.448888888888889e-06,
"loss": 0.2285,
"step": 2100
},
{
"epoch": 1.5033604527767952,
"grad_norm": 3.916525363922119,
"learning_rate": 6.393333333333334e-06,
"loss": 0.2359,
"step": 2125
},
{
"epoch": 1.5210470463388752,
"grad_norm": 4.418327808380127,
"learning_rate": 6.3377777777777786e-06,
"loss": 0.2499,
"step": 2150
},
{
"epoch": 1.538733639900955,
"grad_norm": 4.2388105392456055,
"learning_rate": 6.282222222222223e-06,
"loss": 0.226,
"step": 2175
},
{
"epoch": 1.556420233463035,
"grad_norm": 3.664215564727783,
"learning_rate": 6.2266666666666675e-06,
"loss": 0.2316,
"step": 2200
},
{
"epoch": 1.574106827025115,
"grad_norm": 4.053125858306885,
"learning_rate": 6.171111111111112e-06,
"loss": 0.228,
"step": 2225
},
{
"epoch": 1.5917934205871949,
"grad_norm": 4.452697277069092,
"learning_rate": 6.1155555555555555e-06,
"loss": 0.2196,
"step": 2250
},
{
"epoch": 1.6094800141492749,
"grad_norm": 3.919962167739868,
"learning_rate": 6.0600000000000004e-06,
"loss": 0.2495,
"step": 2275
},
{
"epoch": 1.627166607711355,
"grad_norm": 4.33357572555542,
"learning_rate": 6.004444444444445e-06,
"loss": 0.2245,
"step": 2300
},
{
"epoch": 1.6448532012734347,
"grad_norm": 4.634097099304199,
"learning_rate": 5.948888888888889e-06,
"loss": 0.2438,
"step": 2325
},
{
"epoch": 1.6625397948355147,
"grad_norm": 4.564420223236084,
"learning_rate": 5.893333333333334e-06,
"loss": 0.2047,
"step": 2350
},
{
"epoch": 1.6802263883975948,
"grad_norm": 4.051385879516602,
"learning_rate": 5.837777777777777e-06,
"loss": 0.2392,
"step": 2375
},
{
"epoch": 1.6979129819596745,
"grad_norm": 3.313216209411621,
"learning_rate": 5.782222222222222e-06,
"loss": 0.2286,
"step": 2400
},
{
"epoch": 1.7155995755217543,
"grad_norm": 3.334672451019287,
"learning_rate": 5.726666666666667e-06,
"loss": 0.2296,
"step": 2425
},
{
"epoch": 1.7332861690838346,
"grad_norm": 3.519235134124756,
"learning_rate": 5.671111111111112e-06,
"loss": 0.2107,
"step": 2450
},
{
"epoch": 1.7509727626459144,
"grad_norm": 4.1753339767456055,
"learning_rate": 5.615555555555556e-06,
"loss": 0.2464,
"step": 2475
},
{
"epoch": 1.7686593562079942,
"grad_norm": 3.395427703857422,
"learning_rate": 5.560000000000001e-06,
"loss": 0.2323,
"step": 2500
},
{
"epoch": 1.7863459497700744,
"grad_norm": 3.480686902999878,
"learning_rate": 5.504444444444444e-06,
"loss": 0.2205,
"step": 2525
},
{
"epoch": 1.8040325433321542,
"grad_norm": 4.275322914123535,
"learning_rate": 5.448888888888889e-06,
"loss": 0.2147,
"step": 2550
},
{
"epoch": 1.821719136894234,
"grad_norm": 3.6306569576263428,
"learning_rate": 5.393333333333334e-06,
"loss": 0.2404,
"step": 2575
},
{
"epoch": 1.839405730456314,
"grad_norm": 3.426995038986206,
"learning_rate": 5.337777777777779e-06,
"loss": 0.2259,
"step": 2600
},
{
"epoch": 1.857092324018394,
"grad_norm": 4.106579303741455,
"learning_rate": 5.282222222222223e-06,
"loss": 0.2043,
"step": 2625
},
{
"epoch": 1.8747789175804739,
"grad_norm": 3.365471601486206,
"learning_rate": 5.226666666666667e-06,
"loss": 0.2179,
"step": 2650
},
{
"epoch": 1.892465511142554,
"grad_norm": 3.4525575637817383,
"learning_rate": 5.171111111111111e-06,
"loss": 0.2185,
"step": 2675
},
{
"epoch": 1.910152104704634,
"grad_norm": 3.5960958003997803,
"learning_rate": 5.115555555555556e-06,
"loss": 0.2007,
"step": 2700
},
{
"epoch": 1.9278386982667137,
"grad_norm": 3.91766619682312,
"learning_rate": 5.060000000000001e-06,
"loss": 0.2284,
"step": 2725
},
{
"epoch": 1.9455252918287937,
"grad_norm": 4.662227630615234,
"learning_rate": 5.004444444444445e-06,
"loss": 0.2238,
"step": 2750
},
{
"epoch": 1.9632118853908738,
"grad_norm": 3.227084159851074,
"learning_rate": 4.94888888888889e-06,
"loss": 0.2068,
"step": 2775
},
{
"epoch": 1.9808984789529536,
"grad_norm": 3.7078394889831543,
"learning_rate": 4.893333333333334e-06,
"loss": 0.225,
"step": 2800
},
{
"epoch": 1.9985850725150336,
"grad_norm": 3.540781259536743,
"learning_rate": 4.837777777777778e-06,
"loss": 0.2115,
"step": 2825
},
{
"epoch": 2.0162716660771136,
"grad_norm": 2.6315805912017822,
"learning_rate": 4.7822222222222226e-06,
"loss": 0.1164,
"step": 2850
},
{
"epoch": 2.0339582596391934,
"grad_norm": 3.7077784538269043,
"learning_rate": 4.7266666666666674e-06,
"loss": 0.1052,
"step": 2875
},
{
"epoch": 2.0516448532012737,
"grad_norm": 2.5548624992370605,
"learning_rate": 4.6711111111111115e-06,
"loss": 0.1033,
"step": 2900
},
{
"epoch": 2.0693314467633535,
"grad_norm": 3.648667097091675,
"learning_rate": 4.6155555555555555e-06,
"loss": 0.0954,
"step": 2925
},
{
"epoch": 2.0870180403254333,
"grad_norm": 1.9462636709213257,
"learning_rate": 4.56e-06,
"loss": 0.0907,
"step": 2950
},
{
"epoch": 2.104704633887513,
"grad_norm": 2.353156089782715,
"learning_rate": 4.504444444444444e-06,
"loss": 0.0962,
"step": 2975
},
{
"epoch": 2.1223912274495933,
"grad_norm": 1.8443711996078491,
"learning_rate": 4.448888888888889e-06,
"loss": 0.0867,
"step": 3000
},
{
"epoch": 2.1223912274495933,
"eval_loss": 0.25060394406318665,
"eval_runtime": 4656.2877,
"eval_samples_per_second": 2.311,
"eval_steps_per_second": 0.145,
"eval_wer": 0.15110889462567473,
"step": 3000
},
{
"epoch": 2.140077821011673,
"grad_norm": 2.377577781677246,
"learning_rate": 4.393333333333334e-06,
"loss": 0.0825,
"step": 3025
},
{
"epoch": 2.157764414573753,
"grad_norm": 2.523731231689453,
"learning_rate": 4.337777777777778e-06,
"loss": 0.1005,
"step": 3050
},
{
"epoch": 2.175451008135833,
"grad_norm": 3.5590755939483643,
"learning_rate": 4.282222222222222e-06,
"loss": 0.1085,
"step": 3075
},
{
"epoch": 2.193137601697913,
"grad_norm": 2.7134900093078613,
"learning_rate": 4.226666666666667e-06,
"loss": 0.0957,
"step": 3100
},
{
"epoch": 2.2108241952599927,
"grad_norm": 3.198664903640747,
"learning_rate": 4.171111111111111e-06,
"loss": 0.0898,
"step": 3125
},
{
"epoch": 2.228510788822073,
"grad_norm": 2.207460403442383,
"learning_rate": 4.115555555555556e-06,
"loss": 0.0922,
"step": 3150
},
{
"epoch": 2.246197382384153,
"grad_norm": 3.6293723583221436,
"learning_rate": 4.060000000000001e-06,
"loss": 0.0847,
"step": 3175
},
{
"epoch": 2.2638839759462326,
"grad_norm": 2.8405864238739014,
"learning_rate": 4.004444444444445e-06,
"loss": 0.097,
"step": 3200
},
{
"epoch": 2.281570569508313,
"grad_norm": 3.193572521209717,
"learning_rate": 3.948888888888889e-06,
"loss": 0.0997,
"step": 3225
},
{
"epoch": 2.2992571630703926,
"grad_norm": 2.4953181743621826,
"learning_rate": 3.893333333333333e-06,
"loss": 0.0947,
"step": 3250
},
{
"epoch": 2.3169437566324724,
"grad_norm": 2.287153959274292,
"learning_rate": 3.837777777777778e-06,
"loss": 0.0829,
"step": 3275
},
{
"epoch": 2.3346303501945527,
"grad_norm": 3.0835964679718018,
"learning_rate": 3.782222222222223e-06,
"loss": 0.0986,
"step": 3300
},
{
"epoch": 2.3523169437566325,
"grad_norm": 3.700576066970825,
"learning_rate": 3.726666666666667e-06,
"loss": 0.0938,
"step": 3325
},
{
"epoch": 2.3700035373187123,
"grad_norm": 2.685814380645752,
"learning_rate": 3.6711111111111113e-06,
"loss": 0.0912,
"step": 3350
},
{
"epoch": 2.3876901308807925,
"grad_norm": 2.80168080329895,
"learning_rate": 3.615555555555556e-06,
"loss": 0.1046,
"step": 3375
},
{
"epoch": 2.4053767244428723,
"grad_norm": 1.993931531906128,
"learning_rate": 3.5600000000000002e-06,
"loss": 0.0956,
"step": 3400
},
{
"epoch": 2.423063318004952,
"grad_norm": 2.2069132328033447,
"learning_rate": 3.5044444444444447e-06,
"loss": 0.0897,
"step": 3425
},
{
"epoch": 2.4407499115670324,
"grad_norm": 1.9554574489593506,
"learning_rate": 3.4488888888888896e-06,
"loss": 0.0916,
"step": 3450
},
{
"epoch": 2.458436505129112,
"grad_norm": 2.17236590385437,
"learning_rate": 3.3933333333333336e-06,
"loss": 0.0914,
"step": 3475
},
{
"epoch": 2.476123098691192,
"grad_norm": 1.3875476121902466,
"learning_rate": 3.337777777777778e-06,
"loss": 0.0841,
"step": 3500
},
{
"epoch": 2.493809692253272,
"grad_norm": 2.773275852203369,
"learning_rate": 3.282222222222223e-06,
"loss": 0.0776,
"step": 3525
},
{
"epoch": 2.511496285815352,
"grad_norm": 2.6522185802459717,
"learning_rate": 3.226666666666667e-06,
"loss": 0.1017,
"step": 3550
},
{
"epoch": 2.529182879377432,
"grad_norm": 2.5875155925750732,
"learning_rate": 3.1711111111111114e-06,
"loss": 0.0873,
"step": 3575
},
{
"epoch": 2.5468694729395116,
"grad_norm": 2.877279281616211,
"learning_rate": 3.1155555555555555e-06,
"loss": 0.1071,
"step": 3600
},
{
"epoch": 2.564556066501592,
"grad_norm": 2.6924633979797363,
"learning_rate": 3.0600000000000003e-06,
"loss": 0.0885,
"step": 3625
},
{
"epoch": 2.5822426600636716,
"grad_norm": 3.1496784687042236,
"learning_rate": 3.004444444444445e-06,
"loss": 0.0875,
"step": 3650
},
{
"epoch": 2.599929253625752,
"grad_norm": 2.8723342418670654,
"learning_rate": 2.948888888888889e-06,
"loss": 0.0849,
"step": 3675
},
{
"epoch": 2.6176158471878317,
"grad_norm": 2.9840595722198486,
"learning_rate": 2.8933333333333337e-06,
"loss": 0.085,
"step": 3700
},
{
"epoch": 2.6353024407499115,
"grad_norm": 2.006192684173584,
"learning_rate": 2.837777777777778e-06,
"loss": 0.082,
"step": 3725
},
{
"epoch": 2.6529890343119913,
"grad_norm": 2.020374298095703,
"learning_rate": 2.7822222222222222e-06,
"loss": 0.083,
"step": 3750
},
{
"epoch": 2.6706756278740715,
"grad_norm": 2.8968160152435303,
"learning_rate": 2.726666666666667e-06,
"loss": 0.0896,
"step": 3775
},
{
"epoch": 2.6883622214361513,
"grad_norm": 2.3445401191711426,
"learning_rate": 2.6711111111111116e-06,
"loss": 0.0956,
"step": 3800
},
{
"epoch": 2.7060488149982316,
"grad_norm": 2.18906307220459,
"learning_rate": 2.6155555555555556e-06,
"loss": 0.0968,
"step": 3825
},
{
"epoch": 2.7237354085603114,
"grad_norm": 3.4939792156219482,
"learning_rate": 2.56e-06,
"loss": 0.0807,
"step": 3850
},
{
"epoch": 2.741422002122391,
"grad_norm": 2.769930839538574,
"learning_rate": 2.504444444444445e-06,
"loss": 0.0807,
"step": 3875
},
{
"epoch": 2.759108595684471,
"grad_norm": 2.675159215927124,
"learning_rate": 2.448888888888889e-06,
"loss": 0.0852,
"step": 3900
},
{
"epoch": 2.776795189246551,
"grad_norm": 1.9144662618637085,
"learning_rate": 2.3933333333333334e-06,
"loss": 0.0827,
"step": 3925
},
{
"epoch": 2.794481782808631,
"grad_norm": 2.1294009685516357,
"learning_rate": 2.337777777777778e-06,
"loss": 0.0723,
"step": 3950
},
{
"epoch": 2.8121683763707113,
"grad_norm": 2.5433287620544434,
"learning_rate": 2.2822222222222223e-06,
"loss": 0.0895,
"step": 3975
},
{
"epoch": 2.829854969932791,
"grad_norm": 2.70267391204834,
"learning_rate": 2.226666666666667e-06,
"loss": 0.0973,
"step": 4000
},
{
"epoch": 2.829854969932791,
"eval_loss": 0.24439197778701782,
"eval_runtime": 4587.2812,
"eval_samples_per_second": 2.346,
"eval_steps_per_second": 0.147,
"eval_wer": 0.14902605022295237,
"step": 4000
},
{
"epoch": 2.847541563494871,
"grad_norm": 2.82698392868042,
"learning_rate": 2.1711111111111113e-06,
"loss": 0.087,
"step": 4025
},
{
"epoch": 2.8652281570569507,
"grad_norm": 2.4205260276794434,
"learning_rate": 2.1155555555555557e-06,
"loss": 0.073,
"step": 4050
},
{
"epoch": 2.882914750619031,
"grad_norm": 2.8398914337158203,
"learning_rate": 2.06e-06,
"loss": 0.0829,
"step": 4075
},
{
"epoch": 2.9006013441811107,
"grad_norm": 2.544459581375122,
"learning_rate": 2.0044444444444446e-06,
"loss": 0.0944,
"step": 4100
},
{
"epoch": 2.9182879377431905,
"grad_norm": 3.273761034011841,
"learning_rate": 1.948888888888889e-06,
"loss": 0.0894,
"step": 4125
},
{
"epoch": 2.9359745313052708,
"grad_norm": 2.9628796577453613,
"learning_rate": 1.8933333333333333e-06,
"loss": 0.0779,
"step": 4150
},
{
"epoch": 2.9536611248673506,
"grad_norm": 2.375798225402832,
"learning_rate": 1.837777777777778e-06,
"loss": 0.0887,
"step": 4175
},
{
"epoch": 2.9713477184294304,
"grad_norm": 3.536142587661743,
"learning_rate": 1.7822222222222225e-06,
"loss": 0.0802,
"step": 4200
},
{
"epoch": 2.9890343119915106,
"grad_norm": 2.959638833999634,
"learning_rate": 1.7266666666666667e-06,
"loss": 0.0736,
"step": 4225
},
{
"epoch": 3.0067209055535904,
"grad_norm": 1.6278612613677979,
"learning_rate": 1.6711111111111112e-06,
"loss": 0.0726,
"step": 4250
},
{
"epoch": 3.02440749911567,
"grad_norm": 2.070136308670044,
"learning_rate": 1.6155555555555559e-06,
"loss": 0.0342,
"step": 4275
},
{
"epoch": 3.0420940926777504,
"grad_norm": 1.3914729356765747,
"learning_rate": 1.56e-06,
"loss": 0.0358,
"step": 4300
},
{
"epoch": 3.0597806862398302,
"grad_norm": 2.264988422393799,
"learning_rate": 1.5044444444444446e-06,
"loss": 0.0333,
"step": 4325
},
{
"epoch": 3.07746727980191,
"grad_norm": 2.115565538406372,
"learning_rate": 1.4488888888888892e-06,
"loss": 0.0441,
"step": 4350
},
{
"epoch": 3.0951538733639903,
"grad_norm": 1.555413007736206,
"learning_rate": 1.3933333333333335e-06,
"loss": 0.0333,
"step": 4375
},
{
"epoch": 3.11284046692607,
"grad_norm": 1.6975215673446655,
"learning_rate": 1.337777777777778e-06,
"loss": 0.0347,
"step": 4400
},
{
"epoch": 3.13052706048815,
"grad_norm": 1.0206265449523926,
"learning_rate": 1.2822222222222222e-06,
"loss": 0.0288,
"step": 4425
},
{
"epoch": 3.14821365405023,
"grad_norm": 0.8475884199142456,
"learning_rate": 1.2266666666666666e-06,
"loss": 0.0282,
"step": 4450
},
{
"epoch": 3.16590024761231,
"grad_norm": 1.2225841283798218,
"learning_rate": 1.171111111111111e-06,
"loss": 0.0419,
"step": 4475
},
{
"epoch": 3.1835868411743897,
"grad_norm": 1.5350502729415894,
"learning_rate": 1.1155555555555558e-06,
"loss": 0.033,
"step": 4500
},
{
"epoch": 3.2012734347364695,
"grad_norm": 1.7746856212615967,
"learning_rate": 1.06e-06,
"loss": 0.0314,
"step": 4525
},
{
"epoch": 3.2189600282985498,
"grad_norm": 0.9352214932441711,
"learning_rate": 1.0044444444444445e-06,
"loss": 0.0309,
"step": 4550
},
{
"epoch": 3.2366466218606296,
"grad_norm": 1.1960421800613403,
"learning_rate": 9.488888888888889e-07,
"loss": 0.0432,
"step": 4575
},
{
"epoch": 3.25433321542271,
"grad_norm": 2.057264566421509,
"learning_rate": 8.933333333333334e-07,
"loss": 0.0361,
"step": 4600
},
{
"epoch": 3.2720198089847896,
"grad_norm": 1.3137741088867188,
"learning_rate": 8.37777777777778e-07,
"loss": 0.0274,
"step": 4625
},
{
"epoch": 3.2897064025468694,
"grad_norm": 2.9971253871917725,
"learning_rate": 7.822222222222223e-07,
"loss": 0.0334,
"step": 4650
},
{
"epoch": 3.307392996108949,
"grad_norm": 1.1959739923477173,
"learning_rate": 7.266666666666668e-07,
"loss": 0.0321,
"step": 4675
},
{
"epoch": 3.3250795896710295,
"grad_norm": 1.4771987199783325,
"learning_rate": 6.711111111111111e-07,
"loss": 0.0304,
"step": 4700
},
{
"epoch": 3.3427661832331093,
"grad_norm": 1.407965898513794,
"learning_rate": 6.155555555555556e-07,
"loss": 0.0358,
"step": 4725
},
{
"epoch": 3.360452776795189,
"grad_norm": 1.603348970413208,
"learning_rate": 5.6e-07,
"loss": 0.0253,
"step": 4750
},
{
"epoch": 3.3781393703572693,
"grad_norm": 0.6914874911308289,
"learning_rate": 5.044444444444445e-07,
"loss": 0.0314,
"step": 4775
},
{
"epoch": 3.395825963919349,
"grad_norm": 0.9917079210281372,
"learning_rate": 4.488888888888889e-07,
"loss": 0.0366,
"step": 4800
},
{
"epoch": 3.413512557481429,
"grad_norm": 1.6925675868988037,
"learning_rate": 3.9333333333333336e-07,
"loss": 0.0325,
"step": 4825
},
{
"epoch": 3.431199151043509,
"grad_norm": 1.986444115638733,
"learning_rate": 3.3777777777777777e-07,
"loss": 0.0281,
"step": 4850
},
{
"epoch": 3.448885744605589,
"grad_norm": 0.704007089138031,
"learning_rate": 2.822222222222222e-07,
"loss": 0.0313,
"step": 4875
},
{
"epoch": 3.4665723381676687,
"grad_norm": 1.0404950380325317,
"learning_rate": 2.266666666666667e-07,
"loss": 0.0311,
"step": 4900
},
{
"epoch": 3.484258931729749,
"grad_norm": 1.7548632621765137,
"learning_rate": 1.7111111111111114e-07,
"loss": 0.029,
"step": 4925
},
{
"epoch": 3.501945525291829,
"grad_norm": 2.3233606815338135,
"learning_rate": 1.1555555555555556e-07,
"loss": 0.0364,
"step": 4950
},
{
"epoch": 3.5196321188539086,
"grad_norm": 1.7240501642227173,
"learning_rate": 6.000000000000001e-08,
"loss": 0.035,
"step": 4975
},
{
"epoch": 3.537318712415989,
"grad_norm": 0.9116590023040771,
"learning_rate": 4.444444444444445e-09,
"loss": 0.0303,
"step": 5000
},
{
"epoch": 3.537318712415989,
"eval_loss": 0.27435705065727234,
"eval_runtime": 4637.7857,
"eval_samples_per_second": 2.321,
"eval_steps_per_second": 0.145,
"eval_wer": 0.14744191504341703,
"step": 5000
},
{
"epoch": 3.537318712415989,
"step": 5000,
"total_flos": 5.435079965953229e+20,
"train_loss": 0.26693406739234926,
"train_runtime": 74036.7422,
"train_samples_per_second": 2.161,
"train_steps_per_second": 0.068
}
],
"logging_steps": 25,
"max_steps": 5000,
"num_input_tokens_seen": 0,
"num_train_epochs": 4,
"save_steps": 1000,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 5.435079965953229e+20,
"train_batch_size": 16,
"trial_name": null,
"trial_params": null
}