|
{ |
|
"best_metric": 0.9421904445910342, |
|
"best_model_checkpoint": "./models/bart-base-spelling-de/checkpoint-18000", |
|
"epoch": 1.9999130347526508, |
|
"global_step": 18684, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 0.0002998554913294798, |
|
"loss": 8.3693, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 0.00029971098265895954, |
|
"loss": 4.8353, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 0.0002995504174694926, |
|
"loss": 4.027, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 0.00029938985228002564, |
|
"loss": 3.6199, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.00029922928709055877, |
|
"loss": 2.6901, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.0002990687219010918, |
|
"loss": 1.5175, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.00029890815671162487, |
|
"loss": 1.175, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.000298747591522158, |
|
"loss": 0.9833, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.000298587026332691, |
|
"loss": 0.9042, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.0002984264611432241, |
|
"loss": 0.8678, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.0002982658959537572, |
|
"loss": 0.8539, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.00029810533076429026, |
|
"loss": 0.7903, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.0002979447655748234, |
|
"loss": 0.7743, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 0.0002977842003853564, |
|
"loss": 0.737, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0002976236351958895, |
|
"loss": 0.7497, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.00029746307000642257, |
|
"loss": 0.6934, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.00029730250481695565, |
|
"loss": 0.5994, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0002971419396274887, |
|
"loss": 0.6431, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0002969813744380218, |
|
"loss": 0.5749, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0002968368657675016, |
|
"loss": 0.6268, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.00029667630057803463, |
|
"loss": 0.6121, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0002965157353885677, |
|
"loss": 0.5817, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.00029635517019910084, |
|
"loss": 0.6015, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.00029619460500963387, |
|
"loss": 0.511, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.00029603403982016694, |
|
"loss": 0.5886, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0002958734746307, |
|
"loss": 0.6352, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0002957129094412331, |
|
"loss": 0.4925, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0002955523442517662, |
|
"loss": 0.5772, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.00029539177906229926, |
|
"loss": 0.5069, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.00029523121387283233, |
|
"loss": 0.4974, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0002950706486833654, |
|
"loss": 0.512, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0002949100834938985, |
|
"loss": 0.5083, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.00029474951830443157, |
|
"loss": 0.5309, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.00029458895311496464, |
|
"loss": 0.4753, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0002944283879254977, |
|
"loss": 0.4401, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0002942678227360308, |
|
"loss": 0.4986, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0002941072575465639, |
|
"loss": 0.5574, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.00029394669235709695, |
|
"loss": 0.3952, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.00029378612716763003, |
|
"loss": 0.497, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0002936255619781631, |
|
"loss": 0.3726, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0002934649967886962, |
|
"loss": 0.3782, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.00029330443159922927, |
|
"loss": 0.385, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.00029314386640976234, |
|
"loss": 0.4608, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0002929833012202954, |
|
"loss": 0.4143, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0002928227360308285, |
|
"loss": 0.386, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0002926621708413616, |
|
"loss": 0.3939, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.00029250160565189465, |
|
"loss": 0.3939, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.00029234104046242773, |
|
"loss": 0.4287, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0002921804752729608, |
|
"loss": 0.4586, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0002920199100834939, |
|
"loss": 0.3787, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0002918593448940269, |
|
"loss": 0.3833, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.00029169877970456004, |
|
"loss": 0.3546, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0002915382145150931, |
|
"loss": 0.3287, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0002913776493256262, |
|
"loss": 0.3341, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0002912170841361593, |
|
"loss": 0.3813, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0002910565189466923, |
|
"loss": 0.3462, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.00029089595375722543, |
|
"loss": 0.2947, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0002907353885677585, |
|
"loss": 0.3296, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.00029057482337829153, |
|
"loss": 0.3757, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.00029041425818882466, |
|
"loss": 0.3367, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0002902536929993577, |
|
"loss": 0.3251, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.00029009312780989077, |
|
"loss": 0.3764, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0002899325626204239, |
|
"loss": 0.364, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0002897719974309569, |
|
"loss": 0.3465, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.00028961143224149005, |
|
"loss": 0.3404, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0002894508670520231, |
|
"loss": 0.2828, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.00028929030186255615, |
|
"loss": 0.3096, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0002891297366730893, |
|
"loss": 0.3952, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0002889691714836223, |
|
"loss": 0.3129, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0002888086062941554, |
|
"loss": 0.2965, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.00028864804110468847, |
|
"loss": 0.3447, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.00028848747591522154, |
|
"loss": 0.3014, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0002883269107257546, |
|
"loss": 0.3031, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0002881663455362877, |
|
"loss": 0.3191, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0002880057803468208, |
|
"loss": 0.3159, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.00028784521515735385, |
|
"loss": 0.3031, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.00028768464996788693, |
|
"loss": 0.3029, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.00028752408477842, |
|
"loss": 0.2696, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0002873635195889531, |
|
"loss": 0.2914, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.00028720295439948616, |
|
"loss": 0.2821, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.00028704238921001924, |
|
"loss": 0.3199, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0002868818240205523, |
|
"loss": 0.2796, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0002867212588310854, |
|
"loss": 0.2914, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0002865606936416185, |
|
"loss": 0.3057, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.00028640012845215155, |
|
"loss": 0.307, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.00028623956326268463, |
|
"loss": 0.2883, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0002860789980732177, |
|
"loss": 0.3134, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0002859184328837508, |
|
"loss": 0.2998, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.00028575786769428386, |
|
"loss": 0.2761, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.00028559730250481694, |
|
"loss": 0.3234, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.00028543673731535, |
|
"loss": 0.2495, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0002852761721258831, |
|
"loss": 0.2691, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0002851156069364162, |
|
"loss": 0.2864, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.00028495504174694925, |
|
"loss": 0.2611, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.00028479447655748233, |
|
"loss": 0.3546, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.00028463391136801535, |
|
"loss": 0.3073, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0002844733461785485, |
|
"loss": 0.2859, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.00028431278098908156, |
|
"loss": 0.2192, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0002841522157996146, |
|
"loss": 0.2312, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0002839916506101477, |
|
"loss": 0.2803, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"eval_cer": 0.9428737044619099, |
|
"eval_loss": 0.1978314369916916, |
|
"eval_runtime": 52.7441, |
|
"eval_samples_per_second": 18.959, |
|
"eval_steps_per_second": 4.74, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.00028383108542068074, |
|
"loss": 0.2246, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0002836705202312139, |
|
"loss": 0.2674, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.00028350995504174695, |
|
"loss": 0.2588, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.00028334938985228, |
|
"loss": 0.2511, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0002831888246628131, |
|
"loss": 0.2332, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.00028302825947334613, |
|
"loss": 0.2682, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0002828676942838792, |
|
"loss": 0.2483, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.00028270712909441234, |
|
"loss": 0.2539, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.00028254656390494536, |
|
"loss": 0.2558, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.00028238599871547844, |
|
"loss": 0.2499, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0002822254335260115, |
|
"loss": 0.2851, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0002820648683365446, |
|
"loss": 0.2504, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.00028190430314707773, |
|
"loss": 0.2265, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.00028174373795761075, |
|
"loss": 0.2665, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.00028158317276814383, |
|
"loss": 0.2483, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.0002814226075786769, |
|
"loss": 0.2815, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.00028126204238921, |
|
"loss": 0.2572, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.00028110147719974306, |
|
"loss": 0.1847, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.00028094091201027614, |
|
"loss": 0.2144, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0002807803468208092, |
|
"loss": 0.2308, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0002806197816313423, |
|
"loss": 0.2556, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0002804592164418754, |
|
"loss": 0.2499, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.00028029865125240845, |
|
"loss": 0.2218, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.00028013808606294153, |
|
"loss": 0.2662, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0002799775208734746, |
|
"loss": 0.2323, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0002798169556840077, |
|
"loss": 0.2511, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.00027965639049454076, |
|
"loss": 0.1681, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.00027949582530507384, |
|
"loss": 0.2228, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.0002793352601156069, |
|
"loss": 0.2531, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.00027917469492614, |
|
"loss": 0.227, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.0002790141297366731, |
|
"loss": 0.2348, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.00027885356454720615, |
|
"loss": 0.2283, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.00027869299935773923, |
|
"loss": 0.2743, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.0002785324341682723, |
|
"loss": 0.2838, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.0002783718689788054, |
|
"loss": 0.2158, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.00027821130378933846, |
|
"loss": 0.2157, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.00027805073859987154, |
|
"loss": 0.2308, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.0002778901734104046, |
|
"loss": 0.2416, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.0002777296082209377, |
|
"loss": 0.2231, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.0002775690430314708, |
|
"loss": 0.1801, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.0002774084778420038, |
|
"loss": 0.27, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.00027724791265253693, |
|
"loss": 0.2432, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.00027708734746307, |
|
"loss": 0.2709, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.00027692678227360303, |
|
"loss": 0.2033, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.00027676621708413616, |
|
"loss": 0.2271, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.0002766056518946692, |
|
"loss": 0.2154, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.00027644508670520226, |
|
"loss": 0.2068, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.0002762845215157354, |
|
"loss": 0.2613, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.0002761239563262684, |
|
"loss": 0.253, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.00027596339113680155, |
|
"loss": 0.2211, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.0002758028259473346, |
|
"loss": 0.2149, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.00027564226075786765, |
|
"loss": 0.2356, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.0002754816955684008, |
|
"loss": 0.2204, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.0002753211303789338, |
|
"loss": 0.2477, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.0002751605651894669, |
|
"loss": 0.1956, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.00027499999999999996, |
|
"loss": 0.1617, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.00027483943481053304, |
|
"loss": 0.2347, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.0002746788696210661, |
|
"loss": 0.2258, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.0002745183044315992, |
|
"loss": 0.2028, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.0002743577392421323, |
|
"loss": 0.2026, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.00027419717405266535, |
|
"loss": 0.2073, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.00027403660886319843, |
|
"loss": 0.2004, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.0002738760436737315, |
|
"loss": 0.2157, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.0002737154784842646, |
|
"loss": 0.2135, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.00027355491329479766, |
|
"loss": 0.1822, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.00027339434810533074, |
|
"loss": 0.1787, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.0002732337829158638, |
|
"loss": 0.2042, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.0002730732177263969, |
|
"loss": 0.1842, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.00027291265253692997, |
|
"loss": 0.1897, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.00027275208734746305, |
|
"loss": 0.1914, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.00027259152215799613, |
|
"loss": 0.2059, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.0002724309569685292, |
|
"loss": 0.1567, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.0002722703917790623, |
|
"loss": 0.1549, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.00027210982658959536, |
|
"loss": 0.1554, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.00027194926140012844, |
|
"loss": 0.1744, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.0002717886962106615, |
|
"loss": 0.2383, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.0002716281310211946, |
|
"loss": 0.1904, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.00027146756583172767, |
|
"loss": 0.2276, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.00027130700064226075, |
|
"loss": 0.203, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.00027114643545279383, |
|
"loss": 0.1782, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.00027098587026332685, |
|
"loss": 0.1987, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.00027082530507386, |
|
"loss": 0.1785, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.00027066473988439306, |
|
"loss": 0.2123, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.00027050417469492614, |
|
"loss": 0.2317, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.0002703436095054592, |
|
"loss": 0.2114, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.00027018304431599224, |
|
"loss": 0.2009, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.00027002247912652537, |
|
"loss": 0.1983, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.00026986191393705845, |
|
"loss": 0.2075, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.00026970134874759147, |
|
"loss": 0.1857, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.0002695407835581246, |
|
"loss": 0.1921, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.00026938021836865763, |
|
"loss": 0.2285, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.0002692196531791907, |
|
"loss": 0.2013, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.00026905908798972384, |
|
"loss": 0.2184, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.00026889852280025686, |
|
"loss": 0.2112, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.00026873795761079, |
|
"loss": 0.1988, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.000268577392421323, |
|
"loss": 0.1834, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.0002684168272318561, |
|
"loss": 0.1732, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.0002682562620423892, |
|
"loss": 0.1791, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.00026809569685292225, |
|
"loss": 0.2003, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.00026793513166345533, |
|
"loss": 0.1688, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"eval_cer": 0.9426153509214829, |
|
"eval_loss": 0.147249236702919, |
|
"eval_runtime": 53.7363, |
|
"eval_samples_per_second": 18.609, |
|
"eval_steps_per_second": 4.652, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.0002677745664739884, |
|
"loss": 0.1907, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.0002676140012845215, |
|
"loss": 0.2018, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.00026745343609505456, |
|
"loss": 0.1942, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.00026729287090558764, |
|
"loss": 0.205, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.0002671323057161207, |
|
"loss": 0.1858, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.0002669717405266538, |
|
"loss": 0.2086, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.00026681117533718687, |
|
"loss": 0.201, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.00026665061014771995, |
|
"loss": 0.1849, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.000266490044958253, |
|
"loss": 0.2413, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.0002663294797687861, |
|
"loss": 0.1823, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.0002661689145793192, |
|
"loss": 0.1699, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.00026600834938985226, |
|
"loss": 0.1915, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.00026584778420038534, |
|
"loss": 0.1616, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.0002656872190109184, |
|
"loss": 0.194, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.0002655266538214515, |
|
"loss": 0.1549, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.00026536608863198457, |
|
"loss": 0.1523, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.00026520552344251765, |
|
"loss": 0.2071, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.00026504495825305067, |
|
"loss": 0.1927, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.0002648843930635838, |
|
"loss": 0.1769, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.0002647238278741169, |
|
"loss": 0.1741, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.00026456326268464996, |
|
"loss": 0.1845, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.00026440269749518304, |
|
"loss": 0.1531, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.00026424213230571606, |
|
"loss": 0.1655, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.0002640815671162492, |
|
"loss": 0.185, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.00026392100192678227, |
|
"loss": 0.1772, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.0002637604367373153, |
|
"loss": 0.1772, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.0002635998715478484, |
|
"loss": 0.1714, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.00026343930635838145, |
|
"loss": 0.2066, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.0002632787411689145, |
|
"loss": 0.1741, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.00026311817597944766, |
|
"loss": 0.1804, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.0002629576107899807, |
|
"loss": 0.1985, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.0002627970456005138, |
|
"loss": 0.1781, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.00026263648041104684, |
|
"loss": 0.1784, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.0002624759152215799, |
|
"loss": 0.1783, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.00026231535003211305, |
|
"loss": 0.1675, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.00026215478484264607, |
|
"loss": 0.1788, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.00026199421965317915, |
|
"loss": 0.1897, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.0002618336544637122, |
|
"loss": 0.2146, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.0002616730892742453, |
|
"loss": 0.1779, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.0002615125240847784, |
|
"loss": 0.1666, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.00026135195889531146, |
|
"loss": 0.1735, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.00026119139370584454, |
|
"loss": 0.1775, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.0002610308285163776, |
|
"loss": 0.1672, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.0002608702633269107, |
|
"loss": 0.1636, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.00026070969813744377, |
|
"loss": 0.1952, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.00026054913294797685, |
|
"loss": 0.1802, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.0002603885677585099, |
|
"loss": 0.1692, |
|
"step": 2470 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.000260228002569043, |
|
"loss": 0.1433, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.0002600674373795761, |
|
"loss": 0.2053, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.00025990687219010916, |
|
"loss": 0.1899, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.00025974630700064224, |
|
"loss": 0.1827, |
|
"step": 2510 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.0002595857418111753, |
|
"loss": 0.1473, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.0002594251766217084, |
|
"loss": 0.1792, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.00025926461143224147, |
|
"loss": 0.1671, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.00025910404624277455, |
|
"loss": 0.1911, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.0002589434810533076, |
|
"loss": 0.1745, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.0002587829158638407, |
|
"loss": 0.161, |
|
"step": 2570 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.0002586223506743738, |
|
"loss": 0.1889, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.00025846178548490686, |
|
"loss": 0.1542, |
|
"step": 2590 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.00025830122029543994, |
|
"loss": 0.1865, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.000258140655105973, |
|
"loss": 0.1739, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.0002579800899165061, |
|
"loss": 0.1845, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.0002578195247270391, |
|
"loss": 0.1754, |
|
"step": 2630 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.00025765895953757225, |
|
"loss": 0.191, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.0002574983943481053, |
|
"loss": 0.1645, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.0002573378291586384, |
|
"loss": 0.1544, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.0002571772639691715, |
|
"loss": 0.1903, |
|
"step": 2670 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.0002570166987797045, |
|
"loss": 0.1528, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.00025685613359023764, |
|
"loss": 0.1666, |
|
"step": 2690 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.0002566955684007707, |
|
"loss": 0.1708, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.00025653500321130374, |
|
"loss": 0.1969, |
|
"step": 2710 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.00025637443802183687, |
|
"loss": 0.1639, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.0002562138728323699, |
|
"loss": 0.1905, |
|
"step": 2730 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.00025605330764290297, |
|
"loss": 0.1655, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.0002558927424534361, |
|
"loss": 0.1928, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.0002557321772639691, |
|
"loss": 0.1673, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.0002555716120745022, |
|
"loss": 0.1302, |
|
"step": 2770 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.0002554110468850353, |
|
"loss": 0.2082, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.00025525048169556836, |
|
"loss": 0.1795, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.0002550899165061015, |
|
"loss": 0.1481, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.0002549293513166345, |
|
"loss": 0.1571, |
|
"step": 2810 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.0002547687861271676, |
|
"loss": 0.1523, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.00025460822093770067, |
|
"loss": 0.1813, |
|
"step": 2830 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.00025444765574823375, |
|
"loss": 0.1911, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.0002542870905587668, |
|
"loss": 0.2255, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.0002541265253692999, |
|
"loss": 0.1707, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.000253965960179833, |
|
"loss": 0.1888, |
|
"step": 2870 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.00025380539499036606, |
|
"loss": 0.1745, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.00025364482980089914, |
|
"loss": 0.1476, |
|
"step": 2890 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.0002534842646114322, |
|
"loss": 0.1797, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.0002533236994219653, |
|
"loss": 0.1723, |
|
"step": 2910 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.00025316313423249837, |
|
"loss": 0.1552, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.00025300256904303145, |
|
"loss": 0.1609, |
|
"step": 2930 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.0002528420038535645, |
|
"loss": 0.1942, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.0002526814386640976, |
|
"loss": 0.1649, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.0002525208734746307, |
|
"loss": 0.1598, |
|
"step": 2960 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.00025236030828516376, |
|
"loss": 0.1635, |
|
"step": 2970 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.00025219974309569684, |
|
"loss": 0.1749, |
|
"step": 2980 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.0002520391779062299, |
|
"loss": 0.1531, |
|
"step": 2990 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.000251878612716763, |
|
"loss": 0.121, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"eval_cer": 0.9424291265420887, |
|
"eval_loss": 0.13810260593891144, |
|
"eval_runtime": 53.6672, |
|
"eval_samples_per_second": 18.633, |
|
"eval_steps_per_second": 4.658, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.00025171804752729607, |
|
"loss": 0.1831, |
|
"step": 3010 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.00025155748233782915, |
|
"loss": 0.1663, |
|
"step": 3020 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.0002513969171483622, |
|
"loss": 0.183, |
|
"step": 3030 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.0002512363519588953, |
|
"loss": 0.1515, |
|
"step": 3040 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.0002510757867694284, |
|
"loss": 0.1747, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.00025091522157996146, |
|
"loss": 0.1829, |
|
"step": 3060 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.00025075465639049453, |
|
"loss": 0.1861, |
|
"step": 3070 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.00025059409120102756, |
|
"loss": 0.1474, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.0002504335260115607, |
|
"loss": 0.1613, |
|
"step": 3090 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.00025027296082209377, |
|
"loss": 0.1622, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.0002501123956326268, |
|
"loss": 0.1466, |
|
"step": 3110 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.0002499518304431599, |
|
"loss": 0.1832, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.00024979126525369295, |
|
"loss": 0.1765, |
|
"step": 3130 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.0002496307000642261, |
|
"loss": 0.1741, |
|
"step": 3140 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.00024947013487475916, |
|
"loss": 0.1545, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.0002493095696852922, |
|
"loss": 0.1671, |
|
"step": 3160 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.0002491490044958253, |
|
"loss": 0.1658, |
|
"step": 3170 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.00024898843930635834, |
|
"loss": 0.176, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.0002488278741168914, |
|
"loss": 0.1688, |
|
"step": 3190 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.00024866730892742454, |
|
"loss": 0.1531, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.00024850674373795757, |
|
"loss": 0.1569, |
|
"step": 3210 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.00024834617854849065, |
|
"loss": 0.1483, |
|
"step": 3220 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.0002481856133590237, |
|
"loss": 0.1775, |
|
"step": 3230 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.0002480250481695568, |
|
"loss": 0.1851, |
|
"step": 3240 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.00024786448298008993, |
|
"loss": 0.1765, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.00024770391779062296, |
|
"loss": 0.1816, |
|
"step": 3260 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.00024754335260115603, |
|
"loss": 0.2152, |
|
"step": 3270 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.0002473827874116891, |
|
"loss": 0.2041, |
|
"step": 3280 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.0002472222222222222, |
|
"loss": 0.1415, |
|
"step": 3290 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.00024706165703275527, |
|
"loss": 0.1561, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.00024690109184328835, |
|
"loss": 0.1888, |
|
"step": 3310 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.0002467405266538214, |
|
"loss": 0.1644, |
|
"step": 3320 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.0002465799614643545, |
|
"loss": 0.1583, |
|
"step": 3330 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.0002464193962748876, |
|
"loss": 0.1836, |
|
"step": 3340 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.00024625883108542066, |
|
"loss": 0.1785, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.00024609826589595373, |
|
"loss": 0.1934, |
|
"step": 3360 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.0002459377007064868, |
|
"loss": 0.1718, |
|
"step": 3370 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.0002457771355170199, |
|
"loss": 0.1738, |
|
"step": 3380 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.00024561657032755297, |
|
"loss": 0.1534, |
|
"step": 3390 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.00024545600513808604, |
|
"loss": 0.1295, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.0002452954399486191, |
|
"loss": 0.1433, |
|
"step": 3410 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.0002451348747591522, |
|
"loss": 0.1647, |
|
"step": 3420 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.0002449743095696853, |
|
"loss": 0.1722, |
|
"step": 3430 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.00024481374438021836, |
|
"loss": 0.1237, |
|
"step": 3440 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.00024465317919075143, |
|
"loss": 0.1512, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.0002444926140012845, |
|
"loss": 0.2245, |
|
"step": 3460 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.0002443320488118176, |
|
"loss": 0.1603, |
|
"step": 3470 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.0002441714836223506, |
|
"loss": 0.1443, |
|
"step": 3480 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.00024401091843288372, |
|
"loss": 0.1757, |
|
"step": 3490 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.00024385035324341682, |
|
"loss": 0.1794, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.00024368978805394987, |
|
"loss": 0.1682, |
|
"step": 3510 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.00024352922286448298, |
|
"loss": 0.165, |
|
"step": 3520 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.00024336865767501603, |
|
"loss": 0.2031, |
|
"step": 3530 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.0002432080924855491, |
|
"loss": 0.169, |
|
"step": 3540 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.0002430475272960822, |
|
"loss": 0.1519, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.00024288696210661526, |
|
"loss": 0.2043, |
|
"step": 3560 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.00024272639691714834, |
|
"loss": 0.1556, |
|
"step": 3570 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.00024256583172768142, |
|
"loss": 0.1662, |
|
"step": 3580 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.0002424052665382145, |
|
"loss": 0.1443, |
|
"step": 3590 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.00024224470134874757, |
|
"loss": 0.1618, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.00024208413615928065, |
|
"loss": 0.1326, |
|
"step": 3610 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.00024192357096981373, |
|
"loss": 0.1985, |
|
"step": 3620 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.00024176300578034678, |
|
"loss": 0.1719, |
|
"step": 3630 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.00024160244059087988, |
|
"loss": 0.1616, |
|
"step": 3640 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.00024144187540141296, |
|
"loss": 0.1711, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.000241281310211946, |
|
"loss": 0.1705, |
|
"step": 3660 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.00024112074502247912, |
|
"loss": 0.1791, |
|
"step": 3670 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.00024096017983301217, |
|
"loss": 0.1715, |
|
"step": 3680 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.00024079961464354524, |
|
"loss": 0.1525, |
|
"step": 3690 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00024063904945407835, |
|
"loss": 0.1449, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.0002404784842646114, |
|
"loss": 0.1442, |
|
"step": 3710 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.0002403179190751445, |
|
"loss": 0.1498, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00024015735388567756, |
|
"loss": 0.1461, |
|
"step": 3730 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00023999678869621063, |
|
"loss": 0.1601, |
|
"step": 3740 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00023983622350674374, |
|
"loss": 0.1675, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.0002396756583172768, |
|
"loss": 0.1371, |
|
"step": 3760 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00023951509312780987, |
|
"loss": 0.1788, |
|
"step": 3770 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00023935452793834294, |
|
"loss": 0.1304, |
|
"step": 3780 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.00023919396274887602, |
|
"loss": 0.189, |
|
"step": 3790 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.0002390333975594091, |
|
"loss": 0.1458, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.00023887283236994218, |
|
"loss": 0.1938, |
|
"step": 3810 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.00023871226718047525, |
|
"loss": 0.2327, |
|
"step": 3820 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.0002385517019910083, |
|
"loss": 0.2175, |
|
"step": 3830 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.0002383911368015414, |
|
"loss": 0.1655, |
|
"step": 3840 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.0002382305716120745, |
|
"loss": 0.1814, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.00023807000642260754, |
|
"loss": 0.1633, |
|
"step": 3860 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.00023790944123314064, |
|
"loss": 0.1671, |
|
"step": 3870 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.0002377488760436737, |
|
"loss": 0.164, |
|
"step": 3880 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.0002375883108542068, |
|
"loss": 0.1661, |
|
"step": 3890 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.00023742774566473988, |
|
"loss": 0.1363, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.00023726718047527293, |
|
"loss": 0.1362, |
|
"step": 3910 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.00023710661528580603, |
|
"loss": 0.1939, |
|
"step": 3920 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.00023694605009633908, |
|
"loss": 0.1739, |
|
"step": 3930 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.00023678548490687216, |
|
"loss": 0.1232, |
|
"step": 3940 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.00023662491971740526, |
|
"loss": 0.1703, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.00023646435452793832, |
|
"loss": 0.1683, |
|
"step": 3960 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.0002363037893384714, |
|
"loss": 0.1775, |
|
"step": 3970 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00023614322414900447, |
|
"loss": 0.1596, |
|
"step": 3980 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00023598265895953755, |
|
"loss": 0.156, |
|
"step": 3990 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00023582209377007065, |
|
"loss": 0.1722, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"eval_cer": 0.9424999442638301, |
|
"eval_loss": 0.1339801847934723, |
|
"eval_runtime": 52.8397, |
|
"eval_samples_per_second": 18.925, |
|
"eval_steps_per_second": 4.731, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.0002356615285806037, |
|
"loss": 0.1369, |
|
"step": 4010 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00023550096339113678, |
|
"loss": 0.1532, |
|
"step": 4020 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00023534039820166983, |
|
"loss": 0.1458, |
|
"step": 4030 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00023517983301220294, |
|
"loss": 0.1717, |
|
"step": 4040 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00023501926782273601, |
|
"loss": 0.1698, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.0002348587026332691, |
|
"loss": 0.1457, |
|
"step": 4060 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.00023469813744380217, |
|
"loss": 0.1675, |
|
"step": 4070 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.00023453757225433522, |
|
"loss": 0.1748, |
|
"step": 4080 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.00023437700706486833, |
|
"loss": 0.1657, |
|
"step": 4090 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.0002342164418754014, |
|
"loss": 0.1475, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.00023405587668593445, |
|
"loss": 0.1407, |
|
"step": 4110 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.00023389531149646756, |
|
"loss": 0.1585, |
|
"step": 4120 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.0002337347463070006, |
|
"loss": 0.1542, |
|
"step": 4130 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.0002335741811175337, |
|
"loss": 0.1385, |
|
"step": 4140 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.0002334136159280668, |
|
"loss": 0.1802, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00023325305073859984, |
|
"loss": 0.1727, |
|
"step": 4160 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00023309248554913295, |
|
"loss": 0.1343, |
|
"step": 4170 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.000232931920359666, |
|
"loss": 0.1301, |
|
"step": 4180 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00023277135517019908, |
|
"loss": 0.1505, |
|
"step": 4190 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00023261078998073218, |
|
"loss": 0.1232, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00023245022479126523, |
|
"loss": 0.1351, |
|
"step": 4210 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.0002322896596017983, |
|
"loss": 0.1803, |
|
"step": 4220 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.0002321290944123314, |
|
"loss": 0.1614, |
|
"step": 4230 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00023196852922286446, |
|
"loss": 0.1691, |
|
"step": 4240 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00023180796403339754, |
|
"loss": 0.1715, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00023164739884393062, |
|
"loss": 0.1416, |
|
"step": 4260 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.0002314868336544637, |
|
"loss": 0.1417, |
|
"step": 4270 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00023132626846499675, |
|
"loss": 0.1548, |
|
"step": 4280 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00023116570327552985, |
|
"loss": 0.1952, |
|
"step": 4290 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00023100513808606293, |
|
"loss": 0.1779, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00023084457289659598, |
|
"loss": 0.1775, |
|
"step": 4310 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00023068400770712909, |
|
"loss": 0.1607, |
|
"step": 4320 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00023052344251766214, |
|
"loss": 0.1897, |
|
"step": 4330 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00023036287732819521, |
|
"loss": 0.1861, |
|
"step": 4340 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00023020231213872832, |
|
"loss": 0.1784, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00023004174694926137, |
|
"loss": 0.1791, |
|
"step": 4360 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00022988118175979447, |
|
"loss": 0.1552, |
|
"step": 4370 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00022972061657032753, |
|
"loss": 0.2005, |
|
"step": 4380 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.0002295600513808606, |
|
"loss": 0.1956, |
|
"step": 4390 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.0002293994861913937, |
|
"loss": 0.1321, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00022923892100192676, |
|
"loss": 0.1648, |
|
"step": 4410 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00022907835581245984, |
|
"loss": 0.1609, |
|
"step": 4420 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00022891779062299291, |
|
"loss": 0.1805, |
|
"step": 4430 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.000228757225433526, |
|
"loss": 0.1755, |
|
"step": 4440 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00022859666024405907, |
|
"loss": 0.1424, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00022843609505459215, |
|
"loss": 0.1874, |
|
"step": 4460 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00022827552986512522, |
|
"loss": 0.2036, |
|
"step": 4470 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00022811496467565828, |
|
"loss": 0.1698, |
|
"step": 4480 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00022795439948619138, |
|
"loss": 0.1693, |
|
"step": 4490 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00022779383429672446, |
|
"loss": 0.1396, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.0002276332691072575, |
|
"loss": 0.1788, |
|
"step": 4510 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.0002274727039177906, |
|
"loss": 0.1581, |
|
"step": 4520 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00022731213872832366, |
|
"loss": 0.1442, |
|
"step": 4530 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00022715157353885677, |
|
"loss": 0.1766, |
|
"step": 4540 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00022699100834938985, |
|
"loss": 0.1725, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.0002268304431599229, |
|
"loss": 0.1218, |
|
"step": 4560 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.000226669877970456, |
|
"loss": 0.1336, |
|
"step": 4570 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00022650931278098905, |
|
"loss": 0.1574, |
|
"step": 4580 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00022634874759152213, |
|
"loss": 0.1605, |
|
"step": 4590 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00022618818240205524, |
|
"loss": 0.1647, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00022602761721258829, |
|
"loss": 0.1608, |
|
"step": 4610 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00022586705202312136, |
|
"loss": 0.177, |
|
"step": 4620 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00022570648683365444, |
|
"loss": 0.1431, |
|
"step": 4630 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00022554592164418752, |
|
"loss": 0.1388, |
|
"step": 4640 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00022538535645472062, |
|
"loss": 0.1406, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00022522479126525367, |
|
"loss": 0.1388, |
|
"step": 4660 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00022506422607578675, |
|
"loss": 0.1593, |
|
"step": 4670 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.0002249036608863198, |
|
"loss": 0.1529, |
|
"step": 4680 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.0002247430956968529, |
|
"loss": 0.1463, |
|
"step": 4690 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00022458253050738599, |
|
"loss": 0.1496, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00022442196531791906, |
|
"loss": 0.1474, |
|
"step": 4710 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00022426140012845214, |
|
"loss": 0.1668, |
|
"step": 4720 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.0002241008349389852, |
|
"loss": 0.155, |
|
"step": 4730 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.0002239402697495183, |
|
"loss": 0.1617, |
|
"step": 4740 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00022377970456005137, |
|
"loss": 0.1526, |
|
"step": 4750 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00022361913937058442, |
|
"loss": 0.2071, |
|
"step": 4760 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00022345857418111753, |
|
"loss": 0.1261, |
|
"step": 4770 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00022329800899165058, |
|
"loss": 0.1576, |
|
"step": 4780 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00022313744380218366, |
|
"loss": 0.1692, |
|
"step": 4790 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00022297687861271676, |
|
"loss": 0.1289, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.0002228163134232498, |
|
"loss": 0.1628, |
|
"step": 4810 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.00022265574823378292, |
|
"loss": 0.162, |
|
"step": 4820 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.00022249518304431597, |
|
"loss": 0.1491, |
|
"step": 4830 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.00022233461785484905, |
|
"loss": 0.1744, |
|
"step": 4840 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.00022217405266538215, |
|
"loss": 0.1526, |
|
"step": 4850 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.0002220134874759152, |
|
"loss": 0.1496, |
|
"step": 4860 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.00022185292228644828, |
|
"loss": 0.1635, |
|
"step": 4870 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.00022169235709698136, |
|
"loss": 0.1352, |
|
"step": 4880 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.00022154784842646114, |
|
"loss": 0.1767, |
|
"step": 4890 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 0.0002213872832369942, |
|
"loss": 0.1683, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.00022122671804752727, |
|
"loss": 0.1392, |
|
"step": 4910 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.00022106615285806037, |
|
"loss": 0.1827, |
|
"step": 4920 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.00022090558766859342, |
|
"loss": 0.1415, |
|
"step": 4930 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.0002207450224791265, |
|
"loss": 0.154, |
|
"step": 4940 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.00022058445728965958, |
|
"loss": 0.143, |
|
"step": 4950 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.00022042389210019265, |
|
"loss": 0.121, |
|
"step": 4960 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.00022026332691072576, |
|
"loss": 0.1282, |
|
"step": 4970 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.0002201027617212588, |
|
"loss": 0.1612, |
|
"step": 4980 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 0.0002199421965317919, |
|
"loss": 0.1423, |
|
"step": 4990 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.00021978163134232494, |
|
"loss": 0.1502, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"eval_cer": 0.9423465058667236, |
|
"eval_loss": 0.12921704351902008, |
|
"eval_runtime": 53.2479, |
|
"eval_samples_per_second": 18.78, |
|
"eval_steps_per_second": 4.695, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.00021962106615285804, |
|
"loss": 0.1619, |
|
"step": 5010 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.00021946050096339112, |
|
"loss": 0.1455, |
|
"step": 5020 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.00021929993577392417, |
|
"loss": 0.155, |
|
"step": 5030 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.00021913937058445728, |
|
"loss": 0.1748, |
|
"step": 5040 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.00021897880539499033, |
|
"loss": 0.1428, |
|
"step": 5050 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.00021881824020552343, |
|
"loss": 0.1424, |
|
"step": 5060 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.0002186576750160565, |
|
"loss": 0.186, |
|
"step": 5070 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.00021849710982658956, |
|
"loss": 0.1458, |
|
"step": 5080 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 0.00021833654463712266, |
|
"loss": 0.1358, |
|
"step": 5090 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.00021817597944765572, |
|
"loss": 0.1508, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.0002180154142581888, |
|
"loss": 0.1655, |
|
"step": 5110 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.0002178548490687219, |
|
"loss": 0.1325, |
|
"step": 5120 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.00021769428387925495, |
|
"loss": 0.1791, |
|
"step": 5130 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.00021753371868978803, |
|
"loss": 0.1608, |
|
"step": 5140 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.0002173731535003211, |
|
"loss": 0.1616, |
|
"step": 5150 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.00021721258831085418, |
|
"loss": 0.1388, |
|
"step": 5160 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.0002170520231213873, |
|
"loss": 0.1316, |
|
"step": 5170 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 0.00021689145793192034, |
|
"loss": 0.1581, |
|
"step": 5180 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 0.00021673089274245341, |
|
"loss": 0.1688, |
|
"step": 5190 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 0.00021657032755298647, |
|
"loss": 0.1542, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 0.00021640976236351957, |
|
"loss": 0.1574, |
|
"step": 5210 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 0.00021624919717405265, |
|
"loss": 0.1663, |
|
"step": 5220 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 0.00021608863198458573, |
|
"loss": 0.1376, |
|
"step": 5230 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 0.0002159280667951188, |
|
"loss": 0.1456, |
|
"step": 5240 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 0.00021576750160565185, |
|
"loss": 0.1449, |
|
"step": 5250 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 0.00021560693641618496, |
|
"loss": 0.1313, |
|
"step": 5260 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 0.00021544637122671804, |
|
"loss": 0.1325, |
|
"step": 5270 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 0.0002152858060372511, |
|
"loss": 0.1491, |
|
"step": 5280 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 0.0002151252408477842, |
|
"loss": 0.1208, |
|
"step": 5290 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 0.00021496467565831724, |
|
"loss": 0.1444, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 0.00021480411046885032, |
|
"loss": 0.1456, |
|
"step": 5310 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 0.00021464354527938343, |
|
"loss": 0.1649, |
|
"step": 5320 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 0.00021448298008991648, |
|
"loss": 0.1451, |
|
"step": 5330 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 0.00021432241490044958, |
|
"loss": 0.174, |
|
"step": 5340 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 0.00021416184971098263, |
|
"loss": 0.1651, |
|
"step": 5350 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 0.0002140012845215157, |
|
"loss": 0.16, |
|
"step": 5360 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 0.00021384071933204881, |
|
"loss": 0.1185, |
|
"step": 5370 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 0.00021368015414258186, |
|
"loss": 0.1614, |
|
"step": 5380 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 0.00021351958895311494, |
|
"loss": 0.1762, |
|
"step": 5390 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 0.00021335902376364802, |
|
"loss": 0.1481, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 0.0002131984585741811, |
|
"loss": 0.1495, |
|
"step": 5410 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 0.00021303789338471418, |
|
"loss": 0.14, |
|
"step": 5420 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 0.00021287732819524725, |
|
"loss": 0.1631, |
|
"step": 5430 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 0.00021271676300578033, |
|
"loss": 0.1658, |
|
"step": 5440 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 0.00021255619781631338, |
|
"loss": 0.1269, |
|
"step": 5450 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 0.00021239563262684649, |
|
"loss": 0.1542, |
|
"step": 5460 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 0.00021223506743737956, |
|
"loss": 0.1406, |
|
"step": 5470 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 0.00021207450224791261, |
|
"loss": 0.2154, |
|
"step": 5480 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 0.00021192999357739242, |
|
"loss": 0.1445, |
|
"step": 5490 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 0.00021176942838792547, |
|
"loss": 0.1515, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 0.00021160886319845855, |
|
"loss": 0.1632, |
|
"step": 5510 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 0.00021144829800899166, |
|
"loss": 0.1495, |
|
"step": 5520 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 0.0002112877328195247, |
|
"loss": 0.1308, |
|
"step": 5530 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 0.00021112716763005778, |
|
"loss": 0.1686, |
|
"step": 5540 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 0.00021096660244059086, |
|
"loss": 0.1743, |
|
"step": 5550 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 0.00021080603725112394, |
|
"loss": 0.16, |
|
"step": 5560 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 0.00021064547206165702, |
|
"loss": 0.1191, |
|
"step": 5570 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 0.0002104849068721901, |
|
"loss": 0.1422, |
|
"step": 5580 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 0.00021032434168272317, |
|
"loss": 0.1385, |
|
"step": 5590 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 0.00021016377649325622, |
|
"loss": 0.1342, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 0.00021000321130378933, |
|
"loss": 0.1378, |
|
"step": 5610 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 0.0002098426461143224, |
|
"loss": 0.1868, |
|
"step": 5620 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 0.00020968208092485546, |
|
"loss": 0.1592, |
|
"step": 5630 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 0.00020952151573538856, |
|
"loss": 0.166, |
|
"step": 5640 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 0.0002093609505459216, |
|
"loss": 0.1306, |
|
"step": 5650 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 0.00020920038535645472, |
|
"loss": 0.1608, |
|
"step": 5660 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 0.0002090398201669878, |
|
"loss": 0.1562, |
|
"step": 5670 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 0.00020887925497752084, |
|
"loss": 0.1522, |
|
"step": 5680 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 0.00020871868978805395, |
|
"loss": 0.1804, |
|
"step": 5690 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 0.000208558124598587, |
|
"loss": 0.1383, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 0.00020839755940912008, |
|
"loss": 0.1752, |
|
"step": 5710 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 0.00020823699421965318, |
|
"loss": 0.1389, |
|
"step": 5720 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 0.00020807642903018623, |
|
"loss": 0.1528, |
|
"step": 5730 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 0.0002079158638407193, |
|
"loss": 0.1363, |
|
"step": 5740 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 0.0002077552986512524, |
|
"loss": 0.1268, |
|
"step": 5750 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 0.00020759473346178547, |
|
"loss": 0.1443, |
|
"step": 5760 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 0.00020743416827231854, |
|
"loss": 0.1597, |
|
"step": 5770 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 0.00020727360308285162, |
|
"loss": 0.1609, |
|
"step": 5780 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 0.0002071130378933847, |
|
"loss": 0.1613, |
|
"step": 5790 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 0.00020695247270391775, |
|
"loss": 0.1564, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 0.00020679190751445085, |
|
"loss": 0.1935, |
|
"step": 5810 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 0.00020663134232498393, |
|
"loss": 0.1287, |
|
"step": 5820 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 0.00020647077713551698, |
|
"loss": 0.1108, |
|
"step": 5830 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 0.0002063102119460501, |
|
"loss": 0.1607, |
|
"step": 5840 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 0.00020614964675658314, |
|
"loss": 0.1356, |
|
"step": 5850 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 0.00020598908156711624, |
|
"loss": 0.1507, |
|
"step": 5860 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 0.00020582851637764932, |
|
"loss": 0.1322, |
|
"step": 5870 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 0.00020566795118818237, |
|
"loss": 0.1478, |
|
"step": 5880 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 0.00020550738599871548, |
|
"loss": 0.1293, |
|
"step": 5890 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 0.00020534682080924853, |
|
"loss": 0.1625, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 0.0002051862556197816, |
|
"loss": 0.1499, |
|
"step": 5910 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 0.0002050256904303147, |
|
"loss": 0.1519, |
|
"step": 5920 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 0.00020486512524084776, |
|
"loss": 0.131, |
|
"step": 5930 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 0.00020470456005138084, |
|
"loss": 0.1498, |
|
"step": 5940 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 0.00020454399486191392, |
|
"loss": 0.1338, |
|
"step": 5950 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 0.000204383429672447, |
|
"loss": 0.1398, |
|
"step": 5960 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 0.0002042228644829801, |
|
"loss": 0.1545, |
|
"step": 5970 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 0.00020406229929351315, |
|
"loss": 0.1487, |
|
"step": 5980 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 0.00020390173410404623, |
|
"loss": 0.1431, |
|
"step": 5990 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 0.00020374116891457928, |
|
"loss": 0.1556, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"eval_cer": 0.9424238807849227, |
|
"eval_loss": 0.12604805827140808, |
|
"eval_runtime": 52.638, |
|
"eval_samples_per_second": 18.998, |
|
"eval_steps_per_second": 4.749, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 0.00020358060372511238, |
|
"loss": 0.1882, |
|
"step": 6010 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 0.00020342003853564546, |
|
"loss": 0.163, |
|
"step": 6020 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 0.00020325947334617854, |
|
"loss": 0.148, |
|
"step": 6030 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 0.00020309890815671162, |
|
"loss": 0.1328, |
|
"step": 6040 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 0.00020293834296724467, |
|
"loss": 0.1801, |
|
"step": 6050 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 0.00020277777777777777, |
|
"loss": 0.1564, |
|
"step": 6060 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 0.00020261721258831085, |
|
"loss": 0.1529, |
|
"step": 6070 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 0.0002024566473988439, |
|
"loss": 0.1628, |
|
"step": 6080 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 0.000202296082209377, |
|
"loss": 0.1282, |
|
"step": 6090 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 0.00020213551701991005, |
|
"loss": 0.1674, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 0.00020197495183044313, |
|
"loss": 0.1454, |
|
"step": 6110 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 0.00020181438664097624, |
|
"loss": 0.1427, |
|
"step": 6120 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 0.0002016538214515093, |
|
"loss": 0.1545, |
|
"step": 6130 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 0.0002014932562620424, |
|
"loss": 0.1811, |
|
"step": 6140 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 0.00020133269107257544, |
|
"loss": 0.1572, |
|
"step": 6150 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 0.00020117212588310852, |
|
"loss": 0.1402, |
|
"step": 6160 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 0.00020101156069364163, |
|
"loss": 0.1519, |
|
"step": 6170 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 0.00020085099550417468, |
|
"loss": 0.1157, |
|
"step": 6180 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 0.00020069043031470775, |
|
"loss": 0.1736, |
|
"step": 6190 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 0.00020052986512524083, |
|
"loss": 0.163, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 0.0002003692999357739, |
|
"loss": 0.1319, |
|
"step": 6210 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 0.000200208734746307, |
|
"loss": 0.1593, |
|
"step": 6220 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 0.00020004816955684006, |
|
"loss": 0.1239, |
|
"step": 6230 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 0.00019988760436737314, |
|
"loss": 0.1435, |
|
"step": 6240 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 0.0001997270391779062, |
|
"loss": 0.1543, |
|
"step": 6250 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 0.0001995664739884393, |
|
"loss": 0.1415, |
|
"step": 6260 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 0.00019940590879897238, |
|
"loss": 0.1506, |
|
"step": 6270 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 0.00019924534360950543, |
|
"loss": 0.1772, |
|
"step": 6280 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 0.00019908477842003853, |
|
"loss": 0.1675, |
|
"step": 6290 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 0.00019892421323057158, |
|
"loss": 0.1438, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 0.00019876364804110469, |
|
"loss": 0.1662, |
|
"step": 6310 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 0.00019860308285163776, |
|
"loss": 0.1588, |
|
"step": 6320 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 0.00019844251766217081, |
|
"loss": 0.1464, |
|
"step": 6330 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 0.00019828195247270392, |
|
"loss": 0.1483, |
|
"step": 6340 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 0.00019812138728323697, |
|
"loss": 0.1432, |
|
"step": 6350 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 0.00019796082209377005, |
|
"loss": 0.134, |
|
"step": 6360 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 0.00019780025690430315, |
|
"loss": 0.1596, |
|
"step": 6370 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 0.0001976396917148362, |
|
"loss": 0.1319, |
|
"step": 6380 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 0.00019747912652536928, |
|
"loss": 0.1383, |
|
"step": 6390 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 0.00019731856133590236, |
|
"loss": 0.1604, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 0.00019715799614643544, |
|
"loss": 0.1738, |
|
"step": 6410 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 0.00019699743095696851, |
|
"loss": 0.137, |
|
"step": 6420 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 0.0001968368657675016, |
|
"loss": 0.1832, |
|
"step": 6430 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 0.00019667630057803467, |
|
"loss": 0.1173, |
|
"step": 6440 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 0.00019651573538856772, |
|
"loss": 0.1458, |
|
"step": 6450 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 0.00019635517019910082, |
|
"loss": 0.1674, |
|
"step": 6460 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 0.0001961946050096339, |
|
"loss": 0.108, |
|
"step": 6470 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 0.00019603403982016695, |
|
"loss": 0.1248, |
|
"step": 6480 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 0.00019587347463070006, |
|
"loss": 0.1089, |
|
"step": 6490 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 0.0001957129094412331, |
|
"loss": 0.1422, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 0.0001955523442517662, |
|
"loss": 0.132, |
|
"step": 6510 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 0.0001953917790622993, |
|
"loss": 0.1383, |
|
"step": 6520 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 0.00019523121387283234, |
|
"loss": 0.1285, |
|
"step": 6530 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 0.00019507064868336545, |
|
"loss": 0.1654, |
|
"step": 6540 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 0.0001949100834938985, |
|
"loss": 0.1414, |
|
"step": 6550 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 0.00019474951830443157, |
|
"loss": 0.1488, |
|
"step": 6560 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 0.00019458895311496468, |
|
"loss": 0.1253, |
|
"step": 6570 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 0.00019442838792549773, |
|
"loss": 0.1461, |
|
"step": 6580 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 0.0001942678227360308, |
|
"loss": 0.156, |
|
"step": 6590 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 0.00019410725754656389, |
|
"loss": 0.1537, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 0.00019394669235709696, |
|
"loss": 0.1584, |
|
"step": 6610 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 0.00019378612716763007, |
|
"loss": 0.1335, |
|
"step": 6620 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 0.00019362556197816312, |
|
"loss": 0.1703, |
|
"step": 6630 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 0.0001934649967886962, |
|
"loss": 0.1611, |
|
"step": 6640 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 0.00019330443159922925, |
|
"loss": 0.1458, |
|
"step": 6650 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 0.00019314386640976235, |
|
"loss": 0.1319, |
|
"step": 6660 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 0.00019298330122029543, |
|
"loss": 0.1371, |
|
"step": 6670 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 0.0001928227360308285, |
|
"loss": 0.1294, |
|
"step": 6680 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 0.00019266217084136159, |
|
"loss": 0.1134, |
|
"step": 6690 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 0.00019250160565189464, |
|
"loss": 0.1439, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 0.00019234104046242774, |
|
"loss": 0.1723, |
|
"step": 6710 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 0.00019218047527296082, |
|
"loss": 0.1478, |
|
"step": 6720 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 0.00019201991008349387, |
|
"loss": 0.1309, |
|
"step": 6730 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 0.00019185934489402697, |
|
"loss": 0.121, |
|
"step": 6740 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 0.00019169877970456002, |
|
"loss": 0.131, |
|
"step": 6750 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 0.0001915382145150931, |
|
"loss": 0.1715, |
|
"step": 6760 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 0.0001913776493256262, |
|
"loss": 0.1426, |
|
"step": 6770 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 0.00019121708413615926, |
|
"loss": 0.133, |
|
"step": 6780 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 0.00019105651894669236, |
|
"loss": 0.1594, |
|
"step": 6790 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 0.0001908959537572254, |
|
"loss": 0.1191, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 0.0001907353885677585, |
|
"loss": 0.2101, |
|
"step": 6810 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 0.0001905748233782916, |
|
"loss": 0.1593, |
|
"step": 6820 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 0.00019041425818882465, |
|
"loss": 0.1425, |
|
"step": 6830 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 0.00019025369299935772, |
|
"loss": 0.1542, |
|
"step": 6840 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 0.0001900931278098908, |
|
"loss": 0.1391, |
|
"step": 6850 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 0.00018993256262042388, |
|
"loss": 0.149, |
|
"step": 6860 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 0.00018977199743095696, |
|
"loss": 0.2022, |
|
"step": 6870 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 0.00018961143224149003, |
|
"loss": 0.182, |
|
"step": 6880 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 0.0001894508670520231, |
|
"loss": 0.1914, |
|
"step": 6890 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 0.00018929030186255616, |
|
"loss": 0.1466, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 0.00018912973667308927, |
|
"loss": 0.145, |
|
"step": 6910 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 0.00018896917148362235, |
|
"loss": 0.1507, |
|
"step": 6920 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 0.0001888086062941554, |
|
"loss": 0.1464, |
|
"step": 6930 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 0.0001886480411046885, |
|
"loss": 0.1527, |
|
"step": 6940 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 0.00018848747591522155, |
|
"loss": 0.1466, |
|
"step": 6950 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 0.00018832691072575463, |
|
"loss": 0.1119, |
|
"step": 6960 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 0.00018816634553628773, |
|
"loss": 0.1645, |
|
"step": 6970 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 0.00018800578034682078, |
|
"loss": 0.1619, |
|
"step": 6980 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 0.0001878452151573539, |
|
"loss": 0.158, |
|
"step": 6990 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 0.00018768464996788694, |
|
"loss": 0.1624, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"eval_cer": 0.9425327302461178, |
|
"eval_loss": 0.12462034821510315, |
|
"eval_runtime": 53.629, |
|
"eval_samples_per_second": 18.647, |
|
"eval_steps_per_second": 4.662, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 0.00018752408477842002, |
|
"loss": 0.1716, |
|
"step": 7010 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 0.00018736351958895312, |
|
"loss": 0.1565, |
|
"step": 7020 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 0.00018720295439948617, |
|
"loss": 0.1526, |
|
"step": 7030 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 0.00018704238921001925, |
|
"loss": 0.1555, |
|
"step": 7040 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 0.00018688182402055233, |
|
"loss": 0.1381, |
|
"step": 7050 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 0.0001867212588310854, |
|
"loss": 0.1797, |
|
"step": 7060 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 0.00018656069364161848, |
|
"loss": 0.1567, |
|
"step": 7070 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 0.00018640012845215156, |
|
"loss": 0.1595, |
|
"step": 7080 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 0.00018623956326268464, |
|
"loss": 0.1767, |
|
"step": 7090 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 0.0001860789980732177, |
|
"loss": 0.1406, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 0.0001859184328837508, |
|
"loss": 0.1016, |
|
"step": 7110 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 0.00018575786769428387, |
|
"loss": 0.1569, |
|
"step": 7120 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 0.00018559730250481692, |
|
"loss": 0.1366, |
|
"step": 7130 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 0.00018543673731535003, |
|
"loss": 0.1414, |
|
"step": 7140 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 0.00018527617212588308, |
|
"loss": 0.1463, |
|
"step": 7150 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 0.00018511560693641618, |
|
"loss": 0.1378, |
|
"step": 7160 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 0.00018495504174694926, |
|
"loss": 0.1163, |
|
"step": 7170 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 0.0001847944765574823, |
|
"loss": 0.1179, |
|
"step": 7180 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 0.00018463391136801542, |
|
"loss": 0.111, |
|
"step": 7190 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 0.00018447334617854847, |
|
"loss": 0.1469, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 0.00018431278098908155, |
|
"loss": 0.1177, |
|
"step": 7210 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 0.00018415221579961462, |
|
"loss": 0.1742, |
|
"step": 7220 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 0.0001839916506101477, |
|
"loss": 0.1546, |
|
"step": 7230 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 0.00018383108542068078, |
|
"loss": 0.1275, |
|
"step": 7240 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 0.00018367052023121386, |
|
"loss": 0.1458, |
|
"step": 7250 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 0.00018350995504174693, |
|
"loss": 0.1299, |
|
"step": 7260 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 0.00018334938985227998, |
|
"loss": 0.1107, |
|
"step": 7270 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 0.0001831888246628131, |
|
"loss": 0.1394, |
|
"step": 7280 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 0.00018302825947334617, |
|
"loss": 0.1267, |
|
"step": 7290 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 0.00018286769428387922, |
|
"loss": 0.127, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 0.00018270712909441232, |
|
"loss": 0.1175, |
|
"step": 7310 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 0.00018254656390494537, |
|
"loss": 0.1285, |
|
"step": 7320 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 0.00018238599871547848, |
|
"loss": 0.1552, |
|
"step": 7330 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 0.00018222543352601156, |
|
"loss": 0.1248, |
|
"step": 7340 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 0.0001820648683365446, |
|
"loss": 0.1589, |
|
"step": 7350 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 0.0001819043031470777, |
|
"loss": 0.1457, |
|
"step": 7360 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 0.00018174373795761076, |
|
"loss": 0.1292, |
|
"step": 7370 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 0.00018158317276814384, |
|
"loss": 0.1634, |
|
"step": 7380 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 0.00018142260757867694, |
|
"loss": 0.1407, |
|
"step": 7390 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 0.00018126204238921, |
|
"loss": 0.1196, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 0.00018110147719974307, |
|
"loss": 0.1365, |
|
"step": 7410 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 0.00018094091201027615, |
|
"loss": 0.1412, |
|
"step": 7420 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 0.00018078034682080923, |
|
"loss": 0.1397, |
|
"step": 7430 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 0.00018061978163134233, |
|
"loss": 0.1404, |
|
"step": 7440 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 0.00018045921644187538, |
|
"loss": 0.1599, |
|
"step": 7450 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 0.00018029865125240846, |
|
"loss": 0.1101, |
|
"step": 7460 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 0.0001801380860629415, |
|
"loss": 0.1138, |
|
"step": 7470 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 0.00017997752087347462, |
|
"loss": 0.1709, |
|
"step": 7480 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 0.0001798169556840077, |
|
"loss": 0.1259, |
|
"step": 7490 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 0.00017965639049454077, |
|
"loss": 0.1551, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 0.00017949582530507385, |
|
"loss": 0.1486, |
|
"step": 7510 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 0.0001793352601156069, |
|
"loss": 0.1328, |
|
"step": 7520 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 0.00017917469492614, |
|
"loss": 0.1168, |
|
"step": 7530 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 0.00017901412973667308, |
|
"loss": 0.1355, |
|
"step": 7540 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 0.00017885356454720613, |
|
"loss": 0.1175, |
|
"step": 7550 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 0.00017869299935773924, |
|
"loss": 0.103, |
|
"step": 7560 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 0.0001785324341682723, |
|
"loss": 0.1573, |
|
"step": 7570 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 0.00017837186897880537, |
|
"loss": 0.131, |
|
"step": 7580 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 0.00017821130378933847, |
|
"loss": 0.1511, |
|
"step": 7590 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 0.00017805073859987152, |
|
"loss": 0.131, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 0.0001778901734104046, |
|
"loss": 0.1379, |
|
"step": 7610 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 0.00017772960822093768, |
|
"loss": 0.1397, |
|
"step": 7620 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 0.00017756904303147075, |
|
"loss": 0.1584, |
|
"step": 7630 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 0.00017740847784200386, |
|
"loss": 0.1397, |
|
"step": 7640 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 0.0001772479126525369, |
|
"loss": 0.1382, |
|
"step": 7650 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 0.00017708734746307, |
|
"loss": 0.143, |
|
"step": 7660 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 0.00017692678227360304, |
|
"loss": 0.1418, |
|
"step": 7670 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 0.00017676621708413614, |
|
"loss": 0.153, |
|
"step": 7680 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 0.00017660565189466922, |
|
"loss": 0.1608, |
|
"step": 7690 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 0.0001764450867052023, |
|
"loss": 0.1329, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 0.00017628452151573538, |
|
"loss": 0.1209, |
|
"step": 7710 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 0.00017612395632626843, |
|
"loss": 0.1573, |
|
"step": 7720 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 0.00017596339113680153, |
|
"loss": 0.1354, |
|
"step": 7730 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 0.0001758028259473346, |
|
"loss": 0.1776, |
|
"step": 7740 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 0.00017564226075786766, |
|
"loss": 0.1165, |
|
"step": 7750 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 0.00017548169556840077, |
|
"loss": 0.1452, |
|
"step": 7760 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 0.00017532113037893382, |
|
"loss": 0.139, |
|
"step": 7770 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 0.0001751605651894669, |
|
"loss": 0.1742, |
|
"step": 7780 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 0.000175, |
|
"loss": 0.1332, |
|
"step": 7790 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 0.00017483943481053305, |
|
"loss": 0.1198, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 0.00017467886962106615, |
|
"loss": 0.1556, |
|
"step": 7810 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 0.0001745183044315992, |
|
"loss": 0.1456, |
|
"step": 7820 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 0.00017435773924213228, |
|
"loss": 0.1347, |
|
"step": 7830 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 0.0001741971740526654, |
|
"loss": 0.1291, |
|
"step": 7840 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 0.00017403660886319844, |
|
"loss": 0.1397, |
|
"step": 7850 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 0.00017387604367373152, |
|
"loss": 0.1371, |
|
"step": 7860 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 0.0001737154784842646, |
|
"loss": 0.1214, |
|
"step": 7870 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 0.00017355491329479767, |
|
"loss": 0.1581, |
|
"step": 7880 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 0.00017339434810533075, |
|
"loss": 0.1198, |
|
"step": 7890 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 0.00017323378291586383, |
|
"loss": 0.1322, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 0.0001730732177263969, |
|
"loss": 0.1143, |
|
"step": 7910 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 0.00017291265253692995, |
|
"loss": 0.0978, |
|
"step": 7920 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 0.00017275208734746306, |
|
"loss": 0.1209, |
|
"step": 7930 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 0.00017259152215799614, |
|
"loss": 0.1242, |
|
"step": 7940 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 0.0001724309569685292, |
|
"loss": 0.1667, |
|
"step": 7950 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 0.0001722703917790623, |
|
"loss": 0.1852, |
|
"step": 7960 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 0.00017210982658959534, |
|
"loss": 0.1568, |
|
"step": 7970 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 0.00017194926140012845, |
|
"loss": 0.1364, |
|
"step": 7980 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 0.00017178869621066153, |
|
"loss": 0.1469, |
|
"step": 7990 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 0.00017162813102119458, |
|
"loss": 0.1337, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"eval_cer": 0.9423674888953878, |
|
"eval_loss": 0.12129945307970047, |
|
"eval_runtime": 53.2179, |
|
"eval_samples_per_second": 18.791, |
|
"eval_steps_per_second": 4.698, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 0.00017146756583172768, |
|
"loss": 0.1231, |
|
"step": 8010 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 0.00017130700064226073, |
|
"loss": 0.1527, |
|
"step": 8020 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 0.0001711464354527938, |
|
"loss": 0.142, |
|
"step": 8030 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 0.00017098587026332691, |
|
"loss": 0.1259, |
|
"step": 8040 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 0.00017082530507385996, |
|
"loss": 0.1377, |
|
"step": 8050 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 0.00017066473988439304, |
|
"loss": 0.1577, |
|
"step": 8060 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 0.00017050417469492612, |
|
"loss": 0.1704, |
|
"step": 8070 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 0.0001703436095054592, |
|
"loss": 0.1113, |
|
"step": 8080 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 0.0001701830443159923, |
|
"loss": 0.1418, |
|
"step": 8090 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 0.00017002247912652535, |
|
"loss": 0.1252, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 0.00016986191393705843, |
|
"loss": 0.138, |
|
"step": 8110 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 0.00016970134874759148, |
|
"loss": 0.1408, |
|
"step": 8120 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 0.00016954078355812459, |
|
"loss": 0.1416, |
|
"step": 8130 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 0.00016938021836865766, |
|
"loss": 0.1238, |
|
"step": 8140 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 0.00016921965317919074, |
|
"loss": 0.1236, |
|
"step": 8150 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 0.00016905908798972382, |
|
"loss": 0.1786, |
|
"step": 8160 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 0.00016889852280025687, |
|
"loss": 0.1331, |
|
"step": 8170 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 0.00016873795761078997, |
|
"loss": 0.1481, |
|
"step": 8180 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 0.00016857739242132305, |
|
"loss": 0.1409, |
|
"step": 8190 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 0.0001684168272318561, |
|
"loss": 0.1445, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 0.0001682562620423892, |
|
"loss": 0.1123, |
|
"step": 8210 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 0.00016809569685292226, |
|
"loss": 0.124, |
|
"step": 8220 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 0.00016793513166345534, |
|
"loss": 0.1299, |
|
"step": 8230 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 0.00016777456647398844, |
|
"loss": 0.1343, |
|
"step": 8240 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 0.0001676140012845215, |
|
"loss": 0.1454, |
|
"step": 8250 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 0.00016745343609505457, |
|
"loss": 0.1335, |
|
"step": 8260 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 0.00016729287090558765, |
|
"loss": 0.1427, |
|
"step": 8270 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 0.00016713230571612072, |
|
"loss": 0.1179, |
|
"step": 8280 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 0.00016697174052665383, |
|
"loss": 0.1191, |
|
"step": 8290 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 0.00016681117533718688, |
|
"loss": 0.1388, |
|
"step": 8300 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 0.00016665061014771996, |
|
"loss": 0.1389, |
|
"step": 8310 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 0.000166490044958253, |
|
"loss": 0.1215, |
|
"step": 8320 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 0.0001663294797687861, |
|
"loss": 0.1119, |
|
"step": 8330 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 0.0001661689145793192, |
|
"loss": 0.1141, |
|
"step": 8340 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 0.00016600834938985227, |
|
"loss": 0.153, |
|
"step": 8350 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 0.00016584778420038535, |
|
"loss": 0.1569, |
|
"step": 8360 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 0.0001656872190109184, |
|
"loss": 0.1246, |
|
"step": 8370 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 0.0001655266538214515, |
|
"loss": 0.1263, |
|
"step": 8380 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 0.00016536608863198458, |
|
"loss": 0.1357, |
|
"step": 8390 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 0.00016520552344251763, |
|
"loss": 0.1197, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 0.00016504495825305074, |
|
"loss": 0.1148, |
|
"step": 8410 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 0.00016488439306358379, |
|
"loss": 0.1447, |
|
"step": 8420 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 0.00016472382787411686, |
|
"loss": 0.1348, |
|
"step": 8430 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 0.00016456326268464997, |
|
"loss": 0.1315, |
|
"step": 8440 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 0.00016440269749518302, |
|
"loss": 0.144, |
|
"step": 8450 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 0.00016424213230571612, |
|
"loss": 0.1292, |
|
"step": 8460 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 0.00016408156711624917, |
|
"loss": 0.1371, |
|
"step": 8470 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 0.00016392100192678225, |
|
"loss": 0.1422, |
|
"step": 8480 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 0.00016376043673731536, |
|
"loss": 0.1368, |
|
"step": 8490 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 0.0001635998715478484, |
|
"loss": 0.1805, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 0.00016343930635838149, |
|
"loss": 0.1665, |
|
"step": 8510 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 0.00016327874116891456, |
|
"loss": 0.1416, |
|
"step": 8520 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 0.00016311817597944764, |
|
"loss": 0.1401, |
|
"step": 8530 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 0.00016295761078998072, |
|
"loss": 0.1504, |
|
"step": 8540 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 0.0001627970456005138, |
|
"loss": 0.1411, |
|
"step": 8550 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 0.00016263648041104687, |
|
"loss": 0.1659, |
|
"step": 8560 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 0.00016247591522157992, |
|
"loss": 0.165, |
|
"step": 8570 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 0.00016231535003211303, |
|
"loss": 0.1337, |
|
"step": 8580 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 0.0001621547848426461, |
|
"loss": 0.1296, |
|
"step": 8590 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 0.00016199421965317916, |
|
"loss": 0.1176, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 0.00016183365446371226, |
|
"loss": 0.114, |
|
"step": 8610 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 0.0001616730892742453, |
|
"loss": 0.1396, |
|
"step": 8620 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 0.00016151252408477842, |
|
"loss": 0.1292, |
|
"step": 8630 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 0.0001613519588953115, |
|
"loss": 0.1158, |
|
"step": 8640 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 0.00016119139370584455, |
|
"loss": 0.1323, |
|
"step": 8650 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 0.00016103082851637765, |
|
"loss": 0.1254, |
|
"step": 8660 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 0.0001608702633269107, |
|
"loss": 0.1519, |
|
"step": 8670 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 0.00016070969813744378, |
|
"loss": 0.1308, |
|
"step": 8680 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 0.00016054913294797688, |
|
"loss": 0.1274, |
|
"step": 8690 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 0.00016038856775850993, |
|
"loss": 0.1414, |
|
"step": 8700 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 0.000160228002569043, |
|
"loss": 0.1552, |
|
"step": 8710 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 0.0001600674373795761, |
|
"loss": 0.1316, |
|
"step": 8720 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 0.00015990687219010917, |
|
"loss": 0.1464, |
|
"step": 8730 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 0.00015974630700064227, |
|
"loss": 0.1301, |
|
"step": 8740 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 0.00015958574181117532, |
|
"loss": 0.1399, |
|
"step": 8750 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 0.0001594251766217084, |
|
"loss": 0.1017, |
|
"step": 8760 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 0.00015926461143224145, |
|
"loss": 0.1292, |
|
"step": 8770 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 0.00015910404624277456, |
|
"loss": 0.1087, |
|
"step": 8780 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 0.00015894348105330763, |
|
"loss": 0.182, |
|
"step": 8790 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 0.0001587829158638407, |
|
"loss": 0.1273, |
|
"step": 8800 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 0.0001586223506743738, |
|
"loss": 0.1411, |
|
"step": 8810 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 0.00015846178548490684, |
|
"loss": 0.125, |
|
"step": 8820 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 0.00015830122029543994, |
|
"loss": 0.145, |
|
"step": 8830 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 0.00015814065510597302, |
|
"loss": 0.1278, |
|
"step": 8840 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 0.00015798008991650607, |
|
"loss": 0.1353, |
|
"step": 8850 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 0.00015781952472703918, |
|
"loss": 0.1314, |
|
"step": 8860 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 0.00015765895953757223, |
|
"loss": 0.1316, |
|
"step": 8870 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 0.0001574983943481053, |
|
"loss": 0.1483, |
|
"step": 8880 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 0.0001573378291586384, |
|
"loss": 0.1639, |
|
"step": 8890 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 0.00015717726396917146, |
|
"loss": 0.1411, |
|
"step": 8900 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 0.00015701669877970454, |
|
"loss": 0.1178, |
|
"step": 8910 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 0.00015685613359023762, |
|
"loss": 0.1318, |
|
"step": 8920 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 0.0001566955684007707, |
|
"loss": 0.1453, |
|
"step": 8930 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 0.0001565350032113038, |
|
"loss": 0.142, |
|
"step": 8940 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 0.00015637443802183685, |
|
"loss": 0.1172, |
|
"step": 8950 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 0.00015621387283236993, |
|
"loss": 0.129, |
|
"step": 8960 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 0.00015605330764290298, |
|
"loss": 0.1323, |
|
"step": 8970 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 0.00015589274245343608, |
|
"loss": 0.109, |
|
"step": 8980 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 0.00015573217726396916, |
|
"loss": 0.122, |
|
"step": 8990 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 0.00015557161207450224, |
|
"loss": 0.131, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"eval_cer": 0.9422769995842738, |
|
"eval_loss": 0.11952092498540878, |
|
"eval_runtime": 52.5313, |
|
"eval_samples_per_second": 19.036, |
|
"eval_steps_per_second": 4.759, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 0.00015541104688503532, |
|
"loss": 0.1331, |
|
"step": 9010 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 0.00015525048169556837, |
|
"loss": 0.1607, |
|
"step": 9020 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 0.00015508991650610147, |
|
"loss": 0.1155, |
|
"step": 9030 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 0.00015492935131663455, |
|
"loss": 0.1239, |
|
"step": 9040 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 0.0001547687861271676, |
|
"loss": 0.1601, |
|
"step": 9050 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 0.0001546082209377007, |
|
"loss": 0.1114, |
|
"step": 9060 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 0.00015444765574823376, |
|
"loss": 0.123, |
|
"step": 9070 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 0.00015428709055876683, |
|
"loss": 0.153, |
|
"step": 9080 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 0.00015412652536929994, |
|
"loss": 0.1136, |
|
"step": 9090 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 0.000153965960179833, |
|
"loss": 0.1469, |
|
"step": 9100 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 0.0001538053949903661, |
|
"loss": 0.1355, |
|
"step": 9110 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 0.00015364482980089914, |
|
"loss": 0.127, |
|
"step": 9120 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 0.00015348426461143222, |
|
"loss": 0.1455, |
|
"step": 9130 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 0.00015332369942196533, |
|
"loss": 0.125, |
|
"step": 9140 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 0.00015316313423249838, |
|
"loss": 0.1218, |
|
"step": 9150 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 0.00015300256904303146, |
|
"loss": 0.153, |
|
"step": 9160 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 0.00015284200385356453, |
|
"loss": 0.1332, |
|
"step": 9170 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 0.0001526814386640976, |
|
"loss": 0.1307, |
|
"step": 9180 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 0.0001525208734746307, |
|
"loss": 0.1384, |
|
"step": 9190 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 0.00015236030828516377, |
|
"loss": 0.1618, |
|
"step": 9200 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 0.00015219974309569684, |
|
"loss": 0.1402, |
|
"step": 9210 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 0.0001520391779062299, |
|
"loss": 0.1607, |
|
"step": 9220 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 0.000151878612716763, |
|
"loss": 0.1103, |
|
"step": 9230 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 0.00015171804752729608, |
|
"loss": 0.1248, |
|
"step": 9240 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 0.00015155748233782913, |
|
"loss": 0.1511, |
|
"step": 9250 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 0.00015139691714836223, |
|
"loss": 0.1492, |
|
"step": 9260 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 0.00015123635195889528, |
|
"loss": 0.135, |
|
"step": 9270 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 0.0001510757867694284, |
|
"loss": 0.1448, |
|
"step": 9280 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 0.00015091522157996147, |
|
"loss": 0.1501, |
|
"step": 9290 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 0.00015075465639049452, |
|
"loss": 0.118, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 0.00015059409120102762, |
|
"loss": 0.144, |
|
"step": 9310 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 0.00015043352601156067, |
|
"loss": 0.1299, |
|
"step": 9320 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 0.00015027296082209375, |
|
"loss": 0.1309, |
|
"step": 9330 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 0.00015011239563262685, |
|
"loss": 0.1635, |
|
"step": 9340 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 0.0001499518304431599, |
|
"loss": 0.1189, |
|
"step": 9350 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 0.00014979126525369298, |
|
"loss": 0.1308, |
|
"step": 9360 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 0.00014963070006422606, |
|
"loss": 0.1404, |
|
"step": 9370 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 0.00014947013487475914, |
|
"loss": 0.1304, |
|
"step": 9380 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 0.00014930956968529222, |
|
"loss": 0.1428, |
|
"step": 9390 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 0.0001491490044958253, |
|
"loss": 0.1022, |
|
"step": 9400 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 0.00014898843930635837, |
|
"loss": 0.1143, |
|
"step": 9410 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 0.00014882787411689145, |
|
"loss": 0.146, |
|
"step": 9420 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 0.00014866730892742453, |
|
"loss": 0.1029, |
|
"step": 9430 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 0.0001485067437379576, |
|
"loss": 0.1133, |
|
"step": 9440 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 0.00014834617854849068, |
|
"loss": 0.1399, |
|
"step": 9450 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 0.00014818561335902376, |
|
"loss": 0.113, |
|
"step": 9460 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 0.00014802504816955684, |
|
"loss": 0.1153, |
|
"step": 9470 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 0.00014786448298008992, |
|
"loss": 0.1315, |
|
"step": 9480 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 0.00014770391779062297, |
|
"loss": 0.1063, |
|
"step": 9490 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 0.00014754335260115604, |
|
"loss": 0.1152, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 0.00014738278741168915, |
|
"loss": 0.1193, |
|
"step": 9510 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 0.00014722222222222223, |
|
"loss": 0.0877, |
|
"step": 9520 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 0.00014706165703275528, |
|
"loss": 0.1244, |
|
"step": 9530 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 0.00014690109184328835, |
|
"loss": 0.1254, |
|
"step": 9540 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 0.00014674052665382143, |
|
"loss": 0.1054, |
|
"step": 9550 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 0.0001465799614643545, |
|
"loss": 0.1182, |
|
"step": 9560 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 0.0001464193962748876, |
|
"loss": 0.1267, |
|
"step": 9570 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 0.00014625883108542067, |
|
"loss": 0.1232, |
|
"step": 9580 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 0.00014609826589595374, |
|
"loss": 0.1134, |
|
"step": 9590 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 0.00014593770070648682, |
|
"loss": 0.0977, |
|
"step": 9600 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 0.0001457771355170199, |
|
"loss": 0.1182, |
|
"step": 9610 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 0.00014561657032755298, |
|
"loss": 0.1376, |
|
"step": 9620 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 0.00014545600513808605, |
|
"loss": 0.1474, |
|
"step": 9630 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 0.00014529543994861913, |
|
"loss": 0.1081, |
|
"step": 9640 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 0.0001451348747591522, |
|
"loss": 0.1237, |
|
"step": 9650 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 0.0001449743095696853, |
|
"loss": 0.1433, |
|
"step": 9660 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 0.00014481374438021836, |
|
"loss": 0.1207, |
|
"step": 9670 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 0.00014465317919075144, |
|
"loss": 0.1167, |
|
"step": 9680 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 0.00014449261400128452, |
|
"loss": 0.1344, |
|
"step": 9690 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 0.00014433204881181757, |
|
"loss": 0.1412, |
|
"step": 9700 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 0.00014417148362235068, |
|
"loss": 0.0905, |
|
"step": 9710 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 0.00014401091843288375, |
|
"loss": 0.1156, |
|
"step": 9720 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 0.0001438503532434168, |
|
"loss": 0.1283, |
|
"step": 9730 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 0.00014368978805394988, |
|
"loss": 0.1416, |
|
"step": 9740 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 0.00014352922286448296, |
|
"loss": 0.1204, |
|
"step": 9750 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 0.00014336865767501606, |
|
"loss": 0.1396, |
|
"step": 9760 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 0.00014320809248554911, |
|
"loss": 0.1202, |
|
"step": 9770 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 0.0001430475272960822, |
|
"loss": 0.114, |
|
"step": 9780 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 0.00014288696210661527, |
|
"loss": 0.1043, |
|
"step": 9790 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 0.00014272639691714835, |
|
"loss": 0.1209, |
|
"step": 9800 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 0.00014256583172768143, |
|
"loss": 0.1331, |
|
"step": 9810 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 0.0001424052665382145, |
|
"loss": 0.1103, |
|
"step": 9820 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 0.00014224470134874758, |
|
"loss": 0.1052, |
|
"step": 9830 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 0.00014208413615928066, |
|
"loss": 0.1268, |
|
"step": 9840 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 0.00014192357096981374, |
|
"loss": 0.0955, |
|
"step": 9850 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 0.00014176300578034681, |
|
"loss": 0.1153, |
|
"step": 9860 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 0.0001416024405908799, |
|
"loss": 0.1344, |
|
"step": 9870 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 0.00014144187540141297, |
|
"loss": 0.0937, |
|
"step": 9880 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 0.00014128131021194605, |
|
"loss": 0.0981, |
|
"step": 9890 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 0.0001411207450224791, |
|
"loss": 0.1203, |
|
"step": 9900 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 0.00014096017983301218, |
|
"loss": 0.1067, |
|
"step": 9910 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 0.00014079961464354528, |
|
"loss": 0.1048, |
|
"step": 9920 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 0.00014063904945407836, |
|
"loss": 0.1203, |
|
"step": 9930 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 0.0001404784842646114, |
|
"loss": 0.1052, |
|
"step": 9940 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 0.00014031791907514449, |
|
"loss": 0.0996, |
|
"step": 9950 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 0.00014015735388567756, |
|
"loss": 0.1239, |
|
"step": 9960 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 0.00013999678869621064, |
|
"loss": 0.1313, |
|
"step": 9970 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 0.00013983622350674372, |
|
"loss": 0.1244, |
|
"step": 9980 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 0.0001396756583172768, |
|
"loss": 0.1085, |
|
"step": 9990 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 0.00013951509312780987, |
|
"loss": 0.1137, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"eval_cer": 0.9424448638135867, |
|
"eval_loss": 0.11777233332395554, |
|
"eval_runtime": 53.6735, |
|
"eval_samples_per_second": 18.631, |
|
"eval_steps_per_second": 4.658, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 0.00013935452793834295, |
|
"loss": 0.0858, |
|
"step": 10010 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 0.00013919396274887603, |
|
"loss": 0.1018, |
|
"step": 10020 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 0.0001390333975594091, |
|
"loss": 0.1084, |
|
"step": 10030 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 0.00013887283236994219, |
|
"loss": 0.1274, |
|
"step": 10040 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 0.00013871226718047526, |
|
"loss": 0.1414, |
|
"step": 10050 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 0.00013855170199100834, |
|
"loss": 0.1374, |
|
"step": 10060 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 0.00013839113680154142, |
|
"loss": 0.13, |
|
"step": 10070 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 0.0001382305716120745, |
|
"loss": 0.097, |
|
"step": 10080 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 0.00013807000642260757, |
|
"loss": 0.1264, |
|
"step": 10090 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 0.00013790944123314062, |
|
"loss": 0.1461, |
|
"step": 10100 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 0.0001377488760436737, |
|
"loss": 0.1205, |
|
"step": 10110 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 0.0001375883108542068, |
|
"loss": 0.1065, |
|
"step": 10120 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 0.00013742774566473989, |
|
"loss": 0.1195, |
|
"step": 10130 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 0.00013726718047527294, |
|
"loss": 0.1343, |
|
"step": 10140 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 0.000137106615285806, |
|
"loss": 0.1345, |
|
"step": 10150 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 0.0001369460500963391, |
|
"loss": 0.1146, |
|
"step": 10160 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 0.0001367854849068722, |
|
"loss": 0.118, |
|
"step": 10170 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 0.00013662491971740525, |
|
"loss": 0.15, |
|
"step": 10180 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 0.00013646435452793832, |
|
"loss": 0.1236, |
|
"step": 10190 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 0.0001363037893384714, |
|
"loss": 0.1198, |
|
"step": 10200 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 0.00013614322414900448, |
|
"loss": 0.1396, |
|
"step": 10210 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 0.00013598265895953756, |
|
"loss": 0.1278, |
|
"step": 10220 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 0.00013582209377007064, |
|
"loss": 0.1229, |
|
"step": 10230 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 0.0001356615285806037, |
|
"loss": 0.1138, |
|
"step": 10240 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 0.0001355009633911368, |
|
"loss": 0.0852, |
|
"step": 10250 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 0.00013534039820166987, |
|
"loss": 0.1306, |
|
"step": 10260 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 0.00013517983301220295, |
|
"loss": 0.0935, |
|
"step": 10270 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 0.00013501926782273602, |
|
"loss": 0.1163, |
|
"step": 10280 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 0.0001348587026332691, |
|
"loss": 0.1076, |
|
"step": 10290 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 0.00013469813744380218, |
|
"loss": 0.1184, |
|
"step": 10300 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 0.00013453757225433523, |
|
"loss": 0.0798, |
|
"step": 10310 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 0.00013437700706486833, |
|
"loss": 0.1119, |
|
"step": 10320 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 0.0001342164418754014, |
|
"loss": 0.1794, |
|
"step": 10330 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 0.0001340558766859345, |
|
"loss": 0.1446, |
|
"step": 10340 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 0.00013389531149646754, |
|
"loss": 0.1211, |
|
"step": 10350 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 0.00013373474630700062, |
|
"loss": 0.1392, |
|
"step": 10360 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 0.00013357418111753372, |
|
"loss": 0.1265, |
|
"step": 10370 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 0.00013341361592806677, |
|
"loss": 0.1201, |
|
"step": 10380 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 0.00013325305073859985, |
|
"loss": 0.1068, |
|
"step": 10390 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 0.00013309248554913293, |
|
"loss": 0.1251, |
|
"step": 10400 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 0.000132931920359666, |
|
"loss": 0.1205, |
|
"step": 10410 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 0.00013277135517019908, |
|
"loss": 0.0892, |
|
"step": 10420 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 0.00013261078998073216, |
|
"loss": 0.1053, |
|
"step": 10430 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 0.00013245022479126524, |
|
"loss": 0.1358, |
|
"step": 10440 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 0.00013228965960179832, |
|
"loss": 0.109, |
|
"step": 10450 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 0.0001321290944123314, |
|
"loss": 0.1401, |
|
"step": 10460 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 0.00013196852922286447, |
|
"loss": 0.1456, |
|
"step": 10470 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 0.00013180796403339755, |
|
"loss": 0.1415, |
|
"step": 10480 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 0.00013164739884393063, |
|
"loss": 0.1201, |
|
"step": 10490 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 0.0001314868336544637, |
|
"loss": 0.1287, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 0.00013132626846499676, |
|
"loss": 0.1246, |
|
"step": 10510 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 0.00013116570327552986, |
|
"loss": 0.1411, |
|
"step": 10520 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 0.00013100513808606294, |
|
"loss": 0.1104, |
|
"step": 10530 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 0.00013084457289659602, |
|
"loss": 0.1383, |
|
"step": 10540 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 0.00013068400770712907, |
|
"loss": 0.1179, |
|
"step": 10550 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 0.00013052344251766215, |
|
"loss": 0.1594, |
|
"step": 10560 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 0.00013036287732819525, |
|
"loss": 0.1233, |
|
"step": 10570 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 0.00013020231213872833, |
|
"loss": 0.1091, |
|
"step": 10580 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 0.00013004174694926138, |
|
"loss": 0.1308, |
|
"step": 10590 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 0.00012988118175979446, |
|
"loss": 0.1148, |
|
"step": 10600 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 0.00012972061657032753, |
|
"loss": 0.1658, |
|
"step": 10610 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 0.0001295600513808606, |
|
"loss": 0.1029, |
|
"step": 10620 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 0.0001293994861913937, |
|
"loss": 0.1018, |
|
"step": 10630 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 0.00012923892100192677, |
|
"loss": 0.1382, |
|
"step": 10640 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 0.00012907835581245984, |
|
"loss": 0.1124, |
|
"step": 10650 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 0.00012891779062299292, |
|
"loss": 0.1153, |
|
"step": 10660 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 0.000128757225433526, |
|
"loss": 0.1205, |
|
"step": 10670 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 0.00012859666024405908, |
|
"loss": 0.1178, |
|
"step": 10680 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 0.00012843609505459216, |
|
"loss": 0.1402, |
|
"step": 10690 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 0.00012827552986512523, |
|
"loss": 0.0957, |
|
"step": 10700 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 0.0001281149646756583, |
|
"loss": 0.1108, |
|
"step": 10710 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 0.0001279543994861914, |
|
"loss": 0.1046, |
|
"step": 10720 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 0.00012779383429672447, |
|
"loss": 0.099, |
|
"step": 10730 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 0.00012763326910725754, |
|
"loss": 0.1447, |
|
"step": 10740 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 0.0001274727039177906, |
|
"loss": 0.1565, |
|
"step": 10750 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 0.00012731213872832367, |
|
"loss": 0.1127, |
|
"step": 10760 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 0.00012715157353885678, |
|
"loss": 0.111, |
|
"step": 10770 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 0.00012699100834938986, |
|
"loss": 0.1223, |
|
"step": 10780 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 0.0001268304431599229, |
|
"loss": 0.105, |
|
"step": 10790 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 0.00012666987797045598, |
|
"loss": 0.139, |
|
"step": 10800 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 0.00012650931278098906, |
|
"loss": 0.1143, |
|
"step": 10810 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 0.00012634874759152217, |
|
"loss": 0.0948, |
|
"step": 10820 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 0.00012618818240205522, |
|
"loss": 0.1282, |
|
"step": 10830 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 0.0001260276172125883, |
|
"loss": 0.1267, |
|
"step": 10840 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 0.00012586705202312137, |
|
"loss": 0.1131, |
|
"step": 10850 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 0.00012570648683365445, |
|
"loss": 0.0881, |
|
"step": 10860 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 0.00012554592164418753, |
|
"loss": 0.1122, |
|
"step": 10870 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 0.0001253853564547206, |
|
"loss": 0.1167, |
|
"step": 10880 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 0.00012522479126525368, |
|
"loss": 0.1036, |
|
"step": 10890 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 0.00012506422607578676, |
|
"loss": 0.1162, |
|
"step": 10900 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 0.00012490366088631984, |
|
"loss": 0.1271, |
|
"step": 10910 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 0.00012474309569685292, |
|
"loss": 0.1228, |
|
"step": 10920 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 0.000124582530507386, |
|
"loss": 0.1121, |
|
"step": 10930 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 0.00012442196531791907, |
|
"loss": 0.1168, |
|
"step": 10940 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 0.00012426140012845215, |
|
"loss": 0.1098, |
|
"step": 10950 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 0.0001241008349389852, |
|
"loss": 0.1254, |
|
"step": 10960 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 0.0001239402697495183, |
|
"loss": 0.126, |
|
"step": 10970 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 0.00012377970456005138, |
|
"loss": 0.0993, |
|
"step": 10980 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 0.00012361913937058446, |
|
"loss": 0.1243, |
|
"step": 10990 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 0.0001234585741811175, |
|
"loss": 0.0958, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"eval_cer": 0.9422402792841115, |
|
"eval_loss": 0.11660391092300415, |
|
"eval_runtime": 53.6774, |
|
"eval_samples_per_second": 18.63, |
|
"eval_steps_per_second": 4.657, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 0.0001232980089916506, |
|
"loss": 0.1068, |
|
"step": 11010 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 0.0001231374438021837, |
|
"loss": 0.1095, |
|
"step": 11020 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 0.00012297687861271674, |
|
"loss": 0.1294, |
|
"step": 11030 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 0.00012281631342324982, |
|
"loss": 0.1176, |
|
"step": 11040 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 0.0001226557482337829, |
|
"loss": 0.098, |
|
"step": 11050 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 0.00012249518304431598, |
|
"loss": 0.1304, |
|
"step": 11060 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 0.00012233461785484905, |
|
"loss": 0.1234, |
|
"step": 11070 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 0.00012217405266538213, |
|
"loss": 0.1163, |
|
"step": 11080 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 0.00012201348747591521, |
|
"loss": 0.1267, |
|
"step": 11090 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 0.00012185292228644829, |
|
"loss": 0.1147, |
|
"step": 11100 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 0.00012169235709698135, |
|
"loss": 0.1416, |
|
"step": 11110 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 0.00012153179190751444, |
|
"loss": 0.1308, |
|
"step": 11120 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 0.00012137122671804752, |
|
"loss": 0.135, |
|
"step": 11130 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 0.0001212106615285806, |
|
"loss": 0.1007, |
|
"step": 11140 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 0.00012105009633911366, |
|
"loss": 0.1021, |
|
"step": 11150 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 0.00012088953114964674, |
|
"loss": 0.116, |
|
"step": 11160 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 0.00012072896596017983, |
|
"loss": 0.1111, |
|
"step": 11170 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 0.00012056840077071291, |
|
"loss": 0.1048, |
|
"step": 11180 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 0.00012040783558124597, |
|
"loss": 0.1352, |
|
"step": 11190 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 0.00012024727039177905, |
|
"loss": 0.1181, |
|
"step": 11200 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 0.00012008670520231212, |
|
"loss": 0.1115, |
|
"step": 11210 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 0.00011992614001284521, |
|
"loss": 0.1277, |
|
"step": 11220 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 0.00011976557482337828, |
|
"loss": 0.118, |
|
"step": 11230 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 0.00011960500963391136, |
|
"loss": 0.1369, |
|
"step": 11240 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 0.00011944444444444443, |
|
"loss": 0.1357, |
|
"step": 11250 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 0.0001192838792549775, |
|
"loss": 0.103, |
|
"step": 11260 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 0.0001191233140655106, |
|
"loss": 0.1073, |
|
"step": 11270 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 0.00011896274887604367, |
|
"loss": 0.1222, |
|
"step": 11280 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 0.00011880218368657674, |
|
"loss": 0.1101, |
|
"step": 11290 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 0.00011864161849710982, |
|
"loss": 0.1476, |
|
"step": 11300 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 0.00011848105330764289, |
|
"loss": 0.1097, |
|
"step": 11310 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 0.00011832048811817597, |
|
"loss": 0.1013, |
|
"step": 11320 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 0.00011815992292870905, |
|
"loss": 0.1091, |
|
"step": 11330 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 0.00011799935773924213, |
|
"loss": 0.1332, |
|
"step": 11340 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 0.00011783879254977519, |
|
"loss": 0.1267, |
|
"step": 11350 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 0.00011767822736030827, |
|
"loss": 0.1096, |
|
"step": 11360 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 0.00011751766217084136, |
|
"loss": 0.1007, |
|
"step": 11370 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 0.00011735709698137444, |
|
"loss": 0.1239, |
|
"step": 11380 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 0.0001171965317919075, |
|
"loss": 0.1304, |
|
"step": 11390 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 0.00011703596660244058, |
|
"loss": 0.1183, |
|
"step": 11400 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 0.00011687540141297366, |
|
"loss": 0.1324, |
|
"step": 11410 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 0.00011671483622350675, |
|
"loss": 0.1127, |
|
"step": 11420 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 0.00011655427103403981, |
|
"loss": 0.1376, |
|
"step": 11430 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 0.00011639370584457289, |
|
"loss": 0.1249, |
|
"step": 11440 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 0.00011623314065510597, |
|
"loss": 0.1029, |
|
"step": 11450 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 0.00011607257546563903, |
|
"loss": 0.1056, |
|
"step": 11460 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 0.00011591201027617212, |
|
"loss": 0.1025, |
|
"step": 11470 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 0.0001157514450867052, |
|
"loss": 0.1205, |
|
"step": 11480 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 0.00011559087989723826, |
|
"loss": 0.1479, |
|
"step": 11490 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 0.00011543031470777134, |
|
"loss": 0.1245, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 0.00011526974951830442, |
|
"loss": 0.1031, |
|
"step": 11510 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 0.00011510918432883751, |
|
"loss": 0.1091, |
|
"step": 11520 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 0.00011494861913937058, |
|
"loss": 0.1088, |
|
"step": 11530 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 0.00011478805394990365, |
|
"loss": 0.126, |
|
"step": 11540 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 0.00011462748876043673, |
|
"loss": 0.0949, |
|
"step": 11550 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 0.0001144669235709698, |
|
"loss": 0.0926, |
|
"step": 11560 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 0.00011430635838150289, |
|
"loss": 0.1162, |
|
"step": 11570 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 0.00011414579319203596, |
|
"loss": 0.1343, |
|
"step": 11580 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 0.00011398522800256903, |
|
"loss": 0.1422, |
|
"step": 11590 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 0.0001138246628131021, |
|
"loss": 0.122, |
|
"step": 11600 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 0.00011366409762363518, |
|
"loss": 0.1071, |
|
"step": 11610 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 0.00011350353243416827, |
|
"loss": 0.1223, |
|
"step": 11620 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 0.00011334296724470134, |
|
"loss": 0.1069, |
|
"step": 11630 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 0.0001131984585741811, |
|
"loss": 0.1317, |
|
"step": 11640 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 0.00011305394990366087, |
|
"loss": 0.1488, |
|
"step": 11650 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 0.00011289338471419395, |
|
"loss": 0.1389, |
|
"step": 11660 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 0.00011273281952472703, |
|
"loss": 0.1054, |
|
"step": 11670 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 0.00011257225433526009, |
|
"loss": 0.1037, |
|
"step": 11680 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 0.00011241168914579318, |
|
"loss": 0.116, |
|
"step": 11690 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 0.00011225112395632626, |
|
"loss": 0.118, |
|
"step": 11700 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 0.00011209055876685934, |
|
"loss": 0.127, |
|
"step": 11710 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 0.0001119299935773924, |
|
"loss": 0.1241, |
|
"step": 11720 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 0.00011176942838792548, |
|
"loss": 0.0846, |
|
"step": 11730 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 0.00011160886319845857, |
|
"loss": 0.094, |
|
"step": 11740 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 0.00011144829800899164, |
|
"loss": 0.1492, |
|
"step": 11750 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 0.00011128773281952471, |
|
"loss": 0.11, |
|
"step": 11760 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 0.00011112716763005779, |
|
"loss": 0.1386, |
|
"step": 11770 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 0.00011096660244059086, |
|
"loss": 0.1094, |
|
"step": 11780 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 0.00011080603725112395, |
|
"loss": 0.0972, |
|
"step": 11790 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 0.00011064547206165703, |
|
"loss": 0.1005, |
|
"step": 11800 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 0.0001104849068721901, |
|
"loss": 0.138, |
|
"step": 11810 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 0.00011032434168272317, |
|
"loss": 0.117, |
|
"step": 11820 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 0.00011016377649325625, |
|
"loss": 0.1265, |
|
"step": 11830 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 0.00011000321130378934, |
|
"loss": 0.1045, |
|
"step": 11840 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 0.00010984264611432241, |
|
"loss": 0.116, |
|
"step": 11850 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 0.00010968208092485548, |
|
"loss": 0.1205, |
|
"step": 11860 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 0.00010952151573538856, |
|
"loss": 0.1166, |
|
"step": 11870 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 0.00010936095054592163, |
|
"loss": 0.1283, |
|
"step": 11880 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 0.00010920038535645471, |
|
"loss": 0.1253, |
|
"step": 11890 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 0.00010903982016698779, |
|
"loss": 0.1385, |
|
"step": 11900 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 0.00010887925497752087, |
|
"loss": 0.1002, |
|
"step": 11910 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 0.00010871868978805393, |
|
"loss": 0.1282, |
|
"step": 11920 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 0.00010855812459858701, |
|
"loss": 0.1268, |
|
"step": 11930 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 0.0001083975594091201, |
|
"loss": 0.1202, |
|
"step": 11940 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 0.00010823699421965318, |
|
"loss": 0.1281, |
|
"step": 11950 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 0.00010807642903018624, |
|
"loss": 0.1031, |
|
"step": 11960 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 0.00010791586384071932, |
|
"loss": 0.1406, |
|
"step": 11970 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 0.0001077552986512524, |
|
"loss": 0.1306, |
|
"step": 11980 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 0.00010759473346178548, |
|
"loss": 0.1305, |
|
"step": 11990 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 0.00010743416827231855, |
|
"loss": 0.1067, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"eval_cer": 0.9422219191340304, |
|
"eval_loss": 0.1146860346198082, |
|
"eval_runtime": 53.6921, |
|
"eval_samples_per_second": 18.625, |
|
"eval_steps_per_second": 4.656, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 0.00010727360308285163, |
|
"loss": 0.1096, |
|
"step": 12010 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 0.0001071130378933847, |
|
"loss": 0.1298, |
|
"step": 12020 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 0.00010695247270391777, |
|
"loss": 0.1048, |
|
"step": 12030 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 0.00010679190751445086, |
|
"loss": 0.1385, |
|
"step": 12040 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 0.00010663134232498394, |
|
"loss": 0.1268, |
|
"step": 12050 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 0.000106470777135517, |
|
"loss": 0.1458, |
|
"step": 12060 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 0.00010631021194605008, |
|
"loss": 0.1378, |
|
"step": 12070 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 0.00010614964675658316, |
|
"loss": 0.0785, |
|
"step": 12080 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 0.00010598908156711625, |
|
"loss": 0.1342, |
|
"step": 12090 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 0.00010582851637764932, |
|
"loss": 0.1463, |
|
"step": 12100 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 0.0001056679511881824, |
|
"loss": 0.1109, |
|
"step": 12110 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 0.00010550738599871547, |
|
"loss": 0.1295, |
|
"step": 12120 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 0.00010534682080924854, |
|
"loss": 0.1335, |
|
"step": 12130 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 0.00010518625561978163, |
|
"loss": 0.158, |
|
"step": 12140 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 0.0001050256904303147, |
|
"loss": 0.0963, |
|
"step": 12150 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 0.00010486512524084777, |
|
"loss": 0.1087, |
|
"step": 12160 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 0.00010470456005138085, |
|
"loss": 0.1232, |
|
"step": 12170 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 0.00010454399486191392, |
|
"loss": 0.1136, |
|
"step": 12180 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 0.00010438342967244702, |
|
"loss": 0.1488, |
|
"step": 12190 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 0.00010422286448298008, |
|
"loss": 0.1066, |
|
"step": 12200 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 0.00010406229929351316, |
|
"loss": 0.1015, |
|
"step": 12210 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 0.00010390173410404624, |
|
"loss": 0.0888, |
|
"step": 12220 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 0.0001037411689145793, |
|
"loss": 0.0706, |
|
"step": 12230 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 0.00010358060372511239, |
|
"loss": 0.1065, |
|
"step": 12240 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 0.00010342003853564547, |
|
"loss": 0.1063, |
|
"step": 12250 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 0.00010325947334617855, |
|
"loss": 0.1236, |
|
"step": 12260 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 0.00010309890815671161, |
|
"loss": 0.1465, |
|
"step": 12270 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 0.00010293834296724469, |
|
"loss": 0.1198, |
|
"step": 12280 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 0.00010277777777777778, |
|
"loss": 0.1124, |
|
"step": 12290 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 0.00010261721258831084, |
|
"loss": 0.1323, |
|
"step": 12300 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 0.00010245664739884392, |
|
"loss": 0.1218, |
|
"step": 12310 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 0.000102296082209377, |
|
"loss": 0.1072, |
|
"step": 12320 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 0.00010213551701991006, |
|
"loss": 0.118, |
|
"step": 12330 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 0.00010197495183044315, |
|
"loss": 0.1281, |
|
"step": 12340 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 0.00010181438664097623, |
|
"loss": 0.1169, |
|
"step": 12350 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 0.00010165382145150931, |
|
"loss": 0.1202, |
|
"step": 12360 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 0.00010149325626204237, |
|
"loss": 0.114, |
|
"step": 12370 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 0.00010133269107257545, |
|
"loss": 0.1061, |
|
"step": 12380 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 0.00010117212588310854, |
|
"loss": 0.1157, |
|
"step": 12390 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 0.00010101156069364161, |
|
"loss": 0.1162, |
|
"step": 12400 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 0.00010085099550417468, |
|
"loss": 0.1234, |
|
"step": 12410 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 0.00010069043031470776, |
|
"loss": 0.1368, |
|
"step": 12420 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 0.00010052986512524083, |
|
"loss": 0.1305, |
|
"step": 12430 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 0.00010036929993577392, |
|
"loss": 0.0978, |
|
"step": 12440 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 0.000100208734746307, |
|
"loss": 0.128, |
|
"step": 12450 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 0.00010004816955684007, |
|
"loss": 0.1119, |
|
"step": 12460 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 9.988760436737314e-05, |
|
"loss": 0.1238, |
|
"step": 12470 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 9.972703917790622e-05, |
|
"loss": 0.1357, |
|
"step": 12480 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 9.95664739884393e-05, |
|
"loss": 0.106, |
|
"step": 12490 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 9.940590879897238e-05, |
|
"loss": 0.1098, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 9.924534360950545e-05, |
|
"loss": 0.1173, |
|
"step": 12510 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 9.908477842003853e-05, |
|
"loss": 0.1318, |
|
"step": 12520 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 9.89242132305716e-05, |
|
"loss": 0.131, |
|
"step": 12530 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 9.876364804110468e-05, |
|
"loss": 0.13, |
|
"step": 12540 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 9.860308285163776e-05, |
|
"loss": 0.1739, |
|
"step": 12550 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 9.844251766217084e-05, |
|
"loss": 0.0962, |
|
"step": 12560 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 9.82819524727039e-05, |
|
"loss": 0.1135, |
|
"step": 12570 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 9.812138728323698e-05, |
|
"loss": 0.1068, |
|
"step": 12580 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 9.796082209377007e-05, |
|
"loss": 0.1071, |
|
"step": 12590 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 9.780025690430315e-05, |
|
"loss": 0.1186, |
|
"step": 12600 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 9.763969171483621e-05, |
|
"loss": 0.1195, |
|
"step": 12610 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 9.747912652536929e-05, |
|
"loss": 0.118, |
|
"step": 12620 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 9.731856133590237e-05, |
|
"loss": 0.1419, |
|
"step": 12630 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 9.715799614643545e-05, |
|
"loss": 0.1279, |
|
"step": 12640 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 9.699743095696852e-05, |
|
"loss": 0.1029, |
|
"step": 12650 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 9.68368657675016e-05, |
|
"loss": 0.1166, |
|
"step": 12660 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 9.667630057803466e-05, |
|
"loss": 0.0994, |
|
"step": 12670 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 9.651573538856774e-05, |
|
"loss": 0.119, |
|
"step": 12680 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 9.635517019910083e-05, |
|
"loss": 0.1454, |
|
"step": 12690 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 9.619460500963391e-05, |
|
"loss": 0.115, |
|
"step": 12700 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 9.603403982016698e-05, |
|
"loss": 0.1148, |
|
"step": 12710 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 9.587347463070005e-05, |
|
"loss": 0.1317, |
|
"step": 12720 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 9.571290944123313e-05, |
|
"loss": 0.1138, |
|
"step": 12730 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 9.555234425176622e-05, |
|
"loss": 0.1303, |
|
"step": 12740 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 9.539177906229929e-05, |
|
"loss": 0.1056, |
|
"step": 12750 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 9.523121387283236e-05, |
|
"loss": 0.1023, |
|
"step": 12760 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 9.507064868336544e-05, |
|
"loss": 0.1247, |
|
"step": 12770 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 9.49100834938985e-05, |
|
"loss": 0.1308, |
|
"step": 12780 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 9.47495183044316e-05, |
|
"loss": 0.1406, |
|
"step": 12790 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 9.458895311496467e-05, |
|
"loss": 0.1293, |
|
"step": 12800 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 9.442838792549774e-05, |
|
"loss": 0.1336, |
|
"step": 12810 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 9.426782273603082e-05, |
|
"loss": 0.1266, |
|
"step": 12820 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 9.41072575465639e-05, |
|
"loss": 0.1216, |
|
"step": 12830 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 9.394669235709699e-05, |
|
"loss": 0.0926, |
|
"step": 12840 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 9.378612716763005e-05, |
|
"loss": 0.113, |
|
"step": 12850 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 9.362556197816313e-05, |
|
"loss": 0.1017, |
|
"step": 12860 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 9.34649967886962e-05, |
|
"loss": 0.0958, |
|
"step": 12870 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 9.330443159922927e-05, |
|
"loss": 0.1039, |
|
"step": 12880 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 9.314386640976236e-05, |
|
"loss": 0.1146, |
|
"step": 12890 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 9.298330122029544e-05, |
|
"loss": 0.1076, |
|
"step": 12900 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 9.282273603082852e-05, |
|
"loss": 0.1533, |
|
"step": 12910 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 9.266217084136158e-05, |
|
"loss": 0.1285, |
|
"step": 12920 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 9.250160565189466e-05, |
|
"loss": 0.1328, |
|
"step": 12930 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 9.234104046242775e-05, |
|
"loss": 0.0996, |
|
"step": 12940 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 9.218047527296081e-05, |
|
"loss": 0.1316, |
|
"step": 12950 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 9.201991008349389e-05, |
|
"loss": 0.1259, |
|
"step": 12960 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 9.185934489402697e-05, |
|
"loss": 0.1406, |
|
"step": 12970 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 9.169877970456003e-05, |
|
"loss": 0.1175, |
|
"step": 12980 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 9.153821451509312e-05, |
|
"loss": 0.1066, |
|
"step": 12990 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 9.13776493256262e-05, |
|
"loss": 0.1201, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"eval_cer": 0.9422756881449822, |
|
"eval_loss": 0.11346502602100372, |
|
"eval_runtime": 53.7978, |
|
"eval_samples_per_second": 18.588, |
|
"eval_steps_per_second": 4.647, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 9.121708413615928e-05, |
|
"loss": 0.1418, |
|
"step": 13010 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 9.105651894669234e-05, |
|
"loss": 0.0952, |
|
"step": 13020 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 9.089595375722542e-05, |
|
"loss": 0.158, |
|
"step": 13030 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 9.073538856775851e-05, |
|
"loss": 0.1106, |
|
"step": 13040 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 9.057482337829158e-05, |
|
"loss": 0.1044, |
|
"step": 13050 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 9.041425818882465e-05, |
|
"loss": 0.1143, |
|
"step": 13060 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 9.025369299935773e-05, |
|
"loss": 0.0907, |
|
"step": 13070 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 9.00931278098908e-05, |
|
"loss": 0.1195, |
|
"step": 13080 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 8.993256262042389e-05, |
|
"loss": 0.1292, |
|
"step": 13090 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 8.977199743095697e-05, |
|
"loss": 0.1281, |
|
"step": 13100 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 8.961143224149004e-05, |
|
"loss": 0.1131, |
|
"step": 13110 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 8.945086705202311e-05, |
|
"loss": 0.0931, |
|
"step": 13120 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 8.929030186255619e-05, |
|
"loss": 0.1247, |
|
"step": 13130 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 8.912973667308926e-05, |
|
"loss": 0.1407, |
|
"step": 13140 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 8.896917148362235e-05, |
|
"loss": 0.1199, |
|
"step": 13150 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 8.880860629415542e-05, |
|
"loss": 0.1189, |
|
"step": 13160 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 8.86480411046885e-05, |
|
"loss": 0.1624, |
|
"step": 13170 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 8.848747591522157e-05, |
|
"loss": 0.0905, |
|
"step": 13180 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 8.832691072575464e-05, |
|
"loss": 0.1173, |
|
"step": 13190 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 8.816634553628773e-05, |
|
"loss": 0.1663, |
|
"step": 13200 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 8.800578034682081e-05, |
|
"loss": 0.1127, |
|
"step": 13210 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 8.784521515735387e-05, |
|
"loss": 0.113, |
|
"step": 13220 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 8.768464996788695e-05, |
|
"loss": 0.107, |
|
"step": 13230 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 8.752408477842003e-05, |
|
"loss": 0.1051, |
|
"step": 13240 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 8.736351958895312e-05, |
|
"loss": 0.1251, |
|
"step": 13250 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 8.720295439948618e-05, |
|
"loss": 0.1117, |
|
"step": 13260 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 8.704238921001926e-05, |
|
"loss": 0.1304, |
|
"step": 13270 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 8.688182402055234e-05, |
|
"loss": 0.1114, |
|
"step": 13280 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 8.67212588310854e-05, |
|
"loss": 0.1194, |
|
"step": 13290 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 8.656069364161849e-05, |
|
"loss": 0.1096, |
|
"step": 13300 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 8.640012845215157e-05, |
|
"loss": 0.1263, |
|
"step": 13310 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 8.623956326268463e-05, |
|
"loss": 0.0917, |
|
"step": 13320 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 8.607899807321771e-05, |
|
"loss": 0.097, |
|
"step": 13330 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 8.591843288375079e-05, |
|
"loss": 0.1079, |
|
"step": 13340 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 8.575786769428388e-05, |
|
"loss": 0.1127, |
|
"step": 13350 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 8.559730250481695e-05, |
|
"loss": 0.1152, |
|
"step": 13360 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 8.543673731535002e-05, |
|
"loss": 0.1071, |
|
"step": 13370 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 8.52761721258831e-05, |
|
"loss": 0.0888, |
|
"step": 13380 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 8.511560693641617e-05, |
|
"loss": 0.1302, |
|
"step": 13390 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 8.495504174694926e-05, |
|
"loss": 0.1409, |
|
"step": 13400 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 8.479447655748233e-05, |
|
"loss": 0.1122, |
|
"step": 13410 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 8.463391136801541e-05, |
|
"loss": 0.1108, |
|
"step": 13420 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 8.447334617854848e-05, |
|
"loss": 0.1099, |
|
"step": 13430 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 8.431278098908155e-05, |
|
"loss": 0.1286, |
|
"step": 13440 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 8.415221579961465e-05, |
|
"loss": 0.1205, |
|
"step": 13450 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 8.399165061014771e-05, |
|
"loss": 0.1477, |
|
"step": 13460 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 8.383108542068079e-05, |
|
"loss": 0.1394, |
|
"step": 13470 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 8.367052023121386e-05, |
|
"loss": 0.1222, |
|
"step": 13480 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 8.350995504174693e-05, |
|
"loss": 0.1115, |
|
"step": 13490 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 8.334938985228002e-05, |
|
"loss": 0.0904, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 8.31888246628131e-05, |
|
"loss": 0.1242, |
|
"step": 13510 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 8.302825947334618e-05, |
|
"loss": 0.0992, |
|
"step": 13520 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 8.286769428387924e-05, |
|
"loss": 0.1309, |
|
"step": 13530 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 8.270712909441232e-05, |
|
"loss": 0.1019, |
|
"step": 13540 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 8.254656390494541e-05, |
|
"loss": 0.1237, |
|
"step": 13550 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 8.238599871547847e-05, |
|
"loss": 0.1086, |
|
"step": 13560 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 8.222543352601155e-05, |
|
"loss": 0.1267, |
|
"step": 13570 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 8.206486833654463e-05, |
|
"loss": 0.1214, |
|
"step": 13580 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 8.190430314707769e-05, |
|
"loss": 0.1349, |
|
"step": 13590 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 8.174373795761078e-05, |
|
"loss": 0.1243, |
|
"step": 13600 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 8.158317276814386e-05, |
|
"loss": 0.123, |
|
"step": 13610 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 8.142260757867694e-05, |
|
"loss": 0.1232, |
|
"step": 13620 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 8.126204238921e-05, |
|
"loss": 0.11, |
|
"step": 13630 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 8.110147719974308e-05, |
|
"loss": 0.1073, |
|
"step": 13640 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 8.094091201027617e-05, |
|
"loss": 0.1017, |
|
"step": 13650 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 8.078034682080925e-05, |
|
"loss": 0.0993, |
|
"step": 13660 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 8.061978163134231e-05, |
|
"loss": 0.1092, |
|
"step": 13670 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 8.045921644187539e-05, |
|
"loss": 0.0962, |
|
"step": 13680 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 8.029865125240847e-05, |
|
"loss": 0.098, |
|
"step": 13690 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 8.013808606294155e-05, |
|
"loss": 0.1321, |
|
"step": 13700 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 7.997752087347462e-05, |
|
"loss": 0.1057, |
|
"step": 13710 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 7.98169556840077e-05, |
|
"loss": 0.112, |
|
"step": 13720 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 7.965639049454077e-05, |
|
"loss": 0.1497, |
|
"step": 13730 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 7.949582530507384e-05, |
|
"loss": 0.1073, |
|
"step": 13740 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 7.933526011560694e-05, |
|
"loss": 0.1296, |
|
"step": 13750 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 7.917469492614001e-05, |
|
"loss": 0.1185, |
|
"step": 13760 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 7.901412973667308e-05, |
|
"loss": 0.0927, |
|
"step": 13770 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 7.885356454720616e-05, |
|
"loss": 0.0931, |
|
"step": 13780 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 7.869299935773923e-05, |
|
"loss": 0.098, |
|
"step": 13790 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 7.853243416827232e-05, |
|
"loss": 0.1036, |
|
"step": 13800 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 7.837186897880539e-05, |
|
"loss": 0.124, |
|
"step": 13810 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 7.821130378933847e-05, |
|
"loss": 0.0928, |
|
"step": 13820 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 7.805073859987154e-05, |
|
"loss": 0.1172, |
|
"step": 13830 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 7.789017341040461e-05, |
|
"loss": 0.0884, |
|
"step": 13840 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 7.77296082209377e-05, |
|
"loss": 0.1069, |
|
"step": 13850 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 7.756904303147078e-05, |
|
"loss": 0.0936, |
|
"step": 13860 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 7.740847784200384e-05, |
|
"loss": 0.1041, |
|
"step": 13870 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 7.724791265253692e-05, |
|
"loss": 0.121, |
|
"step": 13880 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 7.708734746307e-05, |
|
"loss": 0.1032, |
|
"step": 13890 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 7.692678227360309e-05, |
|
"loss": 0.1275, |
|
"step": 13900 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 7.676621708413615e-05, |
|
"loss": 0.1451, |
|
"step": 13910 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 7.660565189466923e-05, |
|
"loss": 0.1224, |
|
"step": 13920 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 7.644508670520231e-05, |
|
"loss": 0.1087, |
|
"step": 13930 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 7.628452151573537e-05, |
|
"loss": 0.1409, |
|
"step": 13940 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 7.612395632626846e-05, |
|
"loss": 0.111, |
|
"step": 13950 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 7.596339113680154e-05, |
|
"loss": 0.1126, |
|
"step": 13960 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 7.58028259473346e-05, |
|
"loss": 0.1066, |
|
"step": 13970 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 7.564226075786768e-05, |
|
"loss": 0.1487, |
|
"step": 13980 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 7.548169556840076e-05, |
|
"loss": 0.1041, |
|
"step": 13990 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 7.532113037893385e-05, |
|
"loss": 0.1115, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"eval_cer": 0.942318965641602, |
|
"eval_loss": 0.1111316978931427, |
|
"eval_runtime": 52.79, |
|
"eval_samples_per_second": 18.943, |
|
"eval_steps_per_second": 4.736, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 7.516056518946692e-05, |
|
"loss": 0.1132, |
|
"step": 14010 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 7.5e-05, |
|
"loss": 0.1062, |
|
"step": 14020 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 7.483943481053307e-05, |
|
"loss": 0.148, |
|
"step": 14030 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 7.467886962106615e-05, |
|
"loss": 0.1129, |
|
"step": 14040 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 7.451830443159923e-05, |
|
"loss": 0.1125, |
|
"step": 14050 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 7.43577392421323e-05, |
|
"loss": 0.1099, |
|
"step": 14060 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 7.419717405266538e-05, |
|
"loss": 0.1069, |
|
"step": 14070 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 7.403660886319845e-05, |
|
"loss": 0.1113, |
|
"step": 14080 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 7.387604367373152e-05, |
|
"loss": 0.1137, |
|
"step": 14090 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 7.37154784842646e-05, |
|
"loss": 0.1248, |
|
"step": 14100 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 7.355491329479768e-05, |
|
"loss": 0.1208, |
|
"step": 14110 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 7.339434810533076e-05, |
|
"loss": 0.1152, |
|
"step": 14120 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 7.323378291586383e-05, |
|
"loss": 0.1303, |
|
"step": 14130 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 7.307321772639691e-05, |
|
"loss": 0.1095, |
|
"step": 14140 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 7.291265253692999e-05, |
|
"loss": 0.1175, |
|
"step": 14150 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 7.275208734746307e-05, |
|
"loss": 0.1294, |
|
"step": 14160 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 7.259152215799615e-05, |
|
"loss": 0.1081, |
|
"step": 14170 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 7.243095696852921e-05, |
|
"loss": 0.1188, |
|
"step": 14180 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 7.22703917790623e-05, |
|
"loss": 0.1207, |
|
"step": 14190 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 7.210982658959537e-05, |
|
"loss": 0.1054, |
|
"step": 14200 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 7.194926140012844e-05, |
|
"loss": 0.0846, |
|
"step": 14210 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 7.178869621066152e-05, |
|
"loss": 0.0919, |
|
"step": 14220 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 7.16281310211946e-05, |
|
"loss": 0.1157, |
|
"step": 14230 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 7.146756583172768e-05, |
|
"loss": 0.1248, |
|
"step": 14240 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 7.130700064226075e-05, |
|
"loss": 0.1129, |
|
"step": 14250 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 7.114643545279383e-05, |
|
"loss": 0.1139, |
|
"step": 14260 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 7.10019267822736e-05, |
|
"loss": 0.1211, |
|
"step": 14270 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 7.084136159280668e-05, |
|
"loss": 0.1144, |
|
"step": 14280 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 7.068079640333975e-05, |
|
"loss": 0.0817, |
|
"step": 14290 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 7.052023121387283e-05, |
|
"loss": 0.1169, |
|
"step": 14300 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 7.03596660244059e-05, |
|
"loss": 0.1192, |
|
"step": 14310 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 7.019910083493899e-05, |
|
"loss": 0.1275, |
|
"step": 14320 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 7.003853564547205e-05, |
|
"loss": 0.1125, |
|
"step": 14330 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 6.987797045600513e-05, |
|
"loss": 0.1352, |
|
"step": 14340 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 6.971740526653821e-05, |
|
"loss": 0.0957, |
|
"step": 14350 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 6.955684007707128e-05, |
|
"loss": 0.1357, |
|
"step": 14360 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 6.939627488760436e-05, |
|
"loss": 0.1141, |
|
"step": 14370 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 6.923570969813744e-05, |
|
"loss": 0.1148, |
|
"step": 14380 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 6.90751445086705e-05, |
|
"loss": 0.1114, |
|
"step": 14390 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 6.89145793192036e-05, |
|
"loss": 0.1071, |
|
"step": 14400 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 6.875401412973666e-05, |
|
"loss": 0.1179, |
|
"step": 14410 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 6.859344894026975e-05, |
|
"loss": 0.1379, |
|
"step": 14420 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 6.843288375080282e-05, |
|
"loss": 0.1073, |
|
"step": 14430 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 6.827231856133589e-05, |
|
"loss": 0.1202, |
|
"step": 14440 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 6.811175337186897e-05, |
|
"loss": 0.1093, |
|
"step": 14450 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 6.795118818240205e-05, |
|
"loss": 0.1078, |
|
"step": 14460 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 6.779062299293513e-05, |
|
"loss": 0.1203, |
|
"step": 14470 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 6.76300578034682e-05, |
|
"loss": 0.1339, |
|
"step": 14480 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 6.746949261400128e-05, |
|
"loss": 0.0905, |
|
"step": 14490 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 6.730892742453436e-05, |
|
"loss": 0.1199, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 6.714836223506742e-05, |
|
"loss": 0.1221, |
|
"step": 14510 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 6.698779704560051e-05, |
|
"loss": 0.1032, |
|
"step": 14520 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 6.682723185613358e-05, |
|
"loss": 0.1105, |
|
"step": 14530 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 6.666666666666666e-05, |
|
"loss": 0.1144, |
|
"step": 14540 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 6.650610147719973e-05, |
|
"loss": 0.1185, |
|
"step": 14550 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 6.634553628773281e-05, |
|
"loss": 0.1535, |
|
"step": 14560 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 6.618497109826589e-05, |
|
"loss": 0.1108, |
|
"step": 14570 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 6.602440590879897e-05, |
|
"loss": 0.1361, |
|
"step": 14580 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 6.586384071933204e-05, |
|
"loss": 0.0929, |
|
"step": 14590 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 6.570327552986512e-05, |
|
"loss": 0.1102, |
|
"step": 14600 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 6.55427103403982e-05, |
|
"loss": 0.0789, |
|
"step": 14610 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 6.538214515093128e-05, |
|
"loss": 0.1139, |
|
"step": 14620 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 6.522157996146434e-05, |
|
"loss": 0.1197, |
|
"step": 14630 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 6.506101477199742e-05, |
|
"loss": 0.1174, |
|
"step": 14640 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 6.49004495825305e-05, |
|
"loss": 0.1015, |
|
"step": 14650 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 6.473988439306358e-05, |
|
"loss": 0.1131, |
|
"step": 14660 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 6.457931920359665e-05, |
|
"loss": 0.1256, |
|
"step": 14670 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 6.441875401412973e-05, |
|
"loss": 0.1219, |
|
"step": 14680 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 6.425818882466281e-05, |
|
"loss": 0.1097, |
|
"step": 14690 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 6.409762363519589e-05, |
|
"loss": 0.1084, |
|
"step": 14700 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 6.393705844572896e-05, |
|
"loss": 0.1248, |
|
"step": 14710 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 6.377649325626204e-05, |
|
"loss": 0.1085, |
|
"step": 14720 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 6.361592806679512e-05, |
|
"loss": 0.1329, |
|
"step": 14730 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 6.345536287732818e-05, |
|
"loss": 0.1205, |
|
"step": 14740 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 6.329479768786126e-05, |
|
"loss": 0.1321, |
|
"step": 14750 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 6.313423249839434e-05, |
|
"loss": 0.1434, |
|
"step": 14760 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 6.297366730892742e-05, |
|
"loss": 0.112, |
|
"step": 14770 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 6.28131021194605e-05, |
|
"loss": 0.1453, |
|
"step": 14780 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 6.265253692999357e-05, |
|
"loss": 0.1162, |
|
"step": 14790 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 6.249197174052665e-05, |
|
"loss": 0.1215, |
|
"step": 14800 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 6.233140655105973e-05, |
|
"loss": 0.1115, |
|
"step": 14810 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 6.21708413615928e-05, |
|
"loss": 0.131, |
|
"step": 14820 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 6.201027617212588e-05, |
|
"loss": 0.0925, |
|
"step": 14830 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 6.184971098265895e-05, |
|
"loss": 0.1278, |
|
"step": 14840 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 6.168914579319204e-05, |
|
"loss": 0.1203, |
|
"step": 14850 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 6.15285806037251e-05, |
|
"loss": 0.098, |
|
"step": 14860 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 6.13680154142582e-05, |
|
"loss": 0.1003, |
|
"step": 14870 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 6.120745022479126e-05, |
|
"loss": 0.1363, |
|
"step": 14880 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 6.104688503532434e-05, |
|
"loss": 0.1315, |
|
"step": 14890 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 6.088631984585741e-05, |
|
"loss": 0.1071, |
|
"step": 14900 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 6.072575465639049e-05, |
|
"loss": 0.1178, |
|
"step": 14910 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 6.056518946692357e-05, |
|
"loss": 0.0883, |
|
"step": 14920 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 6.040462427745664e-05, |
|
"loss": 0.1203, |
|
"step": 14930 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 6.024405908798972e-05, |
|
"loss": 0.0994, |
|
"step": 14940 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 6.0083493898522795e-05, |
|
"loss": 0.1157, |
|
"step": 14950 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 5.992292870905587e-05, |
|
"loss": 0.0976, |
|
"step": 14960 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 5.976236351958895e-05, |
|
"loss": 0.0878, |
|
"step": 14970 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 5.960179833012202e-05, |
|
"loss": 0.1364, |
|
"step": 14980 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 5.94412331406551e-05, |
|
"loss": 0.1029, |
|
"step": 14990 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 5.928066795118818e-05, |
|
"loss": 0.1284, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"eval_cer": 0.9422271648911964, |
|
"eval_loss": 0.11006263643503189, |
|
"eval_runtime": 53.6901, |
|
"eval_samples_per_second": 18.625, |
|
"eval_steps_per_second": 4.656, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 5.9120102761721255e-05, |
|
"loss": 0.127, |
|
"step": 15010 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 5.895953757225433e-05, |
|
"loss": 0.1323, |
|
"step": 15020 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 5.879897238278741e-05, |
|
"loss": 0.1159, |
|
"step": 15030 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 5.863840719332048e-05, |
|
"loss": 0.1048, |
|
"step": 15040 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 5.847784200385356e-05, |
|
"loss": 0.1117, |
|
"step": 15050 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 5.8317276814386636e-05, |
|
"loss": 0.1116, |
|
"step": 15060 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 5.8156711624919714e-05, |
|
"loss": 0.0893, |
|
"step": 15070 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 5.799614643545279e-05, |
|
"loss": 0.1263, |
|
"step": 15080 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 5.783558124598586e-05, |
|
"loss": 0.0997, |
|
"step": 15090 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 5.767501605651895e-05, |
|
"loss": 0.0947, |
|
"step": 15100 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 5.751445086705202e-05, |
|
"loss": 0.1168, |
|
"step": 15110 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 5.7353885677585096e-05, |
|
"loss": 0.1406, |
|
"step": 15120 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 5.7193320488118174e-05, |
|
"loss": 0.1103, |
|
"step": 15130 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 5.7032755298651245e-05, |
|
"loss": 0.0813, |
|
"step": 15140 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 5.687219010918433e-05, |
|
"loss": 0.1409, |
|
"step": 15150 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 5.67116249197174e-05, |
|
"loss": 0.1159, |
|
"step": 15160 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 5.655105973025048e-05, |
|
"loss": 0.0965, |
|
"step": 15170 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 5.6390494540783555e-05, |
|
"loss": 0.1093, |
|
"step": 15180 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 5.6229929351316626e-05, |
|
"loss": 0.0891, |
|
"step": 15190 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 5.606936416184971e-05, |
|
"loss": 0.1111, |
|
"step": 15200 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 5.590879897238278e-05, |
|
"loss": 0.1179, |
|
"step": 15210 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 5.574823378291585e-05, |
|
"loss": 0.1003, |
|
"step": 15220 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 5.558766859344894e-05, |
|
"loss": 0.1143, |
|
"step": 15230 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 5.542710340398201e-05, |
|
"loss": 0.0968, |
|
"step": 15240 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 5.526653821451509e-05, |
|
"loss": 0.1236, |
|
"step": 15250 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 5.5105973025048164e-05, |
|
"loss": 0.1191, |
|
"step": 15260 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 5.4945407835581235e-05, |
|
"loss": 0.1276, |
|
"step": 15270 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 5.478484264611432e-05, |
|
"loss": 0.1302, |
|
"step": 15280 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 5.462427745664739e-05, |
|
"loss": 0.1205, |
|
"step": 15290 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 5.4463712267180474e-05, |
|
"loss": 0.0759, |
|
"step": 15300 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 5.4303147077713545e-05, |
|
"loss": 0.0967, |
|
"step": 15310 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 5.4142581888246616e-05, |
|
"loss": 0.1443, |
|
"step": 15320 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 5.39820166987797e-05, |
|
"loss": 0.0762, |
|
"step": 15330 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 5.382145150931277e-05, |
|
"loss": 0.1063, |
|
"step": 15340 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 5.3660886319845856e-05, |
|
"loss": 0.0972, |
|
"step": 15350 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 5.350032113037893e-05, |
|
"loss": 0.134, |
|
"step": 15360 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 5.3339755940912005e-05, |
|
"loss": 0.1279, |
|
"step": 15370 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 5.317919075144508e-05, |
|
"loss": 0.124, |
|
"step": 15380 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 5.3018625561978154e-05, |
|
"loss": 0.0826, |
|
"step": 15390 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 5.285806037251124e-05, |
|
"loss": 0.0884, |
|
"step": 15400 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 5.269749518304431e-05, |
|
"loss": 0.1102, |
|
"step": 15410 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 5.253692999357739e-05, |
|
"loss": 0.0967, |
|
"step": 15420 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 5.2376364804110464e-05, |
|
"loss": 0.1159, |
|
"step": 15430 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 5.2215799614643535e-05, |
|
"loss": 0.0779, |
|
"step": 15440 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 5.205523442517662e-05, |
|
"loss": 0.1137, |
|
"step": 15450 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 5.189466923570969e-05, |
|
"loss": 0.1441, |
|
"step": 15460 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 5.173410404624277e-05, |
|
"loss": 0.098, |
|
"step": 15470 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 5.1573538856775846e-05, |
|
"loss": 0.0864, |
|
"step": 15480 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 5.1412973667308924e-05, |
|
"loss": 0.1088, |
|
"step": 15490 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 5.1252408477842e-05, |
|
"loss": 0.1312, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 5.109184328837507e-05, |
|
"loss": 0.087, |
|
"step": 15510 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 5.093127809890815e-05, |
|
"loss": 0.108, |
|
"step": 15520 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 5.077071290944123e-05, |
|
"loss": 0.1232, |
|
"step": 15530 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 5.0610147719974306e-05, |
|
"loss": 0.0775, |
|
"step": 15540 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 5.0449582530507383e-05, |
|
"loss": 0.1233, |
|
"step": 15550 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 5.028901734104046e-05, |
|
"loss": 0.1048, |
|
"step": 15560 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 5.012845215157353e-05, |
|
"loss": 0.1056, |
|
"step": 15570 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 4.996788696210661e-05, |
|
"loss": 0.1008, |
|
"step": 15580 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 4.980732177263969e-05, |
|
"loss": 0.0913, |
|
"step": 15590 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 4.9646756583172765e-05, |
|
"loss": 0.1192, |
|
"step": 15600 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 4.948619139370584e-05, |
|
"loss": 0.0912, |
|
"step": 15610 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 4.9325626204238914e-05, |
|
"loss": 0.1214, |
|
"step": 15620 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 4.916506101477199e-05, |
|
"loss": 0.1002, |
|
"step": 15630 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 4.900449582530507e-05, |
|
"loss": 0.1194, |
|
"step": 15640 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 4.884393063583815e-05, |
|
"loss": 0.1188, |
|
"step": 15650 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 4.8683365446371225e-05, |
|
"loss": 0.1414, |
|
"step": 15660 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 4.8522800256904296e-05, |
|
"loss": 0.1366, |
|
"step": 15670 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 4.836223506743738e-05, |
|
"loss": 0.0904, |
|
"step": 15680 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 4.820166987797045e-05, |
|
"loss": 0.0985, |
|
"step": 15690 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 4.804110468850353e-05, |
|
"loss": 0.1359, |
|
"step": 15700 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 4.7880539499036607e-05, |
|
"loss": 0.1223, |
|
"step": 15710 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 4.771997430956968e-05, |
|
"loss": 0.1185, |
|
"step": 15720 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 4.755940912010276e-05, |
|
"loss": 0.1452, |
|
"step": 15730 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 4.739884393063583e-05, |
|
"loss": 0.1073, |
|
"step": 15740 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 4.723827874116891e-05, |
|
"loss": 0.127, |
|
"step": 15750 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 4.707771355170199e-05, |
|
"loss": 0.1225, |
|
"step": 15760 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 4.691714836223506e-05, |
|
"loss": 0.1317, |
|
"step": 15770 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 4.6756583172768144e-05, |
|
"loss": 0.0955, |
|
"step": 15780 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 4.6596017983301215e-05, |
|
"loss": 0.1104, |
|
"step": 15790 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 4.64354527938343e-05, |
|
"loss": 0.0951, |
|
"step": 15800 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 4.627488760436737e-05, |
|
"loss": 0.1183, |
|
"step": 15810 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 4.611432241490044e-05, |
|
"loss": 0.1017, |
|
"step": 15820 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 4.5953757225433526e-05, |
|
"loss": 0.091, |
|
"step": 15830 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 4.5793192035966596e-05, |
|
"loss": 0.115, |
|
"step": 15840 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 4.563262684649968e-05, |
|
"loss": 0.1105, |
|
"step": 15850 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 4.547206165703275e-05, |
|
"loss": 0.1203, |
|
"step": 15860 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 4.531149646756582e-05, |
|
"loss": 0.129, |
|
"step": 15870 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 4.515093127809891e-05, |
|
"loss": 0.1048, |
|
"step": 15880 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 4.499036608863198e-05, |
|
"loss": 0.135, |
|
"step": 15890 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 4.482980089916506e-05, |
|
"loss": 0.1113, |
|
"step": 15900 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 4.4669235709698134e-05, |
|
"loss": 0.0992, |
|
"step": 15910 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 4.4508670520231205e-05, |
|
"loss": 0.1455, |
|
"step": 15920 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 4.434810533076429e-05, |
|
"loss": 0.0974, |
|
"step": 15930 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 4.418754014129736e-05, |
|
"loss": 0.0973, |
|
"step": 15940 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 4.4026974951830445e-05, |
|
"loss": 0.1006, |
|
"step": 15950 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 4.3866409762363516e-05, |
|
"loss": 0.1108, |
|
"step": 15960 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 4.3705844572896586e-05, |
|
"loss": 0.1096, |
|
"step": 15970 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 4.354527938342967e-05, |
|
"loss": 0.0909, |
|
"step": 15980 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 4.338471419396274e-05, |
|
"loss": 0.1206, |
|
"step": 15990 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 4.3224149004495826e-05, |
|
"loss": 0.0947, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"eval_cer": 0.9422245420126134, |
|
"eval_loss": 0.10853888839483261, |
|
"eval_runtime": 52.65, |
|
"eval_samples_per_second": 18.993, |
|
"eval_steps_per_second": 4.748, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 4.30635838150289e-05, |
|
"loss": 0.1094, |
|
"step": 16010 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 4.2903018625561975e-05, |
|
"loss": 0.1165, |
|
"step": 16020 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 4.274245343609505e-05, |
|
"loss": 0.1019, |
|
"step": 16030 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 4.2581888246628124e-05, |
|
"loss": 0.0833, |
|
"step": 16040 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 4.242132305716121e-05, |
|
"loss": 0.1146, |
|
"step": 16050 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 4.226075786769428e-05, |
|
"loss": 0.1102, |
|
"step": 16060 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 4.210019267822736e-05, |
|
"loss": 0.1155, |
|
"step": 16070 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 4.1939627488760435e-05, |
|
"loss": 0.0977, |
|
"step": 16080 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 4.1779062299293505e-05, |
|
"loss": 0.1131, |
|
"step": 16090 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 4.161849710982658e-05, |
|
"loss": 0.1458, |
|
"step": 16100 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 4.145793192035966e-05, |
|
"loss": 0.1299, |
|
"step": 16110 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 4.129736673089274e-05, |
|
"loss": 0.0809, |
|
"step": 16120 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 4.1136801541425816e-05, |
|
"loss": 0.0834, |
|
"step": 16130 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 4.0976236351958894e-05, |
|
"loss": 0.1422, |
|
"step": 16140 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 4.0815671162491965e-05, |
|
"loss": 0.1069, |
|
"step": 16150 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 4.065510597302504e-05, |
|
"loss": 0.1079, |
|
"step": 16160 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 4.049454078355812e-05, |
|
"loss": 0.0991, |
|
"step": 16170 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 4.03339755940912e-05, |
|
"loss": 0.0954, |
|
"step": 16180 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 4.0173410404624276e-05, |
|
"loss": 0.1153, |
|
"step": 16190 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 4.001284521515735e-05, |
|
"loss": 0.0932, |
|
"step": 16200 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 3.985228002569043e-05, |
|
"loss": 0.0851, |
|
"step": 16210 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 3.96917148362235e-05, |
|
"loss": 0.098, |
|
"step": 16220 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 3.953114964675658e-05, |
|
"loss": 0.1491, |
|
"step": 16230 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 3.937058445728966e-05, |
|
"loss": 0.1197, |
|
"step": 16240 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 3.921001926782273e-05, |
|
"loss": 0.0998, |
|
"step": 16250 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 3.904945407835581e-05, |
|
"loss": 0.1164, |
|
"step": 16260 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 3.8888888888888884e-05, |
|
"loss": 0.0882, |
|
"step": 16270 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 3.872832369942196e-05, |
|
"loss": 0.1025, |
|
"step": 16280 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 3.856775850995504e-05, |
|
"loss": 0.1193, |
|
"step": 16290 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 3.840719332048811e-05, |
|
"loss": 0.1463, |
|
"step": 16300 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 3.8246628131021195e-05, |
|
"loss": 0.0811, |
|
"step": 16310 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 3.8086062941554266e-05, |
|
"loss": 0.1075, |
|
"step": 16320 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 3.792549775208735e-05, |
|
"loss": 0.0959, |
|
"step": 16330 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 3.776493256262042e-05, |
|
"loss": 0.1149, |
|
"step": 16340 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 3.760436737315349e-05, |
|
"loss": 0.1172, |
|
"step": 16350 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 3.744380218368658e-05, |
|
"loss": 0.0908, |
|
"step": 16360 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 3.728323699421965e-05, |
|
"loss": 0.1142, |
|
"step": 16370 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 3.7122671804752725e-05, |
|
"loss": 0.0922, |
|
"step": 16380 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 3.69621066152858e-05, |
|
"loss": 0.1347, |
|
"step": 16390 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 3.680154142581888e-05, |
|
"loss": 0.1046, |
|
"step": 16400 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 3.664097623635196e-05, |
|
"loss": 0.0856, |
|
"step": 16410 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 3.648041104688503e-05, |
|
"loss": 0.1176, |
|
"step": 16420 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 3.631984585741811e-05, |
|
"loss": 0.1157, |
|
"step": 16430 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 3.6159280667951185e-05, |
|
"loss": 0.0953, |
|
"step": 16440 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 3.599871547848426e-05, |
|
"loss": 0.1246, |
|
"step": 16450 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 3.583815028901734e-05, |
|
"loss": 0.1143, |
|
"step": 16460 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 3.567758509955041e-05, |
|
"loss": 0.1106, |
|
"step": 16470 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 3.551701991008349e-05, |
|
"loss": 0.1065, |
|
"step": 16480 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 3.5356454720616567e-05, |
|
"loss": 0.0998, |
|
"step": 16490 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 3.5195889531149644e-05, |
|
"loss": 0.1224, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 3.503532434168272e-05, |
|
"loss": 0.1208, |
|
"step": 16510 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 3.487475915221579e-05, |
|
"loss": 0.1038, |
|
"step": 16520 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 3.471419396274887e-05, |
|
"loss": 0.1169, |
|
"step": 16530 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 3.455362877328195e-05, |
|
"loss": 0.1366, |
|
"step": 16540 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 3.4393063583815026e-05, |
|
"loss": 0.1132, |
|
"step": 16550 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 3.4232498394348104e-05, |
|
"loss": 0.1224, |
|
"step": 16560 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 3.4071933204881175e-05, |
|
"loss": 0.108, |
|
"step": 16570 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 3.391136801541425e-05, |
|
"loss": 0.1011, |
|
"step": 16580 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 3.375080282594733e-05, |
|
"loss": 0.1055, |
|
"step": 16590 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 3.359023763648041e-05, |
|
"loss": 0.098, |
|
"step": 16600 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 3.3429672447013486e-05, |
|
"loss": 0.1738, |
|
"step": 16610 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 3.3269107257546557e-05, |
|
"loss": 0.0859, |
|
"step": 16620 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 3.3108542068079634e-05, |
|
"loss": 0.1249, |
|
"step": 16630 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 3.294797687861271e-05, |
|
"loss": 0.1253, |
|
"step": 16640 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 3.278741168914579e-05, |
|
"loss": 0.1176, |
|
"step": 16650 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 3.262684649967887e-05, |
|
"loss": 0.1285, |
|
"step": 16660 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 3.2466281310211945e-05, |
|
"loss": 0.1195, |
|
"step": 16670 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 3.2305716120745016e-05, |
|
"loss": 0.1099, |
|
"step": 16680 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 3.2145150931278094e-05, |
|
"loss": 0.1149, |
|
"step": 16690 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 3.198458574181117e-05, |
|
"loss": 0.1167, |
|
"step": 16700 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 3.182402055234425e-05, |
|
"loss": 0.0957, |
|
"step": 16710 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 3.166345536287733e-05, |
|
"loss": 0.1102, |
|
"step": 16720 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 3.1502890173410405e-05, |
|
"loss": 0.1121, |
|
"step": 16730 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 3.1342324983943476e-05, |
|
"loss": 0.119, |
|
"step": 16740 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 3.118175979447655e-05, |
|
"loss": 0.1167, |
|
"step": 16750 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 3.102119460500963e-05, |
|
"loss": 0.1046, |
|
"step": 16760 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 3.086062941554271e-05, |
|
"loss": 0.1322, |
|
"step": 16770 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 3.0700064226075786e-05, |
|
"loss": 0.1136, |
|
"step": 16780 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 3.0539499036608864e-05, |
|
"loss": 0.1062, |
|
"step": 16790 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 3.037893384714194e-05, |
|
"loss": 0.1319, |
|
"step": 16800 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 3.0218368657675013e-05, |
|
"loss": 0.1152, |
|
"step": 16810 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 3.0057803468208087e-05, |
|
"loss": 0.0859, |
|
"step": 16820 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 2.9897238278741165e-05, |
|
"loss": 0.0977, |
|
"step": 16830 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 2.9736673089274243e-05, |
|
"loss": 0.091, |
|
"step": 16840 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 2.957610789980732e-05, |
|
"loss": 0.0985, |
|
"step": 16850 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 2.9415542710340395e-05, |
|
"loss": 0.1222, |
|
"step": 16860 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 2.9254977520873472e-05, |
|
"loss": 0.1225, |
|
"step": 16870 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 2.909441233140655e-05, |
|
"loss": 0.1003, |
|
"step": 16880 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 2.8933847141939624e-05, |
|
"loss": 0.0942, |
|
"step": 16890 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 2.8773281952472702e-05, |
|
"loss": 0.0989, |
|
"step": 16900 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 2.8612716763005776e-05, |
|
"loss": 0.0962, |
|
"step": 16910 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 2.8452151573538854e-05, |
|
"loss": 0.0929, |
|
"step": 16920 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 2.8291586384071932e-05, |
|
"loss": 0.1029, |
|
"step": 16930 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 2.813102119460501e-05, |
|
"loss": 0.1301, |
|
"step": 16940 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 2.7970456005138084e-05, |
|
"loss": 0.1168, |
|
"step": 16950 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 2.7809890815671158e-05, |
|
"loss": 0.0901, |
|
"step": 16960 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 2.7649325626204236e-05, |
|
"loss": 0.102, |
|
"step": 16970 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 2.7488760436737314e-05, |
|
"loss": 0.1136, |
|
"step": 16980 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 2.732819524727039e-05, |
|
"loss": 0.122, |
|
"step": 16990 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 2.716763005780347e-05, |
|
"loss": 0.1081, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"eval_cer": 0.9422206076947389, |
|
"eval_loss": 0.10734660178422928, |
|
"eval_runtime": 52.6881, |
|
"eval_samples_per_second": 18.98, |
|
"eval_steps_per_second": 4.745, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 2.700706486833654e-05, |
|
"loss": 0.1107, |
|
"step": 17010 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 2.6846499678869618e-05, |
|
"loss": 0.09, |
|
"step": 17020 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 2.6685934489402695e-05, |
|
"loss": 0.1052, |
|
"step": 17030 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 2.6525369299935773e-05, |
|
"loss": 0.1187, |
|
"step": 17040 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 2.636480411046885e-05, |
|
"loss": 0.1151, |
|
"step": 17050 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 2.6204238921001922e-05, |
|
"loss": 0.0946, |
|
"step": 17060 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 2.6043673731535e-05, |
|
"loss": 0.125, |
|
"step": 17070 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 2.5883108542068077e-05, |
|
"loss": 0.1142, |
|
"step": 17080 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 2.5722543352601155e-05, |
|
"loss": 0.1194, |
|
"step": 17090 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 2.5561978163134233e-05, |
|
"loss": 0.1252, |
|
"step": 17100 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 2.5401412973667307e-05, |
|
"loss": 0.1224, |
|
"step": 17110 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 2.524084778420038e-05, |
|
"loss": 0.1143, |
|
"step": 17120 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 2.508028259473346e-05, |
|
"loss": 0.101, |
|
"step": 17130 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 2.4919717405266537e-05, |
|
"loss": 0.1219, |
|
"step": 17140 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 2.4759152215799614e-05, |
|
"loss": 0.1392, |
|
"step": 17150 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 2.459858702633269e-05, |
|
"loss": 0.102, |
|
"step": 17160 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 2.4438021836865766e-05, |
|
"loss": 0.0938, |
|
"step": 17170 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 2.427745664739884e-05, |
|
"loss": 0.098, |
|
"step": 17180 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 2.411689145793192e-05, |
|
"loss": 0.0834, |
|
"step": 17190 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 2.3956326268464996e-05, |
|
"loss": 0.0898, |
|
"step": 17200 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 2.379576107899807e-05, |
|
"loss": 0.1023, |
|
"step": 17210 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 2.3635195889531148e-05, |
|
"loss": 0.111, |
|
"step": 17220 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 2.3474630700064226e-05, |
|
"loss": 0.1208, |
|
"step": 17230 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 2.33140655105973e-05, |
|
"loss": 0.0975, |
|
"step": 17240 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 2.3153500321130375e-05, |
|
"loss": 0.1212, |
|
"step": 17250 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 2.2992935131663452e-05, |
|
"loss": 0.114, |
|
"step": 17260 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 2.283236994219653e-05, |
|
"loss": 0.1245, |
|
"step": 17270 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 2.2671804752729608e-05, |
|
"loss": 0.0937, |
|
"step": 17280 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 2.2511239563262685e-05, |
|
"loss": 0.1035, |
|
"step": 17290 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 2.2350674373795756e-05, |
|
"loss": 0.109, |
|
"step": 17300 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 2.2190109184328834e-05, |
|
"loss": 0.1042, |
|
"step": 17310 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 2.2029543994861912e-05, |
|
"loss": 0.123, |
|
"step": 17320 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 2.186897880539499e-05, |
|
"loss": 0.1114, |
|
"step": 17330 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 2.1708413615928067e-05, |
|
"loss": 0.0973, |
|
"step": 17340 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 2.1547848426461138e-05, |
|
"loss": 0.0943, |
|
"step": 17350 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 2.1387283236994216e-05, |
|
"loss": 0.1177, |
|
"step": 17360 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 2.1226718047527294e-05, |
|
"loss": 0.1148, |
|
"step": 17370 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 2.106615285806037e-05, |
|
"loss": 0.1522, |
|
"step": 17380 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 2.090558766859345e-05, |
|
"loss": 0.0969, |
|
"step": 17390 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 2.0745022479126523e-05, |
|
"loss": 0.1249, |
|
"step": 17400 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 2.0584457289659598e-05, |
|
"loss": 0.1181, |
|
"step": 17410 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 2.0423892100192675e-05, |
|
"loss": 0.1077, |
|
"step": 17420 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 2.0263326910725753e-05, |
|
"loss": 0.0898, |
|
"step": 17430 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 2.010276172125883e-05, |
|
"loss": 0.1029, |
|
"step": 17440 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 1.9942196531791905e-05, |
|
"loss": 0.1141, |
|
"step": 17450 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 1.9781631342324983e-05, |
|
"loss": 0.1071, |
|
"step": 17460 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 1.9621066152858057e-05, |
|
"loss": 0.1119, |
|
"step": 17470 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 1.9460500963391135e-05, |
|
"loss": 0.0886, |
|
"step": 17480 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 1.9299935773924213e-05, |
|
"loss": 0.1144, |
|
"step": 17490 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 1.9139370584457287e-05, |
|
"loss": 0.1123, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 1.8978805394990365e-05, |
|
"loss": 0.1319, |
|
"step": 17510 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 1.8818240205523442e-05, |
|
"loss": 0.1147, |
|
"step": 17520 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 1.8657675016056517e-05, |
|
"loss": 0.1136, |
|
"step": 17530 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 1.8513166345536287e-05, |
|
"loss": 0.1201, |
|
"step": 17540 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 1.835260115606936e-05, |
|
"loss": 0.0958, |
|
"step": 17550 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 1.819203596660244e-05, |
|
"loss": 0.1168, |
|
"step": 17560 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 1.8031470777135517e-05, |
|
"loss": 0.0753, |
|
"step": 17570 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 1.787090558766859e-05, |
|
"loss": 0.129, |
|
"step": 17580 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 1.771034039820167e-05, |
|
"loss": 0.1306, |
|
"step": 17590 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 1.7549775208734747e-05, |
|
"loss": 0.11, |
|
"step": 17600 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 1.738921001926782e-05, |
|
"loss": 0.1054, |
|
"step": 17610 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 1.72286448298009e-05, |
|
"loss": 0.117, |
|
"step": 17620 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 1.7068079640333973e-05, |
|
"loss": 0.1114, |
|
"step": 17630 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 1.690751445086705e-05, |
|
"loss": 0.132, |
|
"step": 17640 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 1.674694926140013e-05, |
|
"loss": 0.1099, |
|
"step": 17650 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 1.6586384071933203e-05, |
|
"loss": 0.078, |
|
"step": 17660 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 1.642581888246628e-05, |
|
"loss": 0.1264, |
|
"step": 17670 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 1.6265253692999355e-05, |
|
"loss": 0.115, |
|
"step": 17680 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 1.6104688503532433e-05, |
|
"loss": 0.0968, |
|
"step": 17690 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 1.594412331406551e-05, |
|
"loss": 0.0946, |
|
"step": 17700 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 1.5783558124598585e-05, |
|
"loss": 0.1178, |
|
"step": 17710 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 1.5622992935131662e-05, |
|
"loss": 0.0918, |
|
"step": 17720 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 1.5462427745664737e-05, |
|
"loss": 0.111, |
|
"step": 17730 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 1.5301862556197814e-05, |
|
"loss": 0.0984, |
|
"step": 17740 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 1.5141297366730892e-05, |
|
"loss": 0.0933, |
|
"step": 17750 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 1.4980732177263968e-05, |
|
"loss": 0.0999, |
|
"step": 17760 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 1.4820166987797044e-05, |
|
"loss": 0.129, |
|
"step": 17770 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 1.465960179833012e-05, |
|
"loss": 0.1141, |
|
"step": 17780 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 1.4499036608863198e-05, |
|
"loss": 0.1028, |
|
"step": 17790 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 1.4338471419396274e-05, |
|
"loss": 0.1184, |
|
"step": 17800 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 1.417790622992935e-05, |
|
"loss": 0.1078, |
|
"step": 17810 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 1.4017341040462428e-05, |
|
"loss": 0.1314, |
|
"step": 17820 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 1.3856775850995502e-05, |
|
"loss": 0.0951, |
|
"step": 17830 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 1.369621066152858e-05, |
|
"loss": 0.1322, |
|
"step": 17840 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 1.3535645472061654e-05, |
|
"loss": 0.119, |
|
"step": 17850 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 1.3375080282594732e-05, |
|
"loss": 0.1055, |
|
"step": 17860 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 1.321451509312781e-05, |
|
"loss": 0.1136, |
|
"step": 17870 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 1.3053949903660884e-05, |
|
"loss": 0.0907, |
|
"step": 17880 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 1.2893384714193962e-05, |
|
"loss": 0.1045, |
|
"step": 17890 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 1.2732819524727038e-05, |
|
"loss": 0.0992, |
|
"step": 17900 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 1.2572254335260115e-05, |
|
"loss": 0.0977, |
|
"step": 17910 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 1.2411689145793191e-05, |
|
"loss": 0.1243, |
|
"step": 17920 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 1.2251123956326267e-05, |
|
"loss": 0.087, |
|
"step": 17930 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 1.2090558766859345e-05, |
|
"loss": 0.0978, |
|
"step": 17940 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 1.192999357739242e-05, |
|
"loss": 0.1015, |
|
"step": 17950 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 1.1769428387925497e-05, |
|
"loss": 0.1375, |
|
"step": 17960 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 1.1608863198458575e-05, |
|
"loss": 0.1024, |
|
"step": 17970 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 1.1448298008991649e-05, |
|
"loss": 0.0903, |
|
"step": 17980 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 1.1287732819524727e-05, |
|
"loss": 0.1068, |
|
"step": 17990 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 1.1127167630057801e-05, |
|
"loss": 0.099, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"eval_cer": 0.9421904445910342, |
|
"eval_loss": 0.10652155429124832, |
|
"eval_runtime": 53.6318, |
|
"eval_samples_per_second": 18.646, |
|
"eval_steps_per_second": 4.661, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 1.0966602440590879e-05, |
|
"loss": 0.1304, |
|
"step": 18010 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 1.0806037251123957e-05, |
|
"loss": 0.1089, |
|
"step": 18020 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 1.0645472061657031e-05, |
|
"loss": 0.0976, |
|
"step": 18030 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 1.0484906872190109e-05, |
|
"loss": 0.099, |
|
"step": 18040 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 1.0324341682723185e-05, |
|
"loss": 0.0938, |
|
"step": 18050 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 1.016377649325626e-05, |
|
"loss": 0.109, |
|
"step": 18060 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 1.0003211303789337e-05, |
|
"loss": 0.1032, |
|
"step": 18070 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 9.842646114322414e-06, |
|
"loss": 0.0953, |
|
"step": 18080 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 9.68208092485549e-06, |
|
"loss": 0.1257, |
|
"step": 18090 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 9.521515735388566e-06, |
|
"loss": 0.1235, |
|
"step": 18100 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 9.360950545921644e-06, |
|
"loss": 0.1252, |
|
"step": 18110 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 9.20038535645472e-06, |
|
"loss": 0.111, |
|
"step": 18120 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 9.039820166987796e-06, |
|
"loss": 0.1105, |
|
"step": 18130 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 8.879254977520872e-06, |
|
"loss": 0.1038, |
|
"step": 18140 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 8.718689788053948e-06, |
|
"loss": 0.1159, |
|
"step": 18150 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 8.558124598587026e-06, |
|
"loss": 0.1265, |
|
"step": 18160 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 8.397559409120102e-06, |
|
"loss": 0.1137, |
|
"step": 18170 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 8.236994219653178e-06, |
|
"loss": 0.0975, |
|
"step": 18180 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 8.076429030186254e-06, |
|
"loss": 0.1209, |
|
"step": 18190 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 7.915863840719332e-06, |
|
"loss": 0.1143, |
|
"step": 18200 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 7.755298651252408e-06, |
|
"loss": 0.0847, |
|
"step": 18210 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 7.594733461785485e-06, |
|
"loss": 0.1172, |
|
"step": 18220 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 7.434168272318561e-06, |
|
"loss": 0.1239, |
|
"step": 18230 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 7.2736030828516375e-06, |
|
"loss": 0.1152, |
|
"step": 18240 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 7.1130378933847135e-06, |
|
"loss": 0.0977, |
|
"step": 18250 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 6.9524727039177895e-06, |
|
"loss": 0.1227, |
|
"step": 18260 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 6.791907514450867e-06, |
|
"loss": 0.0985, |
|
"step": 18270 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 6.631342324983943e-06, |
|
"loss": 0.1201, |
|
"step": 18280 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 6.470777135517019e-06, |
|
"loss": 0.1277, |
|
"step": 18290 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 6.310211946050095e-06, |
|
"loss": 0.1083, |
|
"step": 18300 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 6.149646756583172e-06, |
|
"loss": 0.1192, |
|
"step": 18310 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 5.989081567116249e-06, |
|
"loss": 0.0894, |
|
"step": 18320 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 5.828516377649325e-06, |
|
"loss": 0.1127, |
|
"step": 18330 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 5.667951188182402e-06, |
|
"loss": 0.1238, |
|
"step": 18340 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 5.507385998715478e-06, |
|
"loss": 0.1351, |
|
"step": 18350 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 5.346820809248554e-06, |
|
"loss": 0.1259, |
|
"step": 18360 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 5.186255619781631e-06, |
|
"loss": 0.1314, |
|
"step": 18370 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 5.025690430314708e-06, |
|
"loss": 0.0892, |
|
"step": 18380 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 4.865125240847784e-06, |
|
"loss": 0.1321, |
|
"step": 18390 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 4.704560051380861e-06, |
|
"loss": 0.0818, |
|
"step": 18400 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 4.543994861913937e-06, |
|
"loss": 0.0867, |
|
"step": 18410 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 4.3834296724470135e-06, |
|
"loss": 0.0854, |
|
"step": 18420 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 4.2228644829800895e-06, |
|
"loss": 0.1138, |
|
"step": 18430 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 4.0622992935131655e-06, |
|
"loss": 0.1128, |
|
"step": 18440 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 3.901734104046242e-06, |
|
"loss": 0.1083, |
|
"step": 18450 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 3.741168914579319e-06, |
|
"loss": 0.099, |
|
"step": 18460 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 3.5806037251123953e-06, |
|
"loss": 0.1036, |
|
"step": 18470 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 3.4200385356454717e-06, |
|
"loss": 0.0934, |
|
"step": 18480 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 3.259473346178548e-06, |
|
"loss": 0.0919, |
|
"step": 18490 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 3.098908156711625e-06, |
|
"loss": 0.1233, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 2.938342967244701e-06, |
|
"loss": 0.1164, |
|
"step": 18510 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 2.7777777777777775e-06, |
|
"loss": 0.0859, |
|
"step": 18520 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 2.6172125883108544e-06, |
|
"loss": 0.1136, |
|
"step": 18530 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 2.4566473988439304e-06, |
|
"loss": 0.0993, |
|
"step": 18540 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 2.296082209377007e-06, |
|
"loss": 0.121, |
|
"step": 18550 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 2.1355170199100833e-06, |
|
"loss": 0.0918, |
|
"step": 18560 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 1.9749518304431597e-06, |
|
"loss": 0.0692, |
|
"step": 18570 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 1.8143866409762361e-06, |
|
"loss": 0.1291, |
|
"step": 18580 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 1.6538214515093128e-06, |
|
"loss": 0.1559, |
|
"step": 18590 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 1.493256262042389e-06, |
|
"loss": 0.1011, |
|
"step": 18600 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 1.3326910725754655e-06, |
|
"loss": 0.0868, |
|
"step": 18610 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 1.172125883108542e-06, |
|
"loss": 0.1257, |
|
"step": 18620 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 1.0115606936416184e-06, |
|
"loss": 0.102, |
|
"step": 18630 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 8.509955041746949e-07, |
|
"loss": 0.1201, |
|
"step": 18640 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 6.904303147077713e-07, |
|
"loss": 0.1079, |
|
"step": 18650 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 5.298651252408477e-07, |
|
"loss": 0.1304, |
|
"step": 18660 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 3.692999357739242e-07, |
|
"loss": 0.1109, |
|
"step": 18670 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 2.0873474630700063e-07, |
|
"loss": 0.0965, |
|
"step": 18680 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"step": 18684, |
|
"total_flos": 1.426057106632704e+17, |
|
"train_loss": 0.16492035415392825, |
|
"train_runtime": 21638.3485, |
|
"train_samples_per_second": 27.633, |
|
"train_steps_per_second": 0.863 |
|
} |
|
], |
|
"max_steps": 18684, |
|
"num_train_epochs": 2, |
|
"total_flos": 1.426057106632704e+17, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|