|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 1.0, |
|
"global_step": 7508, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.001, |
|
"learning_rate": 8.50153746173217e-06, |
|
"loss": 0.7923, |
|
"r_loss": 0.0, |
|
"step": 10, |
|
"steps": 10 |
|
}, |
|
{ |
|
"epoch": 0.003, |
|
"learning_rate": 1.1057744947017939e-05, |
|
"loss": 0.7252, |
|
"r_loss": 0.0, |
|
"step": 20, |
|
"steps": 20 |
|
}, |
|
{ |
|
"epoch": 0.004, |
|
"learning_rate": 1.255303046997284e-05, |
|
"loss": 0.7285, |
|
"r_loss": 0.0, |
|
"step": 30, |
|
"steps": 30 |
|
}, |
|
{ |
|
"epoch": 0.005, |
|
"learning_rate": 1.3613952432303706e-05, |
|
"loss": 0.7204, |
|
"r_loss": 0.0, |
|
"step": 40, |
|
"steps": 40 |
|
}, |
|
{ |
|
"epoch": 0.007, |
|
"learning_rate": 1.4436867438178569e-05, |
|
"loss": 0.7221, |
|
"r_loss": 0.0, |
|
"step": 50, |
|
"steps": 50 |
|
}, |
|
{ |
|
"epoch": 0.008, |
|
"learning_rate": 1.5109237955258607e-05, |
|
"loss": 0.7153, |
|
"r_loss": 0.0, |
|
"step": 60, |
|
"steps": 60 |
|
}, |
|
{ |
|
"epoch": 0.009, |
|
"learning_rate": 1.5677719127349665e-05, |
|
"loss": 0.6948, |
|
"r_loss": 0.0, |
|
"step": 70, |
|
"steps": 70 |
|
}, |
|
{ |
|
"epoch": 0.011, |
|
"learning_rate": 1.6170159917589475e-05, |
|
"loss": 0.7023, |
|
"r_loss": 0.0, |
|
"step": 80, |
|
"steps": 80 |
|
}, |
|
{ |
|
"epoch": 0.012, |
|
"learning_rate": 1.660452347821351e-05, |
|
"loss": 0.7053, |
|
"r_loss": 0.0, |
|
"step": 90, |
|
"steps": 90 |
|
}, |
|
{ |
|
"epoch": 0.013, |
|
"learning_rate": 1.6993074923464342e-05, |
|
"loss": 0.6908, |
|
"r_loss": 0.0, |
|
"step": 100, |
|
"steps": 100 |
|
}, |
|
{ |
|
"epoch": 0.015, |
|
"learning_rate": 1.7344562460127094e-05, |
|
"loss": 0.7005, |
|
"r_loss": 0.0, |
|
"step": 110, |
|
"steps": 110 |
|
}, |
|
{ |
|
"epoch": 0.016, |
|
"learning_rate": 1.7665445440544375e-05, |
|
"loss": 0.6864, |
|
"r_loss": 0.0, |
|
"step": 120, |
|
"steps": 120 |
|
}, |
|
{ |
|
"epoch": 0.017, |
|
"learning_rate": 1.796062916809319e-05, |
|
"loss": 0.7273, |
|
"r_loss": 0.0, |
|
"step": 130, |
|
"steps": 130 |
|
}, |
|
{ |
|
"epoch": 0.019, |
|
"learning_rate": 1.823392661263543e-05, |
|
"loss": 0.7019, |
|
"r_loss": 0.0, |
|
"step": 140, |
|
"steps": 140 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 1.848836044641924e-05, |
|
"loss": 0.6934, |
|
"r_loss": 0.0, |
|
"step": 150, |
|
"steps": 150 |
|
}, |
|
{ |
|
"epoch": 0.021, |
|
"learning_rate": 1.8726367402875247e-05, |
|
"loss": 0.6908, |
|
"r_loss": 0.0, |
|
"step": 160, |
|
"steps": 160 |
|
}, |
|
{ |
|
"epoch": 0.023, |
|
"learning_rate": 1.8949940572363726e-05, |
|
"loss": 0.677, |
|
"r_loss": 0.0, |
|
"step": 170, |
|
"steps": 170 |
|
}, |
|
{ |
|
"epoch": 0.024, |
|
"learning_rate": 1.916073096349928e-05, |
|
"loss": 0.6889, |
|
"r_loss": 0.0, |
|
"step": 180, |
|
"steps": 180 |
|
}, |
|
{ |
|
"epoch": 0.025, |
|
"learning_rate": 1.936012156854803e-05, |
|
"loss": 0.6831, |
|
"r_loss": 0.0, |
|
"step": 190, |
|
"steps": 190 |
|
}, |
|
{ |
|
"epoch": 0.027, |
|
"learning_rate": 1.9549282408750108e-05, |
|
"loss": 0.6893, |
|
"r_loss": 0.0, |
|
"step": 200, |
|
"steps": 200 |
|
}, |
|
{ |
|
"epoch": 0.028, |
|
"learning_rate": 1.9729212135590334e-05, |
|
"loss": 0.684, |
|
"r_loss": 0.0, |
|
"step": 210, |
|
"steps": 210 |
|
}, |
|
{ |
|
"epoch": 0.029, |
|
"learning_rate": 1.9900769945412862e-05, |
|
"loss": 0.7032, |
|
"r_loss": 0.0, |
|
"step": 220, |
|
"steps": 220 |
|
}, |
|
{ |
|
"epoch": 0.031, |
|
"learning_rate": 1.9991764625102995e-05, |
|
"loss": 0.6922, |
|
"r_loss": 0.0, |
|
"step": 230, |
|
"steps": 230 |
|
}, |
|
{ |
|
"epoch": 0.032, |
|
"learning_rate": 1.9964313375446307e-05, |
|
"loss": 0.6779, |
|
"r_loss": 0.0, |
|
"step": 240, |
|
"steps": 240 |
|
}, |
|
{ |
|
"epoch": 0.033, |
|
"learning_rate": 1.993686212578962e-05, |
|
"loss": 0.6851, |
|
"r_loss": 0.0, |
|
"step": 250, |
|
"steps": 250 |
|
}, |
|
{ |
|
"epoch": 0.035, |
|
"learning_rate": 1.990941087613293e-05, |
|
"loss": 0.6834, |
|
"r_loss": 0.0, |
|
"step": 260, |
|
"steps": 260 |
|
}, |
|
{ |
|
"epoch": 0.036, |
|
"learning_rate": 1.9881959626476245e-05, |
|
"loss": 0.68, |
|
"r_loss": 0.0, |
|
"step": 270, |
|
"steps": 270 |
|
}, |
|
{ |
|
"epoch": 0.037, |
|
"learning_rate": 1.9854508376819557e-05, |
|
"loss": 0.674, |
|
"r_loss": 0.0, |
|
"step": 280, |
|
"steps": 280 |
|
}, |
|
{ |
|
"epoch": 0.039, |
|
"learning_rate": 1.982705712716287e-05, |
|
"loss": 0.6785, |
|
"r_loss": 0.0, |
|
"step": 290, |
|
"steps": 290 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 1.9799605877506184e-05, |
|
"loss": 0.6976, |
|
"r_loss": 0.0, |
|
"step": 300, |
|
"steps": 300 |
|
}, |
|
{ |
|
"epoch": 0.041, |
|
"learning_rate": 1.9772154627849495e-05, |
|
"loss": 0.6639, |
|
"r_loss": 0.0, |
|
"step": 310, |
|
"steps": 310 |
|
}, |
|
{ |
|
"epoch": 0.043, |
|
"learning_rate": 1.9744703378192804e-05, |
|
"loss": 0.6823, |
|
"r_loss": 0.0, |
|
"step": 320, |
|
"steps": 320 |
|
}, |
|
{ |
|
"epoch": 0.044, |
|
"learning_rate": 1.971725212853612e-05, |
|
"loss": 0.6834, |
|
"r_loss": 0.0, |
|
"step": 330, |
|
"steps": 330 |
|
}, |
|
{ |
|
"epoch": 0.045, |
|
"learning_rate": 1.968980087887943e-05, |
|
"loss": 0.6843, |
|
"r_loss": 0.0, |
|
"step": 340, |
|
"steps": 340 |
|
}, |
|
{ |
|
"epoch": 0.047, |
|
"learning_rate": 1.9662349629222742e-05, |
|
"loss": 0.678, |
|
"r_loss": 0.0, |
|
"step": 350, |
|
"steps": 350 |
|
}, |
|
{ |
|
"epoch": 0.048, |
|
"learning_rate": 1.9634898379566057e-05, |
|
"loss": 0.6742, |
|
"r_loss": 0.0, |
|
"step": 360, |
|
"steps": 360 |
|
}, |
|
{ |
|
"epoch": 0.049, |
|
"learning_rate": 1.960744712990937e-05, |
|
"loss": 0.6652, |
|
"r_loss": 0.0, |
|
"step": 370, |
|
"steps": 370 |
|
}, |
|
{ |
|
"epoch": 0.051, |
|
"learning_rate": 1.957999588025268e-05, |
|
"loss": 0.6714, |
|
"r_loss": 0.0, |
|
"step": 380, |
|
"steps": 380 |
|
}, |
|
{ |
|
"epoch": 0.052, |
|
"learning_rate": 1.9552544630595992e-05, |
|
"loss": 0.6664, |
|
"r_loss": 0.0, |
|
"step": 390, |
|
"steps": 390 |
|
}, |
|
{ |
|
"epoch": 0.053, |
|
"learning_rate": 1.9525093380939303e-05, |
|
"loss": 0.6661, |
|
"r_loss": 0.0, |
|
"step": 400, |
|
"steps": 400 |
|
}, |
|
{ |
|
"epoch": 0.055, |
|
"learning_rate": 1.9497642131282615e-05, |
|
"loss": 0.6736, |
|
"r_loss": 0.0, |
|
"step": 410, |
|
"steps": 410 |
|
}, |
|
{ |
|
"epoch": 0.056, |
|
"learning_rate": 1.947019088162593e-05, |
|
"loss": 0.6731, |
|
"r_loss": 0.0, |
|
"step": 420, |
|
"steps": 420 |
|
}, |
|
{ |
|
"epoch": 0.057, |
|
"learning_rate": 1.944273963196924e-05, |
|
"loss": 0.6613, |
|
"r_loss": 0.0, |
|
"step": 430, |
|
"steps": 430 |
|
}, |
|
{ |
|
"epoch": 0.059, |
|
"learning_rate": 1.9415288382312553e-05, |
|
"loss": 0.6651, |
|
"r_loss": 0.0, |
|
"step": 440, |
|
"steps": 440 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 1.9387837132655865e-05, |
|
"loss": 0.6725, |
|
"r_loss": 0.0, |
|
"step": 450, |
|
"steps": 450 |
|
}, |
|
{ |
|
"epoch": 0.061, |
|
"learning_rate": 1.9363131007964846e-05, |
|
"loss": 0.6623, |
|
"r_loss": 0.0, |
|
"step": 460, |
|
"steps": 460 |
|
}, |
|
{ |
|
"epoch": 0.063, |
|
"learning_rate": 1.933567975830816e-05, |
|
"loss": 0.6758, |
|
"r_loss": 0.0, |
|
"step": 470, |
|
"steps": 470 |
|
}, |
|
{ |
|
"epoch": 0.064, |
|
"learning_rate": 1.930822850865147e-05, |
|
"loss": 0.6611, |
|
"r_loss": 0.0, |
|
"step": 480, |
|
"steps": 480 |
|
}, |
|
{ |
|
"epoch": 0.065, |
|
"learning_rate": 1.928077725899478e-05, |
|
"loss": 0.6701, |
|
"r_loss": 0.0, |
|
"step": 490, |
|
"steps": 490 |
|
}, |
|
{ |
|
"epoch": 0.067, |
|
"learning_rate": 1.9253326009338096e-05, |
|
"loss": 0.6657, |
|
"r_loss": 0.0, |
|
"step": 500, |
|
"steps": 500 |
|
}, |
|
{ |
|
"epoch": 0.068, |
|
"learning_rate": 1.9225874759681408e-05, |
|
"loss": 0.6668, |
|
"r_loss": 0.0, |
|
"step": 510, |
|
"steps": 510 |
|
}, |
|
{ |
|
"epoch": 0.069, |
|
"learning_rate": 1.919842351002472e-05, |
|
"loss": 0.6737, |
|
"r_loss": 0.0, |
|
"step": 520, |
|
"steps": 520 |
|
}, |
|
{ |
|
"epoch": 0.071, |
|
"learning_rate": 1.9170972260368035e-05, |
|
"loss": 0.6586, |
|
"r_loss": 0.0, |
|
"step": 530, |
|
"steps": 530 |
|
}, |
|
{ |
|
"epoch": 0.072, |
|
"learning_rate": 1.9143521010711346e-05, |
|
"loss": 0.6704, |
|
"r_loss": 0.0, |
|
"step": 540, |
|
"steps": 540 |
|
}, |
|
{ |
|
"epoch": 0.073, |
|
"learning_rate": 1.9116069761054658e-05, |
|
"loss": 0.6573, |
|
"r_loss": 0.0, |
|
"step": 550, |
|
"steps": 550 |
|
}, |
|
{ |
|
"epoch": 0.075, |
|
"learning_rate": 1.908861851139797e-05, |
|
"loss": 0.6558, |
|
"r_loss": 0.0, |
|
"step": 560, |
|
"steps": 560 |
|
}, |
|
{ |
|
"epoch": 0.076, |
|
"learning_rate": 1.906116726174128e-05, |
|
"loss": 0.6563, |
|
"r_loss": 0.0, |
|
"step": 570, |
|
"steps": 570 |
|
}, |
|
{ |
|
"epoch": 0.077, |
|
"learning_rate": 1.9033716012084593e-05, |
|
"loss": 0.6609, |
|
"r_loss": 0.0, |
|
"step": 580, |
|
"steps": 580 |
|
}, |
|
{ |
|
"epoch": 0.079, |
|
"learning_rate": 1.9006264762427908e-05, |
|
"loss": 0.6551, |
|
"r_loss": 0.0, |
|
"step": 590, |
|
"steps": 590 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 1.897881351277122e-05, |
|
"loss": 0.6599, |
|
"r_loss": 0.0, |
|
"step": 600, |
|
"steps": 600 |
|
}, |
|
{ |
|
"epoch": 0.081, |
|
"learning_rate": 1.895136226311453e-05, |
|
"loss": 0.6557, |
|
"r_loss": 0.0, |
|
"step": 610, |
|
"steps": 610 |
|
}, |
|
{ |
|
"epoch": 0.083, |
|
"learning_rate": 1.8926656138423513e-05, |
|
"loss": 0.6588, |
|
"r_loss": 0.0, |
|
"step": 620, |
|
"steps": 620 |
|
}, |
|
{ |
|
"epoch": 0.084, |
|
"learning_rate": 1.8899204888766824e-05, |
|
"loss": 0.6637, |
|
"r_loss": 0.0, |
|
"step": 630, |
|
"steps": 630 |
|
}, |
|
{ |
|
"epoch": 0.085, |
|
"learning_rate": 1.8871753639110136e-05, |
|
"loss": 0.6464, |
|
"r_loss": 0.0, |
|
"step": 640, |
|
"steps": 640 |
|
}, |
|
{ |
|
"epoch": 0.087, |
|
"learning_rate": 1.8844302389453447e-05, |
|
"loss": 0.6582, |
|
"r_loss": 0.0, |
|
"step": 650, |
|
"steps": 650 |
|
}, |
|
{ |
|
"epoch": 0.088, |
|
"learning_rate": 1.881685113979676e-05, |
|
"loss": 0.6457, |
|
"r_loss": 0.0, |
|
"step": 660, |
|
"steps": 660 |
|
}, |
|
{ |
|
"epoch": 0.089, |
|
"learning_rate": 1.8789399890140074e-05, |
|
"loss": 0.6598, |
|
"r_loss": 0.0, |
|
"step": 670, |
|
"steps": 670 |
|
}, |
|
{ |
|
"epoch": 0.091, |
|
"learning_rate": 1.8761948640483386e-05, |
|
"loss": 0.65, |
|
"r_loss": 0.0, |
|
"step": 680, |
|
"steps": 680 |
|
}, |
|
{ |
|
"epoch": 0.092, |
|
"learning_rate": 1.8734497390826697e-05, |
|
"loss": 0.6348, |
|
"r_loss": 0.0, |
|
"step": 690, |
|
"steps": 690 |
|
}, |
|
{ |
|
"epoch": 0.093, |
|
"learning_rate": 1.8707046141170012e-05, |
|
"loss": 0.6596, |
|
"r_loss": 0.0, |
|
"step": 700, |
|
"steps": 700 |
|
}, |
|
{ |
|
"epoch": 0.095, |
|
"learning_rate": 1.8679594891513324e-05, |
|
"loss": 0.641, |
|
"r_loss": 0.0, |
|
"step": 710, |
|
"steps": 710 |
|
}, |
|
{ |
|
"epoch": 0.096, |
|
"learning_rate": 1.8652143641856632e-05, |
|
"loss": 0.6598, |
|
"r_loss": 0.0, |
|
"step": 720, |
|
"steps": 720 |
|
}, |
|
{ |
|
"epoch": 0.097, |
|
"learning_rate": 1.8624692392199947e-05, |
|
"loss": 0.6606, |
|
"r_loss": 0.0, |
|
"step": 730, |
|
"steps": 730 |
|
}, |
|
{ |
|
"epoch": 0.099, |
|
"learning_rate": 1.859724114254326e-05, |
|
"loss": 0.6471, |
|
"r_loss": 0.0, |
|
"step": 740, |
|
"steps": 740 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 1.856978989288657e-05, |
|
"loss": 0.6613, |
|
"r_loss": 0.0, |
|
"step": 750, |
|
"steps": 750 |
|
}, |
|
{ |
|
"epoch": 0.101, |
|
"learning_rate": 1.8542338643229886e-05, |
|
"loss": 0.6495, |
|
"r_loss": 0.0, |
|
"step": 760, |
|
"steps": 760 |
|
}, |
|
{ |
|
"epoch": 0.103, |
|
"learning_rate": 1.8514887393573197e-05, |
|
"loss": 0.6577, |
|
"r_loss": 0.0, |
|
"step": 770, |
|
"steps": 770 |
|
}, |
|
{ |
|
"epoch": 0.104, |
|
"learning_rate": 1.848743614391651e-05, |
|
"loss": 0.6466, |
|
"r_loss": 0.0, |
|
"step": 780, |
|
"steps": 780 |
|
}, |
|
{ |
|
"epoch": 0.105, |
|
"learning_rate": 1.845998489425982e-05, |
|
"loss": 0.655, |
|
"r_loss": 0.0, |
|
"step": 790, |
|
"steps": 790 |
|
}, |
|
{ |
|
"epoch": 0.107, |
|
"learning_rate": 1.8432533644603132e-05, |
|
"loss": 0.6291, |
|
"r_loss": 0.0, |
|
"step": 800, |
|
"steps": 800 |
|
}, |
|
{ |
|
"epoch": 0.108, |
|
"learning_rate": 1.8405082394946444e-05, |
|
"loss": 0.6574, |
|
"r_loss": 0.0, |
|
"step": 810, |
|
"steps": 810 |
|
}, |
|
{ |
|
"epoch": 0.109, |
|
"learning_rate": 1.837763114528976e-05, |
|
"loss": 0.6576, |
|
"r_loss": 0.0, |
|
"step": 820, |
|
"steps": 820 |
|
}, |
|
{ |
|
"epoch": 0.111, |
|
"learning_rate": 1.835017989563307e-05, |
|
"loss": 0.646, |
|
"r_loss": 0.0, |
|
"step": 830, |
|
"steps": 830 |
|
}, |
|
{ |
|
"epoch": 0.112, |
|
"learning_rate": 1.8322728645976382e-05, |
|
"loss": 0.6479, |
|
"r_loss": 0.0, |
|
"step": 840, |
|
"steps": 840 |
|
}, |
|
{ |
|
"epoch": 0.113, |
|
"learning_rate": 1.8295277396319694e-05, |
|
"loss": 0.6435, |
|
"r_loss": 0.0, |
|
"step": 850, |
|
"steps": 850 |
|
}, |
|
{ |
|
"epoch": 0.115, |
|
"learning_rate": 1.8267826146663005e-05, |
|
"loss": 0.6473, |
|
"r_loss": 0.0, |
|
"step": 860, |
|
"steps": 860 |
|
}, |
|
{ |
|
"epoch": 0.116, |
|
"learning_rate": 1.824037489700632e-05, |
|
"loss": 0.6554, |
|
"r_loss": 0.0, |
|
"step": 870, |
|
"steps": 870 |
|
}, |
|
{ |
|
"epoch": 0.117, |
|
"learning_rate": 1.8212923647349632e-05, |
|
"loss": 0.6362, |
|
"r_loss": 0.0, |
|
"step": 880, |
|
"steps": 880 |
|
}, |
|
{ |
|
"epoch": 0.119, |
|
"learning_rate": 1.8185472397692944e-05, |
|
"loss": 0.6489, |
|
"r_loss": 0.0, |
|
"step": 890, |
|
"steps": 890 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 1.8158021148036255e-05, |
|
"loss": 0.6496, |
|
"r_loss": 0.0, |
|
"step": 900, |
|
"steps": 900 |
|
}, |
|
{ |
|
"epoch": 0.121, |
|
"learning_rate": 1.8130569898379567e-05, |
|
"loss": 0.6358, |
|
"r_loss": 0.0, |
|
"step": 910, |
|
"steps": 910 |
|
}, |
|
{ |
|
"epoch": 0.123, |
|
"learning_rate": 1.810311864872288e-05, |
|
"loss": 0.6461, |
|
"r_loss": 0.0, |
|
"step": 920, |
|
"steps": 920 |
|
}, |
|
{ |
|
"epoch": 0.124, |
|
"learning_rate": 1.8075667399066194e-05, |
|
"loss": 0.6464, |
|
"r_loss": 0.0, |
|
"step": 930, |
|
"steps": 930 |
|
}, |
|
{ |
|
"epoch": 0.125, |
|
"learning_rate": 1.8048216149409505e-05, |
|
"loss": 0.6402, |
|
"r_loss": 0.0, |
|
"step": 940, |
|
"steps": 940 |
|
}, |
|
{ |
|
"epoch": 0.127, |
|
"learning_rate": 1.8020764899752817e-05, |
|
"loss": 0.6423, |
|
"r_loss": 0.0, |
|
"step": 950, |
|
"steps": 950 |
|
}, |
|
{ |
|
"epoch": 0.128, |
|
"learning_rate": 1.799331365009613e-05, |
|
"loss": 0.6318, |
|
"r_loss": 0.0, |
|
"step": 960, |
|
"steps": 960 |
|
}, |
|
{ |
|
"epoch": 0.129, |
|
"learning_rate": 1.796586240043944e-05, |
|
"loss": 0.6353, |
|
"r_loss": 0.0, |
|
"step": 970, |
|
"steps": 970 |
|
}, |
|
{ |
|
"epoch": 0.131, |
|
"learning_rate": 1.793841115078275e-05, |
|
"loss": 0.6379, |
|
"r_loss": 0.0, |
|
"step": 980, |
|
"steps": 980 |
|
}, |
|
{ |
|
"epoch": 0.132, |
|
"learning_rate": 1.7910959901126067e-05, |
|
"loss": 0.6367, |
|
"r_loss": 0.0, |
|
"step": 990, |
|
"steps": 990 |
|
}, |
|
{ |
|
"epoch": 0.133, |
|
"learning_rate": 1.788350865146938e-05, |
|
"loss": 0.6327, |
|
"r_loss": 0.0, |
|
"step": 1000, |
|
"steps": 1000 |
|
}, |
|
{ |
|
"epoch": 0.135, |
|
"learning_rate": 1.785605740181269e-05, |
|
"loss": 0.6357, |
|
"r_loss": 0.0, |
|
"step": 1010, |
|
"steps": 1010 |
|
}, |
|
{ |
|
"epoch": 0.136, |
|
"learning_rate": 1.7828606152156e-05, |
|
"loss": 0.6308, |
|
"r_loss": 0.0, |
|
"step": 1020, |
|
"steps": 1020 |
|
}, |
|
{ |
|
"epoch": 0.137, |
|
"learning_rate": 1.7801154902499313e-05, |
|
"loss": 0.632, |
|
"r_loss": 0.0, |
|
"step": 1030, |
|
"steps": 1030 |
|
}, |
|
{ |
|
"epoch": 0.139, |
|
"learning_rate": 1.7773703652842628e-05, |
|
"loss": 0.6369, |
|
"r_loss": 0.0, |
|
"step": 1040, |
|
"steps": 1040 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 1.774625240318594e-05, |
|
"loss": 0.6308, |
|
"r_loss": 0.0, |
|
"step": 1050, |
|
"steps": 1050 |
|
}, |
|
{ |
|
"epoch": 0.141, |
|
"learning_rate": 1.771880115352925e-05, |
|
"loss": 0.6442, |
|
"r_loss": 0.0, |
|
"step": 1060, |
|
"steps": 1060 |
|
}, |
|
{ |
|
"epoch": 0.143, |
|
"learning_rate": 1.7691349903872567e-05, |
|
"loss": 0.6434, |
|
"r_loss": 0.0, |
|
"step": 1070, |
|
"steps": 1070 |
|
}, |
|
{ |
|
"epoch": 0.144, |
|
"learning_rate": 1.7663898654215878e-05, |
|
"loss": 0.6484, |
|
"r_loss": 0.0, |
|
"step": 1080, |
|
"steps": 1080 |
|
}, |
|
{ |
|
"epoch": 0.145, |
|
"learning_rate": 1.7636447404559186e-05, |
|
"loss": 0.638, |
|
"r_loss": 0.0, |
|
"step": 1090, |
|
"steps": 1090 |
|
}, |
|
{ |
|
"epoch": 0.147, |
|
"learning_rate": 1.76089961549025e-05, |
|
"loss": 0.6337, |
|
"r_loss": 0.0, |
|
"step": 1100, |
|
"steps": 1100 |
|
}, |
|
{ |
|
"epoch": 0.148, |
|
"learning_rate": 1.7581544905245813e-05, |
|
"loss": 0.652, |
|
"r_loss": 0.0, |
|
"step": 1110, |
|
"steps": 1110 |
|
}, |
|
{ |
|
"epoch": 0.149, |
|
"learning_rate": 1.7554093655589125e-05, |
|
"loss": 0.6325, |
|
"r_loss": 0.0, |
|
"step": 1120, |
|
"steps": 1120 |
|
}, |
|
{ |
|
"epoch": 0.151, |
|
"learning_rate": 1.752664240593244e-05, |
|
"loss": 0.6394, |
|
"r_loss": 0.0, |
|
"step": 1130, |
|
"steps": 1130 |
|
}, |
|
{ |
|
"epoch": 0.152, |
|
"learning_rate": 1.749919115627575e-05, |
|
"loss": 0.6251, |
|
"r_loss": 0.0, |
|
"step": 1140, |
|
"steps": 1140 |
|
}, |
|
{ |
|
"epoch": 0.153, |
|
"learning_rate": 1.7471739906619063e-05, |
|
"loss": 0.6304, |
|
"r_loss": 0.0, |
|
"step": 1150, |
|
"steps": 1150 |
|
}, |
|
{ |
|
"epoch": 0.155, |
|
"learning_rate": 1.7444288656962375e-05, |
|
"loss": 0.6251, |
|
"r_loss": 0.0, |
|
"step": 1160, |
|
"steps": 1160 |
|
}, |
|
{ |
|
"epoch": 0.156, |
|
"learning_rate": 1.7416837407305686e-05, |
|
"loss": 0.6169, |
|
"r_loss": 0.0, |
|
"step": 1170, |
|
"steps": 1170 |
|
}, |
|
{ |
|
"epoch": 0.157, |
|
"learning_rate": 1.7389386157648998e-05, |
|
"loss": 0.6268, |
|
"r_loss": 0.0, |
|
"step": 1180, |
|
"steps": 1180 |
|
}, |
|
{ |
|
"epoch": 0.158, |
|
"learning_rate": 1.7361934907992313e-05, |
|
"loss": 0.6305, |
|
"r_loss": 0.0, |
|
"step": 1190, |
|
"steps": 1190 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 1.7334483658335625e-05, |
|
"loss": 0.635, |
|
"r_loss": 0.0, |
|
"step": 1200, |
|
"steps": 1200 |
|
}, |
|
{ |
|
"epoch": 0.161, |
|
"learning_rate": 1.7307032408678936e-05, |
|
"loss": 0.6142, |
|
"r_loss": 0.0, |
|
"step": 1210, |
|
"steps": 1210 |
|
}, |
|
{ |
|
"epoch": 0.162, |
|
"learning_rate": 1.7279581159022248e-05, |
|
"loss": 0.6207, |
|
"r_loss": 0.0, |
|
"step": 1220, |
|
"steps": 1220 |
|
}, |
|
{ |
|
"epoch": 0.164, |
|
"learning_rate": 1.725212990936556e-05, |
|
"loss": 0.6317, |
|
"r_loss": 0.0, |
|
"step": 1230, |
|
"steps": 1230 |
|
}, |
|
{ |
|
"epoch": 0.165, |
|
"learning_rate": 1.7224678659708874e-05, |
|
"loss": 0.6434, |
|
"r_loss": 0.0, |
|
"step": 1240, |
|
"steps": 1240 |
|
}, |
|
{ |
|
"epoch": 0.166, |
|
"learning_rate": 1.7197227410052186e-05, |
|
"loss": 0.6335, |
|
"r_loss": 0.0, |
|
"step": 1250, |
|
"steps": 1250 |
|
}, |
|
{ |
|
"epoch": 0.168, |
|
"learning_rate": 1.7169776160395498e-05, |
|
"loss": 0.6341, |
|
"r_loss": 0.0, |
|
"step": 1260, |
|
"steps": 1260 |
|
}, |
|
{ |
|
"epoch": 0.169, |
|
"learning_rate": 1.714232491073881e-05, |
|
"loss": 0.6226, |
|
"r_loss": 0.0, |
|
"step": 1270, |
|
"steps": 1270 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 1.711487366108212e-05, |
|
"loss": 0.6135, |
|
"r_loss": 0.0, |
|
"step": 1280, |
|
"steps": 1280 |
|
}, |
|
{ |
|
"epoch": 0.172, |
|
"learning_rate": 1.7087422411425433e-05, |
|
"loss": 0.6391, |
|
"r_loss": 0.0, |
|
"step": 1290, |
|
"steps": 1290 |
|
}, |
|
{ |
|
"epoch": 0.173, |
|
"learning_rate": 1.7059971161768748e-05, |
|
"loss": 0.6315, |
|
"r_loss": 0.0, |
|
"step": 1300, |
|
"steps": 1300 |
|
}, |
|
{ |
|
"epoch": 0.174, |
|
"learning_rate": 1.703251991211206e-05, |
|
"loss": 0.6224, |
|
"r_loss": 0.0, |
|
"step": 1310, |
|
"steps": 1310 |
|
}, |
|
{ |
|
"epoch": 0.176, |
|
"learning_rate": 1.700506866245537e-05, |
|
"loss": 0.6376, |
|
"r_loss": 0.0, |
|
"step": 1320, |
|
"steps": 1320 |
|
}, |
|
{ |
|
"epoch": 0.177, |
|
"learning_rate": 1.6977617412798683e-05, |
|
"loss": 0.6263, |
|
"r_loss": 0.0, |
|
"step": 1330, |
|
"steps": 1330 |
|
}, |
|
{ |
|
"epoch": 0.178, |
|
"learning_rate": 1.6950166163141994e-05, |
|
"loss": 0.6222, |
|
"r_loss": 0.0, |
|
"step": 1340, |
|
"steps": 1340 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 1.6922714913485306e-05, |
|
"loss": 0.6213, |
|
"r_loss": 0.0, |
|
"step": 1350, |
|
"steps": 1350 |
|
}, |
|
{ |
|
"epoch": 0.181, |
|
"learning_rate": 1.689526366382862e-05, |
|
"loss": 0.6278, |
|
"r_loss": 0.0, |
|
"step": 1360, |
|
"steps": 1360 |
|
}, |
|
{ |
|
"epoch": 0.182, |
|
"learning_rate": 1.6867812414171932e-05, |
|
"loss": 0.615, |
|
"r_loss": 0.0, |
|
"step": 1370, |
|
"steps": 1370 |
|
}, |
|
{ |
|
"epoch": 0.184, |
|
"learning_rate": 1.6840361164515244e-05, |
|
"loss": 0.6203, |
|
"r_loss": 0.0, |
|
"step": 1380, |
|
"steps": 1380 |
|
}, |
|
{ |
|
"epoch": 0.185, |
|
"learning_rate": 1.6812909914858556e-05, |
|
"loss": 0.6144, |
|
"r_loss": 0.0, |
|
"step": 1390, |
|
"steps": 1390 |
|
}, |
|
{ |
|
"epoch": 0.186, |
|
"learning_rate": 1.6785458665201867e-05, |
|
"loss": 0.6156, |
|
"r_loss": 0.0, |
|
"step": 1400, |
|
"steps": 1400 |
|
}, |
|
{ |
|
"epoch": 0.188, |
|
"learning_rate": 1.6758007415545182e-05, |
|
"loss": 0.6251, |
|
"r_loss": 0.0, |
|
"step": 1410, |
|
"steps": 1410 |
|
}, |
|
{ |
|
"epoch": 0.189, |
|
"learning_rate": 1.6730556165888494e-05, |
|
"loss": 0.6285, |
|
"r_loss": 0.0, |
|
"step": 1420, |
|
"steps": 1420 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 1.6703104916231806e-05, |
|
"loss": 0.6205, |
|
"r_loss": 0.0, |
|
"step": 1430, |
|
"steps": 1430 |
|
}, |
|
{ |
|
"epoch": 0.192, |
|
"learning_rate": 1.667565366657512e-05, |
|
"loss": 0.6238, |
|
"r_loss": 0.0, |
|
"step": 1440, |
|
"steps": 1440 |
|
}, |
|
{ |
|
"epoch": 0.193, |
|
"learning_rate": 1.6648202416918432e-05, |
|
"loss": 0.6257, |
|
"r_loss": 0.0, |
|
"step": 1450, |
|
"steps": 1450 |
|
}, |
|
{ |
|
"epoch": 0.194, |
|
"learning_rate": 1.6620751167261744e-05, |
|
"loss": 0.6261, |
|
"r_loss": 0.0, |
|
"step": 1460, |
|
"steps": 1460 |
|
}, |
|
{ |
|
"epoch": 0.196, |
|
"learning_rate": 1.6593299917605056e-05, |
|
"loss": 0.6316, |
|
"r_loss": 0.0, |
|
"step": 1470, |
|
"steps": 1470 |
|
}, |
|
{ |
|
"epoch": 0.197, |
|
"learning_rate": 1.6565848667948367e-05, |
|
"loss": 0.6324, |
|
"r_loss": 0.0, |
|
"step": 1480, |
|
"steps": 1480 |
|
}, |
|
{ |
|
"epoch": 0.198, |
|
"learning_rate": 1.653839741829168e-05, |
|
"loss": 0.6045, |
|
"r_loss": 0.0, |
|
"step": 1490, |
|
"steps": 1490 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 1.6510946168634994e-05, |
|
"loss": 0.6287, |
|
"r_loss": 0.0, |
|
"step": 1500, |
|
"steps": 1500 |
|
}, |
|
{ |
|
"epoch": 0.201, |
|
"learning_rate": 1.6483494918978305e-05, |
|
"loss": 0.6268, |
|
"r_loss": 0.0, |
|
"step": 1510, |
|
"steps": 1510 |
|
}, |
|
{ |
|
"epoch": 0.202, |
|
"learning_rate": 1.6456043669321617e-05, |
|
"loss": 0.6218, |
|
"r_loss": 0.0, |
|
"step": 1520, |
|
"steps": 1520 |
|
}, |
|
{ |
|
"epoch": 0.204, |
|
"learning_rate": 1.642859241966493e-05, |
|
"loss": 0.6237, |
|
"r_loss": 0.0, |
|
"step": 1530, |
|
"steps": 1530 |
|
}, |
|
{ |
|
"epoch": 0.205, |
|
"learning_rate": 1.640114117000824e-05, |
|
"loss": 0.6134, |
|
"r_loss": 0.0, |
|
"step": 1540, |
|
"steps": 1540 |
|
}, |
|
{ |
|
"epoch": 0.206, |
|
"learning_rate": 1.6373689920351552e-05, |
|
"loss": 0.6046, |
|
"r_loss": 0.0, |
|
"step": 1550, |
|
"steps": 1550 |
|
}, |
|
{ |
|
"epoch": 0.208, |
|
"learning_rate": 1.6346238670694867e-05, |
|
"loss": 0.612, |
|
"r_loss": 0.0, |
|
"step": 1560, |
|
"steps": 1560 |
|
}, |
|
{ |
|
"epoch": 0.209, |
|
"learning_rate": 1.631878742103818e-05, |
|
"loss": 0.6143, |
|
"r_loss": 0.0, |
|
"step": 1570, |
|
"steps": 1570 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 1.629133617138149e-05, |
|
"loss": 0.6069, |
|
"r_loss": 0.0, |
|
"step": 1580, |
|
"steps": 1580 |
|
}, |
|
{ |
|
"epoch": 0.212, |
|
"learning_rate": 1.6263884921724802e-05, |
|
"loss": 0.6014, |
|
"r_loss": 0.0, |
|
"step": 1590, |
|
"steps": 1590 |
|
}, |
|
{ |
|
"epoch": 0.213, |
|
"learning_rate": 1.6236433672068114e-05, |
|
"loss": 0.6092, |
|
"r_loss": 0.0, |
|
"step": 1600, |
|
"steps": 1600 |
|
}, |
|
{ |
|
"epoch": 0.214, |
|
"learning_rate": 1.620898242241143e-05, |
|
"loss": 0.6284, |
|
"r_loss": 0.0, |
|
"step": 1610, |
|
"steps": 1610 |
|
}, |
|
{ |
|
"epoch": 0.216, |
|
"learning_rate": 1.618153117275474e-05, |
|
"loss": 0.6235, |
|
"r_loss": 0.0, |
|
"step": 1620, |
|
"steps": 1620 |
|
}, |
|
{ |
|
"epoch": 0.217, |
|
"learning_rate": 1.6154079923098052e-05, |
|
"loss": 0.6188, |
|
"r_loss": 0.0, |
|
"step": 1630, |
|
"steps": 1630 |
|
}, |
|
{ |
|
"epoch": 0.218, |
|
"learning_rate": 1.6126628673441363e-05, |
|
"loss": 0.6202, |
|
"r_loss": 0.0, |
|
"step": 1640, |
|
"steps": 1640 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 1.6099177423784675e-05, |
|
"loss": 0.6147, |
|
"r_loss": 0.0, |
|
"step": 1650, |
|
"steps": 1650 |
|
}, |
|
{ |
|
"epoch": 0.221, |
|
"learning_rate": 1.6071726174127987e-05, |
|
"loss": 0.6144, |
|
"r_loss": 0.0, |
|
"step": 1660, |
|
"steps": 1660 |
|
}, |
|
{ |
|
"epoch": 0.222, |
|
"learning_rate": 1.6044274924471302e-05, |
|
"loss": 0.6188, |
|
"r_loss": 0.0, |
|
"step": 1670, |
|
"steps": 1670 |
|
}, |
|
{ |
|
"epoch": 0.224, |
|
"learning_rate": 1.6016823674814613e-05, |
|
"loss": 0.611, |
|
"r_loss": 0.0, |
|
"step": 1680, |
|
"steps": 1680 |
|
}, |
|
{ |
|
"epoch": 0.225, |
|
"learning_rate": 1.5989372425157925e-05, |
|
"loss": 0.6108, |
|
"r_loss": 0.0, |
|
"step": 1690, |
|
"steps": 1690 |
|
}, |
|
{ |
|
"epoch": 0.226, |
|
"learning_rate": 1.5961921175501237e-05, |
|
"loss": 0.6087, |
|
"r_loss": 0.0, |
|
"step": 1700, |
|
"steps": 1700 |
|
}, |
|
{ |
|
"epoch": 0.228, |
|
"learning_rate": 1.593446992584455e-05, |
|
"loss": 0.609, |
|
"r_loss": 0.0, |
|
"step": 1710, |
|
"steps": 1710 |
|
}, |
|
{ |
|
"epoch": 0.229, |
|
"learning_rate": 1.590701867618786e-05, |
|
"loss": 0.6164, |
|
"r_loss": 0.0, |
|
"step": 1720, |
|
"steps": 1720 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 1.5879567426531175e-05, |
|
"loss": 0.6118, |
|
"r_loss": 0.0, |
|
"step": 1730, |
|
"steps": 1730 |
|
}, |
|
{ |
|
"epoch": 0.232, |
|
"learning_rate": 1.5852116176874487e-05, |
|
"loss": 0.603, |
|
"r_loss": 0.0, |
|
"step": 1740, |
|
"steps": 1740 |
|
}, |
|
{ |
|
"epoch": 0.233, |
|
"learning_rate": 1.5824664927217798e-05, |
|
"loss": 0.6195, |
|
"r_loss": 0.0, |
|
"step": 1750, |
|
"steps": 1750 |
|
}, |
|
{ |
|
"epoch": 0.234, |
|
"learning_rate": 1.5797213677561113e-05, |
|
"loss": 0.6108, |
|
"r_loss": 0.0, |
|
"step": 1760, |
|
"steps": 1760 |
|
}, |
|
{ |
|
"epoch": 0.236, |
|
"learning_rate": 1.576976242790442e-05, |
|
"loss": 0.6217, |
|
"r_loss": 0.0, |
|
"step": 1770, |
|
"steps": 1770 |
|
}, |
|
{ |
|
"epoch": 0.237, |
|
"learning_rate": 1.5742311178247737e-05, |
|
"loss": 0.6152, |
|
"r_loss": 0.0, |
|
"step": 1780, |
|
"steps": 1780 |
|
}, |
|
{ |
|
"epoch": 0.238, |
|
"learning_rate": 1.5714859928591048e-05, |
|
"loss": 0.6162, |
|
"r_loss": 0.0, |
|
"step": 1790, |
|
"steps": 1790 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 1.568740867893436e-05, |
|
"loss": 0.6152, |
|
"r_loss": 0.0, |
|
"step": 1800, |
|
"steps": 1800 |
|
}, |
|
{ |
|
"epoch": 0.241, |
|
"learning_rate": 1.5659957429277675e-05, |
|
"loss": 0.6155, |
|
"r_loss": 0.0, |
|
"step": 1810, |
|
"steps": 1810 |
|
}, |
|
{ |
|
"epoch": 0.242, |
|
"learning_rate": 1.5632506179620986e-05, |
|
"loss": 0.6119, |
|
"r_loss": 0.0, |
|
"step": 1820, |
|
"steps": 1820 |
|
}, |
|
{ |
|
"epoch": 0.244, |
|
"learning_rate": 1.5605054929964298e-05, |
|
"loss": 0.601, |
|
"r_loss": 0.0, |
|
"step": 1830, |
|
"steps": 1830 |
|
}, |
|
{ |
|
"epoch": 0.245, |
|
"learning_rate": 1.557760368030761e-05, |
|
"loss": 0.6039, |
|
"r_loss": 0.0, |
|
"step": 1840, |
|
"steps": 1840 |
|
}, |
|
{ |
|
"epoch": 0.246, |
|
"learning_rate": 1.555015243065092e-05, |
|
"loss": 0.6093, |
|
"r_loss": 0.0, |
|
"step": 1850, |
|
"steps": 1850 |
|
}, |
|
{ |
|
"epoch": 0.248, |
|
"learning_rate": 1.5522701180994233e-05, |
|
"loss": 0.6041, |
|
"r_loss": 0.0, |
|
"step": 1860, |
|
"steps": 1860 |
|
}, |
|
{ |
|
"epoch": 0.249, |
|
"learning_rate": 1.5495249931337548e-05, |
|
"loss": 0.6093, |
|
"r_loss": 0.0, |
|
"step": 1870, |
|
"steps": 1870 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 1.546779868168086e-05, |
|
"loss": 0.6155, |
|
"r_loss": 0.0, |
|
"step": 1880, |
|
"steps": 1880 |
|
}, |
|
{ |
|
"epoch": 0.252, |
|
"learning_rate": 1.544034743202417e-05, |
|
"loss": 0.6101, |
|
"r_loss": 0.0, |
|
"step": 1890, |
|
"steps": 1890 |
|
}, |
|
{ |
|
"epoch": 0.253, |
|
"learning_rate": 1.5412896182367483e-05, |
|
"loss": 0.6089, |
|
"r_loss": 0.0, |
|
"step": 1900, |
|
"steps": 1900 |
|
}, |
|
{ |
|
"epoch": 0.254, |
|
"learning_rate": 1.5385444932710795e-05, |
|
"loss": 0.6113, |
|
"r_loss": 0.0, |
|
"step": 1910, |
|
"steps": 1910 |
|
}, |
|
{ |
|
"epoch": 0.256, |
|
"learning_rate": 1.5357993683054106e-05, |
|
"loss": 0.6082, |
|
"r_loss": 0.0, |
|
"step": 1920, |
|
"steps": 1920 |
|
}, |
|
{ |
|
"epoch": 0.257, |
|
"learning_rate": 1.533054243339742e-05, |
|
"loss": 0.6079, |
|
"r_loss": 0.0, |
|
"step": 1930, |
|
"steps": 1930 |
|
}, |
|
{ |
|
"epoch": 0.258, |
|
"learning_rate": 1.5303091183740733e-05, |
|
"loss": 0.6023, |
|
"r_loss": 0.0, |
|
"step": 1940, |
|
"steps": 1940 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 1.5275639934084044e-05, |
|
"loss": 0.613, |
|
"r_loss": 0.0, |
|
"step": 1950, |
|
"steps": 1950 |
|
}, |
|
{ |
|
"epoch": 0.261, |
|
"learning_rate": 1.5248188684427356e-05, |
|
"loss": 0.611, |
|
"r_loss": 0.0, |
|
"step": 1960, |
|
"steps": 1960 |
|
}, |
|
{ |
|
"epoch": 0.262, |
|
"learning_rate": 1.5220737434770668e-05, |
|
"loss": 0.621, |
|
"r_loss": 0.0, |
|
"step": 1970, |
|
"steps": 1970 |
|
}, |
|
{ |
|
"epoch": 0.264, |
|
"learning_rate": 1.5193286185113981e-05, |
|
"loss": 0.6062, |
|
"r_loss": 0.0, |
|
"step": 1980, |
|
"steps": 1980 |
|
}, |
|
{ |
|
"epoch": 0.265, |
|
"learning_rate": 1.5165834935457293e-05, |
|
"loss": 0.6073, |
|
"r_loss": 0.0, |
|
"step": 1990, |
|
"steps": 1990 |
|
}, |
|
{ |
|
"epoch": 0.266, |
|
"learning_rate": 1.5138383685800604e-05, |
|
"loss": 0.6124, |
|
"r_loss": 0.0, |
|
"step": 2000, |
|
"steps": 2000 |
|
}, |
|
{ |
|
"epoch": 0.268, |
|
"learning_rate": 1.5110932436143918e-05, |
|
"loss": 0.5875, |
|
"r_loss": 0.0, |
|
"step": 2010, |
|
"steps": 2010 |
|
}, |
|
{ |
|
"epoch": 0.269, |
|
"learning_rate": 1.508348118648723e-05, |
|
"loss": 0.6045, |
|
"r_loss": 0.0, |
|
"step": 2020, |
|
"steps": 2020 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 1.5056029936830541e-05, |
|
"loss": 0.6004, |
|
"r_loss": 0.0, |
|
"step": 2030, |
|
"steps": 2030 |
|
}, |
|
{ |
|
"epoch": 0.272, |
|
"learning_rate": 1.5028578687173854e-05, |
|
"loss": 0.5976, |
|
"r_loss": 0.0, |
|
"step": 2040, |
|
"steps": 2040 |
|
}, |
|
{ |
|
"epoch": 0.273, |
|
"learning_rate": 1.5001127437517166e-05, |
|
"loss": 0.6102, |
|
"r_loss": 0.0, |
|
"step": 2050, |
|
"steps": 2050 |
|
}, |
|
{ |
|
"epoch": 0.274, |
|
"learning_rate": 1.4973676187860477e-05, |
|
"loss": 0.6055, |
|
"r_loss": 0.0, |
|
"step": 2060, |
|
"steps": 2060 |
|
}, |
|
{ |
|
"epoch": 0.276, |
|
"learning_rate": 1.494622493820379e-05, |
|
"loss": 0.6121, |
|
"r_loss": 0.0, |
|
"step": 2070, |
|
"steps": 2070 |
|
}, |
|
{ |
|
"epoch": 0.277, |
|
"learning_rate": 1.4918773688547102e-05, |
|
"loss": 0.5986, |
|
"r_loss": 0.0, |
|
"step": 2080, |
|
"steps": 2080 |
|
}, |
|
{ |
|
"epoch": 0.278, |
|
"learning_rate": 1.4891322438890414e-05, |
|
"loss": 0.5981, |
|
"r_loss": 0.0, |
|
"step": 2090, |
|
"steps": 2090 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 1.4863871189233727e-05, |
|
"loss": 0.6056, |
|
"r_loss": 0.0, |
|
"step": 2100, |
|
"steps": 2100 |
|
}, |
|
{ |
|
"epoch": 0.281, |
|
"learning_rate": 1.4836419939577039e-05, |
|
"loss": 0.6012, |
|
"r_loss": 0.0, |
|
"step": 2110, |
|
"steps": 2110 |
|
}, |
|
{ |
|
"epoch": 0.282, |
|
"learning_rate": 1.480896868992035e-05, |
|
"loss": 0.5978, |
|
"r_loss": 0.0, |
|
"step": 2120, |
|
"steps": 2120 |
|
}, |
|
{ |
|
"epoch": 0.284, |
|
"learning_rate": 1.4781517440263664e-05, |
|
"loss": 0.6054, |
|
"r_loss": 0.0, |
|
"step": 2130, |
|
"steps": 2130 |
|
}, |
|
{ |
|
"epoch": 0.285, |
|
"learning_rate": 1.4754066190606976e-05, |
|
"loss": 0.5973, |
|
"r_loss": 0.0, |
|
"step": 2140, |
|
"steps": 2140 |
|
}, |
|
{ |
|
"epoch": 0.286, |
|
"learning_rate": 1.4726614940950289e-05, |
|
"loss": 0.5924, |
|
"r_loss": 0.0, |
|
"step": 2150, |
|
"steps": 2150 |
|
}, |
|
{ |
|
"epoch": 0.288, |
|
"learning_rate": 1.46991636912936e-05, |
|
"loss": 0.608, |
|
"r_loss": 0.0, |
|
"step": 2160, |
|
"steps": 2160 |
|
}, |
|
{ |
|
"epoch": 0.289, |
|
"learning_rate": 1.4671712441636912e-05, |
|
"loss": 0.613, |
|
"r_loss": 0.0, |
|
"step": 2170, |
|
"steps": 2170 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 1.4644261191980226e-05, |
|
"loss": 0.5915, |
|
"r_loss": 0.0, |
|
"step": 2180, |
|
"steps": 2180 |
|
}, |
|
{ |
|
"epoch": 0.292, |
|
"learning_rate": 1.4616809942323537e-05, |
|
"loss": 0.6014, |
|
"r_loss": 0.0, |
|
"step": 2190, |
|
"steps": 2190 |
|
}, |
|
{ |
|
"epoch": 0.293, |
|
"learning_rate": 1.4589358692666849e-05, |
|
"loss": 0.5866, |
|
"r_loss": 0.0, |
|
"step": 2200, |
|
"steps": 2200 |
|
}, |
|
{ |
|
"epoch": 0.294, |
|
"learning_rate": 1.4561907443010162e-05, |
|
"loss": 0.6016, |
|
"r_loss": 0.0, |
|
"step": 2210, |
|
"steps": 2210 |
|
}, |
|
{ |
|
"epoch": 0.296, |
|
"learning_rate": 1.4534456193353474e-05, |
|
"loss": 0.5869, |
|
"r_loss": 0.0, |
|
"step": 2220, |
|
"steps": 2220 |
|
}, |
|
{ |
|
"epoch": 0.297, |
|
"learning_rate": 1.4507004943696785e-05, |
|
"loss": 0.5988, |
|
"r_loss": 0.0, |
|
"step": 2230, |
|
"steps": 2230 |
|
}, |
|
{ |
|
"epoch": 0.298, |
|
"learning_rate": 1.4479553694040099e-05, |
|
"loss": 0.6068, |
|
"r_loss": 0.0, |
|
"step": 2240, |
|
"steps": 2240 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 1.445210244438341e-05, |
|
"loss": 0.6059, |
|
"r_loss": 0.0, |
|
"step": 2250, |
|
"steps": 2250 |
|
}, |
|
{ |
|
"epoch": 0.301, |
|
"learning_rate": 1.4424651194726722e-05, |
|
"loss": 0.5972, |
|
"r_loss": 0.0, |
|
"step": 2260, |
|
"steps": 2260 |
|
}, |
|
{ |
|
"epoch": 0.302, |
|
"learning_rate": 1.4397199945070037e-05, |
|
"loss": 0.5969, |
|
"r_loss": 0.0, |
|
"step": 2270, |
|
"steps": 2270 |
|
}, |
|
{ |
|
"epoch": 0.304, |
|
"learning_rate": 1.4369748695413347e-05, |
|
"loss": 0.5935, |
|
"r_loss": 0.0, |
|
"step": 2280, |
|
"steps": 2280 |
|
}, |
|
{ |
|
"epoch": 0.305, |
|
"learning_rate": 1.4342297445756659e-05, |
|
"loss": 0.5932, |
|
"r_loss": 0.0, |
|
"step": 2290, |
|
"steps": 2290 |
|
}, |
|
{ |
|
"epoch": 0.306, |
|
"learning_rate": 1.4314846196099974e-05, |
|
"loss": 0.6008, |
|
"r_loss": 0.0, |
|
"step": 2300, |
|
"steps": 2300 |
|
}, |
|
{ |
|
"epoch": 0.308, |
|
"learning_rate": 1.4287394946443284e-05, |
|
"loss": 0.5898, |
|
"r_loss": 0.0, |
|
"step": 2310, |
|
"steps": 2310 |
|
}, |
|
{ |
|
"epoch": 0.309, |
|
"learning_rate": 1.4259943696786599e-05, |
|
"loss": 0.6052, |
|
"r_loss": 0.0, |
|
"step": 2320, |
|
"steps": 2320 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 1.423249244712991e-05, |
|
"loss": 0.5929, |
|
"r_loss": 0.0, |
|
"step": 2330, |
|
"steps": 2330 |
|
}, |
|
{ |
|
"epoch": 0.312, |
|
"learning_rate": 1.4205041197473222e-05, |
|
"loss": 0.5946, |
|
"r_loss": 0.0, |
|
"step": 2340, |
|
"steps": 2340 |
|
}, |
|
{ |
|
"epoch": 0.313, |
|
"learning_rate": 1.4177589947816535e-05, |
|
"loss": 0.5973, |
|
"r_loss": 0.0, |
|
"step": 2350, |
|
"steps": 2350 |
|
}, |
|
{ |
|
"epoch": 0.314, |
|
"learning_rate": 1.4150138698159847e-05, |
|
"loss": 0.6, |
|
"r_loss": 0.0, |
|
"step": 2360, |
|
"steps": 2360 |
|
}, |
|
{ |
|
"epoch": 0.316, |
|
"learning_rate": 1.4122687448503158e-05, |
|
"loss": 0.582, |
|
"r_loss": 0.0, |
|
"step": 2370, |
|
"steps": 2370 |
|
}, |
|
{ |
|
"epoch": 0.317, |
|
"learning_rate": 1.4095236198846472e-05, |
|
"loss": 0.5976, |
|
"r_loss": 0.0, |
|
"step": 2380, |
|
"steps": 2380 |
|
}, |
|
{ |
|
"epoch": 0.318, |
|
"learning_rate": 1.4067784949189783e-05, |
|
"loss": 0.5907, |
|
"r_loss": 0.0, |
|
"step": 2390, |
|
"steps": 2390 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 1.4040333699533095e-05, |
|
"loss": 0.6032, |
|
"r_loss": 0.0, |
|
"step": 2400, |
|
"steps": 2400 |
|
}, |
|
{ |
|
"epoch": 0.321, |
|
"learning_rate": 1.4012882449876408e-05, |
|
"loss": 0.6027, |
|
"r_loss": 0.0, |
|
"step": 2410, |
|
"steps": 2410 |
|
}, |
|
{ |
|
"epoch": 0.322, |
|
"learning_rate": 1.398543120021972e-05, |
|
"loss": 0.5967, |
|
"r_loss": 0.0, |
|
"step": 2420, |
|
"steps": 2420 |
|
}, |
|
{ |
|
"epoch": 0.324, |
|
"learning_rate": 1.3957979950563032e-05, |
|
"loss": 0.5966, |
|
"r_loss": 0.0, |
|
"step": 2430, |
|
"steps": 2430 |
|
}, |
|
{ |
|
"epoch": 0.325, |
|
"learning_rate": 1.3930528700906345e-05, |
|
"loss": 0.5963, |
|
"r_loss": 0.0, |
|
"step": 2440, |
|
"steps": 2440 |
|
}, |
|
{ |
|
"epoch": 0.326, |
|
"learning_rate": 1.3903077451249657e-05, |
|
"loss": 0.5932, |
|
"r_loss": 0.0, |
|
"step": 2450, |
|
"steps": 2450 |
|
}, |
|
{ |
|
"epoch": 0.328, |
|
"learning_rate": 1.3875626201592968e-05, |
|
"loss": 0.5934, |
|
"r_loss": 0.0, |
|
"step": 2460, |
|
"steps": 2460 |
|
}, |
|
{ |
|
"epoch": 0.329, |
|
"learning_rate": 1.3848174951936282e-05, |
|
"loss": 0.588, |
|
"r_loss": 0.0, |
|
"step": 2470, |
|
"steps": 2470 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 1.3820723702279593e-05, |
|
"loss": 0.593, |
|
"r_loss": 0.0, |
|
"step": 2480, |
|
"steps": 2480 |
|
}, |
|
{ |
|
"epoch": 0.332, |
|
"learning_rate": 1.3793272452622905e-05, |
|
"loss": 0.5936, |
|
"r_loss": 0.0, |
|
"step": 2490, |
|
"steps": 2490 |
|
}, |
|
{ |
|
"epoch": 0.333, |
|
"learning_rate": 1.3765821202966218e-05, |
|
"loss": 0.5741, |
|
"r_loss": 0.0, |
|
"step": 2500, |
|
"steps": 2500 |
|
}, |
|
{ |
|
"epoch": 0.334, |
|
"learning_rate": 1.373836995330953e-05, |
|
"loss": 0.6024, |
|
"r_loss": 0.0, |
|
"step": 2510, |
|
"steps": 2510 |
|
}, |
|
{ |
|
"epoch": 0.336, |
|
"learning_rate": 1.3710918703652843e-05, |
|
"loss": 0.5808, |
|
"r_loss": 0.0, |
|
"step": 2520, |
|
"steps": 2520 |
|
}, |
|
{ |
|
"epoch": 0.337, |
|
"learning_rate": 1.3683467453996155e-05, |
|
"loss": 0.6057, |
|
"r_loss": 0.0, |
|
"step": 2530, |
|
"steps": 2530 |
|
}, |
|
{ |
|
"epoch": 0.338, |
|
"learning_rate": 1.3656016204339466e-05, |
|
"loss": 0.589, |
|
"r_loss": 0.0, |
|
"step": 2540, |
|
"steps": 2540 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 1.362856495468278e-05, |
|
"loss": 0.6035, |
|
"r_loss": 0.0, |
|
"step": 2550, |
|
"steps": 2550 |
|
}, |
|
{ |
|
"epoch": 0.341, |
|
"learning_rate": 1.3601113705026091e-05, |
|
"loss": 0.5855, |
|
"r_loss": 0.0, |
|
"step": 2560, |
|
"steps": 2560 |
|
}, |
|
{ |
|
"epoch": 0.342, |
|
"learning_rate": 1.3573662455369403e-05, |
|
"loss": 0.582, |
|
"r_loss": 0.0, |
|
"step": 2570, |
|
"steps": 2570 |
|
}, |
|
{ |
|
"epoch": 0.344, |
|
"learning_rate": 1.3546211205712716e-05, |
|
"loss": 0.5968, |
|
"r_loss": 0.0, |
|
"step": 2580, |
|
"steps": 2580 |
|
}, |
|
{ |
|
"epoch": 0.345, |
|
"learning_rate": 1.3518759956056028e-05, |
|
"loss": 0.58, |
|
"r_loss": 0.0, |
|
"step": 2590, |
|
"steps": 2590 |
|
}, |
|
{ |
|
"epoch": 0.346, |
|
"learning_rate": 1.349130870639934e-05, |
|
"loss": 0.5839, |
|
"r_loss": 0.0, |
|
"step": 2600, |
|
"steps": 2600 |
|
}, |
|
{ |
|
"epoch": 0.348, |
|
"learning_rate": 1.3463857456742653e-05, |
|
"loss": 0.5938, |
|
"r_loss": 0.0, |
|
"step": 2610, |
|
"steps": 2610 |
|
}, |
|
{ |
|
"epoch": 0.349, |
|
"learning_rate": 1.3441896457017303e-05, |
|
"loss": 0.5965, |
|
"r_loss": 0.0, |
|
"step": 2620, |
|
"steps": 2620 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 1.3414445207360614e-05, |
|
"loss": 0.5904, |
|
"r_loss": 0.0, |
|
"step": 2630, |
|
"steps": 2630 |
|
}, |
|
{ |
|
"epoch": 0.352, |
|
"learning_rate": 1.3386993957703927e-05, |
|
"loss": 0.5884, |
|
"r_loss": 0.0, |
|
"step": 2640, |
|
"steps": 2640 |
|
}, |
|
{ |
|
"epoch": 0.353, |
|
"learning_rate": 1.3359542708047239e-05, |
|
"loss": 0.6003, |
|
"r_loss": 0.0, |
|
"step": 2650, |
|
"steps": 2650 |
|
}, |
|
{ |
|
"epoch": 0.354, |
|
"learning_rate": 1.333209145839055e-05, |
|
"loss": 0.6024, |
|
"r_loss": 0.0, |
|
"step": 2660, |
|
"steps": 2660 |
|
}, |
|
{ |
|
"epoch": 0.356, |
|
"learning_rate": 1.3304640208733864e-05, |
|
"loss": 0.5919, |
|
"r_loss": 0.0, |
|
"step": 2670, |
|
"steps": 2670 |
|
}, |
|
{ |
|
"epoch": 0.357, |
|
"learning_rate": 1.3277188959077176e-05, |
|
"loss": 0.5792, |
|
"r_loss": 0.0, |
|
"step": 2680, |
|
"steps": 2680 |
|
}, |
|
{ |
|
"epoch": 0.358, |
|
"learning_rate": 1.3249737709420487e-05, |
|
"loss": 0.5788, |
|
"r_loss": 0.0, |
|
"step": 2690, |
|
"steps": 2690 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 1.3222286459763802e-05, |
|
"loss": 0.5836, |
|
"r_loss": 0.0, |
|
"step": 2700, |
|
"steps": 2700 |
|
}, |
|
{ |
|
"epoch": 0.361, |
|
"learning_rate": 1.3194835210107112e-05, |
|
"loss": 0.5938, |
|
"r_loss": 0.0, |
|
"step": 2710, |
|
"steps": 2710 |
|
}, |
|
{ |
|
"epoch": 0.362, |
|
"learning_rate": 1.3167383960450427e-05, |
|
"loss": 0.5865, |
|
"r_loss": 0.0, |
|
"step": 2720, |
|
"steps": 2720 |
|
}, |
|
{ |
|
"epoch": 0.364, |
|
"learning_rate": 1.3139932710793739e-05, |
|
"loss": 0.585, |
|
"r_loss": 0.0, |
|
"step": 2730, |
|
"steps": 2730 |
|
}, |
|
{ |
|
"epoch": 0.365, |
|
"learning_rate": 1.311248146113705e-05, |
|
"loss": 0.5785, |
|
"r_loss": 0.0, |
|
"step": 2740, |
|
"steps": 2740 |
|
}, |
|
{ |
|
"epoch": 0.366, |
|
"learning_rate": 1.3085030211480364e-05, |
|
"loss": 0.5884, |
|
"r_loss": 0.0, |
|
"step": 2750, |
|
"steps": 2750 |
|
}, |
|
{ |
|
"epoch": 0.368, |
|
"learning_rate": 1.3057578961823676e-05, |
|
"loss": 0.5992, |
|
"r_loss": 0.0, |
|
"step": 2760, |
|
"steps": 2760 |
|
}, |
|
{ |
|
"epoch": 0.369, |
|
"learning_rate": 1.3030127712166987e-05, |
|
"loss": 0.5862, |
|
"r_loss": 0.0, |
|
"step": 2770, |
|
"steps": 2770 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 1.30026764625103e-05, |
|
"loss": 0.5929, |
|
"r_loss": 0.0, |
|
"step": 2780, |
|
"steps": 2780 |
|
}, |
|
{ |
|
"epoch": 0.372, |
|
"learning_rate": 1.2975225212853612e-05, |
|
"loss": 0.5876, |
|
"r_loss": 0.0, |
|
"step": 2790, |
|
"steps": 2790 |
|
}, |
|
{ |
|
"epoch": 0.373, |
|
"learning_rate": 1.2947773963196924e-05, |
|
"loss": 0.5741, |
|
"r_loss": 0.0, |
|
"step": 2800, |
|
"steps": 2800 |
|
}, |
|
{ |
|
"epoch": 0.374, |
|
"learning_rate": 1.2920322713540237e-05, |
|
"loss": 0.5762, |
|
"r_loss": 0.0, |
|
"step": 2810, |
|
"steps": 2810 |
|
}, |
|
{ |
|
"epoch": 0.376, |
|
"learning_rate": 1.2892871463883549e-05, |
|
"loss": 0.5776, |
|
"r_loss": 0.0, |
|
"step": 2820, |
|
"steps": 2820 |
|
}, |
|
{ |
|
"epoch": 0.377, |
|
"learning_rate": 1.286542021422686e-05, |
|
"loss": 0.5848, |
|
"r_loss": 0.0, |
|
"step": 2830, |
|
"steps": 2830 |
|
}, |
|
{ |
|
"epoch": 0.378, |
|
"learning_rate": 1.2837968964570174e-05, |
|
"loss": 0.575, |
|
"r_loss": 0.0, |
|
"step": 2840, |
|
"steps": 2840 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 1.2810517714913485e-05, |
|
"loss": 0.5819, |
|
"r_loss": 0.0, |
|
"step": 2850, |
|
"steps": 2850 |
|
}, |
|
{ |
|
"epoch": 0.381, |
|
"learning_rate": 1.2783066465256797e-05, |
|
"loss": 0.5886, |
|
"r_loss": 0.0, |
|
"step": 2860, |
|
"steps": 2860 |
|
}, |
|
{ |
|
"epoch": 0.382, |
|
"learning_rate": 1.275561521560011e-05, |
|
"loss": 0.5836, |
|
"r_loss": 0.0, |
|
"step": 2870, |
|
"steps": 2870 |
|
}, |
|
{ |
|
"epoch": 0.384, |
|
"learning_rate": 1.2728163965943422e-05, |
|
"loss": 0.5787, |
|
"r_loss": 0.0, |
|
"step": 2880, |
|
"steps": 2880 |
|
}, |
|
{ |
|
"epoch": 0.385, |
|
"learning_rate": 1.2700712716286734e-05, |
|
"loss": 0.582, |
|
"r_loss": 0.0, |
|
"step": 2890, |
|
"steps": 2890 |
|
}, |
|
{ |
|
"epoch": 0.386, |
|
"learning_rate": 1.2673261466630047e-05, |
|
"loss": 0.571, |
|
"r_loss": 0.0, |
|
"step": 2900, |
|
"steps": 2900 |
|
}, |
|
{ |
|
"epoch": 0.388, |
|
"learning_rate": 1.2645810216973359e-05, |
|
"loss": 0.5947, |
|
"r_loss": 0.0, |
|
"step": 2910, |
|
"steps": 2910 |
|
}, |
|
{ |
|
"epoch": 0.389, |
|
"learning_rate": 1.2618358967316672e-05, |
|
"loss": 0.5769, |
|
"r_loss": 0.0, |
|
"step": 2920, |
|
"steps": 2920 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 1.2590907717659983e-05, |
|
"loss": 0.5744, |
|
"r_loss": 0.0, |
|
"step": 2930, |
|
"steps": 2930 |
|
}, |
|
{ |
|
"epoch": 0.392, |
|
"learning_rate": 1.2563456468003295e-05, |
|
"loss": 0.5913, |
|
"r_loss": 0.0, |
|
"step": 2940, |
|
"steps": 2940 |
|
}, |
|
{ |
|
"epoch": 0.393, |
|
"learning_rate": 1.2536005218346608e-05, |
|
"loss": 0.5904, |
|
"r_loss": 0.0, |
|
"step": 2950, |
|
"steps": 2950 |
|
}, |
|
{ |
|
"epoch": 0.394, |
|
"learning_rate": 1.250855396868992e-05, |
|
"loss": 0.5814, |
|
"r_loss": 0.0, |
|
"step": 2960, |
|
"steps": 2960 |
|
}, |
|
{ |
|
"epoch": 0.396, |
|
"learning_rate": 1.2481102719033232e-05, |
|
"loss": 0.5824, |
|
"r_loss": 0.0, |
|
"step": 2970, |
|
"steps": 2970 |
|
}, |
|
{ |
|
"epoch": 0.397, |
|
"learning_rate": 1.2453651469376545e-05, |
|
"loss": 0.5823, |
|
"r_loss": 0.0, |
|
"step": 2980, |
|
"steps": 2980 |
|
}, |
|
{ |
|
"epoch": 0.398, |
|
"learning_rate": 1.2426200219719857e-05, |
|
"loss": 0.586, |
|
"r_loss": 0.0, |
|
"step": 2990, |
|
"steps": 2990 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 1.2398748970063168e-05, |
|
"loss": 0.5777, |
|
"r_loss": 0.0, |
|
"step": 3000, |
|
"steps": 3000 |
|
}, |
|
{ |
|
"epoch": 0.401, |
|
"learning_rate": 1.2371297720406482e-05, |
|
"loss": 0.5626, |
|
"r_loss": 0.0, |
|
"step": 3010, |
|
"steps": 3010 |
|
}, |
|
{ |
|
"epoch": 0.402, |
|
"learning_rate": 1.2343846470749793e-05, |
|
"loss": 0.5889, |
|
"r_loss": 0.0, |
|
"step": 3020, |
|
"steps": 3020 |
|
}, |
|
{ |
|
"epoch": 0.404, |
|
"learning_rate": 1.2316395221093105e-05, |
|
"loss": 0.5844, |
|
"r_loss": 0.0, |
|
"step": 3030, |
|
"steps": 3030 |
|
}, |
|
{ |
|
"epoch": 0.405, |
|
"learning_rate": 1.228894397143642e-05, |
|
"loss": 0.5813, |
|
"r_loss": 0.0, |
|
"step": 3040, |
|
"steps": 3040 |
|
}, |
|
{ |
|
"epoch": 0.406, |
|
"learning_rate": 1.226149272177973e-05, |
|
"loss": 0.5916, |
|
"r_loss": 0.0, |
|
"step": 3050, |
|
"steps": 3050 |
|
}, |
|
{ |
|
"epoch": 0.408, |
|
"learning_rate": 1.2234041472123041e-05, |
|
"loss": 0.5791, |
|
"r_loss": 0.0, |
|
"step": 3060, |
|
"steps": 3060 |
|
}, |
|
{ |
|
"epoch": 0.409, |
|
"learning_rate": 1.2206590222466356e-05, |
|
"loss": 0.5968, |
|
"r_loss": 0.0, |
|
"step": 3070, |
|
"steps": 3070 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 1.2179138972809666e-05, |
|
"loss": 0.5795, |
|
"r_loss": 0.0, |
|
"step": 3080, |
|
"steps": 3080 |
|
}, |
|
{ |
|
"epoch": 0.412, |
|
"learning_rate": 1.2151687723152981e-05, |
|
"loss": 0.5632, |
|
"r_loss": 0.0, |
|
"step": 3090, |
|
"steps": 3090 |
|
}, |
|
{ |
|
"epoch": 0.413, |
|
"learning_rate": 1.2124236473496293e-05, |
|
"loss": 0.5717, |
|
"r_loss": 0.0, |
|
"step": 3100, |
|
"steps": 3100 |
|
}, |
|
{ |
|
"epoch": 0.414, |
|
"learning_rate": 1.2096785223839605e-05, |
|
"loss": 0.5803, |
|
"r_loss": 0.0, |
|
"step": 3110, |
|
"steps": 3110 |
|
}, |
|
{ |
|
"epoch": 0.416, |
|
"learning_rate": 1.2069333974182918e-05, |
|
"loss": 0.585, |
|
"r_loss": 0.0, |
|
"step": 3120, |
|
"steps": 3120 |
|
}, |
|
{ |
|
"epoch": 0.417, |
|
"learning_rate": 1.204188272452623e-05, |
|
"loss": 0.5777, |
|
"r_loss": 0.0, |
|
"step": 3130, |
|
"steps": 3130 |
|
}, |
|
{ |
|
"epoch": 0.418, |
|
"learning_rate": 1.2014431474869541e-05, |
|
"loss": 0.5785, |
|
"r_loss": 0.0, |
|
"step": 3140, |
|
"steps": 3140 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 1.1986980225212855e-05, |
|
"loss": 0.5854, |
|
"r_loss": 0.0, |
|
"step": 3150, |
|
"steps": 3150 |
|
}, |
|
{ |
|
"epoch": 0.421, |
|
"learning_rate": 1.1959528975556166e-05, |
|
"loss": 0.5809, |
|
"r_loss": 0.0, |
|
"step": 3160, |
|
"steps": 3160 |
|
}, |
|
{ |
|
"epoch": 0.422, |
|
"learning_rate": 1.1932077725899478e-05, |
|
"loss": 0.5864, |
|
"r_loss": 0.0, |
|
"step": 3170, |
|
"steps": 3170 |
|
}, |
|
{ |
|
"epoch": 0.424, |
|
"learning_rate": 1.1904626476242791e-05, |
|
"loss": 0.5707, |
|
"r_loss": 0.0, |
|
"step": 3180, |
|
"steps": 3180 |
|
}, |
|
{ |
|
"epoch": 0.425, |
|
"learning_rate": 1.1877175226586103e-05, |
|
"loss": 0.5837, |
|
"r_loss": 0.0, |
|
"step": 3190, |
|
"steps": 3190 |
|
}, |
|
{ |
|
"epoch": 0.426, |
|
"learning_rate": 1.1849723976929414e-05, |
|
"loss": 0.569, |
|
"r_loss": 0.0, |
|
"step": 3200, |
|
"steps": 3200 |
|
}, |
|
{ |
|
"epoch": 0.428, |
|
"learning_rate": 1.1822272727272728e-05, |
|
"loss": 0.5645, |
|
"r_loss": 0.0, |
|
"step": 3210, |
|
"steps": 3210 |
|
}, |
|
{ |
|
"epoch": 0.429, |
|
"learning_rate": 1.179482147761604e-05, |
|
"loss": 0.5774, |
|
"r_loss": 0.0, |
|
"step": 3220, |
|
"steps": 3220 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 1.1767370227959351e-05, |
|
"loss": 0.5816, |
|
"r_loss": 0.0, |
|
"step": 3230, |
|
"steps": 3230 |
|
}, |
|
{ |
|
"epoch": 0.432, |
|
"learning_rate": 1.1739918978302664e-05, |
|
"loss": 0.5741, |
|
"r_loss": 0.0, |
|
"step": 3240, |
|
"steps": 3240 |
|
}, |
|
{ |
|
"epoch": 0.433, |
|
"learning_rate": 1.1712467728645976e-05, |
|
"loss": 0.5811, |
|
"r_loss": 0.0, |
|
"step": 3250, |
|
"steps": 3250 |
|
}, |
|
{ |
|
"epoch": 0.434, |
|
"learning_rate": 1.1685016478989288e-05, |
|
"loss": 0.5806, |
|
"r_loss": 0.0, |
|
"step": 3260, |
|
"steps": 3260 |
|
}, |
|
{ |
|
"epoch": 0.436, |
|
"learning_rate": 1.1657565229332601e-05, |
|
"loss": 0.571, |
|
"r_loss": 0.0, |
|
"step": 3270, |
|
"steps": 3270 |
|
}, |
|
{ |
|
"epoch": 0.437, |
|
"learning_rate": 1.1630113979675913e-05, |
|
"loss": 0.577, |
|
"r_loss": 0.0, |
|
"step": 3280, |
|
"steps": 3280 |
|
}, |
|
{ |
|
"epoch": 0.438, |
|
"learning_rate": 1.1602662730019226e-05, |
|
"loss": 0.5806, |
|
"r_loss": 0.0, |
|
"step": 3290, |
|
"steps": 3290 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 1.1575211480362538e-05, |
|
"loss": 0.5725, |
|
"r_loss": 0.0, |
|
"step": 3300, |
|
"steps": 3300 |
|
}, |
|
{ |
|
"epoch": 0.441, |
|
"learning_rate": 1.154776023070585e-05, |
|
"loss": 0.583, |
|
"r_loss": 0.0, |
|
"step": 3310, |
|
"steps": 3310 |
|
}, |
|
{ |
|
"epoch": 0.442, |
|
"learning_rate": 1.1520308981049163e-05, |
|
"loss": 0.5607, |
|
"r_loss": 0.0, |
|
"step": 3320, |
|
"steps": 3320 |
|
}, |
|
{ |
|
"epoch": 0.444, |
|
"learning_rate": 1.1492857731392474e-05, |
|
"loss": 0.5822, |
|
"r_loss": 0.0, |
|
"step": 3330, |
|
"steps": 3330 |
|
}, |
|
{ |
|
"epoch": 0.445, |
|
"learning_rate": 1.1465406481735786e-05, |
|
"loss": 0.5712, |
|
"r_loss": 0.0, |
|
"step": 3340, |
|
"steps": 3340 |
|
}, |
|
{ |
|
"epoch": 0.446, |
|
"learning_rate": 1.1437955232079099e-05, |
|
"loss": 0.5704, |
|
"r_loss": 0.0, |
|
"step": 3350, |
|
"steps": 3350 |
|
}, |
|
{ |
|
"epoch": 0.448, |
|
"learning_rate": 1.141050398242241e-05, |
|
"loss": 0.5645, |
|
"r_loss": 0.0, |
|
"step": 3360, |
|
"steps": 3360 |
|
}, |
|
{ |
|
"epoch": 0.449, |
|
"learning_rate": 1.1383052732765722e-05, |
|
"loss": 0.5697, |
|
"r_loss": 0.0, |
|
"step": 3370, |
|
"steps": 3370 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 1.1355601483109036e-05, |
|
"loss": 0.5835, |
|
"r_loss": 0.0, |
|
"step": 3380, |
|
"steps": 3380 |
|
}, |
|
{ |
|
"epoch": 0.452, |
|
"learning_rate": 1.1328150233452347e-05, |
|
"loss": 0.5855, |
|
"r_loss": 0.0, |
|
"step": 3390, |
|
"steps": 3390 |
|
}, |
|
{ |
|
"epoch": 0.453, |
|
"learning_rate": 1.1300698983795659e-05, |
|
"loss": 0.5644, |
|
"r_loss": 0.0, |
|
"step": 3400, |
|
"steps": 3400 |
|
}, |
|
{ |
|
"epoch": 0.454, |
|
"learning_rate": 1.1273247734138974e-05, |
|
"loss": 0.5634, |
|
"r_loss": 0.0, |
|
"step": 3410, |
|
"steps": 3410 |
|
}, |
|
{ |
|
"epoch": 0.456, |
|
"learning_rate": 1.1245796484482284e-05, |
|
"loss": 0.5747, |
|
"r_loss": 0.0, |
|
"step": 3420, |
|
"steps": 3420 |
|
}, |
|
{ |
|
"epoch": 0.457, |
|
"learning_rate": 1.1218345234825596e-05, |
|
"loss": 0.5751, |
|
"r_loss": 0.0, |
|
"step": 3430, |
|
"steps": 3430 |
|
}, |
|
{ |
|
"epoch": 0.458, |
|
"learning_rate": 1.119089398516891e-05, |
|
"loss": 0.5728, |
|
"r_loss": 0.0, |
|
"step": 3440, |
|
"steps": 3440 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 1.1163442735512222e-05, |
|
"loss": 0.5642, |
|
"r_loss": 0.0, |
|
"step": 3450, |
|
"steps": 3450 |
|
}, |
|
{ |
|
"epoch": 0.461, |
|
"learning_rate": 1.1135991485855536e-05, |
|
"loss": 0.5725, |
|
"r_loss": 0.0, |
|
"step": 3460, |
|
"steps": 3460 |
|
}, |
|
{ |
|
"epoch": 0.462, |
|
"learning_rate": 1.1108540236198847e-05, |
|
"loss": 0.5699, |
|
"r_loss": 0.0, |
|
"step": 3470, |
|
"steps": 3470 |
|
}, |
|
{ |
|
"epoch": 0.464, |
|
"learning_rate": 1.1081088986542159e-05, |
|
"loss": 0.5726, |
|
"r_loss": 0.0, |
|
"step": 3480, |
|
"steps": 3480 |
|
}, |
|
{ |
|
"epoch": 0.465, |
|
"learning_rate": 1.1053637736885472e-05, |
|
"loss": 0.5708, |
|
"r_loss": 0.0, |
|
"step": 3490, |
|
"steps": 3490 |
|
}, |
|
{ |
|
"epoch": 0.466, |
|
"learning_rate": 1.1026186487228784e-05, |
|
"loss": 0.5739, |
|
"r_loss": 0.0, |
|
"step": 3500, |
|
"steps": 3500 |
|
}, |
|
{ |
|
"epoch": 0.468, |
|
"learning_rate": 1.0998735237572095e-05, |
|
"loss": 0.5696, |
|
"r_loss": 0.0, |
|
"step": 3510, |
|
"steps": 3510 |
|
}, |
|
{ |
|
"epoch": 0.469, |
|
"learning_rate": 1.0971283987915409e-05, |
|
"loss": 0.5792, |
|
"r_loss": 0.0, |
|
"step": 3520, |
|
"steps": 3520 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 1.094383273825872e-05, |
|
"loss": 0.5677, |
|
"r_loss": 0.0, |
|
"step": 3530, |
|
"steps": 3530 |
|
}, |
|
{ |
|
"epoch": 0.471, |
|
"learning_rate": 1.0916381488602032e-05, |
|
"loss": 0.5726, |
|
"r_loss": 0.0, |
|
"step": 3540, |
|
"steps": 3540 |
|
}, |
|
{ |
|
"epoch": 0.473, |
|
"learning_rate": 1.0888930238945345e-05, |
|
"loss": 0.5643, |
|
"r_loss": 0.0, |
|
"step": 3550, |
|
"steps": 3550 |
|
}, |
|
{ |
|
"epoch": 0.474, |
|
"learning_rate": 1.0861478989288657e-05, |
|
"loss": 0.5656, |
|
"r_loss": 0.0, |
|
"step": 3560, |
|
"steps": 3560 |
|
}, |
|
{ |
|
"epoch": 0.475, |
|
"learning_rate": 1.0834027739631969e-05, |
|
"loss": 0.5645, |
|
"r_loss": 0.0, |
|
"step": 3570, |
|
"steps": 3570 |
|
}, |
|
{ |
|
"epoch": 0.477, |
|
"learning_rate": 1.0806576489975282e-05, |
|
"loss": 0.5779, |
|
"r_loss": 0.0, |
|
"step": 3580, |
|
"steps": 3580 |
|
}, |
|
{ |
|
"epoch": 0.478, |
|
"learning_rate": 1.0779125240318594e-05, |
|
"loss": 0.5722, |
|
"r_loss": 0.0, |
|
"step": 3590, |
|
"steps": 3590 |
|
}, |
|
{ |
|
"epoch": 0.479, |
|
"learning_rate": 1.0751673990661905e-05, |
|
"loss": 0.5718, |
|
"r_loss": 0.0, |
|
"step": 3600, |
|
"steps": 3600 |
|
}, |
|
{ |
|
"epoch": 0.481, |
|
"learning_rate": 1.0724222741005219e-05, |
|
"loss": 0.5662, |
|
"r_loss": 0.0, |
|
"step": 3610, |
|
"steps": 3610 |
|
}, |
|
{ |
|
"epoch": 0.482, |
|
"learning_rate": 1.0699516616314198e-05, |
|
"loss": 0.561, |
|
"r_loss": 0.0, |
|
"step": 3620, |
|
"steps": 3620 |
|
}, |
|
{ |
|
"epoch": 0.483, |
|
"learning_rate": 1.067481049162318e-05, |
|
"loss": 0.5697, |
|
"r_loss": 0.0, |
|
"step": 3630, |
|
"steps": 3630 |
|
}, |
|
{ |
|
"epoch": 0.485, |
|
"learning_rate": 1.0647359241966493e-05, |
|
"loss": 0.5617, |
|
"r_loss": 0.0, |
|
"step": 3640, |
|
"steps": 3640 |
|
}, |
|
{ |
|
"epoch": 0.486, |
|
"learning_rate": 1.0619907992309805e-05, |
|
"loss": 0.5542, |
|
"r_loss": 0.0, |
|
"step": 3650, |
|
"steps": 3650 |
|
}, |
|
{ |
|
"epoch": 0.487, |
|
"learning_rate": 1.0592456742653116e-05, |
|
"loss": 0.5873, |
|
"r_loss": 0.0, |
|
"step": 3660, |
|
"steps": 3660 |
|
}, |
|
{ |
|
"epoch": 0.489, |
|
"learning_rate": 1.056500549299643e-05, |
|
"loss": 0.5715, |
|
"r_loss": 0.0, |
|
"step": 3670, |
|
"steps": 3670 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 1.0537554243339741e-05, |
|
"loss": 0.5665, |
|
"r_loss": 0.0, |
|
"step": 3680, |
|
"steps": 3680 |
|
}, |
|
{ |
|
"epoch": 0.491, |
|
"learning_rate": 1.0510102993683055e-05, |
|
"loss": 0.5635, |
|
"r_loss": 0.0, |
|
"step": 3690, |
|
"steps": 3690 |
|
}, |
|
{ |
|
"epoch": 0.493, |
|
"learning_rate": 1.0482651744026366e-05, |
|
"loss": 0.5601, |
|
"r_loss": 0.0, |
|
"step": 3700, |
|
"steps": 3700 |
|
}, |
|
{ |
|
"epoch": 0.494, |
|
"learning_rate": 1.0455200494369678e-05, |
|
"loss": 0.5665, |
|
"r_loss": 0.0, |
|
"step": 3710, |
|
"steps": 3710 |
|
}, |
|
{ |
|
"epoch": 0.495, |
|
"learning_rate": 1.0427749244712991e-05, |
|
"loss": 0.5684, |
|
"r_loss": 0.0, |
|
"step": 3720, |
|
"steps": 3720 |
|
}, |
|
{ |
|
"epoch": 0.497, |
|
"learning_rate": 1.0400297995056303e-05, |
|
"loss": 0.577, |
|
"r_loss": 0.0, |
|
"step": 3730, |
|
"steps": 3730 |
|
}, |
|
{ |
|
"epoch": 0.498, |
|
"learning_rate": 1.0372846745399615e-05, |
|
"loss": 0.563, |
|
"r_loss": 0.0, |
|
"step": 3740, |
|
"steps": 3740 |
|
}, |
|
{ |
|
"epoch": 0.499, |
|
"learning_rate": 1.0345395495742928e-05, |
|
"loss": 0.5563, |
|
"r_loss": 0.0, |
|
"step": 3750, |
|
"steps": 3750 |
|
}, |
|
{ |
|
"epoch": 0.501, |
|
"learning_rate": 1.031794424608624e-05, |
|
"loss": 0.57, |
|
"r_loss": 0.0, |
|
"step": 3760, |
|
"steps": 3760 |
|
}, |
|
{ |
|
"epoch": 0.502, |
|
"learning_rate": 1.0290492996429551e-05, |
|
"loss": 0.5668, |
|
"r_loss": 0.0, |
|
"step": 3770, |
|
"steps": 3770 |
|
}, |
|
{ |
|
"epoch": 0.503, |
|
"learning_rate": 1.0263041746772864e-05, |
|
"loss": 0.5699, |
|
"r_loss": 0.0, |
|
"step": 3780, |
|
"steps": 3780 |
|
}, |
|
{ |
|
"epoch": 0.505, |
|
"learning_rate": 1.0235590497116176e-05, |
|
"loss": 0.5575, |
|
"r_loss": 0.0, |
|
"step": 3790, |
|
"steps": 3790 |
|
}, |
|
{ |
|
"epoch": 0.506, |
|
"learning_rate": 1.0208139247459488e-05, |
|
"loss": 0.5555, |
|
"r_loss": 0.0, |
|
"step": 3800, |
|
"steps": 3800 |
|
}, |
|
{ |
|
"epoch": 0.507, |
|
"learning_rate": 1.0180687997802801e-05, |
|
"loss": 0.5718, |
|
"r_loss": 0.0, |
|
"step": 3810, |
|
"steps": 3810 |
|
}, |
|
{ |
|
"epoch": 0.509, |
|
"learning_rate": 1.0153236748146113e-05, |
|
"loss": 0.5672, |
|
"r_loss": 0.0, |
|
"step": 3820, |
|
"steps": 3820 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 1.0125785498489424e-05, |
|
"loss": 0.5606, |
|
"r_loss": 0.0, |
|
"step": 3830, |
|
"steps": 3830 |
|
}, |
|
{ |
|
"epoch": 0.511, |
|
"learning_rate": 1.009833424883274e-05, |
|
"loss": 0.5656, |
|
"r_loss": 0.0, |
|
"step": 3840, |
|
"steps": 3840 |
|
}, |
|
{ |
|
"epoch": 0.513, |
|
"learning_rate": 1.007088299917605e-05, |
|
"loss": 0.5578, |
|
"r_loss": 0.0, |
|
"step": 3850, |
|
"steps": 3850 |
|
}, |
|
{ |
|
"epoch": 0.514, |
|
"learning_rate": 1.0043431749519364e-05, |
|
"loss": 0.5626, |
|
"r_loss": 0.0, |
|
"step": 3860, |
|
"steps": 3860 |
|
}, |
|
{ |
|
"epoch": 0.515, |
|
"learning_rate": 1.0015980499862676e-05, |
|
"loss": 0.5612, |
|
"r_loss": 0.0, |
|
"step": 3870, |
|
"steps": 3870 |
|
}, |
|
{ |
|
"epoch": 0.517, |
|
"learning_rate": 9.988529250205988e-06, |
|
"loss": 0.5534, |
|
"r_loss": 0.0, |
|
"step": 3880, |
|
"steps": 3880 |
|
}, |
|
{ |
|
"epoch": 0.518, |
|
"learning_rate": 9.9610780005493e-06, |
|
"loss": 0.5467, |
|
"r_loss": 0.0, |
|
"step": 3890, |
|
"steps": 3890 |
|
}, |
|
{ |
|
"epoch": 0.519, |
|
"learning_rate": 9.933626750892613e-06, |
|
"loss": 0.5573, |
|
"r_loss": 0.0, |
|
"step": 3900, |
|
"steps": 3900 |
|
}, |
|
{ |
|
"epoch": 0.521, |
|
"learning_rate": 9.906175501235924e-06, |
|
"loss": 0.5642, |
|
"r_loss": 0.0, |
|
"step": 3910, |
|
"steps": 3910 |
|
}, |
|
{ |
|
"epoch": 0.522, |
|
"learning_rate": 9.878724251579236e-06, |
|
"loss": 0.5621, |
|
"r_loss": 0.0, |
|
"step": 3920, |
|
"steps": 3920 |
|
}, |
|
{ |
|
"epoch": 0.523, |
|
"learning_rate": 9.851273001922549e-06, |
|
"loss": 0.5609, |
|
"r_loss": 0.0, |
|
"step": 3930, |
|
"steps": 3930 |
|
}, |
|
{ |
|
"epoch": 0.525, |
|
"learning_rate": 9.82382175226586e-06, |
|
"loss": 0.5832, |
|
"r_loss": 0.0, |
|
"step": 3940, |
|
"steps": 3940 |
|
}, |
|
{ |
|
"epoch": 0.526, |
|
"learning_rate": 9.796370502609172e-06, |
|
"loss": 0.5658, |
|
"r_loss": 0.0, |
|
"step": 3950, |
|
"steps": 3950 |
|
}, |
|
{ |
|
"epoch": 0.527, |
|
"learning_rate": 9.768919252952486e-06, |
|
"loss": 0.56, |
|
"r_loss": 0.0, |
|
"step": 3960, |
|
"steps": 3960 |
|
}, |
|
{ |
|
"epoch": 0.529, |
|
"learning_rate": 9.741468003295797e-06, |
|
"loss": 0.5683, |
|
"r_loss": 0.0, |
|
"step": 3970, |
|
"steps": 3970 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.714016753639109e-06, |
|
"loss": 0.5619, |
|
"r_loss": 0.0, |
|
"step": 3980, |
|
"steps": 3980 |
|
}, |
|
{ |
|
"epoch": 0.531, |
|
"learning_rate": 9.686565503982422e-06, |
|
"loss": 0.5648, |
|
"r_loss": 0.0, |
|
"step": 3990, |
|
"steps": 3990 |
|
}, |
|
{ |
|
"epoch": 0.533, |
|
"learning_rate": 9.659114254325734e-06, |
|
"loss": 0.5659, |
|
"r_loss": 0.0, |
|
"step": 4000, |
|
"steps": 4000 |
|
}, |
|
{ |
|
"epoch": 0.534, |
|
"learning_rate": 9.631663004669047e-06, |
|
"loss": 0.5652, |
|
"r_loss": 0.0, |
|
"step": 4010, |
|
"steps": 4010 |
|
}, |
|
{ |
|
"epoch": 0.535, |
|
"learning_rate": 9.604211755012359e-06, |
|
"loss": 0.5475, |
|
"r_loss": 0.0, |
|
"step": 4020, |
|
"steps": 4020 |
|
}, |
|
{ |
|
"epoch": 0.537, |
|
"learning_rate": 9.576760505355672e-06, |
|
"loss": 0.5733, |
|
"r_loss": 0.0, |
|
"step": 4030, |
|
"steps": 4030 |
|
}, |
|
{ |
|
"epoch": 0.538, |
|
"learning_rate": 9.549309255698984e-06, |
|
"loss": 0.5537, |
|
"r_loss": 0.0, |
|
"step": 4040, |
|
"steps": 4040 |
|
}, |
|
{ |
|
"epoch": 0.539, |
|
"learning_rate": 9.521858006042296e-06, |
|
"loss": 0.5582, |
|
"r_loss": 0.0, |
|
"step": 4050, |
|
"steps": 4050 |
|
}, |
|
{ |
|
"epoch": 0.541, |
|
"learning_rate": 9.494406756385609e-06, |
|
"loss": 0.5559, |
|
"r_loss": 0.0, |
|
"step": 4060, |
|
"steps": 4060 |
|
}, |
|
{ |
|
"epoch": 0.542, |
|
"learning_rate": 9.46695550672892e-06, |
|
"loss": 0.5698, |
|
"r_loss": 0.0, |
|
"step": 4070, |
|
"steps": 4070 |
|
}, |
|
{ |
|
"epoch": 0.543, |
|
"learning_rate": 9.439504257072232e-06, |
|
"loss": 0.5478, |
|
"r_loss": 0.0, |
|
"step": 4080, |
|
"steps": 4080 |
|
}, |
|
{ |
|
"epoch": 0.545, |
|
"learning_rate": 9.412053007415545e-06, |
|
"loss": 0.5657, |
|
"r_loss": 0.0, |
|
"step": 4090, |
|
"steps": 4090 |
|
}, |
|
{ |
|
"epoch": 0.546, |
|
"learning_rate": 9.384601757758857e-06, |
|
"loss": 0.5695, |
|
"r_loss": 0.0, |
|
"step": 4100, |
|
"steps": 4100 |
|
}, |
|
{ |
|
"epoch": 0.547, |
|
"learning_rate": 9.35715050810217e-06, |
|
"loss": 0.5614, |
|
"r_loss": 0.0, |
|
"step": 4110, |
|
"steps": 4110 |
|
}, |
|
{ |
|
"epoch": 0.549, |
|
"learning_rate": 9.329699258445482e-06, |
|
"loss": 0.5641, |
|
"r_loss": 0.0, |
|
"step": 4120, |
|
"steps": 4120 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 9.302248008788794e-06, |
|
"loss": 0.5645, |
|
"r_loss": 0.0, |
|
"step": 4130, |
|
"steps": 4130 |
|
}, |
|
{ |
|
"epoch": 0.551, |
|
"learning_rate": 9.274796759132107e-06, |
|
"loss": 0.5613, |
|
"r_loss": 0.0, |
|
"step": 4140, |
|
"steps": 4140 |
|
}, |
|
{ |
|
"epoch": 0.553, |
|
"learning_rate": 9.247345509475419e-06, |
|
"loss": 0.5482, |
|
"r_loss": 0.0, |
|
"step": 4150, |
|
"steps": 4150 |
|
}, |
|
{ |
|
"epoch": 0.554, |
|
"learning_rate": 9.21989425981873e-06, |
|
"loss": 0.5579, |
|
"r_loss": 0.0, |
|
"step": 4160, |
|
"steps": 4160 |
|
}, |
|
{ |
|
"epoch": 0.555, |
|
"learning_rate": 9.192443010162044e-06, |
|
"loss": 0.571, |
|
"r_loss": 0.0, |
|
"step": 4170, |
|
"steps": 4170 |
|
}, |
|
{ |
|
"epoch": 0.557, |
|
"learning_rate": 9.164991760505357e-06, |
|
"loss": 0.5565, |
|
"r_loss": 0.0, |
|
"step": 4180, |
|
"steps": 4180 |
|
}, |
|
{ |
|
"epoch": 0.558, |
|
"learning_rate": 9.137540510848667e-06, |
|
"loss": 0.5586, |
|
"r_loss": 0.0, |
|
"step": 4190, |
|
"steps": 4190 |
|
}, |
|
{ |
|
"epoch": 0.559, |
|
"learning_rate": 9.11008926119198e-06, |
|
"loss": 0.554, |
|
"r_loss": 0.0, |
|
"step": 4200, |
|
"steps": 4200 |
|
}, |
|
{ |
|
"epoch": 0.561, |
|
"learning_rate": 9.082638011535293e-06, |
|
"loss": 0.5611, |
|
"r_loss": 0.0, |
|
"step": 4210, |
|
"steps": 4210 |
|
}, |
|
{ |
|
"epoch": 0.562, |
|
"learning_rate": 9.055186761878605e-06, |
|
"loss": 0.5577, |
|
"r_loss": 0.0, |
|
"step": 4220, |
|
"steps": 4220 |
|
}, |
|
{ |
|
"epoch": 0.563, |
|
"learning_rate": 9.027735512221917e-06, |
|
"loss": 0.563, |
|
"r_loss": 0.0, |
|
"step": 4230, |
|
"steps": 4230 |
|
}, |
|
{ |
|
"epoch": 0.565, |
|
"learning_rate": 9.00028426256523e-06, |
|
"loss": 0.5642, |
|
"r_loss": 0.0, |
|
"step": 4240, |
|
"steps": 4240 |
|
}, |
|
{ |
|
"epoch": 0.566, |
|
"learning_rate": 8.972833012908542e-06, |
|
"loss": 0.5513, |
|
"r_loss": 0.0, |
|
"step": 4250, |
|
"steps": 4250 |
|
}, |
|
{ |
|
"epoch": 0.567, |
|
"learning_rate": 8.945381763251853e-06, |
|
"loss": 0.5591, |
|
"r_loss": 0.0, |
|
"step": 4260, |
|
"steps": 4260 |
|
}, |
|
{ |
|
"epoch": 0.569, |
|
"learning_rate": 8.917930513595167e-06, |
|
"loss": 0.5575, |
|
"r_loss": 0.0, |
|
"step": 4270, |
|
"steps": 4270 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 8.890479263938478e-06, |
|
"loss": 0.5618, |
|
"r_loss": 0.0, |
|
"step": 4280, |
|
"steps": 4280 |
|
}, |
|
{ |
|
"epoch": 0.571, |
|
"learning_rate": 8.86302801428179e-06, |
|
"loss": 0.5623, |
|
"r_loss": 0.0, |
|
"step": 4290, |
|
"steps": 4290 |
|
}, |
|
{ |
|
"epoch": 0.573, |
|
"learning_rate": 8.835576764625103e-06, |
|
"loss": 0.5611, |
|
"r_loss": 0.0, |
|
"step": 4300, |
|
"steps": 4300 |
|
}, |
|
{ |
|
"epoch": 0.574, |
|
"learning_rate": 8.808125514968415e-06, |
|
"loss": 0.5552, |
|
"r_loss": 0.0, |
|
"step": 4310, |
|
"steps": 4310 |
|
}, |
|
{ |
|
"epoch": 0.575, |
|
"learning_rate": 8.780674265311727e-06, |
|
"loss": 0.556, |
|
"r_loss": 0.0, |
|
"step": 4320, |
|
"steps": 4320 |
|
}, |
|
{ |
|
"epoch": 0.577, |
|
"learning_rate": 8.75322301565504e-06, |
|
"loss": 0.5585, |
|
"r_loss": 0.0, |
|
"step": 4330, |
|
"steps": 4330 |
|
}, |
|
{ |
|
"epoch": 0.578, |
|
"learning_rate": 8.725771765998351e-06, |
|
"loss": 0.5637, |
|
"r_loss": 0.0, |
|
"step": 4340, |
|
"steps": 4340 |
|
}, |
|
{ |
|
"epoch": 0.579, |
|
"learning_rate": 8.698320516341663e-06, |
|
"loss": 0.5596, |
|
"r_loss": 0.0, |
|
"step": 4350, |
|
"steps": 4350 |
|
}, |
|
{ |
|
"epoch": 0.581, |
|
"learning_rate": 8.670869266684976e-06, |
|
"loss": 0.5531, |
|
"r_loss": 0.0, |
|
"step": 4360, |
|
"steps": 4360 |
|
}, |
|
{ |
|
"epoch": 0.582, |
|
"learning_rate": 8.64341801702829e-06, |
|
"loss": 0.5545, |
|
"r_loss": 0.0, |
|
"step": 4370, |
|
"steps": 4370 |
|
}, |
|
{ |
|
"epoch": 0.583, |
|
"learning_rate": 8.615966767371601e-06, |
|
"loss": 0.5587, |
|
"r_loss": 0.0, |
|
"step": 4380, |
|
"steps": 4380 |
|
}, |
|
{ |
|
"epoch": 0.585, |
|
"learning_rate": 8.588515517714913e-06, |
|
"loss": 0.5584, |
|
"r_loss": 0.0, |
|
"step": 4390, |
|
"steps": 4390 |
|
}, |
|
{ |
|
"epoch": 0.586, |
|
"learning_rate": 8.561064268058226e-06, |
|
"loss": 0.5572, |
|
"r_loss": 0.0, |
|
"step": 4400, |
|
"steps": 4400 |
|
}, |
|
{ |
|
"epoch": 0.587, |
|
"learning_rate": 8.533613018401538e-06, |
|
"loss": 0.5581, |
|
"r_loss": 0.0, |
|
"step": 4410, |
|
"steps": 4410 |
|
}, |
|
{ |
|
"epoch": 0.589, |
|
"learning_rate": 8.50616176874485e-06, |
|
"loss": 0.5568, |
|
"r_loss": 0.0, |
|
"step": 4420, |
|
"steps": 4420 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 8.478710519088163e-06, |
|
"loss": 0.5571, |
|
"r_loss": 0.0, |
|
"step": 4430, |
|
"steps": 4430 |
|
}, |
|
{ |
|
"epoch": 0.591, |
|
"learning_rate": 8.451259269431475e-06, |
|
"loss": 0.5591, |
|
"r_loss": 0.0, |
|
"step": 4440, |
|
"steps": 4440 |
|
}, |
|
{ |
|
"epoch": 0.593, |
|
"learning_rate": 8.423808019774786e-06, |
|
"loss": 0.5545, |
|
"r_loss": 0.0, |
|
"step": 4450, |
|
"steps": 4450 |
|
}, |
|
{ |
|
"epoch": 0.594, |
|
"learning_rate": 8.3963567701181e-06, |
|
"loss": 0.5566, |
|
"r_loss": 0.0, |
|
"step": 4460, |
|
"steps": 4460 |
|
}, |
|
{ |
|
"epoch": 0.595, |
|
"learning_rate": 8.368905520461411e-06, |
|
"loss": 0.5478, |
|
"r_loss": 0.0, |
|
"step": 4470, |
|
"steps": 4470 |
|
}, |
|
{ |
|
"epoch": 0.597, |
|
"learning_rate": 8.341454270804725e-06, |
|
"loss": 0.5597, |
|
"r_loss": 0.0, |
|
"step": 4480, |
|
"steps": 4480 |
|
}, |
|
{ |
|
"epoch": 0.598, |
|
"learning_rate": 8.314003021148036e-06, |
|
"loss": 0.5471, |
|
"r_loss": 0.0, |
|
"step": 4490, |
|
"steps": 4490 |
|
}, |
|
{ |
|
"epoch": 0.599, |
|
"learning_rate": 8.286551771491348e-06, |
|
"loss": 0.5528, |
|
"r_loss": 0.0, |
|
"step": 4500, |
|
"steps": 4500 |
|
}, |
|
{ |
|
"epoch": 0.601, |
|
"learning_rate": 8.259100521834661e-06, |
|
"loss": 0.5452, |
|
"r_loss": 0.0, |
|
"step": 4510, |
|
"steps": 4510 |
|
}, |
|
{ |
|
"epoch": 0.602, |
|
"learning_rate": 8.231649272177973e-06, |
|
"loss": 0.5621, |
|
"r_loss": 0.0, |
|
"step": 4520, |
|
"steps": 4520 |
|
}, |
|
{ |
|
"epoch": 0.603, |
|
"learning_rate": 8.204198022521284e-06, |
|
"loss": 0.5506, |
|
"r_loss": 0.0, |
|
"step": 4530, |
|
"steps": 4530 |
|
}, |
|
{ |
|
"epoch": 0.605, |
|
"learning_rate": 8.176746772864598e-06, |
|
"loss": 0.556, |
|
"r_loss": 0.0, |
|
"step": 4540, |
|
"steps": 4540 |
|
}, |
|
{ |
|
"epoch": 0.606, |
|
"learning_rate": 8.149295523207911e-06, |
|
"loss": 0.5507, |
|
"r_loss": 0.0, |
|
"step": 4550, |
|
"steps": 4550 |
|
}, |
|
{ |
|
"epoch": 0.607, |
|
"learning_rate": 8.121844273551223e-06, |
|
"loss": 0.5598, |
|
"r_loss": 0.0, |
|
"step": 4560, |
|
"steps": 4560 |
|
}, |
|
{ |
|
"epoch": 0.609, |
|
"learning_rate": 8.094393023894534e-06, |
|
"loss": 0.5581, |
|
"r_loss": 0.0, |
|
"step": 4570, |
|
"steps": 4570 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 8.066941774237848e-06, |
|
"loss": 0.5487, |
|
"r_loss": 0.0, |
|
"step": 4580, |
|
"steps": 4580 |
|
}, |
|
{ |
|
"epoch": 0.611, |
|
"learning_rate": 8.03949052458116e-06, |
|
"loss": 0.5549, |
|
"r_loss": 0.0, |
|
"step": 4590, |
|
"steps": 4590 |
|
}, |
|
{ |
|
"epoch": 0.613, |
|
"learning_rate": 8.012039274924471e-06, |
|
"loss": 0.5531, |
|
"r_loss": 0.0, |
|
"step": 4600, |
|
"steps": 4600 |
|
}, |
|
{ |
|
"epoch": 0.614, |
|
"learning_rate": 7.984588025267784e-06, |
|
"loss": 0.5548, |
|
"r_loss": 0.0, |
|
"step": 4610, |
|
"steps": 4610 |
|
}, |
|
{ |
|
"epoch": 0.615, |
|
"learning_rate": 7.957136775611096e-06, |
|
"loss": 0.5549, |
|
"r_loss": 0.0, |
|
"step": 4620, |
|
"steps": 4620 |
|
}, |
|
{ |
|
"epoch": 0.617, |
|
"learning_rate": 7.935175775885745e-06, |
|
"loss": 0.5634, |
|
"r_loss": 0.0, |
|
"step": 4630, |
|
"steps": 4630 |
|
}, |
|
{ |
|
"epoch": 0.618, |
|
"learning_rate": 7.907724526229059e-06, |
|
"loss": 0.5514, |
|
"r_loss": 0.0, |
|
"step": 4640, |
|
"steps": 4640 |
|
}, |
|
{ |
|
"epoch": 0.619, |
|
"learning_rate": 7.88027327657237e-06, |
|
"loss": 0.56, |
|
"r_loss": 0.0, |
|
"step": 4650, |
|
"steps": 4650 |
|
}, |
|
{ |
|
"epoch": 0.621, |
|
"learning_rate": 7.852822026915682e-06, |
|
"loss": 0.544, |
|
"r_loss": 0.0, |
|
"step": 4660, |
|
"steps": 4660 |
|
}, |
|
{ |
|
"epoch": 0.622, |
|
"learning_rate": 7.825370777258995e-06, |
|
"loss": 0.5523, |
|
"r_loss": 0.0, |
|
"step": 4670, |
|
"steps": 4670 |
|
}, |
|
{ |
|
"epoch": 0.623, |
|
"learning_rate": 7.797919527602307e-06, |
|
"loss": 0.5553, |
|
"r_loss": 0.0, |
|
"step": 4680, |
|
"steps": 4680 |
|
}, |
|
{ |
|
"epoch": 0.625, |
|
"learning_rate": 7.770468277945619e-06, |
|
"loss": 0.5566, |
|
"r_loss": 0.0, |
|
"step": 4690, |
|
"steps": 4690 |
|
}, |
|
{ |
|
"epoch": 0.626, |
|
"learning_rate": 7.743017028288932e-06, |
|
"loss": 0.5543, |
|
"r_loss": 0.0, |
|
"step": 4700, |
|
"steps": 4700 |
|
}, |
|
{ |
|
"epoch": 0.627, |
|
"learning_rate": 7.715565778632244e-06, |
|
"loss": 0.555, |
|
"r_loss": 0.0, |
|
"step": 4710, |
|
"steps": 4710 |
|
}, |
|
{ |
|
"epoch": 0.629, |
|
"learning_rate": 7.688114528975555e-06, |
|
"loss": 0.5407, |
|
"r_loss": 0.0, |
|
"step": 4720, |
|
"steps": 4720 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 7.660663279318869e-06, |
|
"loss": 0.5538, |
|
"r_loss": 0.0, |
|
"step": 4730, |
|
"steps": 4730 |
|
}, |
|
{ |
|
"epoch": 0.631, |
|
"learning_rate": 7.633212029662182e-06, |
|
"loss": 0.559, |
|
"r_loss": 0.0, |
|
"step": 4740, |
|
"steps": 4740 |
|
}, |
|
{ |
|
"epoch": 0.633, |
|
"learning_rate": 7.605760780005493e-06, |
|
"loss": 0.5467, |
|
"r_loss": 0.0, |
|
"step": 4750, |
|
"steps": 4750 |
|
}, |
|
{ |
|
"epoch": 0.634, |
|
"learning_rate": 7.578309530348805e-06, |
|
"loss": 0.5422, |
|
"r_loss": 0.0, |
|
"step": 4760, |
|
"steps": 4760 |
|
}, |
|
{ |
|
"epoch": 0.635, |
|
"learning_rate": 7.550858280692118e-06, |
|
"loss": 0.5469, |
|
"r_loss": 0.0, |
|
"step": 4770, |
|
"steps": 4770 |
|
}, |
|
{ |
|
"epoch": 0.637, |
|
"learning_rate": 7.52340703103543e-06, |
|
"loss": 0.5371, |
|
"r_loss": 0.0, |
|
"step": 4780, |
|
"steps": 4780 |
|
}, |
|
{ |
|
"epoch": 0.638, |
|
"learning_rate": 7.495955781378742e-06, |
|
"loss": 0.565, |
|
"r_loss": 0.0, |
|
"step": 4790, |
|
"steps": 4790 |
|
}, |
|
{ |
|
"epoch": 0.639, |
|
"learning_rate": 7.468504531722054e-06, |
|
"loss": 0.5323, |
|
"r_loss": 0.0, |
|
"step": 4800, |
|
"steps": 4800 |
|
}, |
|
{ |
|
"epoch": 0.641, |
|
"learning_rate": 7.441053282065367e-06, |
|
"loss": 0.5507, |
|
"r_loss": 0.0, |
|
"step": 4810, |
|
"steps": 4810 |
|
}, |
|
{ |
|
"epoch": 0.642, |
|
"learning_rate": 7.413602032408678e-06, |
|
"loss": 0.5496, |
|
"r_loss": 0.0, |
|
"step": 4820, |
|
"steps": 4820 |
|
}, |
|
{ |
|
"epoch": 0.643, |
|
"learning_rate": 7.386150782751991e-06, |
|
"loss": 0.5447, |
|
"r_loss": 0.0, |
|
"step": 4830, |
|
"steps": 4830 |
|
}, |
|
{ |
|
"epoch": 0.645, |
|
"learning_rate": 7.358699533095303e-06, |
|
"loss": 0.539, |
|
"r_loss": 0.0, |
|
"step": 4840, |
|
"steps": 4840 |
|
}, |
|
{ |
|
"epoch": 0.646, |
|
"learning_rate": 7.331248283438615e-06, |
|
"loss": 0.5462, |
|
"r_loss": 0.0, |
|
"step": 4850, |
|
"steps": 4850 |
|
}, |
|
{ |
|
"epoch": 0.647, |
|
"learning_rate": 7.3037970337819274e-06, |
|
"loss": 0.5478, |
|
"r_loss": 0.0, |
|
"step": 4860, |
|
"steps": 4860 |
|
}, |
|
{ |
|
"epoch": 0.649, |
|
"learning_rate": 7.276345784125241e-06, |
|
"loss": 0.5394, |
|
"r_loss": 0.0, |
|
"step": 4870, |
|
"steps": 4870 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 7.248894534468553e-06, |
|
"loss": 0.5507, |
|
"r_loss": 0.0, |
|
"step": 4880, |
|
"steps": 4880 |
|
}, |
|
{ |
|
"epoch": 0.651, |
|
"learning_rate": 7.221443284811865e-06, |
|
"loss": 0.5422, |
|
"r_loss": 0.0, |
|
"step": 4890, |
|
"steps": 4890 |
|
}, |
|
{ |
|
"epoch": 0.653, |
|
"learning_rate": 7.193992035155177e-06, |
|
"loss": 0.555, |
|
"r_loss": 0.0, |
|
"step": 4900, |
|
"steps": 4900 |
|
}, |
|
{ |
|
"epoch": 0.654, |
|
"learning_rate": 7.16654078549849e-06, |
|
"loss": 0.5385, |
|
"r_loss": 0.0, |
|
"step": 4910, |
|
"steps": 4910 |
|
}, |
|
{ |
|
"epoch": 0.655, |
|
"learning_rate": 7.1390895358418015e-06, |
|
"loss": 0.5517, |
|
"r_loss": 0.0, |
|
"step": 4920, |
|
"steps": 4920 |
|
}, |
|
{ |
|
"epoch": 0.657, |
|
"learning_rate": 7.111638286185114e-06, |
|
"loss": 0.5419, |
|
"r_loss": 0.0, |
|
"step": 4930, |
|
"steps": 4930 |
|
}, |
|
{ |
|
"epoch": 0.658, |
|
"learning_rate": 7.0841870365284264e-06, |
|
"loss": 0.5396, |
|
"r_loss": 0.0, |
|
"step": 4940, |
|
"steps": 4940 |
|
}, |
|
{ |
|
"epoch": 0.659, |
|
"learning_rate": 7.056735786871739e-06, |
|
"loss": 0.5458, |
|
"r_loss": 0.0, |
|
"step": 4950, |
|
"steps": 4950 |
|
}, |
|
{ |
|
"epoch": 0.661, |
|
"learning_rate": 7.0292845372150506e-06, |
|
"loss": 0.5462, |
|
"r_loss": 0.0, |
|
"step": 4960, |
|
"steps": 4960 |
|
}, |
|
{ |
|
"epoch": 0.662, |
|
"learning_rate": 7.001833287558363e-06, |
|
"loss": 0.5538, |
|
"r_loss": 0.0, |
|
"step": 4970, |
|
"steps": 4970 |
|
}, |
|
{ |
|
"epoch": 0.663, |
|
"learning_rate": 6.9743820379016755e-06, |
|
"loss": 0.5599, |
|
"r_loss": 0.0, |
|
"step": 4980, |
|
"steps": 4980 |
|
}, |
|
{ |
|
"epoch": 0.665, |
|
"learning_rate": 6.946930788244987e-06, |
|
"loss": 0.5595, |
|
"r_loss": 0.0, |
|
"step": 4990, |
|
"steps": 4990 |
|
}, |
|
{ |
|
"epoch": 0.666, |
|
"learning_rate": 6.9194795385883e-06, |
|
"loss": 0.5373, |
|
"r_loss": 0.0, |
|
"step": 5000, |
|
"steps": 5000 |
|
}, |
|
{ |
|
"epoch": 0.667, |
|
"learning_rate": 6.892028288931612e-06, |
|
"loss": 0.5513, |
|
"r_loss": 0.0, |
|
"step": 5010, |
|
"steps": 5010 |
|
}, |
|
{ |
|
"epoch": 0.669, |
|
"learning_rate": 6.864577039274924e-06, |
|
"loss": 0.5394, |
|
"r_loss": 0.0, |
|
"step": 5020, |
|
"steps": 5020 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 6.837125789618236e-06, |
|
"loss": 0.5605, |
|
"r_loss": 0.0, |
|
"step": 5030, |
|
"steps": 5030 |
|
}, |
|
{ |
|
"epoch": 0.671, |
|
"learning_rate": 6.809674539961549e-06, |
|
"loss": 0.5457, |
|
"r_loss": 0.0, |
|
"step": 5040, |
|
"steps": 5040 |
|
}, |
|
{ |
|
"epoch": 0.673, |
|
"learning_rate": 6.782223290304862e-06, |
|
"loss": 0.5463, |
|
"r_loss": 0.0, |
|
"step": 5050, |
|
"steps": 5050 |
|
}, |
|
{ |
|
"epoch": 0.674, |
|
"learning_rate": 6.754772040648173e-06, |
|
"loss": 0.5446, |
|
"r_loss": 0.0, |
|
"step": 5060, |
|
"steps": 5060 |
|
}, |
|
{ |
|
"epoch": 0.675, |
|
"learning_rate": 6.727320790991486e-06, |
|
"loss": 0.5386, |
|
"r_loss": 0.0, |
|
"step": 5070, |
|
"steps": 5070 |
|
}, |
|
{ |
|
"epoch": 0.677, |
|
"learning_rate": 6.699869541334799e-06, |
|
"loss": 0.5416, |
|
"r_loss": 0.0, |
|
"step": 5080, |
|
"steps": 5080 |
|
}, |
|
{ |
|
"epoch": 0.678, |
|
"learning_rate": 6.67241829167811e-06, |
|
"loss": 0.5375, |
|
"r_loss": 0.0, |
|
"step": 5090, |
|
"steps": 5090 |
|
}, |
|
{ |
|
"epoch": 0.679, |
|
"learning_rate": 6.644967042021423e-06, |
|
"loss": 0.5425, |
|
"r_loss": 0.0, |
|
"step": 5100, |
|
"steps": 5100 |
|
}, |
|
{ |
|
"epoch": 0.681, |
|
"learning_rate": 6.617515792364735e-06, |
|
"loss": 0.5397, |
|
"r_loss": 0.0, |
|
"step": 5110, |
|
"steps": 5110 |
|
}, |
|
{ |
|
"epoch": 0.682, |
|
"learning_rate": 6.590064542708047e-06, |
|
"loss": 0.5337, |
|
"r_loss": 0.0, |
|
"step": 5120, |
|
"steps": 5120 |
|
}, |
|
{ |
|
"epoch": 0.683, |
|
"learning_rate": 6.562613293051359e-06, |
|
"loss": 0.5402, |
|
"r_loss": 0.0, |
|
"step": 5130, |
|
"steps": 5130 |
|
}, |
|
{ |
|
"epoch": 0.685, |
|
"learning_rate": 6.535162043394672e-06, |
|
"loss": 0.5358, |
|
"r_loss": 0.0, |
|
"step": 5140, |
|
"steps": 5140 |
|
}, |
|
{ |
|
"epoch": 0.686, |
|
"learning_rate": 6.507710793737984e-06, |
|
"loss": 0.5458, |
|
"r_loss": 0.0, |
|
"step": 5150, |
|
"steps": 5150 |
|
}, |
|
{ |
|
"epoch": 0.687, |
|
"learning_rate": 6.480259544081296e-06, |
|
"loss": 0.5467, |
|
"r_loss": 0.0, |
|
"step": 5160, |
|
"steps": 5160 |
|
}, |
|
{ |
|
"epoch": 0.689, |
|
"learning_rate": 6.452808294424608e-06, |
|
"loss": 0.5457, |
|
"r_loss": 0.0, |
|
"step": 5170, |
|
"steps": 5170 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 6.425357044767921e-06, |
|
"loss": 0.5511, |
|
"r_loss": 0.0, |
|
"step": 5180, |
|
"steps": 5180 |
|
}, |
|
{ |
|
"epoch": 0.691, |
|
"learning_rate": 6.3979057951112325e-06, |
|
"loss": 0.5491, |
|
"r_loss": 0.0, |
|
"step": 5190, |
|
"steps": 5190 |
|
}, |
|
{ |
|
"epoch": 0.693, |
|
"learning_rate": 6.370454545454545e-06, |
|
"loss": 0.5306, |
|
"r_loss": 0.0, |
|
"step": 5200, |
|
"steps": 5200 |
|
}, |
|
{ |
|
"epoch": 0.694, |
|
"learning_rate": 6.3430032957978575e-06, |
|
"loss": 0.5473, |
|
"r_loss": 0.0, |
|
"step": 5210, |
|
"steps": 5210 |
|
}, |
|
{ |
|
"epoch": 0.695, |
|
"learning_rate": 6.315552046141169e-06, |
|
"loss": 0.5433, |
|
"r_loss": 0.0, |
|
"step": 5220, |
|
"steps": 5220 |
|
}, |
|
{ |
|
"epoch": 0.697, |
|
"learning_rate": 6.288100796484482e-06, |
|
"loss": 0.5413, |
|
"r_loss": 0.0, |
|
"step": 5230, |
|
"steps": 5230 |
|
}, |
|
{ |
|
"epoch": 0.698, |
|
"learning_rate": 6.260649546827795e-06, |
|
"loss": 0.5395, |
|
"r_loss": 0.0, |
|
"step": 5240, |
|
"steps": 5240 |
|
}, |
|
{ |
|
"epoch": 0.699, |
|
"learning_rate": 6.233198297171107e-06, |
|
"loss": 0.54, |
|
"r_loss": 0.0, |
|
"step": 5250, |
|
"steps": 5250 |
|
}, |
|
{ |
|
"epoch": 0.701, |
|
"learning_rate": 6.205747047514419e-06, |
|
"loss": 0.5302, |
|
"r_loss": 0.0, |
|
"step": 5260, |
|
"steps": 5260 |
|
}, |
|
{ |
|
"epoch": 0.702, |
|
"learning_rate": 6.1782957978577315e-06, |
|
"loss": 0.5379, |
|
"r_loss": 0.0, |
|
"step": 5270, |
|
"steps": 5270 |
|
}, |
|
{ |
|
"epoch": 0.703, |
|
"learning_rate": 6.150844548201044e-06, |
|
"loss": 0.5371, |
|
"r_loss": 0.0, |
|
"step": 5280, |
|
"steps": 5280 |
|
}, |
|
{ |
|
"epoch": 0.705, |
|
"learning_rate": 6.123393298544356e-06, |
|
"loss": 0.5464, |
|
"r_loss": 0.0, |
|
"step": 5290, |
|
"steps": 5290 |
|
}, |
|
{ |
|
"epoch": 0.706, |
|
"learning_rate": 6.095942048887668e-06, |
|
"loss": 0.5412, |
|
"r_loss": 0.0, |
|
"step": 5300, |
|
"steps": 5300 |
|
}, |
|
{ |
|
"epoch": 0.707, |
|
"learning_rate": 6.0684907992309806e-06, |
|
"loss": 0.5414, |
|
"r_loss": 0.0, |
|
"step": 5310, |
|
"steps": 5310 |
|
}, |
|
{ |
|
"epoch": 0.709, |
|
"learning_rate": 6.041039549574293e-06, |
|
"loss": 0.5508, |
|
"r_loss": 0.0, |
|
"step": 5320, |
|
"steps": 5320 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 6.013588299917605e-06, |
|
"loss": 0.5484, |
|
"r_loss": 0.0, |
|
"step": 5330, |
|
"steps": 5330 |
|
}, |
|
{ |
|
"epoch": 0.711, |
|
"learning_rate": 5.986137050260917e-06, |
|
"loss": 0.548, |
|
"r_loss": 0.0, |
|
"step": 5340, |
|
"steps": 5340 |
|
}, |
|
{ |
|
"epoch": 0.713, |
|
"learning_rate": 5.95868580060423e-06, |
|
"loss": 0.5386, |
|
"r_loss": 0.0, |
|
"step": 5350, |
|
"steps": 5350 |
|
}, |
|
{ |
|
"epoch": 0.714, |
|
"learning_rate": 5.931234550947541e-06, |
|
"loss": 0.534, |
|
"r_loss": 0.0, |
|
"step": 5360, |
|
"steps": 5360 |
|
}, |
|
{ |
|
"epoch": 0.715, |
|
"learning_rate": 5.903783301290854e-06, |
|
"loss": 0.5256, |
|
"r_loss": 0.0, |
|
"step": 5370, |
|
"steps": 5370 |
|
}, |
|
{ |
|
"epoch": 0.717, |
|
"learning_rate": 5.876332051634166e-06, |
|
"loss": 0.5396, |
|
"r_loss": 0.0, |
|
"step": 5380, |
|
"steps": 5380 |
|
}, |
|
{ |
|
"epoch": 0.718, |
|
"learning_rate": 5.848880801977478e-06, |
|
"loss": 0.5547, |
|
"r_loss": 0.0, |
|
"step": 5390, |
|
"steps": 5390 |
|
}, |
|
{ |
|
"epoch": 0.719, |
|
"learning_rate": 5.82142955232079e-06, |
|
"loss": 0.5372, |
|
"r_loss": 0.0, |
|
"step": 5400, |
|
"steps": 5400 |
|
}, |
|
{ |
|
"epoch": 0.721, |
|
"learning_rate": 5.793978302664104e-06, |
|
"loss": 0.5345, |
|
"r_loss": 0.0, |
|
"step": 5410, |
|
"steps": 5410 |
|
}, |
|
{ |
|
"epoch": 0.722, |
|
"learning_rate": 5.766527053007416e-06, |
|
"loss": 0.5487, |
|
"r_loss": 0.0, |
|
"step": 5420, |
|
"steps": 5420 |
|
}, |
|
{ |
|
"epoch": 0.723, |
|
"learning_rate": 5.739075803350728e-06, |
|
"loss": 0.5357, |
|
"r_loss": 0.0, |
|
"step": 5430, |
|
"steps": 5430 |
|
}, |
|
{ |
|
"epoch": 0.725, |
|
"learning_rate": 5.71162455369404e-06, |
|
"loss": 0.5388, |
|
"r_loss": 0.0, |
|
"step": 5440, |
|
"steps": 5440 |
|
}, |
|
{ |
|
"epoch": 0.726, |
|
"learning_rate": 5.684173304037353e-06, |
|
"loss": 0.5274, |
|
"r_loss": 0.0, |
|
"step": 5450, |
|
"steps": 5450 |
|
}, |
|
{ |
|
"epoch": 0.727, |
|
"learning_rate": 5.656722054380664e-06, |
|
"loss": 0.5389, |
|
"r_loss": 0.0, |
|
"step": 5460, |
|
"steps": 5460 |
|
}, |
|
{ |
|
"epoch": 0.729, |
|
"learning_rate": 5.629270804723977e-06, |
|
"loss": 0.5447, |
|
"r_loss": 0.0, |
|
"step": 5470, |
|
"steps": 5470 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 5.601819555067289e-06, |
|
"loss": 0.5535, |
|
"r_loss": 0.0, |
|
"step": 5480, |
|
"steps": 5480 |
|
}, |
|
{ |
|
"epoch": 0.731, |
|
"learning_rate": 5.574368305410601e-06, |
|
"loss": 0.5339, |
|
"r_loss": 0.0, |
|
"step": 5490, |
|
"steps": 5490 |
|
}, |
|
{ |
|
"epoch": 0.733, |
|
"learning_rate": 5.5469170557539135e-06, |
|
"loss": 0.545, |
|
"r_loss": 0.0, |
|
"step": 5500, |
|
"steps": 5500 |
|
}, |
|
{ |
|
"epoch": 0.734, |
|
"learning_rate": 5.519465806097226e-06, |
|
"loss": 0.546, |
|
"r_loss": 0.0, |
|
"step": 5510, |
|
"steps": 5510 |
|
}, |
|
{ |
|
"epoch": 0.735, |
|
"learning_rate": 5.492014556440538e-06, |
|
"loss": 0.5385, |
|
"r_loss": 0.0, |
|
"step": 5520, |
|
"steps": 5520 |
|
}, |
|
{ |
|
"epoch": 0.737, |
|
"learning_rate": 5.46456330678385e-06, |
|
"loss": 0.5448, |
|
"r_loss": 0.0, |
|
"step": 5530, |
|
"steps": 5530 |
|
}, |
|
{ |
|
"epoch": 0.738, |
|
"learning_rate": 5.4371120571271625e-06, |
|
"loss": 0.5352, |
|
"r_loss": 0.0, |
|
"step": 5540, |
|
"steps": 5540 |
|
}, |
|
{ |
|
"epoch": 0.739, |
|
"learning_rate": 5.409660807470475e-06, |
|
"loss": 0.5359, |
|
"r_loss": 0.0, |
|
"step": 5550, |
|
"steps": 5550 |
|
}, |
|
{ |
|
"epoch": 0.741, |
|
"learning_rate": 5.382209557813787e-06, |
|
"loss": 0.5318, |
|
"r_loss": 0.0, |
|
"step": 5560, |
|
"steps": 5560 |
|
}, |
|
{ |
|
"epoch": 0.742, |
|
"learning_rate": 5.354758308157099e-06, |
|
"loss": 0.5354, |
|
"r_loss": 0.0, |
|
"step": 5570, |
|
"steps": 5570 |
|
}, |
|
{ |
|
"epoch": 0.743, |
|
"learning_rate": 5.3273070585004124e-06, |
|
"loss": 0.5344, |
|
"r_loss": 0.0, |
|
"step": 5580, |
|
"steps": 5580 |
|
}, |
|
{ |
|
"epoch": 0.745, |
|
"learning_rate": 5.299855808843725e-06, |
|
"loss": 0.5399, |
|
"r_loss": 0.0, |
|
"step": 5590, |
|
"steps": 5590 |
|
}, |
|
{ |
|
"epoch": 0.746, |
|
"learning_rate": 5.2724045591870366e-06, |
|
"loss": 0.5471, |
|
"r_loss": 0.0, |
|
"step": 5600, |
|
"steps": 5600 |
|
}, |
|
{ |
|
"epoch": 0.747, |
|
"learning_rate": 5.244953309530349e-06, |
|
"loss": 0.5273, |
|
"r_loss": 0.0, |
|
"step": 5610, |
|
"steps": 5610 |
|
}, |
|
{ |
|
"epoch": 0.749, |
|
"learning_rate": 5.2175020598736615e-06, |
|
"loss": 0.5429, |
|
"r_loss": 0.0, |
|
"step": 5620, |
|
"steps": 5620 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 5.19554106014831e-06, |
|
"loss": 0.5176, |
|
"r_loss": 0.0, |
|
"step": 5630, |
|
"steps": 5630 |
|
}, |
|
{ |
|
"epoch": 0.751, |
|
"learning_rate": 5.168089810491624e-06, |
|
"loss": 0.5268, |
|
"r_loss": 0.0, |
|
"step": 5640, |
|
"steps": 5640 |
|
}, |
|
{ |
|
"epoch": 0.753, |
|
"learning_rate": 5.140638560834936e-06, |
|
"loss": 0.5311, |
|
"r_loss": 0.0, |
|
"step": 5650, |
|
"steps": 5650 |
|
}, |
|
{ |
|
"epoch": 0.754, |
|
"learning_rate": 5.113187311178247e-06, |
|
"loss": 0.5309, |
|
"r_loss": 0.0, |
|
"step": 5660, |
|
"steps": 5660 |
|
}, |
|
{ |
|
"epoch": 0.755, |
|
"learning_rate": 5.08573606152156e-06, |
|
"loss": 0.5409, |
|
"r_loss": 0.0, |
|
"step": 5670, |
|
"steps": 5670 |
|
}, |
|
{ |
|
"epoch": 0.757, |
|
"learning_rate": 5.058284811864873e-06, |
|
"loss": 0.5318, |
|
"r_loss": 0.0, |
|
"step": 5680, |
|
"steps": 5680 |
|
}, |
|
{ |
|
"epoch": 0.758, |
|
"learning_rate": 5.030833562208184e-06, |
|
"loss": 0.545, |
|
"r_loss": 0.0, |
|
"step": 5690, |
|
"steps": 5690 |
|
}, |
|
{ |
|
"epoch": 0.759, |
|
"learning_rate": 5.003382312551497e-06, |
|
"loss": 0.5413, |
|
"r_loss": 0.0, |
|
"step": 5700, |
|
"steps": 5700 |
|
}, |
|
{ |
|
"epoch": 0.761, |
|
"learning_rate": 4.975931062894809e-06, |
|
"loss": 0.5307, |
|
"r_loss": 0.0, |
|
"step": 5710, |
|
"steps": 5710 |
|
}, |
|
{ |
|
"epoch": 0.762, |
|
"learning_rate": 4.948479813238121e-06, |
|
"loss": 0.5491, |
|
"r_loss": 0.0, |
|
"step": 5720, |
|
"steps": 5720 |
|
}, |
|
{ |
|
"epoch": 0.763, |
|
"learning_rate": 4.921028563581433e-06, |
|
"loss": 0.5411, |
|
"r_loss": 0.0, |
|
"step": 5730, |
|
"steps": 5730 |
|
}, |
|
{ |
|
"epoch": 0.765, |
|
"learning_rate": 4.893577313924746e-06, |
|
"loss": 0.5456, |
|
"r_loss": 0.0, |
|
"step": 5740, |
|
"steps": 5740 |
|
}, |
|
{ |
|
"epoch": 0.766, |
|
"learning_rate": 4.866126064268058e-06, |
|
"loss": 0.5505, |
|
"r_loss": 0.0, |
|
"step": 5750, |
|
"steps": 5750 |
|
}, |
|
{ |
|
"epoch": 0.767, |
|
"learning_rate": 4.838674814611371e-06, |
|
"loss": 0.5441, |
|
"r_loss": 0.0, |
|
"step": 5760, |
|
"steps": 5760 |
|
}, |
|
{ |
|
"epoch": 0.769, |
|
"learning_rate": 4.8112235649546825e-06, |
|
"loss": 0.5405, |
|
"r_loss": 0.0, |
|
"step": 5770, |
|
"steps": 5770 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 4.786517440263664e-06, |
|
"loss": 0.5445, |
|
"r_loss": 0.0, |
|
"step": 5780, |
|
"steps": 5780 |
|
}, |
|
{ |
|
"epoch": 0.771, |
|
"learning_rate": 4.7590661906069765e-06, |
|
"loss": 0.5338, |
|
"r_loss": 0.0, |
|
"step": 5790, |
|
"steps": 5790 |
|
}, |
|
{ |
|
"epoch": 0.773, |
|
"learning_rate": 4.731614940950288e-06, |
|
"loss": 0.5334, |
|
"r_loss": 0.0, |
|
"step": 5800, |
|
"steps": 5800 |
|
}, |
|
{ |
|
"epoch": 0.774, |
|
"learning_rate": 4.7041636912936006e-06, |
|
"loss": 0.5424, |
|
"r_loss": 0.0, |
|
"step": 5810, |
|
"steps": 5810 |
|
}, |
|
{ |
|
"epoch": 0.775, |
|
"learning_rate": 4.676712441636913e-06, |
|
"loss": 0.5295, |
|
"r_loss": 0.0, |
|
"step": 5820, |
|
"steps": 5820 |
|
}, |
|
{ |
|
"epoch": 0.777, |
|
"learning_rate": 4.649261191980225e-06, |
|
"loss": 0.5504, |
|
"r_loss": 0.0, |
|
"step": 5830, |
|
"steps": 5830 |
|
}, |
|
{ |
|
"epoch": 0.778, |
|
"learning_rate": 4.621809942323537e-06, |
|
"loss": 0.5261, |
|
"r_loss": 0.0, |
|
"step": 5840, |
|
"steps": 5840 |
|
}, |
|
{ |
|
"epoch": 0.779, |
|
"learning_rate": 4.59435869266685e-06, |
|
"loss": 0.5414, |
|
"r_loss": 0.0, |
|
"step": 5850, |
|
"steps": 5850 |
|
}, |
|
{ |
|
"epoch": 0.781, |
|
"learning_rate": 4.566907443010162e-06, |
|
"loss": 0.5431, |
|
"r_loss": 0.0, |
|
"step": 5860, |
|
"steps": 5860 |
|
}, |
|
{ |
|
"epoch": 0.782, |
|
"learning_rate": 4.539456193353475e-06, |
|
"loss": 0.5411, |
|
"r_loss": 0.0, |
|
"step": 5870, |
|
"steps": 5870 |
|
}, |
|
{ |
|
"epoch": 0.783, |
|
"learning_rate": 4.512004943696786e-06, |
|
"loss": 0.5377, |
|
"r_loss": 0.0, |
|
"step": 5880, |
|
"steps": 5880 |
|
}, |
|
{ |
|
"epoch": 0.784, |
|
"learning_rate": 4.484553694040099e-06, |
|
"loss": 0.525, |
|
"r_loss": 0.0, |
|
"step": 5890, |
|
"steps": 5890 |
|
}, |
|
{ |
|
"epoch": 0.786, |
|
"learning_rate": 4.457102444383411e-06, |
|
"loss": 0.5329, |
|
"r_loss": 0.0, |
|
"step": 5900, |
|
"steps": 5900 |
|
}, |
|
{ |
|
"epoch": 0.787, |
|
"learning_rate": 4.429651194726724e-06, |
|
"loss": 0.5438, |
|
"r_loss": 0.0, |
|
"step": 5910, |
|
"steps": 5910 |
|
}, |
|
{ |
|
"epoch": 0.788, |
|
"learning_rate": 4.402199945070035e-06, |
|
"loss": 0.5307, |
|
"r_loss": 0.0, |
|
"step": 5920, |
|
"steps": 5920 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 4.374748695413348e-06, |
|
"loss": 0.5362, |
|
"r_loss": 0.0, |
|
"step": 5930, |
|
"steps": 5930 |
|
}, |
|
{ |
|
"epoch": 0.791, |
|
"learning_rate": 4.34729744575666e-06, |
|
"loss": 0.5474, |
|
"r_loss": 0.0, |
|
"step": 5940, |
|
"steps": 5940 |
|
}, |
|
{ |
|
"epoch": 0.792, |
|
"learning_rate": 4.319846196099972e-06, |
|
"loss": 0.5352, |
|
"r_loss": 0.0, |
|
"step": 5950, |
|
"steps": 5950 |
|
}, |
|
{ |
|
"epoch": 0.794, |
|
"learning_rate": 4.292394946443285e-06, |
|
"loss": 0.5288, |
|
"r_loss": 0.0, |
|
"step": 5960, |
|
"steps": 5960 |
|
}, |
|
{ |
|
"epoch": 0.795, |
|
"learning_rate": 4.264943696786597e-06, |
|
"loss": 0.5431, |
|
"r_loss": 0.0, |
|
"step": 5970, |
|
"steps": 5970 |
|
}, |
|
{ |
|
"epoch": 0.796, |
|
"learning_rate": 4.237492447129909e-06, |
|
"loss": 0.5349, |
|
"r_loss": 0.0, |
|
"step": 5980, |
|
"steps": 5980 |
|
}, |
|
{ |
|
"epoch": 0.798, |
|
"learning_rate": 4.210041197473222e-06, |
|
"loss": 0.5295, |
|
"r_loss": 0.0, |
|
"step": 5990, |
|
"steps": 5990 |
|
}, |
|
{ |
|
"epoch": 0.799, |
|
"learning_rate": 4.1825899478165335e-06, |
|
"loss": 0.5365, |
|
"r_loss": 0.0, |
|
"step": 6000, |
|
"steps": 6000 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 4.155138698159846e-06, |
|
"loss": 0.5356, |
|
"r_loss": 0.0, |
|
"step": 6010, |
|
"steps": 6010 |
|
}, |
|
{ |
|
"epoch": 0.802, |
|
"learning_rate": 4.127687448503158e-06, |
|
"loss": 0.5356, |
|
"r_loss": 0.0, |
|
"step": 6020, |
|
"steps": 6020 |
|
}, |
|
{ |
|
"epoch": 0.803, |
|
"learning_rate": 4.10023619884647e-06, |
|
"loss": 0.548, |
|
"r_loss": 0.0, |
|
"step": 6030, |
|
"steps": 6030 |
|
}, |
|
{ |
|
"epoch": 0.804, |
|
"learning_rate": 4.072784949189783e-06, |
|
"loss": 0.5384, |
|
"r_loss": 0.0, |
|
"step": 6040, |
|
"steps": 6040 |
|
}, |
|
{ |
|
"epoch": 0.806, |
|
"learning_rate": 4.045333699533095e-06, |
|
"loss": 0.5337, |
|
"r_loss": 0.0, |
|
"step": 6050, |
|
"steps": 6050 |
|
}, |
|
{ |
|
"epoch": 0.807, |
|
"learning_rate": 4.0178824498764075e-06, |
|
"loss": 0.5406, |
|
"r_loss": 0.0, |
|
"step": 6060, |
|
"steps": 6060 |
|
}, |
|
{ |
|
"epoch": 0.808, |
|
"learning_rate": 3.99043120021972e-06, |
|
"loss": 0.5345, |
|
"r_loss": 0.0, |
|
"step": 6070, |
|
"steps": 6070 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 3.9629799505630324e-06, |
|
"loss": 0.5253, |
|
"r_loss": 0.0, |
|
"step": 6080, |
|
"steps": 6080 |
|
}, |
|
{ |
|
"epoch": 0.811, |
|
"learning_rate": 3.935528700906344e-06, |
|
"loss": 0.531, |
|
"r_loss": 0.0, |
|
"step": 6090, |
|
"steps": 6090 |
|
}, |
|
{ |
|
"epoch": 0.812, |
|
"learning_rate": 3.9080774512496566e-06, |
|
"loss": 0.5278, |
|
"r_loss": 0.0, |
|
"step": 6100, |
|
"steps": 6100 |
|
}, |
|
{ |
|
"epoch": 0.814, |
|
"learning_rate": 3.880626201592969e-06, |
|
"loss": 0.5297, |
|
"r_loss": 0.0, |
|
"step": 6110, |
|
"steps": 6110 |
|
}, |
|
{ |
|
"epoch": 0.815, |
|
"learning_rate": 3.853174951936281e-06, |
|
"loss": 0.5255, |
|
"r_loss": 0.0, |
|
"step": 6120, |
|
"steps": 6120 |
|
}, |
|
{ |
|
"epoch": 0.816, |
|
"learning_rate": 3.825723702279594e-06, |
|
"loss": 0.5465, |
|
"r_loss": 0.0, |
|
"step": 6130, |
|
"steps": 6130 |
|
}, |
|
{ |
|
"epoch": 0.818, |
|
"learning_rate": 3.7982724526229056e-06, |
|
"loss": 0.5374, |
|
"r_loss": 0.0, |
|
"step": 6140, |
|
"steps": 6140 |
|
}, |
|
{ |
|
"epoch": 0.819, |
|
"learning_rate": 3.7708212029662177e-06, |
|
"loss": 0.5264, |
|
"r_loss": 0.0, |
|
"step": 6150, |
|
"steps": 6150 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 3.7433699533095306e-06, |
|
"loss": 0.5273, |
|
"r_loss": 0.0, |
|
"step": 6160, |
|
"steps": 6160 |
|
}, |
|
{ |
|
"epoch": 0.822, |
|
"learning_rate": 3.7159187036528426e-06, |
|
"loss": 0.512, |
|
"r_loss": 0.0, |
|
"step": 6170, |
|
"steps": 6170 |
|
}, |
|
{ |
|
"epoch": 0.823, |
|
"learning_rate": 3.688467453996155e-06, |
|
"loss": 0.5244, |
|
"r_loss": 0.0, |
|
"step": 6180, |
|
"steps": 6180 |
|
}, |
|
{ |
|
"epoch": 0.824, |
|
"learning_rate": 3.661016204339467e-06, |
|
"loss": 0.5224, |
|
"r_loss": 0.0, |
|
"step": 6190, |
|
"steps": 6190 |
|
}, |
|
{ |
|
"epoch": 0.826, |
|
"learning_rate": 3.6335649546827792e-06, |
|
"loss": 0.5312, |
|
"r_loss": 0.0, |
|
"step": 6200, |
|
"steps": 6200 |
|
}, |
|
{ |
|
"epoch": 0.827, |
|
"learning_rate": 3.6061137050260917e-06, |
|
"loss": 0.5334, |
|
"r_loss": 0.0, |
|
"step": 6210, |
|
"steps": 6210 |
|
}, |
|
{ |
|
"epoch": 0.828, |
|
"learning_rate": 3.5786624553694038e-06, |
|
"loss": 0.5401, |
|
"r_loss": 0.0, |
|
"step": 6220, |
|
"steps": 6220 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 3.5512112057127163e-06, |
|
"loss": 0.5324, |
|
"r_loss": 0.0, |
|
"step": 6230, |
|
"steps": 6230 |
|
}, |
|
{ |
|
"epoch": 0.831, |
|
"learning_rate": 3.5237599560560283e-06, |
|
"loss": 0.5317, |
|
"r_loss": 0.0, |
|
"step": 6240, |
|
"steps": 6240 |
|
}, |
|
{ |
|
"epoch": 0.832, |
|
"learning_rate": 3.4963087063993404e-06, |
|
"loss": 0.5368, |
|
"r_loss": 0.0, |
|
"step": 6250, |
|
"steps": 6250 |
|
}, |
|
{ |
|
"epoch": 0.834, |
|
"learning_rate": 3.4688574567426533e-06, |
|
"loss": 0.5441, |
|
"r_loss": 0.0, |
|
"step": 6260, |
|
"steps": 6260 |
|
}, |
|
{ |
|
"epoch": 0.835, |
|
"learning_rate": 3.4414062070859653e-06, |
|
"loss": 0.5257, |
|
"r_loss": 0.0, |
|
"step": 6270, |
|
"steps": 6270 |
|
}, |
|
{ |
|
"epoch": 0.836, |
|
"learning_rate": 3.413954957429278e-06, |
|
"loss": 0.535, |
|
"r_loss": 0.0, |
|
"step": 6280, |
|
"steps": 6280 |
|
}, |
|
{ |
|
"epoch": 0.838, |
|
"learning_rate": 3.38650370777259e-06, |
|
"loss": 0.5289, |
|
"r_loss": 0.0, |
|
"step": 6290, |
|
"steps": 6290 |
|
}, |
|
{ |
|
"epoch": 0.839, |
|
"learning_rate": 3.359052458115902e-06, |
|
"loss": 0.534, |
|
"r_loss": 0.0, |
|
"step": 6300, |
|
"steps": 6300 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 3.3316012084592144e-06, |
|
"loss": 0.5325, |
|
"r_loss": 0.0, |
|
"step": 6310, |
|
"steps": 6310 |
|
}, |
|
{ |
|
"epoch": 0.842, |
|
"learning_rate": 3.3041499588025265e-06, |
|
"loss": 0.513, |
|
"r_loss": 0.0, |
|
"step": 6320, |
|
"steps": 6320 |
|
}, |
|
{ |
|
"epoch": 0.843, |
|
"learning_rate": 3.2766987091458394e-06, |
|
"loss": 0.5259, |
|
"r_loss": 0.0, |
|
"step": 6330, |
|
"steps": 6330 |
|
}, |
|
{ |
|
"epoch": 0.844, |
|
"learning_rate": 3.249247459489151e-06, |
|
"loss": 0.5192, |
|
"r_loss": 0.0, |
|
"step": 6340, |
|
"steps": 6340 |
|
}, |
|
{ |
|
"epoch": 0.846, |
|
"learning_rate": 3.221796209832463e-06, |
|
"loss": 0.53, |
|
"r_loss": 0.0, |
|
"step": 6350, |
|
"steps": 6350 |
|
}, |
|
{ |
|
"epoch": 0.847, |
|
"learning_rate": 3.194344960175776e-06, |
|
"loss": 0.5325, |
|
"r_loss": 0.0, |
|
"step": 6360, |
|
"steps": 6360 |
|
}, |
|
{ |
|
"epoch": 0.848, |
|
"learning_rate": 3.166893710519088e-06, |
|
"loss": 0.5253, |
|
"r_loss": 0.0, |
|
"step": 6370, |
|
"steps": 6370 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 3.1394424608624005e-06, |
|
"loss": 0.5233, |
|
"r_loss": 0.0, |
|
"step": 6380, |
|
"steps": 6380 |
|
}, |
|
{ |
|
"epoch": 0.851, |
|
"learning_rate": 3.1119912112057125e-06, |
|
"loss": 0.522, |
|
"r_loss": 0.0, |
|
"step": 6390, |
|
"steps": 6390 |
|
}, |
|
{ |
|
"epoch": 0.852, |
|
"learning_rate": 3.084539961549025e-06, |
|
"loss": 0.5424, |
|
"r_loss": 0.0, |
|
"step": 6400, |
|
"steps": 6400 |
|
}, |
|
{ |
|
"epoch": 0.854, |
|
"learning_rate": 3.057088711892337e-06, |
|
"loss": 0.5301, |
|
"r_loss": 0.0, |
|
"step": 6410, |
|
"steps": 6410 |
|
}, |
|
{ |
|
"epoch": 0.855, |
|
"learning_rate": 3.029637462235649e-06, |
|
"loss": 0.5258, |
|
"r_loss": 0.0, |
|
"step": 6420, |
|
"steps": 6420 |
|
}, |
|
{ |
|
"epoch": 0.856, |
|
"learning_rate": 3.002186212578962e-06, |
|
"loss": 0.5255, |
|
"r_loss": 0.0, |
|
"step": 6430, |
|
"steps": 6430 |
|
}, |
|
{ |
|
"epoch": 0.858, |
|
"learning_rate": 2.974734962922274e-06, |
|
"loss": 0.5303, |
|
"r_loss": 0.0, |
|
"step": 6440, |
|
"steps": 6440 |
|
}, |
|
{ |
|
"epoch": 0.859, |
|
"learning_rate": 2.9472837132655866e-06, |
|
"loss": 0.5293, |
|
"r_loss": 0.0, |
|
"step": 6450, |
|
"steps": 6450 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 2.9198324636088986e-06, |
|
"loss": 0.5339, |
|
"r_loss": 0.0, |
|
"step": 6460, |
|
"steps": 6460 |
|
}, |
|
{ |
|
"epoch": 0.862, |
|
"learning_rate": 2.8923812139522107e-06, |
|
"loss": 0.5175, |
|
"r_loss": 0.0, |
|
"step": 6470, |
|
"steps": 6470 |
|
}, |
|
{ |
|
"epoch": 0.863, |
|
"learning_rate": 2.864929964295523e-06, |
|
"loss": 0.5302, |
|
"r_loss": 0.0, |
|
"step": 6480, |
|
"steps": 6480 |
|
}, |
|
{ |
|
"epoch": 0.864, |
|
"learning_rate": 2.8374787146388352e-06, |
|
"loss": 0.5434, |
|
"r_loss": 0.0, |
|
"step": 6490, |
|
"steps": 6490 |
|
}, |
|
{ |
|
"epoch": 0.866, |
|
"learning_rate": 2.8100274649821477e-06, |
|
"loss": 0.5311, |
|
"r_loss": 0.0, |
|
"step": 6500, |
|
"steps": 6500 |
|
}, |
|
{ |
|
"epoch": 0.867, |
|
"learning_rate": 2.7825762153254598e-06, |
|
"loss": 0.5399, |
|
"r_loss": 0.0, |
|
"step": 6510, |
|
"steps": 6510 |
|
}, |
|
{ |
|
"epoch": 0.868, |
|
"learning_rate": 2.755124965668772e-06, |
|
"loss": 0.5403, |
|
"r_loss": 0.0, |
|
"step": 6520, |
|
"steps": 6520 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 2.7276737160120847e-06, |
|
"loss": 0.518, |
|
"r_loss": 0.0, |
|
"step": 6530, |
|
"steps": 6530 |
|
}, |
|
{ |
|
"epoch": 0.871, |
|
"learning_rate": 2.7002224663553968e-06, |
|
"loss": 0.5257, |
|
"r_loss": 0.0, |
|
"step": 6540, |
|
"steps": 6540 |
|
}, |
|
{ |
|
"epoch": 0.872, |
|
"learning_rate": 2.6727712166987093e-06, |
|
"loss": 0.5383, |
|
"r_loss": 0.0, |
|
"step": 6550, |
|
"steps": 6550 |
|
}, |
|
{ |
|
"epoch": 0.874, |
|
"learning_rate": 2.6453199670420213e-06, |
|
"loss": 0.5359, |
|
"r_loss": 0.0, |
|
"step": 6560, |
|
"steps": 6560 |
|
}, |
|
{ |
|
"epoch": 0.875, |
|
"learning_rate": 2.6178687173853334e-06, |
|
"loss": 0.5275, |
|
"r_loss": 0.0, |
|
"step": 6570, |
|
"steps": 6570 |
|
}, |
|
{ |
|
"epoch": 0.876, |
|
"learning_rate": 2.590417467728646e-06, |
|
"loss": 0.5237, |
|
"r_loss": 0.0, |
|
"step": 6580, |
|
"steps": 6580 |
|
}, |
|
{ |
|
"epoch": 0.878, |
|
"learning_rate": 2.562966218071958e-06, |
|
"loss": 0.5238, |
|
"r_loss": 0.0, |
|
"step": 6590, |
|
"steps": 6590 |
|
}, |
|
{ |
|
"epoch": 0.879, |
|
"learning_rate": 2.535514968415271e-06, |
|
"loss": 0.5244, |
|
"r_loss": 0.0, |
|
"step": 6600, |
|
"steps": 6600 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 2.508063718758583e-06, |
|
"loss": 0.5351, |
|
"r_loss": 0.0, |
|
"step": 6610, |
|
"steps": 6610 |
|
}, |
|
{ |
|
"epoch": 0.882, |
|
"learning_rate": 2.480612469101895e-06, |
|
"loss": 0.5365, |
|
"r_loss": 0.0, |
|
"step": 6620, |
|
"steps": 6620 |
|
}, |
|
{ |
|
"epoch": 0.883, |
|
"learning_rate": 2.4531612194452074e-06, |
|
"loss": 0.5302, |
|
"r_loss": 0.0, |
|
"step": 6630, |
|
"steps": 6630 |
|
}, |
|
{ |
|
"epoch": 0.884, |
|
"learning_rate": 2.42570996978852e-06, |
|
"loss": 0.5254, |
|
"r_loss": 0.0, |
|
"step": 6640, |
|
"steps": 6640 |
|
}, |
|
{ |
|
"epoch": 0.886, |
|
"learning_rate": 2.398258720131832e-06, |
|
"loss": 0.5339, |
|
"r_loss": 0.0, |
|
"step": 6650, |
|
"steps": 6650 |
|
}, |
|
{ |
|
"epoch": 0.887, |
|
"learning_rate": 2.370807470475144e-06, |
|
"loss": 0.5343, |
|
"r_loss": 0.0, |
|
"step": 6660, |
|
"steps": 6660 |
|
}, |
|
{ |
|
"epoch": 0.888, |
|
"learning_rate": 2.3433562208184565e-06, |
|
"loss": 0.5367, |
|
"r_loss": 0.0, |
|
"step": 6670, |
|
"steps": 6670 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 2.3159049711617685e-06, |
|
"loss": 0.5354, |
|
"r_loss": 0.0, |
|
"step": 6680, |
|
"steps": 6680 |
|
}, |
|
{ |
|
"epoch": 0.891, |
|
"learning_rate": 2.288453721505081e-06, |
|
"loss": 0.5185, |
|
"r_loss": 0.0, |
|
"step": 6690, |
|
"steps": 6690 |
|
}, |
|
{ |
|
"epoch": 0.892, |
|
"learning_rate": 2.261002471848393e-06, |
|
"loss": 0.5343, |
|
"r_loss": 0.0, |
|
"step": 6700, |
|
"steps": 6700 |
|
}, |
|
{ |
|
"epoch": 0.894, |
|
"learning_rate": 2.2335512221917056e-06, |
|
"loss": 0.5156, |
|
"r_loss": 0.0, |
|
"step": 6710, |
|
"steps": 6710 |
|
}, |
|
{ |
|
"epoch": 0.895, |
|
"learning_rate": 2.2060999725350176e-06, |
|
"loss": 0.5254, |
|
"r_loss": 0.0, |
|
"step": 6720, |
|
"steps": 6720 |
|
}, |
|
{ |
|
"epoch": 0.896, |
|
"learning_rate": 2.17864872287833e-06, |
|
"loss": 0.5123, |
|
"r_loss": 0.0, |
|
"step": 6730, |
|
"steps": 6730 |
|
}, |
|
{ |
|
"epoch": 0.898, |
|
"learning_rate": 2.1511974732216426e-06, |
|
"loss": 0.5281, |
|
"r_loss": 0.0, |
|
"step": 6740, |
|
"steps": 6740 |
|
}, |
|
{ |
|
"epoch": 0.899, |
|
"learning_rate": 2.1237462235649546e-06, |
|
"loss": 0.5234, |
|
"r_loss": 0.0, |
|
"step": 6750, |
|
"steps": 6750 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 2.0962949739082667e-06, |
|
"loss": 0.5221, |
|
"r_loss": 0.0, |
|
"step": 6760, |
|
"steps": 6760 |
|
}, |
|
{ |
|
"epoch": 0.902, |
|
"learning_rate": 2.068843724251579e-06, |
|
"loss": 0.5177, |
|
"r_loss": 0.0, |
|
"step": 6770, |
|
"steps": 6770 |
|
}, |
|
{ |
|
"epoch": 0.903, |
|
"learning_rate": 2.0413924745948916e-06, |
|
"loss": 0.5225, |
|
"r_loss": 0.0, |
|
"step": 6780, |
|
"steps": 6780 |
|
}, |
|
{ |
|
"epoch": 0.904, |
|
"learning_rate": 2.0139412249382037e-06, |
|
"loss": 0.531, |
|
"r_loss": 0.0, |
|
"step": 6790, |
|
"steps": 6790 |
|
}, |
|
{ |
|
"epoch": 0.906, |
|
"learning_rate": 1.986489975281516e-06, |
|
"loss": 0.5303, |
|
"r_loss": 0.0, |
|
"step": 6800, |
|
"steps": 6800 |
|
}, |
|
{ |
|
"epoch": 0.907, |
|
"learning_rate": 1.9590387256248282e-06, |
|
"loss": 0.5236, |
|
"r_loss": 0.0, |
|
"step": 6810, |
|
"steps": 6810 |
|
}, |
|
{ |
|
"epoch": 0.908, |
|
"learning_rate": 1.9315874759681403e-06, |
|
"loss": 0.5192, |
|
"r_loss": 0.0, |
|
"step": 6820, |
|
"steps": 6820 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 1.9041362263114528e-06, |
|
"loss": 0.5255, |
|
"r_loss": 0.0, |
|
"step": 6830, |
|
"steps": 6830 |
|
}, |
|
{ |
|
"epoch": 0.911, |
|
"learning_rate": 1.8766849766547652e-06, |
|
"loss": 0.5121, |
|
"r_loss": 0.0, |
|
"step": 6840, |
|
"steps": 6840 |
|
}, |
|
{ |
|
"epoch": 0.912, |
|
"learning_rate": 1.8492337269980775e-06, |
|
"loss": 0.5227, |
|
"r_loss": 0.0, |
|
"step": 6850, |
|
"steps": 6850 |
|
}, |
|
{ |
|
"epoch": 0.914, |
|
"learning_rate": 1.8217824773413896e-06, |
|
"loss": 0.5183, |
|
"r_loss": 0.0, |
|
"step": 6860, |
|
"steps": 6860 |
|
}, |
|
{ |
|
"epoch": 0.915, |
|
"learning_rate": 1.7943312276847018e-06, |
|
"loss": 0.5269, |
|
"r_loss": 0.0, |
|
"step": 6870, |
|
"steps": 6870 |
|
}, |
|
{ |
|
"epoch": 0.916, |
|
"learning_rate": 1.7668799780280141e-06, |
|
"loss": 0.5287, |
|
"r_loss": 0.0, |
|
"step": 6880, |
|
"steps": 6880 |
|
}, |
|
{ |
|
"epoch": 0.918, |
|
"learning_rate": 1.7394287283713266e-06, |
|
"loss": 0.519, |
|
"r_loss": 0.0, |
|
"step": 6890, |
|
"steps": 6890 |
|
}, |
|
{ |
|
"epoch": 0.919, |
|
"learning_rate": 1.7119774787146389e-06, |
|
"loss": 0.5352, |
|
"r_loss": 0.0, |
|
"step": 6900, |
|
"steps": 6900 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 1.684526229057951e-06, |
|
"loss": 0.5184, |
|
"r_loss": 0.0, |
|
"step": 6910, |
|
"steps": 6910 |
|
}, |
|
{ |
|
"epoch": 0.922, |
|
"learning_rate": 1.6570749794012632e-06, |
|
"loss": 0.524, |
|
"r_loss": 0.0, |
|
"step": 6920, |
|
"steps": 6920 |
|
}, |
|
{ |
|
"epoch": 0.923, |
|
"learning_rate": 1.6296237297445754e-06, |
|
"loss": 0.53, |
|
"r_loss": 0.0, |
|
"step": 6930, |
|
"steps": 6930 |
|
}, |
|
{ |
|
"epoch": 0.924, |
|
"learning_rate": 1.602172480087888e-06, |
|
"loss": 0.5336, |
|
"r_loss": 0.0, |
|
"step": 6940, |
|
"steps": 6940 |
|
}, |
|
{ |
|
"epoch": 0.926, |
|
"learning_rate": 1.5747212304312002e-06, |
|
"loss": 0.5254, |
|
"r_loss": 0.0, |
|
"step": 6950, |
|
"steps": 6950 |
|
}, |
|
{ |
|
"epoch": 0.927, |
|
"learning_rate": 1.5472699807745125e-06, |
|
"loss": 0.5217, |
|
"r_loss": 0.0, |
|
"step": 6960, |
|
"steps": 6960 |
|
}, |
|
{ |
|
"epoch": 0.928, |
|
"learning_rate": 1.5198187311178245e-06, |
|
"loss": 0.5219, |
|
"r_loss": 0.0, |
|
"step": 6970, |
|
"steps": 6970 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 1.492367481461137e-06, |
|
"loss": 0.5325, |
|
"r_loss": 0.0, |
|
"step": 6980, |
|
"steps": 6980 |
|
}, |
|
{ |
|
"epoch": 0.931, |
|
"learning_rate": 1.4649162318044493e-06, |
|
"loss": 0.5206, |
|
"r_loss": 0.0, |
|
"step": 6990, |
|
"steps": 6990 |
|
}, |
|
{ |
|
"epoch": 0.932, |
|
"learning_rate": 1.4374649821477615e-06, |
|
"loss": 0.5361, |
|
"r_loss": 0.0, |
|
"step": 7000, |
|
"steps": 7000 |
|
}, |
|
{ |
|
"epoch": 0.934, |
|
"learning_rate": 1.4100137324910738e-06, |
|
"loss": 0.5135, |
|
"r_loss": 0.0, |
|
"step": 7010, |
|
"steps": 7010 |
|
}, |
|
{ |
|
"epoch": 0.935, |
|
"learning_rate": 1.3825624828343859e-06, |
|
"loss": 0.5374, |
|
"r_loss": 0.0, |
|
"step": 7020, |
|
"steps": 7020 |
|
}, |
|
{ |
|
"epoch": 0.936, |
|
"learning_rate": 1.3551112331776983e-06, |
|
"loss": 0.524, |
|
"r_loss": 0.0, |
|
"step": 7030, |
|
"steps": 7030 |
|
}, |
|
{ |
|
"epoch": 0.938, |
|
"learning_rate": 1.3276599835210106e-06, |
|
"loss": 0.5217, |
|
"r_loss": 0.0, |
|
"step": 7040, |
|
"steps": 7040 |
|
}, |
|
{ |
|
"epoch": 0.939, |
|
"learning_rate": 1.3002087338643229e-06, |
|
"loss": 0.5291, |
|
"r_loss": 0.0, |
|
"step": 7050, |
|
"steps": 7050 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.2727574842076354e-06, |
|
"loss": 0.5251, |
|
"r_loss": 0.0, |
|
"step": 7060, |
|
"steps": 7060 |
|
}, |
|
{ |
|
"epoch": 0.942, |
|
"learning_rate": 1.2453062345509474e-06, |
|
"loss": 0.5211, |
|
"r_loss": 0.0, |
|
"step": 7070, |
|
"steps": 7070 |
|
}, |
|
{ |
|
"epoch": 0.943, |
|
"learning_rate": 1.21785498489426e-06, |
|
"loss": 0.5112, |
|
"r_loss": 0.0, |
|
"step": 7080, |
|
"steps": 7080 |
|
}, |
|
{ |
|
"epoch": 0.944, |
|
"learning_rate": 1.190403735237572e-06, |
|
"loss": 0.5235, |
|
"r_loss": 0.0, |
|
"step": 7090, |
|
"steps": 7090 |
|
}, |
|
{ |
|
"epoch": 0.946, |
|
"learning_rate": 1.1629524855808842e-06, |
|
"loss": 0.5224, |
|
"r_loss": 0.0, |
|
"step": 7100, |
|
"steps": 7100 |
|
}, |
|
{ |
|
"epoch": 0.947, |
|
"learning_rate": 1.1355012359241965e-06, |
|
"loss": 0.5225, |
|
"r_loss": 0.0, |
|
"step": 7110, |
|
"steps": 7110 |
|
}, |
|
{ |
|
"epoch": 0.948, |
|
"learning_rate": 1.1080499862675088e-06, |
|
"loss": 0.5262, |
|
"r_loss": 0.0, |
|
"step": 7120, |
|
"steps": 7120 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.0805987366108212e-06, |
|
"loss": 0.5164, |
|
"r_loss": 0.0, |
|
"step": 7130, |
|
"steps": 7130 |
|
}, |
|
{ |
|
"epoch": 0.951, |
|
"learning_rate": 1.0531474869541333e-06, |
|
"loss": 0.5273, |
|
"r_loss": 0.0, |
|
"step": 7140, |
|
"steps": 7140 |
|
}, |
|
{ |
|
"epoch": 0.952, |
|
"learning_rate": 1.0256962372974458e-06, |
|
"loss": 0.5264, |
|
"r_loss": 0.0, |
|
"step": 7150, |
|
"steps": 7150 |
|
}, |
|
{ |
|
"epoch": 0.954, |
|
"learning_rate": 9.98244987640758e-07, |
|
"loss": 0.5396, |
|
"r_loss": 0.0, |
|
"step": 7160, |
|
"steps": 7160 |
|
}, |
|
{ |
|
"epoch": 0.955, |
|
"learning_rate": 9.7079373798407e-07, |
|
"loss": 0.5188, |
|
"r_loss": 0.0, |
|
"step": 7170, |
|
"steps": 7170 |
|
}, |
|
{ |
|
"epoch": 0.956, |
|
"learning_rate": 9.433424883273827e-07, |
|
"loss": 0.5253, |
|
"r_loss": 0.0, |
|
"step": 7180, |
|
"steps": 7180 |
|
}, |
|
{ |
|
"epoch": 0.958, |
|
"learning_rate": 9.158912386706948e-07, |
|
"loss": 0.5289, |
|
"r_loss": 0.0, |
|
"step": 7190, |
|
"steps": 7190 |
|
}, |
|
{ |
|
"epoch": 0.959, |
|
"learning_rate": 8.884399890140071e-07, |
|
"loss": 0.5341, |
|
"r_loss": 0.0, |
|
"step": 7200, |
|
"steps": 7200 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 8.609887393573195e-07, |
|
"loss": 0.5255, |
|
"r_loss": 0.0, |
|
"step": 7210, |
|
"steps": 7210 |
|
}, |
|
{ |
|
"epoch": 0.962, |
|
"learning_rate": 8.335374897006316e-07, |
|
"loss": 0.5181, |
|
"r_loss": 0.0, |
|
"step": 7220, |
|
"steps": 7220 |
|
}, |
|
{ |
|
"epoch": 0.963, |
|
"learning_rate": 8.06086240043944e-07, |
|
"loss": 0.5258, |
|
"r_loss": 0.0, |
|
"step": 7230, |
|
"steps": 7230 |
|
}, |
|
{ |
|
"epoch": 0.964, |
|
"learning_rate": 7.786349903872563e-07, |
|
"loss": 0.5237, |
|
"r_loss": 0.0, |
|
"step": 7240, |
|
"steps": 7240 |
|
}, |
|
{ |
|
"epoch": 0.966, |
|
"learning_rate": 7.511837407305686e-07, |
|
"loss": 0.5165, |
|
"r_loss": 0.0, |
|
"step": 7250, |
|
"steps": 7250 |
|
}, |
|
{ |
|
"epoch": 0.967, |
|
"learning_rate": 7.237324910738808e-07, |
|
"loss": 0.5336, |
|
"r_loss": 0.0, |
|
"step": 7260, |
|
"steps": 7260 |
|
}, |
|
{ |
|
"epoch": 0.968, |
|
"learning_rate": 6.96281241417193e-07, |
|
"loss": 0.5271, |
|
"r_loss": 0.0, |
|
"step": 7270, |
|
"steps": 7270 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 6.688299917605054e-07, |
|
"loss": 0.5185, |
|
"r_loss": 0.0, |
|
"step": 7280, |
|
"steps": 7280 |
|
}, |
|
{ |
|
"epoch": 0.971, |
|
"learning_rate": 6.413787421038177e-07, |
|
"loss": 0.5241, |
|
"r_loss": 0.0, |
|
"step": 7290, |
|
"steps": 7290 |
|
}, |
|
{ |
|
"epoch": 0.972, |
|
"learning_rate": 6.1392749244713e-07, |
|
"loss": 0.5139, |
|
"r_loss": 0.0, |
|
"step": 7300, |
|
"steps": 7300 |
|
}, |
|
{ |
|
"epoch": 0.974, |
|
"learning_rate": 5.864762427904422e-07, |
|
"loss": 0.5298, |
|
"r_loss": 0.0, |
|
"step": 7310, |
|
"steps": 7310 |
|
}, |
|
{ |
|
"epoch": 0.975, |
|
"learning_rate": 5.590249931337544e-07, |
|
"loss": 0.5104, |
|
"r_loss": 0.0, |
|
"step": 7320, |
|
"steps": 7320 |
|
}, |
|
{ |
|
"epoch": 0.976, |
|
"learning_rate": 5.315737434770667e-07, |
|
"loss": 0.5294, |
|
"r_loss": 0.0, |
|
"step": 7330, |
|
"steps": 7330 |
|
}, |
|
{ |
|
"epoch": 0.978, |
|
"learning_rate": 5.041224938203791e-07, |
|
"loss": 0.5149, |
|
"r_loss": 0.0, |
|
"step": 7340, |
|
"steps": 7340 |
|
}, |
|
{ |
|
"epoch": 0.979, |
|
"learning_rate": 4.7667124416369135e-07, |
|
"loss": 0.5209, |
|
"r_loss": 0.0, |
|
"step": 7350, |
|
"steps": 7350 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 4.4921999450700356e-07, |
|
"loss": 0.5275, |
|
"r_loss": 0.0, |
|
"step": 7360, |
|
"steps": 7360 |
|
}, |
|
{ |
|
"epoch": 0.982, |
|
"learning_rate": 4.2176874485031583e-07, |
|
"loss": 0.5339, |
|
"r_loss": 0.0, |
|
"step": 7370, |
|
"steps": 7370 |
|
}, |
|
{ |
|
"epoch": 0.983, |
|
"learning_rate": 3.9431749519362815e-07, |
|
"loss": 0.5228, |
|
"r_loss": 0.0, |
|
"step": 7380, |
|
"steps": 7380 |
|
}, |
|
{ |
|
"epoch": 0.984, |
|
"learning_rate": 3.668662455369404e-07, |
|
"loss": 0.5294, |
|
"r_loss": 0.0, |
|
"step": 7390, |
|
"steps": 7390 |
|
}, |
|
{ |
|
"epoch": 0.986, |
|
"learning_rate": 3.394149958802527e-07, |
|
"loss": 0.5226, |
|
"r_loss": 0.0, |
|
"step": 7400, |
|
"steps": 7400 |
|
}, |
|
{ |
|
"epoch": 0.987, |
|
"learning_rate": 3.11963746223565e-07, |
|
"loss": 0.5099, |
|
"r_loss": 0.0, |
|
"step": 7410, |
|
"steps": 7410 |
|
}, |
|
{ |
|
"epoch": 0.988, |
|
"learning_rate": 2.845124965668772e-07, |
|
"loss": 0.5312, |
|
"r_loss": 0.0, |
|
"step": 7420, |
|
"steps": 7420 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 2.5706124691018954e-07, |
|
"loss": 0.5155, |
|
"r_loss": 0.0, |
|
"step": 7430, |
|
"steps": 7430 |
|
}, |
|
{ |
|
"epoch": 0.991, |
|
"learning_rate": 2.2960999725350179e-07, |
|
"loss": 0.5098, |
|
"r_loss": 0.0, |
|
"step": 7440, |
|
"steps": 7440 |
|
}, |
|
{ |
|
"epoch": 0.992, |
|
"learning_rate": 2.0215874759681408e-07, |
|
"loss": 0.5206, |
|
"r_loss": 0.0, |
|
"step": 7450, |
|
"steps": 7450 |
|
}, |
|
{ |
|
"epoch": 0.994, |
|
"learning_rate": 1.7470749794012635e-07, |
|
"loss": 0.5252, |
|
"r_loss": 0.0, |
|
"step": 7460, |
|
"steps": 7460 |
|
}, |
|
{ |
|
"epoch": 0.995, |
|
"learning_rate": 1.4725624828343862e-07, |
|
"loss": 0.5238, |
|
"r_loss": 0.0, |
|
"step": 7470, |
|
"steps": 7470 |
|
}, |
|
{ |
|
"epoch": 0.996, |
|
"learning_rate": 1.1980499862675088e-07, |
|
"loss": 0.5272, |
|
"r_loss": 0.0, |
|
"step": 7480, |
|
"steps": 7480 |
|
}, |
|
{ |
|
"epoch": 0.998, |
|
"learning_rate": 9.235374897006317e-08, |
|
"loss": 0.528, |
|
"r_loss": 0.0, |
|
"step": 7490, |
|
"steps": 7490 |
|
}, |
|
{ |
|
"epoch": 0.999, |
|
"learning_rate": 6.490249931337545e-08, |
|
"loss": 0.5232, |
|
"r_loss": 0.0, |
|
"step": 7500, |
|
"steps": 7500 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"step": 7508, |
|
"steps": 7508, |
|
"total_flos": 1.950614777001463e+19, |
|
"train_loss": 0.5782644650881508, |
|
"train_runtime": 47075.3922, |
|
"train_samples_per_second": 20.414, |
|
"train_steps_per_second": 0.159 |
|
} |
|
], |
|
"max_steps": 7508, |
|
"num_train_epochs": 1, |
|
"total_flos": 1.950614777001463e+19, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|