|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 4.979591836734694, |
|
"eval_steps": 500, |
|
"global_step": 915, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 1.0000000000000002e-06, |
|
"loss": 1.7555, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 2.0000000000000003e-06, |
|
"loss": 1.8402, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 3e-06, |
|
"loss": 1.6331, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 4.000000000000001e-06, |
|
"loss": 1.9207, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 5e-06, |
|
"loss": 1.6091, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 6e-06, |
|
"loss": 1.7484, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 7e-06, |
|
"loss": 1.7995, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 8.000000000000001e-06, |
|
"loss": 1.7021, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 9e-06, |
|
"loss": 1.7542, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 1e-05, |
|
"loss": 1.69, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 9.99996987395229e-06, |
|
"loss": 1.6178, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.999879496172195e-06, |
|
"loss": 1.8358, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 9.999728867748802e-06, |
|
"loss": 1.8087, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.999517990497248e-06, |
|
"loss": 1.7274, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 9.999246866958693e-06, |
|
"loss": 1.9479, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.998915500400287e-06, |
|
"loss": 1.6799, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 9.998523894815137e-06, |
|
"loss": 1.657, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.998072054922257e-06, |
|
"loss": 1.6667, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 9.997559986166503e-06, |
|
"loss": 1.7878, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.99698769471852e-06, |
|
"loss": 1.6456, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 9.99635518747466e-06, |
|
"loss": 1.6963, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 9.9956624720569e-06, |
|
"loss": 1.4927, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.99490955681275e-06, |
|
"loss": 1.7288, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 9.994096450815157e-06, |
|
"loss": 1.8489, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.993223163862385e-06, |
|
"loss": 1.6431, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 9.992289706477912e-06, |
|
"loss": 1.5962, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.991296089910289e-06, |
|
"loss": 1.6551, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 9.99024232613301e-06, |
|
"loss": 1.7767, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.989128427844374e-06, |
|
"loss": 1.6875, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 9.98795440846732e-06, |
|
"loss": 1.7293, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.986720282149272e-06, |
|
"loss": 1.8094, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 9.985426063761973e-06, |
|
"loss": 1.7926, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 9.984071768901294e-06, |
|
"loss": 1.5949, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.982657413887055e-06, |
|
"loss": 1.4929, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 9.981183015762831e-06, |
|
"loss": 1.4097, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.979648592295735e-06, |
|
"loss": 1.6776, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 9.978054161976209e-06, |
|
"loss": 1.5844, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.976399744017812e-06, |
|
"loss": 1.6987, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 9.974685358356972e-06, |
|
"loss": 1.5576, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.972911025652754e-06, |
|
"loss": 1.6193, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 9.97107676728661e-06, |
|
"loss": 1.6233, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.969182605362125e-06, |
|
"loss": 1.6317, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 9.96722856270474e-06, |
|
"loss": 1.7902, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.965214662861492e-06, |
|
"loss": 1.4772, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 9.963140930100713e-06, |
|
"loss": 1.5038, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 9.961007389411758e-06, |
|
"loss": 1.5661, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.958814066504684e-06, |
|
"loss": 1.4406, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 9.956560987809948e-06, |
|
"loss": 1.6141, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.954248180478096e-06, |
|
"loss": 1.467, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 9.951875672379424e-06, |
|
"loss": 1.5145, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.94944349210365e-06, |
|
"loss": 1.4695, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 9.946951668959567e-06, |
|
"loss": 1.6004, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.944400232974683e-06, |
|
"loss": 1.4118, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 9.941789214894876e-06, |
|
"loss": 1.6423, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.939118646184007e-06, |
|
"loss": 1.5948, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 9.936388559023547e-06, |
|
"loss": 1.5026, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 9.93359898631219e-06, |
|
"loss": 1.3926, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.930749961665459e-06, |
|
"loss": 1.2179, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 9.92784151941529e-06, |
|
"loss": 1.5675, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.924873694609636e-06, |
|
"loss": 1.5331, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 9.921846523012028e-06, |
|
"loss": 1.4585, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.918760041101152e-06, |
|
"loss": 1.3743, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 9.915614286070408e-06, |
|
"loss": 1.5507, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.912409295827463e-06, |
|
"loss": 1.1745, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 9.909145108993794e-06, |
|
"loss": 1.4218, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 9.905821764904218e-06, |
|
"loss": 1.295, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 9.902439303606424e-06, |
|
"loss": 1.3678, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 9.898997765860492e-06, |
|
"loss": 1.4748, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 9.89549719313839e-06, |
|
"loss": 1.4412, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 9.891937627623486e-06, |
|
"loss": 1.3173, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 9.888319112210041e-06, |
|
"loss": 1.371, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 9.884641690502677e-06, |
|
"loss": 1.2869, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 9.88090540681587e-06, |
|
"loss": 1.3332, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 9.877110306173403e-06, |
|
"loss": 1.4246, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 9.873256434307828e-06, |
|
"loss": 1.4935, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 9.86934383765992e-06, |
|
"loss": 1.2236, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 9.865372563378102e-06, |
|
"loss": 1.3858, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 9.8613426593179e-06, |
|
"loss": 1.3957, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 9.857254174041342e-06, |
|
"loss": 1.295, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 9.853107156816393e-06, |
|
"loss": 1.3521, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 9.848901657616344e-06, |
|
"loss": 1.3231, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 9.844637727119227e-06, |
|
"loss": 1.1221, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 9.840315416707188e-06, |
|
"loss": 1.2744, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 9.83593477846588e-06, |
|
"loss": 1.46, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 9.831495865183832e-06, |
|
"loss": 1.3859, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 9.826998730351806e-06, |
|
"loss": 1.2469, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 9.822443428162165e-06, |
|
"loss": 1.1787, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 9.817830013508207e-06, |
|
"loss": 1.2879, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 9.81315854198351e-06, |
|
"loss": 1.2606, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 9.808429069881267e-06, |
|
"loss": 1.1404, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 9.803641654193599e-06, |
|
"loss": 1.3141, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 9.798796352610869e-06, |
|
"loss": 1.276, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 9.793893223520992e-06, |
|
"loss": 1.1235, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 9.78893232600873e-06, |
|
"loss": 1.25, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.783913719854977e-06, |
|
"loss": 1.3262, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.77883746553604e-06, |
|
"loss": 1.136, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.773703624222908e-06, |
|
"loss": 1.3941, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.768512257780524e-06, |
|
"loss": 1.2803, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.76326342876703e-06, |
|
"loss": 1.3667, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.757957200433011e-06, |
|
"loss": 1.2535, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 9.752593636720744e-06, |
|
"loss": 1.1887, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 9.747172802263423e-06, |
|
"loss": 1.2551, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 9.74169476238437e-06, |
|
"loss": 1.3562, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 9.736159583096262e-06, |
|
"loss": 1.264, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 9.730567331100333e-06, |
|
"loss": 1.2811, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 9.72491807378556e-06, |
|
"loss": 1.2777, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 9.719211879227863e-06, |
|
"loss": 1.1766, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 9.713448816189275e-06, |
|
"loss": 1.2182, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 9.707628954117122e-06, |
|
"loss": 1.2595, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 9.701752363143183e-06, |
|
"loss": 1.1055, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 9.69581911408284e-06, |
|
"loss": 1.1307, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 9.68982927843423e-06, |
|
"loss": 1.2346, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 9.683782928377385e-06, |
|
"loss": 1.1779, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 9.677680136773357e-06, |
|
"loss": 1.0882, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 9.67152097716334e-06, |
|
"loss": 1.1589, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 9.665305523767792e-06, |
|
"loss": 1.2117, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 9.659033851485527e-06, |
|
"loss": 1.234, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 9.652706035892828e-06, |
|
"loss": 1.1928, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 9.64632215324252e-06, |
|
"loss": 1.1914, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 9.639882280463071e-06, |
|
"loss": 1.2158, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 9.633386495157643e-06, |
|
"loss": 1.1415, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 9.62683487560317e-06, |
|
"loss": 1.1528, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 9.620227500749418e-06, |
|
"loss": 0.9916, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 9.613564450218019e-06, |
|
"loss": 1.118, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 9.606845804301523e-06, |
|
"loss": 1.2559, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 9.600071643962433e-06, |
|
"loss": 1.147, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 9.593242050832219e-06, |
|
"loss": 1.1697, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 9.586357107210338e-06, |
|
"loss": 1.2459, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 9.57941689606325e-06, |
|
"loss": 1.1382, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 9.572421501023403e-06, |
|
"loss": 1.3151, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 9.565371006388242e-06, |
|
"loss": 1.0931, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 9.558265497119182e-06, |
|
"loss": 1.0589, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 9.551105058840588e-06, |
|
"loss": 1.1522, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 9.54388977783874e-06, |
|
"loss": 1.0964, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 9.536619741060799e-06, |
|
"loss": 1.0978, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 9.529295036113755e-06, |
|
"loss": 1.1552, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 9.521915751263373e-06, |
|
"loss": 1.1139, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 9.514481975433125e-06, |
|
"loss": 1.1417, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 9.50699379820313e-06, |
|
"loss": 1.0902, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 9.499451309809058e-06, |
|
"loss": 1.1083, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 9.491854601141057e-06, |
|
"loss": 1.0415, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 9.484203763742647e-06, |
|
"loss": 1.1145, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 9.476498889809628e-06, |
|
"loss": 1.0296, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 9.468740072188961e-06, |
|
"loss": 1.1118, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 9.460927404377647e-06, |
|
"loss": 0.9663, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 9.45306098052161e-06, |
|
"loss": 1.0643, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 9.44514089541455e-06, |
|
"loss": 1.1021, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 9.437167244496814e-06, |
|
"loss": 1.091, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 9.429140123854239e-06, |
|
"loss": 1.0772, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 9.421059630216992e-06, |
|
"loss": 1.1406, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 9.412925860958404e-06, |
|
"loss": 1.0528, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 9.404738914093808e-06, |
|
"loss": 1.0789, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 9.396498888279344e-06, |
|
"loss": 1.0843, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 9.388205882810773e-06, |
|
"loss": 1.1047, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 9.37985999762229e-06, |
|
"loss": 1.058, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 9.371461333285308e-06, |
|
"loss": 1.0475, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 9.363009991007252e-06, |
|
"loss": 1.0543, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 9.354506072630337e-06, |
|
"loss": 0.9872, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 9.345949680630347e-06, |
|
"loss": 1.1849, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 9.337340918115385e-06, |
|
"loss": 1.0618, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 9.328679888824654e-06, |
|
"loss": 1.0944, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 9.319966697127181e-06, |
|
"loss": 1.0915, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 9.311201448020582e-06, |
|
"loss": 0.9994, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 9.30238424712978e-06, |
|
"loss": 0.9176, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 9.29351520070574e-06, |
|
"loss": 1.0784, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 9.28459441562419e-06, |
|
"loss": 1.062, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 9.275621999384327e-06, |
|
"loss": 1.0744, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 9.26659806010753e-06, |
|
"loss": 1.0254, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 9.257522706536046e-06, |
|
"loss": 0.9448, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 9.24839604803169e-06, |
|
"loss": 1.1357, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 9.239218194574522e-06, |
|
"loss": 1.0526, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 9.229989256761522e-06, |
|
"loss": 1.0148, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 9.220709345805259e-06, |
|
"loss": 1.0845, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 9.211378573532549e-06, |
|
"loss": 1.1066, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 9.201997052383107e-06, |
|
"loss": 1.011, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 9.192564895408195e-06, |
|
"loss": 1.0926, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 9.183082216269259e-06, |
|
"loss": 0.9629, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 9.173549129236554e-06, |
|
"loss": 1.0738, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 9.163965749187777e-06, |
|
"loss": 1.0841, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 9.154332191606671e-06, |
|
"loss": 1.0281, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 9.144648572581645e-06, |
|
"loss": 1.1063, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 9.134915008804364e-06, |
|
"loss": 1.0057, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 9.12513161756835e-06, |
|
"loss": 1.0011, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 9.11529851676757e-06, |
|
"loss": 1.1683, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 9.105415824895008e-06, |
|
"loss": 0.9943, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 9.095483661041244e-06, |
|
"loss": 1.0805, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 9.085502144893014e-06, |
|
"loss": 1.005, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"learning_rate": 9.075471396731771e-06, |
|
"loss": 1.1173, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 9.065391537432234e-06, |
|
"loss": 1.0501, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"learning_rate": 9.055262688460931e-06, |
|
"loss": 0.9866, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 9.045084971874738e-06, |
|
"loss": 1.1156, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 1.04, |
|
"learning_rate": 9.034858510319404e-06, |
|
"loss": 1.0603, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"learning_rate": 9.02458342702808e-06, |
|
"loss": 0.9592, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 9.014259845819819e-06, |
|
"loss": 0.9943, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 9.003887891098108e-06, |
|
"loss": 0.8909, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 8.993467687849345e-06, |
|
"loss": 0.9349, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"learning_rate": 8.982999361641344e-06, |
|
"loss": 1.1811, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 8.972483038621827e-06, |
|
"loss": 1.0453, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 1.08, |
|
"learning_rate": 8.961918845516893e-06, |
|
"loss": 0.9181, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 8.951306909629492e-06, |
|
"loss": 1.1174, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"learning_rate": 8.940647358837904e-06, |
|
"loss": 0.752, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 8.92994032159418e-06, |
|
"loss": 0.97, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 8.919185926922608e-06, |
|
"loss": 0.9996, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 8.908384304418148e-06, |
|
"loss": 0.9168, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 8.89753558424488e-06, |
|
"loss": 0.9961, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 1.12, |
|
"learning_rate": 8.88663989713443e-06, |
|
"loss": 1.0325, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 8.87569737438439e-06, |
|
"loss": 0.9891, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"learning_rate": 8.864708147856748e-06, |
|
"loss": 0.7389, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 8.85367234997629e-06, |
|
"loss": 0.9575, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"learning_rate": 8.842590113729001e-06, |
|
"loss": 0.9032, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 8.831461572660476e-06, |
|
"loss": 0.9017, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 1.15, |
|
"learning_rate": 8.820286860874298e-06, |
|
"loss": 0.9411, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 8.809066113030427e-06, |
|
"loss": 0.9945, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 8.797799464343575e-06, |
|
"loss": 1.0241, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 1.17, |
|
"learning_rate": 8.786487050581583e-06, |
|
"loss": 1.0131, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 8.775129008063773e-06, |
|
"loss": 1.0196, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 8.763725473659325e-06, |
|
"loss": 1.0947, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 8.752276584785603e-06, |
|
"loss": 1.0206, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 1.19, |
|
"learning_rate": 8.740782479406515e-06, |
|
"loss": 0.9768, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 8.729243296030851e-06, |
|
"loss": 0.9138, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 8.717659173710603e-06, |
|
"loss": 0.9582, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 8.706030252039302e-06, |
|
"loss": 0.8377, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 1.21, |
|
"learning_rate": 8.694356671150327e-06, |
|
"loss": 0.9705, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 8.682638571715221e-06, |
|
"loss": 0.9781, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"learning_rate": 8.670876094941991e-06, |
|
"loss": 0.9982, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 1.23, |
|
"learning_rate": 8.659069382573413e-06, |
|
"loss": 0.9935, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 8.64721857688532e-06, |
|
"loss": 0.9274, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"learning_rate": 8.635323820684884e-06, |
|
"loss": 0.989, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 8.623385257308906e-06, |
|
"loss": 0.8472, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 8.611403030622074e-06, |
|
"loss": 0.9318, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 8.599377285015243e-06, |
|
"loss": 0.9046, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 8.587308165403686e-06, |
|
"loss": 0.812, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 8.575195817225357e-06, |
|
"loss": 0.9221, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"learning_rate": 8.563040386439123e-06, |
|
"loss": 0.9355, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 8.55084201952302e-06, |
|
"loss": 1.0636, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 1.28, |
|
"learning_rate": 8.538600863472481e-06, |
|
"loss": 0.9307, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"learning_rate": 8.526317065798568e-06, |
|
"loss": 0.9033, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 8.51399077452619e-06, |
|
"loss": 1.0066, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 8.501622138192323e-06, |
|
"loss": 0.9879, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 8.489211305844216e-06, |
|
"loss": 0.9377, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"learning_rate": 8.476758427037607e-06, |
|
"loss": 0.9003, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 8.464263651834894e-06, |
|
"loss": 0.9243, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 1.32, |
|
"learning_rate": 8.451727130803362e-06, |
|
"loss": 0.9654, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 8.439149015013343e-06, |
|
"loss": 0.9913, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"learning_rate": 8.4265294560364e-06, |
|
"loss": 0.7543, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 8.41386860594351e-06, |
|
"loss": 0.9519, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 1.34, |
|
"learning_rate": 8.401166617303221e-06, |
|
"loss": 1.0114, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 8.388423643179822e-06, |
|
"loss": 0.8913, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 8.375639837131488e-06, |
|
"loss": 0.8743, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 1.36, |
|
"learning_rate": 8.362815353208441e-06, |
|
"loss": 0.9434, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 8.349950345951086e-06, |
|
"loss": 1.0303, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"learning_rate": 8.337044970388155e-06, |
|
"loss": 0.9072, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 8.324099382034828e-06, |
|
"loss": 0.9545, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"learning_rate": 8.311113736890873e-06, |
|
"loss": 0.8508, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 8.298088191438753e-06, |
|
"loss": 0.9524, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 1.39, |
|
"learning_rate": 8.285022902641753e-06, |
|
"loss": 1.0097, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 8.271918027942076e-06, |
|
"loss": 0.9031, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 8.258773725258955e-06, |
|
"loss": 0.9991, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 8.245590152986746e-06, |
|
"loss": 0.8694, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 1.41, |
|
"learning_rate": 8.23236746999302e-06, |
|
"loss": 1.0646, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"learning_rate": 8.219105835616645e-06, |
|
"loss": 1.0111, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 8.20580540966588e-06, |
|
"loss": 0.9125, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 8.192466352416425e-06, |
|
"loss": 0.9076, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 8.179088824609515e-06, |
|
"loss": 0.937, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"learning_rate": 8.165672987449962e-06, |
|
"loss": 0.905, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 8.15221900260423e-06, |
|
"loss": 0.9514, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 8.138727032198473e-06, |
|
"loss": 1.1232, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 8.125197238816589e-06, |
|
"loss": 0.9331, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"learning_rate": 8.111629785498256e-06, |
|
"loss": 0.8878, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 8.098024835736977e-06, |
|
"loss": 0.9899, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 1.47, |
|
"learning_rate": 8.084382553478095e-06, |
|
"loss": 0.9031, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"learning_rate": 8.070703103116827e-06, |
|
"loss": 0.973, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 8.056986649496288e-06, |
|
"loss": 0.9401, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 8.043233357905488e-06, |
|
"loss": 0.844, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 8.029443394077356e-06, |
|
"loss": 0.9523, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"learning_rate": 8.015616924186736e-06, |
|
"loss": 0.9861, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 8.001754114848382e-06, |
|
"loss": 1.0519, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"learning_rate": 7.987855133114958e-06, |
|
"loss": 0.8462, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 7.97392014647502e-06, |
|
"loss": 0.8905, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 7.959949322850994e-06, |
|
"loss": 0.9227, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 7.945942830597163e-06, |
|
"loss": 0.9107, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"learning_rate": 7.931900838497626e-06, |
|
"loss": 0.9799, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 7.917823515764272e-06, |
|
"loss": 0.9709, |
|
"step": 283 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 7.903711032034742e-06, |
|
"loss": 0.8537, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"learning_rate": 7.889563557370378e-06, |
|
"loss": 0.9063, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 7.87538126225418e-06, |
|
"loss": 0.9615, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 1.56, |
|
"learning_rate": 7.861164317588746e-06, |
|
"loss": 0.9486, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 7.846912894694216e-06, |
|
"loss": 0.8401, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"learning_rate": 7.83262716530621e-06, |
|
"loss": 1.0363, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 7.818307301573757e-06, |
|
"loss": 1.0342, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 1.58, |
|
"learning_rate": 7.80395347605721e-06, |
|
"loss": 0.9348, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 7.789565861726189e-06, |
|
"loss": 0.8611, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 7.77514463195747e-06, |
|
"loss": 0.989, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 7.76068996053292e-06, |
|
"loss": 0.955, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 7.746202021637385e-06, |
|
"loss": 0.9545, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"learning_rate": 7.7316809898566e-06, |
|
"loss": 0.9366, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 7.717127040175084e-06, |
|
"loss": 0.8091, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"learning_rate": 7.70254034797403e-06, |
|
"loss": 0.9747, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 7.687921089029194e-06, |
|
"loss": 1.0307, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 1.63, |
|
"learning_rate": 7.67326943950877e-06, |
|
"loss": 1.0158, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 7.658585575971278e-06, |
|
"loss": 0.8534, |
|
"step": 301 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 7.643869675363425e-06, |
|
"loss": 0.8246, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 7.62912191501798e-06, |
|
"loss": 0.8728, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 1.65, |
|
"learning_rate": 7.614342472651639e-06, |
|
"loss": 0.9968, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"learning_rate": 7.599531526362873e-06, |
|
"loss": 0.9028, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 7.5846892546297925e-06, |
|
"loss": 0.8097, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 7.569815836307994e-06, |
|
"loss": 0.9532, |
|
"step": 307 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 7.5549114506283995e-06, |
|
"loss": 1.0243, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"learning_rate": 7.539976277195102e-06, |
|
"loss": 0.9622, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 7.525010495983202e-06, |
|
"loss": 0.9323, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 7.510014287336638e-06, |
|
"loss": 0.9608, |
|
"step": 311 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 7.494987831966003e-06, |
|
"loss": 0.9235, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"learning_rate": 7.4799313109463844e-06, |
|
"loss": 0.9056, |
|
"step": 313 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 7.464844905715169e-06, |
|
"loss": 0.988, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 1.71, |
|
"learning_rate": 7.449728798069864e-06, |
|
"loss": 0.9618, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"learning_rate": 7.434583170165903e-06, |
|
"loss": 0.9671, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 7.419408204514445e-06, |
|
"loss": 0.9739, |
|
"step": 317 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 7.40420408398019e-06, |
|
"loss": 0.8321, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 7.388970991779159e-06, |
|
"loss": 0.9047, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"learning_rate": 7.373709111476498e-06, |
|
"loss": 1.0659, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 7.358418626984262e-06, |
|
"loss": 0.899, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"learning_rate": 7.343099722559195e-06, |
|
"loss": 0.9948, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 7.327752582800518e-06, |
|
"loss": 0.9291, |
|
"step": 323 |
|
}, |
|
{ |
|
"epoch": 1.76, |
|
"learning_rate": 7.312377392647694e-06, |
|
"loss": 0.9076, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 7.296974337378209e-06, |
|
"loss": 0.829, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 7.281543602605333e-06, |
|
"loss": 0.9306, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 7.266085374275891e-06, |
|
"loss": 0.9055, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 7.250599838668007e-06, |
|
"loss": 0.8165, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"learning_rate": 7.235087182388877e-06, |
|
"loss": 0.939, |
|
"step": 329 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 7.219547592372512e-06, |
|
"loss": 0.9337, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 1.8, |
|
"learning_rate": 7.203981255877482e-06, |
|
"loss": 0.8893, |
|
"step": 331 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 7.188388360484667e-06, |
|
"loss": 0.9036, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"learning_rate": 7.17276909409499e-06, |
|
"loss": 0.9001, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 7.1571236449271575e-06, |
|
"loss": 0.8759, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 1.82, |
|
"learning_rate": 7.141452201515386e-06, |
|
"loss": 0.9875, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 7.125754952707143e-06, |
|
"loss": 0.8603, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 7.110032087660846e-06, |
|
"loss": 0.8247, |
|
"step": 337 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 7.094283795843616e-06, |
|
"loss": 0.8525, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 1.84, |
|
"learning_rate": 7.0785102670289644e-06, |
|
"loss": 0.8645, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"learning_rate": 7.062711691294525e-06, |
|
"loss": 0.8738, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 7.046888259019757e-06, |
|
"loss": 1.0695, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 7.031040160883651e-06, |
|
"loss": 0.8732, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 7.015167587862431e-06, |
|
"loss": 0.892, |
|
"step": 343 |
|
}, |
|
{ |
|
"epoch": 1.87, |
|
"learning_rate": 6.999270731227252e-06, |
|
"loss": 0.8494, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 6.983349782541901e-06, |
|
"loss": 0.9128, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 6.96740493366048e-06, |
|
"loss": 0.8998, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 6.9514363767251024e-06, |
|
"loss": 0.9055, |
|
"step": 347 |
|
}, |
|
{ |
|
"epoch": 1.89, |
|
"learning_rate": 6.9354443041635675e-06, |
|
"loss": 0.8089, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 6.919428908687057e-06, |
|
"loss": 0.891, |
|
"step": 349 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"learning_rate": 6.903390383287795e-06, |
|
"loss": 1.0194, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 1.91, |
|
"learning_rate": 6.887328921236733e-06, |
|
"loss": 0.9762, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 6.871244716081223e-06, |
|
"loss": 0.943, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"learning_rate": 6.855137961642676e-06, |
|
"loss": 0.7557, |
|
"step": 353 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 6.839008852014233e-06, |
|
"loss": 0.8784, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 6.822857581558423e-06, |
|
"loss": 0.8028, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 6.806684344904828e-06, |
|
"loss": 0.7966, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 6.790489336947723e-06, |
|
"loss": 0.9224, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 6.774272752843745e-06, |
|
"loss": 0.7707, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 1.95, |
|
"learning_rate": 6.758034788009523e-06, |
|
"loss": 0.8538, |
|
"step": 359 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 6.741775638119345e-06, |
|
"loss": 0.7929, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"learning_rate": 6.7254954991027765e-06, |
|
"loss": 0.8583, |
|
"step": 361 |
|
}, |
|
{ |
|
"epoch": 1.97, |
|
"learning_rate": 6.709194567142316e-06, |
|
"loss": 0.9552, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 6.6928730386710285e-06, |
|
"loss": 0.8742, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 6.676531110370168e-06, |
|
"loss": 0.8376, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 6.66016897916682e-06, |
|
"loss": 0.8969, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"learning_rate": 6.643786842231523e-06, |
|
"loss": 0.8108, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 6.627384896975896e-06, |
|
"loss": 0.8374, |
|
"step": 367 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"learning_rate": 6.6109633410502475e-06, |
|
"loss": 0.8357, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 6.59452237234121e-06, |
|
"loss": 0.9496, |
|
"step": 369 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"learning_rate": 6.57806218896935e-06, |
|
"loss": 1.0033, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 6.5615829892867715e-06, |
|
"loss": 0.8199, |
|
"step": 371 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 6.545084971874738e-06, |
|
"loss": 0.9517, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"learning_rate": 6.528568335541274e-06, |
|
"loss": 0.9681, |
|
"step": 373 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 6.512033279318768e-06, |
|
"loss": 0.8775, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 2.04, |
|
"learning_rate": 6.495480002461577e-06, |
|
"loss": 0.8583, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 6.478908704443629e-06, |
|
"loss": 0.8624, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"learning_rate": 6.462319584956e-06, |
|
"loss": 0.7446, |
|
"step": 377 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 6.445712843904542e-06, |
|
"loss": 0.7195, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 2.06, |
|
"learning_rate": 6.4290886814074405e-06, |
|
"loss": 0.9782, |
|
"step": 379 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 6.412447297792818e-06, |
|
"loss": 0.8978, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 6.3957888935963265e-06, |
|
"loss": 0.881, |
|
"step": 381 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 6.379113669558713e-06, |
|
"loss": 0.9469, |
|
"step": 382 |
|
}, |
|
{ |
|
"epoch": 2.08, |
|
"learning_rate": 6.3624218266234176e-06, |
|
"loss": 0.8274, |
|
"step": 383 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"learning_rate": 6.345713565934142e-06, |
|
"loss": 0.8372, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 6.328989088832431e-06, |
|
"loss": 1.0378, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"learning_rate": 6.312248596855241e-06, |
|
"loss": 0.8135, |
|
"step": 386 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 6.295492291732519e-06, |
|
"loss": 0.8383, |
|
"step": 387 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 6.278720375384759e-06, |
|
"loss": 0.8324, |
|
"step": 388 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 6.261933049920587e-06, |
|
"loss": 0.8466, |
|
"step": 389 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 6.245130517634307e-06, |
|
"loss": 0.8675, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 6.228312981003476e-06, |
|
"loss": 0.9011, |
|
"step": 391 |
|
}, |
|
{ |
|
"epoch": 2.13, |
|
"learning_rate": 6.211480642686458e-06, |
|
"loss": 0.8772, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 6.194633705519983e-06, |
|
"loss": 1.0046, |
|
"step": 393 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"learning_rate": 6.177772372516706e-06, |
|
"loss": 0.7755, |
|
"step": 394 |
|
}, |
|
{ |
|
"epoch": 2.15, |
|
"learning_rate": 6.160896846862754e-06, |
|
"loss": 0.8375, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 6.1440073319152856e-06, |
|
"loss": 0.8476, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"learning_rate": 6.127104031200032e-06, |
|
"loss": 0.9187, |
|
"step": 397 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 6.110187148408852e-06, |
|
"loss": 0.9529, |
|
"step": 398 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 6.093256887397272e-06, |
|
"loss": 0.9052, |
|
"step": 399 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 6.076313452182033e-06, |
|
"loss": 0.8845, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"learning_rate": 6.05935704693863e-06, |
|
"loss": 0.806, |
|
"step": 401 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 6.042387875998852e-06, |
|
"loss": 0.9245, |
|
"step": 402 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 6.02540614384832e-06, |
|
"loss": 0.9109, |
|
"step": 403 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 6.008412055124024e-06, |
|
"loss": 0.8871, |
|
"step": 404 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"learning_rate": 5.991405814611855e-06, |
|
"loss": 0.9606, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 5.974387627244137e-06, |
|
"loss": 0.9259, |
|
"step": 406 |
|
}, |
|
{ |
|
"epoch": 2.21, |
|
"learning_rate": 5.957357698097163e-06, |
|
"loss": 0.9331, |
|
"step": 407 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 5.940316232388711e-06, |
|
"loss": 0.8568, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 5.923263435475589e-06, |
|
"loss": 0.9379, |
|
"step": 409 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"learning_rate": 5.9061995128511455e-06, |
|
"loss": 0.8497, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 5.889124670142797e-06, |
|
"loss": 0.8845, |
|
"step": 411 |
|
}, |
|
{ |
|
"epoch": 2.24, |
|
"learning_rate": 5.872039113109557e-06, |
|
"loss": 0.9423, |
|
"step": 412 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 5.854943047639547e-06, |
|
"loss": 0.9536, |
|
"step": 413 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"learning_rate": 5.8378366797475184e-06, |
|
"loss": 0.8411, |
|
"step": 414 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 5.820720215572375e-06, |
|
"loss": 0.8476, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 5.803593861374687e-06, |
|
"loss": 0.922, |
|
"step": 416 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 5.786457823534193e-06, |
|
"loss": 0.9568, |
|
"step": 417 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"learning_rate": 5.769312308547334e-06, |
|
"loss": 0.7422, |
|
"step": 418 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 5.752157523024753e-06, |
|
"loss": 0.8765, |
|
"step": 419 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 5.734993673688801e-06, |
|
"loss": 0.9319, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"learning_rate": 5.71782096737106e-06, |
|
"loss": 0.7123, |
|
"step": 421 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 5.7006396110098306e-06, |
|
"loss": 0.8822, |
|
"step": 422 |
|
}, |
|
{ |
|
"epoch": 2.3, |
|
"learning_rate": 5.683449811647664e-06, |
|
"loss": 0.8585, |
|
"step": 423 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 5.666251776428844e-06, |
|
"loss": 0.8678, |
|
"step": 424 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 5.6490457125969035e-06, |
|
"loss": 0.7982, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 5.631831827492121e-06, |
|
"loss": 0.8457, |
|
"step": 426 |
|
}, |
|
{ |
|
"epoch": 2.32, |
|
"learning_rate": 5.614610328549029e-06, |
|
"loss": 0.732, |
|
"step": 427 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 5.5973814232939e-06, |
|
"loss": 0.8866, |
|
"step": 428 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"learning_rate": 5.5801453193422715e-06, |
|
"loss": 0.8772, |
|
"step": 429 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"learning_rate": 5.562902224396416e-06, |
|
"loss": 1.0722, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 5.54565234624285e-06, |
|
"loss": 0.8542, |
|
"step": 431 |
|
}, |
|
{ |
|
"epoch": 2.35, |
|
"learning_rate": 5.52839589274984e-06, |
|
"loss": 0.7729, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 5.511133071864881e-06, |
|
"loss": 0.9159, |
|
"step": 433 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 5.493864091612197e-06, |
|
"loss": 0.9416, |
|
"step": 434 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 5.476589160090238e-06, |
|
"loss": 0.9376, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 2.37, |
|
"learning_rate": 5.459308485469171e-06, |
|
"loss": 0.9561, |
|
"step": 436 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 5.442022275988365e-06, |
|
"loss": 0.881, |
|
"step": 437 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"learning_rate": 5.4247307399538876e-06, |
|
"loss": 0.9091, |
|
"step": 438 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 5.407434085735997e-06, |
|
"loss": 0.847, |
|
"step": 439 |
|
}, |
|
{ |
|
"epoch": 2.39, |
|
"learning_rate": 5.390132521766626e-06, |
|
"loss": 0.9168, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"learning_rate": 5.372826256536867e-06, |
|
"loss": 0.8412, |
|
"step": 441 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 5.355515498594472e-06, |
|
"loss": 0.8875, |
|
"step": 442 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 5.338200456541329e-06, |
|
"loss": 0.8385, |
|
"step": 443 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 5.32088133903095e-06, |
|
"loss": 0.8067, |
|
"step": 444 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"learning_rate": 5.30355835476596e-06, |
|
"loss": 0.8311, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 5.286231712495578e-06, |
|
"loss": 0.9589, |
|
"step": 446 |
|
}, |
|
{ |
|
"epoch": 2.43, |
|
"learning_rate": 5.2689016210131065e-06, |
|
"loss": 0.8565, |
|
"step": 447 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 5.251568289153407e-06, |
|
"loss": 0.862, |
|
"step": 448 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"learning_rate": 5.234231925790396e-06, |
|
"loss": 0.7935, |
|
"step": 449 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 5.216892739834519e-06, |
|
"loss": 0.9223, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 5.199550940230228e-06, |
|
"loss": 0.8139, |
|
"step": 451 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 5.182206735953479e-06, |
|
"loss": 0.926, |
|
"step": 452 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 5.1648603360092034e-06, |
|
"loss": 0.9354, |
|
"step": 453 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"learning_rate": 5.147511949428787e-06, |
|
"loss": 0.9393, |
|
"step": 454 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 5.13016178526756e-06, |
|
"loss": 0.9063, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 2.48, |
|
"learning_rate": 5.112810052602274e-06, |
|
"loss": 0.8157, |
|
"step": 456 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 5.095456960528574e-06, |
|
"loss": 0.7404, |
|
"step": 457 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"learning_rate": 5.078102718158497e-06, |
|
"loss": 0.8, |
|
"step": 458 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 5.060747534617937e-06, |
|
"loss": 0.8425, |
|
"step": 459 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"learning_rate": 5.043391619044122e-06, |
|
"loss": 0.9117, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 5.026035180583116e-06, |
|
"loss": 0.9421, |
|
"step": 461 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 5.008678428387273e-06, |
|
"loss": 0.8304, |
|
"step": 462 |
|
}, |
|
{ |
|
"epoch": 2.52, |
|
"learning_rate": 4.99132157161273e-06, |
|
"loss": 0.9645, |
|
"step": 463 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 4.973964819416885e-06, |
|
"loss": 0.7235, |
|
"step": 464 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 4.956608380955877e-06, |
|
"loss": 0.9902, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 4.939252465382066e-06, |
|
"loss": 0.892, |
|
"step": 466 |
|
}, |
|
{ |
|
"epoch": 2.54, |
|
"learning_rate": 4.9218972818415035e-06, |
|
"loss": 0.7766, |
|
"step": 467 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 4.904543039471427e-06, |
|
"loss": 1.0028, |
|
"step": 468 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 4.887189947397728e-06, |
|
"loss": 0.8625, |
|
"step": 469 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 4.869838214732441e-06, |
|
"loss": 1.0353, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 2.56, |
|
"learning_rate": 4.852488050571214e-06, |
|
"loss": 0.8195, |
|
"step": 471 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 4.8351396639908e-06, |
|
"loss": 0.7516, |
|
"step": 472 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"learning_rate": 4.8177932640465216e-06, |
|
"loss": 0.9172, |
|
"step": 473 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"learning_rate": 4.800449059769774e-06, |
|
"loss": 0.9456, |
|
"step": 474 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 4.783107260165483e-06, |
|
"loss": 1.0406, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 2.59, |
|
"learning_rate": 4.7657680742096044e-06, |
|
"loss": 1.0074, |
|
"step": 476 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 4.748431710846594e-06, |
|
"loss": 0.6638, |
|
"step": 477 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 4.731098378986897e-06, |
|
"loss": 0.9923, |
|
"step": 478 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 4.7137682875044246e-06, |
|
"loss": 0.8311, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 4.696441645234042e-06, |
|
"loss": 0.757, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 4.679118660969051e-06, |
|
"loss": 0.8805, |
|
"step": 481 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"learning_rate": 4.6617995434586714e-06, |
|
"loss": 0.9096, |
|
"step": 482 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 4.64448450140553e-06, |
|
"loss": 0.9217, |
|
"step": 483 |
|
}, |
|
{ |
|
"epoch": 2.63, |
|
"learning_rate": 4.627173743463134e-06, |
|
"loss": 0.8868, |
|
"step": 484 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 4.609867478233377e-06, |
|
"loss": 0.8931, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"learning_rate": 4.592565914264004e-06, |
|
"loss": 0.8445, |
|
"step": 486 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 4.575269260046112e-06, |
|
"loss": 0.8162, |
|
"step": 487 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 4.557977724011636e-06, |
|
"loss": 0.8142, |
|
"step": 488 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"learning_rate": 4.540691514530831e-06, |
|
"loss": 0.8531, |
|
"step": 489 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 4.523410839909764e-06, |
|
"loss": 0.9631, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 2.67, |
|
"learning_rate": 4.506135908387805e-06, |
|
"loss": 0.7765, |
|
"step": 491 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 4.488866928135121e-06, |
|
"loss": 0.9362, |
|
"step": 492 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"learning_rate": 4.4716041072501604e-06, |
|
"loss": 0.8497, |
|
"step": 493 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 4.45434765375715e-06, |
|
"loss": 0.8659, |
|
"step": 494 |
|
}, |
|
{ |
|
"epoch": 2.69, |
|
"learning_rate": 4.437097775603587e-06, |
|
"loss": 0.9575, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 4.41985468065773e-06, |
|
"loss": 0.8198, |
|
"step": 496 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 4.402618576706101e-06, |
|
"loss": 0.8925, |
|
"step": 497 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"learning_rate": 4.385389671450974e-06, |
|
"loss": 0.8702, |
|
"step": 498 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 4.36816817250788e-06, |
|
"loss": 0.7818, |
|
"step": 499 |
|
}, |
|
{ |
|
"epoch": 2.72, |
|
"learning_rate": 4.350954287403099e-06, |
|
"loss": 0.7461, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 4.333748223571158e-06, |
|
"loss": 0.8872, |
|
"step": 501 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"learning_rate": 4.316550188352337e-06, |
|
"loss": 0.9121, |
|
"step": 502 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 4.299360388990171e-06, |
|
"loss": 0.7544, |
|
"step": 503 |
|
}, |
|
{ |
|
"epoch": 2.74, |
|
"learning_rate": 4.282179032628943e-06, |
|
"loss": 0.7822, |
|
"step": 504 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 4.265006326311199e-06, |
|
"loss": 0.9713, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 4.247842476975249e-06, |
|
"loss": 0.8752, |
|
"step": 506 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 4.230687691452667e-06, |
|
"loss": 0.97, |
|
"step": 507 |
|
}, |
|
{ |
|
"epoch": 2.76, |
|
"learning_rate": 4.213542176465809e-06, |
|
"loss": 0.8821, |
|
"step": 508 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"learning_rate": 4.196406138625315e-06, |
|
"loss": 0.8736, |
|
"step": 509 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 4.179279784427625e-06, |
|
"loss": 0.7504, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 4.1621633202524815e-06, |
|
"loss": 0.7806, |
|
"step": 511 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 4.145056952360456e-06, |
|
"loss": 0.7688, |
|
"step": 512 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 4.127960886890445e-06, |
|
"loss": 0.8505, |
|
"step": 513 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 4.110875329857205e-06, |
|
"loss": 0.8255, |
|
"step": 514 |
|
}, |
|
{ |
|
"epoch": 2.8, |
|
"learning_rate": 4.093800487148857e-06, |
|
"loss": 0.8624, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 4.076736564524412e-06, |
|
"loss": 0.9774, |
|
"step": 516 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"learning_rate": 4.0596837676112905e-06, |
|
"loss": 0.734, |
|
"step": 517 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 4.04264230190284e-06, |
|
"loss": 0.8445, |
|
"step": 518 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"learning_rate": 4.025612372755865e-06, |
|
"loss": 0.8627, |
|
"step": 519 |
|
}, |
|
{ |
|
"epoch": 2.83, |
|
"learning_rate": 4.008594185388146e-06, |
|
"loss": 0.9884, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 3.991587944875977e-06, |
|
"loss": 0.8007, |
|
"step": 521 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 3.97459385615168e-06, |
|
"loss": 0.8553, |
|
"step": 522 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 3.957612124001151e-06, |
|
"loss": 0.8161, |
|
"step": 523 |
|
}, |
|
{ |
|
"epoch": 2.85, |
|
"learning_rate": 3.940642953061371e-06, |
|
"loss": 0.9705, |
|
"step": 524 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 3.9236865478179685e-06, |
|
"loss": 0.8085, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"learning_rate": 3.906743112602729e-06, |
|
"loss": 0.8702, |
|
"step": 526 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 3.889812851591149e-06, |
|
"loss": 0.8137, |
|
"step": 527 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 3.872895968799969e-06, |
|
"loss": 0.7632, |
|
"step": 528 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 3.855992668084716e-06, |
|
"loss": 0.8425, |
|
"step": 529 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"learning_rate": 3.839103153137247e-06, |
|
"loss": 0.842, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 3.822227627483295e-06, |
|
"loss": 0.8439, |
|
"step": 531 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 3.8053662944800177e-06, |
|
"loss": 0.8059, |
|
"step": 532 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"learning_rate": 3.7885193573135437e-06, |
|
"loss": 0.8929, |
|
"step": 533 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 3.771687018996525e-06, |
|
"loss": 0.8088, |
|
"step": 534 |
|
}, |
|
{ |
|
"epoch": 2.91, |
|
"learning_rate": 3.7548694823656945e-06, |
|
"loss": 0.8736, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"learning_rate": 3.7380669500794153e-06, |
|
"loss": 0.886, |
|
"step": 536 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"learning_rate": 3.721279624615243e-06, |
|
"loss": 0.8559, |
|
"step": 537 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 3.704507708267483e-06, |
|
"loss": 0.9221, |
|
"step": 538 |
|
}, |
|
{ |
|
"epoch": 2.93, |
|
"learning_rate": 3.6877514031447597e-06, |
|
"loss": 0.8069, |
|
"step": 539 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 3.671010911167572e-06, |
|
"loss": 0.9423, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 3.6542864340658602e-06, |
|
"loss": 0.9176, |
|
"step": 541 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 3.637578173376584e-06, |
|
"loss": 0.8804, |
|
"step": 542 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 3.620886330441289e-06, |
|
"loss": 0.8707, |
|
"step": 543 |
|
}, |
|
{ |
|
"epoch": 2.96, |
|
"learning_rate": 3.6042111064036756e-06, |
|
"loss": 0.7642, |
|
"step": 544 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 3.5875527022071808e-06, |
|
"loss": 1.0431, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"learning_rate": 3.5709113185925615e-06, |
|
"loss": 0.9163, |
|
"step": 546 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"learning_rate": 3.554287156095459e-06, |
|
"loss": 0.8734, |
|
"step": 547 |
|
}, |
|
{ |
|
"epoch": 2.98, |
|
"learning_rate": 3.5376804150440002e-06, |
|
"loss": 0.8863, |
|
"step": 548 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 3.5210912955563735e-06, |
|
"loss": 0.7968, |
|
"step": 549 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 3.5045199975384225e-06, |
|
"loss": 0.876, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 3.4879667206812316e-06, |
|
"loss": 0.8863, |
|
"step": 551 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"learning_rate": 3.471431664458729e-06, |
|
"loss": 0.8485, |
|
"step": 552 |
|
}, |
|
{ |
|
"epoch": 3.01, |
|
"learning_rate": 3.4549150281252635e-06, |
|
"loss": 0.8416, |
|
"step": 553 |
|
}, |
|
{ |
|
"epoch": 3.01, |
|
"learning_rate": 3.43841701071323e-06, |
|
"loss": 0.8251, |
|
"step": 554 |
|
}, |
|
{ |
|
"epoch": 3.02, |
|
"learning_rate": 3.4219378110306523e-06, |
|
"loss": 0.9206, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"learning_rate": 3.4054776276587897e-06, |
|
"loss": 0.8879, |
|
"step": 556 |
|
}, |
|
{ |
|
"epoch": 3.03, |
|
"learning_rate": 3.3890366589497538e-06, |
|
"loss": 1.0349, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 3.04, |
|
"learning_rate": 3.3726151030241074e-06, |
|
"loss": 0.8465, |
|
"step": 558 |
|
}, |
|
{ |
|
"epoch": 3.04, |
|
"learning_rate": 3.3562131577684778e-06, |
|
"loss": 0.8563, |
|
"step": 559 |
|
}, |
|
{ |
|
"epoch": 3.05, |
|
"learning_rate": 3.3398310208331806e-06, |
|
"loss": 0.842, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 3.05, |
|
"learning_rate": 3.323468889629834e-06, |
|
"loss": 0.8472, |
|
"step": 561 |
|
}, |
|
{ |
|
"epoch": 3.06, |
|
"learning_rate": 3.3071269613289735e-06, |
|
"loss": 0.8492, |
|
"step": 562 |
|
}, |
|
{ |
|
"epoch": 3.06, |
|
"learning_rate": 3.2908054328576854e-06, |
|
"loss": 0.8731, |
|
"step": 563 |
|
}, |
|
{ |
|
"epoch": 3.07, |
|
"learning_rate": 3.2745045008972255e-06, |
|
"loss": 0.7911, |
|
"step": 564 |
|
}, |
|
{ |
|
"epoch": 3.07, |
|
"learning_rate": 3.2582243618806574e-06, |
|
"loss": 1.068, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 3.08, |
|
"learning_rate": 3.2419652119904777e-06, |
|
"loss": 0.8996, |
|
"step": 566 |
|
}, |
|
{ |
|
"epoch": 3.09, |
|
"learning_rate": 3.2257272471562574e-06, |
|
"loss": 0.9333, |
|
"step": 567 |
|
}, |
|
{ |
|
"epoch": 3.09, |
|
"learning_rate": 3.2095106630522778e-06, |
|
"loss": 0.8469, |
|
"step": 568 |
|
}, |
|
{ |
|
"epoch": 3.1, |
|
"learning_rate": 3.193315655095175e-06, |
|
"loss": 0.809, |
|
"step": 569 |
|
}, |
|
{ |
|
"epoch": 3.1, |
|
"learning_rate": 3.177142418441578e-06, |
|
"loss": 0.7341, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 3.11, |
|
"learning_rate": 3.160991147985769e-06, |
|
"loss": 0.7984, |
|
"step": 571 |
|
}, |
|
{ |
|
"epoch": 3.11, |
|
"learning_rate": 3.144862038357326e-06, |
|
"loss": 0.8389, |
|
"step": 572 |
|
}, |
|
{ |
|
"epoch": 3.12, |
|
"learning_rate": 3.1287552839187784e-06, |
|
"loss": 0.9403, |
|
"step": 573 |
|
}, |
|
{ |
|
"epoch": 3.12, |
|
"learning_rate": 3.1126710787632662e-06, |
|
"loss": 0.8135, |
|
"step": 574 |
|
}, |
|
{ |
|
"epoch": 3.13, |
|
"learning_rate": 3.096609616712207e-06, |
|
"loss": 0.7741, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 3.13, |
|
"learning_rate": 3.080571091312945e-06, |
|
"loss": 0.835, |
|
"step": 576 |
|
}, |
|
{ |
|
"epoch": 3.14, |
|
"learning_rate": 3.064555695836433e-06, |
|
"loss": 0.9042, |
|
"step": 577 |
|
}, |
|
{ |
|
"epoch": 3.15, |
|
"learning_rate": 3.0485636232748992e-06, |
|
"loss": 0.8505, |
|
"step": 578 |
|
}, |
|
{ |
|
"epoch": 3.15, |
|
"learning_rate": 3.032595066339521e-06, |
|
"loss": 0.8271, |
|
"step": 579 |
|
}, |
|
{ |
|
"epoch": 3.16, |
|
"learning_rate": 3.0166502174581012e-06, |
|
"loss": 0.7945, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 3.16, |
|
"learning_rate": 3.00072926877275e-06, |
|
"loss": 0.8788, |
|
"step": 581 |
|
}, |
|
{ |
|
"epoch": 3.17, |
|
"learning_rate": 2.9848324121375705e-06, |
|
"loss": 0.8791, |
|
"step": 582 |
|
}, |
|
{ |
|
"epoch": 3.17, |
|
"learning_rate": 2.9689598391163497e-06, |
|
"loss": 0.788, |
|
"step": 583 |
|
}, |
|
{ |
|
"epoch": 3.18, |
|
"learning_rate": 2.9531117409802432e-06, |
|
"loss": 0.8701, |
|
"step": 584 |
|
}, |
|
{ |
|
"epoch": 3.18, |
|
"learning_rate": 2.937288308705475e-06, |
|
"loss": 0.8232, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 3.19, |
|
"learning_rate": 2.921489732971038e-06, |
|
"loss": 0.8906, |
|
"step": 586 |
|
}, |
|
{ |
|
"epoch": 3.19, |
|
"learning_rate": 2.9057162041563867e-06, |
|
"loss": 0.8842, |
|
"step": 587 |
|
}, |
|
{ |
|
"epoch": 3.2, |
|
"learning_rate": 2.8899679123391545e-06, |
|
"loss": 0.8328, |
|
"step": 588 |
|
}, |
|
{ |
|
"epoch": 3.21, |
|
"learning_rate": 2.8742450472928595e-06, |
|
"loss": 0.773, |
|
"step": 589 |
|
}, |
|
{ |
|
"epoch": 3.21, |
|
"learning_rate": 2.858547798484613e-06, |
|
"loss": 0.8197, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 3.22, |
|
"learning_rate": 2.8428763550728446e-06, |
|
"loss": 0.8593, |
|
"step": 591 |
|
}, |
|
{ |
|
"epoch": 3.22, |
|
"learning_rate": 2.8272309059050107e-06, |
|
"loss": 0.7967, |
|
"step": 592 |
|
}, |
|
{ |
|
"epoch": 3.23, |
|
"learning_rate": 2.8116116395153343e-06, |
|
"loss": 0.9606, |
|
"step": 593 |
|
}, |
|
{ |
|
"epoch": 3.23, |
|
"learning_rate": 2.7960187441225185e-06, |
|
"loss": 0.9348, |
|
"step": 594 |
|
}, |
|
{ |
|
"epoch": 3.24, |
|
"learning_rate": 2.7804524076274898e-06, |
|
"loss": 0.8325, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 3.24, |
|
"learning_rate": 2.764912817611124e-06, |
|
"loss": 0.8465, |
|
"step": 596 |
|
}, |
|
{ |
|
"epoch": 3.25, |
|
"learning_rate": 2.7494001613319932e-06, |
|
"loss": 0.8464, |
|
"step": 597 |
|
}, |
|
{ |
|
"epoch": 3.25, |
|
"learning_rate": 2.733914625724112e-06, |
|
"loss": 0.969, |
|
"step": 598 |
|
}, |
|
{ |
|
"epoch": 3.26, |
|
"learning_rate": 2.7184563973946687e-06, |
|
"loss": 0.9087, |
|
"step": 599 |
|
}, |
|
{ |
|
"epoch": 3.27, |
|
"learning_rate": 2.7030256626217932e-06, |
|
"loss": 0.9559, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 3.27, |
|
"learning_rate": 2.687622607352307e-06, |
|
"loss": 0.934, |
|
"step": 601 |
|
}, |
|
{ |
|
"epoch": 3.28, |
|
"learning_rate": 2.6722474171994826e-06, |
|
"loss": 0.9529, |
|
"step": 602 |
|
}, |
|
{ |
|
"epoch": 3.28, |
|
"learning_rate": 2.656900277440807e-06, |
|
"loss": 0.9131, |
|
"step": 603 |
|
}, |
|
{ |
|
"epoch": 3.29, |
|
"learning_rate": 2.6415813730157408e-06, |
|
"loss": 0.9768, |
|
"step": 604 |
|
}, |
|
{ |
|
"epoch": 3.29, |
|
"learning_rate": 2.6262908885235046e-06, |
|
"loss": 0.9021, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 3.3, |
|
"learning_rate": 2.611029008220842e-06, |
|
"loss": 0.7654, |
|
"step": 606 |
|
}, |
|
{ |
|
"epoch": 3.3, |
|
"learning_rate": 2.5957959160198115e-06, |
|
"loss": 0.8602, |
|
"step": 607 |
|
}, |
|
{ |
|
"epoch": 3.31, |
|
"learning_rate": 2.580591795485555e-06, |
|
"loss": 0.7933, |
|
"step": 608 |
|
}, |
|
{ |
|
"epoch": 3.31, |
|
"learning_rate": 2.565416829834101e-06, |
|
"loss": 0.9777, |
|
"step": 609 |
|
}, |
|
{ |
|
"epoch": 3.32, |
|
"learning_rate": 2.550271201930136e-06, |
|
"loss": 0.7869, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 3.33, |
|
"learning_rate": 2.535155094284832e-06, |
|
"loss": 0.738, |
|
"step": 611 |
|
}, |
|
{ |
|
"epoch": 3.33, |
|
"learning_rate": 2.5200686890536177e-06, |
|
"loss": 0.9764, |
|
"step": 612 |
|
}, |
|
{ |
|
"epoch": 3.34, |
|
"learning_rate": 2.505012168033999e-06, |
|
"loss": 0.8893, |
|
"step": 613 |
|
}, |
|
{ |
|
"epoch": 3.34, |
|
"learning_rate": 2.489985712663364e-06, |
|
"loss": 0.9448, |
|
"step": 614 |
|
}, |
|
{ |
|
"epoch": 3.35, |
|
"learning_rate": 2.474989504016798e-06, |
|
"loss": 0.8546, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 3.35, |
|
"learning_rate": 2.460023722804899e-06, |
|
"loss": 0.9707, |
|
"step": 616 |
|
}, |
|
{ |
|
"epoch": 3.36, |
|
"learning_rate": 2.4450885493716026e-06, |
|
"loss": 0.7581, |
|
"step": 617 |
|
}, |
|
{ |
|
"epoch": 3.36, |
|
"learning_rate": 2.430184163692008e-06, |
|
"loss": 0.8602, |
|
"step": 618 |
|
}, |
|
{ |
|
"epoch": 3.37, |
|
"learning_rate": 2.4153107453702075e-06, |
|
"loss": 0.8751, |
|
"step": 619 |
|
}, |
|
{ |
|
"epoch": 3.37, |
|
"learning_rate": 2.4004684736371276e-06, |
|
"loss": 0.8511, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 3.38, |
|
"learning_rate": 2.385657527348364e-06, |
|
"loss": 0.9102, |
|
"step": 621 |
|
}, |
|
{ |
|
"epoch": 3.39, |
|
"learning_rate": 2.3708780849820218e-06, |
|
"loss": 0.7458, |
|
"step": 622 |
|
}, |
|
{ |
|
"epoch": 3.39, |
|
"learning_rate": 2.356130324636578e-06, |
|
"loss": 0.8914, |
|
"step": 623 |
|
}, |
|
{ |
|
"epoch": 3.4, |
|
"learning_rate": 2.341414424028723e-06, |
|
"loss": 0.8849, |
|
"step": 624 |
|
}, |
|
{ |
|
"epoch": 3.4, |
|
"learning_rate": 2.32673056049123e-06, |
|
"loss": 0.8613, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 3.41, |
|
"learning_rate": 2.312078910970808e-06, |
|
"loss": 0.8999, |
|
"step": 626 |
|
}, |
|
{ |
|
"epoch": 3.41, |
|
"learning_rate": 2.297459652025972e-06, |
|
"loss": 0.7714, |
|
"step": 627 |
|
}, |
|
{ |
|
"epoch": 3.42, |
|
"learning_rate": 2.2828729598249165e-06, |
|
"loss": 0.7473, |
|
"step": 628 |
|
}, |
|
{ |
|
"epoch": 3.42, |
|
"learning_rate": 2.2683190101434015e-06, |
|
"loss": 0.9195, |
|
"step": 629 |
|
}, |
|
{ |
|
"epoch": 3.43, |
|
"learning_rate": 2.253797978362617e-06, |
|
"loss": 0.8738, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 3.43, |
|
"learning_rate": 2.239310039467082e-06, |
|
"loss": 0.8235, |
|
"step": 631 |
|
}, |
|
{ |
|
"epoch": 3.44, |
|
"learning_rate": 2.2248553680425308e-06, |
|
"loss": 0.8386, |
|
"step": 632 |
|
}, |
|
{ |
|
"epoch": 3.44, |
|
"learning_rate": 2.2104341382738127e-06, |
|
"loss": 1.005, |
|
"step": 633 |
|
}, |
|
{ |
|
"epoch": 3.45, |
|
"learning_rate": 2.19604652394279e-06, |
|
"loss": 0.9029, |
|
"step": 634 |
|
}, |
|
{ |
|
"epoch": 3.46, |
|
"learning_rate": 2.1816926984262454e-06, |
|
"loss": 0.8638, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 3.46, |
|
"learning_rate": 2.1673728346937904e-06, |
|
"loss": 0.8138, |
|
"step": 636 |
|
}, |
|
{ |
|
"epoch": 3.47, |
|
"learning_rate": 2.1530871053057843e-06, |
|
"loss": 0.8106, |
|
"step": 637 |
|
}, |
|
{ |
|
"epoch": 3.47, |
|
"learning_rate": 2.138835682411257e-06, |
|
"loss": 0.9432, |
|
"step": 638 |
|
}, |
|
{ |
|
"epoch": 3.48, |
|
"learning_rate": 2.1246187377458227e-06, |
|
"loss": 0.8596, |
|
"step": 639 |
|
}, |
|
{ |
|
"epoch": 3.48, |
|
"learning_rate": 2.1104364426296237e-06, |
|
"loss": 0.8068, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 3.49, |
|
"learning_rate": 2.0962889679652576e-06, |
|
"loss": 0.7555, |
|
"step": 641 |
|
}, |
|
{ |
|
"epoch": 3.49, |
|
"learning_rate": 2.082176484235728e-06, |
|
"loss": 0.8644, |
|
"step": 642 |
|
}, |
|
{ |
|
"epoch": 3.5, |
|
"learning_rate": 2.0680991615023765e-06, |
|
"loss": 0.8524, |
|
"step": 643 |
|
}, |
|
{ |
|
"epoch": 3.5, |
|
"learning_rate": 2.05405716940284e-06, |
|
"loss": 0.7569, |
|
"step": 644 |
|
}, |
|
{ |
|
"epoch": 3.51, |
|
"learning_rate": 2.040050677149008e-06, |
|
"loss": 0.7284, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 3.52, |
|
"learning_rate": 2.026079853524981e-06, |
|
"loss": 0.9232, |
|
"step": 646 |
|
}, |
|
{ |
|
"epoch": 3.52, |
|
"learning_rate": 2.0121448668850424e-06, |
|
"loss": 0.8451, |
|
"step": 647 |
|
}, |
|
{ |
|
"epoch": 3.53, |
|
"learning_rate": 1.998245885151619e-06, |
|
"loss": 0.9504, |
|
"step": 648 |
|
}, |
|
{ |
|
"epoch": 3.53, |
|
"learning_rate": 1.984383075813268e-06, |
|
"loss": 0.9145, |
|
"step": 649 |
|
}, |
|
{ |
|
"epoch": 3.54, |
|
"learning_rate": 1.970556605922645e-06, |
|
"loss": 0.8912, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 3.54, |
|
"learning_rate": 1.956766642094513e-06, |
|
"loss": 0.8934, |
|
"step": 651 |
|
}, |
|
{ |
|
"epoch": 3.55, |
|
"learning_rate": 1.9430133505037133e-06, |
|
"loss": 0.9124, |
|
"step": 652 |
|
}, |
|
{ |
|
"epoch": 3.55, |
|
"learning_rate": 1.9292968968831734e-06, |
|
"loss": 0.8914, |
|
"step": 653 |
|
}, |
|
{ |
|
"epoch": 3.56, |
|
"learning_rate": 1.9156174465219073e-06, |
|
"loss": 0.8909, |
|
"step": 654 |
|
}, |
|
{ |
|
"epoch": 3.56, |
|
"learning_rate": 1.9019751642630252e-06, |
|
"loss": 0.8434, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 3.57, |
|
"learning_rate": 1.8883702145017452e-06, |
|
"loss": 0.9324, |
|
"step": 656 |
|
}, |
|
{ |
|
"epoch": 3.58, |
|
"learning_rate": 1.8748027611834135e-06, |
|
"loss": 0.9103, |
|
"step": 657 |
|
}, |
|
{ |
|
"epoch": 3.58, |
|
"learning_rate": 1.861272967801529e-06, |
|
"loss": 0.7558, |
|
"step": 658 |
|
}, |
|
{ |
|
"epoch": 3.59, |
|
"learning_rate": 1.84778099739577e-06, |
|
"loss": 0.9089, |
|
"step": 659 |
|
}, |
|
{ |
|
"epoch": 3.59, |
|
"learning_rate": 1.8343270125500379e-06, |
|
"loss": 0.8667, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 3.6, |
|
"learning_rate": 1.820911175390488e-06, |
|
"loss": 0.8894, |
|
"step": 661 |
|
}, |
|
{ |
|
"epoch": 3.6, |
|
"learning_rate": 1.807533647583577e-06, |
|
"loss": 0.8737, |
|
"step": 662 |
|
}, |
|
{ |
|
"epoch": 3.61, |
|
"learning_rate": 1.794194590334123e-06, |
|
"loss": 0.9017, |
|
"step": 663 |
|
}, |
|
{ |
|
"epoch": 3.61, |
|
"learning_rate": 1.780894164383355e-06, |
|
"loss": 0.8644, |
|
"step": 664 |
|
}, |
|
{ |
|
"epoch": 3.62, |
|
"learning_rate": 1.7676325300069824e-06, |
|
"loss": 0.8413, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 3.62, |
|
"learning_rate": 1.754409847013257e-06, |
|
"loss": 0.8635, |
|
"step": 666 |
|
}, |
|
{ |
|
"epoch": 3.63, |
|
"learning_rate": 1.741226274741048e-06, |
|
"loss": 1.03, |
|
"step": 667 |
|
}, |
|
{ |
|
"epoch": 3.64, |
|
"learning_rate": 1.728081972057925e-06, |
|
"loss": 0.8222, |
|
"step": 668 |
|
}, |
|
{ |
|
"epoch": 3.64, |
|
"learning_rate": 1.714977097358248e-06, |
|
"loss": 0.9032, |
|
"step": 669 |
|
}, |
|
{ |
|
"epoch": 3.65, |
|
"learning_rate": 1.7019118085612474e-06, |
|
"loss": 0.9348, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 3.65, |
|
"learning_rate": 1.6888862631091284e-06, |
|
"loss": 0.8573, |
|
"step": 671 |
|
}, |
|
{ |
|
"epoch": 3.66, |
|
"learning_rate": 1.6759006179651726e-06, |
|
"loss": 0.8666, |
|
"step": 672 |
|
}, |
|
{ |
|
"epoch": 3.66, |
|
"learning_rate": 1.6629550296118462e-06, |
|
"loss": 0.897, |
|
"step": 673 |
|
}, |
|
{ |
|
"epoch": 3.67, |
|
"learning_rate": 1.6500496540489142e-06, |
|
"loss": 1.02, |
|
"step": 674 |
|
}, |
|
{ |
|
"epoch": 3.67, |
|
"learning_rate": 1.6371846467915603e-06, |
|
"loss": 0.8099, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 3.68, |
|
"learning_rate": 1.624360162868514e-06, |
|
"loss": 0.7988, |
|
"step": 676 |
|
}, |
|
{ |
|
"epoch": 3.68, |
|
"learning_rate": 1.6115763568201786e-06, |
|
"loss": 0.8619, |
|
"step": 677 |
|
}, |
|
{ |
|
"epoch": 3.69, |
|
"learning_rate": 1.59883338269678e-06, |
|
"loss": 0.8467, |
|
"step": 678 |
|
}, |
|
{ |
|
"epoch": 3.7, |
|
"learning_rate": 1.5861313940564915e-06, |
|
"loss": 0.9343, |
|
"step": 679 |
|
}, |
|
{ |
|
"epoch": 3.7, |
|
"learning_rate": 1.5734705439636017e-06, |
|
"loss": 0.8878, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 3.71, |
|
"learning_rate": 1.5608509849866594e-06, |
|
"loss": 0.7844, |
|
"step": 681 |
|
}, |
|
{ |
|
"epoch": 3.71, |
|
"learning_rate": 1.5482728691966377e-06, |
|
"loss": 0.8583, |
|
"step": 682 |
|
}, |
|
{ |
|
"epoch": 3.72, |
|
"learning_rate": 1.535736348165106e-06, |
|
"loss": 0.8883, |
|
"step": 683 |
|
}, |
|
{ |
|
"epoch": 3.72, |
|
"learning_rate": 1.5232415729623973e-06, |
|
"loss": 0.9003, |
|
"step": 684 |
|
}, |
|
{ |
|
"epoch": 3.73, |
|
"learning_rate": 1.5107886941557853e-06, |
|
"loss": 0.7417, |
|
"step": 685 |
|
}, |
|
{ |
|
"epoch": 3.73, |
|
"learning_rate": 1.4983778618076783e-06, |
|
"loss": 0.842, |
|
"step": 686 |
|
}, |
|
{ |
|
"epoch": 3.74, |
|
"learning_rate": 1.4860092254738111e-06, |
|
"loss": 0.9448, |
|
"step": 687 |
|
}, |
|
{ |
|
"epoch": 3.74, |
|
"learning_rate": 1.4736829342014336e-06, |
|
"loss": 0.8548, |
|
"step": 688 |
|
}, |
|
{ |
|
"epoch": 3.75, |
|
"learning_rate": 1.461399136527522e-06, |
|
"loss": 0.6608, |
|
"step": 689 |
|
}, |
|
{ |
|
"epoch": 3.76, |
|
"learning_rate": 1.4491579804769817e-06, |
|
"loss": 0.8563, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 3.76, |
|
"learning_rate": 1.4369596135608789e-06, |
|
"loss": 0.6762, |
|
"step": 691 |
|
}, |
|
{ |
|
"epoch": 3.77, |
|
"learning_rate": 1.4248041827746445e-06, |
|
"loss": 0.8582, |
|
"step": 692 |
|
}, |
|
{ |
|
"epoch": 3.77, |
|
"learning_rate": 1.4126918345963136e-06, |
|
"loss": 0.9294, |
|
"step": 693 |
|
}, |
|
{ |
|
"epoch": 3.78, |
|
"learning_rate": 1.4006227149847584e-06, |
|
"loss": 0.8519, |
|
"step": 694 |
|
}, |
|
{ |
|
"epoch": 3.78, |
|
"learning_rate": 1.3885969693779277e-06, |
|
"loss": 0.7379, |
|
"step": 695 |
|
}, |
|
{ |
|
"epoch": 3.79, |
|
"learning_rate": 1.3766147426910959e-06, |
|
"loss": 0.796, |
|
"step": 696 |
|
}, |
|
{ |
|
"epoch": 3.79, |
|
"learning_rate": 1.364676179315117e-06, |
|
"loss": 0.7911, |
|
"step": 697 |
|
}, |
|
{ |
|
"epoch": 3.8, |
|
"learning_rate": 1.3527814231146813e-06, |
|
"loss": 0.8022, |
|
"step": 698 |
|
}, |
|
{ |
|
"epoch": 3.8, |
|
"learning_rate": 1.3409306174265873e-06, |
|
"loss": 0.7736, |
|
"step": 699 |
|
}, |
|
{ |
|
"epoch": 3.81, |
|
"learning_rate": 1.3291239050580085e-06, |
|
"loss": 0.8208, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 3.81, |
|
"learning_rate": 1.3173614282847807e-06, |
|
"loss": 0.9625, |
|
"step": 701 |
|
}, |
|
{ |
|
"epoch": 3.82, |
|
"learning_rate": 1.3056433288496739e-06, |
|
"loss": 0.8483, |
|
"step": 702 |
|
}, |
|
{ |
|
"epoch": 3.83, |
|
"learning_rate": 1.2939697479606993e-06, |
|
"loss": 0.8063, |
|
"step": 703 |
|
}, |
|
{ |
|
"epoch": 3.83, |
|
"learning_rate": 1.2823408262893971e-06, |
|
"loss": 0.9138, |
|
"step": 704 |
|
}, |
|
{ |
|
"epoch": 3.84, |
|
"learning_rate": 1.2707567039691505e-06, |
|
"loss": 0.8316, |
|
"step": 705 |
|
}, |
|
{ |
|
"epoch": 3.84, |
|
"learning_rate": 1.2592175205934865e-06, |
|
"loss": 0.7809, |
|
"step": 706 |
|
}, |
|
{ |
|
"epoch": 3.85, |
|
"learning_rate": 1.2477234152143996e-06, |
|
"loss": 0.8831, |
|
"step": 707 |
|
}, |
|
{ |
|
"epoch": 3.85, |
|
"learning_rate": 1.236274526340675e-06, |
|
"loss": 0.7254, |
|
"step": 708 |
|
}, |
|
{ |
|
"epoch": 3.86, |
|
"learning_rate": 1.2248709919362262e-06, |
|
"loss": 0.8232, |
|
"step": 709 |
|
}, |
|
{ |
|
"epoch": 3.86, |
|
"learning_rate": 1.213512949418419e-06, |
|
"loss": 0.7897, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 3.87, |
|
"learning_rate": 1.2022005356564264e-06, |
|
"loss": 0.8565, |
|
"step": 711 |
|
}, |
|
{ |
|
"epoch": 3.87, |
|
"learning_rate": 1.1909338869695747e-06, |
|
"loss": 0.9481, |
|
"step": 712 |
|
}, |
|
{ |
|
"epoch": 3.88, |
|
"learning_rate": 1.1797131391257027e-06, |
|
"loss": 0.8168, |
|
"step": 713 |
|
}, |
|
{ |
|
"epoch": 3.89, |
|
"learning_rate": 1.168538427339524e-06, |
|
"loss": 0.8379, |
|
"step": 714 |
|
}, |
|
{ |
|
"epoch": 3.89, |
|
"learning_rate": 1.1574098862709993e-06, |
|
"loss": 0.9077, |
|
"step": 715 |
|
}, |
|
{ |
|
"epoch": 3.9, |
|
"learning_rate": 1.1463276500237113e-06, |
|
"loss": 0.8559, |
|
"step": 716 |
|
}, |
|
{ |
|
"epoch": 3.9, |
|
"learning_rate": 1.1352918521432515e-06, |
|
"loss": 0.7869, |
|
"step": 717 |
|
}, |
|
{ |
|
"epoch": 3.91, |
|
"learning_rate": 1.1243026256156103e-06, |
|
"loss": 0.7304, |
|
"step": 718 |
|
}, |
|
{ |
|
"epoch": 3.91, |
|
"learning_rate": 1.1133601028655717e-06, |
|
"loss": 0.9448, |
|
"step": 719 |
|
}, |
|
{ |
|
"epoch": 3.92, |
|
"learning_rate": 1.1024644157551206e-06, |
|
"loss": 0.8451, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 3.92, |
|
"learning_rate": 1.091615695581853e-06, |
|
"loss": 0.8134, |
|
"step": 721 |
|
}, |
|
{ |
|
"epoch": 3.93, |
|
"learning_rate": 1.080814073077393e-06, |
|
"loss": 0.8453, |
|
"step": 722 |
|
}, |
|
{ |
|
"epoch": 3.93, |
|
"learning_rate": 1.0700596784058205e-06, |
|
"loss": 0.9098, |
|
"step": 723 |
|
}, |
|
{ |
|
"epoch": 3.94, |
|
"learning_rate": 1.059352641162098e-06, |
|
"loss": 0.8295, |
|
"step": 724 |
|
}, |
|
{ |
|
"epoch": 3.95, |
|
"learning_rate": 1.0486930903705095e-06, |
|
"loss": 0.959, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 3.95, |
|
"learning_rate": 1.0380811544831087e-06, |
|
"loss": 0.9064, |
|
"step": 726 |
|
}, |
|
{ |
|
"epoch": 3.96, |
|
"learning_rate": 1.0275169613781732e-06, |
|
"loss": 0.9544, |
|
"step": 727 |
|
}, |
|
{ |
|
"epoch": 3.96, |
|
"learning_rate": 1.0170006383586561e-06, |
|
"loss": 0.8453, |
|
"step": 728 |
|
}, |
|
{ |
|
"epoch": 3.97, |
|
"learning_rate": 1.006532312150658e-06, |
|
"loss": 0.8531, |
|
"step": 729 |
|
}, |
|
{ |
|
"epoch": 3.97, |
|
"learning_rate": 9.961121089018933e-07, |
|
"loss": 0.8159, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 3.98, |
|
"learning_rate": 9.857401541801814e-07, |
|
"loss": 0.8243, |
|
"step": 731 |
|
}, |
|
{ |
|
"epoch": 3.98, |
|
"learning_rate": 9.754165729719222e-07, |
|
"loss": 0.851, |
|
"step": 732 |
|
}, |
|
{ |
|
"epoch": 3.99, |
|
"learning_rate": 9.651414896805966e-07, |
|
"loss": 0.778, |
|
"step": 733 |
|
}, |
|
{ |
|
"epoch": 3.99, |
|
"learning_rate": 9.549150281252633e-07, |
|
"loss": 0.9498, |
|
"step": 734 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"learning_rate": 9.447373115390702e-07, |
|
"loss": 0.8119, |
|
"step": 735 |
|
}, |
|
{ |
|
"epoch": 4.01, |
|
"learning_rate": 9.346084625677676e-07, |
|
"loss": 0.9721, |
|
"step": 736 |
|
}, |
|
{ |
|
"epoch": 4.01, |
|
"learning_rate": 9.245286032682299e-07, |
|
"loss": 0.8062, |
|
"step": 737 |
|
}, |
|
{ |
|
"epoch": 4.02, |
|
"learning_rate": 9.144978551069867e-07, |
|
"loss": 0.8626, |
|
"step": 738 |
|
}, |
|
{ |
|
"epoch": 4.02, |
|
"learning_rate": 9.045163389587569e-07, |
|
"loss": 0.9832, |
|
"step": 739 |
|
}, |
|
{ |
|
"epoch": 4.03, |
|
"learning_rate": 8.945841751049916e-07, |
|
"loss": 0.8116, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 4.03, |
|
"learning_rate": 8.847014832324313e-07, |
|
"loss": 0.956, |
|
"step": 741 |
|
}, |
|
{ |
|
"epoch": 4.04, |
|
"learning_rate": 8.748683824316518e-07, |
|
"loss": 0.7777, |
|
"step": 742 |
|
}, |
|
{ |
|
"epoch": 4.04, |
|
"learning_rate": 8.650849911956388e-07, |
|
"loss": 0.8544, |
|
"step": 743 |
|
}, |
|
{ |
|
"epoch": 4.05, |
|
"learning_rate": 8.553514274183561e-07, |
|
"loss": 0.7898, |
|
"step": 744 |
|
}, |
|
{ |
|
"epoch": 4.05, |
|
"learning_rate": 8.45667808393329e-07, |
|
"loss": 0.8439, |
|
"step": 745 |
|
}, |
|
{ |
|
"epoch": 4.06, |
|
"learning_rate": 8.360342508122238e-07, |
|
"loss": 0.8586, |
|
"step": 746 |
|
}, |
|
{ |
|
"epoch": 4.07, |
|
"learning_rate": 8.264508707634472e-07, |
|
"loss": 0.9113, |
|
"step": 747 |
|
}, |
|
{ |
|
"epoch": 4.07, |
|
"learning_rate": 8.169177837307418e-07, |
|
"loss": 0.7554, |
|
"step": 748 |
|
}, |
|
{ |
|
"epoch": 4.08, |
|
"learning_rate": 8.074351045918055e-07, |
|
"loss": 1.0145, |
|
"step": 749 |
|
}, |
|
{ |
|
"epoch": 4.08, |
|
"learning_rate": 7.980029476168943e-07, |
|
"loss": 0.8817, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 4.09, |
|
"learning_rate": 7.886214264674525e-07, |
|
"loss": 0.827, |
|
"step": 751 |
|
}, |
|
{ |
|
"epoch": 4.09, |
|
"learning_rate": 7.792906541947421e-07, |
|
"loss": 0.8639, |
|
"step": 752 |
|
}, |
|
{ |
|
"epoch": 4.1, |
|
"learning_rate": 7.700107432384785e-07, |
|
"loss": 0.8534, |
|
"step": 753 |
|
}, |
|
{ |
|
"epoch": 4.1, |
|
"learning_rate": 7.60781805425479e-07, |
|
"loss": 0.88, |
|
"step": 754 |
|
}, |
|
{ |
|
"epoch": 4.11, |
|
"learning_rate": 7.516039519683105e-07, |
|
"loss": 0.8416, |
|
"step": 755 |
|
}, |
|
{ |
|
"epoch": 4.11, |
|
"learning_rate": 7.424772934639552e-07, |
|
"loss": 0.8425, |
|
"step": 756 |
|
}, |
|
{ |
|
"epoch": 4.12, |
|
"learning_rate": 7.334019398924714e-07, |
|
"loss": 0.7865, |
|
"step": 757 |
|
}, |
|
{ |
|
"epoch": 4.13, |
|
"learning_rate": 7.243780006156737e-07, |
|
"loss": 0.937, |
|
"step": 758 |
|
}, |
|
{ |
|
"epoch": 4.13, |
|
"learning_rate": 7.154055843758118e-07, |
|
"loss": 0.884, |
|
"step": 759 |
|
}, |
|
{ |
|
"epoch": 4.14, |
|
"learning_rate": 7.064847992942614e-07, |
|
"loss": 0.884, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 4.14, |
|
"learning_rate": 6.976157528702221e-07, |
|
"loss": 0.9562, |
|
"step": 761 |
|
}, |
|
{ |
|
"epoch": 4.15, |
|
"learning_rate": 6.887985519794188e-07, |
|
"loss": 0.7733, |
|
"step": 762 |
|
}, |
|
{ |
|
"epoch": 4.15, |
|
"learning_rate": 6.80033302872819e-07, |
|
"loss": 0.7185, |
|
"step": 763 |
|
}, |
|
{ |
|
"epoch": 4.16, |
|
"learning_rate": 6.713201111753487e-07, |
|
"loss": 0.884, |
|
"step": 764 |
|
}, |
|
{ |
|
"epoch": 4.16, |
|
"learning_rate": 6.626590818846163e-07, |
|
"loss": 0.7337, |
|
"step": 765 |
|
}, |
|
{ |
|
"epoch": 4.17, |
|
"learning_rate": 6.540503193696551e-07, |
|
"loss": 0.8191, |
|
"step": 766 |
|
}, |
|
{ |
|
"epoch": 4.17, |
|
"learning_rate": 6.454939273696631e-07, |
|
"loss": 0.7203, |
|
"step": 767 |
|
}, |
|
{ |
|
"epoch": 4.18, |
|
"learning_rate": 6.369900089927488e-07, |
|
"loss": 0.8857, |
|
"step": 768 |
|
}, |
|
{ |
|
"epoch": 4.19, |
|
"learning_rate": 6.285386667146937e-07, |
|
"loss": 0.7981, |
|
"step": 769 |
|
}, |
|
{ |
|
"epoch": 4.19, |
|
"learning_rate": 6.201400023777105e-07, |
|
"loss": 0.8905, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 4.2, |
|
"learning_rate": 6.117941171892272e-07, |
|
"loss": 0.8076, |
|
"step": 771 |
|
}, |
|
{ |
|
"epoch": 4.2, |
|
"learning_rate": 6.035011117206574e-07, |
|
"loss": 0.8366, |
|
"step": 772 |
|
}, |
|
{ |
|
"epoch": 4.21, |
|
"learning_rate": 5.952610859061925e-07, |
|
"loss": 0.8544, |
|
"step": 773 |
|
}, |
|
{ |
|
"epoch": 4.21, |
|
"learning_rate": 5.870741390415963e-07, |
|
"loss": 0.9062, |
|
"step": 774 |
|
}, |
|
{ |
|
"epoch": 4.22, |
|
"learning_rate": 5.789403697830104e-07, |
|
"loss": 0.878, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 4.22, |
|
"learning_rate": 5.708598761457623e-07, |
|
"loss": 0.8995, |
|
"step": 776 |
|
}, |
|
{ |
|
"epoch": 4.23, |
|
"learning_rate": 5.628327555031865e-07, |
|
"loss": 0.7554, |
|
"step": 777 |
|
}, |
|
{ |
|
"epoch": 4.23, |
|
"learning_rate": 5.548591045854518e-07, |
|
"loss": 0.8262, |
|
"step": 778 |
|
}, |
|
{ |
|
"epoch": 4.24, |
|
"learning_rate": 5.46939019478393e-07, |
|
"loss": 0.8744, |
|
"step": 779 |
|
}, |
|
{ |
|
"epoch": 4.24, |
|
"learning_rate": 5.390725956223531e-07, |
|
"loss": 0.8344, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 4.25, |
|
"learning_rate": 5.312599278110403e-07, |
|
"loss": 1.0051, |
|
"step": 781 |
|
}, |
|
{ |
|
"epoch": 4.26, |
|
"learning_rate": 5.235011101903725e-07, |
|
"loss": 0.9111, |
|
"step": 782 |
|
}, |
|
{ |
|
"epoch": 4.26, |
|
"learning_rate": 5.157962362573543e-07, |
|
"loss": 0.8366, |
|
"step": 783 |
|
}, |
|
{ |
|
"epoch": 4.27, |
|
"learning_rate": 5.081453988589447e-07, |
|
"loss": 0.7421, |
|
"step": 784 |
|
}, |
|
{ |
|
"epoch": 4.27, |
|
"learning_rate": 5.005486901909429e-07, |
|
"loss": 0.924, |
|
"step": 785 |
|
}, |
|
{ |
|
"epoch": 4.28, |
|
"learning_rate": 4.930062017968706e-07, |
|
"loss": 0.7788, |
|
"step": 786 |
|
}, |
|
{ |
|
"epoch": 4.28, |
|
"learning_rate": 4.855180245668755e-07, |
|
"loss": 0.7902, |
|
"step": 787 |
|
}, |
|
{ |
|
"epoch": 4.29, |
|
"learning_rate": 4.780842487366283e-07, |
|
"loss": 0.7286, |
|
"step": 788 |
|
}, |
|
{ |
|
"epoch": 4.29, |
|
"learning_rate": 4.7070496388624544e-07, |
|
"loss": 0.9672, |
|
"step": 789 |
|
}, |
|
{ |
|
"epoch": 4.3, |
|
"learning_rate": 4.6338025893920167e-07, |
|
"loss": 0.7559, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 4.3, |
|
"learning_rate": 4.5611022216126124e-07, |
|
"loss": 0.8186, |
|
"step": 791 |
|
}, |
|
{ |
|
"epoch": 4.31, |
|
"learning_rate": 4.488949411594135e-07, |
|
"loss": 0.7844, |
|
"step": 792 |
|
}, |
|
{ |
|
"epoch": 4.32, |
|
"learning_rate": 4.4173450288081844e-07, |
|
"loss": 0.7213, |
|
"step": 793 |
|
}, |
|
{ |
|
"epoch": 4.32, |
|
"learning_rate": 4.346289936117587e-07, |
|
"loss": 0.9005, |
|
"step": 794 |
|
}, |
|
{ |
|
"epoch": 4.33, |
|
"learning_rate": 4.275784989765985e-07, |
|
"loss": 0.9292, |
|
"step": 795 |
|
}, |
|
{ |
|
"epoch": 4.33, |
|
"learning_rate": 4.205831039367525e-07, |
|
"loss": 0.8719, |
|
"step": 796 |
|
}, |
|
{ |
|
"epoch": 4.34, |
|
"learning_rate": 4.13642892789663e-07, |
|
"loss": 0.8004, |
|
"step": 797 |
|
}, |
|
{ |
|
"epoch": 4.34, |
|
"learning_rate": 4.067579491677831e-07, |
|
"loss": 0.9129, |
|
"step": 798 |
|
}, |
|
{ |
|
"epoch": 4.35, |
|
"learning_rate": 3.999283560375683e-07, |
|
"loss": 0.8006, |
|
"step": 799 |
|
}, |
|
{ |
|
"epoch": 4.35, |
|
"learning_rate": 3.93154195698478e-07, |
|
"loss": 0.9297, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 4.36, |
|
"learning_rate": 3.8643554978198385e-07, |
|
"loss": 0.974, |
|
"step": 801 |
|
}, |
|
{ |
|
"epoch": 4.36, |
|
"learning_rate": 3.7977249925058303e-07, |
|
"loss": 0.9562, |
|
"step": 802 |
|
}, |
|
{ |
|
"epoch": 4.37, |
|
"learning_rate": 3.7316512439682926e-07, |
|
"loss": 0.9122, |
|
"step": 803 |
|
}, |
|
{ |
|
"epoch": 4.38, |
|
"learning_rate": 3.6661350484235857e-07, |
|
"loss": 0.8722, |
|
"step": 804 |
|
}, |
|
{ |
|
"epoch": 4.38, |
|
"learning_rate": 3.6011771953693044e-07, |
|
"loss": 0.9109, |
|
"step": 805 |
|
}, |
|
{ |
|
"epoch": 4.39, |
|
"learning_rate": 3.5367784675747975e-07, |
|
"loss": 0.9652, |
|
"step": 806 |
|
}, |
|
{ |
|
"epoch": 4.39, |
|
"learning_rate": 3.472939641071743e-07, |
|
"loss": 0.8378, |
|
"step": 807 |
|
}, |
|
{ |
|
"epoch": 4.4, |
|
"learning_rate": 3.409661485144744e-07, |
|
"loss": 0.8497, |
|
"step": 808 |
|
}, |
|
{ |
|
"epoch": 4.4, |
|
"learning_rate": 3.346944762322102e-07, |
|
"loss": 0.9019, |
|
"step": 809 |
|
}, |
|
{ |
|
"epoch": 4.41, |
|
"learning_rate": 3.284790228366602e-07, |
|
"loss": 0.7724, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 4.41, |
|
"learning_rate": 3.2231986322664386e-07, |
|
"loss": 0.7944, |
|
"step": 811 |
|
}, |
|
{ |
|
"epoch": 4.42, |
|
"learning_rate": 3.162170716226148e-07, |
|
"loss": 0.9376, |
|
"step": 812 |
|
}, |
|
{ |
|
"epoch": 4.42, |
|
"learning_rate": 3.1017072156576957e-07, |
|
"loss": 0.8481, |
|
"step": 813 |
|
}, |
|
{ |
|
"epoch": 4.43, |
|
"learning_rate": 3.0418088591716076e-07, |
|
"loss": 0.87, |
|
"step": 814 |
|
}, |
|
{ |
|
"epoch": 4.44, |
|
"learning_rate": 2.982476368568177e-07, |
|
"loss": 0.9193, |
|
"step": 815 |
|
}, |
|
{ |
|
"epoch": 4.44, |
|
"learning_rate": 2.923710458828777e-07, |
|
"loss": 0.9044, |
|
"step": 816 |
|
}, |
|
{ |
|
"epoch": 4.45, |
|
"learning_rate": 2.86551183810726e-07, |
|
"loss": 0.8131, |
|
"step": 817 |
|
}, |
|
{ |
|
"epoch": 4.45, |
|
"learning_rate": 2.80788120772138e-07, |
|
"loss": 0.7668, |
|
"step": 818 |
|
}, |
|
{ |
|
"epoch": 4.46, |
|
"learning_rate": 2.7508192621443994e-07, |
|
"loss": 0.8382, |
|
"step": 819 |
|
}, |
|
{ |
|
"epoch": 4.46, |
|
"learning_rate": 2.6943266889966624e-07, |
|
"loss": 0.8403, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 4.47, |
|
"learning_rate": 2.638404169037373e-07, |
|
"loss": 0.923, |
|
"step": 821 |
|
}, |
|
{ |
|
"epoch": 4.47, |
|
"learning_rate": 2.583052376156314e-07, |
|
"loss": 0.8127, |
|
"step": 822 |
|
}, |
|
{ |
|
"epoch": 4.48, |
|
"learning_rate": 2.528271977365787e-07, |
|
"loss": 0.8425, |
|
"step": 823 |
|
}, |
|
{ |
|
"epoch": 4.48, |
|
"learning_rate": 2.474063632792556e-07, |
|
"loss": 0.8664, |
|
"step": 824 |
|
}, |
|
{ |
|
"epoch": 4.49, |
|
"learning_rate": 2.4204279956698994e-07, |
|
"loss": 0.9305, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 4.5, |
|
"learning_rate": 2.3673657123297166e-07, |
|
"loss": 0.779, |
|
"step": 826 |
|
}, |
|
{ |
|
"epoch": 4.5, |
|
"learning_rate": 2.3148774221947667e-07, |
|
"loss": 0.8542, |
|
"step": 827 |
|
}, |
|
{ |
|
"epoch": 4.51, |
|
"learning_rate": 2.2629637577709252e-07, |
|
"loss": 0.8933, |
|
"step": 828 |
|
}, |
|
{ |
|
"epoch": 4.51, |
|
"learning_rate": 2.2116253446396175e-07, |
|
"loss": 0.8068, |
|
"step": 829 |
|
}, |
|
{ |
|
"epoch": 4.52, |
|
"learning_rate": 2.1608628014502364e-07, |
|
"loss": 0.75, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 4.52, |
|
"learning_rate": 2.1106767399126982e-07, |
|
"loss": 0.9549, |
|
"step": 831 |
|
}, |
|
{ |
|
"epoch": 4.53, |
|
"learning_rate": 2.061067764790087e-07, |
|
"loss": 0.9055, |
|
"step": 832 |
|
}, |
|
{ |
|
"epoch": 4.53, |
|
"learning_rate": 2.0120364738913212e-07, |
|
"loss": 0.8934, |
|
"step": 833 |
|
}, |
|
{ |
|
"epoch": 4.54, |
|
"learning_rate": 1.9635834580640223e-07, |
|
"loss": 0.8489, |
|
"step": 834 |
|
}, |
|
{ |
|
"epoch": 4.54, |
|
"learning_rate": 1.915709301187335e-07, |
|
"loss": 0.7928, |
|
"step": 835 |
|
}, |
|
{ |
|
"epoch": 4.55, |
|
"learning_rate": 1.8684145801649067e-07, |
|
"loss": 0.9472, |
|
"step": 836 |
|
}, |
|
{ |
|
"epoch": 4.56, |
|
"learning_rate": 1.821699864917953e-07, |
|
"loss": 0.9573, |
|
"step": 837 |
|
}, |
|
{ |
|
"epoch": 4.56, |
|
"learning_rate": 1.7755657183783638e-07, |
|
"loss": 0.8264, |
|
"step": 838 |
|
}, |
|
{ |
|
"epoch": 4.57, |
|
"learning_rate": 1.7300126964819363e-07, |
|
"loss": 0.8467, |
|
"step": 839 |
|
}, |
|
{ |
|
"epoch": 4.57, |
|
"learning_rate": 1.6850413481616868e-07, |
|
"loss": 0.9277, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 4.58, |
|
"learning_rate": 1.6406522153411985e-07, |
|
"loss": 0.9041, |
|
"step": 841 |
|
}, |
|
{ |
|
"epoch": 4.58, |
|
"learning_rate": 1.596845832928129e-07, |
|
"loss": 0.8498, |
|
"step": 842 |
|
}, |
|
{ |
|
"epoch": 4.59, |
|
"learning_rate": 1.5536227288077466e-07, |
|
"loss": 0.8126, |
|
"step": 843 |
|
}, |
|
{ |
|
"epoch": 4.59, |
|
"learning_rate": 1.510983423836576e-07, |
|
"loss": 0.9155, |
|
"step": 844 |
|
}, |
|
{ |
|
"epoch": 4.6, |
|
"learning_rate": 1.468928431836092e-07, |
|
"loss": 0.8811, |
|
"step": 845 |
|
}, |
|
{ |
|
"epoch": 4.6, |
|
"learning_rate": 1.427458259586584e-07, |
|
"loss": 0.7399, |
|
"step": 846 |
|
}, |
|
{ |
|
"epoch": 4.61, |
|
"learning_rate": 1.3865734068210124e-07, |
|
"loss": 0.8392, |
|
"step": 847 |
|
}, |
|
{ |
|
"epoch": 4.61, |
|
"learning_rate": 1.3462743662189802e-07, |
|
"loss": 0.8258, |
|
"step": 848 |
|
}, |
|
{ |
|
"epoch": 4.62, |
|
"learning_rate": 1.3065616234008204e-07, |
|
"loss": 0.9331, |
|
"step": 849 |
|
}, |
|
{ |
|
"epoch": 4.63, |
|
"learning_rate": 1.2674356569217282e-07, |
|
"loss": 0.839, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 4.63, |
|
"learning_rate": 1.228896938265983e-07, |
|
"loss": 0.9558, |
|
"step": 851 |
|
}, |
|
{ |
|
"epoch": 4.64, |
|
"learning_rate": 1.1909459318413086e-07, |
|
"loss": 0.8337, |
|
"step": 852 |
|
}, |
|
{ |
|
"epoch": 4.64, |
|
"learning_rate": 1.1535830949732318e-07, |
|
"loss": 0.8119, |
|
"step": 853 |
|
}, |
|
{ |
|
"epoch": 4.65, |
|
"learning_rate": 1.1168088778996e-07, |
|
"loss": 0.9339, |
|
"step": 854 |
|
}, |
|
{ |
|
"epoch": 4.65, |
|
"learning_rate": 1.080623723765134e-07, |
|
"loss": 0.878, |
|
"step": 855 |
|
}, |
|
{ |
|
"epoch": 4.66, |
|
"learning_rate": 1.0450280686161163e-07, |
|
"loss": 0.8696, |
|
"step": 856 |
|
}, |
|
{ |
|
"epoch": 4.66, |
|
"learning_rate": 1.010022341395095e-07, |
|
"loss": 0.798, |
|
"step": 857 |
|
}, |
|
{ |
|
"epoch": 4.67, |
|
"learning_rate": 9.756069639357602e-08, |
|
"loss": 0.7726, |
|
"step": 858 |
|
}, |
|
{ |
|
"epoch": 4.67, |
|
"learning_rate": 9.417823509578316e-08, |
|
"loss": 0.9084, |
|
"step": 859 |
|
}, |
|
{ |
|
"epoch": 4.68, |
|
"learning_rate": 9.085489100620737e-08, |
|
"loss": 0.8672, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 4.69, |
|
"learning_rate": 8.759070417253768e-08, |
|
"loss": 0.9089, |
|
"step": 861 |
|
}, |
|
{ |
|
"epoch": 4.69, |
|
"learning_rate": 8.438571392959338e-08, |
|
"loss": 0.8619, |
|
"step": 862 |
|
}, |
|
{ |
|
"epoch": 4.7, |
|
"learning_rate": 8.123995889884995e-08, |
|
"loss": 0.919, |
|
"step": 863 |
|
}, |
|
{ |
|
"epoch": 4.7, |
|
"learning_rate": 7.815347698797327e-08, |
|
"loss": 0.9736, |
|
"step": 864 |
|
}, |
|
{ |
|
"epoch": 4.71, |
|
"learning_rate": 7.512630539036502e-08, |
|
"loss": 0.8875, |
|
"step": 865 |
|
}, |
|
{ |
|
"epoch": 4.71, |
|
"learning_rate": 7.215848058471086e-08, |
|
"loss": 0.8318, |
|
"step": 866 |
|
}, |
|
{ |
|
"epoch": 4.72, |
|
"learning_rate": 6.925003833454402e-08, |
|
"loss": 0.8584, |
|
"step": 867 |
|
}, |
|
{ |
|
"epoch": 4.72, |
|
"learning_rate": 6.640101368781126e-08, |
|
"loss": 0.9158, |
|
"step": 868 |
|
}, |
|
{ |
|
"epoch": 4.73, |
|
"learning_rate": 6.361144097645489e-08, |
|
"loss": 0.7842, |
|
"step": 869 |
|
}, |
|
{ |
|
"epoch": 4.73, |
|
"learning_rate": 6.088135381599414e-08, |
|
"loss": 0.6236, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 4.74, |
|
"learning_rate": 5.821078510512446e-08, |
|
"loss": 0.8197, |
|
"step": 871 |
|
}, |
|
{ |
|
"epoch": 4.75, |
|
"learning_rate": 5.559976702531722e-08, |
|
"loss": 0.8129, |
|
"step": 872 |
|
}, |
|
{ |
|
"epoch": 4.75, |
|
"learning_rate": 5.304833104043505e-08, |
|
"loss": 0.9605, |
|
"step": 873 |
|
}, |
|
{ |
|
"epoch": 4.76, |
|
"learning_rate": 5.0556507896350473e-08, |
|
"loss": 0.8278, |
|
"step": 874 |
|
}, |
|
{ |
|
"epoch": 4.76, |
|
"learning_rate": 4.8124327620576726e-08, |
|
"loss": 0.8557, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 4.77, |
|
"learning_rate": 4.575181952190533e-08, |
|
"loss": 0.8224, |
|
"step": 876 |
|
}, |
|
{ |
|
"epoch": 4.77, |
|
"learning_rate": 4.3439012190053534e-08, |
|
"loss": 0.5978, |
|
"step": 877 |
|
}, |
|
{ |
|
"epoch": 4.78, |
|
"learning_rate": 4.1185933495317965e-08, |
|
"loss": 0.8474, |
|
"step": 878 |
|
}, |
|
{ |
|
"epoch": 4.78, |
|
"learning_rate": 3.899261058824266e-08, |
|
"loss": 0.7566, |
|
"step": 879 |
|
}, |
|
{ |
|
"epoch": 4.79, |
|
"learning_rate": 3.685906989928656e-08, |
|
"loss": 0.8055, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 4.79, |
|
"learning_rate": 3.478533713850929e-08, |
|
"loss": 0.8079, |
|
"step": 881 |
|
}, |
|
{ |
|
"epoch": 4.8, |
|
"learning_rate": 3.277143729526033e-08, |
|
"loss": 0.8823, |
|
"step": 882 |
|
}, |
|
{ |
|
"epoch": 4.81, |
|
"learning_rate": 3.081739463787592e-08, |
|
"loss": 0.8067, |
|
"step": 883 |
|
}, |
|
{ |
|
"epoch": 4.81, |
|
"learning_rate": 2.892323271339037e-08, |
|
"loss": 0.8689, |
|
"step": 884 |
|
}, |
|
{ |
|
"epoch": 4.82, |
|
"learning_rate": 2.7088974347246888e-08, |
|
"loss": 0.9433, |
|
"step": 885 |
|
}, |
|
{ |
|
"epoch": 4.82, |
|
"learning_rate": 2.531464164302888e-08, |
|
"loss": 0.8466, |
|
"step": 886 |
|
}, |
|
{ |
|
"epoch": 4.83, |
|
"learning_rate": 2.3600255982187958e-08, |
|
"loss": 0.9036, |
|
"step": 887 |
|
}, |
|
{ |
|
"epoch": 4.83, |
|
"learning_rate": 2.19458380237908e-08, |
|
"loss": 0.9569, |
|
"step": 888 |
|
}, |
|
{ |
|
"epoch": 4.84, |
|
"learning_rate": 2.0351407704267134e-08, |
|
"loss": 0.9518, |
|
"step": 889 |
|
}, |
|
{ |
|
"epoch": 4.84, |
|
"learning_rate": 1.8816984237169378e-08, |
|
"loss": 0.8609, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 4.85, |
|
"learning_rate": 1.734258611294448e-08, |
|
"loss": 0.9466, |
|
"step": 891 |
|
}, |
|
{ |
|
"epoch": 4.85, |
|
"learning_rate": 1.5928231098706893e-08, |
|
"loss": 0.8711, |
|
"step": 892 |
|
}, |
|
{ |
|
"epoch": 4.86, |
|
"learning_rate": 1.4573936238028163e-08, |
|
"loss": 0.94, |
|
"step": 893 |
|
}, |
|
{ |
|
"epoch": 4.87, |
|
"learning_rate": 1.3279717850728236e-08, |
|
"loss": 0.8015, |
|
"step": 894 |
|
}, |
|
{ |
|
"epoch": 4.87, |
|
"learning_rate": 1.2045591532681145e-08, |
|
"loss": 0.7746, |
|
"step": 895 |
|
}, |
|
{ |
|
"epoch": 4.88, |
|
"learning_rate": 1.0871572155626841e-08, |
|
"loss": 0.9479, |
|
"step": 896 |
|
}, |
|
{ |
|
"epoch": 4.88, |
|
"learning_rate": 9.757673866990225e-09, |
|
"loss": 0.878, |
|
"step": 897 |
|
}, |
|
{ |
|
"epoch": 4.89, |
|
"learning_rate": 8.703910089712387e-09, |
|
"loss": 0.9255, |
|
"step": 898 |
|
}, |
|
{ |
|
"epoch": 4.89, |
|
"learning_rate": 7.710293522088518e-09, |
|
"loss": 0.9318, |
|
"step": 899 |
|
}, |
|
{ |
|
"epoch": 4.9, |
|
"learning_rate": 6.7768361376152616e-09, |
|
"loss": 0.8398, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 4.9, |
|
"learning_rate": 5.903549184844703e-09, |
|
"loss": 0.7752, |
|
"step": 901 |
|
}, |
|
{ |
|
"epoch": 4.91, |
|
"learning_rate": 5.090443187251159e-09, |
|
"loss": 0.8166, |
|
"step": 902 |
|
}, |
|
{ |
|
"epoch": 4.91, |
|
"learning_rate": 4.337527943101827e-09, |
|
"loss": 0.713, |
|
"step": 903 |
|
}, |
|
{ |
|
"epoch": 4.92, |
|
"learning_rate": 3.6448125253413236e-09, |
|
"loss": 0.9034, |
|
"step": 904 |
|
}, |
|
{ |
|
"epoch": 4.93, |
|
"learning_rate": 3.0123052814812203e-09, |
|
"loss": 0.867, |
|
"step": 905 |
|
}, |
|
{ |
|
"epoch": 4.93, |
|
"learning_rate": 2.440013833498456e-09, |
|
"loss": 0.9285, |
|
"step": 906 |
|
}, |
|
{ |
|
"epoch": 4.94, |
|
"learning_rate": 1.9279450777442976e-09, |
|
"loss": 0.8003, |
|
"step": 907 |
|
}, |
|
{ |
|
"epoch": 4.94, |
|
"learning_rate": 1.4761051848627417e-09, |
|
"loss": 0.8849, |
|
"step": 908 |
|
}, |
|
{ |
|
"epoch": 4.95, |
|
"learning_rate": 1.0844995997139064e-09, |
|
"loss": 0.8478, |
|
"step": 909 |
|
}, |
|
{ |
|
"epoch": 4.95, |
|
"learning_rate": 7.53133041307974e-10, |
|
"loss": 0.8967, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 4.96, |
|
"learning_rate": 4.820095027524563e-10, |
|
"loss": 0.9835, |
|
"step": 911 |
|
}, |
|
{ |
|
"epoch": 4.96, |
|
"learning_rate": 2.7113225119834717e-10, |
|
"loss": 0.8379, |
|
"step": 912 |
|
}, |
|
{ |
|
"epoch": 4.97, |
|
"learning_rate": 1.205038278051518e-10, |
|
"loss": 0.9426, |
|
"step": 913 |
|
}, |
|
{ |
|
"epoch": 4.97, |
|
"learning_rate": 3.0126047709244654e-11, |
|
"loss": 0.8376, |
|
"step": 914 |
|
}, |
|
{ |
|
"epoch": 4.98, |
|
"learning_rate": 0.0, |
|
"loss": 0.9024, |
|
"step": 915 |
|
}, |
|
{ |
|
"epoch": 4.98, |
|
"step": 915, |
|
"total_flos": 8.22738094695383e+17, |
|
"train_loss": 0.9721869942920456, |
|
"train_runtime": 16666.1068, |
|
"train_samples_per_second": 0.441, |
|
"train_steps_per_second": 0.055 |
|
} |
|
], |
|
"logging_steps": 1.0, |
|
"max_steps": 915, |
|
"num_train_epochs": 5, |
|
"save_steps": 1000, |
|
"total_flos": 8.22738094695383e+17, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|