{ "best_metric": null, "best_model_checkpoint": null, "epoch": 20.0, "eval_steps": 500, "global_step": 27700, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.004332129963898917, "grad_norm": 2.0469717979431152, "learning_rate": 9.997833935018051e-05, "loss": 2.6564, "step": 6 }, { "epoch": 0.008664259927797834, "grad_norm": 1.5152959823608398, "learning_rate": 9.995667870036101e-05, "loss": 2.5081, "step": 12 }, { "epoch": 0.012996389891696752, "grad_norm": 1.9685180187225342, "learning_rate": 9.993501805054151e-05, "loss": 2.4962, "step": 18 }, { "epoch": 0.017328519855595668, "grad_norm": 1.3475664854049683, "learning_rate": 9.991335740072203e-05, "loss": 2.3351, "step": 24 }, { "epoch": 0.021660649819494584, "grad_norm": 1.3752148151397705, "learning_rate": 9.989169675090253e-05, "loss": 2.3886, "step": 30 }, { "epoch": 0.025992779783393503, "grad_norm": 1.3871607780456543, "learning_rate": 9.987003610108304e-05, "loss": 2.3351, "step": 36 }, { "epoch": 0.03032490974729242, "grad_norm": 1.3841423988342285, "learning_rate": 9.984837545126354e-05, "loss": 2.3754, "step": 42 }, { "epoch": 0.034657039711191336, "grad_norm": 1.3603386878967285, "learning_rate": 9.982671480144404e-05, "loss": 2.3838, "step": 48 }, { "epoch": 0.03898916967509025, "grad_norm": 1.6379190683364868, "learning_rate": 9.980505415162456e-05, "loss": 2.1723, "step": 54 }, { "epoch": 0.04332129963898917, "grad_norm": 1.4946362972259521, "learning_rate": 9.978339350180505e-05, "loss": 2.2127, "step": 60 }, { "epoch": 0.047653429602888084, "grad_norm": 1.305691123008728, "learning_rate": 9.976173285198557e-05, "loss": 2.3638, "step": 66 }, { "epoch": 0.05198555956678701, "grad_norm": 1.2658382654190063, "learning_rate": 9.974007220216607e-05, "loss": 2.1491, "step": 72 }, { "epoch": 0.05631768953068592, "grad_norm": 1.3745174407958984, "learning_rate": 9.971841155234657e-05, "loss": 2.3112, "step": 78 }, { "epoch": 0.06064981949458484, "grad_norm": 1.2713632583618164, "learning_rate": 9.969675090252708e-05, "loss": 2.3263, "step": 84 }, { "epoch": 0.06498194945848375, "grad_norm": 1.545251488685608, "learning_rate": 9.967509025270758e-05, "loss": 2.1061, "step": 90 }, { "epoch": 0.06931407942238267, "grad_norm": 1.369242787361145, "learning_rate": 9.96534296028881e-05, "loss": 2.2089, "step": 96 }, { "epoch": 0.0736462093862816, "grad_norm": 1.4624031782150269, "learning_rate": 9.96317689530686e-05, "loss": 2.2092, "step": 102 }, { "epoch": 0.0779783393501805, "grad_norm": 1.425108551979065, "learning_rate": 9.96101083032491e-05, "loss": 2.2893, "step": 108 }, { "epoch": 0.08231046931407943, "grad_norm": 1.4925994873046875, "learning_rate": 9.95884476534296e-05, "loss": 2.2243, "step": 114 }, { "epoch": 0.08664259927797834, "grad_norm": 1.3201707601547241, "learning_rate": 9.956678700361012e-05, "loss": 2.1551, "step": 120 }, { "epoch": 0.09097472924187726, "grad_norm": 1.3067225217819214, "learning_rate": 9.954512635379061e-05, "loss": 2.1861, "step": 126 }, { "epoch": 0.09530685920577617, "grad_norm": 1.3277055025100708, "learning_rate": 9.952346570397112e-05, "loss": 2.1544, "step": 132 }, { "epoch": 0.09963898916967509, "grad_norm": 1.2822017669677734, "learning_rate": 9.950180505415163e-05, "loss": 2.2172, "step": 138 }, { "epoch": 0.10397111913357401, "grad_norm": 1.2318449020385742, "learning_rate": 9.948014440433214e-05, "loss": 2.2287, "step": 144 }, { "epoch": 0.10830324909747292, "grad_norm": 1.230015754699707, "learning_rate": 9.945848375451264e-05, "loss": 2.3068, "step": 150 }, { "epoch": 0.11263537906137185, "grad_norm": 1.599064826965332, "learning_rate": 9.943682310469314e-05, "loss": 2.2334, "step": 156 }, { "epoch": 0.11696750902527075, "grad_norm": 1.2962054014205933, "learning_rate": 9.941516245487366e-05, "loss": 1.9842, "step": 162 }, { "epoch": 0.12129963898916968, "grad_norm": 1.3339725732803345, "learning_rate": 9.939350180505416e-05, "loss": 2.1986, "step": 168 }, { "epoch": 0.1256317689530686, "grad_norm": 1.356671929359436, "learning_rate": 9.937184115523465e-05, "loss": 2.171, "step": 174 }, { "epoch": 0.1299638989169675, "grad_norm": 1.276473879814148, "learning_rate": 9.935018050541517e-05, "loss": 2.2829, "step": 180 }, { "epoch": 0.13429602888086642, "grad_norm": 1.3547333478927612, "learning_rate": 9.932851985559567e-05, "loss": 2.1693, "step": 186 }, { "epoch": 0.13862815884476534, "grad_norm": 1.246497631072998, "learning_rate": 9.930685920577618e-05, "loss": 2.1175, "step": 192 }, { "epoch": 0.14296028880866427, "grad_norm": 1.1589373350143433, "learning_rate": 9.928519855595668e-05, "loss": 2.179, "step": 198 }, { "epoch": 0.1472924187725632, "grad_norm": 1.2698945999145508, "learning_rate": 9.92635379061372e-05, "loss": 2.1903, "step": 204 }, { "epoch": 0.15162454873646208, "grad_norm": 1.1812002658843994, "learning_rate": 9.92418772563177e-05, "loss": 2.2477, "step": 210 }, { "epoch": 0.155956678700361, "grad_norm": 1.3712854385375977, "learning_rate": 9.922021660649819e-05, "loss": 2.1653, "step": 216 }, { "epoch": 0.16028880866425993, "grad_norm": 1.1667898893356323, "learning_rate": 9.91985559566787e-05, "loss": 2.1272, "step": 222 }, { "epoch": 0.16462093862815885, "grad_norm": 1.2409676313400269, "learning_rate": 9.917689530685921e-05, "loss": 2.2061, "step": 228 }, { "epoch": 0.16895306859205775, "grad_norm": 1.3557971715927124, "learning_rate": 9.915523465703973e-05, "loss": 2.1441, "step": 234 }, { "epoch": 0.17328519855595667, "grad_norm": 1.34516179561615, "learning_rate": 9.913357400722022e-05, "loss": 2.1164, "step": 240 }, { "epoch": 0.1776173285198556, "grad_norm": 1.2913334369659424, "learning_rate": 9.911191335740073e-05, "loss": 2.1744, "step": 246 }, { "epoch": 0.18194945848375452, "grad_norm": 1.3375498056411743, "learning_rate": 9.909025270758124e-05, "loss": 2.0828, "step": 252 }, { "epoch": 0.18628158844765344, "grad_norm": 1.3266654014587402, "learning_rate": 9.906859205776174e-05, "loss": 2.0818, "step": 258 }, { "epoch": 0.19061371841155234, "grad_norm": 1.2326065301895142, "learning_rate": 9.904693140794224e-05, "loss": 2.0148, "step": 264 }, { "epoch": 0.19494584837545126, "grad_norm": 1.2848244905471802, "learning_rate": 9.902527075812274e-05, "loss": 2.1217, "step": 270 }, { "epoch": 0.19927797833935018, "grad_norm": 1.1974108219146729, "learning_rate": 9.900361010830326e-05, "loss": 2.0768, "step": 276 }, { "epoch": 0.2036101083032491, "grad_norm": 1.2478421926498413, "learning_rate": 9.898194945848376e-05, "loss": 2.11, "step": 282 }, { "epoch": 0.20794223826714803, "grad_norm": 1.5355215072631836, "learning_rate": 9.896028880866427e-05, "loss": 2.0641, "step": 288 }, { "epoch": 0.21227436823104692, "grad_norm": 1.321541428565979, "learning_rate": 9.893862815884477e-05, "loss": 2.3098, "step": 294 }, { "epoch": 0.21660649819494585, "grad_norm": 1.0981677770614624, "learning_rate": 9.891696750902527e-05, "loss": 2.1822, "step": 300 }, { "epoch": 0.22093862815884477, "grad_norm": 1.2628662586212158, "learning_rate": 9.889530685920578e-05, "loss": 2.0979, "step": 306 }, { "epoch": 0.2252707581227437, "grad_norm": 1.1747299432754517, "learning_rate": 9.887364620938628e-05, "loss": 2.1004, "step": 312 }, { "epoch": 0.2296028880866426, "grad_norm": 1.1220794916152954, "learning_rate": 9.88519855595668e-05, "loss": 2.0453, "step": 318 }, { "epoch": 0.2339350180505415, "grad_norm": 1.1372015476226807, "learning_rate": 9.88303249097473e-05, "loss": 2.1821, "step": 324 }, { "epoch": 0.23826714801444043, "grad_norm": 1.2937543392181396, "learning_rate": 9.88086642599278e-05, "loss": 2.1234, "step": 330 }, { "epoch": 0.24259927797833936, "grad_norm": 1.3127963542938232, "learning_rate": 9.878700361010831e-05, "loss": 2.1206, "step": 336 }, { "epoch": 0.24693140794223828, "grad_norm": 1.1687551736831665, "learning_rate": 9.876534296028881e-05, "loss": 2.1772, "step": 342 }, { "epoch": 0.2512635379061372, "grad_norm": 1.157422661781311, "learning_rate": 9.874368231046933e-05, "loss": 2.0222, "step": 348 }, { "epoch": 0.2555956678700361, "grad_norm": 1.1806762218475342, "learning_rate": 9.872202166064982e-05, "loss": 2.097, "step": 354 }, { "epoch": 0.259927797833935, "grad_norm": 1.2087692022323608, "learning_rate": 9.870036101083033e-05, "loss": 2.1732, "step": 360 }, { "epoch": 0.2642599277978339, "grad_norm": 1.197892189025879, "learning_rate": 9.867870036101084e-05, "loss": 2.0973, "step": 366 }, { "epoch": 0.26859205776173284, "grad_norm": 1.2604889869689941, "learning_rate": 9.865703971119134e-05, "loss": 2.154, "step": 372 }, { "epoch": 0.27292418772563176, "grad_norm": 1.193973183631897, "learning_rate": 9.863537906137184e-05, "loss": 2.0974, "step": 378 }, { "epoch": 0.2772563176895307, "grad_norm": 1.198718547821045, "learning_rate": 9.861371841155235e-05, "loss": 2.194, "step": 384 }, { "epoch": 0.2815884476534296, "grad_norm": 1.2242847681045532, "learning_rate": 9.859205776173286e-05, "loss": 2.164, "step": 390 }, { "epoch": 0.28592057761732853, "grad_norm": 1.243475317955017, "learning_rate": 9.857039711191335e-05, "loss": 2.0351, "step": 396 }, { "epoch": 0.29025270758122745, "grad_norm": 1.2920485734939575, "learning_rate": 9.854873646209387e-05, "loss": 2.1773, "step": 402 }, { "epoch": 0.2945848375451264, "grad_norm": 1.1937910318374634, "learning_rate": 9.852707581227437e-05, "loss": 1.9981, "step": 408 }, { "epoch": 0.29891696750902524, "grad_norm": 1.1500145196914673, "learning_rate": 9.850541516245489e-05, "loss": 2.0822, "step": 414 }, { "epoch": 0.30324909747292417, "grad_norm": 1.3043928146362305, "learning_rate": 9.848375451263538e-05, "loss": 2.0387, "step": 420 }, { "epoch": 0.3075812274368231, "grad_norm": 1.3072212934494019, "learning_rate": 9.846209386281588e-05, "loss": 2.0051, "step": 426 }, { "epoch": 0.311913357400722, "grad_norm": 1.3109694719314575, "learning_rate": 9.84404332129964e-05, "loss": 2.0945, "step": 432 }, { "epoch": 0.31624548736462094, "grad_norm": 1.2839062213897705, "learning_rate": 9.84187725631769e-05, "loss": 1.9797, "step": 438 }, { "epoch": 0.32057761732851986, "grad_norm": 1.413023591041565, "learning_rate": 9.83971119133574e-05, "loss": 1.9028, "step": 444 }, { "epoch": 0.3249097472924188, "grad_norm": 1.2802754640579224, "learning_rate": 9.837545126353791e-05, "loss": 1.9921, "step": 450 }, { "epoch": 0.3292418772563177, "grad_norm": 1.2412651777267456, "learning_rate": 9.835379061371843e-05, "loss": 2.0128, "step": 456 }, { "epoch": 0.33357400722021663, "grad_norm": 1.2361631393432617, "learning_rate": 9.833212996389892e-05, "loss": 2.0943, "step": 462 }, { "epoch": 0.3379061371841155, "grad_norm": 1.1547577381134033, "learning_rate": 9.831046931407942e-05, "loss": 1.9736, "step": 468 }, { "epoch": 0.3422382671480144, "grad_norm": 1.2576320171356201, "learning_rate": 9.828880866425994e-05, "loss": 2.1898, "step": 474 }, { "epoch": 0.34657039711191334, "grad_norm": 1.2113392353057861, "learning_rate": 9.826714801444044e-05, "loss": 2.0663, "step": 480 }, { "epoch": 0.35090252707581226, "grad_norm": 1.1123450994491577, "learning_rate": 9.824548736462094e-05, "loss": 2.1913, "step": 486 }, { "epoch": 0.3552346570397112, "grad_norm": 1.1093122959136963, "learning_rate": 9.822382671480145e-05, "loss": 1.9647, "step": 492 }, { "epoch": 0.3595667870036101, "grad_norm": 1.148982286453247, "learning_rate": 9.820216606498196e-05, "loss": 2.0398, "step": 498 }, { "epoch": 0.36389891696750903, "grad_norm": 1.0545357465744019, "learning_rate": 9.818050541516247e-05, "loss": 2.0518, "step": 504 }, { "epoch": 0.36823104693140796, "grad_norm": 1.1357444524765015, "learning_rate": 9.815884476534296e-05, "loss": 1.9665, "step": 510 }, { "epoch": 0.3725631768953069, "grad_norm": 1.2326693534851074, "learning_rate": 9.813718411552347e-05, "loss": 2.0741, "step": 516 }, { "epoch": 0.3768953068592058, "grad_norm": 1.2234129905700684, "learning_rate": 9.811552346570398e-05, "loss": 2.0192, "step": 522 }, { "epoch": 0.38122743682310467, "grad_norm": 1.1305060386657715, "learning_rate": 9.809386281588449e-05, "loss": 2.0265, "step": 528 }, { "epoch": 0.3855595667870036, "grad_norm": 1.2020946741104126, "learning_rate": 9.807220216606498e-05, "loss": 2.0858, "step": 534 }, { "epoch": 0.3898916967509025, "grad_norm": 1.155094027519226, "learning_rate": 9.805054151624548e-05, "loss": 2.0225, "step": 540 }, { "epoch": 0.39422382671480144, "grad_norm": 1.1797747611999512, "learning_rate": 9.8028880866426e-05, "loss": 1.9967, "step": 546 }, { "epoch": 0.39855595667870036, "grad_norm": 1.1708471775054932, "learning_rate": 9.80072202166065e-05, "loss": 2.0566, "step": 552 }, { "epoch": 0.4028880866425993, "grad_norm": 1.1178022623062134, "learning_rate": 9.798555956678701e-05, "loss": 2.0352, "step": 558 }, { "epoch": 0.4072202166064982, "grad_norm": 1.1828281879425049, "learning_rate": 9.796389891696751e-05, "loss": 2.1046, "step": 564 }, { "epoch": 0.41155234657039713, "grad_norm": 1.218628168106079, "learning_rate": 9.794223826714803e-05, "loss": 2.1466, "step": 570 }, { "epoch": 0.41588447653429605, "grad_norm": 1.1755589246749878, "learning_rate": 9.792057761732852e-05, "loss": 2.1087, "step": 576 }, { "epoch": 0.4202166064981949, "grad_norm": 1.2978154420852661, "learning_rate": 9.789891696750902e-05, "loss": 2.0252, "step": 582 }, { "epoch": 0.42454873646209385, "grad_norm": 1.2139396667480469, "learning_rate": 9.787725631768954e-05, "loss": 2.0352, "step": 588 }, { "epoch": 0.42888086642599277, "grad_norm": 1.0754166841506958, "learning_rate": 9.785559566787004e-05, "loss": 2.0927, "step": 594 }, { "epoch": 0.4332129963898917, "grad_norm": 1.5046592950820923, "learning_rate": 9.783393501805054e-05, "loss": 1.9956, "step": 600 }, { "epoch": 0.4375451263537906, "grad_norm": 1.1988701820373535, "learning_rate": 9.781227436823105e-05, "loss": 2.0618, "step": 606 }, { "epoch": 0.44187725631768954, "grad_norm": 1.267943263053894, "learning_rate": 9.779061371841156e-05, "loss": 1.998, "step": 612 }, { "epoch": 0.44620938628158846, "grad_norm": 1.171288251876831, "learning_rate": 9.776895306859207e-05, "loss": 1.8313, "step": 618 }, { "epoch": 0.4505415162454874, "grad_norm": 1.11545991897583, "learning_rate": 9.774729241877256e-05, "loss": 2.0237, "step": 624 }, { "epoch": 0.4548736462093863, "grad_norm": 1.227541446685791, "learning_rate": 9.772563176895307e-05, "loss": 2.0524, "step": 630 }, { "epoch": 0.4592057761732852, "grad_norm": 1.139235496520996, "learning_rate": 9.770397111913358e-05, "loss": 2.0968, "step": 636 }, { "epoch": 0.4635379061371841, "grad_norm": 1.20332670211792, "learning_rate": 9.768231046931408e-05, "loss": 2.154, "step": 642 }, { "epoch": 0.467870036101083, "grad_norm": 1.1637028455734253, "learning_rate": 9.766064981949458e-05, "loss": 1.9189, "step": 648 }, { "epoch": 0.47220216606498194, "grad_norm": 1.1110986471176147, "learning_rate": 9.76389891696751e-05, "loss": 2.0349, "step": 654 }, { "epoch": 0.47653429602888087, "grad_norm": 1.1178441047668457, "learning_rate": 9.76173285198556e-05, "loss": 1.9554, "step": 660 }, { "epoch": 0.4808664259927798, "grad_norm": 1.1718621253967285, "learning_rate": 9.75956678700361e-05, "loss": 2.0165, "step": 666 }, { "epoch": 0.4851985559566787, "grad_norm": 1.1984493732452393, "learning_rate": 9.757400722021661e-05, "loss": 2.0551, "step": 672 }, { "epoch": 0.48953068592057764, "grad_norm": 1.2409266233444214, "learning_rate": 9.755234657039711e-05, "loss": 2.083, "step": 678 }, { "epoch": 0.49386281588447656, "grad_norm": 1.16851806640625, "learning_rate": 9.753068592057763e-05, "loss": 1.9648, "step": 684 }, { "epoch": 0.4981949458483754, "grad_norm": 1.2248796224594116, "learning_rate": 9.750902527075812e-05, "loss": 2.0572, "step": 690 }, { "epoch": 0.5025270758122744, "grad_norm": 1.1372920274734497, "learning_rate": 9.748736462093864e-05, "loss": 2.1241, "step": 696 }, { "epoch": 0.5068592057761733, "grad_norm": 1.342521071434021, "learning_rate": 9.746570397111914e-05, "loss": 1.946, "step": 702 }, { "epoch": 0.5111913357400723, "grad_norm": 1.1315120458602905, "learning_rate": 9.744404332129964e-05, "loss": 2.1177, "step": 708 }, { "epoch": 0.5155234657039711, "grad_norm": 1.2093828916549683, "learning_rate": 9.742238267148015e-05, "loss": 2.0665, "step": 714 }, { "epoch": 0.51985559566787, "grad_norm": 1.1943097114562988, "learning_rate": 9.740072202166065e-05, "loss": 1.8964, "step": 720 }, { "epoch": 0.524187725631769, "grad_norm": 1.2171800136566162, "learning_rate": 9.737906137184117e-05, "loss": 2.0082, "step": 726 }, { "epoch": 0.5285198555956678, "grad_norm": 1.1622673273086548, "learning_rate": 9.735740072202167e-05, "loss": 2.028, "step": 732 }, { "epoch": 0.5328519855595668, "grad_norm": 1.1388583183288574, "learning_rate": 9.733574007220217e-05, "loss": 2.0015, "step": 738 }, { "epoch": 0.5371841155234657, "grad_norm": 1.1073946952819824, "learning_rate": 9.731407942238268e-05, "loss": 2.0479, "step": 744 }, { "epoch": 0.5415162454873647, "grad_norm": 1.118569254875183, "learning_rate": 9.729241877256318e-05, "loss": 1.9313, "step": 750 }, { "epoch": 0.5458483754512635, "grad_norm": 1.1694623231887817, "learning_rate": 9.727075812274368e-05, "loss": 1.9148, "step": 756 }, { "epoch": 0.5501805054151625, "grad_norm": 1.1155989170074463, "learning_rate": 9.724909747292419e-05, "loss": 1.9899, "step": 762 }, { "epoch": 0.5545126353790614, "grad_norm": 1.2128525972366333, "learning_rate": 9.72274368231047e-05, "loss": 1.9564, "step": 768 }, { "epoch": 0.5588447653429602, "grad_norm": 1.4020377397537231, "learning_rate": 9.72057761732852e-05, "loss": 2.0213, "step": 774 }, { "epoch": 0.5631768953068592, "grad_norm": 1.125792384147644, "learning_rate": 9.718411552346571e-05, "loss": 2.0906, "step": 780 }, { "epoch": 0.5675090252707581, "grad_norm": 1.1749944686889648, "learning_rate": 9.716245487364621e-05, "loss": 2.0595, "step": 786 }, { "epoch": 0.5718411552346571, "grad_norm": 1.1687344312667847, "learning_rate": 9.714079422382672e-05, "loss": 2.0686, "step": 792 }, { "epoch": 0.5761732851985559, "grad_norm": 1.2126874923706055, "learning_rate": 9.711913357400723e-05, "loss": 2.0585, "step": 798 }, { "epoch": 0.5805054151624549, "grad_norm": 1.2012944221496582, "learning_rate": 9.709747292418772e-05, "loss": 1.8554, "step": 804 }, { "epoch": 0.5848375451263538, "grad_norm": 1.2034316062927246, "learning_rate": 9.707581227436824e-05, "loss": 2.0038, "step": 810 }, { "epoch": 0.5891696750902528, "grad_norm": 1.3076980113983154, "learning_rate": 9.705415162454874e-05, "loss": 2.0851, "step": 816 }, { "epoch": 0.5935018050541516, "grad_norm": 1.1185672283172607, "learning_rate": 9.703249097472924e-05, "loss": 1.9179, "step": 822 }, { "epoch": 0.5978339350180505, "grad_norm": 1.1259865760803223, "learning_rate": 9.701083032490975e-05, "loss": 2.017, "step": 828 }, { "epoch": 0.6021660649819495, "grad_norm": 1.31527578830719, "learning_rate": 9.698916967509025e-05, "loss": 2.0019, "step": 834 }, { "epoch": 0.6064981949458483, "grad_norm": 1.1191191673278809, "learning_rate": 9.696750902527077e-05, "loss": 2.0379, "step": 840 }, { "epoch": 0.6108303249097473, "grad_norm": 1.1169673204421997, "learning_rate": 9.694584837545127e-05, "loss": 2.0543, "step": 846 }, { "epoch": 0.6151624548736462, "grad_norm": 1.1496541500091553, "learning_rate": 9.692418772563177e-05, "loss": 1.8252, "step": 852 }, { "epoch": 0.6194945848375452, "grad_norm": 1.236804723739624, "learning_rate": 9.690252707581228e-05, "loss": 2.0078, "step": 858 }, { "epoch": 0.623826714801444, "grad_norm": 1.1014999151229858, "learning_rate": 9.68808664259928e-05, "loss": 2.1038, "step": 864 }, { "epoch": 0.628158844765343, "grad_norm": 1.2501580715179443, "learning_rate": 9.685920577617328e-05, "loss": 2.155, "step": 870 }, { "epoch": 0.6324909747292419, "grad_norm": 1.2189114093780518, "learning_rate": 9.683754512635379e-05, "loss": 2.0065, "step": 876 }, { "epoch": 0.6368231046931407, "grad_norm": 1.159293293952942, "learning_rate": 9.68158844765343e-05, "loss": 1.966, "step": 882 }, { "epoch": 0.6411552346570397, "grad_norm": 1.1827335357666016, "learning_rate": 9.679422382671481e-05, "loss": 2.07, "step": 888 }, { "epoch": 0.6454873646209386, "grad_norm": 1.1940498352050781, "learning_rate": 9.677256317689531e-05, "loss": 2.0417, "step": 894 }, { "epoch": 0.6498194945848376, "grad_norm": 1.18551766872406, "learning_rate": 9.675090252707581e-05, "loss": 2.002, "step": 900 }, { "epoch": 0.6541516245487364, "grad_norm": 1.1575868129730225, "learning_rate": 9.672924187725633e-05, "loss": 2.0083, "step": 906 }, { "epoch": 0.6584837545126354, "grad_norm": 1.178809642791748, "learning_rate": 9.670758122743683e-05, "loss": 2.1272, "step": 912 }, { "epoch": 0.6628158844765343, "grad_norm": 1.2386428117752075, "learning_rate": 9.668592057761732e-05, "loss": 2.1054, "step": 918 }, { "epoch": 0.6671480144404333, "grad_norm": 1.1959928274154663, "learning_rate": 9.666425992779784e-05, "loss": 1.942, "step": 924 }, { "epoch": 0.6714801444043321, "grad_norm": 1.2338539361953735, "learning_rate": 9.664259927797834e-05, "loss": 1.9534, "step": 930 }, { "epoch": 0.675812274368231, "grad_norm": 1.2257753610610962, "learning_rate": 9.662093862815885e-05, "loss": 2.017, "step": 936 }, { "epoch": 0.68014440433213, "grad_norm": 1.3591161966323853, "learning_rate": 9.659927797833935e-05, "loss": 2.0901, "step": 942 }, { "epoch": 0.6844765342960288, "grad_norm": 1.1570501327514648, "learning_rate": 9.657761732851987e-05, "loss": 2.0913, "step": 948 }, { "epoch": 0.6888086642599278, "grad_norm": 1.1551520824432373, "learning_rate": 9.655595667870037e-05, "loss": 1.8852, "step": 954 }, { "epoch": 0.6931407942238267, "grad_norm": 1.3439853191375732, "learning_rate": 9.653429602888086e-05, "loss": 1.9244, "step": 960 }, { "epoch": 0.6974729241877257, "grad_norm": 1.3010165691375732, "learning_rate": 9.651263537906138e-05, "loss": 2.0164, "step": 966 }, { "epoch": 0.7018050541516245, "grad_norm": 1.2460235357284546, "learning_rate": 9.649097472924188e-05, "loss": 2.0116, "step": 972 }, { "epoch": 0.7061371841155235, "grad_norm": 1.3394955396652222, "learning_rate": 9.64693140794224e-05, "loss": 2.0993, "step": 978 }, { "epoch": 0.7104693140794224, "grad_norm": 1.4588391780853271, "learning_rate": 9.644765342960289e-05, "loss": 1.9974, "step": 984 }, { "epoch": 0.7148014440433214, "grad_norm": 1.1528292894363403, "learning_rate": 9.64259927797834e-05, "loss": 2.0257, "step": 990 }, { "epoch": 0.7191335740072202, "grad_norm": 1.1834160089492798, "learning_rate": 9.64043321299639e-05, "loss": 2.0899, "step": 996 }, { "epoch": 0.7234657039711191, "grad_norm": 1.1591521501541138, "learning_rate": 9.638267148014441e-05, "loss": 1.816, "step": 1002 }, { "epoch": 0.7277978339350181, "grad_norm": 1.1841281652450562, "learning_rate": 9.636101083032491e-05, "loss": 2.1597, "step": 1008 }, { "epoch": 0.7321299638989169, "grad_norm": 1.1212106943130493, "learning_rate": 9.633935018050542e-05, "loss": 1.9621, "step": 1014 }, { "epoch": 0.7364620938628159, "grad_norm": 1.1139216423034668, "learning_rate": 9.631768953068593e-05, "loss": 2.0353, "step": 1020 }, { "epoch": 0.7407942238267148, "grad_norm": 1.31769859790802, "learning_rate": 9.629602888086644e-05, "loss": 1.9996, "step": 1026 }, { "epoch": 0.7451263537906138, "grad_norm": 1.1716049909591675, "learning_rate": 9.627436823104694e-05, "loss": 1.8254, "step": 1032 }, { "epoch": 0.7494584837545126, "grad_norm": 1.7301291227340698, "learning_rate": 9.625270758122744e-05, "loss": 1.8714, "step": 1038 }, { "epoch": 0.7537906137184116, "grad_norm": 1.014140248298645, "learning_rate": 9.623104693140795e-05, "loss": 1.8709, "step": 1044 }, { "epoch": 0.7581227436823105, "grad_norm": 1.2106646299362183, "learning_rate": 9.620938628158845e-05, "loss": 2.0586, "step": 1050 }, { "epoch": 0.7624548736462093, "grad_norm": 1.1773916482925415, "learning_rate": 9.618772563176895e-05, "loss": 2.0346, "step": 1056 }, { "epoch": 0.7667870036101083, "grad_norm": 1.2546076774597168, "learning_rate": 9.616606498194947e-05, "loss": 2.0213, "step": 1062 }, { "epoch": 0.7711191335740072, "grad_norm": 1.2344430685043335, "learning_rate": 9.614440433212997e-05, "loss": 2.0278, "step": 1068 }, { "epoch": 0.7754512635379062, "grad_norm": 1.3171497583389282, "learning_rate": 9.612274368231048e-05, "loss": 1.9998, "step": 1074 }, { "epoch": 0.779783393501805, "grad_norm": 1.2410316467285156, "learning_rate": 9.610108303249098e-05, "loss": 2.0032, "step": 1080 }, { "epoch": 0.784115523465704, "grad_norm": 1.1531325578689575, "learning_rate": 9.607942238267148e-05, "loss": 2.0195, "step": 1086 }, { "epoch": 0.7884476534296029, "grad_norm": 1.141191840171814, "learning_rate": 9.6057761732852e-05, "loss": 2.0206, "step": 1092 }, { "epoch": 0.7927797833935019, "grad_norm": 1.2847638130187988, "learning_rate": 9.603610108303249e-05, "loss": 1.9647, "step": 1098 }, { "epoch": 0.7971119133574007, "grad_norm": 1.2062815427780151, "learning_rate": 9.6014440433213e-05, "loss": 2.0029, "step": 1104 }, { "epoch": 0.8014440433212996, "grad_norm": 1.2531037330627441, "learning_rate": 9.599277978339351e-05, "loss": 1.9356, "step": 1110 }, { "epoch": 0.8057761732851986, "grad_norm": 1.12994384765625, "learning_rate": 9.597111913357401e-05, "loss": 1.9396, "step": 1116 }, { "epoch": 0.8101083032490974, "grad_norm": 1.1987286806106567, "learning_rate": 9.594945848375451e-05, "loss": 2.1702, "step": 1122 }, { "epoch": 0.8144404332129964, "grad_norm": 1.1836192607879639, "learning_rate": 9.592779783393502e-05, "loss": 1.8822, "step": 1128 }, { "epoch": 0.8187725631768953, "grad_norm": 1.3749895095825195, "learning_rate": 9.590613718411553e-05, "loss": 2.0836, "step": 1134 }, { "epoch": 0.8231046931407943, "grad_norm": 1.214987874031067, "learning_rate": 9.588447653429602e-05, "loss": 2.1629, "step": 1140 }, { "epoch": 0.8274368231046931, "grad_norm": 1.2281479835510254, "learning_rate": 9.586281588447654e-05, "loss": 2.0441, "step": 1146 }, { "epoch": 0.8317689530685921, "grad_norm": 1.1435974836349487, "learning_rate": 9.584115523465704e-05, "loss": 2.0224, "step": 1152 }, { "epoch": 0.836101083032491, "grad_norm": 1.2288541793823242, "learning_rate": 9.581949458483756e-05, "loss": 2.0462, "step": 1158 }, { "epoch": 0.8404332129963898, "grad_norm": 1.2351398468017578, "learning_rate": 9.579783393501805e-05, "loss": 1.9496, "step": 1164 }, { "epoch": 0.8447653429602888, "grad_norm": 1.2272454500198364, "learning_rate": 9.577617328519855e-05, "loss": 2.0906, "step": 1170 }, { "epoch": 0.8490974729241877, "grad_norm": 1.248745322227478, "learning_rate": 9.575451263537907e-05, "loss": 2.0224, "step": 1176 }, { "epoch": 0.8534296028880867, "grad_norm": 1.1659197807312012, "learning_rate": 9.573285198555957e-05, "loss": 1.9237, "step": 1182 }, { "epoch": 0.8577617328519855, "grad_norm": 1.3459186553955078, "learning_rate": 9.571119133574008e-05, "loss": 2.0447, "step": 1188 }, { "epoch": 0.8620938628158845, "grad_norm": 1.1550065279006958, "learning_rate": 9.568953068592058e-05, "loss": 1.9671, "step": 1194 }, { "epoch": 0.8664259927797834, "grad_norm": 1.1686569452285767, "learning_rate": 9.56678700361011e-05, "loss": 1.9366, "step": 1200 }, { "epoch": 0.8707581227436824, "grad_norm": 1.2944872379302979, "learning_rate": 9.564620938628159e-05, "loss": 1.9976, "step": 1206 }, { "epoch": 0.8750902527075812, "grad_norm": 1.165549874305725, "learning_rate": 9.562454873646209e-05, "loss": 1.8846, "step": 1212 }, { "epoch": 0.8794223826714801, "grad_norm": 1.1984751224517822, "learning_rate": 9.56028880866426e-05, "loss": 2.1572, "step": 1218 }, { "epoch": 0.8837545126353791, "grad_norm": 1.136785864830017, "learning_rate": 9.558122743682311e-05, "loss": 1.9633, "step": 1224 }, { "epoch": 0.8880866425992779, "grad_norm": 1.1618266105651855, "learning_rate": 9.555956678700361e-05, "loss": 2.2701, "step": 1230 }, { "epoch": 0.8924187725631769, "grad_norm": 1.1402859687805176, "learning_rate": 9.553790613718412e-05, "loss": 1.8744, "step": 1236 }, { "epoch": 0.8967509025270758, "grad_norm": 1.2060097455978394, "learning_rate": 9.551624548736463e-05, "loss": 2.1191, "step": 1242 }, { "epoch": 0.9010830324909748, "grad_norm": 1.1903290748596191, "learning_rate": 9.549458483754514e-05, "loss": 2.0506, "step": 1248 }, { "epoch": 0.9054151624548736, "grad_norm": 1.2682914733886719, "learning_rate": 9.547292418772563e-05, "loss": 1.9598, "step": 1254 }, { "epoch": 0.9097472924187726, "grad_norm": 1.2742027044296265, "learning_rate": 9.545126353790614e-05, "loss": 1.9782, "step": 1260 }, { "epoch": 0.9140794223826715, "grad_norm": 1.241673231124878, "learning_rate": 9.542960288808665e-05, "loss": 1.9066, "step": 1266 }, { "epoch": 0.9184115523465703, "grad_norm": 1.171004056930542, "learning_rate": 9.540794223826716e-05, "loss": 1.923, "step": 1272 }, { "epoch": 0.9227436823104693, "grad_norm": 1.213433027267456, "learning_rate": 9.538628158844765e-05, "loss": 2.0032, "step": 1278 }, { "epoch": 0.9270758122743682, "grad_norm": 1.2507683038711548, "learning_rate": 9.536462093862817e-05, "loss": 1.9679, "step": 1284 }, { "epoch": 0.9314079422382672, "grad_norm": 1.338147759437561, "learning_rate": 9.534296028880867e-05, "loss": 2.0193, "step": 1290 }, { "epoch": 0.935740072202166, "grad_norm": 1.1887259483337402, "learning_rate": 9.532129963898918e-05, "loss": 2.092, "step": 1296 }, { "epoch": 0.940072202166065, "grad_norm": 1.2374497652053833, "learning_rate": 9.529963898916968e-05, "loss": 2.0363, "step": 1302 }, { "epoch": 0.9444043321299639, "grad_norm": 1.1295220851898193, "learning_rate": 9.527797833935018e-05, "loss": 1.9713, "step": 1308 }, { "epoch": 0.9487364620938629, "grad_norm": 1.1744009256362915, "learning_rate": 9.52563176895307e-05, "loss": 2.1263, "step": 1314 }, { "epoch": 0.9530685920577617, "grad_norm": 1.1340572834014893, "learning_rate": 9.523465703971119e-05, "loss": 2.006, "step": 1320 }, { "epoch": 0.9574007220216606, "grad_norm": 1.376057505607605, "learning_rate": 9.52129963898917e-05, "loss": 2.1207, "step": 1326 }, { "epoch": 0.9617328519855596, "grad_norm": 1.2292835712432861, "learning_rate": 9.519133574007221e-05, "loss": 2.0226, "step": 1332 }, { "epoch": 0.9660649819494584, "grad_norm": 1.3123277425765991, "learning_rate": 9.516967509025271e-05, "loss": 1.9821, "step": 1338 }, { "epoch": 0.9703971119133574, "grad_norm": 1.2093147039413452, "learning_rate": 9.514801444043322e-05, "loss": 1.8834, "step": 1344 }, { "epoch": 0.9747292418772563, "grad_norm": 1.2096234560012817, "learning_rate": 9.512635379061372e-05, "loss": 1.9443, "step": 1350 }, { "epoch": 0.9790613718411553, "grad_norm": 1.186706304550171, "learning_rate": 9.510469314079424e-05, "loss": 2.134, "step": 1356 }, { "epoch": 0.9833935018050541, "grad_norm": 1.2176740169525146, "learning_rate": 9.508303249097474e-05, "loss": 1.8772, "step": 1362 }, { "epoch": 0.9877256317689531, "grad_norm": 1.220944881439209, "learning_rate": 9.506137184115524e-05, "loss": 2.0044, "step": 1368 }, { "epoch": 0.992057761732852, "grad_norm": 1.25131094455719, "learning_rate": 9.503971119133574e-05, "loss": 2.0716, "step": 1374 }, { "epoch": 0.9963898916967509, "grad_norm": 1.2535934448242188, "learning_rate": 9.501805054151625e-05, "loss": 2.039, "step": 1380 }, { "epoch": 1.0007220216606498, "grad_norm": 1.2755100727081299, "learning_rate": 9.499638989169675e-05, "loss": 2.014, "step": 1386 }, { "epoch": 1.0050541516245488, "grad_norm": 1.2141451835632324, "learning_rate": 9.497472924187725e-05, "loss": 2.0238, "step": 1392 }, { "epoch": 1.0093862815884476, "grad_norm": 1.2714251279830933, "learning_rate": 9.495306859205777e-05, "loss": 1.9093, "step": 1398 }, { "epoch": 1.0137184115523465, "grad_norm": 1.1801530122756958, "learning_rate": 9.493140794223827e-05, "loss": 1.8519, "step": 1404 }, { "epoch": 1.0180505415162455, "grad_norm": 1.3398112058639526, "learning_rate": 9.490974729241878e-05, "loss": 1.8443, "step": 1410 }, { "epoch": 1.0223826714801445, "grad_norm": 1.241417407989502, "learning_rate": 9.488808664259928e-05, "loss": 1.8288, "step": 1416 }, { "epoch": 1.0267148014440433, "grad_norm": 1.2556840181350708, "learning_rate": 9.486642599277978e-05, "loss": 1.8862, "step": 1422 }, { "epoch": 1.0310469314079422, "grad_norm": 1.4902241230010986, "learning_rate": 9.48447653429603e-05, "loss": 1.8716, "step": 1428 }, { "epoch": 1.0353790613718412, "grad_norm": 1.161116123199463, "learning_rate": 9.482310469314079e-05, "loss": 1.9243, "step": 1434 }, { "epoch": 1.03971119133574, "grad_norm": 1.232774019241333, "learning_rate": 9.480144404332131e-05, "loss": 2.0974, "step": 1440 }, { "epoch": 1.044043321299639, "grad_norm": 1.2913951873779297, "learning_rate": 9.477978339350181e-05, "loss": 1.8333, "step": 1446 }, { "epoch": 1.048375451263538, "grad_norm": 1.1886059045791626, "learning_rate": 9.475812274368231e-05, "loss": 1.9832, "step": 1452 }, { "epoch": 1.052707581227437, "grad_norm": 1.2127817869186401, "learning_rate": 9.473646209386282e-05, "loss": 1.9609, "step": 1458 }, { "epoch": 1.0570397111913357, "grad_norm": 1.3016777038574219, "learning_rate": 9.471480144404332e-05, "loss": 1.9077, "step": 1464 }, { "epoch": 1.0613718411552346, "grad_norm": 1.4022756814956665, "learning_rate": 9.469314079422384e-05, "loss": 1.9374, "step": 1470 }, { "epoch": 1.0657039711191336, "grad_norm": 1.1350409984588623, "learning_rate": 9.467148014440434e-05, "loss": 1.7653, "step": 1476 }, { "epoch": 1.0700361010830326, "grad_norm": 1.2347087860107422, "learning_rate": 9.464981949458484e-05, "loss": 1.9825, "step": 1482 }, { "epoch": 1.0743682310469314, "grad_norm": 1.2201688289642334, "learning_rate": 9.462815884476535e-05, "loss": 1.8954, "step": 1488 }, { "epoch": 1.0787003610108303, "grad_norm": 1.4032177925109863, "learning_rate": 9.460649819494586e-05, "loss": 1.9189, "step": 1494 }, { "epoch": 1.0830324909747293, "grad_norm": 1.320712685585022, "learning_rate": 9.458483754512635e-05, "loss": 2.0273, "step": 1500 }, { "epoch": 1.087364620938628, "grad_norm": 1.2774354219436646, "learning_rate": 9.456317689530686e-05, "loss": 1.8891, "step": 1506 }, { "epoch": 1.091696750902527, "grad_norm": 1.3788686990737915, "learning_rate": 9.454151624548737e-05, "loss": 1.8306, "step": 1512 }, { "epoch": 1.096028880866426, "grad_norm": 1.1408787965774536, "learning_rate": 9.451985559566788e-05, "loss": 1.8413, "step": 1518 }, { "epoch": 1.100361010830325, "grad_norm": 1.23151695728302, "learning_rate": 9.449819494584838e-05, "loss": 1.9466, "step": 1524 }, { "epoch": 1.1046931407942238, "grad_norm": 1.396835207939148, "learning_rate": 9.447653429602888e-05, "loss": 1.9767, "step": 1530 }, { "epoch": 1.1090252707581227, "grad_norm": 1.3142069578170776, "learning_rate": 9.445487364620939e-05, "loss": 2.0415, "step": 1536 }, { "epoch": 1.1133574007220217, "grad_norm": 1.2366503477096558, "learning_rate": 9.44332129963899e-05, "loss": 2.0136, "step": 1542 }, { "epoch": 1.1176895306859205, "grad_norm": 1.2139068841934204, "learning_rate": 9.441155234657039e-05, "loss": 1.9694, "step": 1548 }, { "epoch": 1.1220216606498195, "grad_norm": 1.3151137828826904, "learning_rate": 9.438989169675091e-05, "loss": 1.8374, "step": 1554 }, { "epoch": 1.1263537906137184, "grad_norm": 1.3529072999954224, "learning_rate": 9.436823104693141e-05, "loss": 1.9917, "step": 1560 }, { "epoch": 1.1306859205776174, "grad_norm": 1.2185949087142944, "learning_rate": 9.434657039711192e-05, "loss": 1.9176, "step": 1566 }, { "epoch": 1.1350180505415162, "grad_norm": 1.370305061340332, "learning_rate": 9.432490974729242e-05, "loss": 1.9232, "step": 1572 }, { "epoch": 1.1393501805054151, "grad_norm": 1.333575963973999, "learning_rate": 9.430324909747292e-05, "loss": 2.0012, "step": 1578 }, { "epoch": 1.1436823104693141, "grad_norm": 1.3591889142990112, "learning_rate": 9.428158844765344e-05, "loss": 2.0528, "step": 1584 }, { "epoch": 1.1480144404332129, "grad_norm": 1.2027857303619385, "learning_rate": 9.425992779783394e-05, "loss": 1.8401, "step": 1590 }, { "epoch": 1.1523465703971119, "grad_norm": 1.228775978088379, "learning_rate": 9.423826714801445e-05, "loss": 1.8756, "step": 1596 }, { "epoch": 1.1566787003610108, "grad_norm": 1.3899282217025757, "learning_rate": 9.421660649819495e-05, "loss": 1.9529, "step": 1602 }, { "epoch": 1.1610108303249098, "grad_norm": 1.2614527940750122, "learning_rate": 9.419494584837547e-05, "loss": 1.9193, "step": 1608 }, { "epoch": 1.1653429602888086, "grad_norm": 1.4220877885818481, "learning_rate": 9.417328519855596e-05, "loss": 1.8371, "step": 1614 }, { "epoch": 1.1696750902527075, "grad_norm": 1.2857218980789185, "learning_rate": 9.415162454873646e-05, "loss": 1.9575, "step": 1620 }, { "epoch": 1.1740072202166065, "grad_norm": 1.1691901683807373, "learning_rate": 9.412996389891698e-05, "loss": 1.86, "step": 1626 }, { "epoch": 1.1783393501805055, "grad_norm": 1.3953442573547363, "learning_rate": 9.410830324909748e-05, "loss": 1.8213, "step": 1632 }, { "epoch": 1.1826714801444043, "grad_norm": 1.3189011812210083, "learning_rate": 9.408664259927798e-05, "loss": 1.9942, "step": 1638 }, { "epoch": 1.1870036101083032, "grad_norm": 1.2242608070373535, "learning_rate": 9.406498194945848e-05, "loss": 1.8726, "step": 1644 }, { "epoch": 1.1913357400722022, "grad_norm": 1.4473179578781128, "learning_rate": 9.4043321299639e-05, "loss": 1.9162, "step": 1650 }, { "epoch": 1.1956678700361012, "grad_norm": 1.2710192203521729, "learning_rate": 9.40216606498195e-05, "loss": 1.9565, "step": 1656 }, { "epoch": 1.2, "grad_norm": 1.3035579919815063, "learning_rate": 9.4e-05, "loss": 2.042, "step": 1662 }, { "epoch": 1.204332129963899, "grad_norm": 1.344599723815918, "learning_rate": 9.397833935018051e-05, "loss": 1.9206, "step": 1668 }, { "epoch": 1.208664259927798, "grad_norm": 1.2555919885635376, "learning_rate": 9.395667870036101e-05, "loss": 1.9444, "step": 1674 }, { "epoch": 1.2129963898916967, "grad_norm": 1.2901116609573364, "learning_rate": 9.393501805054152e-05, "loss": 1.8836, "step": 1680 }, { "epoch": 1.2173285198555956, "grad_norm": 1.2449308633804321, "learning_rate": 9.391335740072202e-05, "loss": 1.9751, "step": 1686 }, { "epoch": 1.2216606498194946, "grad_norm": 1.2739218473434448, "learning_rate": 9.389169675090254e-05, "loss": 1.7649, "step": 1692 }, { "epoch": 1.2259927797833936, "grad_norm": 1.3208640813827515, "learning_rate": 9.387003610108304e-05, "loss": 1.8745, "step": 1698 }, { "epoch": 1.2303249097472924, "grad_norm": 1.2186170816421509, "learning_rate": 9.384837545126353e-05, "loss": 1.8014, "step": 1704 }, { "epoch": 1.2346570397111913, "grad_norm": 1.3435767889022827, "learning_rate": 9.382671480144405e-05, "loss": 1.9541, "step": 1710 }, { "epoch": 1.2389891696750903, "grad_norm": 1.2266908884048462, "learning_rate": 9.380505415162455e-05, "loss": 1.9695, "step": 1716 }, { "epoch": 1.243321299638989, "grad_norm": 1.2332193851470947, "learning_rate": 9.378339350180507e-05, "loss": 1.9336, "step": 1722 }, { "epoch": 1.247653429602888, "grad_norm": 1.2644178867340088, "learning_rate": 9.376173285198556e-05, "loss": 1.8651, "step": 1728 }, { "epoch": 1.251985559566787, "grad_norm": 1.351762294769287, "learning_rate": 9.374007220216607e-05, "loss": 2.0106, "step": 1734 }, { "epoch": 1.256317689530686, "grad_norm": 1.2501542568206787, "learning_rate": 9.371841155234658e-05, "loss": 2.0765, "step": 1740 }, { "epoch": 1.2606498194945848, "grad_norm": 1.19627046585083, "learning_rate": 9.369675090252708e-05, "loss": 1.9411, "step": 1746 }, { "epoch": 1.2649819494584837, "grad_norm": 1.2568250894546509, "learning_rate": 9.367509025270758e-05, "loss": 1.7795, "step": 1752 }, { "epoch": 1.2693140794223827, "grad_norm": 1.3179750442504883, "learning_rate": 9.365342960288809e-05, "loss": 1.9808, "step": 1758 }, { "epoch": 1.2736462093862815, "grad_norm": 1.504164695739746, "learning_rate": 9.36317689530686e-05, "loss": 1.942, "step": 1764 }, { "epoch": 1.2779783393501805, "grad_norm": 1.2818363904953003, "learning_rate": 9.36101083032491e-05, "loss": 1.7266, "step": 1770 }, { "epoch": 1.2823104693140794, "grad_norm": 1.221922755241394, "learning_rate": 9.358844765342961e-05, "loss": 1.9315, "step": 1776 }, { "epoch": 1.2866425992779784, "grad_norm": 1.4442038536071777, "learning_rate": 9.356678700361011e-05, "loss": 1.8555, "step": 1782 }, { "epoch": 1.2909747292418774, "grad_norm": 1.4088420867919922, "learning_rate": 9.354512635379062e-05, "loss": 1.7873, "step": 1788 }, { "epoch": 1.2953068592057762, "grad_norm": 1.255726933479309, "learning_rate": 9.352346570397112e-05, "loss": 1.8974, "step": 1794 }, { "epoch": 1.2996389891696751, "grad_norm": 1.2514727115631104, "learning_rate": 9.350180505415162e-05, "loss": 1.8451, "step": 1800 }, { "epoch": 1.3039711191335739, "grad_norm": 1.3951563835144043, "learning_rate": 9.348014440433214e-05, "loss": 1.9974, "step": 1806 }, { "epoch": 1.3083032490974729, "grad_norm": 1.6221004724502563, "learning_rate": 9.345848375451264e-05, "loss": 1.9724, "step": 1812 }, { "epoch": 1.3126353790613718, "grad_norm": 1.1781755685806274, "learning_rate": 9.343682310469315e-05, "loss": 1.833, "step": 1818 }, { "epoch": 1.3169675090252708, "grad_norm": 1.3637298345565796, "learning_rate": 9.341516245487365e-05, "loss": 1.8304, "step": 1824 }, { "epoch": 1.3212996389891698, "grad_norm": 1.455683708190918, "learning_rate": 9.339350180505415e-05, "loss": 1.9416, "step": 1830 }, { "epoch": 1.3256317689530686, "grad_norm": 1.3598400354385376, "learning_rate": 9.337184115523467e-05, "loss": 1.8714, "step": 1836 }, { "epoch": 1.3299638989169675, "grad_norm": 1.2072970867156982, "learning_rate": 9.335018050541516e-05, "loss": 1.8395, "step": 1842 }, { "epoch": 1.3342960288808663, "grad_norm": 1.3141353130340576, "learning_rate": 9.332851985559568e-05, "loss": 1.9345, "step": 1848 }, { "epoch": 1.3386281588447653, "grad_norm": 1.326292872428894, "learning_rate": 9.330685920577618e-05, "loss": 1.8088, "step": 1854 }, { "epoch": 1.3429602888086642, "grad_norm": 1.365944266319275, "learning_rate": 9.328519855595668e-05, "loss": 1.9742, "step": 1860 }, { "epoch": 1.3472924187725632, "grad_norm": 1.2215503454208374, "learning_rate": 9.326353790613719e-05, "loss": 1.9639, "step": 1866 }, { "epoch": 1.3516245487364622, "grad_norm": 1.2887612581253052, "learning_rate": 9.324187725631769e-05, "loss": 1.7591, "step": 1872 }, { "epoch": 1.355956678700361, "grad_norm": 1.4575281143188477, "learning_rate": 9.32202166064982e-05, "loss": 2.0431, "step": 1878 }, { "epoch": 1.36028880866426, "grad_norm": 1.4112777709960938, "learning_rate": 9.31985559566787e-05, "loss": 1.9848, "step": 1884 }, { "epoch": 1.364620938628159, "grad_norm": 1.2822797298431396, "learning_rate": 9.317689530685921e-05, "loss": 1.9236, "step": 1890 }, { "epoch": 1.3689530685920577, "grad_norm": 1.4721968173980713, "learning_rate": 9.315523465703972e-05, "loss": 1.9023, "step": 1896 }, { "epoch": 1.3732851985559567, "grad_norm": 1.3576946258544922, "learning_rate": 9.313357400722023e-05, "loss": 1.9435, "step": 1902 }, { "epoch": 1.3776173285198556, "grad_norm": 1.3061134815216064, "learning_rate": 9.311191335740072e-05, "loss": 1.821, "step": 1908 }, { "epoch": 1.3819494584837546, "grad_norm": 1.1953153610229492, "learning_rate": 9.309025270758122e-05, "loss": 1.8688, "step": 1914 }, { "epoch": 1.3862815884476534, "grad_norm": 1.2458395957946777, "learning_rate": 9.306859205776174e-05, "loss": 1.8953, "step": 1920 }, { "epoch": 1.3906137184115523, "grad_norm": 1.531686782836914, "learning_rate": 9.304693140794224e-05, "loss": 1.9373, "step": 1926 }, { "epoch": 1.3949458483754513, "grad_norm": 1.192873239517212, "learning_rate": 9.302527075812275e-05, "loss": 1.7467, "step": 1932 }, { "epoch": 1.39927797833935, "grad_norm": 1.2905805110931396, "learning_rate": 9.300361010830325e-05, "loss": 2.1259, "step": 1938 }, { "epoch": 1.403610108303249, "grad_norm": 1.3752537965774536, "learning_rate": 9.298194945848377e-05, "loss": 1.981, "step": 1944 }, { "epoch": 1.407942238267148, "grad_norm": 1.468444585800171, "learning_rate": 9.296028880866426e-05, "loss": 2.0496, "step": 1950 }, { "epoch": 1.412274368231047, "grad_norm": 1.2377848625183105, "learning_rate": 9.293862815884476e-05, "loss": 1.9239, "step": 1956 }, { "epoch": 1.4166064981949458, "grad_norm": 1.3616564273834229, "learning_rate": 9.291696750902528e-05, "loss": 1.8354, "step": 1962 }, { "epoch": 1.4209386281588448, "grad_norm": 1.3243838548660278, "learning_rate": 9.289530685920578e-05, "loss": 2.051, "step": 1968 }, { "epoch": 1.4252707581227437, "grad_norm": 1.3544048070907593, "learning_rate": 9.287364620938628e-05, "loss": 1.816, "step": 1974 }, { "epoch": 1.4296028880866425, "grad_norm": 1.366278886795044, "learning_rate": 9.285198555956679e-05, "loss": 1.8745, "step": 1980 }, { "epoch": 1.4339350180505415, "grad_norm": 1.2942559719085693, "learning_rate": 9.28303249097473e-05, "loss": 1.9236, "step": 1986 }, { "epoch": 1.4382671480144404, "grad_norm": 1.3861033916473389, "learning_rate": 9.280866425992781e-05, "loss": 1.9159, "step": 1992 }, { "epoch": 1.4425992779783394, "grad_norm": 1.421248197555542, "learning_rate": 9.27870036101083e-05, "loss": 1.922, "step": 1998 }, { "epoch": 1.4469314079422384, "grad_norm": 1.4333752393722534, "learning_rate": 9.276534296028881e-05, "loss": 1.8728, "step": 2004 }, { "epoch": 1.4512635379061372, "grad_norm": 1.2770198583602905, "learning_rate": 9.274368231046932e-05, "loss": 1.8797, "step": 2010 }, { "epoch": 1.4555956678700361, "grad_norm": 1.2614632844924927, "learning_rate": 9.272202166064983e-05, "loss": 1.8771, "step": 2016 }, { "epoch": 1.459927797833935, "grad_norm": 1.4901973009109497, "learning_rate": 9.270036101083032e-05, "loss": 1.8138, "step": 2022 }, { "epoch": 1.4642599277978339, "grad_norm": 1.4118350744247437, "learning_rate": 9.267870036101084e-05, "loss": 1.8602, "step": 2028 }, { "epoch": 1.4685920577617328, "grad_norm": 1.2848551273345947, "learning_rate": 9.265703971119134e-05, "loss": 1.9042, "step": 2034 }, { "epoch": 1.4729241877256318, "grad_norm": 1.3482563495635986, "learning_rate": 9.263537906137185e-05, "loss": 1.9577, "step": 2040 }, { "epoch": 1.4772563176895308, "grad_norm": 1.2884601354599, "learning_rate": 9.261371841155235e-05, "loss": 1.9122, "step": 2046 }, { "epoch": 1.4815884476534296, "grad_norm": 1.3487848043441772, "learning_rate": 9.259205776173285e-05, "loss": 1.9049, "step": 2052 }, { "epoch": 1.4859205776173285, "grad_norm": 1.3778682947158813, "learning_rate": 9.257039711191337e-05, "loss": 1.936, "step": 2058 }, { "epoch": 1.4902527075812273, "grad_norm": 1.2957931756973267, "learning_rate": 9.254873646209386e-05, "loss": 1.9486, "step": 2064 }, { "epoch": 1.4945848375451263, "grad_norm": 1.56313955783844, "learning_rate": 9.252707581227438e-05, "loss": 1.8609, "step": 2070 }, { "epoch": 1.4989169675090253, "grad_norm": 1.3535372018814087, "learning_rate": 9.250541516245488e-05, "loss": 1.9203, "step": 2076 }, { "epoch": 1.5032490974729242, "grad_norm": 1.4306648969650269, "learning_rate": 9.248375451263538e-05, "loss": 1.9128, "step": 2082 }, { "epoch": 1.5075812274368232, "grad_norm": 1.3628528118133545, "learning_rate": 9.246209386281589e-05, "loss": 1.9285, "step": 2088 }, { "epoch": 1.511913357400722, "grad_norm": 1.3860634565353394, "learning_rate": 9.244043321299639e-05, "loss": 1.971, "step": 2094 }, { "epoch": 1.516245487364621, "grad_norm": 1.4176313877105713, "learning_rate": 9.24187725631769e-05, "loss": 1.9204, "step": 2100 }, { "epoch": 1.5205776173285197, "grad_norm": 1.3192576169967651, "learning_rate": 9.239711191335741e-05, "loss": 2.022, "step": 2106 }, { "epoch": 1.5249097472924187, "grad_norm": 1.3229833841323853, "learning_rate": 9.237545126353791e-05, "loss": 1.8993, "step": 2112 }, { "epoch": 1.5292418772563177, "grad_norm": 1.303503155708313, "learning_rate": 9.235379061371842e-05, "loss": 1.7945, "step": 2118 }, { "epoch": 1.5335740072202166, "grad_norm": 1.2761913537979126, "learning_rate": 9.233212996389892e-05, "loss": 1.7972, "step": 2124 }, { "epoch": 1.5379061371841156, "grad_norm": 1.2907836437225342, "learning_rate": 9.231046931407942e-05, "loss": 1.8863, "step": 2130 }, { "epoch": 1.5422382671480146, "grad_norm": 1.275059461593628, "learning_rate": 9.228880866425993e-05, "loss": 1.8729, "step": 2136 }, { "epoch": 1.5465703971119134, "grad_norm": 1.4963499307632446, "learning_rate": 9.226714801444044e-05, "loss": 1.8299, "step": 2142 }, { "epoch": 1.550902527075812, "grad_norm": 1.3415898084640503, "learning_rate": 9.224548736462095e-05, "loss": 1.9118, "step": 2148 }, { "epoch": 1.555234657039711, "grad_norm": 1.3052208423614502, "learning_rate": 9.222382671480145e-05, "loss": 1.9045, "step": 2154 }, { "epoch": 1.55956678700361, "grad_norm": 1.2212722301483154, "learning_rate": 9.220216606498195e-05, "loss": 1.8814, "step": 2160 }, { "epoch": 1.563898916967509, "grad_norm": 1.2563961744308472, "learning_rate": 9.218050541516246e-05, "loss": 2.0174, "step": 2166 }, { "epoch": 1.568231046931408, "grad_norm": 1.2717846632003784, "learning_rate": 9.215884476534297e-05, "loss": 1.7432, "step": 2172 }, { "epoch": 1.572563176895307, "grad_norm": 1.3389899730682373, "learning_rate": 9.213718411552346e-05, "loss": 1.8488, "step": 2178 }, { "epoch": 1.5768953068592058, "grad_norm": 1.2358067035675049, "learning_rate": 9.211552346570398e-05, "loss": 1.9089, "step": 2184 }, { "epoch": 1.5812274368231047, "grad_norm": 1.283457636833191, "learning_rate": 9.209386281588448e-05, "loss": 1.9667, "step": 2190 }, { "epoch": 1.5855595667870035, "grad_norm": 1.2990509271621704, "learning_rate": 9.207220216606498e-05, "loss": 1.7825, "step": 2196 }, { "epoch": 1.5898916967509025, "grad_norm": 1.3877605199813843, "learning_rate": 9.205054151624549e-05, "loss": 1.9402, "step": 2202 }, { "epoch": 1.5942238267148015, "grad_norm": 1.3258841037750244, "learning_rate": 9.202888086642599e-05, "loss": 1.8899, "step": 2208 }, { "epoch": 1.5985559566787004, "grad_norm": 1.2759488821029663, "learning_rate": 9.200722021660651e-05, "loss": 1.7603, "step": 2214 }, { "epoch": 1.6028880866425994, "grad_norm": 1.3308206796646118, "learning_rate": 9.198555956678701e-05, "loss": 1.873, "step": 2220 }, { "epoch": 1.6072202166064982, "grad_norm": 1.3291231393814087, "learning_rate": 9.196389891696751e-05, "loss": 1.9033, "step": 2226 }, { "epoch": 1.6115523465703971, "grad_norm": 1.3440351486206055, "learning_rate": 9.194223826714802e-05, "loss": 1.888, "step": 2232 }, { "epoch": 1.615884476534296, "grad_norm": 1.3693597316741943, "learning_rate": 9.192057761732853e-05, "loss": 1.8977, "step": 2238 }, { "epoch": 1.6202166064981949, "grad_norm": 1.3133327960968018, "learning_rate": 9.189891696750902e-05, "loss": 1.9615, "step": 2244 }, { "epoch": 1.6245487364620939, "grad_norm": 1.3309358358383179, "learning_rate": 9.187725631768953e-05, "loss": 1.8919, "step": 2250 }, { "epoch": 1.6288808664259928, "grad_norm": 1.4147940874099731, "learning_rate": 9.185559566787004e-05, "loss": 2.0169, "step": 2256 }, { "epoch": 1.6332129963898918, "grad_norm": 1.4273767471313477, "learning_rate": 9.183393501805055e-05, "loss": 2.059, "step": 2262 }, { "epoch": 1.6375451263537906, "grad_norm": 1.4377713203430176, "learning_rate": 9.181227436823105e-05, "loss": 1.7384, "step": 2268 }, { "epoch": 1.6418772563176895, "grad_norm": 1.2909488677978516, "learning_rate": 9.179061371841155e-05, "loss": 1.9039, "step": 2274 }, { "epoch": 1.6462093862815883, "grad_norm": 1.2630620002746582, "learning_rate": 9.176895306859207e-05, "loss": 1.7842, "step": 2280 }, { "epoch": 1.6505415162454873, "grad_norm": 1.4811687469482422, "learning_rate": 9.174729241877257e-05, "loss": 1.9603, "step": 2286 }, { "epoch": 1.6548736462093863, "grad_norm": 1.3755275011062622, "learning_rate": 9.172563176895306e-05, "loss": 1.9636, "step": 2292 }, { "epoch": 1.6592057761732852, "grad_norm": 1.2757568359375, "learning_rate": 9.170397111913358e-05, "loss": 1.7121, "step": 2298 }, { "epoch": 1.6635379061371842, "grad_norm": 1.3673440217971802, "learning_rate": 9.168231046931408e-05, "loss": 1.9408, "step": 2304 }, { "epoch": 1.6678700361010832, "grad_norm": 1.4030228853225708, "learning_rate": 9.166064981949459e-05, "loss": 1.8929, "step": 2310 }, { "epoch": 1.672202166064982, "grad_norm": 1.2310054302215576, "learning_rate": 9.163898916967509e-05, "loss": 1.7409, "step": 2316 }, { "epoch": 1.6765342960288807, "grad_norm": 1.30148446559906, "learning_rate": 9.16173285198556e-05, "loss": 2.0004, "step": 2322 }, { "epoch": 1.6808664259927797, "grad_norm": 1.2718123197555542, "learning_rate": 9.159566787003611e-05, "loss": 1.8552, "step": 2328 }, { "epoch": 1.6851985559566787, "grad_norm": 1.1873555183410645, "learning_rate": 9.157400722021661e-05, "loss": 1.8867, "step": 2334 }, { "epoch": 1.6895306859205776, "grad_norm": 1.3232542276382446, "learning_rate": 9.155234657039712e-05, "loss": 1.8476, "step": 2340 }, { "epoch": 1.6938628158844766, "grad_norm": 1.264496922492981, "learning_rate": 9.153068592057762e-05, "loss": 1.854, "step": 2346 }, { "epoch": 1.6981949458483756, "grad_norm": 1.275537133216858, "learning_rate": 9.150902527075814e-05, "loss": 1.842, "step": 2352 }, { "epoch": 1.7025270758122744, "grad_norm": 1.3770458698272705, "learning_rate": 9.148736462093863e-05, "loss": 1.8246, "step": 2358 }, { "epoch": 1.7068592057761733, "grad_norm": 1.2376651763916016, "learning_rate": 9.146570397111914e-05, "loss": 1.9399, "step": 2364 }, { "epoch": 1.711191335740072, "grad_norm": 1.2842133045196533, "learning_rate": 9.144404332129965e-05, "loss": 1.9039, "step": 2370 }, { "epoch": 1.715523465703971, "grad_norm": 1.6136068105697632, "learning_rate": 9.142238267148015e-05, "loss": 1.9108, "step": 2376 }, { "epoch": 1.71985559566787, "grad_norm": 1.64150071144104, "learning_rate": 9.140072202166065e-05, "loss": 1.8083, "step": 2382 }, { "epoch": 1.724187725631769, "grad_norm": 1.4390571117401123, "learning_rate": 9.137906137184116e-05, "loss": 1.9031, "step": 2388 }, { "epoch": 1.728519855595668, "grad_norm": 1.3347324132919312, "learning_rate": 9.135740072202167e-05, "loss": 1.7869, "step": 2394 }, { "epoch": 1.7328519855595668, "grad_norm": 1.4086651802062988, "learning_rate": 9.133574007220218e-05, "loss": 1.7073, "step": 2400 }, { "epoch": 1.7371841155234657, "grad_norm": 1.523418664932251, "learning_rate": 9.131407942238268e-05, "loss": 1.9545, "step": 2406 }, { "epoch": 1.7415162454873645, "grad_norm": 1.2368996143341064, "learning_rate": 9.129241877256318e-05, "loss": 1.7923, "step": 2412 }, { "epoch": 1.7458483754512635, "grad_norm": 1.3054194450378418, "learning_rate": 9.127075812274369e-05, "loss": 1.9953, "step": 2418 }, { "epoch": 1.7501805054151625, "grad_norm": 1.3370704650878906, "learning_rate": 9.124909747292419e-05, "loss": 1.9025, "step": 2424 }, { "epoch": 1.7545126353790614, "grad_norm": 1.315619945526123, "learning_rate": 9.122743682310469e-05, "loss": 1.8349, "step": 2430 }, { "epoch": 1.7588447653429604, "grad_norm": 1.3070690631866455, "learning_rate": 9.120577617328521e-05, "loss": 1.8585, "step": 2436 }, { "epoch": 1.7631768953068592, "grad_norm": 1.2747665643692017, "learning_rate": 9.118411552346571e-05, "loss": 1.8521, "step": 2442 }, { "epoch": 1.7675090252707581, "grad_norm": 1.2534335851669312, "learning_rate": 9.116245487364622e-05, "loss": 1.9366, "step": 2448 }, { "epoch": 1.771841155234657, "grad_norm": 1.4453128576278687, "learning_rate": 9.114079422382672e-05, "loss": 1.7639, "step": 2454 }, { "epoch": 1.7761732851985559, "grad_norm": 1.3489632606506348, "learning_rate": 9.111913357400722e-05, "loss": 2.0262, "step": 2460 }, { "epoch": 1.7805054151624549, "grad_norm": 1.367490530014038, "learning_rate": 9.109747292418774e-05, "loss": 1.7964, "step": 2466 }, { "epoch": 1.7848375451263538, "grad_norm": 1.3858834505081177, "learning_rate": 9.107581227436823e-05, "loss": 1.9456, "step": 2472 }, { "epoch": 1.7891696750902528, "grad_norm": 1.223004937171936, "learning_rate": 9.105415162454874e-05, "loss": 1.9534, "step": 2478 }, { "epoch": 1.7935018050541516, "grad_norm": 1.2788199186325073, "learning_rate": 9.103249097472925e-05, "loss": 1.8763, "step": 2484 }, { "epoch": 1.7978339350180506, "grad_norm": 1.2852994203567505, "learning_rate": 9.101083032490975e-05, "loss": 1.9037, "step": 2490 }, { "epoch": 1.8021660649819493, "grad_norm": 1.2604200839996338, "learning_rate": 9.098916967509025e-05, "loss": 1.9646, "step": 2496 }, { "epoch": 1.8064981949458483, "grad_norm": 1.2987477779388428, "learning_rate": 9.096750902527076e-05, "loss": 1.9412, "step": 2502 }, { "epoch": 1.8108303249097473, "grad_norm": 1.3709746599197388, "learning_rate": 9.094584837545127e-05, "loss": 1.8534, "step": 2508 }, { "epoch": 1.8151624548736462, "grad_norm": 1.2641104459762573, "learning_rate": 9.092418772563176e-05, "loss": 1.8065, "step": 2514 }, { "epoch": 1.8194945848375452, "grad_norm": 1.4036431312561035, "learning_rate": 9.090252707581228e-05, "loss": 1.927, "step": 2520 }, { "epoch": 1.8238267148014442, "grad_norm": 1.3780423402786255, "learning_rate": 9.088086642599278e-05, "loss": 1.8835, "step": 2526 }, { "epoch": 1.828158844765343, "grad_norm": 1.4659676551818848, "learning_rate": 9.08592057761733e-05, "loss": 1.7986, "step": 2532 }, { "epoch": 1.8324909747292417, "grad_norm": 1.5334960222244263, "learning_rate": 9.083754512635379e-05, "loss": 1.8977, "step": 2538 }, { "epoch": 1.8368231046931407, "grad_norm": 1.247686743736267, "learning_rate": 9.08158844765343e-05, "loss": 1.9287, "step": 2544 }, { "epoch": 1.8411552346570397, "grad_norm": 1.3799062967300415, "learning_rate": 9.079422382671481e-05, "loss": 1.8705, "step": 2550 }, { "epoch": 1.8454873646209387, "grad_norm": 1.277754306793213, "learning_rate": 9.077256317689531e-05, "loss": 1.913, "step": 2556 }, { "epoch": 1.8498194945848376, "grad_norm": 1.3164188861846924, "learning_rate": 9.075090252707582e-05, "loss": 1.7444, "step": 2562 }, { "epoch": 1.8541516245487366, "grad_norm": 1.4048140048980713, "learning_rate": 9.072924187725632e-05, "loss": 1.7661, "step": 2568 }, { "epoch": 1.8584837545126354, "grad_norm": 1.391715168952942, "learning_rate": 9.070758122743682e-05, "loss": 1.74, "step": 2574 }, { "epoch": 1.8628158844765343, "grad_norm": 1.4624733924865723, "learning_rate": 9.068592057761734e-05, "loss": 1.8304, "step": 2580 }, { "epoch": 1.867148014440433, "grad_norm": 1.3607451915740967, "learning_rate": 9.066425992779783e-05, "loss": 2.0284, "step": 2586 }, { "epoch": 1.871480144404332, "grad_norm": 1.5483511686325073, "learning_rate": 9.064259927797835e-05, "loss": 1.8919, "step": 2592 }, { "epoch": 1.875812274368231, "grad_norm": 1.4410533905029297, "learning_rate": 9.062093862815885e-05, "loss": 1.9486, "step": 2598 }, { "epoch": 1.88014440433213, "grad_norm": 1.4013601541519165, "learning_rate": 9.059927797833935e-05, "loss": 1.8391, "step": 2604 }, { "epoch": 1.884476534296029, "grad_norm": 1.4957237243652344, "learning_rate": 9.057761732851986e-05, "loss": 1.8612, "step": 2610 }, { "epoch": 1.8888086642599278, "grad_norm": 1.3556795120239258, "learning_rate": 9.055595667870036e-05, "loss": 1.9386, "step": 2616 }, { "epoch": 1.8931407942238268, "grad_norm": 1.4698115587234497, "learning_rate": 9.053429602888088e-05, "loss": 1.997, "step": 2622 }, { "epoch": 1.8974729241877255, "grad_norm": 1.2698191404342651, "learning_rate": 9.051263537906137e-05, "loss": 1.8205, "step": 2628 }, { "epoch": 1.9018050541516245, "grad_norm": 1.3431501388549805, "learning_rate": 9.049097472924188e-05, "loss": 1.8836, "step": 2634 }, { "epoch": 1.9061371841155235, "grad_norm": 1.473340630531311, "learning_rate": 9.046931407942239e-05, "loss": 1.8185, "step": 2640 }, { "epoch": 1.9104693140794224, "grad_norm": 1.5168447494506836, "learning_rate": 9.04476534296029e-05, "loss": 1.9628, "step": 2646 }, { "epoch": 1.9148014440433214, "grad_norm": 1.4048134088516235, "learning_rate": 9.042599277978339e-05, "loss": 2.0115, "step": 2652 }, { "epoch": 1.9191335740072202, "grad_norm": 1.2985855340957642, "learning_rate": 9.04043321299639e-05, "loss": 1.9677, "step": 2658 }, { "epoch": 1.9234657039711192, "grad_norm": 1.2353999614715576, "learning_rate": 9.038267148014441e-05, "loss": 1.9889, "step": 2664 }, { "epoch": 1.927797833935018, "grad_norm": 1.32314133644104, "learning_rate": 9.036101083032492e-05, "loss": 1.8117, "step": 2670 }, { "epoch": 1.932129963898917, "grad_norm": 1.4457571506500244, "learning_rate": 9.033935018050542e-05, "loss": 2.0364, "step": 2676 }, { "epoch": 1.9364620938628159, "grad_norm": 1.4590137004852295, "learning_rate": 9.031768953068592e-05, "loss": 1.7373, "step": 2682 }, { "epoch": 1.9407942238267148, "grad_norm": 1.3767893314361572, "learning_rate": 9.029602888086644e-05, "loss": 2.0722, "step": 2688 }, { "epoch": 1.9451263537906138, "grad_norm": 1.3204545974731445, "learning_rate": 9.027436823104693e-05, "loss": 1.7903, "step": 2694 }, { "epoch": 1.9494584837545126, "grad_norm": 1.3157508373260498, "learning_rate": 9.025270758122743e-05, "loss": 1.9187, "step": 2700 }, { "epoch": 1.9537906137184116, "grad_norm": 1.2535635232925415, "learning_rate": 9.023104693140795e-05, "loss": 1.8674, "step": 2706 }, { "epoch": 1.9581227436823103, "grad_norm": 1.4080469608306885, "learning_rate": 9.020938628158845e-05, "loss": 1.8129, "step": 2712 }, { "epoch": 1.9624548736462093, "grad_norm": 1.2955127954483032, "learning_rate": 9.018772563176896e-05, "loss": 1.9712, "step": 2718 }, { "epoch": 1.9667870036101083, "grad_norm": 1.2763488292694092, "learning_rate": 9.016606498194946e-05, "loss": 1.8394, "step": 2724 }, { "epoch": 1.9711191335740073, "grad_norm": 1.4496939182281494, "learning_rate": 9.014440433212998e-05, "loss": 1.8482, "step": 2730 }, { "epoch": 1.9754512635379062, "grad_norm": 1.4020060300827026, "learning_rate": 9.012274368231048e-05, "loss": 1.8241, "step": 2736 }, { "epoch": 1.9797833935018052, "grad_norm": 1.425187110900879, "learning_rate": 9.010108303249097e-05, "loss": 1.8704, "step": 2742 }, { "epoch": 1.984115523465704, "grad_norm": 1.2587729692459106, "learning_rate": 9.007942238267148e-05, "loss": 1.7192, "step": 2748 }, { "epoch": 1.9884476534296027, "grad_norm": 1.4463845491409302, "learning_rate": 9.005776173285199e-05, "loss": 1.8134, "step": 2754 }, { "epoch": 1.9927797833935017, "grad_norm": 1.336258888244629, "learning_rate": 9.003610108303249e-05, "loss": 1.7518, "step": 2760 }, { "epoch": 1.9971119133574007, "grad_norm": 1.3635334968566895, "learning_rate": 9.0014440433213e-05, "loss": 1.9093, "step": 2766 }, { "epoch": 2.0014440433212997, "grad_norm": 1.3204944133758545, "learning_rate": 8.999277978339351e-05, "loss": 1.9955, "step": 2772 }, { "epoch": 2.0057761732851986, "grad_norm": 1.299351453781128, "learning_rate": 8.997111913357401e-05, "loss": 1.7993, "step": 2778 }, { "epoch": 2.0101083032490976, "grad_norm": 1.4241260290145874, "learning_rate": 8.994945848375452e-05, "loss": 1.9424, "step": 2784 }, { "epoch": 2.0144404332129966, "grad_norm": 1.2731657028198242, "learning_rate": 8.992779783393502e-05, "loss": 1.8076, "step": 2790 }, { "epoch": 2.018772563176895, "grad_norm": 1.3109437227249146, "learning_rate": 8.990613718411552e-05, "loss": 1.6988, "step": 2796 }, { "epoch": 2.023104693140794, "grad_norm": 1.3174623250961304, "learning_rate": 8.988447653429604e-05, "loss": 1.8106, "step": 2802 }, { "epoch": 2.027436823104693, "grad_norm": 1.2370026111602783, "learning_rate": 8.986281588447653e-05, "loss": 1.8164, "step": 2808 }, { "epoch": 2.031768953068592, "grad_norm": 1.5429061651229858, "learning_rate": 8.984115523465705e-05, "loss": 1.8465, "step": 2814 }, { "epoch": 2.036101083032491, "grad_norm": 1.3175725936889648, "learning_rate": 8.981949458483755e-05, "loss": 1.8006, "step": 2820 }, { "epoch": 2.04043321299639, "grad_norm": 1.3450121879577637, "learning_rate": 8.979783393501805e-05, "loss": 1.6827, "step": 2826 }, { "epoch": 2.044765342960289, "grad_norm": 1.2900246381759644, "learning_rate": 8.977617328519856e-05, "loss": 1.7363, "step": 2832 }, { "epoch": 2.0490974729241875, "grad_norm": 1.3868024349212646, "learning_rate": 8.975451263537906e-05, "loss": 1.7826, "step": 2838 }, { "epoch": 2.0534296028880865, "grad_norm": 1.456239104270935, "learning_rate": 8.973285198555958e-05, "loss": 1.8087, "step": 2844 }, { "epoch": 2.0577617328519855, "grad_norm": 1.393865704536438, "learning_rate": 8.971119133574008e-05, "loss": 1.9399, "step": 2850 }, { "epoch": 2.0620938628158845, "grad_norm": 1.307541847229004, "learning_rate": 8.968953068592058e-05, "loss": 1.8167, "step": 2856 }, { "epoch": 2.0664259927797834, "grad_norm": 1.3942677974700928, "learning_rate": 8.966787003610109e-05, "loss": 1.8816, "step": 2862 }, { "epoch": 2.0707581227436824, "grad_norm": 1.2913978099822998, "learning_rate": 8.964620938628159e-05, "loss": 1.7868, "step": 2868 }, { "epoch": 2.0750902527075814, "grad_norm": 1.4937732219696045, "learning_rate": 8.962454873646209e-05, "loss": 1.6915, "step": 2874 }, { "epoch": 2.07942238267148, "grad_norm": 1.4330472946166992, "learning_rate": 8.96028880866426e-05, "loss": 1.875, "step": 2880 }, { "epoch": 2.083754512635379, "grad_norm": 1.3210880756378174, "learning_rate": 8.958122743682311e-05, "loss": 1.7833, "step": 2886 }, { "epoch": 2.088086642599278, "grad_norm": 1.3707735538482666, "learning_rate": 8.955956678700362e-05, "loss": 1.8286, "step": 2892 }, { "epoch": 2.092418772563177, "grad_norm": 1.4147958755493164, "learning_rate": 8.953790613718412e-05, "loss": 1.9962, "step": 2898 }, { "epoch": 2.096750902527076, "grad_norm": 1.211733102798462, "learning_rate": 8.951624548736462e-05, "loss": 1.8222, "step": 2904 }, { "epoch": 2.101083032490975, "grad_norm": 1.4813019037246704, "learning_rate": 8.949458483754513e-05, "loss": 1.8782, "step": 2910 }, { "epoch": 2.105415162454874, "grad_norm": 1.3542702198028564, "learning_rate": 8.947292418772564e-05, "loss": 1.8716, "step": 2916 }, { "epoch": 2.1097472924187723, "grad_norm": 1.321266770362854, "learning_rate": 8.945126353790613e-05, "loss": 1.8747, "step": 2922 }, { "epoch": 2.1140794223826713, "grad_norm": 1.4037319421768188, "learning_rate": 8.942960288808665e-05, "loss": 1.8693, "step": 2928 }, { "epoch": 2.1184115523465703, "grad_norm": 1.3415311574935913, "learning_rate": 8.940794223826715e-05, "loss": 1.7625, "step": 2934 }, { "epoch": 2.1227436823104693, "grad_norm": 1.645894169807434, "learning_rate": 8.938628158844766e-05, "loss": 1.7546, "step": 2940 }, { "epoch": 2.1270758122743683, "grad_norm": 1.552162766456604, "learning_rate": 8.936462093862816e-05, "loss": 1.8146, "step": 2946 }, { "epoch": 2.1314079422382672, "grad_norm": 1.4145435094833374, "learning_rate": 8.934296028880866e-05, "loss": 1.8956, "step": 2952 }, { "epoch": 2.135740072202166, "grad_norm": 1.3878324031829834, "learning_rate": 8.932129963898918e-05, "loss": 1.8807, "step": 2958 }, { "epoch": 2.140072202166065, "grad_norm": 1.613510012626648, "learning_rate": 8.929963898916968e-05, "loss": 1.6833, "step": 2964 }, { "epoch": 2.1444043321299637, "grad_norm": 1.3004451990127563, "learning_rate": 8.927797833935019e-05, "loss": 1.88, "step": 2970 }, { "epoch": 2.1487364620938627, "grad_norm": 1.4385720491409302, "learning_rate": 8.925631768953069e-05, "loss": 1.7543, "step": 2976 }, { "epoch": 2.1530685920577617, "grad_norm": 1.306748628616333, "learning_rate": 8.92346570397112e-05, "loss": 1.6954, "step": 2982 }, { "epoch": 2.1574007220216607, "grad_norm": 1.3268028497695923, "learning_rate": 8.92129963898917e-05, "loss": 1.8243, "step": 2988 }, { "epoch": 2.1617328519855596, "grad_norm": 1.3416894674301147, "learning_rate": 8.91913357400722e-05, "loss": 1.6302, "step": 2994 }, { "epoch": 2.1660649819494586, "grad_norm": 1.5698119401931763, "learning_rate": 8.916967509025272e-05, "loss": 1.8843, "step": 3000 }, { "epoch": 2.1703971119133576, "grad_norm": 1.485539197921753, "learning_rate": 8.914801444043322e-05, "loss": 1.849, "step": 3006 }, { "epoch": 2.174729241877256, "grad_norm": 1.4647479057312012, "learning_rate": 8.912635379061372e-05, "loss": 1.7756, "step": 3012 }, { "epoch": 2.179061371841155, "grad_norm": 1.5629044771194458, "learning_rate": 8.910469314079422e-05, "loss": 1.8325, "step": 3018 }, { "epoch": 2.183393501805054, "grad_norm": 1.6540532112121582, "learning_rate": 8.908303249097474e-05, "loss": 1.7513, "step": 3024 }, { "epoch": 2.187725631768953, "grad_norm": 1.4427595138549805, "learning_rate": 8.906137184115524e-05, "loss": 1.8796, "step": 3030 }, { "epoch": 2.192057761732852, "grad_norm": 1.3385310173034668, "learning_rate": 8.903971119133573e-05, "loss": 1.8149, "step": 3036 }, { "epoch": 2.196389891696751, "grad_norm": 1.4092869758605957, "learning_rate": 8.901805054151625e-05, "loss": 1.7242, "step": 3042 }, { "epoch": 2.20072202166065, "grad_norm": 1.5199534893035889, "learning_rate": 8.899638989169675e-05, "loss": 1.8425, "step": 3048 }, { "epoch": 2.2050541516245485, "grad_norm": 1.5366950035095215, "learning_rate": 8.897472924187726e-05, "loss": 1.7446, "step": 3054 }, { "epoch": 2.2093862815884475, "grad_norm": 1.3262131214141846, "learning_rate": 8.895306859205776e-05, "loss": 1.6771, "step": 3060 }, { "epoch": 2.2137184115523465, "grad_norm": 1.5029282569885254, "learning_rate": 8.893140794223828e-05, "loss": 1.9294, "step": 3066 }, { "epoch": 2.2180505415162455, "grad_norm": 1.469660758972168, "learning_rate": 8.890974729241878e-05, "loss": 1.7583, "step": 3072 }, { "epoch": 2.2223826714801445, "grad_norm": 1.432472586631775, "learning_rate": 8.888808664259928e-05, "loss": 1.7202, "step": 3078 }, { "epoch": 2.2267148014440434, "grad_norm": 1.4789842367172241, "learning_rate": 8.886642599277979e-05, "loss": 1.8158, "step": 3084 }, { "epoch": 2.2310469314079424, "grad_norm": 1.3391679525375366, "learning_rate": 8.884476534296029e-05, "loss": 1.7642, "step": 3090 }, { "epoch": 2.235379061371841, "grad_norm": 1.5485726594924927, "learning_rate": 8.882310469314081e-05, "loss": 1.7391, "step": 3096 }, { "epoch": 2.23971119133574, "grad_norm": 1.4782353639602661, "learning_rate": 8.88014440433213e-05, "loss": 1.9485, "step": 3102 }, { "epoch": 2.244043321299639, "grad_norm": 1.3726975917816162, "learning_rate": 8.877978339350181e-05, "loss": 2.0078, "step": 3108 }, { "epoch": 2.248375451263538, "grad_norm": 1.5260320901870728, "learning_rate": 8.875812274368232e-05, "loss": 1.6579, "step": 3114 }, { "epoch": 2.252707581227437, "grad_norm": 1.611466884613037, "learning_rate": 8.873646209386282e-05, "loss": 1.7774, "step": 3120 }, { "epoch": 2.257039711191336, "grad_norm": 1.424183964729309, "learning_rate": 8.871480144404332e-05, "loss": 1.7318, "step": 3126 }, { "epoch": 2.261371841155235, "grad_norm": 1.3302850723266602, "learning_rate": 8.869314079422383e-05, "loss": 1.8385, "step": 3132 }, { "epoch": 2.265703971119134, "grad_norm": 1.3189494609832764, "learning_rate": 8.867148014440434e-05, "loss": 1.6771, "step": 3138 }, { "epoch": 2.2700361010830323, "grad_norm": 1.550413727760315, "learning_rate": 8.864981949458485e-05, "loss": 1.74, "step": 3144 }, { "epoch": 2.2743682310469313, "grad_norm": 1.2629214525222778, "learning_rate": 8.862815884476535e-05, "loss": 1.656, "step": 3150 }, { "epoch": 2.2787003610108303, "grad_norm": 1.4548317193984985, "learning_rate": 8.860649819494585e-05, "loss": 1.8073, "step": 3156 }, { "epoch": 2.2830324909747293, "grad_norm": 1.4483304023742676, "learning_rate": 8.858483754512636e-05, "loss": 1.8435, "step": 3162 }, { "epoch": 2.2873646209386282, "grad_norm": 1.4440531730651855, "learning_rate": 8.856317689530686e-05, "loss": 1.9712, "step": 3168 }, { "epoch": 2.2916967509025272, "grad_norm": 1.4899183511734009, "learning_rate": 8.854151624548736e-05, "loss": 1.9456, "step": 3174 }, { "epoch": 2.2960288808664258, "grad_norm": 1.5519587993621826, "learning_rate": 8.851985559566788e-05, "loss": 1.9127, "step": 3180 }, { "epoch": 2.3003610108303247, "grad_norm": 1.4153159856796265, "learning_rate": 8.849819494584838e-05, "loss": 1.7935, "step": 3186 }, { "epoch": 2.3046931407942237, "grad_norm": 1.473517894744873, "learning_rate": 8.847653429602889e-05, "loss": 1.8843, "step": 3192 }, { "epoch": 2.3090252707581227, "grad_norm": 1.3581178188323975, "learning_rate": 8.845487364620939e-05, "loss": 1.8765, "step": 3198 }, { "epoch": 2.3133574007220217, "grad_norm": 1.5811293125152588, "learning_rate": 8.843321299638989e-05, "loss": 1.7181, "step": 3204 }, { "epoch": 2.3176895306859207, "grad_norm": 1.523474931716919, "learning_rate": 8.841155234657041e-05, "loss": 1.8577, "step": 3210 }, { "epoch": 2.3220216606498196, "grad_norm": 1.563397765159607, "learning_rate": 8.83898916967509e-05, "loss": 1.7523, "step": 3216 }, { "epoch": 2.3263537906137186, "grad_norm": 1.4686119556427002, "learning_rate": 8.836823104693142e-05, "loss": 1.8704, "step": 3222 }, { "epoch": 2.330685920577617, "grad_norm": 1.4558385610580444, "learning_rate": 8.834657039711192e-05, "loss": 1.96, "step": 3228 }, { "epoch": 2.335018050541516, "grad_norm": 1.3353931903839111, "learning_rate": 8.832490974729242e-05, "loss": 1.9344, "step": 3234 }, { "epoch": 2.339350180505415, "grad_norm": 1.4872876405715942, "learning_rate": 8.830324909747293e-05, "loss": 1.8539, "step": 3240 }, { "epoch": 2.343682310469314, "grad_norm": 1.5594154596328735, "learning_rate": 8.828158844765343e-05, "loss": 1.8172, "step": 3246 }, { "epoch": 2.348014440433213, "grad_norm": 1.2777477502822876, "learning_rate": 8.825992779783395e-05, "loss": 1.8698, "step": 3252 }, { "epoch": 2.352346570397112, "grad_norm": 1.4094089269638062, "learning_rate": 8.823826714801444e-05, "loss": 1.7086, "step": 3258 }, { "epoch": 2.356678700361011, "grad_norm": 1.4454020261764526, "learning_rate": 8.821660649819495e-05, "loss": 1.8578, "step": 3264 }, { "epoch": 2.3610108303249095, "grad_norm": 1.4724870920181274, "learning_rate": 8.819494584837546e-05, "loss": 1.8504, "step": 3270 }, { "epoch": 2.3653429602888085, "grad_norm": 1.3311761617660522, "learning_rate": 8.817328519855597e-05, "loss": 1.8584, "step": 3276 }, { "epoch": 2.3696750902527075, "grad_norm": 1.4777659177780151, "learning_rate": 8.815162454873646e-05, "loss": 1.7093, "step": 3282 }, { "epoch": 2.3740072202166065, "grad_norm": 1.47633695602417, "learning_rate": 8.812996389891696e-05, "loss": 1.7541, "step": 3288 }, { "epoch": 2.3783393501805055, "grad_norm": 1.4523942470550537, "learning_rate": 8.810830324909748e-05, "loss": 1.7763, "step": 3294 }, { "epoch": 2.3826714801444044, "grad_norm": 1.517761468887329, "learning_rate": 8.808664259927798e-05, "loss": 1.7365, "step": 3300 }, { "epoch": 2.3870036101083034, "grad_norm": 1.5118472576141357, "learning_rate": 8.806498194945849e-05, "loss": 1.8087, "step": 3306 }, { "epoch": 2.3913357400722024, "grad_norm": 1.576972484588623, "learning_rate": 8.804332129963899e-05, "loss": 1.8659, "step": 3312 }, { "epoch": 2.395667870036101, "grad_norm": 1.5188674926757812, "learning_rate": 8.802166064981951e-05, "loss": 1.7156, "step": 3318 }, { "epoch": 2.4, "grad_norm": 1.4951331615447998, "learning_rate": 8.800000000000001e-05, "loss": 1.8251, "step": 3324 }, { "epoch": 2.404332129963899, "grad_norm": 1.4529703855514526, "learning_rate": 8.79783393501805e-05, "loss": 1.7554, "step": 3330 }, { "epoch": 2.408664259927798, "grad_norm": 1.4625014066696167, "learning_rate": 8.795667870036102e-05, "loss": 1.7585, "step": 3336 }, { "epoch": 2.412996389891697, "grad_norm": 1.4499711990356445, "learning_rate": 8.793501805054152e-05, "loss": 1.8358, "step": 3342 }, { "epoch": 2.417328519855596, "grad_norm": 1.5352270603179932, "learning_rate": 8.791335740072202e-05, "loss": 1.7966, "step": 3348 }, { "epoch": 2.4216606498194944, "grad_norm": 1.5651659965515137, "learning_rate": 8.789169675090253e-05, "loss": 1.7946, "step": 3354 }, { "epoch": 2.4259927797833933, "grad_norm": 1.4316205978393555, "learning_rate": 8.787003610108304e-05, "loss": 1.8841, "step": 3360 }, { "epoch": 2.4303249097472923, "grad_norm": 1.547731876373291, "learning_rate": 8.784837545126355e-05, "loss": 1.9856, "step": 3366 }, { "epoch": 2.4346570397111913, "grad_norm": 1.4234883785247803, "learning_rate": 8.782671480144404e-05, "loss": 1.9239, "step": 3372 }, { "epoch": 2.4389891696750903, "grad_norm": 1.4130299091339111, "learning_rate": 8.780505415162455e-05, "loss": 1.9368, "step": 3378 }, { "epoch": 2.4433212996389893, "grad_norm": 1.6666849851608276, "learning_rate": 8.778339350180506e-05, "loss": 1.9196, "step": 3384 }, { "epoch": 2.4476534296028882, "grad_norm": 1.328849196434021, "learning_rate": 8.776173285198557e-05, "loss": 1.8679, "step": 3390 }, { "epoch": 2.451985559566787, "grad_norm": 1.4613603353500366, "learning_rate": 8.774007220216606e-05, "loss": 1.7502, "step": 3396 }, { "epoch": 2.4563176895306857, "grad_norm": 1.6586109399795532, "learning_rate": 8.771841155234658e-05, "loss": 1.82, "step": 3402 }, { "epoch": 2.4606498194945847, "grad_norm": 1.5979253053665161, "learning_rate": 8.769675090252708e-05, "loss": 1.9547, "step": 3408 }, { "epoch": 2.4649819494584837, "grad_norm": 1.5491660833358765, "learning_rate": 8.767509025270759e-05, "loss": 1.7877, "step": 3414 }, { "epoch": 2.4693140794223827, "grad_norm": 1.3452082872390747, "learning_rate": 8.765342960288809e-05, "loss": 1.9054, "step": 3420 }, { "epoch": 2.4736462093862817, "grad_norm": 1.5472066402435303, "learning_rate": 8.763176895306859e-05, "loss": 1.7345, "step": 3426 }, { "epoch": 2.4779783393501806, "grad_norm": 1.336403250694275, "learning_rate": 8.761010830324911e-05, "loss": 1.7745, "step": 3432 }, { "epoch": 2.4823104693140796, "grad_norm": 1.5414328575134277, "learning_rate": 8.75884476534296e-05, "loss": 1.7927, "step": 3438 }, { "epoch": 2.486642599277978, "grad_norm": 1.6150093078613281, "learning_rate": 8.756678700361012e-05, "loss": 1.9031, "step": 3444 }, { "epoch": 2.490974729241877, "grad_norm": 1.6289031505584717, "learning_rate": 8.754512635379062e-05, "loss": 1.7763, "step": 3450 }, { "epoch": 2.495306859205776, "grad_norm": 1.493942379951477, "learning_rate": 8.752346570397112e-05, "loss": 1.7914, "step": 3456 }, { "epoch": 2.499638989169675, "grad_norm": 1.5335317850112915, "learning_rate": 8.750180505415163e-05, "loss": 1.8566, "step": 3462 }, { "epoch": 2.503971119133574, "grad_norm": 1.445523738861084, "learning_rate": 8.748014440433213e-05, "loss": 1.7351, "step": 3468 }, { "epoch": 2.508303249097473, "grad_norm": 1.2315399646759033, "learning_rate": 8.745848375451265e-05, "loss": 1.725, "step": 3474 }, { "epoch": 2.512635379061372, "grad_norm": 1.618829607963562, "learning_rate": 8.743682310469315e-05, "loss": 1.8279, "step": 3480 }, { "epoch": 2.516967509025271, "grad_norm": 1.4977085590362549, "learning_rate": 8.741516245487365e-05, "loss": 1.7463, "step": 3486 }, { "epoch": 2.5212996389891695, "grad_norm": 1.3427156209945679, "learning_rate": 8.739350180505416e-05, "loss": 1.8968, "step": 3492 }, { "epoch": 2.5256317689530685, "grad_norm": 1.4139822721481323, "learning_rate": 8.737184115523466e-05, "loss": 1.9395, "step": 3498 }, { "epoch": 2.5299638989169675, "grad_norm": 1.4320950508117676, "learning_rate": 8.735018050541516e-05, "loss": 1.8103, "step": 3504 }, { "epoch": 2.5342960288808665, "grad_norm": 1.496302843093872, "learning_rate": 8.732851985559567e-05, "loss": 1.979, "step": 3510 }, { "epoch": 2.5386281588447654, "grad_norm": 1.4391734600067139, "learning_rate": 8.730685920577618e-05, "loss": 1.783, "step": 3516 }, { "epoch": 2.5429602888086644, "grad_norm": 1.4863210916519165, "learning_rate": 8.728519855595669e-05, "loss": 1.8357, "step": 3522 }, { "epoch": 2.547292418772563, "grad_norm": 1.4426627159118652, "learning_rate": 8.726353790613719e-05, "loss": 1.8779, "step": 3528 }, { "epoch": 2.551624548736462, "grad_norm": 1.5277155637741089, "learning_rate": 8.724187725631769e-05, "loss": 1.7025, "step": 3534 }, { "epoch": 2.555956678700361, "grad_norm": 1.422000527381897, "learning_rate": 8.72202166064982e-05, "loss": 1.8269, "step": 3540 }, { "epoch": 2.56028880866426, "grad_norm": 1.555419683456421, "learning_rate": 8.719855595667871e-05, "loss": 1.9054, "step": 3546 }, { "epoch": 2.564620938628159, "grad_norm": 1.516588807106018, "learning_rate": 8.71768953068592e-05, "loss": 1.8149, "step": 3552 }, { "epoch": 2.568953068592058, "grad_norm": 1.4617726802825928, "learning_rate": 8.715523465703972e-05, "loss": 1.5867, "step": 3558 }, { "epoch": 2.573285198555957, "grad_norm": 1.4879695177078247, "learning_rate": 8.713357400722022e-05, "loss": 1.7458, "step": 3564 }, { "epoch": 2.577617328519856, "grad_norm": 1.3446221351623535, "learning_rate": 8.711191335740072e-05, "loss": 1.7134, "step": 3570 }, { "epoch": 2.581949458483755, "grad_norm": 1.4569180011749268, "learning_rate": 8.709025270758123e-05, "loss": 1.8212, "step": 3576 }, { "epoch": 2.5862815884476533, "grad_norm": 1.6838642358779907, "learning_rate": 8.706859205776173e-05, "loss": 1.7387, "step": 3582 }, { "epoch": 2.5906137184115523, "grad_norm": 1.3476632833480835, "learning_rate": 8.704693140794225e-05, "loss": 1.6771, "step": 3588 }, { "epoch": 2.5949458483754513, "grad_norm": 1.6001485586166382, "learning_rate": 8.702527075812275e-05, "loss": 1.8737, "step": 3594 }, { "epoch": 2.5992779783393503, "grad_norm": 1.5055811405181885, "learning_rate": 8.700361010830325e-05, "loss": 1.8475, "step": 3600 }, { "epoch": 2.6036101083032492, "grad_norm": 1.4596132040023804, "learning_rate": 8.698194945848376e-05, "loss": 1.8348, "step": 3606 }, { "epoch": 2.6079422382671478, "grad_norm": 1.541390299797058, "learning_rate": 8.696028880866426e-05, "loss": 1.8862, "step": 3612 }, { "epoch": 2.6122743682310468, "grad_norm": 1.6278433799743652, "learning_rate": 8.693862815884476e-05, "loss": 1.9699, "step": 3618 }, { "epoch": 2.6166064981949457, "grad_norm": 1.5377955436706543, "learning_rate": 8.691696750902527e-05, "loss": 1.9127, "step": 3624 }, { "epoch": 2.6209386281588447, "grad_norm": 1.4964910745620728, "learning_rate": 8.689530685920578e-05, "loss": 1.7638, "step": 3630 }, { "epoch": 2.6252707581227437, "grad_norm": 1.3137484788894653, "learning_rate": 8.687364620938629e-05, "loss": 1.7191, "step": 3636 }, { "epoch": 2.6296028880866427, "grad_norm": 1.5137817859649658, "learning_rate": 8.685198555956679e-05, "loss": 1.8831, "step": 3642 }, { "epoch": 2.6339350180505416, "grad_norm": 1.3826507329940796, "learning_rate": 8.68303249097473e-05, "loss": 1.7801, "step": 3648 }, { "epoch": 2.6382671480144406, "grad_norm": 1.5147032737731934, "learning_rate": 8.68086642599278e-05, "loss": 1.7678, "step": 3654 }, { "epoch": 2.6425992779783396, "grad_norm": 1.545626163482666, "learning_rate": 8.678700361010831e-05, "loss": 1.7927, "step": 3660 }, { "epoch": 2.646931407942238, "grad_norm": 1.6036428213119507, "learning_rate": 8.67653429602888e-05, "loss": 1.7683, "step": 3666 }, { "epoch": 2.651263537906137, "grad_norm": 1.4055842161178589, "learning_rate": 8.674368231046932e-05, "loss": 1.8374, "step": 3672 }, { "epoch": 2.655595667870036, "grad_norm": 1.4374827146530151, "learning_rate": 8.672202166064982e-05, "loss": 1.8827, "step": 3678 }, { "epoch": 2.659927797833935, "grad_norm": 1.3665235042572021, "learning_rate": 8.670036101083033e-05, "loss": 1.7861, "step": 3684 }, { "epoch": 2.664259927797834, "grad_norm": 1.4032800197601318, "learning_rate": 8.667870036101083e-05, "loss": 1.8567, "step": 3690 }, { "epoch": 2.6685920577617326, "grad_norm": 1.4242801666259766, "learning_rate": 8.665703971119133e-05, "loss": 1.8175, "step": 3696 }, { "epoch": 2.6729241877256316, "grad_norm": 1.541653037071228, "learning_rate": 8.663537906137185e-05, "loss": 1.8124, "step": 3702 }, { "epoch": 2.6772563176895305, "grad_norm": 1.5858688354492188, "learning_rate": 8.661371841155235e-05, "loss": 1.7549, "step": 3708 }, { "epoch": 2.6815884476534295, "grad_norm": 1.63284432888031, "learning_rate": 8.659205776173286e-05, "loss": 1.7369, "step": 3714 }, { "epoch": 2.6859205776173285, "grad_norm": 1.5337514877319336, "learning_rate": 8.657039711191336e-05, "loss": 1.7082, "step": 3720 }, { "epoch": 2.6902527075812275, "grad_norm": 1.5949983596801758, "learning_rate": 8.654873646209388e-05, "loss": 1.7704, "step": 3726 }, { "epoch": 2.6945848375451265, "grad_norm": 1.3026549816131592, "learning_rate": 8.652707581227437e-05, "loss": 1.7037, "step": 3732 }, { "epoch": 2.6989169675090254, "grad_norm": 1.533652424812317, "learning_rate": 8.650541516245487e-05, "loss": 1.7338, "step": 3738 }, { "epoch": 2.7032490974729244, "grad_norm": 1.599485993385315, "learning_rate": 8.648375451263539e-05, "loss": 1.7463, "step": 3744 }, { "epoch": 2.707581227436823, "grad_norm": 1.5458048582077026, "learning_rate": 8.646209386281589e-05, "loss": 1.7586, "step": 3750 }, { "epoch": 2.711913357400722, "grad_norm": 1.491220474243164, "learning_rate": 8.644043321299639e-05, "loss": 1.8196, "step": 3756 }, { "epoch": 2.716245487364621, "grad_norm": 1.508008599281311, "learning_rate": 8.64187725631769e-05, "loss": 1.9138, "step": 3762 }, { "epoch": 2.72057761732852, "grad_norm": 1.3923741579055786, "learning_rate": 8.639711191335741e-05, "loss": 1.7389, "step": 3768 }, { "epoch": 2.724909747292419, "grad_norm": 1.3440769910812378, "learning_rate": 8.637545126353792e-05, "loss": 1.6845, "step": 3774 }, { "epoch": 2.729241877256318, "grad_norm": 1.4081510305404663, "learning_rate": 8.63537906137184e-05, "loss": 1.8043, "step": 3780 }, { "epoch": 2.7335740072202164, "grad_norm": 1.5444183349609375, "learning_rate": 8.633212996389892e-05, "loss": 1.8649, "step": 3786 }, { "epoch": 2.7379061371841154, "grad_norm": 1.346382737159729, "learning_rate": 8.631046931407943e-05, "loss": 1.7967, "step": 3792 }, { "epoch": 2.7422382671480143, "grad_norm": 1.384224534034729, "learning_rate": 8.628880866425993e-05, "loss": 1.6876, "step": 3798 }, { "epoch": 2.7465703971119133, "grad_norm": 1.6237854957580566, "learning_rate": 8.626714801444043e-05, "loss": 1.7972, "step": 3804 }, { "epoch": 2.7509025270758123, "grad_norm": 1.4574010372161865, "learning_rate": 8.624548736462095e-05, "loss": 1.8551, "step": 3810 }, { "epoch": 2.7552346570397113, "grad_norm": 1.7082672119140625, "learning_rate": 8.622382671480145e-05, "loss": 1.9006, "step": 3816 }, { "epoch": 2.7595667870036102, "grad_norm": 1.628898024559021, "learning_rate": 8.620216606498196e-05, "loss": 2.0072, "step": 3822 }, { "epoch": 2.7638989169675092, "grad_norm": 1.5239311456680298, "learning_rate": 8.618050541516246e-05, "loss": 1.8311, "step": 3828 }, { "epoch": 2.768231046931408, "grad_norm": 1.3905366659164429, "learning_rate": 8.615884476534296e-05, "loss": 1.6844, "step": 3834 }, { "epoch": 2.7725631768953067, "grad_norm": 1.6084203720092773, "learning_rate": 8.613718411552348e-05, "loss": 1.8446, "step": 3840 }, { "epoch": 2.7768953068592057, "grad_norm": 1.514164686203003, "learning_rate": 8.611552346570397e-05, "loss": 1.8022, "step": 3846 }, { "epoch": 2.7812274368231047, "grad_norm": 1.3984124660491943, "learning_rate": 8.609386281588448e-05, "loss": 1.7041, "step": 3852 }, { "epoch": 2.7855595667870037, "grad_norm": 1.3913060426712036, "learning_rate": 8.607220216606499e-05, "loss": 1.7604, "step": 3858 }, { "epoch": 2.7898916967509026, "grad_norm": 1.3518754243850708, "learning_rate": 8.605054151624549e-05, "loss": 1.7285, "step": 3864 }, { "epoch": 2.794223826714801, "grad_norm": 1.4763267040252686, "learning_rate": 8.6028880866426e-05, "loss": 1.7959, "step": 3870 }, { "epoch": 2.7985559566787, "grad_norm": 1.4982695579528809, "learning_rate": 8.60072202166065e-05, "loss": 1.8187, "step": 3876 }, { "epoch": 2.802888086642599, "grad_norm": 1.4512739181518555, "learning_rate": 8.598555956678701e-05, "loss": 1.9409, "step": 3882 }, { "epoch": 2.807220216606498, "grad_norm": 1.4035935401916504, "learning_rate": 8.596389891696752e-05, "loss": 1.8048, "step": 3888 }, { "epoch": 2.811552346570397, "grad_norm": 1.4693052768707275, "learning_rate": 8.594223826714802e-05, "loss": 1.7355, "step": 3894 }, { "epoch": 2.815884476534296, "grad_norm": 1.4883086681365967, "learning_rate": 8.592057761732852e-05, "loss": 1.9013, "step": 3900 }, { "epoch": 2.820216606498195, "grad_norm": 1.808141827583313, "learning_rate": 8.589891696750903e-05, "loss": 1.7993, "step": 3906 }, { "epoch": 2.824548736462094, "grad_norm": 1.3361742496490479, "learning_rate": 8.587725631768953e-05, "loss": 1.8508, "step": 3912 }, { "epoch": 2.828880866425993, "grad_norm": 1.355735421180725, "learning_rate": 8.585559566787003e-05, "loss": 1.7837, "step": 3918 }, { "epoch": 2.8332129963898915, "grad_norm": 1.4334018230438232, "learning_rate": 8.583393501805055e-05, "loss": 1.8473, "step": 3924 }, { "epoch": 2.8375451263537905, "grad_norm": 1.5884405374526978, "learning_rate": 8.581227436823105e-05, "loss": 1.7645, "step": 3930 }, { "epoch": 2.8418772563176895, "grad_norm": 1.6245017051696777, "learning_rate": 8.579061371841156e-05, "loss": 1.8379, "step": 3936 }, { "epoch": 2.8462093862815885, "grad_norm": 1.7909353971481323, "learning_rate": 8.576895306859206e-05, "loss": 1.9469, "step": 3942 }, { "epoch": 2.8505415162454875, "grad_norm": 1.4635123014450073, "learning_rate": 8.574729241877256e-05, "loss": 1.7914, "step": 3948 }, { "epoch": 2.8548736462093864, "grad_norm": 1.3702229261398315, "learning_rate": 8.572563176895308e-05, "loss": 1.7946, "step": 3954 }, { "epoch": 2.859205776173285, "grad_norm": 1.7010841369628906, "learning_rate": 8.570397111913357e-05, "loss": 1.7905, "step": 3960 }, { "epoch": 2.863537906137184, "grad_norm": 1.5447696447372437, "learning_rate": 8.568231046931409e-05, "loss": 1.8085, "step": 3966 }, { "epoch": 2.867870036101083, "grad_norm": 1.5590651035308838, "learning_rate": 8.566064981949459e-05, "loss": 1.858, "step": 3972 }, { "epoch": 2.872202166064982, "grad_norm": 1.5558878183364868, "learning_rate": 8.563898916967509e-05, "loss": 1.9387, "step": 3978 }, { "epoch": 2.876534296028881, "grad_norm": 1.45836341381073, "learning_rate": 8.56173285198556e-05, "loss": 1.6581, "step": 3984 }, { "epoch": 2.88086642599278, "grad_norm": 1.5232306718826294, "learning_rate": 8.55956678700361e-05, "loss": 1.7459, "step": 3990 }, { "epoch": 2.885198555956679, "grad_norm": 1.4140738248825073, "learning_rate": 8.557400722021662e-05, "loss": 1.8146, "step": 3996 }, { "epoch": 2.889530685920578, "grad_norm": 1.6213436126708984, "learning_rate": 8.55523465703971e-05, "loss": 1.8202, "step": 4002 }, { "epoch": 2.893862815884477, "grad_norm": 1.4864338636398315, "learning_rate": 8.553068592057762e-05, "loss": 1.7771, "step": 4008 }, { "epoch": 2.8981949458483753, "grad_norm": 1.643946647644043, "learning_rate": 8.550902527075813e-05, "loss": 1.9363, "step": 4014 }, { "epoch": 2.9025270758122743, "grad_norm": 1.447617769241333, "learning_rate": 8.548736462093864e-05, "loss": 1.8174, "step": 4020 }, { "epoch": 2.9068592057761733, "grad_norm": 1.4886562824249268, "learning_rate": 8.546570397111913e-05, "loss": 1.8508, "step": 4026 }, { "epoch": 2.9111913357400723, "grad_norm": 1.5560579299926758, "learning_rate": 8.544404332129964e-05, "loss": 1.6819, "step": 4032 }, { "epoch": 2.9155234657039713, "grad_norm": 1.4020459651947021, "learning_rate": 8.542238267148015e-05, "loss": 1.8833, "step": 4038 }, { "epoch": 2.91985559566787, "grad_norm": 1.3136491775512695, "learning_rate": 8.540072202166066e-05, "loss": 1.7149, "step": 4044 }, { "epoch": 2.9241877256317688, "grad_norm": 1.564806342124939, "learning_rate": 8.537906137184116e-05, "loss": 1.9677, "step": 4050 }, { "epoch": 2.9285198555956677, "grad_norm": 1.3823758363723755, "learning_rate": 8.535740072202166e-05, "loss": 1.8151, "step": 4056 }, { "epoch": 2.9328519855595667, "grad_norm": 1.47847580909729, "learning_rate": 8.533574007220218e-05, "loss": 1.8068, "step": 4062 }, { "epoch": 2.9371841155234657, "grad_norm": 1.5501521825790405, "learning_rate": 8.531407942238268e-05, "loss": 1.953, "step": 4068 }, { "epoch": 2.9415162454873647, "grad_norm": 1.3352515697479248, "learning_rate": 8.529241877256317e-05, "loss": 1.7806, "step": 4074 }, { "epoch": 2.9458483754512637, "grad_norm": 1.4805330038070679, "learning_rate": 8.527075812274369e-05, "loss": 1.8453, "step": 4080 }, { "epoch": 2.9501805054151626, "grad_norm": 1.563120722770691, "learning_rate": 8.524909747292419e-05, "loss": 1.8172, "step": 4086 }, { "epoch": 2.9545126353790616, "grad_norm": 1.5237318277359009, "learning_rate": 8.52274368231047e-05, "loss": 1.7736, "step": 4092 }, { "epoch": 2.95884476534296, "grad_norm": 1.4789378643035889, "learning_rate": 8.52057761732852e-05, "loss": 1.7867, "step": 4098 }, { "epoch": 2.963176895306859, "grad_norm": 1.4557100534439087, "learning_rate": 8.518411552346572e-05, "loss": 1.741, "step": 4104 }, { "epoch": 2.967509025270758, "grad_norm": 1.7376407384872437, "learning_rate": 8.516245487364622e-05, "loss": 1.8227, "step": 4110 }, { "epoch": 2.971841155234657, "grad_norm": 1.4459553956985474, "learning_rate": 8.514079422382671e-05, "loss": 1.8737, "step": 4116 }, { "epoch": 2.976173285198556, "grad_norm": 1.4566750526428223, "learning_rate": 8.511913357400722e-05, "loss": 1.7253, "step": 4122 }, { "epoch": 2.9805054151624546, "grad_norm": 1.4124408960342407, "learning_rate": 8.509747292418773e-05, "loss": 1.707, "step": 4128 }, { "epoch": 2.9848375451263536, "grad_norm": 1.6454405784606934, "learning_rate": 8.507581227436824e-05, "loss": 1.806, "step": 4134 }, { "epoch": 2.9891696750902526, "grad_norm": 1.5051449537277222, "learning_rate": 8.505415162454873e-05, "loss": 1.8316, "step": 4140 }, { "epoch": 2.9935018050541515, "grad_norm": 1.5145840644836426, "learning_rate": 8.503249097472925e-05, "loss": 1.7164, "step": 4146 }, { "epoch": 2.9978339350180505, "grad_norm": 1.3991270065307617, "learning_rate": 8.501083032490975e-05, "loss": 1.864, "step": 4152 }, { "epoch": 3.0021660649819495, "grad_norm": 1.419814944267273, "learning_rate": 8.498916967509026e-05, "loss": 1.8375, "step": 4158 }, { "epoch": 3.0064981949458485, "grad_norm": 1.7082300186157227, "learning_rate": 8.496750902527076e-05, "loss": 1.7882, "step": 4164 }, { "epoch": 3.0108303249097474, "grad_norm": 1.423844814300537, "learning_rate": 8.494584837545126e-05, "loss": 1.656, "step": 4170 }, { "epoch": 3.0151624548736464, "grad_norm": 1.4101111888885498, "learning_rate": 8.492418772563178e-05, "loss": 1.7283, "step": 4176 }, { "epoch": 3.019494584837545, "grad_norm": 1.371246576309204, "learning_rate": 8.490252707581227e-05, "loss": 1.6889, "step": 4182 }, { "epoch": 3.023826714801444, "grad_norm": 1.4183197021484375, "learning_rate": 8.488086642599279e-05, "loss": 1.707, "step": 4188 }, { "epoch": 3.028158844765343, "grad_norm": 1.4114634990692139, "learning_rate": 8.485920577617329e-05, "loss": 1.6719, "step": 4194 }, { "epoch": 3.032490974729242, "grad_norm": 1.5585206747055054, "learning_rate": 8.48375451263538e-05, "loss": 1.6871, "step": 4200 }, { "epoch": 3.036823104693141, "grad_norm": 1.5855401754379272, "learning_rate": 8.48158844765343e-05, "loss": 1.7406, "step": 4206 }, { "epoch": 3.04115523465704, "grad_norm": 1.450364351272583, "learning_rate": 8.47942238267148e-05, "loss": 1.7457, "step": 4212 }, { "epoch": 3.045487364620939, "grad_norm": 1.5571187734603882, "learning_rate": 8.477256317689532e-05, "loss": 1.9011, "step": 4218 }, { "epoch": 3.0498194945848374, "grad_norm": 1.4603110551834106, "learning_rate": 8.475090252707582e-05, "loss": 1.6843, "step": 4224 }, { "epoch": 3.0541516245487363, "grad_norm": 1.548563003540039, "learning_rate": 8.472924187725632e-05, "loss": 1.8329, "step": 4230 }, { "epoch": 3.0584837545126353, "grad_norm": 1.4152315855026245, "learning_rate": 8.470758122743683e-05, "loss": 1.7472, "step": 4236 }, { "epoch": 3.0628158844765343, "grad_norm": 1.4308110475540161, "learning_rate": 8.468592057761733e-05, "loss": 1.7025, "step": 4242 }, { "epoch": 3.0671480144404333, "grad_norm": 1.6021742820739746, "learning_rate": 8.466425992779783e-05, "loss": 1.7973, "step": 4248 }, { "epoch": 3.0714801444043323, "grad_norm": 1.4094674587249756, "learning_rate": 8.464259927797834e-05, "loss": 1.6397, "step": 4254 }, { "epoch": 3.0758122743682312, "grad_norm": 1.7090513706207275, "learning_rate": 8.462093862815885e-05, "loss": 1.7928, "step": 4260 }, { "epoch": 3.0801444043321298, "grad_norm": 1.5686172246932983, "learning_rate": 8.459927797833936e-05, "loss": 1.8277, "step": 4266 }, { "epoch": 3.0844765342960287, "grad_norm": 1.4457476139068604, "learning_rate": 8.457761732851986e-05, "loss": 1.7757, "step": 4272 }, { "epoch": 3.0888086642599277, "grad_norm": 1.8065993785858154, "learning_rate": 8.455595667870036e-05, "loss": 1.734, "step": 4278 }, { "epoch": 3.0931407942238267, "grad_norm": 1.4931447505950928, "learning_rate": 8.453429602888087e-05, "loss": 1.7614, "step": 4284 }, { "epoch": 3.0974729241877257, "grad_norm": 1.7065285444259644, "learning_rate": 8.451263537906138e-05, "loss": 1.7306, "step": 4290 }, { "epoch": 3.1018050541516247, "grad_norm": 1.6127516031265259, "learning_rate": 8.449097472924187e-05, "loss": 1.8682, "step": 4296 }, { "epoch": 3.1061371841155236, "grad_norm": 1.5586177110671997, "learning_rate": 8.446931407942239e-05, "loss": 1.7209, "step": 4302 }, { "epoch": 3.110469314079422, "grad_norm": 1.5729008913040161, "learning_rate": 8.444765342960289e-05, "loss": 1.7559, "step": 4308 }, { "epoch": 3.114801444043321, "grad_norm": 1.6049374341964722, "learning_rate": 8.442599277978341e-05, "loss": 1.8671, "step": 4314 }, { "epoch": 3.11913357400722, "grad_norm": 1.61104416847229, "learning_rate": 8.44043321299639e-05, "loss": 1.7105, "step": 4320 }, { "epoch": 3.123465703971119, "grad_norm": 1.564556360244751, "learning_rate": 8.43826714801444e-05, "loss": 1.8528, "step": 4326 }, { "epoch": 3.127797833935018, "grad_norm": 1.5766047239303589, "learning_rate": 8.436101083032492e-05, "loss": 1.7691, "step": 4332 }, { "epoch": 3.132129963898917, "grad_norm": 1.5864530801773071, "learning_rate": 8.433935018050542e-05, "loss": 1.9316, "step": 4338 }, { "epoch": 3.136462093862816, "grad_norm": 1.6648482084274292, "learning_rate": 8.431768953068593e-05, "loss": 1.7536, "step": 4344 }, { "epoch": 3.140794223826715, "grad_norm": 1.5193685293197632, "learning_rate": 8.429602888086643e-05, "loss": 1.6894, "step": 4350 }, { "epoch": 3.1451263537906136, "grad_norm": 1.605458378791809, "learning_rate": 8.427436823104695e-05, "loss": 1.7705, "step": 4356 }, { "epoch": 3.1494584837545125, "grad_norm": 1.6133873462677002, "learning_rate": 8.425270758122744e-05, "loss": 1.751, "step": 4362 }, { "epoch": 3.1537906137184115, "grad_norm": 1.5689691305160522, "learning_rate": 8.423104693140794e-05, "loss": 1.7249, "step": 4368 }, { "epoch": 3.1581227436823105, "grad_norm": 1.5737971067428589, "learning_rate": 8.420938628158846e-05, "loss": 1.8715, "step": 4374 }, { "epoch": 3.1624548736462095, "grad_norm": 1.5358664989471436, "learning_rate": 8.418772563176896e-05, "loss": 1.696, "step": 4380 }, { "epoch": 3.1667870036101085, "grad_norm": 1.448867917060852, "learning_rate": 8.416606498194946e-05, "loss": 1.6911, "step": 4386 }, { "epoch": 3.171119133574007, "grad_norm": 1.601750373840332, "learning_rate": 8.414440433212996e-05, "loss": 1.6646, "step": 4392 }, { "epoch": 3.175451263537906, "grad_norm": 1.6414759159088135, "learning_rate": 8.412274368231048e-05, "loss": 1.7651, "step": 4398 }, { "epoch": 3.179783393501805, "grad_norm": 1.6494015455245972, "learning_rate": 8.410108303249098e-05, "loss": 1.7758, "step": 4404 }, { "epoch": 3.184115523465704, "grad_norm": 1.5052547454833984, "learning_rate": 8.407942238267147e-05, "loss": 1.7046, "step": 4410 }, { "epoch": 3.188447653429603, "grad_norm": 1.5180132389068604, "learning_rate": 8.405776173285199e-05, "loss": 1.7376, "step": 4416 }, { "epoch": 3.192779783393502, "grad_norm": 1.5165789127349854, "learning_rate": 8.40361010830325e-05, "loss": 1.8118, "step": 4422 }, { "epoch": 3.197111913357401, "grad_norm": 1.5024259090423584, "learning_rate": 8.4014440433213e-05, "loss": 1.7324, "step": 4428 }, { "epoch": 3.2014440433213, "grad_norm": 1.5724279880523682, "learning_rate": 8.39927797833935e-05, "loss": 1.7317, "step": 4434 }, { "epoch": 3.2057761732851984, "grad_norm": 1.4864689111709595, "learning_rate": 8.397111913357402e-05, "loss": 1.7826, "step": 4440 }, { "epoch": 3.2101083032490974, "grad_norm": 1.4951560497283936, "learning_rate": 8.394945848375452e-05, "loss": 1.7605, "step": 4446 }, { "epoch": 3.2144404332129963, "grad_norm": 1.4976807832717896, "learning_rate": 8.392779783393502e-05, "loss": 1.7622, "step": 4452 }, { "epoch": 3.2187725631768953, "grad_norm": 1.666466236114502, "learning_rate": 8.390613718411553e-05, "loss": 1.666, "step": 4458 }, { "epoch": 3.2231046931407943, "grad_norm": 1.5471858978271484, "learning_rate": 8.388447653429603e-05, "loss": 1.743, "step": 4464 }, { "epoch": 3.2274368231046933, "grad_norm": 1.5564380884170532, "learning_rate": 8.386281588447655e-05, "loss": 1.8763, "step": 4470 }, { "epoch": 3.2317689530685922, "grad_norm": 1.6341413259506226, "learning_rate": 8.384115523465704e-05, "loss": 1.6384, "step": 4476 }, { "epoch": 3.2361010830324908, "grad_norm": 1.5988564491271973, "learning_rate": 8.381949458483755e-05, "loss": 1.6961, "step": 4482 }, { "epoch": 3.2404332129963898, "grad_norm": 1.5455330610275269, "learning_rate": 8.379783393501806e-05, "loss": 1.6866, "step": 4488 }, { "epoch": 3.2447653429602887, "grad_norm": 1.617074966430664, "learning_rate": 8.377617328519856e-05, "loss": 1.8105, "step": 4494 }, { "epoch": 3.2490974729241877, "grad_norm": 1.6742558479309082, "learning_rate": 8.375451263537906e-05, "loss": 1.7833, "step": 4500 }, { "epoch": 3.2534296028880867, "grad_norm": 1.453832745552063, "learning_rate": 8.373285198555957e-05, "loss": 1.6767, "step": 4506 }, { "epoch": 3.2577617328519857, "grad_norm": 1.5320348739624023, "learning_rate": 8.371119133574008e-05, "loss": 1.6795, "step": 4512 }, { "epoch": 3.2620938628158846, "grad_norm": 1.7345879077911377, "learning_rate": 8.368953068592059e-05, "loss": 1.8145, "step": 4518 }, { "epoch": 3.2664259927797836, "grad_norm": 1.6064505577087402, "learning_rate": 8.366787003610109e-05, "loss": 1.7741, "step": 4524 }, { "epoch": 3.270758122743682, "grad_norm": 1.4818639755249023, "learning_rate": 8.364620938628159e-05, "loss": 1.7354, "step": 4530 }, { "epoch": 3.275090252707581, "grad_norm": 1.5519332885742188, "learning_rate": 8.36245487364621e-05, "loss": 1.7606, "step": 4536 }, { "epoch": 3.27942238267148, "grad_norm": 1.6139543056488037, "learning_rate": 8.36028880866426e-05, "loss": 1.6633, "step": 4542 }, { "epoch": 3.283754512635379, "grad_norm": 1.7811957597732544, "learning_rate": 8.35812274368231e-05, "loss": 1.757, "step": 4548 }, { "epoch": 3.288086642599278, "grad_norm": 1.6479711532592773, "learning_rate": 8.355956678700362e-05, "loss": 1.7119, "step": 4554 }, { "epoch": 3.292418772563177, "grad_norm": 1.5185586214065552, "learning_rate": 8.353790613718412e-05, "loss": 1.7213, "step": 4560 }, { "epoch": 3.2967509025270756, "grad_norm": 1.6819757223129272, "learning_rate": 8.351624548736463e-05, "loss": 1.7524, "step": 4566 }, { "epoch": 3.3010830324909746, "grad_norm": 1.6251552104949951, "learning_rate": 8.349458483754513e-05, "loss": 1.8034, "step": 4572 }, { "epoch": 3.3054151624548735, "grad_norm": 1.5681416988372803, "learning_rate": 8.347292418772563e-05, "loss": 1.6505, "step": 4578 }, { "epoch": 3.3097472924187725, "grad_norm": 1.6041322946548462, "learning_rate": 8.345126353790615e-05, "loss": 1.6664, "step": 4584 }, { "epoch": 3.3140794223826715, "grad_norm": 1.3140677213668823, "learning_rate": 8.342960288808664e-05, "loss": 1.6174, "step": 4590 }, { "epoch": 3.3184115523465705, "grad_norm": 1.5076295137405396, "learning_rate": 8.340794223826716e-05, "loss": 1.6726, "step": 4596 }, { "epoch": 3.3227436823104695, "grad_norm": 1.7478344440460205, "learning_rate": 8.338628158844766e-05, "loss": 1.8007, "step": 4602 }, { "epoch": 3.3270758122743684, "grad_norm": 1.845562219619751, "learning_rate": 8.336462093862816e-05, "loss": 1.7447, "step": 4608 }, { "epoch": 3.331407942238267, "grad_norm": 1.5124542713165283, "learning_rate": 8.334296028880867e-05, "loss": 1.6686, "step": 4614 }, { "epoch": 3.335740072202166, "grad_norm": 1.529046893119812, "learning_rate": 8.332129963898917e-05, "loss": 1.706, "step": 4620 }, { "epoch": 3.340072202166065, "grad_norm": 1.7531447410583496, "learning_rate": 8.329963898916969e-05, "loss": 1.7703, "step": 4626 }, { "epoch": 3.344404332129964, "grad_norm": 1.3565945625305176, "learning_rate": 8.327797833935019e-05, "loss": 1.7822, "step": 4632 }, { "epoch": 3.348736462093863, "grad_norm": 1.4938759803771973, "learning_rate": 8.325631768953069e-05, "loss": 1.7522, "step": 4638 }, { "epoch": 3.353068592057762, "grad_norm": 1.765376329421997, "learning_rate": 8.32346570397112e-05, "loss": 1.8021, "step": 4644 }, { "epoch": 3.357400722021661, "grad_norm": 1.8293818235397339, "learning_rate": 8.32129963898917e-05, "loss": 1.774, "step": 4650 }, { "epoch": 3.3617328519855594, "grad_norm": 1.484188199043274, "learning_rate": 8.31913357400722e-05, "loss": 1.7034, "step": 4656 }, { "epoch": 3.3660649819494584, "grad_norm": 1.6173304319381714, "learning_rate": 8.31696750902527e-05, "loss": 1.8063, "step": 4662 }, { "epoch": 3.3703971119133573, "grad_norm": 1.6482347249984741, "learning_rate": 8.314801444043322e-05, "loss": 1.8161, "step": 4668 }, { "epoch": 3.3747292418772563, "grad_norm": 1.5435242652893066, "learning_rate": 8.312635379061372e-05, "loss": 1.8121, "step": 4674 }, { "epoch": 3.3790613718411553, "grad_norm": 1.5043394565582275, "learning_rate": 8.310469314079423e-05, "loss": 1.8491, "step": 4680 }, { "epoch": 3.3833935018050543, "grad_norm": 1.600873589515686, "learning_rate": 8.308303249097473e-05, "loss": 1.8228, "step": 4686 }, { "epoch": 3.3877256317689532, "grad_norm": 1.6737322807312012, "learning_rate": 8.306137184115523e-05, "loss": 1.6796, "step": 4692 }, { "epoch": 3.3920577617328522, "grad_norm": 1.496095895767212, "learning_rate": 8.303971119133575e-05, "loss": 1.8529, "step": 4698 }, { "epoch": 3.3963898916967508, "grad_norm": 1.3883488178253174, "learning_rate": 8.301805054151624e-05, "loss": 1.5785, "step": 4704 }, { "epoch": 3.4007220216606497, "grad_norm": 1.548928141593933, "learning_rate": 8.299638989169676e-05, "loss": 1.6687, "step": 4710 }, { "epoch": 3.4050541516245487, "grad_norm": 1.54681396484375, "learning_rate": 8.297472924187726e-05, "loss": 1.7442, "step": 4716 }, { "epoch": 3.4093862815884477, "grad_norm": 1.548289179801941, "learning_rate": 8.295306859205776e-05, "loss": 1.731, "step": 4722 }, { "epoch": 3.4137184115523467, "grad_norm": 1.5993109941482544, "learning_rate": 8.293140794223827e-05, "loss": 1.7801, "step": 4728 }, { "epoch": 3.4180505415162457, "grad_norm": 1.6489711999893188, "learning_rate": 8.290974729241877e-05, "loss": 1.6428, "step": 4734 }, { "epoch": 3.422382671480144, "grad_norm": 1.4876315593719482, "learning_rate": 8.288808664259929e-05, "loss": 1.6462, "step": 4740 }, { "epoch": 3.426714801444043, "grad_norm": 1.7176752090454102, "learning_rate": 8.286642599277978e-05, "loss": 1.6393, "step": 4746 }, { "epoch": 3.431046931407942, "grad_norm": 1.5196412801742554, "learning_rate": 8.28447653429603e-05, "loss": 1.8994, "step": 4752 }, { "epoch": 3.435379061371841, "grad_norm": 1.658115267753601, "learning_rate": 8.28231046931408e-05, "loss": 1.8208, "step": 4758 }, { "epoch": 3.43971119133574, "grad_norm": 1.4496105909347534, "learning_rate": 8.280144404332131e-05, "loss": 1.8013, "step": 4764 }, { "epoch": 3.444043321299639, "grad_norm": 1.4678930044174194, "learning_rate": 8.27797833935018e-05, "loss": 1.7208, "step": 4770 }, { "epoch": 3.448375451263538, "grad_norm": 1.7084873914718628, "learning_rate": 8.27581227436823e-05, "loss": 1.6976, "step": 4776 }, { "epoch": 3.452707581227437, "grad_norm": 1.543615460395813, "learning_rate": 8.273646209386282e-05, "loss": 1.786, "step": 4782 }, { "epoch": 3.4570397111913356, "grad_norm": 1.6778939962387085, "learning_rate": 8.271480144404333e-05, "loss": 1.7775, "step": 4788 }, { "epoch": 3.4613718411552346, "grad_norm": 1.695090651512146, "learning_rate": 8.269314079422383e-05, "loss": 1.8155, "step": 4794 }, { "epoch": 3.4657039711191335, "grad_norm": 1.7225008010864258, "learning_rate": 8.267148014440433e-05, "loss": 1.702, "step": 4800 }, { "epoch": 3.4700361010830325, "grad_norm": 1.507054328918457, "learning_rate": 8.264981949458485e-05, "loss": 1.6995, "step": 4806 }, { "epoch": 3.4743682310469315, "grad_norm": 1.6530815362930298, "learning_rate": 8.262815884476535e-05, "loss": 1.7483, "step": 4812 }, { "epoch": 3.4787003610108305, "grad_norm": 1.6047580242156982, "learning_rate": 8.260649819494584e-05, "loss": 1.5983, "step": 4818 }, { "epoch": 3.483032490974729, "grad_norm": 1.5870919227600098, "learning_rate": 8.258483754512636e-05, "loss": 1.9267, "step": 4824 }, { "epoch": 3.487364620938628, "grad_norm": 1.3833556175231934, "learning_rate": 8.256317689530686e-05, "loss": 1.7854, "step": 4830 }, { "epoch": 3.491696750902527, "grad_norm": 1.5011537075042725, "learning_rate": 8.254151624548737e-05, "loss": 1.7391, "step": 4836 }, { "epoch": 3.496028880866426, "grad_norm": 1.5946050882339478, "learning_rate": 8.251985559566787e-05, "loss": 1.6968, "step": 4842 }, { "epoch": 3.500361010830325, "grad_norm": 1.6181870698928833, "learning_rate": 8.249819494584839e-05, "loss": 1.7225, "step": 4848 }, { "epoch": 3.504693140794224, "grad_norm": 1.3880207538604736, "learning_rate": 8.247653429602889e-05, "loss": 1.5755, "step": 4854 }, { "epoch": 3.509025270758123, "grad_norm": 1.9227714538574219, "learning_rate": 8.245487364620938e-05, "loss": 1.739, "step": 4860 }, { "epoch": 3.513357400722022, "grad_norm": 1.431706428527832, "learning_rate": 8.24332129963899e-05, "loss": 1.6947, "step": 4866 }, { "epoch": 3.517689530685921, "grad_norm": 1.7447054386138916, "learning_rate": 8.24115523465704e-05, "loss": 1.729, "step": 4872 }, { "epoch": 3.5220216606498194, "grad_norm": 1.614373803138733, "learning_rate": 8.238989169675092e-05, "loss": 1.8143, "step": 4878 }, { "epoch": 3.5263537906137183, "grad_norm": 1.4877253770828247, "learning_rate": 8.23682310469314e-05, "loss": 1.7277, "step": 4884 }, { "epoch": 3.5306859205776173, "grad_norm": 1.7306504249572754, "learning_rate": 8.234657039711192e-05, "loss": 1.9047, "step": 4890 }, { "epoch": 3.5350180505415163, "grad_norm": 1.654218077659607, "learning_rate": 8.232490974729243e-05, "loss": 1.691, "step": 4896 }, { "epoch": 3.5393501805054153, "grad_norm": 1.531161904335022, "learning_rate": 8.230324909747293e-05, "loss": 1.6893, "step": 4902 }, { "epoch": 3.543682310469314, "grad_norm": 1.6672862768173218, "learning_rate": 8.228158844765343e-05, "loss": 1.7313, "step": 4908 }, { "epoch": 3.548014440433213, "grad_norm": 1.5840322971343994, "learning_rate": 8.225992779783394e-05, "loss": 1.6674, "step": 4914 }, { "epoch": 3.5523465703971118, "grad_norm": 1.5155513286590576, "learning_rate": 8.223826714801445e-05, "loss": 1.8361, "step": 4920 }, { "epoch": 3.5566787003610107, "grad_norm": 1.6411936283111572, "learning_rate": 8.221660649819494e-05, "loss": 1.7938, "step": 4926 }, { "epoch": 3.5610108303249097, "grad_norm": 1.6390342712402344, "learning_rate": 8.219494584837546e-05, "loss": 1.7905, "step": 4932 }, { "epoch": 3.5653429602888087, "grad_norm": 1.7438820600509644, "learning_rate": 8.217328519855596e-05, "loss": 1.8307, "step": 4938 }, { "epoch": 3.5696750902527077, "grad_norm": 1.6976786851882935, "learning_rate": 8.215162454873646e-05, "loss": 1.7834, "step": 4944 }, { "epoch": 3.5740072202166067, "grad_norm": 1.6116440296173096, "learning_rate": 8.212996389891697e-05, "loss": 1.7684, "step": 4950 }, { "epoch": 3.5783393501805056, "grad_norm": 1.6113998889923096, "learning_rate": 8.210830324909747e-05, "loss": 1.7979, "step": 4956 }, { "epoch": 3.582671480144404, "grad_norm": 1.6461843252182007, "learning_rate": 8.208664259927799e-05, "loss": 1.7212, "step": 4962 }, { "epoch": 3.587003610108303, "grad_norm": 1.662481427192688, "learning_rate": 8.206498194945849e-05, "loss": 1.7977, "step": 4968 }, { "epoch": 3.591335740072202, "grad_norm": 1.5998501777648926, "learning_rate": 8.2043321299639e-05, "loss": 1.7017, "step": 4974 }, { "epoch": 3.595667870036101, "grad_norm": 1.62797212600708, "learning_rate": 8.20216606498195e-05, "loss": 1.8844, "step": 4980 }, { "epoch": 3.6, "grad_norm": 1.568451166152954, "learning_rate": 8.2e-05, "loss": 1.7324, "step": 4986 }, { "epoch": 3.604332129963899, "grad_norm": 1.5148454904556274, "learning_rate": 8.19783393501805e-05, "loss": 1.6724, "step": 4992 }, { "epoch": 3.6086642599277976, "grad_norm": 1.6874289512634277, "learning_rate": 8.195667870036101e-05, "loss": 1.8757, "step": 4998 }, { "epoch": 3.6129963898916966, "grad_norm": 1.6691734790802002, "learning_rate": 8.193501805054152e-05, "loss": 1.7703, "step": 5004 }, { "epoch": 3.6173285198555956, "grad_norm": 1.4583778381347656, "learning_rate": 8.191335740072203e-05, "loss": 1.8674, "step": 5010 }, { "epoch": 3.6216606498194945, "grad_norm": 1.6717382669448853, "learning_rate": 8.189169675090253e-05, "loss": 1.8673, "step": 5016 }, { "epoch": 3.6259927797833935, "grad_norm": 1.5273691415786743, "learning_rate": 8.187003610108303e-05, "loss": 1.707, "step": 5022 }, { "epoch": 3.6303249097472925, "grad_norm": 1.6732550859451294, "learning_rate": 8.184837545126354e-05, "loss": 1.6873, "step": 5028 }, { "epoch": 3.6346570397111915, "grad_norm": 1.4828213453292847, "learning_rate": 8.182671480144405e-05, "loss": 1.7422, "step": 5034 }, { "epoch": 3.6389891696750905, "grad_norm": 1.600921392440796, "learning_rate": 8.180505415162454e-05, "loss": 1.6914, "step": 5040 }, { "epoch": 3.6433212996389894, "grad_norm": 1.4632253646850586, "learning_rate": 8.178339350180506e-05, "loss": 1.6547, "step": 5046 }, { "epoch": 3.647653429602888, "grad_norm": 1.534578561782837, "learning_rate": 8.176173285198556e-05, "loss": 1.7686, "step": 5052 }, { "epoch": 3.651985559566787, "grad_norm": 1.621065378189087, "learning_rate": 8.174007220216608e-05, "loss": 1.776, "step": 5058 }, { "epoch": 3.656317689530686, "grad_norm": 1.5999253988265991, "learning_rate": 8.171841155234657e-05, "loss": 1.6162, "step": 5064 }, { "epoch": 3.660649819494585, "grad_norm": 1.7034722566604614, "learning_rate": 8.169675090252707e-05, "loss": 1.7845, "step": 5070 }, { "epoch": 3.664981949458484, "grad_norm": 1.48503839969635, "learning_rate": 8.167509025270759e-05, "loss": 1.7941, "step": 5076 }, { "epoch": 3.6693140794223824, "grad_norm": 1.6809515953063965, "learning_rate": 8.165342960288809e-05, "loss": 1.7361, "step": 5082 }, { "epoch": 3.6736462093862814, "grad_norm": 1.6679285764694214, "learning_rate": 8.16317689530686e-05, "loss": 1.696, "step": 5088 }, { "epoch": 3.6779783393501804, "grad_norm": 1.618849515914917, "learning_rate": 8.16101083032491e-05, "loss": 1.7964, "step": 5094 }, { "epoch": 3.6823104693140793, "grad_norm": 1.6021450757980347, "learning_rate": 8.158844765342962e-05, "loss": 1.6914, "step": 5100 }, { "epoch": 3.6866425992779783, "grad_norm": 1.5046802759170532, "learning_rate": 8.15667870036101e-05, "loss": 1.8463, "step": 5106 }, { "epoch": 3.6909747292418773, "grad_norm": 1.7005822658538818, "learning_rate": 8.154512635379061e-05, "loss": 1.7816, "step": 5112 }, { "epoch": 3.6953068592057763, "grad_norm": 1.55636727809906, "learning_rate": 8.152346570397113e-05, "loss": 1.6798, "step": 5118 }, { "epoch": 3.6996389891696753, "grad_norm": 1.6602107286453247, "learning_rate": 8.150180505415163e-05, "loss": 1.7763, "step": 5124 }, { "epoch": 3.7039711191335742, "grad_norm": 1.6221524477005005, "learning_rate": 8.148014440433213e-05, "loss": 1.8559, "step": 5130 }, { "epoch": 3.7083032490974728, "grad_norm": 1.5085234642028809, "learning_rate": 8.145848375451264e-05, "loss": 1.7613, "step": 5136 }, { "epoch": 3.7126353790613718, "grad_norm": 1.4024304151535034, "learning_rate": 8.143682310469315e-05, "loss": 1.6659, "step": 5142 }, { "epoch": 3.7169675090252707, "grad_norm": 1.567247986793518, "learning_rate": 8.141516245487366e-05, "loss": 1.7419, "step": 5148 }, { "epoch": 3.7212996389891697, "grad_norm": 1.352175235748291, "learning_rate": 8.139350180505415e-05, "loss": 1.7854, "step": 5154 }, { "epoch": 3.7256317689530687, "grad_norm": 1.6537840366363525, "learning_rate": 8.137184115523466e-05, "loss": 1.715, "step": 5160 }, { "epoch": 3.7299638989169677, "grad_norm": 1.4055896997451782, "learning_rate": 8.135018050541517e-05, "loss": 1.8849, "step": 5166 }, { "epoch": 3.734296028880866, "grad_norm": 1.5697423219680786, "learning_rate": 8.132851985559567e-05, "loss": 1.7579, "step": 5172 }, { "epoch": 3.738628158844765, "grad_norm": 1.566031813621521, "learning_rate": 8.130685920577617e-05, "loss": 1.7499, "step": 5178 }, { "epoch": 3.742960288808664, "grad_norm": 1.5549131631851196, "learning_rate": 8.128519855595669e-05, "loss": 1.6799, "step": 5184 }, { "epoch": 3.747292418772563, "grad_norm": 1.5588487386703491, "learning_rate": 8.126353790613719e-05, "loss": 1.6834, "step": 5190 }, { "epoch": 3.751624548736462, "grad_norm": 1.5660828351974487, "learning_rate": 8.12418772563177e-05, "loss": 1.7901, "step": 5196 }, { "epoch": 3.755956678700361, "grad_norm": 1.4894384145736694, "learning_rate": 8.12202166064982e-05, "loss": 1.7638, "step": 5202 }, { "epoch": 3.76028880866426, "grad_norm": 1.5432465076446533, "learning_rate": 8.11985559566787e-05, "loss": 1.64, "step": 5208 }, { "epoch": 3.764620938628159, "grad_norm": 1.6729116439819336, "learning_rate": 8.117689530685922e-05, "loss": 1.7878, "step": 5214 }, { "epoch": 3.768953068592058, "grad_norm": 1.5913641452789307, "learning_rate": 8.115523465703971e-05, "loss": 1.9421, "step": 5220 }, { "epoch": 3.7732851985559566, "grad_norm": 1.684883952140808, "learning_rate": 8.113357400722022e-05, "loss": 1.7975, "step": 5226 }, { "epoch": 3.7776173285198555, "grad_norm": 1.5214558839797974, "learning_rate": 8.111191335740073e-05, "loss": 1.7517, "step": 5232 }, { "epoch": 3.7819494584837545, "grad_norm": 1.489880084991455, "learning_rate": 8.109025270758123e-05, "loss": 1.741, "step": 5238 }, { "epoch": 3.7862815884476535, "grad_norm": 1.763386607170105, "learning_rate": 8.106859205776173e-05, "loss": 1.6486, "step": 5244 }, { "epoch": 3.7906137184115525, "grad_norm": 1.538878321647644, "learning_rate": 8.104693140794224e-05, "loss": 1.6525, "step": 5250 }, { "epoch": 3.794945848375451, "grad_norm": 1.8253357410430908, "learning_rate": 8.102527075812275e-05, "loss": 1.8435, "step": 5256 }, { "epoch": 3.79927797833935, "grad_norm": 1.6297709941864014, "learning_rate": 8.100361010830326e-05, "loss": 1.7643, "step": 5262 }, { "epoch": 3.803610108303249, "grad_norm": 1.6747561693191528, "learning_rate": 8.098194945848376e-05, "loss": 1.796, "step": 5268 }, { "epoch": 3.807942238267148, "grad_norm": 1.475718379020691, "learning_rate": 8.096028880866426e-05, "loss": 1.7305, "step": 5274 }, { "epoch": 3.812274368231047, "grad_norm": 1.6068952083587646, "learning_rate": 8.093862815884477e-05, "loss": 1.7595, "step": 5280 }, { "epoch": 3.816606498194946, "grad_norm": 1.4879449605941772, "learning_rate": 8.091696750902527e-05, "loss": 1.7774, "step": 5286 }, { "epoch": 3.820938628158845, "grad_norm": 1.5454704761505127, "learning_rate": 8.089530685920577e-05, "loss": 1.8487, "step": 5292 }, { "epoch": 3.825270758122744, "grad_norm": 1.7022215127944946, "learning_rate": 8.087364620938629e-05, "loss": 1.7101, "step": 5298 }, { "epoch": 3.829602888086643, "grad_norm": 1.5859121084213257, "learning_rate": 8.08519855595668e-05, "loss": 1.6937, "step": 5304 }, { "epoch": 3.8339350180505414, "grad_norm": 1.6372172832489014, "learning_rate": 8.08303249097473e-05, "loss": 1.7182, "step": 5310 }, { "epoch": 3.8382671480144404, "grad_norm": 1.8666167259216309, "learning_rate": 8.08086642599278e-05, "loss": 1.8689, "step": 5316 }, { "epoch": 3.8425992779783393, "grad_norm": 1.6148149967193604, "learning_rate": 8.07870036101083e-05, "loss": 1.7419, "step": 5322 }, { "epoch": 3.8469314079422383, "grad_norm": 1.627740502357483, "learning_rate": 8.076534296028882e-05, "loss": 1.819, "step": 5328 }, { "epoch": 3.8512635379061373, "grad_norm": 1.5696738958358765, "learning_rate": 8.074368231046931e-05, "loss": 1.8273, "step": 5334 }, { "epoch": 3.855595667870036, "grad_norm": 1.7058870792388916, "learning_rate": 8.072202166064983e-05, "loss": 1.6968, "step": 5340 }, { "epoch": 3.859927797833935, "grad_norm": 1.6600971221923828, "learning_rate": 8.070036101083033e-05, "loss": 1.8064, "step": 5346 }, { "epoch": 3.864259927797834, "grad_norm": 1.4972425699234009, "learning_rate": 8.067870036101083e-05, "loss": 1.7593, "step": 5352 }, { "epoch": 3.8685920577617328, "grad_norm": 1.5542570352554321, "learning_rate": 8.065703971119134e-05, "loss": 1.8333, "step": 5358 }, { "epoch": 3.8729241877256317, "grad_norm": 1.5219500064849854, "learning_rate": 8.063537906137184e-05, "loss": 1.8287, "step": 5364 }, { "epoch": 3.8772563176895307, "grad_norm": 1.4573575258255005, "learning_rate": 8.061371841155236e-05, "loss": 1.7942, "step": 5370 }, { "epoch": 3.8815884476534297, "grad_norm": 1.5931671857833862, "learning_rate": 8.059205776173286e-05, "loss": 1.8272, "step": 5376 }, { "epoch": 3.8859205776173287, "grad_norm": 1.5799790620803833, "learning_rate": 8.057039711191336e-05, "loss": 1.7582, "step": 5382 }, { "epoch": 3.8902527075812277, "grad_norm": 1.6085264682769775, "learning_rate": 8.054873646209387e-05, "loss": 1.6493, "step": 5388 }, { "epoch": 3.894584837545126, "grad_norm": 1.5485328435897827, "learning_rate": 8.052707581227438e-05, "loss": 1.653, "step": 5394 }, { "epoch": 3.898916967509025, "grad_norm": 1.70017409324646, "learning_rate": 8.050541516245487e-05, "loss": 1.8206, "step": 5400 }, { "epoch": 3.903249097472924, "grad_norm": 1.5802083015441895, "learning_rate": 8.048375451263538e-05, "loss": 1.6315, "step": 5406 }, { "epoch": 3.907581227436823, "grad_norm": 1.584958553314209, "learning_rate": 8.046209386281589e-05, "loss": 1.8065, "step": 5412 }, { "epoch": 3.911913357400722, "grad_norm": 1.6847898960113525, "learning_rate": 8.04404332129964e-05, "loss": 1.7355, "step": 5418 }, { "epoch": 3.916245487364621, "grad_norm": 1.6820820569992065, "learning_rate": 8.04187725631769e-05, "loss": 1.6307, "step": 5424 }, { "epoch": 3.9205776173285196, "grad_norm": 1.484824538230896, "learning_rate": 8.03971119133574e-05, "loss": 1.7884, "step": 5430 }, { "epoch": 3.9249097472924186, "grad_norm": 1.8032629489898682, "learning_rate": 8.037545126353792e-05, "loss": 1.7711, "step": 5436 }, { "epoch": 3.9292418772563176, "grad_norm": 1.5050805807113647, "learning_rate": 8.035379061371842e-05, "loss": 1.6476, "step": 5442 }, { "epoch": 3.9335740072202166, "grad_norm": 1.680967092514038, "learning_rate": 8.033212996389891e-05, "loss": 1.8872, "step": 5448 }, { "epoch": 3.9379061371841155, "grad_norm": 1.4918149709701538, "learning_rate": 8.031046931407943e-05, "loss": 1.6246, "step": 5454 }, { "epoch": 3.9422382671480145, "grad_norm": 1.685973048210144, "learning_rate": 8.028880866425993e-05, "loss": 1.6502, "step": 5460 }, { "epoch": 3.9465703971119135, "grad_norm": 1.5754412412643433, "learning_rate": 8.026714801444044e-05, "loss": 1.5816, "step": 5466 }, { "epoch": 3.9509025270758125, "grad_norm": 1.4636616706848145, "learning_rate": 8.024548736462094e-05, "loss": 1.7326, "step": 5472 }, { "epoch": 3.9552346570397114, "grad_norm": 1.4628803730010986, "learning_rate": 8.022382671480146e-05, "loss": 1.6701, "step": 5478 }, { "epoch": 3.95956678700361, "grad_norm": 1.5943557024002075, "learning_rate": 8.020216606498196e-05, "loss": 1.6756, "step": 5484 }, { "epoch": 3.963898916967509, "grad_norm": 1.5845967531204224, "learning_rate": 8.018050541516245e-05, "loss": 1.8297, "step": 5490 }, { "epoch": 3.968231046931408, "grad_norm": 1.5994409322738647, "learning_rate": 8.015884476534296e-05, "loss": 1.8136, "step": 5496 }, { "epoch": 3.972563176895307, "grad_norm": 1.6412502527236938, "learning_rate": 8.013718411552347e-05, "loss": 1.8397, "step": 5502 }, { "epoch": 3.976895306859206, "grad_norm": 1.5223716497421265, "learning_rate": 8.011552346570398e-05, "loss": 1.6729, "step": 5508 }, { "epoch": 3.9812274368231044, "grad_norm": 1.6280032396316528, "learning_rate": 8.009386281588447e-05, "loss": 1.893, "step": 5514 }, { "epoch": 3.9855595667870034, "grad_norm": 1.6926552057266235, "learning_rate": 8.007220216606499e-05, "loss": 1.5927, "step": 5520 }, { "epoch": 3.9898916967509024, "grad_norm": 1.5570157766342163, "learning_rate": 8.00505415162455e-05, "loss": 1.64, "step": 5526 }, { "epoch": 3.9942238267148014, "grad_norm": 1.6186165809631348, "learning_rate": 8.0028880866426e-05, "loss": 1.7016, "step": 5532 }, { "epoch": 3.9985559566787003, "grad_norm": 1.7071150541305542, "learning_rate": 8.00072202166065e-05, "loss": 1.7091, "step": 5538 }, { "epoch": 4.002888086642599, "grad_norm": 1.4383127689361572, "learning_rate": 7.9985559566787e-05, "loss": 1.6498, "step": 5544 }, { "epoch": 4.007220216606498, "grad_norm": 1.6690502166748047, "learning_rate": 7.996389891696752e-05, "loss": 1.6714, "step": 5550 }, { "epoch": 4.011552346570397, "grad_norm": 1.6955111026763916, "learning_rate": 7.994223826714802e-05, "loss": 1.6188, "step": 5556 }, { "epoch": 4.015884476534296, "grad_norm": 1.5323209762573242, "learning_rate": 7.992057761732853e-05, "loss": 1.8299, "step": 5562 }, { "epoch": 4.020216606498195, "grad_norm": 1.4871385097503662, "learning_rate": 7.989891696750903e-05, "loss": 1.6109, "step": 5568 }, { "epoch": 4.024548736462094, "grad_norm": 1.7349871397018433, "learning_rate": 7.987725631768953e-05, "loss": 1.7277, "step": 5574 }, { "epoch": 4.028880866425993, "grad_norm": 1.5424202680587769, "learning_rate": 7.985559566787004e-05, "loss": 1.7427, "step": 5580 }, { "epoch": 4.033212996389891, "grad_norm": 1.6044737100601196, "learning_rate": 7.983393501805054e-05, "loss": 1.7218, "step": 5586 }, { "epoch": 4.03754512635379, "grad_norm": 1.7512400150299072, "learning_rate": 7.981227436823106e-05, "loss": 1.687, "step": 5592 }, { "epoch": 4.041877256317689, "grad_norm": 1.7815380096435547, "learning_rate": 7.979061371841156e-05, "loss": 1.7195, "step": 5598 }, { "epoch": 4.046209386281588, "grad_norm": 1.6185029745101929, "learning_rate": 7.976895306859205e-05, "loss": 1.5878, "step": 5604 }, { "epoch": 4.050541516245487, "grad_norm": 1.5741853713989258, "learning_rate": 7.974729241877257e-05, "loss": 1.6049, "step": 5610 }, { "epoch": 4.054873646209386, "grad_norm": 1.594574213027954, "learning_rate": 7.972563176895307e-05, "loss": 1.7445, "step": 5616 }, { "epoch": 4.059205776173285, "grad_norm": 1.5438119173049927, "learning_rate": 7.970397111913359e-05, "loss": 1.6961, "step": 5622 }, { "epoch": 4.063537906137184, "grad_norm": 1.7096153497695923, "learning_rate": 7.968231046931408e-05, "loss": 1.7878, "step": 5628 }, { "epoch": 4.067870036101083, "grad_norm": 1.4786994457244873, "learning_rate": 7.966064981949459e-05, "loss": 1.6616, "step": 5634 }, { "epoch": 4.072202166064982, "grad_norm": 1.68770432472229, "learning_rate": 7.96389891696751e-05, "loss": 1.8264, "step": 5640 }, { "epoch": 4.076534296028881, "grad_norm": 2.120720386505127, "learning_rate": 7.96173285198556e-05, "loss": 1.7825, "step": 5646 }, { "epoch": 4.08086642599278, "grad_norm": 1.7693120241165161, "learning_rate": 7.95956678700361e-05, "loss": 1.6907, "step": 5652 }, { "epoch": 4.085198555956679, "grad_norm": 1.782841682434082, "learning_rate": 7.95740072202166e-05, "loss": 1.7197, "step": 5658 }, { "epoch": 4.089530685920578, "grad_norm": 1.697631597518921, "learning_rate": 7.955234657039712e-05, "loss": 1.7071, "step": 5664 }, { "epoch": 4.093862815884476, "grad_norm": 1.5238312482833862, "learning_rate": 7.953068592057761e-05, "loss": 1.7049, "step": 5670 }, { "epoch": 4.098194945848375, "grad_norm": 1.7160964012145996, "learning_rate": 7.950902527075813e-05, "loss": 1.6148, "step": 5676 }, { "epoch": 4.102527075812274, "grad_norm": 1.6714850664138794, "learning_rate": 7.948736462093863e-05, "loss": 1.721, "step": 5682 }, { "epoch": 4.106859205776173, "grad_norm": 1.5852328538894653, "learning_rate": 7.946570397111914e-05, "loss": 1.612, "step": 5688 }, { "epoch": 4.111191335740072, "grad_norm": 1.5865859985351562, "learning_rate": 7.944404332129964e-05, "loss": 1.5919, "step": 5694 }, { "epoch": 4.115523465703971, "grad_norm": 1.4504220485687256, "learning_rate": 7.942238267148014e-05, "loss": 1.6702, "step": 5700 }, { "epoch": 4.11985559566787, "grad_norm": 1.7038459777832031, "learning_rate": 7.940072202166066e-05, "loss": 1.6766, "step": 5706 }, { "epoch": 4.124187725631769, "grad_norm": 1.7308201789855957, "learning_rate": 7.937906137184116e-05, "loss": 1.7537, "step": 5712 }, { "epoch": 4.128519855595668, "grad_norm": 1.7125248908996582, "learning_rate": 7.935740072202167e-05, "loss": 1.6221, "step": 5718 }, { "epoch": 4.132851985559567, "grad_norm": 1.5896432399749756, "learning_rate": 7.933574007220217e-05, "loss": 1.6507, "step": 5724 }, { "epoch": 4.137184115523466, "grad_norm": 1.8399280309677124, "learning_rate": 7.931407942238267e-05, "loss": 1.7452, "step": 5730 }, { "epoch": 4.141516245487365, "grad_norm": 1.669938325881958, "learning_rate": 7.929241877256318e-05, "loss": 1.6962, "step": 5736 }, { "epoch": 4.145848375451264, "grad_norm": 1.6299235820770264, "learning_rate": 7.927075812274368e-05, "loss": 1.7517, "step": 5742 }, { "epoch": 4.150180505415163, "grad_norm": 1.6390680074691772, "learning_rate": 7.92490974729242e-05, "loss": 1.7144, "step": 5748 }, { "epoch": 4.154512635379062, "grad_norm": 1.6312898397445679, "learning_rate": 7.92274368231047e-05, "loss": 1.7346, "step": 5754 }, { "epoch": 4.15884476534296, "grad_norm": 1.8234426975250244, "learning_rate": 7.92057761732852e-05, "loss": 1.7874, "step": 5760 }, { "epoch": 4.163176895306859, "grad_norm": 1.7259629964828491, "learning_rate": 7.91841155234657e-05, "loss": 1.6702, "step": 5766 }, { "epoch": 4.167509025270758, "grad_norm": 1.8502966165542603, "learning_rate": 7.916245487364621e-05, "loss": 1.6571, "step": 5772 }, { "epoch": 4.171841155234657, "grad_norm": 1.6337683200836182, "learning_rate": 7.914079422382672e-05, "loss": 1.6197, "step": 5778 }, { "epoch": 4.176173285198556, "grad_norm": 1.7422279119491577, "learning_rate": 7.911913357400721e-05, "loss": 1.6108, "step": 5784 }, { "epoch": 4.180505415162455, "grad_norm": 1.734500527381897, "learning_rate": 7.909747292418773e-05, "loss": 1.5691, "step": 5790 }, { "epoch": 4.184837545126354, "grad_norm": 1.7699047327041626, "learning_rate": 7.907581227436823e-05, "loss": 1.7755, "step": 5796 }, { "epoch": 4.189169675090253, "grad_norm": 1.4985324144363403, "learning_rate": 7.905415162454875e-05, "loss": 1.6601, "step": 5802 }, { "epoch": 4.193501805054152, "grad_norm": 1.7704424858093262, "learning_rate": 7.903249097472924e-05, "loss": 1.6596, "step": 5808 }, { "epoch": 4.197833935018051, "grad_norm": 1.5737330913543701, "learning_rate": 7.901083032490974e-05, "loss": 1.7079, "step": 5814 }, { "epoch": 4.20216606498195, "grad_norm": 1.5857319831848145, "learning_rate": 7.898916967509026e-05, "loss": 1.6686, "step": 5820 }, { "epoch": 4.206498194945849, "grad_norm": 1.7224262952804565, "learning_rate": 7.896750902527076e-05, "loss": 1.6042, "step": 5826 }, { "epoch": 4.210830324909748, "grad_norm": 1.69579017162323, "learning_rate": 7.894584837545127e-05, "loss": 1.7127, "step": 5832 }, { "epoch": 4.215162454873647, "grad_norm": 1.8030771017074585, "learning_rate": 7.892418772563177e-05, "loss": 1.7148, "step": 5838 }, { "epoch": 4.219494584837545, "grad_norm": 1.8374943733215332, "learning_rate": 7.890252707581229e-05, "loss": 1.7161, "step": 5844 }, { "epoch": 4.223826714801444, "grad_norm": 1.744922161102295, "learning_rate": 7.888086642599278e-05, "loss": 1.7109, "step": 5850 }, { "epoch": 4.228158844765343, "grad_norm": 1.6115013360977173, "learning_rate": 7.885920577617328e-05, "loss": 1.7863, "step": 5856 }, { "epoch": 4.232490974729242, "grad_norm": 1.6358370780944824, "learning_rate": 7.88375451263538e-05, "loss": 1.7325, "step": 5862 }, { "epoch": 4.236823104693141, "grad_norm": 1.7013176679611206, "learning_rate": 7.88158844765343e-05, "loss": 1.6526, "step": 5868 }, { "epoch": 4.24115523465704, "grad_norm": 1.924896240234375, "learning_rate": 7.87942238267148e-05, "loss": 1.7905, "step": 5874 }, { "epoch": 4.245487364620939, "grad_norm": 1.737907886505127, "learning_rate": 7.87725631768953e-05, "loss": 1.7574, "step": 5880 }, { "epoch": 4.2498194945848375, "grad_norm": 1.4201635122299194, "learning_rate": 7.875090252707582e-05, "loss": 1.667, "step": 5886 }, { "epoch": 4.2541516245487365, "grad_norm": 1.6169565916061401, "learning_rate": 7.872924187725633e-05, "loss": 1.5995, "step": 5892 }, { "epoch": 4.2584837545126355, "grad_norm": 1.7920392751693726, "learning_rate": 7.870758122743682e-05, "loss": 1.8019, "step": 5898 }, { "epoch": 4.2628158844765345, "grad_norm": 1.6815756559371948, "learning_rate": 7.868592057761733e-05, "loss": 1.7667, "step": 5904 }, { "epoch": 4.2671480144404335, "grad_norm": 1.6518813371658325, "learning_rate": 7.866425992779784e-05, "loss": 1.69, "step": 5910 }, { "epoch": 4.271480144404332, "grad_norm": 1.7253786325454712, "learning_rate": 7.864259927797834e-05, "loss": 1.7221, "step": 5916 }, { "epoch": 4.275812274368231, "grad_norm": 1.5930122137069702, "learning_rate": 7.862093862815884e-05, "loss": 1.6923, "step": 5922 }, { "epoch": 4.28014440433213, "grad_norm": 1.6287250518798828, "learning_rate": 7.859927797833936e-05, "loss": 1.6752, "step": 5928 }, { "epoch": 4.2844765342960285, "grad_norm": 1.6803449392318726, "learning_rate": 7.857761732851986e-05, "loss": 1.5622, "step": 5934 }, { "epoch": 4.2888086642599275, "grad_norm": 1.7550612688064575, "learning_rate": 7.855595667870037e-05, "loss": 1.7048, "step": 5940 }, { "epoch": 4.293140794223826, "grad_norm": 1.7911068201065063, "learning_rate": 7.853429602888087e-05, "loss": 1.8035, "step": 5946 }, { "epoch": 4.297472924187725, "grad_norm": 1.7555116415023804, "learning_rate": 7.851263537906137e-05, "loss": 1.6963, "step": 5952 }, { "epoch": 4.301805054151624, "grad_norm": 1.9857341051101685, "learning_rate": 7.849097472924189e-05, "loss": 1.8744, "step": 5958 }, { "epoch": 4.306137184115523, "grad_norm": 1.604610562324524, "learning_rate": 7.846931407942238e-05, "loss": 1.5674, "step": 5964 }, { "epoch": 4.310469314079422, "grad_norm": 1.711546778678894, "learning_rate": 7.84476534296029e-05, "loss": 1.7386, "step": 5970 }, { "epoch": 4.314801444043321, "grad_norm": 1.8602561950683594, "learning_rate": 7.84259927797834e-05, "loss": 1.7205, "step": 5976 }, { "epoch": 4.31913357400722, "grad_norm": 1.6639668941497803, "learning_rate": 7.84043321299639e-05, "loss": 1.7825, "step": 5982 }, { "epoch": 4.323465703971119, "grad_norm": 1.4736350774765015, "learning_rate": 7.83826714801444e-05, "loss": 1.6392, "step": 5988 }, { "epoch": 4.327797833935018, "grad_norm": 1.7398465871810913, "learning_rate": 7.836101083032491e-05, "loss": 1.7183, "step": 5994 }, { "epoch": 4.332129963898917, "grad_norm": 1.6983509063720703, "learning_rate": 7.833935018050543e-05, "loss": 1.6364, "step": 6000 }, { "epoch": 4.336462093862816, "grad_norm": 2.000410318374634, "learning_rate": 7.831768953068593e-05, "loss": 1.6712, "step": 6006 }, { "epoch": 4.340794223826715, "grad_norm": 1.662174940109253, "learning_rate": 7.829602888086643e-05, "loss": 1.7144, "step": 6012 }, { "epoch": 4.345126353790613, "grad_norm": 1.5986292362213135, "learning_rate": 7.827436823104694e-05, "loss": 1.7143, "step": 6018 }, { "epoch": 4.349458483754512, "grad_norm": 1.841261863708496, "learning_rate": 7.825270758122744e-05, "loss": 1.7214, "step": 6024 }, { "epoch": 4.353790613718411, "grad_norm": 1.8462514877319336, "learning_rate": 7.823104693140794e-05, "loss": 1.6322, "step": 6030 }, { "epoch": 4.35812274368231, "grad_norm": 1.6749650239944458, "learning_rate": 7.820938628158844e-05, "loss": 1.5668, "step": 6036 }, { "epoch": 4.362454873646209, "grad_norm": 1.5724742412567139, "learning_rate": 7.818772563176896e-05, "loss": 1.6526, "step": 6042 }, { "epoch": 4.366787003610108, "grad_norm": 1.9240885972976685, "learning_rate": 7.816606498194946e-05, "loss": 1.6989, "step": 6048 }, { "epoch": 4.371119133574007, "grad_norm": 1.804455280303955, "learning_rate": 7.814440433212997e-05, "loss": 1.6803, "step": 6054 }, { "epoch": 4.375451263537906, "grad_norm": 1.671733021736145, "learning_rate": 7.812274368231047e-05, "loss": 1.7184, "step": 6060 }, { "epoch": 4.379783393501805, "grad_norm": 1.6550850868225098, "learning_rate": 7.810108303249097e-05, "loss": 1.59, "step": 6066 }, { "epoch": 4.384115523465704, "grad_norm": 1.6431257724761963, "learning_rate": 7.807942238267149e-05, "loss": 1.7578, "step": 6072 }, { "epoch": 4.388447653429603, "grad_norm": 1.7343659400939941, "learning_rate": 7.805776173285198e-05, "loss": 1.7324, "step": 6078 }, { "epoch": 4.392779783393502, "grad_norm": 1.7962208986282349, "learning_rate": 7.80361010830325e-05, "loss": 1.843, "step": 6084 }, { "epoch": 4.397111913357401, "grad_norm": 1.7537035942077637, "learning_rate": 7.8014440433213e-05, "loss": 1.7055, "step": 6090 }, { "epoch": 4.4014440433213, "grad_norm": 1.7747939825057983, "learning_rate": 7.79927797833935e-05, "loss": 1.7326, "step": 6096 }, { "epoch": 4.405776173285199, "grad_norm": 1.643624186515808, "learning_rate": 7.797111913357401e-05, "loss": 1.6692, "step": 6102 }, { "epoch": 4.410108303249097, "grad_norm": 1.53639554977417, "learning_rate": 7.794945848375451e-05, "loss": 1.655, "step": 6108 }, { "epoch": 4.414440433212996, "grad_norm": 1.8310946226119995, "learning_rate": 7.792779783393503e-05, "loss": 1.7255, "step": 6114 }, { "epoch": 4.418772563176895, "grad_norm": 1.8531361818313599, "learning_rate": 7.790613718411553e-05, "loss": 1.7894, "step": 6120 }, { "epoch": 4.423104693140794, "grad_norm": 1.5105721950531006, "learning_rate": 7.788447653429603e-05, "loss": 1.5875, "step": 6126 }, { "epoch": 4.427436823104693, "grad_norm": 1.6683555841445923, "learning_rate": 7.786281588447654e-05, "loss": 1.6446, "step": 6132 }, { "epoch": 4.431768953068592, "grad_norm": 1.7124634981155396, "learning_rate": 7.784115523465705e-05, "loss": 1.5304, "step": 6138 }, { "epoch": 4.436101083032491, "grad_norm": 1.7297983169555664, "learning_rate": 7.781949458483754e-05, "loss": 1.6405, "step": 6144 }, { "epoch": 4.44043321299639, "grad_norm": 1.7115557193756104, "learning_rate": 7.779783393501805e-05, "loss": 1.7894, "step": 6150 }, { "epoch": 4.444765342960289, "grad_norm": 1.8467234373092651, "learning_rate": 7.777617328519856e-05, "loss": 1.7545, "step": 6156 }, { "epoch": 4.449097472924188, "grad_norm": 1.7932335138320923, "learning_rate": 7.775451263537907e-05, "loss": 1.7432, "step": 6162 }, { "epoch": 4.453429602888087, "grad_norm": 1.8021281957626343, "learning_rate": 7.773285198555957e-05, "loss": 1.755, "step": 6168 }, { "epoch": 4.457761732851986, "grad_norm": 1.6702615022659302, "learning_rate": 7.771119133574007e-05, "loss": 1.6085, "step": 6174 }, { "epoch": 4.462093862815885, "grad_norm": 1.7161496877670288, "learning_rate": 7.768953068592059e-05, "loss": 1.7449, "step": 6180 }, { "epoch": 4.466425992779784, "grad_norm": 1.775092363357544, "learning_rate": 7.766787003610109e-05, "loss": 1.8322, "step": 6186 }, { "epoch": 4.470758122743682, "grad_norm": 1.693324327468872, "learning_rate": 7.764620938628158e-05, "loss": 1.694, "step": 6192 }, { "epoch": 4.475090252707581, "grad_norm": 1.5456526279449463, "learning_rate": 7.76245487364621e-05, "loss": 1.7554, "step": 6198 }, { "epoch": 4.47942238267148, "grad_norm": 1.7538892030715942, "learning_rate": 7.76028880866426e-05, "loss": 1.7445, "step": 6204 }, { "epoch": 4.483754512635379, "grad_norm": 1.6797393560409546, "learning_rate": 7.75812274368231e-05, "loss": 1.6421, "step": 6210 }, { "epoch": 4.488086642599278, "grad_norm": 1.5838606357574463, "learning_rate": 7.755956678700361e-05, "loss": 1.7783, "step": 6216 }, { "epoch": 4.492418772563177, "grad_norm": 1.5768022537231445, "learning_rate": 7.753790613718413e-05, "loss": 1.7, "step": 6222 }, { "epoch": 4.496750902527076, "grad_norm": 1.6184730529785156, "learning_rate": 7.751624548736463e-05, "loss": 1.7369, "step": 6228 }, { "epoch": 4.501083032490975, "grad_norm": 1.6981490850448608, "learning_rate": 7.749458483754512e-05, "loss": 1.6827, "step": 6234 }, { "epoch": 4.505415162454874, "grad_norm": 1.6360697746276855, "learning_rate": 7.747292418772564e-05, "loss": 1.6882, "step": 6240 }, { "epoch": 4.509747292418773, "grad_norm": 1.7539596557617188, "learning_rate": 7.745126353790614e-05, "loss": 1.7492, "step": 6246 }, { "epoch": 4.514079422382672, "grad_norm": 1.6087987422943115, "learning_rate": 7.742960288808666e-05, "loss": 1.5621, "step": 6252 }, { "epoch": 4.518411552346571, "grad_norm": 1.9913228750228882, "learning_rate": 7.740794223826715e-05, "loss": 1.783, "step": 6258 }, { "epoch": 4.52274368231047, "grad_norm": 1.6948806047439575, "learning_rate": 7.738628158844766e-05, "loss": 1.7258, "step": 6264 }, { "epoch": 4.527075812274369, "grad_norm": 1.54940927028656, "learning_rate": 7.736462093862817e-05, "loss": 1.7179, "step": 6270 }, { "epoch": 4.531407942238268, "grad_norm": 1.9192289113998413, "learning_rate": 7.734296028880867e-05, "loss": 1.7595, "step": 6276 }, { "epoch": 4.535740072202166, "grad_norm": 1.8084369897842407, "learning_rate": 7.732129963898917e-05, "loss": 1.6911, "step": 6282 }, { "epoch": 4.540072202166065, "grad_norm": 1.8283580541610718, "learning_rate": 7.729963898916968e-05, "loss": 1.6888, "step": 6288 }, { "epoch": 4.544404332129964, "grad_norm": 1.681138277053833, "learning_rate": 7.727797833935019e-05, "loss": 1.7081, "step": 6294 }, { "epoch": 4.548736462093863, "grad_norm": 1.7156180143356323, "learning_rate": 7.72563176895307e-05, "loss": 1.76, "step": 6300 }, { "epoch": 4.553068592057762, "grad_norm": 1.5604966878890991, "learning_rate": 7.72346570397112e-05, "loss": 1.7476, "step": 6306 }, { "epoch": 4.557400722021661, "grad_norm": 1.5356557369232178, "learning_rate": 7.72129963898917e-05, "loss": 1.6155, "step": 6312 }, { "epoch": 4.5617328519855596, "grad_norm": 1.5977967977523804, "learning_rate": 7.71913357400722e-05, "loss": 1.7924, "step": 6318 }, { "epoch": 4.5660649819494585, "grad_norm": 1.7557899951934814, "learning_rate": 7.716967509025271e-05, "loss": 1.6878, "step": 6324 }, { "epoch": 4.5703971119133575, "grad_norm": 1.8078233003616333, "learning_rate": 7.714801444043321e-05, "loss": 1.789, "step": 6330 }, { "epoch": 4.5747292418772565, "grad_norm": 1.6310216188430786, "learning_rate": 7.712635379061373e-05, "loss": 1.7325, "step": 6336 }, { "epoch": 4.5790613718411555, "grad_norm": 1.6880606412887573, "learning_rate": 7.710469314079423e-05, "loss": 1.6669, "step": 6342 }, { "epoch": 4.5833935018050544, "grad_norm": 1.713457465171814, "learning_rate": 7.708303249097473e-05, "loss": 1.6319, "step": 6348 }, { "epoch": 4.587725631768953, "grad_norm": 1.6874985694885254, "learning_rate": 7.706137184115524e-05, "loss": 1.7157, "step": 6354 }, { "epoch": 4.5920577617328515, "grad_norm": 1.5871684551239014, "learning_rate": 7.703971119133574e-05, "loss": 1.7207, "step": 6360 }, { "epoch": 4.5963898916967505, "grad_norm": 1.6265347003936768, "learning_rate": 7.701805054151626e-05, "loss": 1.6748, "step": 6366 }, { "epoch": 4.6007220216606495, "grad_norm": 1.7521090507507324, "learning_rate": 7.699638989169675e-05, "loss": 1.5997, "step": 6372 }, { "epoch": 4.6050541516245485, "grad_norm": 1.8057323694229126, "learning_rate": 7.697472924187726e-05, "loss": 1.7103, "step": 6378 }, { "epoch": 4.609386281588447, "grad_norm": 1.7194278240203857, "learning_rate": 7.695306859205777e-05, "loss": 1.6147, "step": 6384 }, { "epoch": 4.613718411552346, "grad_norm": 1.504032850265503, "learning_rate": 7.693140794223827e-05, "loss": 1.6684, "step": 6390 }, { "epoch": 4.618050541516245, "grad_norm": 1.8034778833389282, "learning_rate": 7.690974729241877e-05, "loss": 1.6635, "step": 6396 }, { "epoch": 4.622382671480144, "grad_norm": 1.631672739982605, "learning_rate": 7.688808664259928e-05, "loss": 1.7039, "step": 6402 }, { "epoch": 4.626714801444043, "grad_norm": 1.8789169788360596, "learning_rate": 7.68664259927798e-05, "loss": 1.6211, "step": 6408 }, { "epoch": 4.631046931407942, "grad_norm": 1.874737024307251, "learning_rate": 7.684476534296028e-05, "loss": 1.8656, "step": 6414 }, { "epoch": 4.635379061371841, "grad_norm": 1.8787630796432495, "learning_rate": 7.68231046931408e-05, "loss": 1.6178, "step": 6420 }, { "epoch": 4.63971119133574, "grad_norm": 1.638251781463623, "learning_rate": 7.68014440433213e-05, "loss": 1.6484, "step": 6426 }, { "epoch": 4.644043321299639, "grad_norm": 1.7027487754821777, "learning_rate": 7.677978339350182e-05, "loss": 1.5933, "step": 6432 }, { "epoch": 4.648375451263538, "grad_norm": 1.8651024103164673, "learning_rate": 7.675812274368231e-05, "loss": 1.8201, "step": 6438 }, { "epoch": 4.652707581227437, "grad_norm": 1.7897592782974243, "learning_rate": 7.673646209386281e-05, "loss": 1.7217, "step": 6444 }, { "epoch": 4.657039711191336, "grad_norm": 1.7765439748764038, "learning_rate": 7.671480144404333e-05, "loss": 1.7937, "step": 6450 }, { "epoch": 4.661371841155234, "grad_norm": 1.7860112190246582, "learning_rate": 7.669314079422383e-05, "loss": 1.7321, "step": 6456 }, { "epoch": 4.665703971119133, "grad_norm": 1.6773638725280762, "learning_rate": 7.667148014440434e-05, "loss": 1.6566, "step": 6462 }, { "epoch": 4.670036101083032, "grad_norm": 1.7579811811447144, "learning_rate": 7.664981949458484e-05, "loss": 1.7255, "step": 6468 }, { "epoch": 4.674368231046931, "grad_norm": 1.6748288869857788, "learning_rate": 7.662815884476536e-05, "loss": 1.6507, "step": 6474 }, { "epoch": 4.67870036101083, "grad_norm": 1.5777277946472168, "learning_rate": 7.660649819494585e-05, "loss": 1.6931, "step": 6480 }, { "epoch": 4.683032490974729, "grad_norm": 1.597028374671936, "learning_rate": 7.658483754512635e-05, "loss": 1.6442, "step": 6486 }, { "epoch": 4.687364620938628, "grad_norm": 1.8182154893875122, "learning_rate": 7.656317689530687e-05, "loss": 1.7048, "step": 6492 }, { "epoch": 4.691696750902527, "grad_norm": 1.7548727989196777, "learning_rate": 7.654151624548737e-05, "loss": 1.6717, "step": 6498 }, { "epoch": 4.696028880866426, "grad_norm": 1.9332393407821655, "learning_rate": 7.651985559566787e-05, "loss": 1.7035, "step": 6504 }, { "epoch": 4.700361010830325, "grad_norm": 1.922795295715332, "learning_rate": 7.649819494584838e-05, "loss": 1.6489, "step": 6510 }, { "epoch": 4.704693140794224, "grad_norm": 1.7331790924072266, "learning_rate": 7.647653429602889e-05, "loss": 1.6752, "step": 6516 }, { "epoch": 4.709025270758123, "grad_norm": 1.663933277130127, "learning_rate": 7.64548736462094e-05, "loss": 1.6847, "step": 6522 }, { "epoch": 4.713357400722022, "grad_norm": 1.7516316175460815, "learning_rate": 7.643321299638989e-05, "loss": 1.6548, "step": 6528 }, { "epoch": 4.71768953068592, "grad_norm": 1.7978991270065308, "learning_rate": 7.64115523465704e-05, "loss": 1.6855, "step": 6534 }, { "epoch": 4.722021660649819, "grad_norm": 1.8310818672180176, "learning_rate": 7.63898916967509e-05, "loss": 1.7529, "step": 6540 }, { "epoch": 4.726353790613718, "grad_norm": 1.4914026260375977, "learning_rate": 7.636823104693142e-05, "loss": 1.7497, "step": 6546 }, { "epoch": 4.730685920577617, "grad_norm": 1.7645925283432007, "learning_rate": 7.634657039711191e-05, "loss": 1.7729, "step": 6552 }, { "epoch": 4.735018050541516, "grad_norm": 1.8393405675888062, "learning_rate": 7.632490974729243e-05, "loss": 1.6792, "step": 6558 }, { "epoch": 4.739350180505415, "grad_norm": 1.692963719367981, "learning_rate": 7.630324909747293e-05, "loss": 1.7304, "step": 6564 }, { "epoch": 4.743682310469314, "grad_norm": 1.646930456161499, "learning_rate": 7.628158844765344e-05, "loss": 1.7616, "step": 6570 }, { "epoch": 4.748014440433213, "grad_norm": 1.7866606712341309, "learning_rate": 7.625992779783394e-05, "loss": 1.6773, "step": 6576 }, { "epoch": 4.752346570397112, "grad_norm": 1.8032617568969727, "learning_rate": 7.623826714801444e-05, "loss": 1.6835, "step": 6582 }, { "epoch": 4.756678700361011, "grad_norm": 1.6876643896102905, "learning_rate": 7.621660649819496e-05, "loss": 1.6728, "step": 6588 }, { "epoch": 4.76101083032491, "grad_norm": 1.9263309240341187, "learning_rate": 7.619494584837545e-05, "loss": 1.6752, "step": 6594 }, { "epoch": 4.765342960288809, "grad_norm": 1.620936393737793, "learning_rate": 7.617328519855595e-05, "loss": 1.5934, "step": 6600 }, { "epoch": 4.769675090252708, "grad_norm": 1.8353679180145264, "learning_rate": 7.615162454873647e-05, "loss": 1.7914, "step": 6606 }, { "epoch": 4.774007220216607, "grad_norm": 1.6800568103790283, "learning_rate": 7.612996389891697e-05, "loss": 1.6834, "step": 6612 }, { "epoch": 4.778339350180506, "grad_norm": 1.7549793720245361, "learning_rate": 7.610830324909747e-05, "loss": 1.6395, "step": 6618 }, { "epoch": 4.782671480144405, "grad_norm": 1.9787557125091553, "learning_rate": 7.608664259927798e-05, "loss": 1.8339, "step": 6624 }, { "epoch": 4.787003610108303, "grad_norm": 1.8703575134277344, "learning_rate": 7.60649819494585e-05, "loss": 1.6711, "step": 6630 }, { "epoch": 4.791335740072202, "grad_norm": 1.637927770614624, "learning_rate": 7.6043321299639e-05, "loss": 1.8339, "step": 6636 }, { "epoch": 4.795667870036101, "grad_norm": 1.5050417184829712, "learning_rate": 7.602166064981949e-05, "loss": 1.6588, "step": 6642 }, { "epoch": 4.8, "grad_norm": 1.642702341079712, "learning_rate": 7.6e-05, "loss": 1.688, "step": 6648 }, { "epoch": 4.804332129963899, "grad_norm": 1.6545082330703735, "learning_rate": 7.597833935018051e-05, "loss": 1.576, "step": 6654 }, { "epoch": 4.808664259927798, "grad_norm": 1.5883156061172485, "learning_rate": 7.595667870036101e-05, "loss": 1.8414, "step": 6660 }, { "epoch": 4.812996389891697, "grad_norm": 1.5606862306594849, "learning_rate": 7.593501805054151e-05, "loss": 1.6663, "step": 6666 }, { "epoch": 4.817328519855596, "grad_norm": 1.863951563835144, "learning_rate": 7.591335740072203e-05, "loss": 1.8012, "step": 6672 }, { "epoch": 4.821660649819495, "grad_norm": 1.8710488080978394, "learning_rate": 7.589169675090253e-05, "loss": 1.7692, "step": 6678 }, { "epoch": 4.825992779783394, "grad_norm": 1.5971882343292236, "learning_rate": 7.587003610108304e-05, "loss": 1.7444, "step": 6684 }, { "epoch": 4.830324909747293, "grad_norm": 1.6892380714416504, "learning_rate": 7.584837545126354e-05, "loss": 1.6682, "step": 6690 }, { "epoch": 4.834657039711192, "grad_norm": 1.7566777467727661, "learning_rate": 7.582671480144404e-05, "loss": 1.6699, "step": 6696 }, { "epoch": 4.838989169675091, "grad_norm": 1.7562795877456665, "learning_rate": 7.580505415162456e-05, "loss": 1.6834, "step": 6702 }, { "epoch": 4.843321299638989, "grad_norm": 1.8356683254241943, "learning_rate": 7.578339350180505e-05, "loss": 1.727, "step": 6708 }, { "epoch": 4.847653429602888, "grad_norm": 1.6499661207199097, "learning_rate": 7.576173285198557e-05, "loss": 1.7079, "step": 6714 }, { "epoch": 4.851985559566787, "grad_norm": 1.4443529844284058, "learning_rate": 7.574007220216607e-05, "loss": 1.5479, "step": 6720 }, { "epoch": 4.856317689530686, "grad_norm": 1.633927583694458, "learning_rate": 7.571841155234657e-05, "loss": 1.7522, "step": 6726 }, { "epoch": 4.860649819494585, "grad_norm": 1.4663947820663452, "learning_rate": 7.569675090252708e-05, "loss": 1.5697, "step": 6732 }, { "epoch": 4.864981949458484, "grad_norm": 1.5066217184066772, "learning_rate": 7.567509025270758e-05, "loss": 1.8245, "step": 6738 }, { "epoch": 4.869314079422383, "grad_norm": 1.6470338106155396, "learning_rate": 7.56534296028881e-05, "loss": 1.6159, "step": 6744 }, { "epoch": 4.873646209386282, "grad_norm": 1.8459632396697998, "learning_rate": 7.56317689530686e-05, "loss": 1.7362, "step": 6750 }, { "epoch": 4.8779783393501805, "grad_norm": 1.7923812866210938, "learning_rate": 7.56101083032491e-05, "loss": 1.747, "step": 6756 }, { "epoch": 4.8823104693140795, "grad_norm": 1.6283410787582397, "learning_rate": 7.55884476534296e-05, "loss": 1.7072, "step": 6762 }, { "epoch": 4.8866425992779785, "grad_norm": 1.5675694942474365, "learning_rate": 7.556678700361011e-05, "loss": 1.6127, "step": 6768 }, { "epoch": 4.8909747292418775, "grad_norm": 1.8150726556777954, "learning_rate": 7.554512635379061e-05, "loss": 1.7059, "step": 6774 }, { "epoch": 4.8953068592057765, "grad_norm": 1.7699037790298462, "learning_rate": 7.552346570397112e-05, "loss": 1.7278, "step": 6780 }, { "epoch": 4.899638989169675, "grad_norm": 1.8475871086120605, "learning_rate": 7.550180505415163e-05, "loss": 1.679, "step": 6786 }, { "epoch": 4.903971119133574, "grad_norm": 1.7242374420166016, "learning_rate": 7.548014440433214e-05, "loss": 1.635, "step": 6792 }, { "epoch": 4.908303249097473, "grad_norm": 1.7109843492507935, "learning_rate": 7.545848375451264e-05, "loss": 1.707, "step": 6798 }, { "epoch": 4.9126353790613715, "grad_norm": 1.7471641302108765, "learning_rate": 7.543682310469314e-05, "loss": 1.6949, "step": 6804 }, { "epoch": 4.9169675090252705, "grad_norm": 1.7381162643432617, "learning_rate": 7.541516245487365e-05, "loss": 1.8183, "step": 6810 }, { "epoch": 4.9212996389891694, "grad_norm": 1.6778373718261719, "learning_rate": 7.539350180505416e-05, "loss": 1.7191, "step": 6816 }, { "epoch": 4.925631768953068, "grad_norm": 1.8151276111602783, "learning_rate": 7.537184115523465e-05, "loss": 1.7579, "step": 6822 }, { "epoch": 4.929963898916967, "grad_norm": 1.880372166633606, "learning_rate": 7.535018050541517e-05, "loss": 1.7072, "step": 6828 }, { "epoch": 4.934296028880866, "grad_norm": 1.7363860607147217, "learning_rate": 7.532851985559567e-05, "loss": 1.7868, "step": 6834 }, { "epoch": 4.938628158844765, "grad_norm": 1.7393163442611694, "learning_rate": 7.530685920577618e-05, "loss": 1.6876, "step": 6840 }, { "epoch": 4.942960288808664, "grad_norm": 1.916064739227295, "learning_rate": 7.528519855595668e-05, "loss": 1.8757, "step": 6846 }, { "epoch": 4.947292418772563, "grad_norm": 1.7780601978302002, "learning_rate": 7.526353790613718e-05, "loss": 1.6976, "step": 6852 }, { "epoch": 4.951624548736462, "grad_norm": 1.901220679283142, "learning_rate": 7.52418772563177e-05, "loss": 1.7292, "step": 6858 }, { "epoch": 4.955956678700361, "grad_norm": 1.6308395862579346, "learning_rate": 7.52202166064982e-05, "loss": 1.6714, "step": 6864 }, { "epoch": 4.96028880866426, "grad_norm": 1.7202249765396118, "learning_rate": 7.51985559566787e-05, "loss": 1.8019, "step": 6870 }, { "epoch": 4.964620938628159, "grad_norm": 1.6657606363296509, "learning_rate": 7.517689530685921e-05, "loss": 1.6038, "step": 6876 }, { "epoch": 4.968953068592057, "grad_norm": 1.6859421730041504, "learning_rate": 7.515523465703972e-05, "loss": 1.8133, "step": 6882 }, { "epoch": 4.973285198555956, "grad_norm": 1.6139439344406128, "learning_rate": 7.513357400722021e-05, "loss": 1.7444, "step": 6888 }, { "epoch": 4.977617328519855, "grad_norm": 1.8989307880401611, "learning_rate": 7.511191335740072e-05, "loss": 1.5637, "step": 6894 }, { "epoch": 4.981949458483754, "grad_norm": 1.7194766998291016, "learning_rate": 7.509025270758123e-05, "loss": 1.8026, "step": 6900 }, { "epoch": 4.986281588447653, "grad_norm": 1.5287039279937744, "learning_rate": 7.506859205776174e-05, "loss": 1.5804, "step": 6906 }, { "epoch": 4.990613718411552, "grad_norm": 1.8419551849365234, "learning_rate": 7.504693140794224e-05, "loss": 1.5655, "step": 6912 }, { "epoch": 4.994945848375451, "grad_norm": 1.7731727361679077, "learning_rate": 7.502527075812274e-05, "loss": 1.861, "step": 6918 }, { "epoch": 4.99927797833935, "grad_norm": 1.6581676006317139, "learning_rate": 7.500361010830326e-05, "loss": 1.6952, "step": 6924 }, { "epoch": 5.003610108303249, "grad_norm": 1.5256121158599854, "learning_rate": 7.498194945848376e-05, "loss": 1.5832, "step": 6930 }, { "epoch": 5.007942238267148, "grad_norm": 1.817798137664795, "learning_rate": 7.496028880866425e-05, "loss": 1.6775, "step": 6936 }, { "epoch": 5.012274368231047, "grad_norm": 1.6862815618515015, "learning_rate": 7.493862815884477e-05, "loss": 1.5916, "step": 6942 }, { "epoch": 5.016606498194946, "grad_norm": 1.6804051399230957, "learning_rate": 7.491696750902527e-05, "loss": 1.5872, "step": 6948 }, { "epoch": 5.020938628158845, "grad_norm": 1.8733099699020386, "learning_rate": 7.489530685920578e-05, "loss": 1.6872, "step": 6954 }, { "epoch": 5.025270758122744, "grad_norm": 1.5743608474731445, "learning_rate": 7.487364620938628e-05, "loss": 1.6647, "step": 6960 }, { "epoch": 5.029602888086643, "grad_norm": 1.5953441858291626, "learning_rate": 7.48519855595668e-05, "loss": 1.6023, "step": 6966 }, { "epoch": 5.033935018050541, "grad_norm": 1.4180645942687988, "learning_rate": 7.48303249097473e-05, "loss": 1.5667, "step": 6972 }, { "epoch": 5.03826714801444, "grad_norm": 1.8498455286026, "learning_rate": 7.480866425992779e-05, "loss": 1.7609, "step": 6978 }, { "epoch": 5.042599277978339, "grad_norm": 1.8996055126190186, "learning_rate": 7.47870036101083e-05, "loss": 1.6613, "step": 6984 }, { "epoch": 5.046931407942238, "grad_norm": 1.893189787864685, "learning_rate": 7.476534296028881e-05, "loss": 1.7123, "step": 6990 }, { "epoch": 5.051263537906137, "grad_norm": 1.5711545944213867, "learning_rate": 7.474368231046933e-05, "loss": 1.7261, "step": 6996 }, { "epoch": 5.055595667870036, "grad_norm": 1.9621397256851196, "learning_rate": 7.472202166064982e-05, "loss": 1.7501, "step": 7002 }, { "epoch": 5.059927797833935, "grad_norm": 1.932073950767517, "learning_rate": 7.470036101083033e-05, "loss": 1.6054, "step": 7008 }, { "epoch": 5.064259927797834, "grad_norm": 1.6758652925491333, "learning_rate": 7.467870036101084e-05, "loss": 1.627, "step": 7014 }, { "epoch": 5.068592057761733, "grad_norm": 1.860435128211975, "learning_rate": 7.465703971119134e-05, "loss": 1.6329, "step": 7020 }, { "epoch": 5.072924187725632, "grad_norm": 1.716552734375, "learning_rate": 7.463537906137184e-05, "loss": 1.7657, "step": 7026 }, { "epoch": 5.077256317689531, "grad_norm": 1.7217508554458618, "learning_rate": 7.461371841155235e-05, "loss": 1.7242, "step": 7032 }, { "epoch": 5.08158844765343, "grad_norm": 1.8037688732147217, "learning_rate": 7.459205776173286e-05, "loss": 1.5894, "step": 7038 }, { "epoch": 5.085920577617329, "grad_norm": 1.7853631973266602, "learning_rate": 7.457039711191335e-05, "loss": 1.629, "step": 7044 }, { "epoch": 5.090252707581228, "grad_norm": 1.810818076133728, "learning_rate": 7.454873646209387e-05, "loss": 1.5012, "step": 7050 }, { "epoch": 5.094584837545126, "grad_norm": 2.02217960357666, "learning_rate": 7.452707581227437e-05, "loss": 1.7267, "step": 7056 }, { "epoch": 5.098916967509025, "grad_norm": 1.8041623830795288, "learning_rate": 7.450541516245488e-05, "loss": 1.672, "step": 7062 }, { "epoch": 5.103249097472924, "grad_norm": 1.7721242904663086, "learning_rate": 7.448375451263538e-05, "loss": 1.6051, "step": 7068 }, { "epoch": 5.107581227436823, "grad_norm": 1.8781174421310425, "learning_rate": 7.446209386281588e-05, "loss": 1.6718, "step": 7074 }, { "epoch": 5.111913357400722, "grad_norm": 1.722411036491394, "learning_rate": 7.44404332129964e-05, "loss": 1.62, "step": 7080 }, { "epoch": 5.116245487364621, "grad_norm": 1.7854832410812378, "learning_rate": 7.44187725631769e-05, "loss": 1.7063, "step": 7086 }, { "epoch": 5.12057761732852, "grad_norm": 1.7524381875991821, "learning_rate": 7.43971119133574e-05, "loss": 1.6288, "step": 7092 }, { "epoch": 5.124909747292419, "grad_norm": 1.629851222038269, "learning_rate": 7.437545126353791e-05, "loss": 1.6309, "step": 7098 }, { "epoch": 5.129241877256318, "grad_norm": 1.6489806175231934, "learning_rate": 7.435379061371841e-05, "loss": 1.6196, "step": 7104 }, { "epoch": 5.133574007220217, "grad_norm": 1.6421762704849243, "learning_rate": 7.433212996389893e-05, "loss": 1.5791, "step": 7110 }, { "epoch": 5.137906137184116, "grad_norm": 1.9120826721191406, "learning_rate": 7.431046931407942e-05, "loss": 1.6528, "step": 7116 }, { "epoch": 5.142238267148015, "grad_norm": 1.7292189598083496, "learning_rate": 7.428880866425994e-05, "loss": 1.6484, "step": 7122 }, { "epoch": 5.146570397111914, "grad_norm": 2.049380302429199, "learning_rate": 7.426714801444044e-05, "loss": 1.746, "step": 7128 }, { "epoch": 5.150902527075813, "grad_norm": 1.8253040313720703, "learning_rate": 7.424548736462094e-05, "loss": 1.5459, "step": 7134 }, { "epoch": 5.155234657039712, "grad_norm": 1.730258822441101, "learning_rate": 7.422382671480144e-05, "loss": 1.7053, "step": 7140 }, { "epoch": 5.15956678700361, "grad_norm": 1.8754324913024902, "learning_rate": 7.420216606498195e-05, "loss": 1.6323, "step": 7146 }, { "epoch": 5.163898916967509, "grad_norm": 1.7928742170333862, "learning_rate": 7.418050541516246e-05, "loss": 1.6472, "step": 7152 }, { "epoch": 5.168231046931408, "grad_norm": 1.8704208135604858, "learning_rate": 7.415884476534295e-05, "loss": 1.5742, "step": 7158 }, { "epoch": 5.172563176895307, "grad_norm": 1.7730602025985718, "learning_rate": 7.413718411552347e-05, "loss": 1.7508, "step": 7164 }, { "epoch": 5.176895306859206, "grad_norm": 1.924324631690979, "learning_rate": 7.411552346570397e-05, "loss": 1.6422, "step": 7170 }, { "epoch": 5.181227436823105, "grad_norm": 1.827392339706421, "learning_rate": 7.409386281588449e-05, "loss": 1.7292, "step": 7176 }, { "epoch": 5.185559566787004, "grad_norm": 1.8377982378005981, "learning_rate": 7.407220216606498e-05, "loss": 1.6518, "step": 7182 }, { "epoch": 5.189891696750903, "grad_norm": 1.8795768022537231, "learning_rate": 7.405054151624548e-05, "loss": 1.7012, "step": 7188 }, { "epoch": 5.1942238267148015, "grad_norm": 1.7151986360549927, "learning_rate": 7.4028880866426e-05, "loss": 1.5495, "step": 7194 }, { "epoch": 5.1985559566787005, "grad_norm": 1.8887649774551392, "learning_rate": 7.40072202166065e-05, "loss": 1.6469, "step": 7200 }, { "epoch": 5.2028880866425995, "grad_norm": 1.8095066547393799, "learning_rate": 7.398555956678701e-05, "loss": 1.7178, "step": 7206 }, { "epoch": 5.2072202166064985, "grad_norm": 1.7813700437545776, "learning_rate": 7.396389891696751e-05, "loss": 1.7442, "step": 7212 }, { "epoch": 5.2115523465703975, "grad_norm": 1.7976068258285522, "learning_rate": 7.394223826714803e-05, "loss": 1.5481, "step": 7218 }, { "epoch": 5.215884476534296, "grad_norm": 1.932558536529541, "learning_rate": 7.392057761732852e-05, "loss": 1.8, "step": 7224 }, { "epoch": 5.2202166064981945, "grad_norm": 1.8538416624069214, "learning_rate": 7.389891696750902e-05, "loss": 1.5893, "step": 7230 }, { "epoch": 5.2245487364620935, "grad_norm": 1.8784427642822266, "learning_rate": 7.387725631768954e-05, "loss": 1.7175, "step": 7236 }, { "epoch": 5.2288808664259925, "grad_norm": 1.9263156652450562, "learning_rate": 7.385559566787004e-05, "loss": 1.6373, "step": 7242 }, { "epoch": 5.2332129963898915, "grad_norm": 1.674646258354187, "learning_rate": 7.383393501805054e-05, "loss": 1.6637, "step": 7248 }, { "epoch": 5.23754512635379, "grad_norm": 1.8438295125961304, "learning_rate": 7.381227436823105e-05, "loss": 1.6833, "step": 7254 }, { "epoch": 5.241877256317689, "grad_norm": 1.9228692054748535, "learning_rate": 7.379061371841156e-05, "loss": 1.6445, "step": 7260 }, { "epoch": 5.246209386281588, "grad_norm": 1.5970263481140137, "learning_rate": 7.376895306859207e-05, "loss": 1.7147, "step": 7266 }, { "epoch": 5.250541516245487, "grad_norm": 1.5604310035705566, "learning_rate": 7.374729241877256e-05, "loss": 1.5272, "step": 7272 }, { "epoch": 5.254873646209386, "grad_norm": 1.8667677640914917, "learning_rate": 7.372563176895307e-05, "loss": 1.6316, "step": 7278 }, { "epoch": 5.259205776173285, "grad_norm": 1.6557292938232422, "learning_rate": 7.370397111913358e-05, "loss": 1.6048, "step": 7284 }, { "epoch": 5.263537906137184, "grad_norm": 1.7370303869247437, "learning_rate": 7.368231046931408e-05, "loss": 1.585, "step": 7290 }, { "epoch": 5.267870036101083, "grad_norm": 1.8234481811523438, "learning_rate": 7.366064981949458e-05, "loss": 1.5885, "step": 7296 }, { "epoch": 5.272202166064982, "grad_norm": 1.775384545326233, "learning_rate": 7.36389891696751e-05, "loss": 1.646, "step": 7302 }, { "epoch": 5.276534296028881, "grad_norm": 1.8459120988845825, "learning_rate": 7.36173285198556e-05, "loss": 1.686, "step": 7308 }, { "epoch": 5.28086642599278, "grad_norm": 1.7361106872558594, "learning_rate": 7.35956678700361e-05, "loss": 1.6757, "step": 7314 }, { "epoch": 5.285198555956678, "grad_norm": 2.0872950553894043, "learning_rate": 7.357400722021661e-05, "loss": 1.688, "step": 7320 }, { "epoch": 5.289530685920577, "grad_norm": 1.786478042602539, "learning_rate": 7.355234657039711e-05, "loss": 1.6614, "step": 7326 }, { "epoch": 5.293862815884476, "grad_norm": 1.7402349710464478, "learning_rate": 7.353068592057763e-05, "loss": 1.7019, "step": 7332 }, { "epoch": 5.298194945848375, "grad_norm": 1.9275758266448975, "learning_rate": 7.350902527075812e-05, "loss": 1.7315, "step": 7338 }, { "epoch": 5.302527075812274, "grad_norm": 1.8186196088790894, "learning_rate": 7.348736462093864e-05, "loss": 1.7114, "step": 7344 }, { "epoch": 5.306859205776173, "grad_norm": 1.7823747396469116, "learning_rate": 7.346570397111914e-05, "loss": 1.778, "step": 7350 }, { "epoch": 5.311191335740072, "grad_norm": 2.0236849784851074, "learning_rate": 7.344404332129964e-05, "loss": 1.5547, "step": 7356 }, { "epoch": 5.315523465703971, "grad_norm": 1.694535732269287, "learning_rate": 7.342238267148015e-05, "loss": 1.6938, "step": 7362 }, { "epoch": 5.31985559566787, "grad_norm": 1.741994857788086, "learning_rate": 7.340072202166065e-05, "loss": 1.6043, "step": 7368 }, { "epoch": 5.324187725631769, "grad_norm": 1.8013322353363037, "learning_rate": 7.337906137184117e-05, "loss": 1.6539, "step": 7374 }, { "epoch": 5.328519855595668, "grad_norm": 1.973013997077942, "learning_rate": 7.335740072202167e-05, "loss": 1.6952, "step": 7380 }, { "epoch": 5.332851985559567, "grad_norm": 1.904317021369934, "learning_rate": 7.333574007220217e-05, "loss": 1.7021, "step": 7386 }, { "epoch": 5.337184115523466, "grad_norm": 1.7099312543869019, "learning_rate": 7.331407942238268e-05, "loss": 1.5844, "step": 7392 }, { "epoch": 5.341516245487365, "grad_norm": 1.9053092002868652, "learning_rate": 7.329241877256318e-05, "loss": 1.6197, "step": 7398 }, { "epoch": 5.345848375451263, "grad_norm": 1.7344249486923218, "learning_rate": 7.327075812274368e-05, "loss": 1.6015, "step": 7404 }, { "epoch": 5.350180505415162, "grad_norm": 1.8829872608184814, "learning_rate": 7.324909747292418e-05, "loss": 1.7673, "step": 7410 }, { "epoch": 5.354512635379061, "grad_norm": 1.9475154876708984, "learning_rate": 7.32274368231047e-05, "loss": 1.7136, "step": 7416 }, { "epoch": 5.35884476534296, "grad_norm": 1.8083956241607666, "learning_rate": 7.32057761732852e-05, "loss": 1.7685, "step": 7422 }, { "epoch": 5.363176895306859, "grad_norm": 1.7618496417999268, "learning_rate": 7.318411552346571e-05, "loss": 1.6847, "step": 7428 }, { "epoch": 5.367509025270758, "grad_norm": 1.8150408267974854, "learning_rate": 7.316245487364621e-05, "loss": 1.6235, "step": 7434 }, { "epoch": 5.371841155234657, "grad_norm": 1.9297406673431396, "learning_rate": 7.314079422382671e-05, "loss": 1.6719, "step": 7440 }, { "epoch": 5.376173285198556, "grad_norm": 1.9739959239959717, "learning_rate": 7.311913357400723e-05, "loss": 1.7206, "step": 7446 }, { "epoch": 5.380505415162455, "grad_norm": 1.958534836769104, "learning_rate": 7.309747292418772e-05, "loss": 1.7447, "step": 7452 }, { "epoch": 5.384837545126354, "grad_norm": 1.6005042791366577, "learning_rate": 7.307581227436824e-05, "loss": 1.5394, "step": 7458 }, { "epoch": 5.389169675090253, "grad_norm": 1.7943589687347412, "learning_rate": 7.305415162454874e-05, "loss": 1.5379, "step": 7464 }, { "epoch": 5.393501805054152, "grad_norm": 1.76604163646698, "learning_rate": 7.303249097472924e-05, "loss": 1.6006, "step": 7470 }, { "epoch": 5.397833935018051, "grad_norm": 1.8200541734695435, "learning_rate": 7.301083032490975e-05, "loss": 1.6881, "step": 7476 }, { "epoch": 5.40216606498195, "grad_norm": 1.815521478652954, "learning_rate": 7.298916967509025e-05, "loss": 1.6486, "step": 7482 }, { "epoch": 5.406498194945849, "grad_norm": 1.9071468114852905, "learning_rate": 7.296750902527077e-05, "loss": 1.6581, "step": 7488 }, { "epoch": 5.410830324909747, "grad_norm": 1.782283902168274, "learning_rate": 7.294584837545127e-05, "loss": 1.6952, "step": 7494 }, { "epoch": 5.415162454873646, "grad_norm": 1.9334980249404907, "learning_rate": 7.292418772563177e-05, "loss": 1.6283, "step": 7500 }, { "epoch": 5.419494584837545, "grad_norm": 1.5638436079025269, "learning_rate": 7.290252707581228e-05, "loss": 1.5591, "step": 7506 }, { "epoch": 5.423826714801444, "grad_norm": 1.8555253744125366, "learning_rate": 7.28808664259928e-05, "loss": 1.7766, "step": 7512 }, { "epoch": 5.428158844765343, "grad_norm": 2.130786418914795, "learning_rate": 7.285920577617328e-05, "loss": 1.6685, "step": 7518 }, { "epoch": 5.432490974729242, "grad_norm": 1.6943217515945435, "learning_rate": 7.283754512635379e-05, "loss": 1.6988, "step": 7524 }, { "epoch": 5.436823104693141, "grad_norm": 1.907670021057129, "learning_rate": 7.28158844765343e-05, "loss": 1.7262, "step": 7530 }, { "epoch": 5.44115523465704, "grad_norm": 1.8848294019699097, "learning_rate": 7.27942238267148e-05, "loss": 1.607, "step": 7536 }, { "epoch": 5.445487364620939, "grad_norm": 1.799243450164795, "learning_rate": 7.277256317689531e-05, "loss": 1.5934, "step": 7542 }, { "epoch": 5.449819494584838, "grad_norm": 1.8209720849990845, "learning_rate": 7.275090252707581e-05, "loss": 1.6877, "step": 7548 }, { "epoch": 5.454151624548737, "grad_norm": 1.8905701637268066, "learning_rate": 7.272924187725633e-05, "loss": 1.5687, "step": 7554 }, { "epoch": 5.458483754512636, "grad_norm": 1.8209524154663086, "learning_rate": 7.270758122743683e-05, "loss": 1.6224, "step": 7560 }, { "epoch": 5.462815884476535, "grad_norm": 1.8583896160125732, "learning_rate": 7.268592057761732e-05, "loss": 1.746, "step": 7566 }, { "epoch": 5.467148014440433, "grad_norm": 1.9084917306900024, "learning_rate": 7.266425992779784e-05, "loss": 1.5956, "step": 7572 }, { "epoch": 5.471480144404332, "grad_norm": 1.7360036373138428, "learning_rate": 7.264259927797834e-05, "loss": 1.5694, "step": 7578 }, { "epoch": 5.475812274368231, "grad_norm": 1.686075210571289, "learning_rate": 7.262093862815885e-05, "loss": 1.6773, "step": 7584 }, { "epoch": 5.48014440433213, "grad_norm": 1.7724977731704712, "learning_rate": 7.259927797833935e-05, "loss": 1.587, "step": 7590 }, { "epoch": 5.484476534296029, "grad_norm": 1.8738068342208862, "learning_rate": 7.257761732851987e-05, "loss": 1.658, "step": 7596 }, { "epoch": 5.488808664259928, "grad_norm": 1.7755991220474243, "learning_rate": 7.255595667870037e-05, "loss": 1.562, "step": 7602 }, { "epoch": 5.493140794223827, "grad_norm": 1.8585394620895386, "learning_rate": 7.253429602888087e-05, "loss": 1.6631, "step": 7608 }, { "epoch": 5.497472924187726, "grad_norm": 1.6825475692749023, "learning_rate": 7.251263537906138e-05, "loss": 1.6236, "step": 7614 }, { "epoch": 5.501805054151625, "grad_norm": 1.6755540370941162, "learning_rate": 7.249097472924188e-05, "loss": 1.5867, "step": 7620 }, { "epoch": 5.5061371841155236, "grad_norm": 1.7259434461593628, "learning_rate": 7.24693140794224e-05, "loss": 1.5601, "step": 7626 }, { "epoch": 5.5104693140794225, "grad_norm": 1.8260570764541626, "learning_rate": 7.244765342960289e-05, "loss": 1.6677, "step": 7632 }, { "epoch": 5.5148014440433215, "grad_norm": 1.7593762874603271, "learning_rate": 7.242599277978339e-05, "loss": 1.6109, "step": 7638 }, { "epoch": 5.5191335740072205, "grad_norm": 1.8089516162872314, "learning_rate": 7.24043321299639e-05, "loss": 1.7958, "step": 7644 }, { "epoch": 5.5234657039711195, "grad_norm": 1.8030314445495605, "learning_rate": 7.238267148014441e-05, "loss": 1.6041, "step": 7650 }, { "epoch": 5.5277978339350184, "grad_norm": 1.8451223373413086, "learning_rate": 7.236101083032491e-05, "loss": 1.6399, "step": 7656 }, { "epoch": 5.532129963898917, "grad_norm": 1.8362971544265747, "learning_rate": 7.233935018050542e-05, "loss": 1.7763, "step": 7662 }, { "epoch": 5.5364620938628155, "grad_norm": 1.8358193635940552, "learning_rate": 7.231768953068593e-05, "loss": 1.5676, "step": 7668 }, { "epoch": 5.5407942238267145, "grad_norm": 1.663753628730774, "learning_rate": 7.229602888086644e-05, "loss": 1.5245, "step": 7674 }, { "epoch": 5.5451263537906135, "grad_norm": 1.8764318227767944, "learning_rate": 7.227436823104692e-05, "loss": 1.8124, "step": 7680 }, { "epoch": 5.5494584837545125, "grad_norm": 1.6874598264694214, "learning_rate": 7.225270758122744e-05, "loss": 1.6458, "step": 7686 }, { "epoch": 5.553790613718411, "grad_norm": 2.0758495330810547, "learning_rate": 7.223104693140794e-05, "loss": 1.651, "step": 7692 }, { "epoch": 5.55812274368231, "grad_norm": 1.9171005487442017, "learning_rate": 7.220938628158845e-05, "loss": 1.6694, "step": 7698 }, { "epoch": 5.562454873646209, "grad_norm": 1.9406453371047974, "learning_rate": 7.218772563176895e-05, "loss": 1.7205, "step": 7704 }, { "epoch": 5.566787003610108, "grad_norm": 1.8588297367095947, "learning_rate": 7.216606498194947e-05, "loss": 1.6424, "step": 7710 }, { "epoch": 5.571119133574007, "grad_norm": 1.8383994102478027, "learning_rate": 7.214440433212997e-05, "loss": 1.6255, "step": 7716 }, { "epoch": 5.575451263537906, "grad_norm": 1.8067480325698853, "learning_rate": 7.212274368231046e-05, "loss": 1.6918, "step": 7722 }, { "epoch": 5.579783393501805, "grad_norm": 1.840858817100525, "learning_rate": 7.210108303249098e-05, "loss": 1.7001, "step": 7728 }, { "epoch": 5.584115523465704, "grad_norm": 1.687282681465149, "learning_rate": 7.207942238267148e-05, "loss": 1.6512, "step": 7734 }, { "epoch": 5.588447653429603, "grad_norm": 1.6026248931884766, "learning_rate": 7.2057761732852e-05, "loss": 1.5609, "step": 7740 }, { "epoch": 5.592779783393501, "grad_norm": 1.7752634286880493, "learning_rate": 7.203610108303249e-05, "loss": 1.7611, "step": 7746 }, { "epoch": 5.5971119133574, "grad_norm": 1.7897206544876099, "learning_rate": 7.2014440433213e-05, "loss": 1.5656, "step": 7752 }, { "epoch": 5.601444043321299, "grad_norm": 1.64107084274292, "learning_rate": 7.199277978339351e-05, "loss": 1.7079, "step": 7758 }, { "epoch": 5.605776173285198, "grad_norm": 1.8135124444961548, "learning_rate": 7.197111913357401e-05, "loss": 1.7439, "step": 7764 }, { "epoch": 5.610108303249097, "grad_norm": 1.7337638139724731, "learning_rate": 7.194945848375451e-05, "loss": 1.5577, "step": 7770 }, { "epoch": 5.614440433212996, "grad_norm": 2.1683082580566406, "learning_rate": 7.192779783393502e-05, "loss": 1.7033, "step": 7776 }, { "epoch": 5.618772563176895, "grad_norm": 1.9333600997924805, "learning_rate": 7.190613718411553e-05, "loss": 1.7166, "step": 7782 }, { "epoch": 5.623104693140794, "grad_norm": 1.9663941860198975, "learning_rate": 7.188447653429602e-05, "loss": 1.7377, "step": 7788 }, { "epoch": 5.627436823104693, "grad_norm": 1.6376093626022339, "learning_rate": 7.186281588447654e-05, "loss": 1.6656, "step": 7794 }, { "epoch": 5.631768953068592, "grad_norm": 1.6083195209503174, "learning_rate": 7.184115523465704e-05, "loss": 1.6294, "step": 7800 }, { "epoch": 5.636101083032491, "grad_norm": 1.7283453941345215, "learning_rate": 7.181949458483755e-05, "loss": 1.6287, "step": 7806 }, { "epoch": 5.64043321299639, "grad_norm": 2.0410704612731934, "learning_rate": 7.179783393501805e-05, "loss": 1.753, "step": 7812 }, { "epoch": 5.644765342960289, "grad_norm": 1.7562671899795532, "learning_rate": 7.177617328519855e-05, "loss": 1.7759, "step": 7818 }, { "epoch": 5.649097472924188, "grad_norm": 1.967004418373108, "learning_rate": 7.175451263537907e-05, "loss": 1.636, "step": 7824 }, { "epoch": 5.653429602888087, "grad_norm": 1.9438207149505615, "learning_rate": 7.173285198555957e-05, "loss": 1.6113, "step": 7830 }, { "epoch": 5.657761732851986, "grad_norm": 2.168013334274292, "learning_rate": 7.171119133574008e-05, "loss": 1.6963, "step": 7836 }, { "epoch": 5.662093862815884, "grad_norm": 1.8814265727996826, "learning_rate": 7.168953068592058e-05, "loss": 1.7216, "step": 7842 }, { "epoch": 5.666425992779783, "grad_norm": 1.8556036949157715, "learning_rate": 7.166787003610108e-05, "loss": 1.7003, "step": 7848 }, { "epoch": 5.670758122743682, "grad_norm": 1.807923436164856, "learning_rate": 7.16462093862816e-05, "loss": 1.6412, "step": 7854 }, { "epoch": 5.675090252707581, "grad_norm": 1.6672875881195068, "learning_rate": 7.162454873646209e-05, "loss": 1.6897, "step": 7860 }, { "epoch": 5.67942238267148, "grad_norm": 1.8167040348052979, "learning_rate": 7.16028880866426e-05, "loss": 1.7618, "step": 7866 }, { "epoch": 5.683754512635379, "grad_norm": 1.7749176025390625, "learning_rate": 7.158122743682311e-05, "loss": 1.6582, "step": 7872 }, { "epoch": 5.688086642599278, "grad_norm": 1.9250788688659668, "learning_rate": 7.155956678700361e-05, "loss": 1.623, "step": 7878 }, { "epoch": 5.692418772563177, "grad_norm": 1.6440633535385132, "learning_rate": 7.153790613718412e-05, "loss": 1.4309, "step": 7884 }, { "epoch": 5.696750902527076, "grad_norm": 1.8783047199249268, "learning_rate": 7.151624548736462e-05, "loss": 1.7492, "step": 7890 }, { "epoch": 5.701083032490975, "grad_norm": 1.7556110620498657, "learning_rate": 7.149458483754514e-05, "loss": 1.5701, "step": 7896 }, { "epoch": 5.705415162454874, "grad_norm": 1.8307605981826782, "learning_rate": 7.147292418772563e-05, "loss": 1.6755, "step": 7902 }, { "epoch": 5.709747292418773, "grad_norm": 2.0365285873413086, "learning_rate": 7.145126353790614e-05, "loss": 1.7028, "step": 7908 }, { "epoch": 5.714079422382672, "grad_norm": 1.7807424068450928, "learning_rate": 7.142960288808665e-05, "loss": 1.571, "step": 7914 }, { "epoch": 5.71841155234657, "grad_norm": 1.7400039434432983, "learning_rate": 7.140794223826716e-05, "loss": 1.7081, "step": 7920 }, { "epoch": 5.722743682310469, "grad_norm": 2.335845708847046, "learning_rate": 7.138628158844765e-05, "loss": 1.7127, "step": 7926 }, { "epoch": 5.727075812274368, "grad_norm": 1.9361305236816406, "learning_rate": 7.136462093862816e-05, "loss": 1.731, "step": 7932 }, { "epoch": 5.731407942238267, "grad_norm": 1.8917917013168335, "learning_rate": 7.134296028880867e-05, "loss": 1.7231, "step": 7938 }, { "epoch": 5.735740072202166, "grad_norm": 1.9247841835021973, "learning_rate": 7.132129963898918e-05, "loss": 1.7132, "step": 7944 }, { "epoch": 5.740072202166065, "grad_norm": 1.9171816110610962, "learning_rate": 7.129963898916968e-05, "loss": 1.6622, "step": 7950 }, { "epoch": 5.744404332129964, "grad_norm": 1.876046895980835, "learning_rate": 7.127797833935018e-05, "loss": 1.7324, "step": 7956 }, { "epoch": 5.748736462093863, "grad_norm": 1.9420334100723267, "learning_rate": 7.12563176895307e-05, "loss": 1.7295, "step": 7962 }, { "epoch": 5.753068592057762, "grad_norm": 1.7340840101242065, "learning_rate": 7.123465703971119e-05, "loss": 1.7608, "step": 7968 }, { "epoch": 5.757400722021661, "grad_norm": 1.791069746017456, "learning_rate": 7.121299638989169e-05, "loss": 1.6578, "step": 7974 }, { "epoch": 5.76173285198556, "grad_norm": 1.9585387706756592, "learning_rate": 7.119133574007221e-05, "loss": 1.6334, "step": 7980 }, { "epoch": 5.766064981949459, "grad_norm": 1.79752516746521, "learning_rate": 7.116967509025271e-05, "loss": 1.802, "step": 7986 }, { "epoch": 5.770397111913358, "grad_norm": 1.9339418411254883, "learning_rate": 7.114801444043321e-05, "loss": 1.6311, "step": 7992 }, { "epoch": 5.774729241877257, "grad_norm": 2.121168613433838, "learning_rate": 7.112635379061372e-05, "loss": 1.621, "step": 7998 }, { "epoch": 5.779061371841156, "grad_norm": 1.822784185409546, "learning_rate": 7.110469314079423e-05, "loss": 1.7591, "step": 8004 }, { "epoch": 5.783393501805055, "grad_norm": 1.7117409706115723, "learning_rate": 7.108303249097474e-05, "loss": 1.5201, "step": 8010 }, { "epoch": 5.787725631768953, "grad_norm": 1.9525624513626099, "learning_rate": 7.106137184115523e-05, "loss": 1.7174, "step": 8016 }, { "epoch": 5.792057761732852, "grad_norm": 1.8775008916854858, "learning_rate": 7.103971119133574e-05, "loss": 1.7766, "step": 8022 }, { "epoch": 5.796389891696751, "grad_norm": 1.6939162015914917, "learning_rate": 7.101805054151625e-05, "loss": 1.6193, "step": 8028 }, { "epoch": 5.80072202166065, "grad_norm": 1.7726097106933594, "learning_rate": 7.099638989169675e-05, "loss": 1.6948, "step": 8034 }, { "epoch": 5.805054151624549, "grad_norm": 1.87083899974823, "learning_rate": 7.097472924187725e-05, "loss": 1.6468, "step": 8040 }, { "epoch": 5.809386281588448, "grad_norm": 1.7086987495422363, "learning_rate": 7.095306859205777e-05, "loss": 1.6489, "step": 8046 }, { "epoch": 5.813718411552347, "grad_norm": 1.8037338256835938, "learning_rate": 7.093140794223827e-05, "loss": 1.695, "step": 8052 }, { "epoch": 5.818050541516246, "grad_norm": 1.9314862489700317, "learning_rate": 7.090974729241878e-05, "loss": 1.4714, "step": 8058 }, { "epoch": 5.8223826714801445, "grad_norm": 1.8704792261123657, "learning_rate": 7.088808664259928e-05, "loss": 1.5266, "step": 8064 }, { "epoch": 5.8267148014440435, "grad_norm": 1.85940420627594, "learning_rate": 7.086642599277978e-05, "loss": 1.8204, "step": 8070 }, { "epoch": 5.8310469314079425, "grad_norm": 1.9251248836517334, "learning_rate": 7.08447653429603e-05, "loss": 1.6081, "step": 8076 }, { "epoch": 5.8353790613718415, "grad_norm": 1.8482950925827026, "learning_rate": 7.082310469314079e-05, "loss": 1.7889, "step": 8082 }, { "epoch": 5.8397111913357405, "grad_norm": 1.8394749164581299, "learning_rate": 7.08014440433213e-05, "loss": 1.724, "step": 8088 }, { "epoch": 5.8440433212996385, "grad_norm": 1.8953851461410522, "learning_rate": 7.077978339350181e-05, "loss": 1.759, "step": 8094 }, { "epoch": 5.8483754512635375, "grad_norm": 1.7887818813323975, "learning_rate": 7.075812274368231e-05, "loss": 1.6606, "step": 8100 }, { "epoch": 5.8527075812274365, "grad_norm": 1.7751671075820923, "learning_rate": 7.073646209386282e-05, "loss": 1.7733, "step": 8106 }, { "epoch": 5.8570397111913355, "grad_norm": 1.620985746383667, "learning_rate": 7.071480144404332e-05, "loss": 1.5908, "step": 8112 }, { "epoch": 5.8613718411552345, "grad_norm": 1.8627665042877197, "learning_rate": 7.069314079422384e-05, "loss": 1.5613, "step": 8118 }, { "epoch": 5.865703971119133, "grad_norm": 1.9402095079421997, "learning_rate": 7.067148014440434e-05, "loss": 1.5396, "step": 8124 }, { "epoch": 5.870036101083032, "grad_norm": 1.9350756406784058, "learning_rate": 7.064981949458484e-05, "loss": 1.7733, "step": 8130 }, { "epoch": 5.874368231046931, "grad_norm": 2.00177264213562, "learning_rate": 7.062815884476535e-05, "loss": 1.7011, "step": 8136 }, { "epoch": 5.87870036101083, "grad_norm": 1.7651840448379517, "learning_rate": 7.060649819494585e-05, "loss": 1.8156, "step": 8142 }, { "epoch": 5.883032490974729, "grad_norm": 1.8686625957489014, "learning_rate": 7.058483754512635e-05, "loss": 1.7584, "step": 8148 }, { "epoch": 5.887364620938628, "grad_norm": 1.8333086967468262, "learning_rate": 7.056317689530686e-05, "loss": 1.7358, "step": 8154 }, { "epoch": 5.891696750902527, "grad_norm": 1.694111704826355, "learning_rate": 7.054151624548737e-05, "loss": 1.6503, "step": 8160 }, { "epoch": 5.896028880866426, "grad_norm": 1.888282299041748, "learning_rate": 7.051985559566788e-05, "loss": 1.8015, "step": 8166 }, { "epoch": 5.900361010830325, "grad_norm": 2.0100536346435547, "learning_rate": 7.049819494584838e-05, "loss": 1.6999, "step": 8172 }, { "epoch": 5.904693140794224, "grad_norm": 1.877898097038269, "learning_rate": 7.047653429602888e-05, "loss": 1.6838, "step": 8178 }, { "epoch": 5.909025270758123, "grad_norm": 1.8017489910125732, "learning_rate": 7.045487364620939e-05, "loss": 1.6549, "step": 8184 }, { "epoch": 5.913357400722021, "grad_norm": 1.7196335792541504, "learning_rate": 7.04332129963899e-05, "loss": 1.5867, "step": 8190 }, { "epoch": 5.91768953068592, "grad_norm": 1.8160486221313477, "learning_rate": 7.041155234657039e-05, "loss": 1.6474, "step": 8196 }, { "epoch": 5.922021660649819, "grad_norm": 1.8755321502685547, "learning_rate": 7.038989169675091e-05, "loss": 1.6976, "step": 8202 }, { "epoch": 5.926353790613718, "grad_norm": 1.6077752113342285, "learning_rate": 7.036823104693141e-05, "loss": 1.6773, "step": 8208 }, { "epoch": 5.930685920577617, "grad_norm": 1.8785219192504883, "learning_rate": 7.034657039711192e-05, "loss": 1.576, "step": 8214 }, { "epoch": 5.935018050541516, "grad_norm": 1.7722182273864746, "learning_rate": 7.032490974729242e-05, "loss": 1.6845, "step": 8220 }, { "epoch": 5.939350180505415, "grad_norm": 2.0409274101257324, "learning_rate": 7.030324909747292e-05, "loss": 1.5676, "step": 8226 }, { "epoch": 5.943682310469314, "grad_norm": 2.0133161544799805, "learning_rate": 7.028158844765344e-05, "loss": 1.7019, "step": 8232 }, { "epoch": 5.948014440433213, "grad_norm": 1.8310308456420898, "learning_rate": 7.025992779783394e-05, "loss": 1.6667, "step": 8238 }, { "epoch": 5.952346570397112, "grad_norm": 2.0395634174346924, "learning_rate": 7.023826714801444e-05, "loss": 1.5532, "step": 8244 }, { "epoch": 5.956678700361011, "grad_norm": 1.7982017993927002, "learning_rate": 7.021660649819495e-05, "loss": 1.7057, "step": 8250 }, { "epoch": 5.96101083032491, "grad_norm": 1.9239319562911987, "learning_rate": 7.019494584837546e-05, "loss": 1.6507, "step": 8256 }, { "epoch": 5.965342960288808, "grad_norm": 1.9067988395690918, "learning_rate": 7.017328519855595e-05, "loss": 1.6504, "step": 8262 }, { "epoch": 5.969675090252707, "grad_norm": 1.7637206315994263, "learning_rate": 7.015162454873646e-05, "loss": 1.5793, "step": 8268 }, { "epoch": 5.974007220216606, "grad_norm": 1.8445377349853516, "learning_rate": 7.012996389891697e-05, "loss": 1.6177, "step": 8274 }, { "epoch": 5.978339350180505, "grad_norm": 1.8682135343551636, "learning_rate": 7.010830324909748e-05, "loss": 1.6786, "step": 8280 }, { "epoch": 5.982671480144404, "grad_norm": 2.1961538791656494, "learning_rate": 7.008664259927798e-05, "loss": 1.7511, "step": 8286 }, { "epoch": 5.987003610108303, "grad_norm": 1.8215526342391968, "learning_rate": 7.006498194945848e-05, "loss": 1.5423, "step": 8292 }, { "epoch": 5.991335740072202, "grad_norm": 2.055919647216797, "learning_rate": 7.0043321299639e-05, "loss": 1.5947, "step": 8298 }, { "epoch": 5.995667870036101, "grad_norm": 1.9543720483779907, "learning_rate": 7.00216606498195e-05, "loss": 1.5636, "step": 8304 }, { "epoch": 6.0, "grad_norm": 2.067350387573242, "learning_rate": 7e-05, "loss": 1.579, "step": 8310 }, { "epoch": 6.004332129963899, "grad_norm": 2.082796573638916, "learning_rate": 6.997833935018051e-05, "loss": 1.5392, "step": 8316 }, { "epoch": 6.008664259927798, "grad_norm": 1.7453832626342773, "learning_rate": 6.995667870036101e-05, "loss": 1.6619, "step": 8322 }, { "epoch": 6.012996389891697, "grad_norm": 1.6898438930511475, "learning_rate": 6.993501805054152e-05, "loss": 1.5839, "step": 8328 }, { "epoch": 6.017328519855596, "grad_norm": 1.9538588523864746, "learning_rate": 6.991335740072202e-05, "loss": 1.6751, "step": 8334 }, { "epoch": 6.021660649819495, "grad_norm": 1.8587323427200317, "learning_rate": 6.989169675090254e-05, "loss": 1.4655, "step": 8340 }, { "epoch": 6.025992779783394, "grad_norm": 1.9880495071411133, "learning_rate": 6.987003610108304e-05, "loss": 1.744, "step": 8346 }, { "epoch": 6.030324909747293, "grad_norm": 1.5169285535812378, "learning_rate": 6.984837545126354e-05, "loss": 1.6552, "step": 8352 }, { "epoch": 6.034657039711191, "grad_norm": 1.8923704624176025, "learning_rate": 6.982671480144405e-05, "loss": 1.6882, "step": 8358 }, { "epoch": 6.03898916967509, "grad_norm": 1.8967580795288086, "learning_rate": 6.980505415162455e-05, "loss": 1.6144, "step": 8364 }, { "epoch": 6.043321299638989, "grad_norm": 1.8046070337295532, "learning_rate": 6.978339350180507e-05, "loss": 1.5404, "step": 8370 }, { "epoch": 6.047653429602888, "grad_norm": 1.611784815788269, "learning_rate": 6.976173285198556e-05, "loss": 1.5261, "step": 8376 }, { "epoch": 6.051985559566787, "grad_norm": 1.9831805229187012, "learning_rate": 6.974007220216607e-05, "loss": 1.5396, "step": 8382 }, { "epoch": 6.056317689530686, "grad_norm": 1.9910296201705933, "learning_rate": 6.971841155234658e-05, "loss": 1.605, "step": 8388 }, { "epoch": 6.060649819494585, "grad_norm": 2.048849582672119, "learning_rate": 6.969675090252708e-05, "loss": 1.5576, "step": 8394 }, { "epoch": 6.064981949458484, "grad_norm": 2.022496461868286, "learning_rate": 6.967509025270758e-05, "loss": 1.6527, "step": 8400 }, { "epoch": 6.069314079422383, "grad_norm": 1.9454243183135986, "learning_rate": 6.965342960288809e-05, "loss": 1.6315, "step": 8406 }, { "epoch": 6.073646209386282, "grad_norm": 2.026052474975586, "learning_rate": 6.96317689530686e-05, "loss": 1.4864, "step": 8412 }, { "epoch": 6.077978339350181, "grad_norm": 1.978685975074768, "learning_rate": 6.96101083032491e-05, "loss": 1.6385, "step": 8418 }, { "epoch": 6.08231046931408, "grad_norm": 2.0477688312530518, "learning_rate": 6.958844765342961e-05, "loss": 1.582, "step": 8424 }, { "epoch": 6.086642599277979, "grad_norm": 1.9148788452148438, "learning_rate": 6.956678700361011e-05, "loss": 1.5711, "step": 8430 }, { "epoch": 6.090974729241878, "grad_norm": 1.978842854499817, "learning_rate": 6.954512635379062e-05, "loss": 1.5797, "step": 8436 }, { "epoch": 6.095306859205776, "grad_norm": 1.8769779205322266, "learning_rate": 6.952346570397112e-05, "loss": 1.6348, "step": 8442 }, { "epoch": 6.099638989169675, "grad_norm": 1.8636958599090576, "learning_rate": 6.950180505415162e-05, "loss": 1.6488, "step": 8448 }, { "epoch": 6.103971119133574, "grad_norm": 1.875897765159607, "learning_rate": 6.948014440433214e-05, "loss": 1.4985, "step": 8454 }, { "epoch": 6.108303249097473, "grad_norm": 1.8161344528198242, "learning_rate": 6.945848375451264e-05, "loss": 1.6491, "step": 8460 }, { "epoch": 6.112635379061372, "grad_norm": 1.8347136974334717, "learning_rate": 6.943682310469315e-05, "loss": 1.4884, "step": 8466 }, { "epoch": 6.116967509025271, "grad_norm": 1.9021079540252686, "learning_rate": 6.941516245487365e-05, "loss": 1.6232, "step": 8472 }, { "epoch": 6.12129963898917, "grad_norm": 1.7561743259429932, "learning_rate": 6.939350180505415e-05, "loss": 1.6266, "step": 8478 }, { "epoch": 6.125631768953069, "grad_norm": 1.7779520750045776, "learning_rate": 6.937184115523467e-05, "loss": 1.6627, "step": 8484 }, { "epoch": 6.129963898916968, "grad_norm": 1.8438737392425537, "learning_rate": 6.935018050541516e-05, "loss": 1.6294, "step": 8490 }, { "epoch": 6.134296028880867, "grad_norm": 1.9633796215057373, "learning_rate": 6.932851985559568e-05, "loss": 1.6617, "step": 8496 }, { "epoch": 6.1386281588447655, "grad_norm": 1.9247891902923584, "learning_rate": 6.930685920577618e-05, "loss": 1.7389, "step": 8502 }, { "epoch": 6.1429602888086645, "grad_norm": 1.8127788305282593, "learning_rate": 6.928519855595668e-05, "loss": 1.5572, "step": 8508 }, { "epoch": 6.1472924187725635, "grad_norm": 1.7081775665283203, "learning_rate": 6.926353790613718e-05, "loss": 1.6465, "step": 8514 }, { "epoch": 6.1516245487364625, "grad_norm": 1.823668122291565, "learning_rate": 6.924187725631769e-05, "loss": 1.629, "step": 8520 }, { "epoch": 6.1559566787003615, "grad_norm": 1.651544213294983, "learning_rate": 6.92202166064982e-05, "loss": 1.6163, "step": 8526 }, { "epoch": 6.1602888086642595, "grad_norm": 2.1959152221679688, "learning_rate": 6.91985559566787e-05, "loss": 1.5075, "step": 8532 }, { "epoch": 6.1646209386281585, "grad_norm": 1.8243144750595093, "learning_rate": 6.917689530685921e-05, "loss": 1.4587, "step": 8538 }, { "epoch": 6.1689530685920575, "grad_norm": 1.6942427158355713, "learning_rate": 6.915523465703971e-05, "loss": 1.6224, "step": 8544 }, { "epoch": 6.1732851985559565, "grad_norm": 1.8657515048980713, "learning_rate": 6.913357400722023e-05, "loss": 1.6801, "step": 8550 }, { "epoch": 6.1776173285198555, "grad_norm": 1.8314465284347534, "learning_rate": 6.911191335740072e-05, "loss": 1.6042, "step": 8556 }, { "epoch": 6.181949458483754, "grad_norm": 1.8294448852539062, "learning_rate": 6.909025270758122e-05, "loss": 1.6886, "step": 8562 }, { "epoch": 6.186281588447653, "grad_norm": 2.0211520195007324, "learning_rate": 6.906859205776174e-05, "loss": 1.6093, "step": 8568 }, { "epoch": 6.190613718411552, "grad_norm": 1.743374228477478, "learning_rate": 6.904693140794224e-05, "loss": 1.5789, "step": 8574 }, { "epoch": 6.194945848375451, "grad_norm": 1.955727458000183, "learning_rate": 6.902527075812275e-05, "loss": 1.5903, "step": 8580 }, { "epoch": 6.19927797833935, "grad_norm": 1.9525341987609863, "learning_rate": 6.900361010830325e-05, "loss": 1.5908, "step": 8586 }, { "epoch": 6.203610108303249, "grad_norm": 1.9689329862594604, "learning_rate": 6.898194945848377e-05, "loss": 1.5535, "step": 8592 }, { "epoch": 6.207942238267148, "grad_norm": 1.86732816696167, "learning_rate": 6.896028880866427e-05, "loss": 1.5747, "step": 8598 }, { "epoch": 6.212274368231047, "grad_norm": 1.9417610168457031, "learning_rate": 6.893862815884476e-05, "loss": 1.6257, "step": 8604 }, { "epoch": 6.216606498194946, "grad_norm": 1.813456416130066, "learning_rate": 6.891696750902528e-05, "loss": 1.5718, "step": 8610 }, { "epoch": 6.220938628158844, "grad_norm": 1.9809274673461914, "learning_rate": 6.889530685920578e-05, "loss": 1.5563, "step": 8616 }, { "epoch": 6.225270758122743, "grad_norm": 1.867325782775879, "learning_rate": 6.887364620938628e-05, "loss": 1.63, "step": 8622 }, { "epoch": 6.229602888086642, "grad_norm": 1.8962361812591553, "learning_rate": 6.885198555956679e-05, "loss": 1.6065, "step": 8628 }, { "epoch": 6.233935018050541, "grad_norm": 1.80984628200531, "learning_rate": 6.883032490974729e-05, "loss": 1.6844, "step": 8634 }, { "epoch": 6.23826714801444, "grad_norm": 1.9162397384643555, "learning_rate": 6.880866425992781e-05, "loss": 1.5853, "step": 8640 }, { "epoch": 6.242599277978339, "grad_norm": 1.8071461915969849, "learning_rate": 6.87870036101083e-05, "loss": 1.6482, "step": 8646 }, { "epoch": 6.246931407942238, "grad_norm": 1.9177442789077759, "learning_rate": 6.876534296028881e-05, "loss": 1.7703, "step": 8652 }, { "epoch": 6.251263537906137, "grad_norm": 1.7440309524536133, "learning_rate": 6.874368231046932e-05, "loss": 1.6117, "step": 8658 }, { "epoch": 6.255595667870036, "grad_norm": 1.8345636129379272, "learning_rate": 6.872202166064983e-05, "loss": 1.5718, "step": 8664 }, { "epoch": 6.259927797833935, "grad_norm": 1.8380975723266602, "learning_rate": 6.870036101083032e-05, "loss": 1.625, "step": 8670 }, { "epoch": 6.264259927797834, "grad_norm": 2.0022976398468018, "learning_rate": 6.867870036101083e-05, "loss": 1.6157, "step": 8676 }, { "epoch": 6.268592057761733, "grad_norm": 2.23987078666687, "learning_rate": 6.865703971119134e-05, "loss": 1.5732, "step": 8682 }, { "epoch": 6.272924187725632, "grad_norm": 1.9172440767288208, "learning_rate": 6.863537906137185e-05, "loss": 1.6478, "step": 8688 }, { "epoch": 6.277256317689531, "grad_norm": 1.7828422784805298, "learning_rate": 6.861371841155235e-05, "loss": 1.5883, "step": 8694 }, { "epoch": 6.28158844765343, "grad_norm": 2.020284652709961, "learning_rate": 6.859205776173285e-05, "loss": 1.7764, "step": 8700 }, { "epoch": 6.285920577617328, "grad_norm": 2.255997657775879, "learning_rate": 6.857039711191337e-05, "loss": 1.632, "step": 8706 }, { "epoch": 6.290252707581227, "grad_norm": 2.049473285675049, "learning_rate": 6.854873646209386e-05, "loss": 1.6649, "step": 8712 }, { "epoch": 6.294584837545126, "grad_norm": 2.04447603225708, "learning_rate": 6.852707581227436e-05, "loss": 1.7226, "step": 8718 }, { "epoch": 6.298916967509025, "grad_norm": 1.674225091934204, "learning_rate": 6.850541516245488e-05, "loss": 1.6309, "step": 8724 }, { "epoch": 6.303249097472924, "grad_norm": 1.9397810697555542, "learning_rate": 6.848375451263538e-05, "loss": 1.5018, "step": 8730 }, { "epoch": 6.307581227436823, "grad_norm": 1.943455696105957, "learning_rate": 6.846209386281589e-05, "loss": 1.6552, "step": 8736 }, { "epoch": 6.311913357400722, "grad_norm": 1.88715660572052, "learning_rate": 6.844043321299639e-05, "loss": 1.5014, "step": 8742 }, { "epoch": 6.316245487364621, "grad_norm": 1.8923841714859009, "learning_rate": 6.84187725631769e-05, "loss": 1.6985, "step": 8748 }, { "epoch": 6.32057761732852, "grad_norm": 1.9054538011550903, "learning_rate": 6.839711191335741e-05, "loss": 1.7143, "step": 8754 }, { "epoch": 6.324909747292419, "grad_norm": 1.8695518970489502, "learning_rate": 6.83754512635379e-05, "loss": 1.6057, "step": 8760 }, { "epoch": 6.329241877256318, "grad_norm": 1.814996361732483, "learning_rate": 6.835379061371842e-05, "loss": 1.8065, "step": 8766 }, { "epoch": 6.333574007220217, "grad_norm": 1.928772211074829, "learning_rate": 6.833212996389892e-05, "loss": 1.5736, "step": 8772 }, { "epoch": 6.337906137184116, "grad_norm": 1.9476656913757324, "learning_rate": 6.831046931407942e-05, "loss": 1.5939, "step": 8778 }, { "epoch": 6.342238267148014, "grad_norm": 1.7593858242034912, "learning_rate": 6.828880866425992e-05, "loss": 1.7132, "step": 8784 }, { "epoch": 6.346570397111913, "grad_norm": 1.877399206161499, "learning_rate": 6.826714801444044e-05, "loss": 1.7548, "step": 8790 }, { "epoch": 6.350902527075812, "grad_norm": 1.9802852869033813, "learning_rate": 6.824548736462094e-05, "loss": 1.6746, "step": 8796 }, { "epoch": 6.355234657039711, "grad_norm": 2.052037000656128, "learning_rate": 6.822382671480145e-05, "loss": 1.6229, "step": 8802 }, { "epoch": 6.35956678700361, "grad_norm": 1.7751929759979248, "learning_rate": 6.820216606498195e-05, "loss": 1.5719, "step": 8808 }, { "epoch": 6.363898916967509, "grad_norm": 1.6754508018493652, "learning_rate": 6.818050541516245e-05, "loss": 1.6498, "step": 8814 }, { "epoch": 6.368231046931408, "grad_norm": 1.7863086462020874, "learning_rate": 6.815884476534297e-05, "loss": 1.6083, "step": 8820 }, { "epoch": 6.372563176895307, "grad_norm": 2.056918144226074, "learning_rate": 6.813718411552346e-05, "loss": 1.6401, "step": 8826 }, { "epoch": 6.376895306859206, "grad_norm": 1.8217085599899292, "learning_rate": 6.811552346570398e-05, "loss": 1.704, "step": 8832 }, { "epoch": 6.381227436823105, "grad_norm": 1.9700219631195068, "learning_rate": 6.809386281588448e-05, "loss": 1.7144, "step": 8838 }, { "epoch": 6.385559566787004, "grad_norm": 1.9669568538665771, "learning_rate": 6.807220216606498e-05, "loss": 1.6877, "step": 8844 }, { "epoch": 6.389891696750903, "grad_norm": 2.051335096359253, "learning_rate": 6.805054151624549e-05, "loss": 1.784, "step": 8850 }, { "epoch": 6.394223826714802, "grad_norm": 1.8551676273345947, "learning_rate": 6.802888086642599e-05, "loss": 1.5433, "step": 8856 }, { "epoch": 6.398555956678701, "grad_norm": 1.9324445724487305, "learning_rate": 6.800722021660651e-05, "loss": 1.6051, "step": 8862 }, { "epoch": 6.4028880866426, "grad_norm": 2.040347099304199, "learning_rate": 6.798555956678701e-05, "loss": 1.6213, "step": 8868 }, { "epoch": 6.407220216606499, "grad_norm": 1.758521556854248, "learning_rate": 6.796389891696751e-05, "loss": 1.632, "step": 8874 }, { "epoch": 6.411552346570397, "grad_norm": 1.698585033416748, "learning_rate": 6.794223826714802e-05, "loss": 1.4973, "step": 8880 }, { "epoch": 6.415884476534296, "grad_norm": 1.9676415920257568, "learning_rate": 6.792057761732852e-05, "loss": 1.5476, "step": 8886 }, { "epoch": 6.420216606498195, "grad_norm": 1.7044501304626465, "learning_rate": 6.789891696750902e-05, "loss": 1.6078, "step": 8892 }, { "epoch": 6.424548736462094, "grad_norm": 2.077646255493164, "learning_rate": 6.787725631768953e-05, "loss": 1.6776, "step": 8898 }, { "epoch": 6.428880866425993, "grad_norm": 1.9515453577041626, "learning_rate": 6.785559566787004e-05, "loss": 1.603, "step": 8904 }, { "epoch": 6.433212996389892, "grad_norm": 1.8762223720550537, "learning_rate": 6.783393501805055e-05, "loss": 1.6841, "step": 8910 }, { "epoch": 6.437545126353791, "grad_norm": 2.109537124633789, "learning_rate": 6.781227436823105e-05, "loss": 1.625, "step": 8916 }, { "epoch": 6.44187725631769, "grad_norm": 1.8960450887680054, "learning_rate": 6.779061371841155e-05, "loss": 1.4738, "step": 8922 }, { "epoch": 6.446209386281589, "grad_norm": 1.9257605075836182, "learning_rate": 6.776895306859206e-05, "loss": 1.7297, "step": 8928 }, { "epoch": 6.4505415162454875, "grad_norm": 2.020355463027954, "learning_rate": 6.774729241877257e-05, "loss": 1.5502, "step": 8934 }, { "epoch": 6.4548736462093865, "grad_norm": 1.9276117086410522, "learning_rate": 6.772563176895306e-05, "loss": 1.6851, "step": 8940 }, { "epoch": 6.4592057761732855, "grad_norm": 1.8518027067184448, "learning_rate": 6.770397111913358e-05, "loss": 1.7678, "step": 8946 }, { "epoch": 6.4635379061371845, "grad_norm": 1.8601495027542114, "learning_rate": 6.768231046931408e-05, "loss": 1.7406, "step": 8952 }, { "epoch": 6.467870036101083, "grad_norm": 1.8449751138687134, "learning_rate": 6.766064981949459e-05, "loss": 1.747, "step": 8958 }, { "epoch": 6.4722021660649816, "grad_norm": 2.1348469257354736, "learning_rate": 6.763898916967509e-05, "loss": 1.6354, "step": 8964 }, { "epoch": 6.4765342960288805, "grad_norm": 1.878867745399475, "learning_rate": 6.761732851985559e-05, "loss": 1.5962, "step": 8970 }, { "epoch": 6.4808664259927795, "grad_norm": 2.0197739601135254, "learning_rate": 6.759566787003611e-05, "loss": 1.647, "step": 8976 }, { "epoch": 6.4851985559566785, "grad_norm": 1.7973934412002563, "learning_rate": 6.757400722021661e-05, "loss": 1.5183, "step": 8982 }, { "epoch": 6.4895306859205775, "grad_norm": 1.891770839691162, "learning_rate": 6.755234657039712e-05, "loss": 1.606, "step": 8988 }, { "epoch": 6.4938628158844764, "grad_norm": 1.9712717533111572, "learning_rate": 6.753068592057762e-05, "loss": 1.7466, "step": 8994 }, { "epoch": 6.498194945848375, "grad_norm": 2.2113778591156006, "learning_rate": 6.750902527075814e-05, "loss": 1.5073, "step": 9000 }, { "epoch": 6.502527075812274, "grad_norm": 1.8791521787643433, "learning_rate": 6.748736462093863e-05, "loss": 1.5587, "step": 9006 }, { "epoch": 6.506859205776173, "grad_norm": 1.8182092905044556, "learning_rate": 6.746570397111913e-05, "loss": 1.7694, "step": 9012 }, { "epoch": 6.511191335740072, "grad_norm": 1.6269865036010742, "learning_rate": 6.744404332129965e-05, "loss": 1.6184, "step": 9018 }, { "epoch": 6.515523465703971, "grad_norm": 1.8318636417388916, "learning_rate": 6.742238267148015e-05, "loss": 1.5954, "step": 9024 }, { "epoch": 6.51985559566787, "grad_norm": 1.7524454593658447, "learning_rate": 6.740072202166065e-05, "loss": 1.5757, "step": 9030 }, { "epoch": 6.524187725631769, "grad_norm": 1.817063808441162, "learning_rate": 6.737906137184116e-05, "loss": 1.6276, "step": 9036 }, { "epoch": 6.528519855595668, "grad_norm": 1.9085137844085693, "learning_rate": 6.735740072202167e-05, "loss": 1.6832, "step": 9042 }, { "epoch": 6.532851985559567, "grad_norm": 1.7129305601119995, "learning_rate": 6.733574007220218e-05, "loss": 1.6891, "step": 9048 }, { "epoch": 6.537184115523465, "grad_norm": 2.0315263271331787, "learning_rate": 6.731407942238266e-05, "loss": 1.7309, "step": 9054 }, { "epoch": 6.541516245487364, "grad_norm": 1.986680269241333, "learning_rate": 6.729241877256318e-05, "loss": 1.6228, "step": 9060 }, { "epoch": 6.545848375451263, "grad_norm": 1.9846749305725098, "learning_rate": 6.727075812274368e-05, "loss": 1.6593, "step": 9066 }, { "epoch": 6.550180505415162, "grad_norm": 1.9363188743591309, "learning_rate": 6.724909747292419e-05, "loss": 1.6043, "step": 9072 }, { "epoch": 6.554512635379061, "grad_norm": 1.9625585079193115, "learning_rate": 6.722743682310469e-05, "loss": 1.609, "step": 9078 }, { "epoch": 6.55884476534296, "grad_norm": 1.833266258239746, "learning_rate": 6.720577617328521e-05, "loss": 1.5569, "step": 9084 }, { "epoch": 6.563176895306859, "grad_norm": 1.6867882013320923, "learning_rate": 6.718411552346571e-05, "loss": 1.4681, "step": 9090 }, { "epoch": 6.567509025270758, "grad_norm": 2.290539264678955, "learning_rate": 6.716245487364621e-05, "loss": 1.6535, "step": 9096 }, { "epoch": 6.571841155234657, "grad_norm": 1.773980736732483, "learning_rate": 6.714079422382672e-05, "loss": 1.6891, "step": 9102 }, { "epoch": 6.576173285198556, "grad_norm": 1.8579740524291992, "learning_rate": 6.711913357400722e-05, "loss": 1.8043, "step": 9108 }, { "epoch": 6.580505415162455, "grad_norm": 1.8755165338516235, "learning_rate": 6.709747292418774e-05, "loss": 1.5558, "step": 9114 }, { "epoch": 6.584837545126354, "grad_norm": 1.9752403497695923, "learning_rate": 6.707581227436823e-05, "loss": 1.6012, "step": 9120 }, { "epoch": 6.589169675090253, "grad_norm": 1.8332537412643433, "learning_rate": 6.705415162454874e-05, "loss": 1.7184, "step": 9126 }, { "epoch": 6.593501805054151, "grad_norm": 1.9886561632156372, "learning_rate": 6.703249097472925e-05, "loss": 1.6171, "step": 9132 }, { "epoch": 6.59783393501805, "grad_norm": 1.8968223333358765, "learning_rate": 6.701083032490975e-05, "loss": 1.6253, "step": 9138 }, { "epoch": 6.602166064981949, "grad_norm": 1.8025635480880737, "learning_rate": 6.698916967509025e-05, "loss": 1.5397, "step": 9144 }, { "epoch": 6.606498194945848, "grad_norm": 2.214709520339966, "learning_rate": 6.696750902527076e-05, "loss": 1.8002, "step": 9150 }, { "epoch": 6.610830324909747, "grad_norm": 2.0061392784118652, "learning_rate": 6.694584837545127e-05, "loss": 1.6233, "step": 9156 }, { "epoch": 6.615162454873646, "grad_norm": 1.9379751682281494, "learning_rate": 6.692418772563178e-05, "loss": 1.6302, "step": 9162 }, { "epoch": 6.619494584837545, "grad_norm": 1.9627916812896729, "learning_rate": 6.690252707581228e-05, "loss": 1.6174, "step": 9168 }, { "epoch": 6.623826714801444, "grad_norm": 1.8913462162017822, "learning_rate": 6.688086642599278e-05, "loss": 1.6563, "step": 9174 }, { "epoch": 6.628158844765343, "grad_norm": 1.9336371421813965, "learning_rate": 6.685920577617329e-05, "loss": 1.6232, "step": 9180 }, { "epoch": 6.632490974729242, "grad_norm": 2.0348851680755615, "learning_rate": 6.683754512635379e-05, "loss": 1.6193, "step": 9186 }, { "epoch": 6.636823104693141, "grad_norm": 1.9607568979263306, "learning_rate": 6.68158844765343e-05, "loss": 1.6672, "step": 9192 }, { "epoch": 6.64115523465704, "grad_norm": 2.1042988300323486, "learning_rate": 6.679422382671481e-05, "loss": 1.5483, "step": 9198 }, { "epoch": 6.645487364620939, "grad_norm": 1.8489630222320557, "learning_rate": 6.677256317689531e-05, "loss": 1.506, "step": 9204 }, { "epoch": 6.649819494584838, "grad_norm": 1.7229171991348267, "learning_rate": 6.675090252707582e-05, "loss": 1.6085, "step": 9210 }, { "epoch": 6.654151624548737, "grad_norm": 1.9341528415679932, "learning_rate": 6.672924187725632e-05, "loss": 1.6384, "step": 9216 }, { "epoch": 6.658483754512636, "grad_norm": 1.7496291399002075, "learning_rate": 6.670758122743682e-05, "loss": 1.3975, "step": 9222 }, { "epoch": 6.662815884476534, "grad_norm": 1.8554860353469849, "learning_rate": 6.668592057761734e-05, "loss": 1.5555, "step": 9228 }, { "epoch": 6.667148014440433, "grad_norm": 1.975528359413147, "learning_rate": 6.666425992779783e-05, "loss": 1.5521, "step": 9234 }, { "epoch": 6.671480144404332, "grad_norm": 1.8701903820037842, "learning_rate": 6.664259927797835e-05, "loss": 1.5925, "step": 9240 }, { "epoch": 6.675812274368231, "grad_norm": 1.7620608806610107, "learning_rate": 6.662093862815885e-05, "loss": 1.6952, "step": 9246 }, { "epoch": 6.68014440433213, "grad_norm": 1.9155280590057373, "learning_rate": 6.659927797833935e-05, "loss": 1.6238, "step": 9252 }, { "epoch": 6.684476534296029, "grad_norm": 1.8315355777740479, "learning_rate": 6.657761732851986e-05, "loss": 1.5846, "step": 9258 }, { "epoch": 6.688808664259928, "grad_norm": 1.900761365890503, "learning_rate": 6.655595667870036e-05, "loss": 1.6277, "step": 9264 }, { "epoch": 6.693140794223827, "grad_norm": 1.9963537454605103, "learning_rate": 6.653429602888088e-05, "loss": 1.7139, "step": 9270 }, { "epoch": 6.697472924187726, "grad_norm": 1.8544288873672485, "learning_rate": 6.651263537906137e-05, "loss": 1.6635, "step": 9276 }, { "epoch": 6.701805054151625, "grad_norm": 2.12528657913208, "learning_rate": 6.649097472924188e-05, "loss": 1.7027, "step": 9282 }, { "epoch": 6.706137184115524, "grad_norm": 2.026744842529297, "learning_rate": 6.646931407942239e-05, "loss": 1.6495, "step": 9288 }, { "epoch": 6.710469314079423, "grad_norm": 1.8314309120178223, "learning_rate": 6.64476534296029e-05, "loss": 1.6397, "step": 9294 }, { "epoch": 6.714801444043322, "grad_norm": 2.120835304260254, "learning_rate": 6.642599277978339e-05, "loss": 1.6582, "step": 9300 }, { "epoch": 6.71913357400722, "grad_norm": 1.7651792764663696, "learning_rate": 6.64043321299639e-05, "loss": 1.5786, "step": 9306 }, { "epoch": 6.723465703971119, "grad_norm": 1.9016015529632568, "learning_rate": 6.638267148014441e-05, "loss": 1.6149, "step": 9312 }, { "epoch": 6.727797833935018, "grad_norm": 2.0158910751342773, "learning_rate": 6.636101083032492e-05, "loss": 1.5648, "step": 9318 }, { "epoch": 6.732129963898917, "grad_norm": 1.9612637758255005, "learning_rate": 6.633935018050542e-05, "loss": 1.5953, "step": 9324 }, { "epoch": 6.736462093862816, "grad_norm": 2.1548373699188232, "learning_rate": 6.631768953068592e-05, "loss": 1.579, "step": 9330 }, { "epoch": 6.740794223826715, "grad_norm": 1.8903685808181763, "learning_rate": 6.629602888086644e-05, "loss": 1.7388, "step": 9336 }, { "epoch": 6.745126353790614, "grad_norm": 2.003770112991333, "learning_rate": 6.627436823104694e-05, "loss": 1.5853, "step": 9342 }, { "epoch": 6.749458483754513, "grad_norm": 1.9280061721801758, "learning_rate": 6.625270758122743e-05, "loss": 1.6198, "step": 9348 }, { "epoch": 6.753790613718412, "grad_norm": 1.9483157396316528, "learning_rate": 6.623104693140795e-05, "loss": 1.5245, "step": 9354 }, { "epoch": 6.758122743682311, "grad_norm": 2.012418270111084, "learning_rate": 6.620938628158845e-05, "loss": 1.5906, "step": 9360 }, { "epoch": 6.76245487364621, "grad_norm": 1.9264795780181885, "learning_rate": 6.618772563176895e-05, "loss": 1.5913, "step": 9366 }, { "epoch": 6.7667870036101085, "grad_norm": 1.846164584159851, "learning_rate": 6.616606498194946e-05, "loss": 1.6123, "step": 9372 }, { "epoch": 6.7711191335740075, "grad_norm": 1.946144700050354, "learning_rate": 6.614440433212997e-05, "loss": 1.5974, "step": 9378 }, { "epoch": 6.7754512635379065, "grad_norm": 1.83589768409729, "learning_rate": 6.612274368231048e-05, "loss": 1.5937, "step": 9384 }, { "epoch": 6.7797833935018055, "grad_norm": 1.694117784500122, "learning_rate": 6.610108303249097e-05, "loss": 1.637, "step": 9390 }, { "epoch": 6.7841155234657045, "grad_norm": 1.7106013298034668, "learning_rate": 6.607942238267148e-05, "loss": 1.535, "step": 9396 }, { "epoch": 6.7884476534296025, "grad_norm": 1.7247287034988403, "learning_rate": 6.605776173285199e-05, "loss": 1.6729, "step": 9402 }, { "epoch": 6.7927797833935015, "grad_norm": 1.809632420539856, "learning_rate": 6.60361010830325e-05, "loss": 1.5617, "step": 9408 }, { "epoch": 6.7971119133574005, "grad_norm": 1.9668958187103271, "learning_rate": 6.6014440433213e-05, "loss": 1.7769, "step": 9414 }, { "epoch": 6.8014440433212995, "grad_norm": 1.934672474861145, "learning_rate": 6.599277978339351e-05, "loss": 1.6644, "step": 9420 }, { "epoch": 6.8057761732851985, "grad_norm": 1.9261488914489746, "learning_rate": 6.597111913357401e-05, "loss": 1.5862, "step": 9426 }, { "epoch": 6.810108303249097, "grad_norm": 1.6786755323410034, "learning_rate": 6.594945848375452e-05, "loss": 1.6962, "step": 9432 }, { "epoch": 6.814440433212996, "grad_norm": 1.9339308738708496, "learning_rate": 6.592779783393502e-05, "loss": 1.6653, "step": 9438 }, { "epoch": 6.818772563176895, "grad_norm": 1.854048728942871, "learning_rate": 6.590613718411552e-05, "loss": 1.6179, "step": 9444 }, { "epoch": 6.823104693140794, "grad_norm": 1.9270546436309814, "learning_rate": 6.588447653429604e-05, "loss": 1.5964, "step": 9450 }, { "epoch": 6.827436823104693, "grad_norm": 1.9039897918701172, "learning_rate": 6.586281588447653e-05, "loss": 1.6344, "step": 9456 }, { "epoch": 6.831768953068592, "grad_norm": 1.9300066232681274, "learning_rate": 6.584115523465705e-05, "loss": 1.6282, "step": 9462 }, { "epoch": 6.836101083032491, "grad_norm": 1.9988186359405518, "learning_rate": 6.581949458483755e-05, "loss": 1.7048, "step": 9468 }, { "epoch": 6.840433212996389, "grad_norm": 2.1415481567382812, "learning_rate": 6.579783393501805e-05, "loss": 1.7227, "step": 9474 }, { "epoch": 6.844765342960288, "grad_norm": 1.6863512992858887, "learning_rate": 6.577617328519856e-05, "loss": 1.6355, "step": 9480 }, { "epoch": 6.849097472924187, "grad_norm": 1.9070063829421997, "learning_rate": 6.575451263537906e-05, "loss": 1.6095, "step": 9486 }, { "epoch": 6.853429602888086, "grad_norm": 2.0130062103271484, "learning_rate": 6.573285198555958e-05, "loss": 1.7036, "step": 9492 }, { "epoch": 6.857761732851985, "grad_norm": 1.894051432609558, "learning_rate": 6.571119133574008e-05, "loss": 1.7516, "step": 9498 }, { "epoch": 6.862093862815884, "grad_norm": 2.095449447631836, "learning_rate": 6.568953068592058e-05, "loss": 1.4731, "step": 9504 }, { "epoch": 6.866425992779783, "grad_norm": 1.998462438583374, "learning_rate": 6.566787003610109e-05, "loss": 1.6074, "step": 9510 }, { "epoch": 6.870758122743682, "grad_norm": 1.7657119035720825, "learning_rate": 6.564620938628159e-05, "loss": 1.5289, "step": 9516 }, { "epoch": 6.875090252707581, "grad_norm": 1.9929907321929932, "learning_rate": 6.562454873646209e-05, "loss": 1.6988, "step": 9522 }, { "epoch": 6.87942238267148, "grad_norm": 1.887926697731018, "learning_rate": 6.56028880866426e-05, "loss": 1.7523, "step": 9528 }, { "epoch": 6.883754512635379, "grad_norm": 1.7682982683181763, "learning_rate": 6.558122743682311e-05, "loss": 1.5947, "step": 9534 }, { "epoch": 6.888086642599278, "grad_norm": 1.8641672134399414, "learning_rate": 6.555956678700362e-05, "loss": 1.6332, "step": 9540 }, { "epoch": 6.892418772563177, "grad_norm": 1.9179068803787231, "learning_rate": 6.553790613718412e-05, "loss": 1.6045, "step": 9546 }, { "epoch": 6.896750902527076, "grad_norm": 1.953715205192566, "learning_rate": 6.551624548736462e-05, "loss": 1.7341, "step": 9552 }, { "epoch": 6.901083032490975, "grad_norm": 1.7356258630752563, "learning_rate": 6.549458483754513e-05, "loss": 1.5945, "step": 9558 }, { "epoch": 6.905415162454874, "grad_norm": 1.8389780521392822, "learning_rate": 6.547292418772564e-05, "loss": 1.5545, "step": 9564 }, { "epoch": 6.909747292418773, "grad_norm": 1.9269793033599854, "learning_rate": 6.545126353790613e-05, "loss": 1.5318, "step": 9570 }, { "epoch": 6.914079422382671, "grad_norm": 1.6845364570617676, "learning_rate": 6.542960288808665e-05, "loss": 1.5643, "step": 9576 }, { "epoch": 6.91841155234657, "grad_norm": 1.7713559865951538, "learning_rate": 6.540794223826715e-05, "loss": 1.6762, "step": 9582 }, { "epoch": 6.922743682310469, "grad_norm": 1.7227777242660522, "learning_rate": 6.538628158844767e-05, "loss": 1.671, "step": 9588 }, { "epoch": 6.927075812274368, "grad_norm": 1.954413652420044, "learning_rate": 6.536462093862816e-05, "loss": 1.5898, "step": 9594 }, { "epoch": 6.931407942238267, "grad_norm": 1.7818641662597656, "learning_rate": 6.534296028880866e-05, "loss": 1.6825, "step": 9600 }, { "epoch": 6.935740072202166, "grad_norm": 1.8194085359573364, "learning_rate": 6.532129963898918e-05, "loss": 1.5489, "step": 9606 }, { "epoch": 6.940072202166065, "grad_norm": 2.059938907623291, "learning_rate": 6.529963898916968e-05, "loss": 1.5819, "step": 9612 }, { "epoch": 6.944404332129964, "grad_norm": 1.6424115896224976, "learning_rate": 6.527797833935018e-05, "loss": 1.7047, "step": 9618 }, { "epoch": 6.948736462093863, "grad_norm": 1.7853312492370605, "learning_rate": 6.525631768953069e-05, "loss": 1.6392, "step": 9624 }, { "epoch": 6.953068592057762, "grad_norm": 1.7465037107467651, "learning_rate": 6.523465703971119e-05, "loss": 1.6903, "step": 9630 }, { "epoch": 6.957400722021661, "grad_norm": 2.140524387359619, "learning_rate": 6.52129963898917e-05, "loss": 1.699, "step": 9636 }, { "epoch": 6.96173285198556, "grad_norm": 1.849637508392334, "learning_rate": 6.51913357400722e-05, "loss": 1.6892, "step": 9642 }, { "epoch": 6.966064981949458, "grad_norm": 2.054187536239624, "learning_rate": 6.516967509025271e-05, "loss": 1.7087, "step": 9648 }, { "epoch": 6.970397111913357, "grad_norm": 1.9495089054107666, "learning_rate": 6.514801444043322e-05, "loss": 1.6835, "step": 9654 }, { "epoch": 6.974729241877256, "grad_norm": 1.8121434450149536, "learning_rate": 6.512635379061372e-05, "loss": 1.5588, "step": 9660 }, { "epoch": 6.979061371841155, "grad_norm": 1.8349846601486206, "learning_rate": 6.510469314079422e-05, "loss": 1.6254, "step": 9666 }, { "epoch": 6.983393501805054, "grad_norm": 1.8339335918426514, "learning_rate": 6.508303249097473e-05, "loss": 1.7498, "step": 9672 }, { "epoch": 6.987725631768953, "grad_norm": 1.8546510934829712, "learning_rate": 6.506137184115524e-05, "loss": 1.7396, "step": 9678 }, { "epoch": 6.992057761732852, "grad_norm": 1.6735299825668335, "learning_rate": 6.503971119133573e-05, "loss": 1.6569, "step": 9684 }, { "epoch": 6.996389891696751, "grad_norm": 1.9428467750549316, "learning_rate": 6.501805054151625e-05, "loss": 1.6058, "step": 9690 }, { "epoch": 7.00072202166065, "grad_norm": 1.6741836071014404, "learning_rate": 6.499638989169675e-05, "loss": 1.7519, "step": 9696 }, { "epoch": 7.005054151624549, "grad_norm": 1.9725664854049683, "learning_rate": 6.497472924187726e-05, "loss": 1.5968, "step": 9702 }, { "epoch": 7.009386281588448, "grad_norm": 1.953238606452942, "learning_rate": 6.495306859205776e-05, "loss": 1.459, "step": 9708 }, { "epoch": 7.013718411552347, "grad_norm": 1.916232943534851, "learning_rate": 6.493140794223826e-05, "loss": 1.5953, "step": 9714 }, { "epoch": 7.018050541516246, "grad_norm": 1.9370999336242676, "learning_rate": 6.490974729241878e-05, "loss": 1.4744, "step": 9720 }, { "epoch": 7.022382671480145, "grad_norm": 2.077833652496338, "learning_rate": 6.488808664259928e-05, "loss": 1.5515, "step": 9726 }, { "epoch": 7.026714801444044, "grad_norm": 1.9519095420837402, "learning_rate": 6.486642599277979e-05, "loss": 1.6417, "step": 9732 }, { "epoch": 7.031046931407943, "grad_norm": 1.8029378652572632, "learning_rate": 6.484476534296029e-05, "loss": 1.578, "step": 9738 }, { "epoch": 7.035379061371841, "grad_norm": 2.0262856483459473, "learning_rate": 6.482310469314081e-05, "loss": 1.6254, "step": 9744 }, { "epoch": 7.03971119133574, "grad_norm": 1.9552550315856934, "learning_rate": 6.48014440433213e-05, "loss": 1.6448, "step": 9750 }, { "epoch": 7.044043321299639, "grad_norm": 2.0212457180023193, "learning_rate": 6.47797833935018e-05, "loss": 1.5362, "step": 9756 }, { "epoch": 7.048375451263538, "grad_norm": 1.9013935327529907, "learning_rate": 6.475812274368232e-05, "loss": 1.4803, "step": 9762 }, { "epoch": 7.052707581227437, "grad_norm": 1.9728074073791504, "learning_rate": 6.473646209386282e-05, "loss": 1.7513, "step": 9768 }, { "epoch": 7.057039711191336, "grad_norm": 1.9400535821914673, "learning_rate": 6.471480144404332e-05, "loss": 1.5614, "step": 9774 }, { "epoch": 7.061371841155235, "grad_norm": 1.8947036266326904, "learning_rate": 6.469314079422383e-05, "loss": 1.5275, "step": 9780 }, { "epoch": 7.065703971119134, "grad_norm": 1.8430978059768677, "learning_rate": 6.467148014440434e-05, "loss": 1.5661, "step": 9786 }, { "epoch": 7.070036101083033, "grad_norm": 2.047276496887207, "learning_rate": 6.464981949458485e-05, "loss": 1.5472, "step": 9792 }, { "epoch": 7.074368231046932, "grad_norm": 1.8033576011657715, "learning_rate": 6.462815884476534e-05, "loss": 1.5336, "step": 9798 }, { "epoch": 7.0787003610108306, "grad_norm": 1.9022314548492432, "learning_rate": 6.460649819494585e-05, "loss": 1.64, "step": 9804 }, { "epoch": 7.0830324909747295, "grad_norm": 2.0333640575408936, "learning_rate": 6.458483754512636e-05, "loss": 1.6599, "step": 9810 }, { "epoch": 7.0873646209386285, "grad_norm": 2.368354320526123, "learning_rate": 6.456317689530686e-05, "loss": 1.6435, "step": 9816 }, { "epoch": 7.0916967509025275, "grad_norm": 1.7389042377471924, "learning_rate": 6.454151624548736e-05, "loss": 1.517, "step": 9822 }, { "epoch": 7.096028880866426, "grad_norm": 1.7489292621612549, "learning_rate": 6.451985559566788e-05, "loss": 1.5783, "step": 9828 }, { "epoch": 7.100361010830325, "grad_norm": 1.896857738494873, "learning_rate": 6.449819494584838e-05, "loss": 1.6548, "step": 9834 }, { "epoch": 7.1046931407942235, "grad_norm": 1.8568108081817627, "learning_rate": 6.447653429602889e-05, "loss": 1.6391, "step": 9840 }, { "epoch": 7.1090252707581225, "grad_norm": 2.120563268661499, "learning_rate": 6.445487364620939e-05, "loss": 1.6806, "step": 9846 }, { "epoch": 7.1133574007220215, "grad_norm": 1.9083834886550903, "learning_rate": 6.443321299638989e-05, "loss": 1.5679, "step": 9852 }, { "epoch": 7.1176895306859205, "grad_norm": 2.0347933769226074, "learning_rate": 6.441155234657041e-05, "loss": 1.5182, "step": 9858 }, { "epoch": 7.1220216606498195, "grad_norm": 1.9091744422912598, "learning_rate": 6.43898916967509e-05, "loss": 1.6323, "step": 9864 }, { "epoch": 7.126353790613718, "grad_norm": 1.9169838428497314, "learning_rate": 6.436823104693142e-05, "loss": 1.6561, "step": 9870 }, { "epoch": 7.130685920577617, "grad_norm": 1.838072419166565, "learning_rate": 6.434657039711192e-05, "loss": 1.5616, "step": 9876 }, { "epoch": 7.135018050541516, "grad_norm": 1.902567744255066, "learning_rate": 6.432490974729242e-05, "loss": 1.5747, "step": 9882 }, { "epoch": 7.139350180505415, "grad_norm": 2.016021490097046, "learning_rate": 6.430324909747292e-05, "loss": 1.549, "step": 9888 }, { "epoch": 7.143682310469314, "grad_norm": 2.2396724224090576, "learning_rate": 6.428158844765343e-05, "loss": 1.6492, "step": 9894 }, { "epoch": 7.148014440433213, "grad_norm": 2.106273651123047, "learning_rate": 6.425992779783394e-05, "loss": 1.6973, "step": 9900 }, { "epoch": 7.152346570397112, "grad_norm": 1.8459246158599854, "learning_rate": 6.423826714801445e-05, "loss": 1.6326, "step": 9906 }, { "epoch": 7.156678700361011, "grad_norm": 1.9374557733535767, "learning_rate": 6.421660649819495e-05, "loss": 1.6553, "step": 9912 }, { "epoch": 7.161010830324909, "grad_norm": 2.0185606479644775, "learning_rate": 6.419494584837545e-05, "loss": 1.6337, "step": 9918 }, { "epoch": 7.165342960288808, "grad_norm": 1.6122630834579468, "learning_rate": 6.417328519855596e-05, "loss": 1.5229, "step": 9924 }, { "epoch": 7.169675090252707, "grad_norm": 1.8983932733535767, "learning_rate": 6.415162454873646e-05, "loss": 1.6132, "step": 9930 }, { "epoch": 7.174007220216606, "grad_norm": 1.8443197011947632, "learning_rate": 6.412996389891696e-05, "loss": 1.4938, "step": 9936 }, { "epoch": 7.178339350180505, "grad_norm": 1.8580642938613892, "learning_rate": 6.410830324909748e-05, "loss": 1.5286, "step": 9942 }, { "epoch": 7.182671480144404, "grad_norm": 1.7848647832870483, "learning_rate": 6.408664259927798e-05, "loss": 1.4799, "step": 9948 }, { "epoch": 7.187003610108303, "grad_norm": 1.7907495498657227, "learning_rate": 6.406498194945849e-05, "loss": 1.6383, "step": 9954 }, { "epoch": 7.191335740072202, "grad_norm": 1.9623292684555054, "learning_rate": 6.404332129963899e-05, "loss": 1.5931, "step": 9960 }, { "epoch": 7.195667870036101, "grad_norm": 2.069399118423462, "learning_rate": 6.40216606498195e-05, "loss": 1.5333, "step": 9966 }, { "epoch": 7.2, "grad_norm": 1.8380491733551025, "learning_rate": 6.400000000000001e-05, "loss": 1.5995, "step": 9972 }, { "epoch": 7.204332129963899, "grad_norm": 2.164081335067749, "learning_rate": 6.39783393501805e-05, "loss": 1.6551, "step": 9978 }, { "epoch": 7.208664259927798, "grad_norm": 1.9332292079925537, "learning_rate": 6.395667870036102e-05, "loss": 1.6189, "step": 9984 }, { "epoch": 7.212996389891697, "grad_norm": 1.7069607973098755, "learning_rate": 6.393501805054152e-05, "loss": 1.4511, "step": 9990 }, { "epoch": 7.217328519855596, "grad_norm": 2.2620248794555664, "learning_rate": 6.391335740072202e-05, "loss": 1.7671, "step": 9996 }, { "epoch": 7.221660649819494, "grad_norm": 1.914498209953308, "learning_rate": 6.389169675090253e-05, "loss": 1.6286, "step": 10002 }, { "epoch": 7.225992779783393, "grad_norm": 1.9255021810531616, "learning_rate": 6.387003610108303e-05, "loss": 1.7153, "step": 10008 }, { "epoch": 7.230324909747292, "grad_norm": 2.003011465072632, "learning_rate": 6.384837545126355e-05, "loss": 1.6262, "step": 10014 }, { "epoch": 7.234657039711191, "grad_norm": 2.1006407737731934, "learning_rate": 6.382671480144404e-05, "loss": 1.6362, "step": 10020 }, { "epoch": 7.23898916967509, "grad_norm": 1.9823328256607056, "learning_rate": 6.380505415162455e-05, "loss": 1.6379, "step": 10026 }, { "epoch": 7.243321299638989, "grad_norm": 2.03825044631958, "learning_rate": 6.378339350180506e-05, "loss": 1.5465, "step": 10032 }, { "epoch": 7.247653429602888, "grad_norm": 1.892566442489624, "learning_rate": 6.376173285198557e-05, "loss": 1.595, "step": 10038 }, { "epoch": 7.251985559566787, "grad_norm": 1.8918653726577759, "learning_rate": 6.374007220216606e-05, "loss": 1.6795, "step": 10044 }, { "epoch": 7.256317689530686, "grad_norm": 1.8963112831115723, "learning_rate": 6.371841155234657e-05, "loss": 1.5935, "step": 10050 }, { "epoch": 7.260649819494585, "grad_norm": 1.8471248149871826, "learning_rate": 6.369675090252708e-05, "loss": 1.5707, "step": 10056 }, { "epoch": 7.264981949458484, "grad_norm": 2.0307610034942627, "learning_rate": 6.367509025270759e-05, "loss": 1.5557, "step": 10062 }, { "epoch": 7.269314079422383, "grad_norm": 1.9639112949371338, "learning_rate": 6.365342960288809e-05, "loss": 1.5185, "step": 10068 }, { "epoch": 7.273646209386282, "grad_norm": 1.9077099561691284, "learning_rate": 6.363176895306859e-05, "loss": 1.6522, "step": 10074 }, { "epoch": 7.277978339350181, "grad_norm": 2.047818422317505, "learning_rate": 6.361010830324911e-05, "loss": 1.6564, "step": 10080 }, { "epoch": 7.28231046931408, "grad_norm": 2.343204975128174, "learning_rate": 6.358844765342961e-05, "loss": 1.6037, "step": 10086 }, { "epoch": 7.286642599277978, "grad_norm": 1.7131836414337158, "learning_rate": 6.35667870036101e-05, "loss": 1.5792, "step": 10092 }, { "epoch": 7.290974729241877, "grad_norm": 1.9226739406585693, "learning_rate": 6.354512635379062e-05, "loss": 1.5127, "step": 10098 }, { "epoch": 7.295306859205776, "grad_norm": 2.0165350437164307, "learning_rate": 6.352346570397112e-05, "loss": 1.5691, "step": 10104 }, { "epoch": 7.299638989169675, "grad_norm": 2.057013750076294, "learning_rate": 6.350180505415163e-05, "loss": 1.6338, "step": 10110 }, { "epoch": 7.303971119133574, "grad_norm": 1.9843287467956543, "learning_rate": 6.348014440433213e-05, "loss": 1.5139, "step": 10116 }, { "epoch": 7.308303249097473, "grad_norm": 1.9190555810928345, "learning_rate": 6.345848375451265e-05, "loss": 1.6107, "step": 10122 }, { "epoch": 7.312635379061372, "grad_norm": 1.943626046180725, "learning_rate": 6.343682310469315e-05, "loss": 1.642, "step": 10128 }, { "epoch": 7.316967509025271, "grad_norm": 2.110901117324829, "learning_rate": 6.341516245487364e-05, "loss": 1.4998, "step": 10134 }, { "epoch": 7.32129963898917, "grad_norm": 2.2016537189483643, "learning_rate": 6.339350180505416e-05, "loss": 1.5429, "step": 10140 }, { "epoch": 7.325631768953069, "grad_norm": 1.9564350843429565, "learning_rate": 6.337184115523466e-05, "loss": 1.6528, "step": 10146 }, { "epoch": 7.329963898916968, "grad_norm": 1.755905270576477, "learning_rate": 6.335018050541518e-05, "loss": 1.596, "step": 10152 }, { "epoch": 7.334296028880867, "grad_norm": 2.0203404426574707, "learning_rate": 6.332851985559566e-05, "loss": 1.6137, "step": 10158 }, { "epoch": 7.338628158844766, "grad_norm": 1.9278435707092285, "learning_rate": 6.330685920577618e-05, "loss": 1.5382, "step": 10164 }, { "epoch": 7.342960288808664, "grad_norm": 1.8637179136276245, "learning_rate": 6.328519855595668e-05, "loss": 1.6832, "step": 10170 }, { "epoch": 7.347292418772563, "grad_norm": 1.933638334274292, "learning_rate": 6.326353790613719e-05, "loss": 1.6726, "step": 10176 }, { "epoch": 7.351624548736462, "grad_norm": 1.9859590530395508, "learning_rate": 6.324187725631769e-05, "loss": 1.4695, "step": 10182 }, { "epoch": 7.355956678700361, "grad_norm": 2.138709783554077, "learning_rate": 6.32202166064982e-05, "loss": 1.6057, "step": 10188 }, { "epoch": 7.36028880866426, "grad_norm": 2.0569567680358887, "learning_rate": 6.319855595667871e-05, "loss": 1.4707, "step": 10194 }, { "epoch": 7.364620938628159, "grad_norm": 1.760939121246338, "learning_rate": 6.31768953068592e-05, "loss": 1.6315, "step": 10200 }, { "epoch": 7.368953068592058, "grad_norm": 1.9426867961883545, "learning_rate": 6.315523465703972e-05, "loss": 1.5678, "step": 10206 }, { "epoch": 7.373285198555957, "grad_norm": 2.101731538772583, "learning_rate": 6.313357400722022e-05, "loss": 1.5604, "step": 10212 }, { "epoch": 7.377617328519856, "grad_norm": 2.144639015197754, "learning_rate": 6.311191335740072e-05, "loss": 1.7942, "step": 10218 }, { "epoch": 7.381949458483755, "grad_norm": 1.8070834875106812, "learning_rate": 6.309025270758123e-05, "loss": 1.5736, "step": 10224 }, { "epoch": 7.386281588447654, "grad_norm": 1.9638293981552124, "learning_rate": 6.306859205776173e-05, "loss": 1.5538, "step": 10230 }, { "epoch": 7.390613718411553, "grad_norm": 1.959932565689087, "learning_rate": 6.304693140794225e-05, "loss": 1.6188, "step": 10236 }, { "epoch": 7.3949458483754515, "grad_norm": 1.92802894115448, "learning_rate": 6.302527075812275e-05, "loss": 1.5675, "step": 10242 }, { "epoch": 7.3992779783393505, "grad_norm": 2.0230445861816406, "learning_rate": 6.300361010830325e-05, "loss": 1.5541, "step": 10248 }, { "epoch": 7.4036101083032495, "grad_norm": 1.9767804145812988, "learning_rate": 6.298194945848376e-05, "loss": 1.6401, "step": 10254 }, { "epoch": 7.4079422382671485, "grad_norm": 2.07725191116333, "learning_rate": 6.296028880866426e-05, "loss": 1.5281, "step": 10260 }, { "epoch": 7.412274368231047, "grad_norm": 1.8108367919921875, "learning_rate": 6.293862815884476e-05, "loss": 1.5252, "step": 10266 }, { "epoch": 7.4166064981949456, "grad_norm": 1.89191734790802, "learning_rate": 6.291696750902527e-05, "loss": 1.4494, "step": 10272 }, { "epoch": 7.4209386281588445, "grad_norm": 1.9573228359222412, "learning_rate": 6.289530685920578e-05, "loss": 1.5215, "step": 10278 }, { "epoch": 7.4252707581227435, "grad_norm": 1.9287657737731934, "learning_rate": 6.287364620938629e-05, "loss": 1.5399, "step": 10284 }, { "epoch": 7.4296028880866425, "grad_norm": 1.7987972497940063, "learning_rate": 6.285198555956679e-05, "loss": 1.4723, "step": 10290 }, { "epoch": 7.4339350180505415, "grad_norm": 1.7862412929534912, "learning_rate": 6.28303249097473e-05, "loss": 1.6119, "step": 10296 }, { "epoch": 7.4382671480144404, "grad_norm": 2.0575904846191406, "learning_rate": 6.28086642599278e-05, "loss": 1.7715, "step": 10302 }, { "epoch": 7.442599277978339, "grad_norm": 2.125060558319092, "learning_rate": 6.278700361010831e-05, "loss": 1.6651, "step": 10308 }, { "epoch": 7.446931407942238, "grad_norm": 2.1584155559539795, "learning_rate": 6.27653429602888e-05, "loss": 1.6454, "step": 10314 }, { "epoch": 7.451263537906137, "grad_norm": 2.3380367755889893, "learning_rate": 6.274368231046932e-05, "loss": 1.7215, "step": 10320 }, { "epoch": 7.455595667870036, "grad_norm": 2.09023380279541, "learning_rate": 6.272202166064982e-05, "loss": 1.6722, "step": 10326 }, { "epoch": 7.459927797833935, "grad_norm": 1.9373562335968018, "learning_rate": 6.270036101083034e-05, "loss": 1.7093, "step": 10332 }, { "epoch": 7.464259927797834, "grad_norm": 2.2724509239196777, "learning_rate": 6.267870036101083e-05, "loss": 1.5977, "step": 10338 }, { "epoch": 7.468592057761732, "grad_norm": 2.130654811859131, "learning_rate": 6.265703971119133e-05, "loss": 1.5381, "step": 10344 }, { "epoch": 7.472924187725631, "grad_norm": 2.0568888187408447, "learning_rate": 6.263537906137185e-05, "loss": 1.5245, "step": 10350 }, { "epoch": 7.47725631768953, "grad_norm": 1.8806312084197998, "learning_rate": 6.261371841155235e-05, "loss": 1.6097, "step": 10356 }, { "epoch": 7.481588447653429, "grad_norm": 2.0314013957977295, "learning_rate": 6.259205776173286e-05, "loss": 1.5998, "step": 10362 }, { "epoch": 7.485920577617328, "grad_norm": 1.9467524290084839, "learning_rate": 6.257039711191336e-05, "loss": 1.5382, "step": 10368 }, { "epoch": 7.490252707581227, "grad_norm": 1.9848229885101318, "learning_rate": 6.254873646209388e-05, "loss": 1.524, "step": 10374 }, { "epoch": 7.494584837545126, "grad_norm": 2.045520067214966, "learning_rate": 6.252707581227437e-05, "loss": 1.6324, "step": 10380 }, { "epoch": 7.498916967509025, "grad_norm": 1.9880856275558472, "learning_rate": 6.250541516245487e-05, "loss": 1.569, "step": 10386 }, { "epoch": 7.503249097472924, "grad_norm": 1.9321931600570679, "learning_rate": 6.248375451263539e-05, "loss": 1.6197, "step": 10392 }, { "epoch": 7.507581227436823, "grad_norm": 2.009622097015381, "learning_rate": 6.246209386281589e-05, "loss": 1.5413, "step": 10398 }, { "epoch": 7.511913357400722, "grad_norm": 2.0088963508605957, "learning_rate": 6.244043321299639e-05, "loss": 1.5224, "step": 10404 }, { "epoch": 7.516245487364621, "grad_norm": 1.8381723165512085, "learning_rate": 6.24187725631769e-05, "loss": 1.5975, "step": 10410 }, { "epoch": 7.52057761732852, "grad_norm": 1.8380820751190186, "learning_rate": 6.239711191335741e-05, "loss": 1.5983, "step": 10416 }, { "epoch": 7.524909747292419, "grad_norm": 1.7824277877807617, "learning_rate": 6.237545126353792e-05, "loss": 1.4569, "step": 10422 }, { "epoch": 7.529241877256318, "grad_norm": 2.0210444927215576, "learning_rate": 6.23537906137184e-05, "loss": 1.5027, "step": 10428 }, { "epoch": 7.533574007220217, "grad_norm": 2.1629116535186768, "learning_rate": 6.233212996389892e-05, "loss": 1.6645, "step": 10434 }, { "epoch": 7.537906137184115, "grad_norm": 1.9851205348968506, "learning_rate": 6.231046931407942e-05, "loss": 1.65, "step": 10440 }, { "epoch": 7.542238267148014, "grad_norm": 1.7558478116989136, "learning_rate": 6.228880866425993e-05, "loss": 1.5519, "step": 10446 }, { "epoch": 7.546570397111913, "grad_norm": 1.9845895767211914, "learning_rate": 6.226714801444043e-05, "loss": 1.5678, "step": 10452 }, { "epoch": 7.550902527075812, "grad_norm": 1.857239842414856, "learning_rate": 6.224548736462095e-05, "loss": 1.5648, "step": 10458 }, { "epoch": 7.555234657039711, "grad_norm": 1.9618619680404663, "learning_rate": 6.222382671480145e-05, "loss": 1.5044, "step": 10464 }, { "epoch": 7.55956678700361, "grad_norm": 2.0372681617736816, "learning_rate": 6.220216606498195e-05, "loss": 1.5644, "step": 10470 }, { "epoch": 7.563898916967509, "grad_norm": 2.0167410373687744, "learning_rate": 6.218050541516246e-05, "loss": 1.5791, "step": 10476 }, { "epoch": 7.568231046931408, "grad_norm": 1.8650460243225098, "learning_rate": 6.215884476534296e-05, "loss": 1.5948, "step": 10482 }, { "epoch": 7.572563176895307, "grad_norm": 2.210024356842041, "learning_rate": 6.213718411552348e-05, "loss": 1.5915, "step": 10488 }, { "epoch": 7.576895306859206, "grad_norm": 2.1486449241638184, "learning_rate": 6.211552346570397e-05, "loss": 1.5708, "step": 10494 }, { "epoch": 7.581227436823105, "grad_norm": 1.828852891921997, "learning_rate": 6.209386281588448e-05, "loss": 1.4854, "step": 10500 }, { "epoch": 7.585559566787004, "grad_norm": 1.87925386428833, "learning_rate": 6.207220216606499e-05, "loss": 1.5893, "step": 10506 }, { "epoch": 7.589891696750902, "grad_norm": 1.9495599269866943, "learning_rate": 6.205054151624549e-05, "loss": 1.5665, "step": 10512 }, { "epoch": 7.594223826714801, "grad_norm": 1.9881877899169922, "learning_rate": 6.2028880866426e-05, "loss": 1.5365, "step": 10518 }, { "epoch": 7.5985559566787, "grad_norm": 1.9813497066497803, "learning_rate": 6.20072202166065e-05, "loss": 1.4599, "step": 10524 }, { "epoch": 7.602888086642599, "grad_norm": 1.6117018461227417, "learning_rate": 6.198555956678701e-05, "loss": 1.5803, "step": 10530 }, { "epoch": 7.607220216606498, "grad_norm": 2.2445154190063477, "learning_rate": 6.196389891696752e-05, "loss": 1.6351, "step": 10536 }, { "epoch": 7.611552346570397, "grad_norm": 1.7345025539398193, "learning_rate": 6.194223826714802e-05, "loss": 1.6364, "step": 10542 }, { "epoch": 7.615884476534296, "grad_norm": 2.0090601444244385, "learning_rate": 6.192057761732852e-05, "loss": 1.6132, "step": 10548 }, { "epoch": 7.620216606498195, "grad_norm": 2.008115768432617, "learning_rate": 6.189891696750903e-05, "loss": 1.6983, "step": 10554 }, { "epoch": 7.624548736462094, "grad_norm": 2.1193201541900635, "learning_rate": 6.187725631768953e-05, "loss": 1.6762, "step": 10560 }, { "epoch": 7.628880866425993, "grad_norm": 2.232187509536743, "learning_rate": 6.185559566787003e-05, "loss": 1.664, "step": 10566 }, { "epoch": 7.633212996389892, "grad_norm": 1.9040619134902954, "learning_rate": 6.183393501805055e-05, "loss": 1.6146, "step": 10572 }, { "epoch": 7.637545126353791, "grad_norm": 1.9088313579559326, "learning_rate": 6.181227436823105e-05, "loss": 1.6437, "step": 10578 }, { "epoch": 7.64187725631769, "grad_norm": 2.119981527328491, "learning_rate": 6.179061371841156e-05, "loss": 1.6708, "step": 10584 }, { "epoch": 7.646209386281589, "grad_norm": 2.0685009956359863, "learning_rate": 6.176895306859206e-05, "loss": 1.6098, "step": 10590 }, { "epoch": 7.650541516245488, "grad_norm": 2.3060097694396973, "learning_rate": 6.174729241877256e-05, "loss": 1.6379, "step": 10596 }, { "epoch": 7.654873646209387, "grad_norm": 1.7250210046768188, "learning_rate": 6.172563176895308e-05, "loss": 1.6474, "step": 10602 }, { "epoch": 7.659205776173286, "grad_norm": 2.2664644718170166, "learning_rate": 6.170397111913357e-05, "loss": 1.6054, "step": 10608 }, { "epoch": 7.663537906137184, "grad_norm": 2.0107505321502686, "learning_rate": 6.168231046931409e-05, "loss": 1.598, "step": 10614 }, { "epoch": 7.667870036101083, "grad_norm": 1.782798409461975, "learning_rate": 6.166064981949459e-05, "loss": 1.6317, "step": 10620 }, { "epoch": 7.672202166064982, "grad_norm": 1.9036554098129272, "learning_rate": 6.163898916967509e-05, "loss": 1.5942, "step": 10626 }, { "epoch": 7.676534296028881, "grad_norm": 1.8150752782821655, "learning_rate": 6.16173285198556e-05, "loss": 1.6959, "step": 10632 }, { "epoch": 7.68086642599278, "grad_norm": 1.8104013204574585, "learning_rate": 6.15956678700361e-05, "loss": 1.4215, "step": 10638 }, { "epoch": 7.685198555956679, "grad_norm": 1.981524109840393, "learning_rate": 6.157400722021662e-05, "loss": 1.6146, "step": 10644 }, { "epoch": 7.689530685920578, "grad_norm": 2.139376640319824, "learning_rate": 6.155234657039712e-05, "loss": 1.5501, "step": 10650 }, { "epoch": 7.693862815884477, "grad_norm": 2.0124704837799072, "learning_rate": 6.153068592057762e-05, "loss": 1.5574, "step": 10656 }, { "epoch": 7.698194945848376, "grad_norm": 1.8820339441299438, "learning_rate": 6.150902527075813e-05, "loss": 1.6346, "step": 10662 }, { "epoch": 7.702527075812275, "grad_norm": 1.7963639497756958, "learning_rate": 6.148736462093863e-05, "loss": 1.7213, "step": 10668 }, { "epoch": 7.706859205776174, "grad_norm": 2.1451642513275146, "learning_rate": 6.146570397111913e-05, "loss": 1.5143, "step": 10674 }, { "epoch": 7.7111913357400725, "grad_norm": 1.8841915130615234, "learning_rate": 6.144404332129964e-05, "loss": 1.6922, "step": 10680 }, { "epoch": 7.715523465703971, "grad_norm": 2.1581995487213135, "learning_rate": 6.142238267148015e-05, "loss": 1.6259, "step": 10686 }, { "epoch": 7.71985559566787, "grad_norm": 2.1226205825805664, "learning_rate": 6.140072202166066e-05, "loss": 1.6109, "step": 10692 }, { "epoch": 7.724187725631769, "grad_norm": 2.1035304069519043, "learning_rate": 6.137906137184116e-05, "loss": 1.5709, "step": 10698 }, { "epoch": 7.728519855595668, "grad_norm": 2.276719093322754, "learning_rate": 6.135740072202166e-05, "loss": 1.5757, "step": 10704 }, { "epoch": 7.7328519855595665, "grad_norm": 2.1249918937683105, "learning_rate": 6.133574007220216e-05, "loss": 1.6298, "step": 10710 }, { "epoch": 7.7371841155234655, "grad_norm": 1.8760122060775757, "learning_rate": 6.131407942238268e-05, "loss": 1.5457, "step": 10716 }, { "epoch": 7.7415162454873645, "grad_norm": 1.8666199445724487, "learning_rate": 6.129241877256317e-05, "loss": 1.6453, "step": 10722 }, { "epoch": 7.7458483754512635, "grad_norm": 2.003375291824341, "learning_rate": 6.127075812274369e-05, "loss": 1.6251, "step": 10728 }, { "epoch": 7.7501805054151625, "grad_norm": 1.869892954826355, "learning_rate": 6.124909747292419e-05, "loss": 1.5698, "step": 10734 }, { "epoch": 7.754512635379061, "grad_norm": 1.829566478729248, "learning_rate": 6.12274368231047e-05, "loss": 1.484, "step": 10740 }, { "epoch": 7.75884476534296, "grad_norm": 2.0275700092315674, "learning_rate": 6.12057761732852e-05, "loss": 1.699, "step": 10746 }, { "epoch": 7.763176895306859, "grad_norm": 2.1538848876953125, "learning_rate": 6.11841155234657e-05, "loss": 1.6091, "step": 10752 }, { "epoch": 7.767509025270758, "grad_norm": 2.0093483924865723, "learning_rate": 6.116245487364622e-05, "loss": 1.615, "step": 10758 }, { "epoch": 7.771841155234657, "grad_norm": 2.0728471279144287, "learning_rate": 6.114079422382671e-05, "loss": 1.5566, "step": 10764 }, { "epoch": 7.776173285198556, "grad_norm": 2.1147305965423584, "learning_rate": 6.111913357400722e-05, "loss": 1.7157, "step": 10770 }, { "epoch": 7.780505415162455, "grad_norm": 2.472635507583618, "learning_rate": 6.109747292418773e-05, "loss": 1.6391, "step": 10776 }, { "epoch": 7.784837545126354, "grad_norm": 2.1076276302337646, "learning_rate": 6.107581227436824e-05, "loss": 1.5417, "step": 10782 }, { "epoch": 7.789169675090252, "grad_norm": 2.0170931816101074, "learning_rate": 6.105415162454873e-05, "loss": 1.6105, "step": 10788 }, { "epoch": 7.793501805054151, "grad_norm": 2.109231472015381, "learning_rate": 6.103249097472924e-05, "loss": 1.6668, "step": 10794 }, { "epoch": 7.79783393501805, "grad_norm": 1.8297489881515503, "learning_rate": 6.1010830324909754e-05, "loss": 1.5667, "step": 10800 }, { "epoch": 7.802166064981949, "grad_norm": 1.9658536911010742, "learning_rate": 6.098916967509025e-05, "loss": 1.5554, "step": 10806 }, { "epoch": 7.806498194945848, "grad_norm": 2.0890586376190186, "learning_rate": 6.096750902527076e-05, "loss": 1.6113, "step": 10812 }, { "epoch": 7.810830324909747, "grad_norm": 1.903466820716858, "learning_rate": 6.0945848375451264e-05, "loss": 1.5607, "step": 10818 }, { "epoch": 7.815162454873646, "grad_norm": 2.3064017295837402, "learning_rate": 6.0924187725631774e-05, "loss": 1.7176, "step": 10824 }, { "epoch": 7.819494584837545, "grad_norm": 2.0083649158477783, "learning_rate": 6.090252707581228e-05, "loss": 1.7302, "step": 10830 }, { "epoch": 7.823826714801444, "grad_norm": 2.1693575382232666, "learning_rate": 6.088086642599278e-05, "loss": 1.6031, "step": 10836 }, { "epoch": 7.828158844765343, "grad_norm": 1.9792218208312988, "learning_rate": 6.085920577617329e-05, "loss": 1.5552, "step": 10842 }, { "epoch": 7.832490974729242, "grad_norm": 1.992863416671753, "learning_rate": 6.083754512635379e-05, "loss": 1.5557, "step": 10848 }, { "epoch": 7.836823104693141, "grad_norm": 1.84101402759552, "learning_rate": 6.08158844765343e-05, "loss": 1.5051, "step": 10854 }, { "epoch": 7.841155234657039, "grad_norm": 1.678145170211792, "learning_rate": 6.07942238267148e-05, "loss": 1.5521, "step": 10860 }, { "epoch": 7.845487364620938, "grad_norm": 2.0923948287963867, "learning_rate": 6.0772563176895316e-05, "loss": 1.6662, "step": 10866 }, { "epoch": 7.849819494584837, "grad_norm": 1.6822552680969238, "learning_rate": 6.075090252707581e-05, "loss": 1.6011, "step": 10872 }, { "epoch": 7.854151624548736, "grad_norm": 1.9894970655441284, "learning_rate": 6.0729241877256316e-05, "loss": 1.572, "step": 10878 }, { "epoch": 7.858483754512635, "grad_norm": 1.701906681060791, "learning_rate": 6.0707581227436826e-05, "loss": 1.5007, "step": 10884 }, { "epoch": 7.862815884476534, "grad_norm": 2.056612014770508, "learning_rate": 6.068592057761733e-05, "loss": 1.7267, "step": 10890 }, { "epoch": 7.867148014440433, "grad_norm": 1.9778732061386108, "learning_rate": 6.066425992779784e-05, "loss": 1.7176, "step": 10896 }, { "epoch": 7.871480144404332, "grad_norm": 2.1813700199127197, "learning_rate": 6.064259927797834e-05, "loss": 1.6213, "step": 10902 }, { "epoch": 7.875812274368231, "grad_norm": 1.7770721912384033, "learning_rate": 6.062093862815885e-05, "loss": 1.538, "step": 10908 }, { "epoch": 7.88014440433213, "grad_norm": 2.1265599727630615, "learning_rate": 6.0599277978339356e-05, "loss": 1.6709, "step": 10914 }, { "epoch": 7.884476534296029, "grad_norm": 1.8953309059143066, "learning_rate": 6.057761732851985e-05, "loss": 1.6015, "step": 10920 }, { "epoch": 7.888808664259928, "grad_norm": 1.8109713792800903, "learning_rate": 6.055595667870036e-05, "loss": 1.5837, "step": 10926 }, { "epoch": 7.893140794223827, "grad_norm": 1.8425462245941162, "learning_rate": 6.0534296028880865e-05, "loss": 1.6699, "step": 10932 }, { "epoch": 7.897472924187726, "grad_norm": 1.9291131496429443, "learning_rate": 6.0512635379061375e-05, "loss": 1.6056, "step": 10938 }, { "epoch": 7.901805054151625, "grad_norm": 1.855167269706726, "learning_rate": 6.049097472924188e-05, "loss": 1.5875, "step": 10944 }, { "epoch": 7.906137184115524, "grad_norm": 2.134678840637207, "learning_rate": 6.046931407942239e-05, "loss": 1.7164, "step": 10950 }, { "epoch": 7.910469314079422, "grad_norm": 1.8993396759033203, "learning_rate": 6.044765342960289e-05, "loss": 1.4896, "step": 10956 }, { "epoch": 7.914801444043321, "grad_norm": 2.0269241333007812, "learning_rate": 6.042599277978339e-05, "loss": 1.6412, "step": 10962 }, { "epoch": 7.91913357400722, "grad_norm": 1.9453972578048706, "learning_rate": 6.0404332129963905e-05, "loss": 1.6129, "step": 10968 }, { "epoch": 7.923465703971119, "grad_norm": 2.081911325454712, "learning_rate": 6.03826714801444e-05, "loss": 1.5552, "step": 10974 }, { "epoch": 7.927797833935018, "grad_norm": 1.952131748199463, "learning_rate": 6.036101083032492e-05, "loss": 1.5776, "step": 10980 }, { "epoch": 7.932129963898917, "grad_norm": 2.283252716064453, "learning_rate": 6.0339350180505415e-05, "loss": 1.6545, "step": 10986 }, { "epoch": 7.936462093862816, "grad_norm": 2.098613977432251, "learning_rate": 6.0317689530685925e-05, "loss": 1.8481, "step": 10992 }, { "epoch": 7.940794223826715, "grad_norm": 2.073064088821411, "learning_rate": 6.029602888086643e-05, "loss": 1.6761, "step": 10998 }, { "epoch": 7.945126353790614, "grad_norm": 1.869102954864502, "learning_rate": 6.027436823104693e-05, "loss": 1.4836, "step": 11004 }, { "epoch": 7.949458483754513, "grad_norm": 1.9034520387649536, "learning_rate": 6.025270758122744e-05, "loss": 1.5849, "step": 11010 }, { "epoch": 7.953790613718412, "grad_norm": 1.8845574855804443, "learning_rate": 6.0231046931407945e-05, "loss": 1.6196, "step": 11016 }, { "epoch": 7.958122743682311, "grad_norm": 1.8274184465408325, "learning_rate": 6.0209386281588455e-05, "loss": 1.5683, "step": 11022 }, { "epoch": 7.96245487364621, "grad_norm": 1.8724125623703003, "learning_rate": 6.018772563176895e-05, "loss": 1.7753, "step": 11028 }, { "epoch": 7.966787003610108, "grad_norm": 2.194159507751465, "learning_rate": 6.016606498194947e-05, "loss": 1.6556, "step": 11034 }, { "epoch": 7.971119133574007, "grad_norm": 2.0170297622680664, "learning_rate": 6.0144404332129964e-05, "loss": 1.6219, "step": 11040 }, { "epoch": 7.975451263537906, "grad_norm": 1.8317772150039673, "learning_rate": 6.012274368231047e-05, "loss": 1.5428, "step": 11046 }, { "epoch": 7.979783393501805, "grad_norm": 2.0846071243286133, "learning_rate": 6.010108303249098e-05, "loss": 1.5823, "step": 11052 }, { "epoch": 7.984115523465704, "grad_norm": 2.0731499195098877, "learning_rate": 6.007942238267148e-05, "loss": 1.6037, "step": 11058 }, { "epoch": 7.988447653429603, "grad_norm": 1.797060251235962, "learning_rate": 6.005776173285199e-05, "loss": 1.6691, "step": 11064 }, { "epoch": 7.992779783393502, "grad_norm": 1.9171465635299683, "learning_rate": 6.0036101083032494e-05, "loss": 1.5229, "step": 11070 }, { "epoch": 7.997111913357401, "grad_norm": 1.9261317253112793, "learning_rate": 6.0014440433213004e-05, "loss": 1.6413, "step": 11076 }, { "epoch": 8.0014440433213, "grad_norm": 1.8272626399993896, "learning_rate": 5.999277978339351e-05, "loss": 1.6679, "step": 11082 }, { "epoch": 8.005776173285199, "grad_norm": 2.0036661624908447, "learning_rate": 5.9971119133574004e-05, "loss": 1.6102, "step": 11088 }, { "epoch": 8.010108303249098, "grad_norm": 2.122319221496582, "learning_rate": 5.994945848375452e-05, "loss": 1.5124, "step": 11094 }, { "epoch": 8.014440433212997, "grad_norm": 2.1506876945495605, "learning_rate": 5.992779783393502e-05, "loss": 1.611, "step": 11100 }, { "epoch": 8.018772563176896, "grad_norm": 1.8982925415039062, "learning_rate": 5.990613718411553e-05, "loss": 1.5408, "step": 11106 }, { "epoch": 8.023104693140795, "grad_norm": 2.2911598682403564, "learning_rate": 5.988447653429603e-05, "loss": 1.6883, "step": 11112 }, { "epoch": 8.027436823104694, "grad_norm": 2.1696829795837402, "learning_rate": 5.986281588447654e-05, "loss": 1.641, "step": 11118 }, { "epoch": 8.031768953068593, "grad_norm": 2.224719762802124, "learning_rate": 5.984115523465704e-05, "loss": 1.5022, "step": 11124 }, { "epoch": 8.036101083032491, "grad_norm": 1.8557778596878052, "learning_rate": 5.9819494584837546e-05, "loss": 1.5154, "step": 11130 }, { "epoch": 8.04043321299639, "grad_norm": 1.937695860862732, "learning_rate": 5.9797833935018056e-05, "loss": 1.5522, "step": 11136 }, { "epoch": 8.04476534296029, "grad_norm": 2.061955213546753, "learning_rate": 5.977617328519855e-05, "loss": 1.5278, "step": 11142 }, { "epoch": 8.049097472924188, "grad_norm": 1.9772926568984985, "learning_rate": 5.975451263537907e-05, "loss": 1.5052, "step": 11148 }, { "epoch": 8.053429602888087, "grad_norm": 2.0794177055358887, "learning_rate": 5.9732851985559566e-05, "loss": 1.517, "step": 11154 }, { "epoch": 8.057761732851986, "grad_norm": 2.2162559032440186, "learning_rate": 5.971119133574008e-05, "loss": 1.5846, "step": 11160 }, { "epoch": 8.062093862815885, "grad_norm": 1.8168580532073975, "learning_rate": 5.968953068592058e-05, "loss": 1.5282, "step": 11166 }, { "epoch": 8.066425992779783, "grad_norm": 2.0684266090393066, "learning_rate": 5.966787003610108e-05, "loss": 1.4407, "step": 11172 }, { "epoch": 8.070758122743682, "grad_norm": 1.940142035484314, "learning_rate": 5.964620938628159e-05, "loss": 1.5806, "step": 11178 }, { "epoch": 8.07509025270758, "grad_norm": 2.1622023582458496, "learning_rate": 5.9624548736462096e-05, "loss": 1.5582, "step": 11184 }, { "epoch": 8.07942238267148, "grad_norm": 1.8186097145080566, "learning_rate": 5.9602888086642606e-05, "loss": 1.4254, "step": 11190 }, { "epoch": 8.083754512635378, "grad_norm": 1.9948668479919434, "learning_rate": 5.958122743682311e-05, "loss": 1.4751, "step": 11196 }, { "epoch": 8.088086642599277, "grad_norm": 2.1061763763427734, "learning_rate": 5.955956678700362e-05, "loss": 1.4555, "step": 11202 }, { "epoch": 8.092418772563176, "grad_norm": 2.0797479152679443, "learning_rate": 5.9537906137184115e-05, "loss": 1.6217, "step": 11208 }, { "epoch": 8.096750902527075, "grad_norm": 2.029334545135498, "learning_rate": 5.951624548736462e-05, "loss": 1.6108, "step": 11214 }, { "epoch": 8.101083032490974, "grad_norm": 1.969943642616272, "learning_rate": 5.949458483754513e-05, "loss": 1.5756, "step": 11220 }, { "epoch": 8.105415162454873, "grad_norm": 1.9565093517303467, "learning_rate": 5.947292418772563e-05, "loss": 1.5602, "step": 11226 }, { "epoch": 8.109747292418772, "grad_norm": 2.0324833393096924, "learning_rate": 5.945126353790614e-05, "loss": 1.6337, "step": 11232 }, { "epoch": 8.114079422382671, "grad_norm": 2.0298426151275635, "learning_rate": 5.9429602888086645e-05, "loss": 1.6121, "step": 11238 }, { "epoch": 8.11841155234657, "grad_norm": 2.2827260494232178, "learning_rate": 5.9407942238267155e-05, "loss": 1.4887, "step": 11244 }, { "epoch": 8.12274368231047, "grad_norm": 2.066521644592285, "learning_rate": 5.938628158844766e-05, "loss": 1.5625, "step": 11250 }, { "epoch": 8.127075812274368, "grad_norm": 1.8890362977981567, "learning_rate": 5.9364620938628155e-05, "loss": 1.6109, "step": 11256 }, { "epoch": 8.131407942238267, "grad_norm": 2.3060524463653564, "learning_rate": 5.934296028880867e-05, "loss": 1.6421, "step": 11262 }, { "epoch": 8.135740072202166, "grad_norm": 2.2063376903533936, "learning_rate": 5.932129963898917e-05, "loss": 1.5249, "step": 11268 }, { "epoch": 8.140072202166065, "grad_norm": 2.159181594848633, "learning_rate": 5.929963898916968e-05, "loss": 1.6417, "step": 11274 }, { "epoch": 8.144404332129964, "grad_norm": 2.1451761722564697, "learning_rate": 5.927797833935018e-05, "loss": 1.5534, "step": 11280 }, { "epoch": 8.148736462093863, "grad_norm": 2.065324544906616, "learning_rate": 5.925631768953069e-05, "loss": 1.613, "step": 11286 }, { "epoch": 8.153068592057762, "grad_norm": 2.0713107585906982, "learning_rate": 5.9234657039711195e-05, "loss": 1.5501, "step": 11292 }, { "epoch": 8.157400722021661, "grad_norm": 1.9833269119262695, "learning_rate": 5.92129963898917e-05, "loss": 1.5981, "step": 11298 }, { "epoch": 8.16173285198556, "grad_norm": 1.949151635169983, "learning_rate": 5.919133574007221e-05, "loss": 1.6369, "step": 11304 }, { "epoch": 8.166064981949459, "grad_norm": 1.8646470308303833, "learning_rate": 5.9169675090252704e-05, "loss": 1.4796, "step": 11310 }, { "epoch": 8.170397111913358, "grad_norm": 2.150998830795288, "learning_rate": 5.914801444043322e-05, "loss": 1.5343, "step": 11316 }, { "epoch": 8.174729241877257, "grad_norm": 2.0839314460754395, "learning_rate": 5.912635379061372e-05, "loss": 1.517, "step": 11322 }, { "epoch": 8.179061371841156, "grad_norm": 2.08762264251709, "learning_rate": 5.9104693140794234e-05, "loss": 1.5851, "step": 11328 }, { "epoch": 8.183393501805055, "grad_norm": 2.095771551132202, "learning_rate": 5.908303249097473e-05, "loss": 1.5717, "step": 11334 }, { "epoch": 8.187725631768952, "grad_norm": 1.9823826551437378, "learning_rate": 5.9061371841155234e-05, "loss": 1.6017, "step": 11340 }, { "epoch": 8.192057761732851, "grad_norm": 2.0679867267608643, "learning_rate": 5.9039711191335744e-05, "loss": 1.6253, "step": 11346 }, { "epoch": 8.19638989169675, "grad_norm": 1.9685328006744385, "learning_rate": 5.901805054151625e-05, "loss": 1.5276, "step": 11352 }, { "epoch": 8.20072202166065, "grad_norm": 2.174042224884033, "learning_rate": 5.899638989169676e-05, "loss": 1.4446, "step": 11358 }, { "epoch": 8.205054151624548, "grad_norm": 2.1932425498962402, "learning_rate": 5.897472924187726e-05, "loss": 1.6214, "step": 11364 }, { "epoch": 8.209386281588447, "grad_norm": 2.4115912914276123, "learning_rate": 5.895306859205777e-05, "loss": 1.6955, "step": 11370 }, { "epoch": 8.213718411552346, "grad_norm": 2.0577683448791504, "learning_rate": 5.8931407942238274e-05, "loss": 1.4657, "step": 11376 }, { "epoch": 8.218050541516245, "grad_norm": 2.1235978603363037, "learning_rate": 5.890974729241877e-05, "loss": 1.5695, "step": 11382 }, { "epoch": 8.222382671480144, "grad_norm": 2.1620993614196777, "learning_rate": 5.888808664259928e-05, "loss": 1.6755, "step": 11388 }, { "epoch": 8.226714801444043, "grad_norm": 1.9189422130584717, "learning_rate": 5.886642599277978e-05, "loss": 1.5757, "step": 11394 }, { "epoch": 8.231046931407942, "grad_norm": 1.9492719173431396, "learning_rate": 5.884476534296029e-05, "loss": 1.5506, "step": 11400 }, { "epoch": 8.235379061371841, "grad_norm": 1.9618265628814697, "learning_rate": 5.8823104693140796e-05, "loss": 1.6278, "step": 11406 }, { "epoch": 8.23971119133574, "grad_norm": 1.954938530921936, "learning_rate": 5.8801444043321306e-05, "loss": 1.5866, "step": 11412 }, { "epoch": 8.244043321299639, "grad_norm": 2.088529586791992, "learning_rate": 5.877978339350181e-05, "loss": 1.6109, "step": 11418 }, { "epoch": 8.248375451263538, "grad_norm": 2.0383214950561523, "learning_rate": 5.8758122743682306e-05, "loss": 1.5336, "step": 11424 }, { "epoch": 8.252707581227437, "grad_norm": 2.119114875793457, "learning_rate": 5.873646209386282e-05, "loss": 1.572, "step": 11430 }, { "epoch": 8.257039711191336, "grad_norm": 2.094831943511963, "learning_rate": 5.871480144404332e-05, "loss": 1.5194, "step": 11436 }, { "epoch": 8.261371841155235, "grad_norm": 2.1534481048583984, "learning_rate": 5.8693140794223836e-05, "loss": 1.477, "step": 11442 }, { "epoch": 8.265703971119134, "grad_norm": 2.0001220703125, "learning_rate": 5.867148014440433e-05, "loss": 1.5057, "step": 11448 }, { "epoch": 8.270036101083033, "grad_norm": 2.1555051803588867, "learning_rate": 5.864981949458484e-05, "loss": 1.5392, "step": 11454 }, { "epoch": 8.274368231046932, "grad_norm": 2.257235288619995, "learning_rate": 5.8628158844765346e-05, "loss": 1.6864, "step": 11460 }, { "epoch": 8.27870036101083, "grad_norm": 1.798761010169983, "learning_rate": 5.860649819494585e-05, "loss": 1.5412, "step": 11466 }, { "epoch": 8.28303249097473, "grad_norm": 1.9732658863067627, "learning_rate": 5.858483754512636e-05, "loss": 1.5023, "step": 11472 }, { "epoch": 8.287364620938629, "grad_norm": 1.733834147453308, "learning_rate": 5.856317689530686e-05, "loss": 1.5444, "step": 11478 }, { "epoch": 8.291696750902528, "grad_norm": 2.1837189197540283, "learning_rate": 5.854151624548737e-05, "loss": 1.6302, "step": 11484 }, { "epoch": 8.296028880866427, "grad_norm": 1.9477901458740234, "learning_rate": 5.851985559566787e-05, "loss": 1.6041, "step": 11490 }, { "epoch": 8.300361010830326, "grad_norm": 2.217151165008545, "learning_rate": 5.8498194945848386e-05, "loss": 1.5583, "step": 11496 }, { "epoch": 8.304693140794225, "grad_norm": 1.6501661539077759, "learning_rate": 5.847653429602888e-05, "loss": 1.4924, "step": 11502 }, { "epoch": 8.309025270758124, "grad_norm": 2.0548157691955566, "learning_rate": 5.8454873646209385e-05, "loss": 1.6299, "step": 11508 }, { "epoch": 8.313357400722023, "grad_norm": 1.902255654335022, "learning_rate": 5.8433212996389895e-05, "loss": 1.592, "step": 11514 }, { "epoch": 8.31768953068592, "grad_norm": 2.324254035949707, "learning_rate": 5.84115523465704e-05, "loss": 1.5092, "step": 11520 }, { "epoch": 8.322021660649819, "grad_norm": 1.706589698791504, "learning_rate": 5.838989169675091e-05, "loss": 1.5052, "step": 11526 }, { "epoch": 8.326353790613718, "grad_norm": 2.038691759109497, "learning_rate": 5.836823104693141e-05, "loss": 1.3964, "step": 11532 }, { "epoch": 8.330685920577617, "grad_norm": 2.135855197906494, "learning_rate": 5.834657039711192e-05, "loss": 1.4959, "step": 11538 }, { "epoch": 8.335018050541516, "grad_norm": 2.21952748298645, "learning_rate": 5.8324909747292425e-05, "loss": 1.5692, "step": 11544 }, { "epoch": 8.339350180505415, "grad_norm": 2.1005823612213135, "learning_rate": 5.830324909747292e-05, "loss": 1.6535, "step": 11550 }, { "epoch": 8.343682310469314, "grad_norm": 2.060222625732422, "learning_rate": 5.828158844765343e-05, "loss": 1.7063, "step": 11556 }, { "epoch": 8.348014440433213, "grad_norm": 2.0887339115142822, "learning_rate": 5.8259927797833935e-05, "loss": 1.6636, "step": 11562 }, { "epoch": 8.352346570397112, "grad_norm": 1.8827074766159058, "learning_rate": 5.8238267148014445e-05, "loss": 1.5626, "step": 11568 }, { "epoch": 8.35667870036101, "grad_norm": 2.00748872756958, "learning_rate": 5.821660649819495e-05, "loss": 1.5048, "step": 11574 }, { "epoch": 8.36101083032491, "grad_norm": 2.3993868827819824, "learning_rate": 5.819494584837546e-05, "loss": 1.5274, "step": 11580 }, { "epoch": 8.365342960288809, "grad_norm": 2.0397188663482666, "learning_rate": 5.817328519855596e-05, "loss": 1.5678, "step": 11586 }, { "epoch": 8.369675090252708, "grad_norm": 2.2466461658477783, "learning_rate": 5.815162454873646e-05, "loss": 1.5559, "step": 11592 }, { "epoch": 8.374007220216606, "grad_norm": 2.221811532974243, "learning_rate": 5.8129963898916974e-05, "loss": 1.605, "step": 11598 }, { "epoch": 8.378339350180505, "grad_norm": 2.067548990249634, "learning_rate": 5.810830324909747e-05, "loss": 1.446, "step": 11604 }, { "epoch": 8.382671480144404, "grad_norm": 1.8565523624420166, "learning_rate": 5.808664259927799e-05, "loss": 1.5368, "step": 11610 }, { "epoch": 8.387003610108303, "grad_norm": 1.8101376295089722, "learning_rate": 5.8064981949458484e-05, "loss": 1.52, "step": 11616 }, { "epoch": 8.391335740072202, "grad_norm": 1.9795160293579102, "learning_rate": 5.8043321299639e-05, "loss": 1.5956, "step": 11622 }, { "epoch": 8.395667870036101, "grad_norm": 1.8990445137023926, "learning_rate": 5.80216606498195e-05, "loss": 1.4701, "step": 11628 }, { "epoch": 8.4, "grad_norm": 1.831107258796692, "learning_rate": 5.8e-05, "loss": 1.6135, "step": 11634 }, { "epoch": 8.4043321299639, "grad_norm": 2.0172243118286133, "learning_rate": 5.797833935018051e-05, "loss": 1.4908, "step": 11640 }, { "epoch": 8.408664259927798, "grad_norm": 1.996693730354309, "learning_rate": 5.7956678700361014e-05, "loss": 1.5407, "step": 11646 }, { "epoch": 8.412996389891697, "grad_norm": 2.1123733520507812, "learning_rate": 5.7935018050541524e-05, "loss": 1.5437, "step": 11652 }, { "epoch": 8.417328519855596, "grad_norm": 1.8570284843444824, "learning_rate": 5.791335740072203e-05, "loss": 1.5701, "step": 11658 }, { "epoch": 8.421660649819495, "grad_norm": 1.9769119024276733, "learning_rate": 5.789169675090252e-05, "loss": 1.6194, "step": 11664 }, { "epoch": 8.425992779783394, "grad_norm": 1.9768295288085938, "learning_rate": 5.787003610108303e-05, "loss": 1.5525, "step": 11670 }, { "epoch": 8.430324909747293, "grad_norm": 1.9675571918487549, "learning_rate": 5.7848375451263536e-05, "loss": 1.6081, "step": 11676 }, { "epoch": 8.434657039711192, "grad_norm": 2.023153066635132, "learning_rate": 5.7826714801444046e-05, "loss": 1.5192, "step": 11682 }, { "epoch": 8.43898916967509, "grad_norm": 1.854264736175537, "learning_rate": 5.780505415162455e-05, "loss": 1.5111, "step": 11688 }, { "epoch": 8.443321299638988, "grad_norm": 2.2083215713500977, "learning_rate": 5.778339350180506e-05, "loss": 1.5639, "step": 11694 }, { "epoch": 8.447653429602887, "grad_norm": 1.8538720607757568, "learning_rate": 5.776173285198556e-05, "loss": 1.5984, "step": 11700 }, { "epoch": 8.451985559566786, "grad_norm": 2.2890172004699707, "learning_rate": 5.774007220216606e-05, "loss": 1.6641, "step": 11706 }, { "epoch": 8.456317689530685, "grad_norm": 1.921978235244751, "learning_rate": 5.7718411552346576e-05, "loss": 1.6243, "step": 11712 }, { "epoch": 8.460649819494584, "grad_norm": 1.9094078540802002, "learning_rate": 5.769675090252707e-05, "loss": 1.5168, "step": 11718 }, { "epoch": 8.464981949458483, "grad_norm": 2.1210663318634033, "learning_rate": 5.767509025270759e-05, "loss": 1.5782, "step": 11724 }, { "epoch": 8.469314079422382, "grad_norm": 1.9536869525909424, "learning_rate": 5.7653429602888086e-05, "loss": 1.535, "step": 11730 }, { "epoch": 8.473646209386281, "grad_norm": 2.162128448486328, "learning_rate": 5.7631768953068596e-05, "loss": 1.5982, "step": 11736 }, { "epoch": 8.47797833935018, "grad_norm": 2.126727342605591, "learning_rate": 5.76101083032491e-05, "loss": 1.6226, "step": 11742 }, { "epoch": 8.48231046931408, "grad_norm": 2.018458604812622, "learning_rate": 5.75884476534296e-05, "loss": 1.5638, "step": 11748 }, { "epoch": 8.486642599277978, "grad_norm": 2.17450213432312, "learning_rate": 5.756678700361011e-05, "loss": 1.6051, "step": 11754 }, { "epoch": 8.490974729241877, "grad_norm": 2.3495187759399414, "learning_rate": 5.7545126353790616e-05, "loss": 1.5442, "step": 11760 }, { "epoch": 8.495306859205776, "grad_norm": 2.095604419708252, "learning_rate": 5.7523465703971126e-05, "loss": 1.4513, "step": 11766 }, { "epoch": 8.499638989169675, "grad_norm": 2.017148017883301, "learning_rate": 5.750180505415162e-05, "loss": 1.5887, "step": 11772 }, { "epoch": 8.503971119133574, "grad_norm": 1.7833319902420044, "learning_rate": 5.748014440433214e-05, "loss": 1.6949, "step": 11778 }, { "epoch": 8.508303249097473, "grad_norm": 2.059739828109741, "learning_rate": 5.7458483754512635e-05, "loss": 1.6202, "step": 11784 }, { "epoch": 8.512635379061372, "grad_norm": 2.1304452419281006, "learning_rate": 5.743682310469314e-05, "loss": 1.6174, "step": 11790 }, { "epoch": 8.516967509025271, "grad_norm": 2.0080196857452393, "learning_rate": 5.741516245487365e-05, "loss": 1.5667, "step": 11796 }, { "epoch": 8.52129963898917, "grad_norm": 2.1829562187194824, "learning_rate": 5.739350180505415e-05, "loss": 1.6181, "step": 11802 }, { "epoch": 8.525631768953069, "grad_norm": 2.1255247592926025, "learning_rate": 5.737184115523466e-05, "loss": 1.5236, "step": 11808 }, { "epoch": 8.529963898916968, "grad_norm": 2.0499765872955322, "learning_rate": 5.7350180505415165e-05, "loss": 1.5657, "step": 11814 }, { "epoch": 8.534296028880867, "grad_norm": 2.1238856315612793, "learning_rate": 5.7328519855595675e-05, "loss": 1.4999, "step": 11820 }, { "epoch": 8.538628158844766, "grad_norm": 2.0005481243133545, "learning_rate": 5.730685920577618e-05, "loss": 1.5441, "step": 11826 }, { "epoch": 8.542960288808665, "grad_norm": 2.1007657051086426, "learning_rate": 5.7285198555956675e-05, "loss": 1.584, "step": 11832 }, { "epoch": 8.547292418772564, "grad_norm": 1.8820791244506836, "learning_rate": 5.726353790613719e-05, "loss": 1.4961, "step": 11838 }, { "epoch": 8.551624548736463, "grad_norm": 2.0996758937835693, "learning_rate": 5.724187725631769e-05, "loss": 1.6876, "step": 11844 }, { "epoch": 8.555956678700362, "grad_norm": 2.4986958503723145, "learning_rate": 5.72202166064982e-05, "loss": 1.5418, "step": 11850 }, { "epoch": 8.56028880866426, "grad_norm": 2.170224666595459, "learning_rate": 5.71985559566787e-05, "loss": 1.6699, "step": 11856 }, { "epoch": 8.56462093862816, "grad_norm": 2.2982773780822754, "learning_rate": 5.717689530685921e-05, "loss": 1.5557, "step": 11862 }, { "epoch": 8.568953068592057, "grad_norm": 2.0949974060058594, "learning_rate": 5.7155234657039714e-05, "loss": 1.5083, "step": 11868 }, { "epoch": 8.573285198555956, "grad_norm": 2.113118886947632, "learning_rate": 5.713357400722022e-05, "loss": 1.4918, "step": 11874 }, { "epoch": 8.577617328519855, "grad_norm": 2.160568952560425, "learning_rate": 5.711191335740073e-05, "loss": 1.5602, "step": 11880 }, { "epoch": 8.581949458483754, "grad_norm": 2.210313558578491, "learning_rate": 5.7090252707581224e-05, "loss": 1.6646, "step": 11886 }, { "epoch": 8.586281588447653, "grad_norm": 2.11845064163208, "learning_rate": 5.706859205776174e-05, "loss": 1.5534, "step": 11892 }, { "epoch": 8.590613718411552, "grad_norm": 1.9661785364151, "learning_rate": 5.704693140794224e-05, "loss": 1.6078, "step": 11898 }, { "epoch": 8.59494584837545, "grad_norm": 2.1539909839630127, "learning_rate": 5.7025270758122754e-05, "loss": 1.5653, "step": 11904 }, { "epoch": 8.59927797833935, "grad_norm": 2.0429370403289795, "learning_rate": 5.700361010830325e-05, "loss": 1.595, "step": 11910 }, { "epoch": 8.603610108303249, "grad_norm": 1.8456283807754517, "learning_rate": 5.6981949458483754e-05, "loss": 1.6057, "step": 11916 }, { "epoch": 8.607942238267148, "grad_norm": 2.176992416381836, "learning_rate": 5.6960288808664264e-05, "loss": 1.672, "step": 11922 }, { "epoch": 8.612274368231047, "grad_norm": 1.8578511476516724, "learning_rate": 5.693862815884477e-05, "loss": 1.5512, "step": 11928 }, { "epoch": 8.616606498194946, "grad_norm": 2.263883113861084, "learning_rate": 5.691696750902528e-05, "loss": 1.4793, "step": 11934 }, { "epoch": 8.620938628158845, "grad_norm": 2.0699145793914795, "learning_rate": 5.689530685920578e-05, "loss": 1.5853, "step": 11940 }, { "epoch": 8.625270758122744, "grad_norm": 2.0254311561584473, "learning_rate": 5.687364620938629e-05, "loss": 1.6316, "step": 11946 }, { "epoch": 8.629602888086643, "grad_norm": 2.233401298522949, "learning_rate": 5.6851985559566786e-05, "loss": 1.4985, "step": 11952 }, { "epoch": 8.633935018050542, "grad_norm": 2.0759716033935547, "learning_rate": 5.683032490974729e-05, "loss": 1.5214, "step": 11958 }, { "epoch": 8.63826714801444, "grad_norm": 2.0121567249298096, "learning_rate": 5.68086642599278e-05, "loss": 1.545, "step": 11964 }, { "epoch": 8.64259927797834, "grad_norm": 2.0106287002563477, "learning_rate": 5.67870036101083e-05, "loss": 1.4955, "step": 11970 }, { "epoch": 8.646931407942239, "grad_norm": 2.1340906620025635, "learning_rate": 5.676534296028881e-05, "loss": 1.5953, "step": 11976 }, { "epoch": 8.651263537906138, "grad_norm": 1.9547375440597534, "learning_rate": 5.6743682310469316e-05, "loss": 1.6638, "step": 11982 }, { "epoch": 8.655595667870037, "grad_norm": 2.1421329975128174, "learning_rate": 5.6722021660649826e-05, "loss": 1.6057, "step": 11988 }, { "epoch": 8.659927797833936, "grad_norm": 2.1595137119293213, "learning_rate": 5.670036101083033e-05, "loss": 1.6692, "step": 11994 }, { "epoch": 8.664259927797834, "grad_norm": 1.9696604013442993, "learning_rate": 5.6678700361010826e-05, "loss": 1.5232, "step": 12000 }, { "epoch": 8.668592057761733, "grad_norm": 1.9718834161758423, "learning_rate": 5.665703971119134e-05, "loss": 1.5327, "step": 12006 }, { "epoch": 8.672924187725632, "grad_norm": 2.196185827255249, "learning_rate": 5.663537906137184e-05, "loss": 1.6018, "step": 12012 }, { "epoch": 8.677256317689531, "grad_norm": 2.04217267036438, "learning_rate": 5.661371841155235e-05, "loss": 1.4912, "step": 12018 }, { "epoch": 8.68158844765343, "grad_norm": 2.301579713821411, "learning_rate": 5.659205776173285e-05, "loss": 1.5519, "step": 12024 }, { "epoch": 8.685920577617328, "grad_norm": 2.0910651683807373, "learning_rate": 5.657039711191336e-05, "loss": 1.6468, "step": 12030 }, { "epoch": 8.690252707581227, "grad_norm": 2.062088966369629, "learning_rate": 5.6548736462093866e-05, "loss": 1.6402, "step": 12036 }, { "epoch": 8.694584837545126, "grad_norm": 2.048952102661133, "learning_rate": 5.652707581227437e-05, "loss": 1.4277, "step": 12042 }, { "epoch": 8.698916967509025, "grad_norm": 2.1816112995147705, "learning_rate": 5.650541516245488e-05, "loss": 1.5632, "step": 12048 }, { "epoch": 8.703249097472924, "grad_norm": 2.0629289150238037, "learning_rate": 5.6483754512635375e-05, "loss": 1.5535, "step": 12054 }, { "epoch": 8.707581227436823, "grad_norm": 2.0252602100372314, "learning_rate": 5.646209386281589e-05, "loss": 1.5853, "step": 12060 }, { "epoch": 8.711913357400721, "grad_norm": 2.2511627674102783, "learning_rate": 5.644043321299639e-05, "loss": 1.7716, "step": 12066 }, { "epoch": 8.71624548736462, "grad_norm": 2.0534067153930664, "learning_rate": 5.6418772563176905e-05, "loss": 1.529, "step": 12072 }, { "epoch": 8.72057761732852, "grad_norm": 2.1222469806671143, "learning_rate": 5.63971119133574e-05, "loss": 1.541, "step": 12078 }, { "epoch": 8.724909747292418, "grad_norm": 2.1659841537475586, "learning_rate": 5.6375451263537905e-05, "loss": 1.5732, "step": 12084 }, { "epoch": 8.729241877256317, "grad_norm": 2.2196736335754395, "learning_rate": 5.6353790613718415e-05, "loss": 1.6208, "step": 12090 }, { "epoch": 8.733574007220216, "grad_norm": 2.0393598079681396, "learning_rate": 5.633212996389892e-05, "loss": 1.5989, "step": 12096 }, { "epoch": 8.737906137184115, "grad_norm": 1.8484115600585938, "learning_rate": 5.631046931407943e-05, "loss": 1.6617, "step": 12102 }, { "epoch": 8.742238267148014, "grad_norm": 2.0516180992126465, "learning_rate": 5.628880866425993e-05, "loss": 1.6474, "step": 12108 }, { "epoch": 8.746570397111913, "grad_norm": 1.9849053621292114, "learning_rate": 5.626714801444044e-05, "loss": 1.7633, "step": 12114 }, { "epoch": 8.750902527075812, "grad_norm": 2.139012575149536, "learning_rate": 5.6245487364620945e-05, "loss": 1.5389, "step": 12120 }, { "epoch": 8.755234657039711, "grad_norm": 2.2067975997924805, "learning_rate": 5.622382671480144e-05, "loss": 1.6243, "step": 12126 }, { "epoch": 8.75956678700361, "grad_norm": 2.0330746173858643, "learning_rate": 5.620216606498195e-05, "loss": 1.5138, "step": 12132 }, { "epoch": 8.76389891696751, "grad_norm": 1.8802597522735596, "learning_rate": 5.6180505415162454e-05, "loss": 1.443, "step": 12138 }, { "epoch": 8.768231046931408, "grad_norm": 2.3848633766174316, "learning_rate": 5.6158844765342964e-05, "loss": 1.6711, "step": 12144 }, { "epoch": 8.772563176895307, "grad_norm": 1.793905258178711, "learning_rate": 5.613718411552347e-05, "loss": 1.5492, "step": 12150 }, { "epoch": 8.776895306859206, "grad_norm": 2.0590438842773438, "learning_rate": 5.611552346570398e-05, "loss": 1.5998, "step": 12156 }, { "epoch": 8.781227436823105, "grad_norm": 2.243889093399048, "learning_rate": 5.609386281588448e-05, "loss": 1.5771, "step": 12162 }, { "epoch": 8.785559566787004, "grad_norm": 1.8767045736312866, "learning_rate": 5.607220216606498e-05, "loss": 1.5903, "step": 12168 }, { "epoch": 8.789891696750903, "grad_norm": 1.8150548934936523, "learning_rate": 5.6050541516245494e-05, "loss": 1.6558, "step": 12174 }, { "epoch": 8.794223826714802, "grad_norm": 2.2269723415374756, "learning_rate": 5.602888086642599e-05, "loss": 1.5827, "step": 12180 }, { "epoch": 8.798555956678701, "grad_norm": 2.0509777069091797, "learning_rate": 5.600722021660651e-05, "loss": 1.5888, "step": 12186 }, { "epoch": 8.8028880866426, "grad_norm": 2.127096176147461, "learning_rate": 5.5985559566787004e-05, "loss": 1.5584, "step": 12192 }, { "epoch": 8.807220216606499, "grad_norm": 2.0060501098632812, "learning_rate": 5.5963898916967514e-05, "loss": 1.6377, "step": 12198 }, { "epoch": 8.811552346570398, "grad_norm": 2.284513473510742, "learning_rate": 5.594223826714802e-05, "loss": 1.5131, "step": 12204 }, { "epoch": 8.815884476534297, "grad_norm": 1.9891473054885864, "learning_rate": 5.592057761732852e-05, "loss": 1.623, "step": 12210 }, { "epoch": 8.820216606498194, "grad_norm": 2.045672655105591, "learning_rate": 5.589891696750903e-05, "loss": 1.6536, "step": 12216 }, { "epoch": 8.824548736462093, "grad_norm": 2.2415339946746826, "learning_rate": 5.587725631768953e-05, "loss": 1.5442, "step": 12222 }, { "epoch": 8.828880866425992, "grad_norm": 1.8184646368026733, "learning_rate": 5.585559566787004e-05, "loss": 1.4681, "step": 12228 }, { "epoch": 8.833212996389891, "grad_norm": 2.0282974243164062, "learning_rate": 5.583393501805054e-05, "loss": 1.6448, "step": 12234 }, { "epoch": 8.83754512635379, "grad_norm": 1.9600356817245483, "learning_rate": 5.5812274368231056e-05, "loss": 1.5986, "step": 12240 }, { "epoch": 8.841877256317689, "grad_norm": 2.2426252365112305, "learning_rate": 5.579061371841155e-05, "loss": 1.6482, "step": 12246 }, { "epoch": 8.846209386281588, "grad_norm": 1.9804201126098633, "learning_rate": 5.5768953068592056e-05, "loss": 1.4968, "step": 12252 }, { "epoch": 8.850541516245487, "grad_norm": 1.9789375066757202, "learning_rate": 5.5747292418772566e-05, "loss": 1.57, "step": 12258 }, { "epoch": 8.854873646209386, "grad_norm": 1.8503862619400024, "learning_rate": 5.572563176895307e-05, "loss": 1.4, "step": 12264 }, { "epoch": 8.859205776173285, "grad_norm": 1.9574944972991943, "learning_rate": 5.570397111913358e-05, "loss": 1.4399, "step": 12270 }, { "epoch": 8.863537906137184, "grad_norm": 2.0485284328460693, "learning_rate": 5.568231046931408e-05, "loss": 1.648, "step": 12276 }, { "epoch": 8.867870036101083, "grad_norm": 1.9036632776260376, "learning_rate": 5.566064981949459e-05, "loss": 1.6258, "step": 12282 }, { "epoch": 8.872202166064982, "grad_norm": 1.934434413909912, "learning_rate": 5.5638989169675096e-05, "loss": 1.4941, "step": 12288 }, { "epoch": 8.876534296028881, "grad_norm": 2.0873844623565674, "learning_rate": 5.561732851985559e-05, "loss": 1.6602, "step": 12294 }, { "epoch": 8.88086642599278, "grad_norm": 2.232616901397705, "learning_rate": 5.55956678700361e-05, "loss": 1.5938, "step": 12300 }, { "epoch": 8.885198555956679, "grad_norm": 2.1324994564056396, "learning_rate": 5.5574007220216606e-05, "loss": 1.5847, "step": 12306 }, { "epoch": 8.889530685920578, "grad_norm": 1.967127799987793, "learning_rate": 5.5552346570397116e-05, "loss": 1.5908, "step": 12312 }, { "epoch": 8.893862815884477, "grad_norm": 1.9688254594802856, "learning_rate": 5.553068592057762e-05, "loss": 1.6439, "step": 12318 }, { "epoch": 8.898194945848376, "grad_norm": 2.1430814266204834, "learning_rate": 5.550902527075813e-05, "loss": 1.5696, "step": 12324 }, { "epoch": 8.902527075812275, "grad_norm": 2.318126678466797, "learning_rate": 5.548736462093863e-05, "loss": 1.4418, "step": 12330 }, { "epoch": 8.906859205776174, "grad_norm": 2.1443912982940674, "learning_rate": 5.546570397111913e-05, "loss": 1.4632, "step": 12336 }, { "epoch": 8.911191335740073, "grad_norm": 1.9208513498306274, "learning_rate": 5.5444043321299645e-05, "loss": 1.5503, "step": 12342 }, { "epoch": 8.915523465703972, "grad_norm": 1.8692078590393066, "learning_rate": 5.542238267148014e-05, "loss": 1.582, "step": 12348 }, { "epoch": 8.91985559566787, "grad_norm": 2.0721914768218994, "learning_rate": 5.540072202166066e-05, "loss": 1.551, "step": 12354 }, { "epoch": 8.92418772563177, "grad_norm": 2.0585639476776123, "learning_rate": 5.5379061371841155e-05, "loss": 1.6428, "step": 12360 }, { "epoch": 8.928519855595669, "grad_norm": 2.1103789806365967, "learning_rate": 5.535740072202167e-05, "loss": 1.5457, "step": 12366 }, { "epoch": 8.932851985559568, "grad_norm": 1.770367980003357, "learning_rate": 5.533574007220217e-05, "loss": 1.6276, "step": 12372 }, { "epoch": 8.937184115523465, "grad_norm": 1.880644679069519, "learning_rate": 5.531407942238267e-05, "loss": 1.6856, "step": 12378 }, { "epoch": 8.941516245487364, "grad_norm": 1.9495809078216553, "learning_rate": 5.529241877256318e-05, "loss": 1.6542, "step": 12384 }, { "epoch": 8.945848375451263, "grad_norm": 1.9409406185150146, "learning_rate": 5.5270758122743685e-05, "loss": 1.5636, "step": 12390 }, { "epoch": 8.950180505415162, "grad_norm": 2.1270132064819336, "learning_rate": 5.5249097472924195e-05, "loss": 1.6509, "step": 12396 }, { "epoch": 8.95451263537906, "grad_norm": 2.21099853515625, "learning_rate": 5.52274368231047e-05, "loss": 1.6899, "step": 12402 }, { "epoch": 8.95884476534296, "grad_norm": 2.196564197540283, "learning_rate": 5.520577617328521e-05, "loss": 1.5537, "step": 12408 }, { "epoch": 8.963176895306859, "grad_norm": 2.01851487159729, "learning_rate": 5.5184115523465704e-05, "loss": 1.5956, "step": 12414 }, { "epoch": 8.967509025270758, "grad_norm": 2.170071840286255, "learning_rate": 5.516245487364621e-05, "loss": 1.5559, "step": 12420 }, { "epoch": 8.971841155234657, "grad_norm": 1.8896069526672363, "learning_rate": 5.514079422382672e-05, "loss": 1.5063, "step": 12426 }, { "epoch": 8.976173285198556, "grad_norm": 1.9678043127059937, "learning_rate": 5.511913357400722e-05, "loss": 1.5667, "step": 12432 }, { "epoch": 8.980505415162455, "grad_norm": 1.9223374128341675, "learning_rate": 5.509747292418773e-05, "loss": 1.6787, "step": 12438 }, { "epoch": 8.984837545126354, "grad_norm": 2.357966899871826, "learning_rate": 5.5075812274368234e-05, "loss": 1.6528, "step": 12444 }, { "epoch": 8.989169675090253, "grad_norm": 2.1825504302978516, "learning_rate": 5.5054151624548744e-05, "loss": 1.6705, "step": 12450 }, { "epoch": 8.993501805054152, "grad_norm": 2.177602767944336, "learning_rate": 5.503249097472925e-05, "loss": 1.746, "step": 12456 }, { "epoch": 8.99783393501805, "grad_norm": 2.0957930088043213, "learning_rate": 5.5010830324909744e-05, "loss": 1.6501, "step": 12462 }, { "epoch": 9.00216606498195, "grad_norm": 1.9863684177398682, "learning_rate": 5.498916967509026e-05, "loss": 1.6171, "step": 12468 }, { "epoch": 9.006498194945848, "grad_norm": 2.1859960556030273, "learning_rate": 5.496750902527076e-05, "loss": 1.5901, "step": 12474 }, { "epoch": 9.010830324909747, "grad_norm": 2.1732120513916016, "learning_rate": 5.494584837545127e-05, "loss": 1.5503, "step": 12480 }, { "epoch": 9.015162454873646, "grad_norm": 2.083592176437378, "learning_rate": 5.492418772563177e-05, "loss": 1.4592, "step": 12486 }, { "epoch": 9.019494584837545, "grad_norm": 2.0516748428344727, "learning_rate": 5.490252707581228e-05, "loss": 1.4909, "step": 12492 }, { "epoch": 9.023826714801444, "grad_norm": 2.310063123703003, "learning_rate": 5.488086642599278e-05, "loss": 1.6034, "step": 12498 }, { "epoch": 9.028158844765343, "grad_norm": 2.1975409984588623, "learning_rate": 5.4859205776173286e-05, "loss": 1.6516, "step": 12504 }, { "epoch": 9.032490974729242, "grad_norm": 2.1255884170532227, "learning_rate": 5.4837545126353796e-05, "loss": 1.5109, "step": 12510 }, { "epoch": 9.036823104693141, "grad_norm": 1.9011567831039429, "learning_rate": 5.481588447653429e-05, "loss": 1.497, "step": 12516 }, { "epoch": 9.04115523465704, "grad_norm": 2.1564526557922363, "learning_rate": 5.479422382671481e-05, "loss": 1.638, "step": 12522 }, { "epoch": 9.04548736462094, "grad_norm": 2.00880765914917, "learning_rate": 5.4772563176895306e-05, "loss": 1.6221, "step": 12528 }, { "epoch": 9.049819494584838, "grad_norm": 2.27095890045166, "learning_rate": 5.475090252707582e-05, "loss": 1.5806, "step": 12534 }, { "epoch": 9.054151624548737, "grad_norm": 1.9507887363433838, "learning_rate": 5.472924187725632e-05, "loss": 1.5285, "step": 12540 }, { "epoch": 9.058483754512636, "grad_norm": 2.3929736614227295, "learning_rate": 5.470758122743682e-05, "loss": 1.6088, "step": 12546 }, { "epoch": 9.062815884476535, "grad_norm": 2.312147617340088, "learning_rate": 5.468592057761733e-05, "loss": 1.4541, "step": 12552 }, { "epoch": 9.067148014440432, "grad_norm": 2.188136100769043, "learning_rate": 5.4664259927797836e-05, "loss": 1.5814, "step": 12558 }, { "epoch": 9.071480144404331, "grad_norm": 2.363081455230713, "learning_rate": 5.4642599277978346e-05, "loss": 1.6123, "step": 12564 }, { "epoch": 9.07581227436823, "grad_norm": 2.4169912338256836, "learning_rate": 5.462093862815885e-05, "loss": 1.5758, "step": 12570 }, { "epoch": 9.08014440433213, "grad_norm": 2.036259651184082, "learning_rate": 5.459927797833936e-05, "loss": 1.4533, "step": 12576 }, { "epoch": 9.084476534296028, "grad_norm": 2.053438425064087, "learning_rate": 5.4577617328519856e-05, "loss": 1.5363, "step": 12582 }, { "epoch": 9.088808664259927, "grad_norm": 2.3502440452575684, "learning_rate": 5.455595667870036e-05, "loss": 1.5271, "step": 12588 }, { "epoch": 9.093140794223826, "grad_norm": 2.051758050918579, "learning_rate": 5.453429602888087e-05, "loss": 1.5609, "step": 12594 }, { "epoch": 9.097472924187725, "grad_norm": 1.8842295408248901, "learning_rate": 5.451263537906137e-05, "loss": 1.5319, "step": 12600 }, { "epoch": 9.101805054151624, "grad_norm": 2.0824687480926514, "learning_rate": 5.449097472924188e-05, "loss": 1.5581, "step": 12606 }, { "epoch": 9.106137184115523, "grad_norm": 2.197455406188965, "learning_rate": 5.4469314079422385e-05, "loss": 1.4412, "step": 12612 }, { "epoch": 9.110469314079422, "grad_norm": 2.2163093090057373, "learning_rate": 5.4447653429602895e-05, "loss": 1.5916, "step": 12618 }, { "epoch": 9.114801444043321, "grad_norm": 2.0774595737457275, "learning_rate": 5.44259927797834e-05, "loss": 1.5015, "step": 12624 }, { "epoch": 9.11913357400722, "grad_norm": 2.1481270790100098, "learning_rate": 5.4404332129963895e-05, "loss": 1.5394, "step": 12630 }, { "epoch": 9.123465703971119, "grad_norm": 1.9280171394348145, "learning_rate": 5.438267148014441e-05, "loss": 1.5304, "step": 12636 }, { "epoch": 9.127797833935018, "grad_norm": 2.0632050037384033, "learning_rate": 5.436101083032491e-05, "loss": 1.4845, "step": 12642 }, { "epoch": 9.132129963898917, "grad_norm": 2.328723907470703, "learning_rate": 5.4339350180505425e-05, "loss": 1.4635, "step": 12648 }, { "epoch": 9.136462093862816, "grad_norm": 1.8843615055084229, "learning_rate": 5.431768953068592e-05, "loss": 1.5197, "step": 12654 }, { "epoch": 9.140794223826715, "grad_norm": 1.9417107105255127, "learning_rate": 5.429602888086643e-05, "loss": 1.5213, "step": 12660 }, { "epoch": 9.145126353790614, "grad_norm": 2.2190942764282227, "learning_rate": 5.4274368231046935e-05, "loss": 1.5363, "step": 12666 }, { "epoch": 9.149458483754513, "grad_norm": 2.2251298427581787, "learning_rate": 5.425270758122744e-05, "loss": 1.4987, "step": 12672 }, { "epoch": 9.153790613718412, "grad_norm": 2.1819283962249756, "learning_rate": 5.423104693140795e-05, "loss": 1.6094, "step": 12678 }, { "epoch": 9.158122743682311, "grad_norm": 2.0244951248168945, "learning_rate": 5.420938628158845e-05, "loss": 1.5268, "step": 12684 }, { "epoch": 9.16245487364621, "grad_norm": 2.0540409088134766, "learning_rate": 5.418772563176896e-05, "loss": 1.6291, "step": 12690 }, { "epoch": 9.166787003610109, "grad_norm": 1.936043620109558, "learning_rate": 5.416606498194946e-05, "loss": 1.4985, "step": 12696 }, { "epoch": 9.171119133574008, "grad_norm": 2.203028678894043, "learning_rate": 5.414440433212996e-05, "loss": 1.407, "step": 12702 }, { "epoch": 9.175451263537907, "grad_norm": 2.05773663520813, "learning_rate": 5.412274368231047e-05, "loss": 1.5097, "step": 12708 }, { "epoch": 9.179783393501806, "grad_norm": 2.3270835876464844, "learning_rate": 5.4101083032490974e-05, "loss": 1.6164, "step": 12714 }, { "epoch": 9.184115523465705, "grad_norm": 2.117050886154175, "learning_rate": 5.4079422382671484e-05, "loss": 1.5851, "step": 12720 }, { "epoch": 9.188447653429602, "grad_norm": 2.211069345474243, "learning_rate": 5.405776173285199e-05, "loss": 1.5681, "step": 12726 }, { "epoch": 9.192779783393501, "grad_norm": 1.9528687000274658, "learning_rate": 5.40361010830325e-05, "loss": 1.5351, "step": 12732 }, { "epoch": 9.1971119133574, "grad_norm": 2.42067813873291, "learning_rate": 5.4014440433213e-05, "loss": 1.4854, "step": 12738 }, { "epoch": 9.201444043321299, "grad_norm": 2.384049892425537, "learning_rate": 5.39927797833935e-05, "loss": 1.5955, "step": 12744 }, { "epoch": 9.205776173285198, "grad_norm": 2.295511245727539, "learning_rate": 5.3971119133574014e-05, "loss": 1.6262, "step": 12750 }, { "epoch": 9.210108303249097, "grad_norm": 2.0357980728149414, "learning_rate": 5.394945848375451e-05, "loss": 1.5284, "step": 12756 }, { "epoch": 9.214440433212996, "grad_norm": 1.9795677661895752, "learning_rate": 5.392779783393502e-05, "loss": 1.5644, "step": 12762 }, { "epoch": 9.218772563176895, "grad_norm": 1.9499012231826782, "learning_rate": 5.390613718411552e-05, "loss": 1.5659, "step": 12768 }, { "epoch": 9.223104693140794, "grad_norm": 2.088865041732788, "learning_rate": 5.388447653429603e-05, "loss": 1.5076, "step": 12774 }, { "epoch": 9.227436823104693, "grad_norm": 2.0630903244018555, "learning_rate": 5.3862815884476536e-05, "loss": 1.5464, "step": 12780 }, { "epoch": 9.231768953068592, "grad_norm": 2.7943851947784424, "learning_rate": 5.384115523465704e-05, "loss": 1.5582, "step": 12786 }, { "epoch": 9.23610108303249, "grad_norm": 1.939738392829895, "learning_rate": 5.381949458483755e-05, "loss": 1.6067, "step": 12792 }, { "epoch": 9.24043321299639, "grad_norm": 2.174177408218384, "learning_rate": 5.3797833935018046e-05, "loss": 1.5875, "step": 12798 }, { "epoch": 9.244765342960289, "grad_norm": 2.255918502807617, "learning_rate": 5.377617328519856e-05, "loss": 1.5678, "step": 12804 }, { "epoch": 9.249097472924188, "grad_norm": 2.288457155227661, "learning_rate": 5.375451263537906e-05, "loss": 1.5989, "step": 12810 }, { "epoch": 9.253429602888087, "grad_norm": 2.3024749755859375, "learning_rate": 5.3732851985559576e-05, "loss": 1.4943, "step": 12816 }, { "epoch": 9.257761732851986, "grad_norm": 1.9192348718643188, "learning_rate": 5.371119133574007e-05, "loss": 1.5822, "step": 12822 }, { "epoch": 9.262093862815885, "grad_norm": 2.1705594062805176, "learning_rate": 5.3689530685920576e-05, "loss": 1.5675, "step": 12828 }, { "epoch": 9.266425992779784, "grad_norm": 2.134631872177124, "learning_rate": 5.3667870036101086e-05, "loss": 1.5644, "step": 12834 }, { "epoch": 9.270758122743683, "grad_norm": 2.2905690670013428, "learning_rate": 5.364620938628159e-05, "loss": 1.504, "step": 12840 }, { "epoch": 9.275090252707582, "grad_norm": 2.1851017475128174, "learning_rate": 5.36245487364621e-05, "loss": 1.5641, "step": 12846 }, { "epoch": 9.27942238267148, "grad_norm": 2.116050958633423, "learning_rate": 5.36028880866426e-05, "loss": 1.5678, "step": 12852 }, { "epoch": 9.28375451263538, "grad_norm": 1.9505747556686401, "learning_rate": 5.358122743682311e-05, "loss": 1.4904, "step": 12858 }, { "epoch": 9.288086642599279, "grad_norm": 1.902115821838379, "learning_rate": 5.3559566787003616e-05, "loss": 1.509, "step": 12864 }, { "epoch": 9.292418772563177, "grad_norm": 1.8676198720932007, "learning_rate": 5.353790613718411e-05, "loss": 1.4685, "step": 12870 }, { "epoch": 9.296750902527076, "grad_norm": 2.012260675430298, "learning_rate": 5.351624548736462e-05, "loss": 1.4603, "step": 12876 }, { "epoch": 9.301083032490975, "grad_norm": 2.0526819229125977, "learning_rate": 5.3494584837545125e-05, "loss": 1.5827, "step": 12882 }, { "epoch": 9.305415162454874, "grad_norm": 2.1591522693634033, "learning_rate": 5.3472924187725635e-05, "loss": 1.4875, "step": 12888 }, { "epoch": 9.309747292418773, "grad_norm": 2.151298999786377, "learning_rate": 5.345126353790614e-05, "loss": 1.5174, "step": 12894 }, { "epoch": 9.314079422382672, "grad_norm": 1.9836814403533936, "learning_rate": 5.342960288808665e-05, "loss": 1.492, "step": 12900 }, { "epoch": 9.31841155234657, "grad_norm": 1.9636588096618652, "learning_rate": 5.340794223826715e-05, "loss": 1.5203, "step": 12906 }, { "epoch": 9.322743682310469, "grad_norm": 2.334902286529541, "learning_rate": 5.338628158844765e-05, "loss": 1.6132, "step": 12912 }, { "epoch": 9.327075812274368, "grad_norm": 2.164987325668335, "learning_rate": 5.3364620938628165e-05, "loss": 1.5888, "step": 12918 }, { "epoch": 9.331407942238267, "grad_norm": 2.0883188247680664, "learning_rate": 5.334296028880866e-05, "loss": 1.5133, "step": 12924 }, { "epoch": 9.335740072202166, "grad_norm": 1.9443188905715942, "learning_rate": 5.332129963898918e-05, "loss": 1.5559, "step": 12930 }, { "epoch": 9.340072202166064, "grad_norm": 2.375842332839966, "learning_rate": 5.3299638989169675e-05, "loss": 1.5627, "step": 12936 }, { "epoch": 9.344404332129963, "grad_norm": 2.3309121131896973, "learning_rate": 5.3277978339350185e-05, "loss": 1.5225, "step": 12942 }, { "epoch": 9.348736462093862, "grad_norm": 2.536954879760742, "learning_rate": 5.325631768953069e-05, "loss": 1.6839, "step": 12948 }, { "epoch": 9.353068592057761, "grad_norm": 2.337256908416748, "learning_rate": 5.323465703971119e-05, "loss": 1.5738, "step": 12954 }, { "epoch": 9.35740072202166, "grad_norm": 2.345672369003296, "learning_rate": 5.32129963898917e-05, "loss": 1.5596, "step": 12960 }, { "epoch": 9.36173285198556, "grad_norm": 1.8097978830337524, "learning_rate": 5.3191335740072204e-05, "loss": 1.4824, "step": 12966 }, { "epoch": 9.366064981949458, "grad_norm": 2.0242931842803955, "learning_rate": 5.3169675090252714e-05, "loss": 1.5615, "step": 12972 }, { "epoch": 9.370397111913357, "grad_norm": 2.108421564102173, "learning_rate": 5.314801444043321e-05, "loss": 1.518, "step": 12978 }, { "epoch": 9.374729241877256, "grad_norm": 2.092357635498047, "learning_rate": 5.312635379061373e-05, "loss": 1.4969, "step": 12984 }, { "epoch": 9.379061371841155, "grad_norm": 2.383894443511963, "learning_rate": 5.3104693140794224e-05, "loss": 1.5128, "step": 12990 }, { "epoch": 9.383393501805054, "grad_norm": 2.1744842529296875, "learning_rate": 5.308303249097473e-05, "loss": 1.5703, "step": 12996 }, { "epoch": 9.387725631768953, "grad_norm": 2.1576921939849854, "learning_rate": 5.306137184115524e-05, "loss": 1.483, "step": 13002 }, { "epoch": 9.392057761732852, "grad_norm": 2.067983627319336, "learning_rate": 5.303971119133574e-05, "loss": 1.5408, "step": 13008 }, { "epoch": 9.396389891696751, "grad_norm": 2.3359529972076416, "learning_rate": 5.301805054151625e-05, "loss": 1.597, "step": 13014 }, { "epoch": 9.40072202166065, "grad_norm": 1.9858944416046143, "learning_rate": 5.2996389891696754e-05, "loss": 1.4092, "step": 13020 }, { "epoch": 9.40505415162455, "grad_norm": 2.259436845779419, "learning_rate": 5.2974729241877264e-05, "loss": 1.5777, "step": 13026 }, { "epoch": 9.409386281588448, "grad_norm": 2.045910358428955, "learning_rate": 5.295306859205777e-05, "loss": 1.5632, "step": 13032 }, { "epoch": 9.413718411552347, "grad_norm": 2.13466477394104, "learning_rate": 5.293140794223826e-05, "loss": 1.5976, "step": 13038 }, { "epoch": 9.418050541516246, "grad_norm": 2.2947425842285156, "learning_rate": 5.290974729241877e-05, "loss": 1.6431, "step": 13044 }, { "epoch": 9.422382671480145, "grad_norm": 1.8938108682632446, "learning_rate": 5.2888086642599276e-05, "loss": 1.5767, "step": 13050 }, { "epoch": 9.426714801444044, "grad_norm": 2.257636547088623, "learning_rate": 5.2866425992779786e-05, "loss": 1.6279, "step": 13056 }, { "epoch": 9.431046931407943, "grad_norm": 2.2670786380767822, "learning_rate": 5.284476534296029e-05, "loss": 1.5788, "step": 13062 }, { "epoch": 9.435379061371842, "grad_norm": 2.2394309043884277, "learning_rate": 5.28231046931408e-05, "loss": 1.5029, "step": 13068 }, { "epoch": 9.43971119133574, "grad_norm": 2.22761869430542, "learning_rate": 5.28014440433213e-05, "loss": 1.5512, "step": 13074 }, { "epoch": 9.444043321299638, "grad_norm": 2.2470757961273193, "learning_rate": 5.27797833935018e-05, "loss": 1.5553, "step": 13080 }, { "epoch": 9.448375451263537, "grad_norm": 2.194565534591675, "learning_rate": 5.2758122743682316e-05, "loss": 1.5708, "step": 13086 }, { "epoch": 9.452707581227436, "grad_norm": 2.0176570415496826, "learning_rate": 5.273646209386281e-05, "loss": 1.5839, "step": 13092 }, { "epoch": 9.457039711191335, "grad_norm": 2.3660902976989746, "learning_rate": 5.271480144404333e-05, "loss": 1.5662, "step": 13098 }, { "epoch": 9.461371841155234, "grad_norm": 2.247149705886841, "learning_rate": 5.2693140794223826e-05, "loss": 1.6475, "step": 13104 }, { "epoch": 9.465703971119133, "grad_norm": 2.199258327484131, "learning_rate": 5.267148014440434e-05, "loss": 1.6218, "step": 13110 }, { "epoch": 9.470036101083032, "grad_norm": 2.117887258529663, "learning_rate": 5.264981949458484e-05, "loss": 1.5864, "step": 13116 }, { "epoch": 9.474368231046931, "grad_norm": 2.1368136405944824, "learning_rate": 5.262815884476534e-05, "loss": 1.5912, "step": 13122 }, { "epoch": 9.47870036101083, "grad_norm": 2.3343889713287354, "learning_rate": 5.260649819494585e-05, "loss": 1.5719, "step": 13128 }, { "epoch": 9.483032490974729, "grad_norm": 2.0993223190307617, "learning_rate": 5.2584837545126356e-05, "loss": 1.6086, "step": 13134 }, { "epoch": 9.487364620938628, "grad_norm": 2.220853328704834, "learning_rate": 5.2563176895306866e-05, "loss": 1.5919, "step": 13140 }, { "epoch": 9.491696750902527, "grad_norm": 2.113405227661133, "learning_rate": 5.254151624548737e-05, "loss": 1.4954, "step": 13146 }, { "epoch": 9.496028880866426, "grad_norm": 1.9651336669921875, "learning_rate": 5.251985559566788e-05, "loss": 1.5602, "step": 13152 }, { "epoch": 9.500361010830325, "grad_norm": 2.178856372833252, "learning_rate": 5.2498194945848375e-05, "loss": 1.5066, "step": 13158 }, { "epoch": 9.504693140794224, "grad_norm": 2.2245545387268066, "learning_rate": 5.247653429602888e-05, "loss": 1.5387, "step": 13164 }, { "epoch": 9.509025270758123, "grad_norm": 1.906387209892273, "learning_rate": 5.245487364620939e-05, "loss": 1.3326, "step": 13170 }, { "epoch": 9.513357400722022, "grad_norm": 2.555107593536377, "learning_rate": 5.243321299638989e-05, "loss": 1.5273, "step": 13176 }, { "epoch": 9.51768953068592, "grad_norm": 2.107290744781494, "learning_rate": 5.24115523465704e-05, "loss": 1.5979, "step": 13182 }, { "epoch": 9.52202166064982, "grad_norm": 1.9489107131958008, "learning_rate": 5.2389891696750905e-05, "loss": 1.6088, "step": 13188 }, { "epoch": 9.526353790613719, "grad_norm": 2.0959253311157227, "learning_rate": 5.2368231046931415e-05, "loss": 1.5144, "step": 13194 }, { "epoch": 9.530685920577618, "grad_norm": 1.9230812788009644, "learning_rate": 5.234657039711192e-05, "loss": 1.6095, "step": 13200 }, { "epoch": 9.535018050541517, "grad_norm": 1.9981838464736938, "learning_rate": 5.2324909747292415e-05, "loss": 1.6271, "step": 13206 }, { "epoch": 9.539350180505416, "grad_norm": 1.8735038042068481, "learning_rate": 5.230324909747293e-05, "loss": 1.5099, "step": 13212 }, { "epoch": 9.543682310469315, "grad_norm": 2.1915688514709473, "learning_rate": 5.228158844765343e-05, "loss": 1.5051, "step": 13218 }, { "epoch": 9.548014440433214, "grad_norm": 2.3255093097686768, "learning_rate": 5.225992779783394e-05, "loss": 1.49, "step": 13224 }, { "epoch": 9.552346570397113, "grad_norm": 2.2508187294006348, "learning_rate": 5.223826714801444e-05, "loss": 1.5666, "step": 13230 }, { "epoch": 9.556678700361012, "grad_norm": 2.228929281234741, "learning_rate": 5.221660649819495e-05, "loss": 1.5021, "step": 13236 }, { "epoch": 9.56101083032491, "grad_norm": 2.1795029640197754, "learning_rate": 5.2194945848375454e-05, "loss": 1.5731, "step": 13242 }, { "epoch": 9.56534296028881, "grad_norm": 2.166821241378784, "learning_rate": 5.217328519855596e-05, "loss": 1.4715, "step": 13248 }, { "epoch": 9.569675090252707, "grad_norm": 2.0718955993652344, "learning_rate": 5.215162454873647e-05, "loss": 1.4873, "step": 13254 }, { "epoch": 9.574007220216606, "grad_norm": 2.412121057510376, "learning_rate": 5.2129963898916964e-05, "loss": 1.5767, "step": 13260 }, { "epoch": 9.578339350180505, "grad_norm": 1.9756791591644287, "learning_rate": 5.210830324909748e-05, "loss": 1.685, "step": 13266 }, { "epoch": 9.582671480144404, "grad_norm": 2.2172303199768066, "learning_rate": 5.208664259927798e-05, "loss": 1.5767, "step": 13272 }, { "epoch": 9.587003610108303, "grad_norm": 2.583319664001465, "learning_rate": 5.2064981949458494e-05, "loss": 1.5772, "step": 13278 }, { "epoch": 9.591335740072202, "grad_norm": 2.0782053470611572, "learning_rate": 5.204332129963899e-05, "loss": 1.5706, "step": 13284 }, { "epoch": 9.5956678700361, "grad_norm": 1.954258918762207, "learning_rate": 5.2021660649819494e-05, "loss": 1.5329, "step": 13290 }, { "epoch": 9.6, "grad_norm": 2.1231424808502197, "learning_rate": 5.2000000000000004e-05, "loss": 1.6598, "step": 13296 }, { "epoch": 9.604332129963899, "grad_norm": 2.234015941619873, "learning_rate": 5.197833935018051e-05, "loss": 1.4928, "step": 13302 }, { "epoch": 9.608664259927798, "grad_norm": 2.2357542514801025, "learning_rate": 5.195667870036102e-05, "loss": 1.4771, "step": 13308 }, { "epoch": 9.612996389891697, "grad_norm": 2.1566619873046875, "learning_rate": 5.193501805054152e-05, "loss": 1.6454, "step": 13314 }, { "epoch": 9.617328519855596, "grad_norm": 2.2505266666412354, "learning_rate": 5.191335740072203e-05, "loss": 1.5158, "step": 13320 }, { "epoch": 9.621660649819495, "grad_norm": 2.1636545658111572, "learning_rate": 5.1891696750902526e-05, "loss": 1.5332, "step": 13326 }, { "epoch": 9.625992779783394, "grad_norm": 2.0250983238220215, "learning_rate": 5.187003610108303e-05, "loss": 1.5711, "step": 13332 }, { "epoch": 9.630324909747292, "grad_norm": 1.9677085876464844, "learning_rate": 5.184837545126354e-05, "loss": 1.5925, "step": 13338 }, { "epoch": 9.634657039711191, "grad_norm": 2.363617181777954, "learning_rate": 5.182671480144404e-05, "loss": 1.6139, "step": 13344 }, { "epoch": 9.63898916967509, "grad_norm": 2.1360225677490234, "learning_rate": 5.180505415162455e-05, "loss": 1.5761, "step": 13350 }, { "epoch": 9.64332129963899, "grad_norm": 1.9913398027420044, "learning_rate": 5.1783393501805056e-05, "loss": 1.5361, "step": 13356 }, { "epoch": 9.647653429602888, "grad_norm": 2.1897294521331787, "learning_rate": 5.1761732851985566e-05, "loss": 1.4688, "step": 13362 }, { "epoch": 9.651985559566787, "grad_norm": 2.0441946983337402, "learning_rate": 5.174007220216607e-05, "loss": 1.5965, "step": 13368 }, { "epoch": 9.656317689530686, "grad_norm": 2.1288411617279053, "learning_rate": 5.1718411552346566e-05, "loss": 1.4506, "step": 13374 }, { "epoch": 9.660649819494585, "grad_norm": 1.9829185009002686, "learning_rate": 5.169675090252708e-05, "loss": 1.4903, "step": 13380 }, { "epoch": 9.664981949458484, "grad_norm": 1.93166184425354, "learning_rate": 5.167509025270758e-05, "loss": 1.5095, "step": 13386 }, { "epoch": 9.669314079422383, "grad_norm": 2.301772117614746, "learning_rate": 5.1653429602888096e-05, "loss": 1.5824, "step": 13392 }, { "epoch": 9.673646209386282, "grad_norm": 2.1365838050842285, "learning_rate": 5.163176895306859e-05, "loss": 1.4405, "step": 13398 }, { "epoch": 9.677978339350181, "grad_norm": 2.0197858810424805, "learning_rate": 5.16101083032491e-05, "loss": 1.4811, "step": 13404 }, { "epoch": 9.68231046931408, "grad_norm": 2.463376522064209, "learning_rate": 5.1588447653429606e-05, "loss": 1.6446, "step": 13410 }, { "epoch": 9.686642599277977, "grad_norm": 1.9436571598052979, "learning_rate": 5.156678700361011e-05, "loss": 1.4295, "step": 13416 }, { "epoch": 9.690974729241876, "grad_norm": 2.1034350395202637, "learning_rate": 5.154512635379062e-05, "loss": 1.5683, "step": 13422 }, { "epoch": 9.695306859205775, "grad_norm": 1.8237329721450806, "learning_rate": 5.152346570397112e-05, "loss": 1.472, "step": 13428 }, { "epoch": 9.699638989169674, "grad_norm": 1.9748588800430298, "learning_rate": 5.150180505415163e-05, "loss": 1.6486, "step": 13434 }, { "epoch": 9.703971119133573, "grad_norm": 2.1246743202209473, "learning_rate": 5.148014440433213e-05, "loss": 1.5887, "step": 13440 }, { "epoch": 9.708303249097472, "grad_norm": 2.411581516265869, "learning_rate": 5.1458483754512645e-05, "loss": 1.5412, "step": 13446 }, { "epoch": 9.712635379061371, "grad_norm": 2.290353775024414, "learning_rate": 5.143682310469314e-05, "loss": 1.441, "step": 13452 }, { "epoch": 9.71696750902527, "grad_norm": 2.1770200729370117, "learning_rate": 5.1415162454873645e-05, "loss": 1.5483, "step": 13458 }, { "epoch": 9.72129963898917, "grad_norm": 2.0860326290130615, "learning_rate": 5.1393501805054155e-05, "loss": 1.6362, "step": 13464 }, { "epoch": 9.725631768953068, "grad_norm": 2.049882650375366, "learning_rate": 5.137184115523466e-05, "loss": 1.4656, "step": 13470 }, { "epoch": 9.729963898916967, "grad_norm": 1.991666316986084, "learning_rate": 5.135018050541517e-05, "loss": 1.6103, "step": 13476 }, { "epoch": 9.734296028880866, "grad_norm": 2.1093249320983887, "learning_rate": 5.132851985559567e-05, "loss": 1.4801, "step": 13482 }, { "epoch": 9.738628158844765, "grad_norm": 2.103339910507202, "learning_rate": 5.130685920577618e-05, "loss": 1.514, "step": 13488 }, { "epoch": 9.742960288808664, "grad_norm": 2.180771827697754, "learning_rate": 5.1285198555956685e-05, "loss": 1.5963, "step": 13494 }, { "epoch": 9.747292418772563, "grad_norm": 2.1355159282684326, "learning_rate": 5.126353790613718e-05, "loss": 1.6741, "step": 13500 }, { "epoch": 9.751624548736462, "grad_norm": 2.086791753768921, "learning_rate": 5.124187725631769e-05, "loss": 1.5457, "step": 13506 }, { "epoch": 9.755956678700361, "grad_norm": 2.364572525024414, "learning_rate": 5.1220216606498194e-05, "loss": 1.5785, "step": 13512 }, { "epoch": 9.76028880866426, "grad_norm": 2.278076648712158, "learning_rate": 5.1198555956678704e-05, "loss": 1.4476, "step": 13518 }, { "epoch": 9.764620938628159, "grad_norm": 2.2016143798828125, "learning_rate": 5.117689530685921e-05, "loss": 1.5826, "step": 13524 }, { "epoch": 9.768953068592058, "grad_norm": 2.1963164806365967, "learning_rate": 5.115523465703972e-05, "loss": 1.5769, "step": 13530 }, { "epoch": 9.773285198555957, "grad_norm": 2.218479633331299, "learning_rate": 5.113357400722022e-05, "loss": 1.584, "step": 13536 }, { "epoch": 9.777617328519856, "grad_norm": 1.8303182125091553, "learning_rate": 5.111191335740072e-05, "loss": 1.5654, "step": 13542 }, { "epoch": 9.781949458483755, "grad_norm": 2.0453314781188965, "learning_rate": 5.1090252707581234e-05, "loss": 1.5926, "step": 13548 }, { "epoch": 9.786281588447654, "grad_norm": 2.244978904724121, "learning_rate": 5.106859205776173e-05, "loss": 1.5693, "step": 13554 }, { "epoch": 9.790613718411553, "grad_norm": 2.185175657272339, "learning_rate": 5.104693140794225e-05, "loss": 1.5633, "step": 13560 }, { "epoch": 9.794945848375452, "grad_norm": 2.183323860168457, "learning_rate": 5.1025270758122744e-05, "loss": 1.4956, "step": 13566 }, { "epoch": 9.79927797833935, "grad_norm": 2.0931267738342285, "learning_rate": 5.1003610108303254e-05, "loss": 1.6251, "step": 13572 }, { "epoch": 9.80361010830325, "grad_norm": 2.154728651046753, "learning_rate": 5.098194945848376e-05, "loss": 1.6026, "step": 13578 }, { "epoch": 9.807942238267149, "grad_norm": 2.4687490463256836, "learning_rate": 5.096028880866426e-05, "loss": 1.471, "step": 13584 }, { "epoch": 9.812274368231048, "grad_norm": 2.169205904006958, "learning_rate": 5.093862815884477e-05, "loss": 1.677, "step": 13590 }, { "epoch": 9.816606498194945, "grad_norm": 2.4284868240356445, "learning_rate": 5.091696750902527e-05, "loss": 1.6476, "step": 13596 }, { "epoch": 9.820938628158844, "grad_norm": 2.0832717418670654, "learning_rate": 5.089530685920578e-05, "loss": 1.5533, "step": 13602 }, { "epoch": 9.825270758122743, "grad_norm": 2.0670180320739746, "learning_rate": 5.0873646209386286e-05, "loss": 1.5354, "step": 13608 }, { "epoch": 9.829602888086642, "grad_norm": 2.289077043533325, "learning_rate": 5.0851985559566796e-05, "loss": 1.5016, "step": 13614 }, { "epoch": 9.833935018050541, "grad_norm": 2.1111600399017334, "learning_rate": 5.083032490974729e-05, "loss": 1.5037, "step": 13620 }, { "epoch": 9.83826714801444, "grad_norm": 2.098095655441284, "learning_rate": 5.0808664259927796e-05, "loss": 1.457, "step": 13626 }, { "epoch": 9.842599277978339, "grad_norm": 2.2200539112091064, "learning_rate": 5.0787003610108306e-05, "loss": 1.4925, "step": 13632 }, { "epoch": 9.846931407942238, "grad_norm": 2.062950849533081, "learning_rate": 5.076534296028881e-05, "loss": 1.6782, "step": 13638 }, { "epoch": 9.851263537906137, "grad_norm": 2.0816991329193115, "learning_rate": 5.074368231046932e-05, "loss": 1.5813, "step": 13644 }, { "epoch": 9.855595667870036, "grad_norm": 2.240861654281616, "learning_rate": 5.072202166064982e-05, "loss": 1.4739, "step": 13650 }, { "epoch": 9.859927797833935, "grad_norm": 2.522345542907715, "learning_rate": 5.070036101083033e-05, "loss": 1.7921, "step": 13656 }, { "epoch": 9.864259927797834, "grad_norm": 2.149214744567871, "learning_rate": 5.0678700361010836e-05, "loss": 1.4394, "step": 13662 }, { "epoch": 9.868592057761733, "grad_norm": 2.3839828968048096, "learning_rate": 5.065703971119133e-05, "loss": 1.6192, "step": 13668 }, { "epoch": 9.872924187725632, "grad_norm": 2.1648545265197754, "learning_rate": 5.063537906137185e-05, "loss": 1.4609, "step": 13674 }, { "epoch": 9.87725631768953, "grad_norm": 2.057360887527466, "learning_rate": 5.0613718411552346e-05, "loss": 1.538, "step": 13680 }, { "epoch": 9.88158844765343, "grad_norm": 2.1901419162750244, "learning_rate": 5.0592057761732856e-05, "loss": 1.5883, "step": 13686 }, { "epoch": 9.885920577617329, "grad_norm": 2.068713426589966, "learning_rate": 5.057039711191336e-05, "loss": 1.594, "step": 13692 }, { "epoch": 9.890252707581228, "grad_norm": 2.231536388397217, "learning_rate": 5.054873646209386e-05, "loss": 1.5794, "step": 13698 }, { "epoch": 9.894584837545127, "grad_norm": 2.1959140300750732, "learning_rate": 5.052707581227437e-05, "loss": 1.5656, "step": 13704 }, { "epoch": 9.898916967509026, "grad_norm": 2.1210460662841797, "learning_rate": 5.0505415162454875e-05, "loss": 1.6297, "step": 13710 }, { "epoch": 9.903249097472925, "grad_norm": 1.9860336780548096, "learning_rate": 5.0483754512635385e-05, "loss": 1.5767, "step": 13716 }, { "epoch": 9.907581227436824, "grad_norm": 2.004159927368164, "learning_rate": 5.046209386281588e-05, "loss": 1.6298, "step": 13722 }, { "epoch": 9.911913357400723, "grad_norm": 1.8768541812896729, "learning_rate": 5.04404332129964e-05, "loss": 1.5109, "step": 13728 }, { "epoch": 9.916245487364622, "grad_norm": 1.9933325052261353, "learning_rate": 5.0418772563176895e-05, "loss": 1.516, "step": 13734 }, { "epoch": 9.92057761732852, "grad_norm": 2.251406192779541, "learning_rate": 5.03971119133574e-05, "loss": 1.6774, "step": 13740 }, { "epoch": 9.92490974729242, "grad_norm": 2.1851210594177246, "learning_rate": 5.037545126353791e-05, "loss": 1.6144, "step": 13746 }, { "epoch": 9.929241877256318, "grad_norm": 1.9815512895584106, "learning_rate": 5.035379061371841e-05, "loss": 1.4826, "step": 13752 }, { "epoch": 9.933574007220217, "grad_norm": 2.1019887924194336, "learning_rate": 5.033212996389892e-05, "loss": 1.4469, "step": 13758 }, { "epoch": 9.937906137184115, "grad_norm": 1.9896489381790161, "learning_rate": 5.0310469314079425e-05, "loss": 1.5667, "step": 13764 }, { "epoch": 9.942238267148014, "grad_norm": 2.091355085372925, "learning_rate": 5.0288808664259935e-05, "loss": 1.4222, "step": 13770 }, { "epoch": 9.946570397111913, "grad_norm": 1.9639573097229004, "learning_rate": 5.026714801444044e-05, "loss": 1.4558, "step": 13776 }, { "epoch": 9.950902527075812, "grad_norm": 2.2421727180480957, "learning_rate": 5.0245487364620934e-05, "loss": 1.5996, "step": 13782 }, { "epoch": 9.95523465703971, "grad_norm": 1.9946539402008057, "learning_rate": 5.0223826714801444e-05, "loss": 1.522, "step": 13788 }, { "epoch": 9.95956678700361, "grad_norm": 2.238999843597412, "learning_rate": 5.020216606498195e-05, "loss": 1.5197, "step": 13794 }, { "epoch": 9.963898916967509, "grad_norm": 2.2011122703552246, "learning_rate": 5.018050541516246e-05, "loss": 1.6195, "step": 13800 }, { "epoch": 9.968231046931407, "grad_norm": 2.19651198387146, "learning_rate": 5.015884476534296e-05, "loss": 1.5721, "step": 13806 }, { "epoch": 9.972563176895306, "grad_norm": 1.7406957149505615, "learning_rate": 5.013718411552347e-05, "loss": 1.479, "step": 13812 }, { "epoch": 9.976895306859205, "grad_norm": 2.1590170860290527, "learning_rate": 5.0115523465703974e-05, "loss": 1.6992, "step": 13818 }, { "epoch": 9.981227436823104, "grad_norm": 2.1502208709716797, "learning_rate": 5.009386281588447e-05, "loss": 1.5604, "step": 13824 }, { "epoch": 9.985559566787003, "grad_norm": 2.018888473510742, "learning_rate": 5.007220216606499e-05, "loss": 1.5623, "step": 13830 }, { "epoch": 9.989891696750902, "grad_norm": 2.106571912765503, "learning_rate": 5.0050541516245484e-05, "loss": 1.5127, "step": 13836 }, { "epoch": 9.994223826714801, "grad_norm": 2.0180845260620117, "learning_rate": 5.0028880866426e-05, "loss": 1.6298, "step": 13842 }, { "epoch": 9.9985559566787, "grad_norm": 2.4180965423583984, "learning_rate": 5.00072202166065e-05, "loss": 1.5977, "step": 13848 }, { "epoch": 10.0028880866426, "grad_norm": 2.097599983215332, "learning_rate": 4.998555956678701e-05, "loss": 1.5094, "step": 13854 }, { "epoch": 10.007220216606498, "grad_norm": 2.1146388053894043, "learning_rate": 4.996389891696751e-05, "loss": 1.6093, "step": 13860 }, { "epoch": 10.011552346570397, "grad_norm": 2.129944324493408, "learning_rate": 4.994223826714802e-05, "loss": 1.486, "step": 13866 }, { "epoch": 10.015884476534296, "grad_norm": 2.254216194152832, "learning_rate": 4.992057761732852e-05, "loss": 1.5158, "step": 13872 }, { "epoch": 10.020216606498195, "grad_norm": 2.456321954727173, "learning_rate": 4.9898916967509026e-05, "loss": 1.6612, "step": 13878 }, { "epoch": 10.024548736462094, "grad_norm": 2.1398701667785645, "learning_rate": 4.9877256317689536e-05, "loss": 1.5492, "step": 13884 }, { "epoch": 10.028880866425993, "grad_norm": 2.178299903869629, "learning_rate": 4.985559566787004e-05, "loss": 1.5966, "step": 13890 }, { "epoch": 10.033212996389892, "grad_norm": 2.1067919731140137, "learning_rate": 4.983393501805054e-05, "loss": 1.545, "step": 13896 }, { "epoch": 10.037545126353791, "grad_norm": 2.0320487022399902, "learning_rate": 4.9812274368231046e-05, "loss": 1.4649, "step": 13902 }, { "epoch": 10.04187725631769, "grad_norm": 2.026679039001465, "learning_rate": 4.9790613718411556e-05, "loss": 1.5045, "step": 13908 }, { "epoch": 10.046209386281589, "grad_norm": 2.4859981536865234, "learning_rate": 4.976895306859206e-05, "loss": 1.5548, "step": 13914 }, { "epoch": 10.050541516245488, "grad_norm": 2.190380096435547, "learning_rate": 4.974729241877257e-05, "loss": 1.4744, "step": 13920 }, { "epoch": 10.054873646209387, "grad_norm": 2.06523060798645, "learning_rate": 4.972563176895307e-05, "loss": 1.5488, "step": 13926 }, { "epoch": 10.059205776173286, "grad_norm": 2.1824891567230225, "learning_rate": 4.9703971119133576e-05, "loss": 1.468, "step": 13932 }, { "epoch": 10.063537906137183, "grad_norm": 2.116769790649414, "learning_rate": 4.968231046931408e-05, "loss": 1.5049, "step": 13938 }, { "epoch": 10.067870036101082, "grad_norm": 2.045421600341797, "learning_rate": 4.966064981949459e-05, "loss": 1.5119, "step": 13944 }, { "epoch": 10.072202166064981, "grad_norm": 2.0766656398773193, "learning_rate": 4.963898916967509e-05, "loss": 1.4791, "step": 13950 }, { "epoch": 10.07653429602888, "grad_norm": 2.2448534965515137, "learning_rate": 4.96173285198556e-05, "loss": 1.5268, "step": 13956 }, { "epoch": 10.08086642599278, "grad_norm": 2.257699489593506, "learning_rate": 4.9595667870036106e-05, "loss": 1.4489, "step": 13962 }, { "epoch": 10.085198555956678, "grad_norm": 2.2425789833068848, "learning_rate": 4.957400722021661e-05, "loss": 1.5685, "step": 13968 }, { "epoch": 10.089530685920577, "grad_norm": 2.2534611225128174, "learning_rate": 4.955234657039711e-05, "loss": 1.5684, "step": 13974 }, { "epoch": 10.093862815884476, "grad_norm": 2.469127655029297, "learning_rate": 4.953068592057762e-05, "loss": 1.4865, "step": 13980 }, { "epoch": 10.098194945848375, "grad_norm": 2.1506285667419434, "learning_rate": 4.9509025270758125e-05, "loss": 1.4545, "step": 13986 }, { "epoch": 10.102527075812274, "grad_norm": 2.226968288421631, "learning_rate": 4.948736462093863e-05, "loss": 1.5335, "step": 13992 }, { "epoch": 10.106859205776173, "grad_norm": 2.0292999744415283, "learning_rate": 4.946570397111914e-05, "loss": 1.5052, "step": 13998 }, { "epoch": 10.111191335740072, "grad_norm": 2.124643325805664, "learning_rate": 4.944404332129964e-05, "loss": 1.4483, "step": 14004 }, { "epoch": 10.115523465703971, "grad_norm": 2.3071084022521973, "learning_rate": 4.9422382671480145e-05, "loss": 1.5052, "step": 14010 }, { "epoch": 10.11985559566787, "grad_norm": 2.249065399169922, "learning_rate": 4.940072202166065e-05, "loss": 1.5693, "step": 14016 }, { "epoch": 10.124187725631769, "grad_norm": 2.125871181488037, "learning_rate": 4.937906137184116e-05, "loss": 1.5449, "step": 14022 }, { "epoch": 10.128519855595668, "grad_norm": 2.3416812419891357, "learning_rate": 4.935740072202166e-05, "loss": 1.5731, "step": 14028 }, { "epoch": 10.132851985559567, "grad_norm": 2.2656147480010986, "learning_rate": 4.933574007220217e-05, "loss": 1.5357, "step": 14034 }, { "epoch": 10.137184115523466, "grad_norm": 2.2251172065734863, "learning_rate": 4.9314079422382675e-05, "loss": 1.4525, "step": 14040 }, { "epoch": 10.141516245487365, "grad_norm": 2.25722074508667, "learning_rate": 4.9292418772563185e-05, "loss": 1.46, "step": 14046 }, { "epoch": 10.145848375451264, "grad_norm": 2.4635400772094727, "learning_rate": 4.927075812274368e-05, "loss": 1.5427, "step": 14052 }, { "epoch": 10.150180505415163, "grad_norm": 2.2485978603363037, "learning_rate": 4.924909747292419e-05, "loss": 1.5631, "step": 14058 }, { "epoch": 10.154512635379062, "grad_norm": 2.315094470977783, "learning_rate": 4.9227436823104694e-05, "loss": 1.4875, "step": 14064 }, { "epoch": 10.15884476534296, "grad_norm": 2.2360565662384033, "learning_rate": 4.92057761732852e-05, "loss": 1.4682, "step": 14070 }, { "epoch": 10.16317689530686, "grad_norm": 2.2100956439971924, "learning_rate": 4.918411552346571e-05, "loss": 1.5783, "step": 14076 }, { "epoch": 10.167509025270759, "grad_norm": 2.2891037464141846, "learning_rate": 4.916245487364621e-05, "loss": 1.5247, "step": 14082 }, { "epoch": 10.171841155234658, "grad_norm": 1.976851463317871, "learning_rate": 4.914079422382672e-05, "loss": 1.4681, "step": 14088 }, { "epoch": 10.176173285198557, "grad_norm": 2.2144930362701416, "learning_rate": 4.911913357400722e-05, "loss": 1.604, "step": 14094 }, { "epoch": 10.180505415162456, "grad_norm": 2.438720703125, "learning_rate": 4.909747292418773e-05, "loss": 1.4123, "step": 14100 }, { "epoch": 10.184837545126355, "grad_norm": 2.2263474464416504, "learning_rate": 4.907581227436823e-05, "loss": 1.5206, "step": 14106 }, { "epoch": 10.189169675090252, "grad_norm": 2.2284207344055176, "learning_rate": 4.905415162454874e-05, "loss": 1.536, "step": 14112 }, { "epoch": 10.19350180505415, "grad_norm": 2.214881420135498, "learning_rate": 4.9032490974729244e-05, "loss": 1.5108, "step": 14118 }, { "epoch": 10.19783393501805, "grad_norm": 1.9738929271697998, "learning_rate": 4.9010830324909754e-05, "loss": 1.4482, "step": 14124 }, { "epoch": 10.202166064981949, "grad_norm": 1.9565082788467407, "learning_rate": 4.898916967509026e-05, "loss": 1.392, "step": 14130 }, { "epoch": 10.206498194945848, "grad_norm": 1.8968371152877808, "learning_rate": 4.896750902527076e-05, "loss": 1.5608, "step": 14136 }, { "epoch": 10.210830324909747, "grad_norm": 2.1367743015289307, "learning_rate": 4.894584837545126e-05, "loss": 1.4428, "step": 14142 }, { "epoch": 10.215162454873646, "grad_norm": 2.2396178245544434, "learning_rate": 4.892418772563177e-05, "loss": 1.5671, "step": 14148 }, { "epoch": 10.219494584837545, "grad_norm": 2.2189340591430664, "learning_rate": 4.8902527075812276e-05, "loss": 1.6153, "step": 14154 }, { "epoch": 10.223826714801444, "grad_norm": 2.2881016731262207, "learning_rate": 4.888086642599278e-05, "loss": 1.44, "step": 14160 }, { "epoch": 10.228158844765343, "grad_norm": 2.2663698196411133, "learning_rate": 4.885920577617329e-05, "loss": 1.5967, "step": 14166 }, { "epoch": 10.232490974729242, "grad_norm": 2.349099636077881, "learning_rate": 4.883754512635379e-05, "loss": 1.4604, "step": 14172 }, { "epoch": 10.23682310469314, "grad_norm": 2.0991592407226562, "learning_rate": 4.8815884476534296e-05, "loss": 1.4677, "step": 14178 }, { "epoch": 10.24115523465704, "grad_norm": 2.426813840866089, "learning_rate": 4.87942238267148e-05, "loss": 1.4539, "step": 14184 }, { "epoch": 10.245487364620939, "grad_norm": 2.1847646236419678, "learning_rate": 4.877256317689531e-05, "loss": 1.4608, "step": 14190 }, { "epoch": 10.249819494584838, "grad_norm": 2.277043581008911, "learning_rate": 4.875090252707581e-05, "loss": 1.5169, "step": 14196 }, { "epoch": 10.254151624548737, "grad_norm": 2.1350855827331543, "learning_rate": 4.872924187725632e-05, "loss": 1.4253, "step": 14202 }, { "epoch": 10.258483754512635, "grad_norm": 2.074939489364624, "learning_rate": 4.8707581227436826e-05, "loss": 1.4562, "step": 14208 }, { "epoch": 10.262815884476534, "grad_norm": 2.1449482440948486, "learning_rate": 4.8685920577617336e-05, "loss": 1.5381, "step": 14214 }, { "epoch": 10.267148014440433, "grad_norm": 2.019440174102783, "learning_rate": 4.866425992779783e-05, "loss": 1.4151, "step": 14220 }, { "epoch": 10.271480144404332, "grad_norm": 2.219475746154785, "learning_rate": 4.864259927797834e-05, "loss": 1.5607, "step": 14226 }, { "epoch": 10.275812274368231, "grad_norm": 2.029355525970459, "learning_rate": 4.8620938628158846e-05, "loss": 1.501, "step": 14232 }, { "epoch": 10.28014440433213, "grad_norm": 2.1489741802215576, "learning_rate": 4.8599277978339356e-05, "loss": 1.5974, "step": 14238 }, { "epoch": 10.28447653429603, "grad_norm": 2.0712509155273438, "learning_rate": 4.857761732851986e-05, "loss": 1.5094, "step": 14244 }, { "epoch": 10.288808664259928, "grad_norm": 2.4740278720855713, "learning_rate": 4.855595667870036e-05, "loss": 1.5692, "step": 14250 }, { "epoch": 10.293140794223827, "grad_norm": 1.9916423559188843, "learning_rate": 4.853429602888087e-05, "loss": 1.6092, "step": 14256 }, { "epoch": 10.297472924187726, "grad_norm": 2.1331324577331543, "learning_rate": 4.8512635379061375e-05, "loss": 1.566, "step": 14262 }, { "epoch": 10.301805054151625, "grad_norm": 1.9371330738067627, "learning_rate": 4.849097472924188e-05, "loss": 1.4796, "step": 14268 }, { "epoch": 10.306137184115524, "grad_norm": 2.2808384895324707, "learning_rate": 4.846931407942238e-05, "loss": 1.507, "step": 14274 }, { "epoch": 10.310469314079423, "grad_norm": 2.0304155349731445, "learning_rate": 4.844765342960289e-05, "loss": 1.537, "step": 14280 }, { "epoch": 10.31480144404332, "grad_norm": 2.2128026485443115, "learning_rate": 4.8425992779783395e-05, "loss": 1.5757, "step": 14286 }, { "epoch": 10.31913357400722, "grad_norm": 2.2290079593658447, "learning_rate": 4.8404332129963905e-05, "loss": 1.5695, "step": 14292 }, { "epoch": 10.323465703971118, "grad_norm": 2.403942823410034, "learning_rate": 4.838267148014441e-05, "loss": 1.5852, "step": 14298 }, { "epoch": 10.327797833935017, "grad_norm": 2.418081045150757, "learning_rate": 4.836101083032491e-05, "loss": 1.5038, "step": 14304 }, { "epoch": 10.332129963898916, "grad_norm": 2.372572898864746, "learning_rate": 4.8339350180505415e-05, "loss": 1.6048, "step": 14310 }, { "epoch": 10.336462093862815, "grad_norm": 1.9152146577835083, "learning_rate": 4.8317689530685925e-05, "loss": 1.416, "step": 14316 }, { "epoch": 10.340794223826714, "grad_norm": 2.2300024032592773, "learning_rate": 4.829602888086643e-05, "loss": 1.6113, "step": 14322 }, { "epoch": 10.345126353790613, "grad_norm": 2.2526259422302246, "learning_rate": 4.827436823104694e-05, "loss": 1.4662, "step": 14328 }, { "epoch": 10.349458483754512, "grad_norm": 2.2190520763397217, "learning_rate": 4.825270758122744e-05, "loss": 1.6051, "step": 14334 }, { "epoch": 10.353790613718411, "grad_norm": 2.149430990219116, "learning_rate": 4.8231046931407944e-05, "loss": 1.5232, "step": 14340 }, { "epoch": 10.35812274368231, "grad_norm": 2.261610984802246, "learning_rate": 4.820938628158845e-05, "loss": 1.5806, "step": 14346 }, { "epoch": 10.36245487364621, "grad_norm": 2.2039120197296143, "learning_rate": 4.818772563176895e-05, "loss": 1.5934, "step": 14352 }, { "epoch": 10.366787003610108, "grad_norm": 2.4596381187438965, "learning_rate": 4.816606498194946e-05, "loss": 1.5337, "step": 14358 }, { "epoch": 10.371119133574007, "grad_norm": 2.309692859649658, "learning_rate": 4.8144404332129964e-05, "loss": 1.4678, "step": 14364 }, { "epoch": 10.375451263537906, "grad_norm": 2.160098075866699, "learning_rate": 4.8122743682310474e-05, "loss": 1.494, "step": 14370 }, { "epoch": 10.379783393501805, "grad_norm": 2.2386412620544434, "learning_rate": 4.810108303249098e-05, "loss": 1.5314, "step": 14376 }, { "epoch": 10.384115523465704, "grad_norm": 2.1582720279693604, "learning_rate": 4.807942238267149e-05, "loss": 1.5075, "step": 14382 }, { "epoch": 10.388447653429603, "grad_norm": 2.243748664855957, "learning_rate": 4.8057761732851984e-05, "loss": 1.4555, "step": 14388 }, { "epoch": 10.392779783393502, "grad_norm": 2.001035690307617, "learning_rate": 4.8036101083032494e-05, "loss": 1.513, "step": 14394 }, { "epoch": 10.397111913357401, "grad_norm": 2.2524359226226807, "learning_rate": 4.8014440433213e-05, "loss": 1.5622, "step": 14400 }, { "epoch": 10.4014440433213, "grad_norm": 1.9950720071792603, "learning_rate": 4.799277978339351e-05, "loss": 1.4504, "step": 14406 }, { "epoch": 10.405776173285199, "grad_norm": 2.261556386947632, "learning_rate": 4.797111913357401e-05, "loss": 1.4831, "step": 14412 }, { "epoch": 10.410108303249098, "grad_norm": 2.202493667602539, "learning_rate": 4.794945848375452e-05, "loss": 1.4733, "step": 14418 }, { "epoch": 10.414440433212997, "grad_norm": 2.1312716007232666, "learning_rate": 4.792779783393502e-05, "loss": 1.4621, "step": 14424 }, { "epoch": 10.418772563176896, "grad_norm": 2.3765435218811035, "learning_rate": 4.7906137184115526e-05, "loss": 1.5296, "step": 14430 }, { "epoch": 10.423104693140795, "grad_norm": 2.1512646675109863, "learning_rate": 4.788447653429603e-05, "loss": 1.4733, "step": 14436 }, { "epoch": 10.427436823104694, "grad_norm": 2.183589458465576, "learning_rate": 4.786281588447653e-05, "loss": 1.6072, "step": 14442 }, { "epoch": 10.431768953068593, "grad_norm": 2.1902120113372803, "learning_rate": 4.784115523465704e-05, "loss": 1.4624, "step": 14448 }, { "epoch": 10.43610108303249, "grad_norm": 2.3661692142486572, "learning_rate": 4.7819494584837546e-05, "loss": 1.4674, "step": 14454 }, { "epoch": 10.440433212996389, "grad_norm": 2.146878480911255, "learning_rate": 4.7797833935018056e-05, "loss": 1.5223, "step": 14460 }, { "epoch": 10.444765342960288, "grad_norm": 2.1378908157348633, "learning_rate": 4.777617328519855e-05, "loss": 1.5596, "step": 14466 }, { "epoch": 10.449097472924187, "grad_norm": 2.0796916484832764, "learning_rate": 4.775451263537906e-05, "loss": 1.4857, "step": 14472 }, { "epoch": 10.453429602888086, "grad_norm": 2.2940833568573, "learning_rate": 4.7732851985559566e-05, "loss": 1.4515, "step": 14478 }, { "epoch": 10.457761732851985, "grad_norm": 2.141904354095459, "learning_rate": 4.7711191335740076e-05, "loss": 1.556, "step": 14484 }, { "epoch": 10.462093862815884, "grad_norm": 2.3469555377960205, "learning_rate": 4.768953068592058e-05, "loss": 1.6509, "step": 14490 }, { "epoch": 10.466425992779783, "grad_norm": 2.2802910804748535, "learning_rate": 4.766787003610109e-05, "loss": 1.4606, "step": 14496 }, { "epoch": 10.470758122743682, "grad_norm": 2.1517157554626465, "learning_rate": 4.764620938628159e-05, "loss": 1.6175, "step": 14502 }, { "epoch": 10.47509025270758, "grad_norm": 2.1798458099365234, "learning_rate": 4.7624548736462096e-05, "loss": 1.5565, "step": 14508 }, { "epoch": 10.47942238267148, "grad_norm": 2.286890983581543, "learning_rate": 4.76028880866426e-05, "loss": 1.5415, "step": 14514 }, { "epoch": 10.483754512635379, "grad_norm": 2.038008451461792, "learning_rate": 4.758122743682311e-05, "loss": 1.5035, "step": 14520 }, { "epoch": 10.488086642599278, "grad_norm": 2.220285177230835, "learning_rate": 4.755956678700361e-05, "loss": 1.5053, "step": 14526 }, { "epoch": 10.492418772563177, "grad_norm": 2.2478389739990234, "learning_rate": 4.7537906137184115e-05, "loss": 1.4932, "step": 14532 }, { "epoch": 10.496750902527076, "grad_norm": 2.3517367839813232, "learning_rate": 4.7516245487364625e-05, "loss": 1.5308, "step": 14538 }, { "epoch": 10.501083032490975, "grad_norm": 2.1038191318511963, "learning_rate": 4.749458483754513e-05, "loss": 1.5104, "step": 14544 }, { "epoch": 10.505415162454874, "grad_norm": 2.2862415313720703, "learning_rate": 4.747292418772563e-05, "loss": 1.5217, "step": 14550 }, { "epoch": 10.509747292418773, "grad_norm": 2.264819622039795, "learning_rate": 4.7451263537906135e-05, "loss": 1.5782, "step": 14556 }, { "epoch": 10.514079422382672, "grad_norm": 2.1319563388824463, "learning_rate": 4.7429602888086645e-05, "loss": 1.553, "step": 14562 }, { "epoch": 10.51841155234657, "grad_norm": 2.194624662399292, "learning_rate": 4.740794223826715e-05, "loss": 1.4513, "step": 14568 }, { "epoch": 10.52274368231047, "grad_norm": 2.0167458057403564, "learning_rate": 4.738628158844766e-05, "loss": 1.5998, "step": 14574 }, { "epoch": 10.527075812274369, "grad_norm": 2.1254210472106934, "learning_rate": 4.736462093862816e-05, "loss": 1.542, "step": 14580 }, { "epoch": 10.531407942238268, "grad_norm": 2.160515308380127, "learning_rate": 4.734296028880867e-05, "loss": 1.57, "step": 14586 }, { "epoch": 10.535740072202167, "grad_norm": 2.157892942428589, "learning_rate": 4.732129963898917e-05, "loss": 1.4215, "step": 14592 }, { "epoch": 10.540072202166066, "grad_norm": 2.240523338317871, "learning_rate": 4.729963898916968e-05, "loss": 1.4986, "step": 14598 }, { "epoch": 10.544404332129965, "grad_norm": 2.3580009937286377, "learning_rate": 4.727797833935018e-05, "loss": 1.6088, "step": 14604 }, { "epoch": 10.548736462093864, "grad_norm": 2.3860573768615723, "learning_rate": 4.725631768953069e-05, "loss": 1.6449, "step": 14610 }, { "epoch": 10.553068592057762, "grad_norm": 2.131627321243286, "learning_rate": 4.7234657039711194e-05, "loss": 1.5646, "step": 14616 }, { "epoch": 10.557400722021661, "grad_norm": 2.485670328140259, "learning_rate": 4.72129963898917e-05, "loss": 1.5436, "step": 14622 }, { "epoch": 10.56173285198556, "grad_norm": 2.336902618408203, "learning_rate": 4.719133574007221e-05, "loss": 1.5995, "step": 14628 }, { "epoch": 10.56606498194946, "grad_norm": 2.2055654525756836, "learning_rate": 4.716967509025271e-05, "loss": 1.6548, "step": 14634 }, { "epoch": 10.570397111913357, "grad_norm": 2.2093563079833984, "learning_rate": 4.7148014440433214e-05, "loss": 1.6205, "step": 14640 }, { "epoch": 10.574729241877256, "grad_norm": 1.9775214195251465, "learning_rate": 4.712635379061372e-05, "loss": 1.4921, "step": 14646 }, { "epoch": 10.579061371841155, "grad_norm": 2.43571400642395, "learning_rate": 4.710469314079423e-05, "loss": 1.5104, "step": 14652 }, { "epoch": 10.583393501805054, "grad_norm": 2.392577648162842, "learning_rate": 4.708303249097473e-05, "loss": 1.662, "step": 14658 }, { "epoch": 10.587725631768953, "grad_norm": 1.904500961303711, "learning_rate": 4.706137184115524e-05, "loss": 1.7144, "step": 14664 }, { "epoch": 10.592057761732852, "grad_norm": 2.226088523864746, "learning_rate": 4.7039711191335744e-05, "loss": 1.588, "step": 14670 }, { "epoch": 10.59638989169675, "grad_norm": 2.016425848007202, "learning_rate": 4.701805054151625e-05, "loss": 1.4743, "step": 14676 }, { "epoch": 10.60072202166065, "grad_norm": 2.069317579269409, "learning_rate": 4.699638989169675e-05, "loss": 1.5438, "step": 14682 }, { "epoch": 10.605054151624548, "grad_norm": 2.261540412902832, "learning_rate": 4.697472924187726e-05, "loss": 1.5336, "step": 14688 }, { "epoch": 10.609386281588447, "grad_norm": 2.0085887908935547, "learning_rate": 4.695306859205776e-05, "loss": 1.4842, "step": 14694 }, { "epoch": 10.613718411552346, "grad_norm": 2.1954185962677, "learning_rate": 4.693140794223827e-05, "loss": 1.5123, "step": 14700 }, { "epoch": 10.618050541516245, "grad_norm": 2.4554195404052734, "learning_rate": 4.6909747292418776e-05, "loss": 1.5982, "step": 14706 }, { "epoch": 10.622382671480144, "grad_norm": 2.0059781074523926, "learning_rate": 4.688808664259928e-05, "loss": 1.579, "step": 14712 }, { "epoch": 10.626714801444043, "grad_norm": 2.2027430534362793, "learning_rate": 4.686642599277978e-05, "loss": 1.6012, "step": 14718 }, { "epoch": 10.631046931407942, "grad_norm": 2.389864921569824, "learning_rate": 4.6844765342960286e-05, "loss": 1.4343, "step": 14724 }, { "epoch": 10.635379061371841, "grad_norm": 2.331190347671509, "learning_rate": 4.6823104693140796e-05, "loss": 1.5549, "step": 14730 }, { "epoch": 10.63971119133574, "grad_norm": 2.0765187740325928, "learning_rate": 4.68014440433213e-05, "loss": 1.5407, "step": 14736 }, { "epoch": 10.64404332129964, "grad_norm": 2.1831672191619873, "learning_rate": 4.677978339350181e-05, "loss": 1.5324, "step": 14742 }, { "epoch": 10.648375451263538, "grad_norm": 2.193617582321167, "learning_rate": 4.675812274368231e-05, "loss": 1.6076, "step": 14748 }, { "epoch": 10.652707581227437, "grad_norm": 2.3070316314697266, "learning_rate": 4.673646209386282e-05, "loss": 1.4007, "step": 14754 }, { "epoch": 10.657039711191336, "grad_norm": 2.0610344409942627, "learning_rate": 4.671480144404332e-05, "loss": 1.5536, "step": 14760 }, { "epoch": 10.661371841155235, "grad_norm": 2.0428500175476074, "learning_rate": 4.669314079422383e-05, "loss": 1.5332, "step": 14766 }, { "epoch": 10.665703971119134, "grad_norm": 2.156804323196411, "learning_rate": 4.667148014440433e-05, "loss": 1.5204, "step": 14772 }, { "epoch": 10.670036101083033, "grad_norm": 2.032644510269165, "learning_rate": 4.664981949458484e-05, "loss": 1.5854, "step": 14778 }, { "epoch": 10.674368231046932, "grad_norm": 2.2474727630615234, "learning_rate": 4.6628158844765346e-05, "loss": 1.5727, "step": 14784 }, { "epoch": 10.678700361010831, "grad_norm": 2.4174461364746094, "learning_rate": 4.6606498194945856e-05, "loss": 1.6159, "step": 14790 }, { "epoch": 10.68303249097473, "grad_norm": 2.1526730060577393, "learning_rate": 4.658483754512636e-05, "loss": 1.5033, "step": 14796 }, { "epoch": 10.687364620938627, "grad_norm": 1.9475619792938232, "learning_rate": 4.656317689530686e-05, "loss": 1.4692, "step": 14802 }, { "epoch": 10.691696750902526, "grad_norm": 1.9876539707183838, "learning_rate": 4.6541516245487365e-05, "loss": 1.5212, "step": 14808 }, { "epoch": 10.696028880866425, "grad_norm": 1.7102696895599365, "learning_rate": 4.651985559566787e-05, "loss": 1.5123, "step": 14814 }, { "epoch": 10.700361010830324, "grad_norm": 2.2595696449279785, "learning_rate": 4.649819494584838e-05, "loss": 1.6271, "step": 14820 }, { "epoch": 10.704693140794223, "grad_norm": 2.3320350646972656, "learning_rate": 4.647653429602888e-05, "loss": 1.5026, "step": 14826 }, { "epoch": 10.709025270758122, "grad_norm": 2.0772905349731445, "learning_rate": 4.645487364620939e-05, "loss": 1.5337, "step": 14832 }, { "epoch": 10.713357400722021, "grad_norm": 2.315115451812744, "learning_rate": 4.6433212996389895e-05, "loss": 1.4573, "step": 14838 }, { "epoch": 10.71768953068592, "grad_norm": 2.0279335975646973, "learning_rate": 4.64115523465704e-05, "loss": 1.5277, "step": 14844 }, { "epoch": 10.722021660649819, "grad_norm": 2.1763696670532227, "learning_rate": 4.63898916967509e-05, "loss": 1.5823, "step": 14850 }, { "epoch": 10.726353790613718, "grad_norm": 2.3330092430114746, "learning_rate": 4.636823104693141e-05, "loss": 1.4785, "step": 14856 }, { "epoch": 10.730685920577617, "grad_norm": 2.066272735595703, "learning_rate": 4.6346570397111915e-05, "loss": 1.5107, "step": 14862 }, { "epoch": 10.735018050541516, "grad_norm": 2.1472957134246826, "learning_rate": 4.6324909747292425e-05, "loss": 1.5077, "step": 14868 }, { "epoch": 10.739350180505415, "grad_norm": 2.4863221645355225, "learning_rate": 4.630324909747293e-05, "loss": 1.5165, "step": 14874 }, { "epoch": 10.743682310469314, "grad_norm": 2.064725399017334, "learning_rate": 4.628158844765344e-05, "loss": 1.5571, "step": 14880 }, { "epoch": 10.748014440433213, "grad_norm": 2.3333542346954346, "learning_rate": 4.6259927797833934e-05, "loss": 1.493, "step": 14886 }, { "epoch": 10.752346570397112, "grad_norm": 2.268975257873535, "learning_rate": 4.6238267148014444e-05, "loss": 1.4852, "step": 14892 }, { "epoch": 10.756678700361011, "grad_norm": 1.925182580947876, "learning_rate": 4.621660649819495e-05, "loss": 1.5063, "step": 14898 }, { "epoch": 10.76101083032491, "grad_norm": 2.451669454574585, "learning_rate": 4.619494584837545e-05, "loss": 1.4916, "step": 14904 }, { "epoch": 10.765342960288809, "grad_norm": 2.18601131439209, "learning_rate": 4.617328519855596e-05, "loss": 1.5013, "step": 14910 }, { "epoch": 10.769675090252708, "grad_norm": 2.019944429397583, "learning_rate": 4.6151624548736464e-05, "loss": 1.5969, "step": 14916 }, { "epoch": 10.774007220216607, "grad_norm": 2.2905373573303223, "learning_rate": 4.6129963898916974e-05, "loss": 1.6258, "step": 14922 }, { "epoch": 10.778339350180506, "grad_norm": 2.1885321140289307, "learning_rate": 4.610830324909747e-05, "loss": 1.5251, "step": 14928 }, { "epoch": 10.782671480144405, "grad_norm": 2.0212032794952393, "learning_rate": 4.608664259927798e-05, "loss": 1.5296, "step": 14934 }, { "epoch": 10.787003610108304, "grad_norm": 2.1675620079040527, "learning_rate": 4.6064981949458484e-05, "loss": 1.5413, "step": 14940 }, { "epoch": 10.791335740072203, "grad_norm": 1.9909799098968506, "learning_rate": 4.6043321299638994e-05, "loss": 1.5837, "step": 14946 }, { "epoch": 10.795667870036102, "grad_norm": 2.3776793479919434, "learning_rate": 4.60216606498195e-05, "loss": 1.5619, "step": 14952 }, { "epoch": 10.8, "grad_norm": 2.211695909500122, "learning_rate": 4.600000000000001e-05, "loss": 1.5581, "step": 14958 }, { "epoch": 10.8043321299639, "grad_norm": 2.378602981567383, "learning_rate": 4.59783393501805e-05, "loss": 1.4348, "step": 14964 }, { "epoch": 10.808664259927799, "grad_norm": 2.0610949993133545, "learning_rate": 4.595667870036101e-05, "loss": 1.5238, "step": 14970 }, { "epoch": 10.812996389891698, "grad_norm": 2.5737009048461914, "learning_rate": 4.5935018050541516e-05, "loss": 1.6609, "step": 14976 }, { "epoch": 10.817328519855595, "grad_norm": 2.1638596057891846, "learning_rate": 4.5913357400722026e-05, "loss": 1.5953, "step": 14982 }, { "epoch": 10.821660649819494, "grad_norm": 2.179776906967163, "learning_rate": 4.589169675090253e-05, "loss": 1.4913, "step": 14988 }, { "epoch": 10.825992779783393, "grad_norm": 2.219437599182129, "learning_rate": 4.587003610108303e-05, "loss": 1.613, "step": 14994 }, { "epoch": 10.830324909747292, "grad_norm": 2.3498172760009766, "learning_rate": 4.584837545126354e-05, "loss": 1.589, "step": 15000 }, { "epoch": 10.83465703971119, "grad_norm": 2.4759721755981445, "learning_rate": 4.5826714801444046e-05, "loss": 1.5121, "step": 15006 }, { "epoch": 10.83898916967509, "grad_norm": 2.1936328411102295, "learning_rate": 4.580505415162455e-05, "loss": 1.5868, "step": 15012 }, { "epoch": 10.843321299638989, "grad_norm": 1.9705636501312256, "learning_rate": 4.578339350180505e-05, "loss": 1.4468, "step": 15018 }, { "epoch": 10.847653429602888, "grad_norm": 2.188265085220337, "learning_rate": 4.576173285198556e-05, "loss": 1.5215, "step": 15024 }, { "epoch": 10.851985559566787, "grad_norm": 2.340402603149414, "learning_rate": 4.5740072202166066e-05, "loss": 1.5149, "step": 15030 }, { "epoch": 10.856317689530686, "grad_norm": 2.0141420364379883, "learning_rate": 4.5718411552346576e-05, "loss": 1.6231, "step": 15036 }, { "epoch": 10.860649819494585, "grad_norm": 2.277353286743164, "learning_rate": 4.569675090252708e-05, "loss": 1.4035, "step": 15042 }, { "epoch": 10.864981949458484, "grad_norm": 2.330378532409668, "learning_rate": 4.567509025270758e-05, "loss": 1.5979, "step": 15048 }, { "epoch": 10.869314079422383, "grad_norm": 2.4388082027435303, "learning_rate": 4.5653429602888086e-05, "loss": 1.6202, "step": 15054 }, { "epoch": 10.873646209386282, "grad_norm": 2.4540131092071533, "learning_rate": 4.5631768953068596e-05, "loss": 1.5686, "step": 15060 }, { "epoch": 10.87797833935018, "grad_norm": 2.2919905185699463, "learning_rate": 4.56101083032491e-05, "loss": 1.4361, "step": 15066 }, { "epoch": 10.88231046931408, "grad_norm": 2.231090784072876, "learning_rate": 4.558844765342961e-05, "loss": 1.664, "step": 15072 }, { "epoch": 10.886642599277979, "grad_norm": 2.2143871784210205, "learning_rate": 4.556678700361011e-05, "loss": 1.6361, "step": 15078 }, { "epoch": 10.890974729241877, "grad_norm": 2.17541766166687, "learning_rate": 4.5545126353790615e-05, "loss": 1.522, "step": 15084 }, { "epoch": 10.895306859205776, "grad_norm": 1.9845707416534424, "learning_rate": 4.552346570397112e-05, "loss": 1.5024, "step": 15090 }, { "epoch": 10.899638989169675, "grad_norm": 2.0888164043426514, "learning_rate": 4.550180505415162e-05, "loss": 1.5733, "step": 15096 }, { "epoch": 10.903971119133574, "grad_norm": 2.2577767372131348, "learning_rate": 4.548014440433213e-05, "loss": 1.6968, "step": 15102 }, { "epoch": 10.908303249097473, "grad_norm": 2.215456247329712, "learning_rate": 4.5458483754512635e-05, "loss": 1.5721, "step": 15108 }, { "epoch": 10.912635379061372, "grad_norm": 2.4743905067443848, "learning_rate": 4.5436823104693145e-05, "loss": 1.5302, "step": 15114 }, { "epoch": 10.916967509025271, "grad_norm": 2.572481870651245, "learning_rate": 4.541516245487365e-05, "loss": 1.5198, "step": 15120 }, { "epoch": 10.92129963898917, "grad_norm": 2.727876901626587, "learning_rate": 4.539350180505416e-05, "loss": 1.6291, "step": 15126 }, { "epoch": 10.92563176895307, "grad_norm": 2.2732632160186768, "learning_rate": 4.5371841155234655e-05, "loss": 1.5564, "step": 15132 }, { "epoch": 10.929963898916968, "grad_norm": 2.2454631328582764, "learning_rate": 4.5350180505415165e-05, "loss": 1.5943, "step": 15138 }, { "epoch": 10.934296028880865, "grad_norm": 2.0467138290405273, "learning_rate": 4.532851985559567e-05, "loss": 1.5185, "step": 15144 }, { "epoch": 10.938628158844764, "grad_norm": 2.3243658542633057, "learning_rate": 4.530685920577618e-05, "loss": 1.6004, "step": 15150 }, { "epoch": 10.942960288808663, "grad_norm": 2.3442320823669434, "learning_rate": 4.528519855595668e-05, "loss": 1.642, "step": 15156 }, { "epoch": 10.947292418772562, "grad_norm": 1.9998376369476318, "learning_rate": 4.526353790613719e-05, "loss": 1.4626, "step": 15162 }, { "epoch": 10.951624548736461, "grad_norm": 2.1948249340057373, "learning_rate": 4.5241877256317694e-05, "loss": 1.4938, "step": 15168 }, { "epoch": 10.95595667870036, "grad_norm": 1.7371089458465576, "learning_rate": 4.52202166064982e-05, "loss": 1.3219, "step": 15174 }, { "epoch": 10.96028880866426, "grad_norm": 1.9729772806167603, "learning_rate": 4.51985559566787e-05, "loss": 1.5172, "step": 15180 }, { "epoch": 10.964620938628158, "grad_norm": 2.0509915351867676, "learning_rate": 4.5176895306859204e-05, "loss": 1.5131, "step": 15186 }, { "epoch": 10.968953068592057, "grad_norm": 2.0599217414855957, "learning_rate": 4.5155234657039714e-05, "loss": 1.4427, "step": 15192 }, { "epoch": 10.973285198555956, "grad_norm": 2.2235631942749023, "learning_rate": 4.513357400722022e-05, "loss": 1.5479, "step": 15198 }, { "epoch": 10.977617328519855, "grad_norm": 2.0170695781707764, "learning_rate": 4.511191335740073e-05, "loss": 1.5767, "step": 15204 }, { "epoch": 10.981949458483754, "grad_norm": 2.2101223468780518, "learning_rate": 4.509025270758123e-05, "loss": 1.4957, "step": 15210 }, { "epoch": 10.986281588447653, "grad_norm": 2.0225045680999756, "learning_rate": 4.5068592057761734e-05, "loss": 1.5358, "step": 15216 }, { "epoch": 10.990613718411552, "grad_norm": 2.433509349822998, "learning_rate": 4.504693140794224e-05, "loss": 1.5279, "step": 15222 }, { "epoch": 10.994945848375451, "grad_norm": 2.1479883193969727, "learning_rate": 4.502527075812275e-05, "loss": 1.5726, "step": 15228 }, { "epoch": 10.99927797833935, "grad_norm": 2.12069034576416, "learning_rate": 4.500361010830325e-05, "loss": 1.5868, "step": 15234 }, { "epoch": 11.00361010830325, "grad_norm": 2.1374289989471436, "learning_rate": 4.498194945848376e-05, "loss": 1.4779, "step": 15240 }, { "epoch": 11.007942238267148, "grad_norm": 2.315962553024292, "learning_rate": 4.496028880866426e-05, "loss": 1.4936, "step": 15246 }, { "epoch": 11.012274368231047, "grad_norm": 2.1487767696380615, "learning_rate": 4.493862815884477e-05, "loss": 1.5854, "step": 15252 }, { "epoch": 11.016606498194946, "grad_norm": 2.25765061378479, "learning_rate": 4.491696750902527e-05, "loss": 1.4117, "step": 15258 }, { "epoch": 11.020938628158845, "grad_norm": 2.2447614669799805, "learning_rate": 4.489530685920578e-05, "loss": 1.5477, "step": 15264 }, { "epoch": 11.025270758122744, "grad_norm": 2.0913798809051514, "learning_rate": 4.487364620938628e-05, "loss": 1.5145, "step": 15270 }, { "epoch": 11.029602888086643, "grad_norm": 2.334491491317749, "learning_rate": 4.4851985559566786e-05, "loss": 1.5059, "step": 15276 }, { "epoch": 11.033935018050542, "grad_norm": 2.045881509780884, "learning_rate": 4.4830324909747296e-05, "loss": 1.4712, "step": 15282 }, { "epoch": 11.038267148014441, "grad_norm": 2.3344874382019043, "learning_rate": 4.48086642599278e-05, "loss": 1.5097, "step": 15288 }, { "epoch": 11.04259927797834, "grad_norm": 2.3502137660980225, "learning_rate": 4.478700361010831e-05, "loss": 1.5215, "step": 15294 }, { "epoch": 11.046931407942239, "grad_norm": 2.5239827632904053, "learning_rate": 4.4765342960288806e-05, "loss": 1.4611, "step": 15300 }, { "epoch": 11.051263537906138, "grad_norm": 2.1256630420684814, "learning_rate": 4.4743682310469316e-05, "loss": 1.3788, "step": 15306 }, { "epoch": 11.055595667870037, "grad_norm": 2.1252267360687256, "learning_rate": 4.472202166064982e-05, "loss": 1.5886, "step": 15312 }, { "epoch": 11.059927797833936, "grad_norm": 2.4861669540405273, "learning_rate": 4.470036101083033e-05, "loss": 1.5902, "step": 15318 }, { "epoch": 11.064259927797833, "grad_norm": 2.416262626647949, "learning_rate": 4.467870036101083e-05, "loss": 1.5052, "step": 15324 }, { "epoch": 11.068592057761732, "grad_norm": 2.1898701190948486, "learning_rate": 4.465703971119134e-05, "loss": 1.4501, "step": 15330 }, { "epoch": 11.072924187725631, "grad_norm": 2.505957841873169, "learning_rate": 4.4635379061371846e-05, "loss": 1.2749, "step": 15336 }, { "epoch": 11.07725631768953, "grad_norm": 2.3380439281463623, "learning_rate": 4.461371841155235e-05, "loss": 1.4431, "step": 15342 }, { "epoch": 11.081588447653429, "grad_norm": 2.345998764038086, "learning_rate": 4.459205776173285e-05, "loss": 1.5262, "step": 15348 }, { "epoch": 11.085920577617328, "grad_norm": 2.6114602088928223, "learning_rate": 4.457039711191336e-05, "loss": 1.4853, "step": 15354 }, { "epoch": 11.090252707581227, "grad_norm": 2.4611072540283203, "learning_rate": 4.4548736462093865e-05, "loss": 1.4951, "step": 15360 }, { "epoch": 11.094584837545126, "grad_norm": 2.3082168102264404, "learning_rate": 4.452707581227437e-05, "loss": 1.5798, "step": 15366 }, { "epoch": 11.098916967509025, "grad_norm": 2.6160953044891357, "learning_rate": 4.450541516245488e-05, "loss": 1.5514, "step": 15372 }, { "epoch": 11.103249097472924, "grad_norm": 2.138272523880005, "learning_rate": 4.448375451263538e-05, "loss": 1.4168, "step": 15378 }, { "epoch": 11.107581227436823, "grad_norm": 2.640043020248413, "learning_rate": 4.4462093862815885e-05, "loss": 1.4768, "step": 15384 }, { "epoch": 11.111913357400722, "grad_norm": 2.5113210678100586, "learning_rate": 4.444043321299639e-05, "loss": 1.412, "step": 15390 }, { "epoch": 11.11624548736462, "grad_norm": 2.515587329864502, "learning_rate": 4.44187725631769e-05, "loss": 1.6086, "step": 15396 }, { "epoch": 11.12057761732852, "grad_norm": 2.325913429260254, "learning_rate": 4.43971119133574e-05, "loss": 1.6003, "step": 15402 }, { "epoch": 11.124909747292419, "grad_norm": 2.2703311443328857, "learning_rate": 4.437545126353791e-05, "loss": 1.5787, "step": 15408 }, { "epoch": 11.129241877256318, "grad_norm": 2.234112024307251, "learning_rate": 4.4353790613718415e-05, "loss": 1.5046, "step": 15414 }, { "epoch": 11.133574007220217, "grad_norm": 2.218562602996826, "learning_rate": 4.4332129963898925e-05, "loss": 1.5648, "step": 15420 }, { "epoch": 11.137906137184116, "grad_norm": 1.9541754722595215, "learning_rate": 4.431046931407942e-05, "loss": 1.4895, "step": 15426 }, { "epoch": 11.142238267148015, "grad_norm": 2.070208787918091, "learning_rate": 4.428880866425993e-05, "loss": 1.4545, "step": 15432 }, { "epoch": 11.146570397111914, "grad_norm": 2.3400328159332275, "learning_rate": 4.4267148014440434e-05, "loss": 1.5221, "step": 15438 }, { "epoch": 11.150902527075813, "grad_norm": 2.094608783721924, "learning_rate": 4.4245487364620944e-05, "loss": 1.4026, "step": 15444 }, { "epoch": 11.155234657039712, "grad_norm": 2.3299169540405273, "learning_rate": 4.422382671480145e-05, "loss": 1.587, "step": 15450 }, { "epoch": 11.15956678700361, "grad_norm": 2.3144798278808594, "learning_rate": 4.420216606498195e-05, "loss": 1.6061, "step": 15456 }, { "epoch": 11.16389891696751, "grad_norm": 2.173572063446045, "learning_rate": 4.4180505415162454e-05, "loss": 1.477, "step": 15462 }, { "epoch": 11.168231046931409, "grad_norm": 2.0429654121398926, "learning_rate": 4.415884476534296e-05, "loss": 1.5118, "step": 15468 }, { "epoch": 11.172563176895308, "grad_norm": 2.2627367973327637, "learning_rate": 4.413718411552347e-05, "loss": 1.5021, "step": 15474 }, { "epoch": 11.176895306859207, "grad_norm": 2.1072797775268555, "learning_rate": 4.411552346570397e-05, "loss": 1.3537, "step": 15480 }, { "epoch": 11.181227436823105, "grad_norm": 2.1526246070861816, "learning_rate": 4.409386281588448e-05, "loss": 1.5332, "step": 15486 }, { "epoch": 11.185559566787004, "grad_norm": 2.2886269092559814, "learning_rate": 4.4072202166064984e-05, "loss": 1.5667, "step": 15492 }, { "epoch": 11.189891696750902, "grad_norm": 2.378934621810913, "learning_rate": 4.4050541516245494e-05, "loss": 1.5079, "step": 15498 }, { "epoch": 11.1942238267148, "grad_norm": 2.5023727416992188, "learning_rate": 4.402888086642599e-05, "loss": 1.4649, "step": 15504 }, { "epoch": 11.1985559566787, "grad_norm": 2.3180222511291504, "learning_rate": 4.40072202166065e-05, "loss": 1.364, "step": 15510 }, { "epoch": 11.202888086642599, "grad_norm": 2.272477865219116, "learning_rate": 4.3985559566787e-05, "loss": 1.5352, "step": 15516 }, { "epoch": 11.207220216606498, "grad_norm": 2.335594654083252, "learning_rate": 4.396389891696751e-05, "loss": 1.4869, "step": 15522 }, { "epoch": 11.211552346570397, "grad_norm": 2.0763213634490967, "learning_rate": 4.3942238267148016e-05, "loss": 1.5232, "step": 15528 }, { "epoch": 11.215884476534296, "grad_norm": 2.109048366546631, "learning_rate": 4.3920577617328526e-05, "loss": 1.4703, "step": 15534 }, { "epoch": 11.220216606498195, "grad_norm": 2.1441471576690674, "learning_rate": 4.389891696750903e-05, "loss": 1.4665, "step": 15540 }, { "epoch": 11.224548736462094, "grad_norm": 2.1786439418792725, "learning_rate": 4.387725631768953e-05, "loss": 1.4546, "step": 15546 }, { "epoch": 11.228880866425992, "grad_norm": 2.4305412769317627, "learning_rate": 4.3855595667870036e-05, "loss": 1.4393, "step": 15552 }, { "epoch": 11.233212996389891, "grad_norm": 2.3540971279144287, "learning_rate": 4.383393501805054e-05, "loss": 1.4694, "step": 15558 }, { "epoch": 11.23754512635379, "grad_norm": 2.1233770847320557, "learning_rate": 4.381227436823105e-05, "loss": 1.5326, "step": 15564 }, { "epoch": 11.24187725631769, "grad_norm": 2.223475933074951, "learning_rate": 4.379061371841155e-05, "loss": 1.6594, "step": 15570 }, { "epoch": 11.246209386281588, "grad_norm": 2.351036787033081, "learning_rate": 4.376895306859206e-05, "loss": 1.4773, "step": 15576 }, { "epoch": 11.250541516245487, "grad_norm": 2.423201560974121, "learning_rate": 4.3747292418772566e-05, "loss": 1.5332, "step": 15582 }, { "epoch": 11.254873646209386, "grad_norm": 2.0613913536071777, "learning_rate": 4.372563176895307e-05, "loss": 1.4869, "step": 15588 }, { "epoch": 11.259205776173285, "grad_norm": 2.2268927097320557, "learning_rate": 4.370397111913357e-05, "loss": 1.5267, "step": 15594 }, { "epoch": 11.263537906137184, "grad_norm": 2.3538248538970947, "learning_rate": 4.368231046931408e-05, "loss": 1.4366, "step": 15600 }, { "epoch": 11.267870036101083, "grad_norm": 2.2724766731262207, "learning_rate": 4.3660649819494586e-05, "loss": 1.6545, "step": 15606 }, { "epoch": 11.272202166064982, "grad_norm": 2.2519938945770264, "learning_rate": 4.3638989169675096e-05, "loss": 1.6246, "step": 15612 }, { "epoch": 11.276534296028881, "grad_norm": 2.1057333946228027, "learning_rate": 4.36173285198556e-05, "loss": 1.4555, "step": 15618 }, { "epoch": 11.28086642599278, "grad_norm": 2.656663656234741, "learning_rate": 4.359566787003611e-05, "loss": 1.4168, "step": 15624 }, { "epoch": 11.28519855595668, "grad_norm": 2.3049299716949463, "learning_rate": 4.3574007220216605e-05, "loss": 1.5977, "step": 15630 }, { "epoch": 11.289530685920578, "grad_norm": 2.5886118412017822, "learning_rate": 4.3552346570397115e-05, "loss": 1.5583, "step": 15636 }, { "epoch": 11.293862815884477, "grad_norm": 2.234684467315674, "learning_rate": 4.353068592057762e-05, "loss": 1.4811, "step": 15642 }, { "epoch": 11.298194945848376, "grad_norm": 1.9752253293991089, "learning_rate": 4.350902527075812e-05, "loss": 1.5534, "step": 15648 }, { "epoch": 11.302527075812275, "grad_norm": 2.204495668411255, "learning_rate": 4.348736462093863e-05, "loss": 1.4766, "step": 15654 }, { "epoch": 11.306859205776174, "grad_norm": 2.302417755126953, "learning_rate": 4.3465703971119135e-05, "loss": 1.404, "step": 15660 }, { "epoch": 11.311191335740073, "grad_norm": 2.1613476276397705, "learning_rate": 4.3444043321299645e-05, "loss": 1.5372, "step": 15666 }, { "epoch": 11.31552346570397, "grad_norm": 2.3813555240631104, "learning_rate": 4.342238267148014e-05, "loss": 1.4994, "step": 15672 }, { "epoch": 11.31985559566787, "grad_norm": 1.9674543142318726, "learning_rate": 4.340072202166065e-05, "loss": 1.4832, "step": 15678 }, { "epoch": 11.324187725631768, "grad_norm": 2.654719114303589, "learning_rate": 4.3379061371841155e-05, "loss": 1.4948, "step": 15684 }, { "epoch": 11.328519855595667, "grad_norm": 2.405287742614746, "learning_rate": 4.3357400722021665e-05, "loss": 1.4797, "step": 15690 }, { "epoch": 11.332851985559566, "grad_norm": 2.1810829639434814, "learning_rate": 4.333574007220217e-05, "loss": 1.6421, "step": 15696 }, { "epoch": 11.337184115523465, "grad_norm": 2.340005397796631, "learning_rate": 4.331407942238268e-05, "loss": 1.4552, "step": 15702 }, { "epoch": 11.341516245487364, "grad_norm": 2.083414077758789, "learning_rate": 4.329241877256318e-05, "loss": 1.5849, "step": 15708 }, { "epoch": 11.345848375451263, "grad_norm": 2.240737199783325, "learning_rate": 4.3270758122743684e-05, "loss": 1.43, "step": 15714 }, { "epoch": 11.350180505415162, "grad_norm": 2.3946454524993896, "learning_rate": 4.324909747292419e-05, "loss": 1.5706, "step": 15720 }, { "epoch": 11.354512635379061, "grad_norm": 2.4826531410217285, "learning_rate": 4.32274368231047e-05, "loss": 1.4589, "step": 15726 }, { "epoch": 11.35884476534296, "grad_norm": 2.574298858642578, "learning_rate": 4.32057761732852e-05, "loss": 1.5459, "step": 15732 }, { "epoch": 11.363176895306859, "grad_norm": 2.4555885791778564, "learning_rate": 4.3184115523465704e-05, "loss": 1.5861, "step": 15738 }, { "epoch": 11.367509025270758, "grad_norm": 2.2050609588623047, "learning_rate": 4.3162454873646214e-05, "loss": 1.5238, "step": 15744 }, { "epoch": 11.371841155234657, "grad_norm": 2.1177525520324707, "learning_rate": 4.314079422382672e-05, "loss": 1.5372, "step": 15750 }, { "epoch": 11.376173285198556, "grad_norm": 1.9905301332473755, "learning_rate": 4.311913357400722e-05, "loss": 1.5107, "step": 15756 }, { "epoch": 11.380505415162455, "grad_norm": 2.3256425857543945, "learning_rate": 4.3097472924187724e-05, "loss": 1.5032, "step": 15762 }, { "epoch": 11.384837545126354, "grad_norm": 1.9738303422927856, "learning_rate": 4.3075812274368234e-05, "loss": 1.4525, "step": 15768 }, { "epoch": 11.389169675090253, "grad_norm": 2.4473276138305664, "learning_rate": 4.305415162454874e-05, "loss": 1.5666, "step": 15774 }, { "epoch": 11.393501805054152, "grad_norm": 2.5174367427825928, "learning_rate": 4.303249097472925e-05, "loss": 1.5253, "step": 15780 }, { "epoch": 11.39783393501805, "grad_norm": 2.013636827468872, "learning_rate": 4.301083032490975e-05, "loss": 1.4658, "step": 15786 }, { "epoch": 11.40216606498195, "grad_norm": 1.9617146253585815, "learning_rate": 4.298916967509026e-05, "loss": 1.6329, "step": 15792 }, { "epoch": 11.406498194945849, "grad_norm": 2.427856922149658, "learning_rate": 4.2967509025270756e-05, "loss": 1.4999, "step": 15798 }, { "epoch": 11.410830324909748, "grad_norm": 2.1759066581726074, "learning_rate": 4.2945848375451266e-05, "loss": 1.4862, "step": 15804 }, { "epoch": 11.415162454873647, "grad_norm": 2.3450963497161865, "learning_rate": 4.292418772563177e-05, "loss": 1.4922, "step": 15810 }, { "epoch": 11.419494584837546, "grad_norm": 2.1881420612335205, "learning_rate": 4.290252707581228e-05, "loss": 1.4891, "step": 15816 }, { "epoch": 11.423826714801445, "grad_norm": 2.3364782333374023, "learning_rate": 4.288086642599278e-05, "loss": 1.5469, "step": 15822 }, { "epoch": 11.428158844765344, "grad_norm": 2.300335168838501, "learning_rate": 4.2859205776173286e-05, "loss": 1.5234, "step": 15828 }, { "epoch": 11.432490974729243, "grad_norm": 2.3899519443511963, "learning_rate": 4.2837545126353796e-05, "loss": 1.5343, "step": 15834 }, { "epoch": 11.43682310469314, "grad_norm": 2.2038729190826416, "learning_rate": 4.281588447653429e-05, "loss": 1.5948, "step": 15840 }, { "epoch": 11.441155234657039, "grad_norm": 2.3149425983428955, "learning_rate": 4.27942238267148e-05, "loss": 1.5276, "step": 15846 }, { "epoch": 11.445487364620938, "grad_norm": 2.43754243850708, "learning_rate": 4.2772563176895306e-05, "loss": 1.6963, "step": 15852 }, { "epoch": 11.449819494584837, "grad_norm": 2.266249179840088, "learning_rate": 4.2750902527075816e-05, "loss": 1.5067, "step": 15858 }, { "epoch": 11.454151624548736, "grad_norm": 2.3913543224334717, "learning_rate": 4.272924187725632e-05, "loss": 1.5928, "step": 15864 }, { "epoch": 11.458483754512635, "grad_norm": 2.4679126739501953, "learning_rate": 4.270758122743683e-05, "loss": 1.5493, "step": 15870 }, { "epoch": 11.462815884476534, "grad_norm": 2.0237786769866943, "learning_rate": 4.268592057761733e-05, "loss": 1.4547, "step": 15876 }, { "epoch": 11.467148014440433, "grad_norm": 2.241983652114868, "learning_rate": 4.2664259927797836e-05, "loss": 1.5985, "step": 15882 }, { "epoch": 11.471480144404332, "grad_norm": 2.19942307472229, "learning_rate": 4.264259927797834e-05, "loss": 1.5299, "step": 15888 }, { "epoch": 11.47581227436823, "grad_norm": 2.3135366439819336, "learning_rate": 4.262093862815885e-05, "loss": 1.589, "step": 15894 }, { "epoch": 11.48014440433213, "grad_norm": 2.3352584838867188, "learning_rate": 4.259927797833935e-05, "loss": 1.5883, "step": 15900 }, { "epoch": 11.484476534296029, "grad_norm": 2.291890859603882, "learning_rate": 4.257761732851986e-05, "loss": 1.4795, "step": 15906 }, { "epoch": 11.488808664259928, "grad_norm": 2.373717784881592, "learning_rate": 4.2555956678700365e-05, "loss": 1.5638, "step": 15912 }, { "epoch": 11.493140794223827, "grad_norm": 2.4225003719329834, "learning_rate": 4.253429602888087e-05, "loss": 1.529, "step": 15918 }, { "epoch": 11.497472924187726, "grad_norm": 2.123438835144043, "learning_rate": 4.251263537906137e-05, "loss": 1.5343, "step": 15924 }, { "epoch": 11.501805054151625, "grad_norm": 2.2829504013061523, "learning_rate": 4.2490974729241875e-05, "loss": 1.4629, "step": 15930 }, { "epoch": 11.506137184115524, "grad_norm": 2.516819715499878, "learning_rate": 4.2469314079422385e-05, "loss": 1.5566, "step": 15936 }, { "epoch": 11.510469314079423, "grad_norm": 1.9219509363174438, "learning_rate": 4.244765342960289e-05, "loss": 1.5728, "step": 15942 }, { "epoch": 11.514801444043322, "grad_norm": 1.9886771440505981, "learning_rate": 4.24259927797834e-05, "loss": 1.5238, "step": 15948 }, { "epoch": 11.51913357400722, "grad_norm": 2.363318920135498, "learning_rate": 4.24043321299639e-05, "loss": 1.4716, "step": 15954 }, { "epoch": 11.52346570397112, "grad_norm": 2.2942428588867188, "learning_rate": 4.238267148014441e-05, "loss": 1.5195, "step": 15960 }, { "epoch": 11.527797833935018, "grad_norm": 2.4192826747894287, "learning_rate": 4.236101083032491e-05, "loss": 1.5166, "step": 15966 }, { "epoch": 11.532129963898917, "grad_norm": 2.370823860168457, "learning_rate": 4.233935018050542e-05, "loss": 1.5091, "step": 15972 }, { "epoch": 11.536462093862816, "grad_norm": 2.5061731338500977, "learning_rate": 4.231768953068592e-05, "loss": 1.5791, "step": 15978 }, { "epoch": 11.540794223826715, "grad_norm": 1.9836598634719849, "learning_rate": 4.229602888086643e-05, "loss": 1.4992, "step": 15984 }, { "epoch": 11.545126353790614, "grad_norm": 2.0596415996551514, "learning_rate": 4.2274368231046934e-05, "loss": 1.4134, "step": 15990 }, { "epoch": 11.549458483754513, "grad_norm": 2.3466851711273193, "learning_rate": 4.2252707581227444e-05, "loss": 1.5384, "step": 15996 }, { "epoch": 11.553790613718412, "grad_norm": 2.5305137634277344, "learning_rate": 4.223104693140794e-05, "loss": 1.5287, "step": 16002 }, { "epoch": 11.558122743682311, "grad_norm": 2.284517526626587, "learning_rate": 4.220938628158845e-05, "loss": 1.6115, "step": 16008 }, { "epoch": 11.56245487364621, "grad_norm": 2.3423550128936768, "learning_rate": 4.2187725631768954e-05, "loss": 1.3774, "step": 16014 }, { "epoch": 11.566787003610107, "grad_norm": 2.326418161392212, "learning_rate": 4.216606498194946e-05, "loss": 1.6033, "step": 16020 }, { "epoch": 11.571119133574006, "grad_norm": 2.348857879638672, "learning_rate": 4.214440433212997e-05, "loss": 1.4876, "step": 16026 }, { "epoch": 11.575451263537905, "grad_norm": 2.1456034183502197, "learning_rate": 4.212274368231047e-05, "loss": 1.4181, "step": 16032 }, { "epoch": 11.579783393501804, "grad_norm": 2.185635566711426, "learning_rate": 4.210108303249098e-05, "loss": 1.462, "step": 16038 }, { "epoch": 11.584115523465703, "grad_norm": 2.043097734451294, "learning_rate": 4.207942238267148e-05, "loss": 1.4851, "step": 16044 }, { "epoch": 11.588447653429602, "grad_norm": 2.6713595390319824, "learning_rate": 4.205776173285199e-05, "loss": 1.55, "step": 16050 }, { "epoch": 11.592779783393501, "grad_norm": 2.3096189498901367, "learning_rate": 4.203610108303249e-05, "loss": 1.6498, "step": 16056 }, { "epoch": 11.5971119133574, "grad_norm": 2.464040517807007, "learning_rate": 4.2014440433213e-05, "loss": 1.5445, "step": 16062 }, { "epoch": 11.6014440433213, "grad_norm": 2.30250883102417, "learning_rate": 4.19927797833935e-05, "loss": 1.639, "step": 16068 }, { "epoch": 11.605776173285198, "grad_norm": 2.2857167720794678, "learning_rate": 4.197111913357401e-05, "loss": 1.4283, "step": 16074 }, { "epoch": 11.610108303249097, "grad_norm": 2.087151527404785, "learning_rate": 4.1949458483754516e-05, "loss": 1.4869, "step": 16080 }, { "epoch": 11.614440433212996, "grad_norm": 2.2440643310546875, "learning_rate": 4.192779783393502e-05, "loss": 1.4737, "step": 16086 }, { "epoch": 11.618772563176895, "grad_norm": 2.140232801437378, "learning_rate": 4.190613718411552e-05, "loss": 1.554, "step": 16092 }, { "epoch": 11.623104693140794, "grad_norm": 2.4576361179351807, "learning_rate": 4.188447653429603e-05, "loss": 1.5099, "step": 16098 }, { "epoch": 11.627436823104693, "grad_norm": 2.2120659351348877, "learning_rate": 4.1862815884476536e-05, "loss": 1.4677, "step": 16104 }, { "epoch": 11.631768953068592, "grad_norm": 2.1863322257995605, "learning_rate": 4.184115523465704e-05, "loss": 1.5178, "step": 16110 }, { "epoch": 11.636101083032491, "grad_norm": 2.261352062225342, "learning_rate": 4.181949458483755e-05, "loss": 1.4494, "step": 16116 }, { "epoch": 11.64043321299639, "grad_norm": 2.6109344959259033, "learning_rate": 4.179783393501805e-05, "loss": 1.6287, "step": 16122 }, { "epoch": 11.644765342960289, "grad_norm": 2.363722801208496, "learning_rate": 4.1776173285198556e-05, "loss": 1.4226, "step": 16128 }, { "epoch": 11.649097472924188, "grad_norm": 2.3942453861236572, "learning_rate": 4.175451263537906e-05, "loss": 1.4338, "step": 16134 }, { "epoch": 11.653429602888087, "grad_norm": 2.342080593109131, "learning_rate": 4.173285198555957e-05, "loss": 1.4993, "step": 16140 }, { "epoch": 11.657761732851986, "grad_norm": 2.2077221870422363, "learning_rate": 4.171119133574007e-05, "loss": 1.5108, "step": 16146 }, { "epoch": 11.662093862815885, "grad_norm": 2.0875508785247803, "learning_rate": 4.168953068592058e-05, "loss": 1.4382, "step": 16152 }, { "epoch": 11.666425992779784, "grad_norm": 2.068708658218384, "learning_rate": 4.1667870036101086e-05, "loss": 1.529, "step": 16158 }, { "epoch": 11.670758122743683, "grad_norm": 2.303392171859741, "learning_rate": 4.1646209386281596e-05, "loss": 1.4666, "step": 16164 }, { "epoch": 11.675090252707582, "grad_norm": 2.2947206497192383, "learning_rate": 4.162454873646209e-05, "loss": 1.6981, "step": 16170 }, { "epoch": 11.679422382671481, "grad_norm": 2.106077194213867, "learning_rate": 4.16028880866426e-05, "loss": 1.6258, "step": 16176 }, { "epoch": 11.683754512635378, "grad_norm": 2.257936716079712, "learning_rate": 4.1581227436823105e-05, "loss": 1.4419, "step": 16182 }, { "epoch": 11.688086642599277, "grad_norm": 2.0072076320648193, "learning_rate": 4.1559566787003615e-05, "loss": 1.383, "step": 16188 }, { "epoch": 11.692418772563176, "grad_norm": 2.3076629638671875, "learning_rate": 4.153790613718412e-05, "loss": 1.5833, "step": 16194 }, { "epoch": 11.696750902527075, "grad_norm": 2.3034493923187256, "learning_rate": 4.151624548736462e-05, "loss": 1.5137, "step": 16200 }, { "epoch": 11.701083032490974, "grad_norm": 2.0517067909240723, "learning_rate": 4.149458483754513e-05, "loss": 1.4668, "step": 16206 }, { "epoch": 11.705415162454873, "grad_norm": 2.5194671154022217, "learning_rate": 4.147292418772563e-05, "loss": 1.5466, "step": 16212 }, { "epoch": 11.709747292418772, "grad_norm": 2.1545422077178955, "learning_rate": 4.145126353790614e-05, "loss": 1.4959, "step": 16218 }, { "epoch": 11.714079422382671, "grad_norm": 2.197841167449951, "learning_rate": 4.142960288808664e-05, "loss": 1.4723, "step": 16224 }, { "epoch": 11.71841155234657, "grad_norm": 2.4157893657684326, "learning_rate": 4.140794223826715e-05, "loss": 1.6583, "step": 16230 }, { "epoch": 11.722743682310469, "grad_norm": 2.560800313949585, "learning_rate": 4.1386281588447655e-05, "loss": 1.4512, "step": 16236 }, { "epoch": 11.727075812274368, "grad_norm": 1.909978985786438, "learning_rate": 4.1364620938628165e-05, "loss": 1.4832, "step": 16242 }, { "epoch": 11.731407942238267, "grad_norm": 2.4167137145996094, "learning_rate": 4.134296028880867e-05, "loss": 1.4591, "step": 16248 }, { "epoch": 11.735740072202166, "grad_norm": 2.065486192703247, "learning_rate": 4.132129963898917e-05, "loss": 1.4684, "step": 16254 }, { "epoch": 11.740072202166065, "grad_norm": 2.3221023082733154, "learning_rate": 4.1299638989169674e-05, "loss": 1.4521, "step": 16260 }, { "epoch": 11.744404332129964, "grad_norm": 2.3037092685699463, "learning_rate": 4.1277978339350184e-05, "loss": 1.5635, "step": 16266 }, { "epoch": 11.748736462093863, "grad_norm": 2.2895114421844482, "learning_rate": 4.125631768953069e-05, "loss": 1.5481, "step": 16272 }, { "epoch": 11.753068592057762, "grad_norm": 2.2859103679656982, "learning_rate": 4.12346570397112e-05, "loss": 1.5632, "step": 16278 }, { "epoch": 11.75740072202166, "grad_norm": 2.192972421646118, "learning_rate": 4.12129963898917e-05, "loss": 1.5171, "step": 16284 }, { "epoch": 11.76173285198556, "grad_norm": 2.2066755294799805, "learning_rate": 4.1191335740072204e-05, "loss": 1.5626, "step": 16290 }, { "epoch": 11.766064981949459, "grad_norm": 2.2874319553375244, "learning_rate": 4.116967509025271e-05, "loss": 1.5101, "step": 16296 }, { "epoch": 11.770397111913358, "grad_norm": 1.9382129907608032, "learning_rate": 4.114801444043321e-05, "loss": 1.4293, "step": 16302 }, { "epoch": 11.774729241877257, "grad_norm": 2.54685640335083, "learning_rate": 4.112635379061372e-05, "loss": 1.6579, "step": 16308 }, { "epoch": 11.779061371841156, "grad_norm": 2.6492302417755127, "learning_rate": 4.1104693140794224e-05, "loss": 1.5432, "step": 16314 }, { "epoch": 11.783393501805055, "grad_norm": 2.069462776184082, "learning_rate": 4.1083032490974734e-05, "loss": 1.5094, "step": 16320 }, { "epoch": 11.787725631768954, "grad_norm": 2.420814037322998, "learning_rate": 4.106137184115524e-05, "loss": 1.4918, "step": 16326 }, { "epoch": 11.792057761732853, "grad_norm": 2.334379196166992, "learning_rate": 4.103971119133575e-05, "loss": 1.5528, "step": 16332 }, { "epoch": 11.796389891696752, "grad_norm": 2.104404926300049, "learning_rate": 4.101805054151624e-05, "loss": 1.4861, "step": 16338 }, { "epoch": 11.80072202166065, "grad_norm": 2.083563804626465, "learning_rate": 4.099638989169675e-05, "loss": 1.5357, "step": 16344 }, { "epoch": 11.80505415162455, "grad_norm": 2.231351375579834, "learning_rate": 4.0974729241877256e-05, "loss": 1.4473, "step": 16350 }, { "epoch": 11.809386281588448, "grad_norm": 2.4597725868225098, "learning_rate": 4.0953068592057766e-05, "loss": 1.4417, "step": 16356 }, { "epoch": 11.813718411552347, "grad_norm": 2.2959349155426025, "learning_rate": 4.093140794223827e-05, "loss": 1.5014, "step": 16362 }, { "epoch": 11.818050541516245, "grad_norm": 2.106027841567993, "learning_rate": 4.090974729241878e-05, "loss": 1.4395, "step": 16368 }, { "epoch": 11.822382671480144, "grad_norm": 2.264627695083618, "learning_rate": 4.088808664259928e-05, "loss": 1.5575, "step": 16374 }, { "epoch": 11.826714801444043, "grad_norm": 2.3997995853424072, "learning_rate": 4.0866425992779786e-05, "loss": 1.5603, "step": 16380 }, { "epoch": 11.831046931407942, "grad_norm": 2.1090292930603027, "learning_rate": 4.084476534296029e-05, "loss": 1.5471, "step": 16386 }, { "epoch": 11.83537906137184, "grad_norm": 2.1749658584594727, "learning_rate": 4.082310469314079e-05, "loss": 1.4751, "step": 16392 }, { "epoch": 11.83971119133574, "grad_norm": 2.5059728622436523, "learning_rate": 4.08014440433213e-05, "loss": 1.4696, "step": 16398 }, { "epoch": 11.844043321299639, "grad_norm": 2.5433342456817627, "learning_rate": 4.0779783393501806e-05, "loss": 1.5069, "step": 16404 }, { "epoch": 11.848375451263538, "grad_norm": 2.333730936050415, "learning_rate": 4.0758122743682316e-05, "loss": 1.6446, "step": 16410 }, { "epoch": 11.852707581227437, "grad_norm": 2.230637311935425, "learning_rate": 4.073646209386282e-05, "loss": 1.4372, "step": 16416 }, { "epoch": 11.857039711191335, "grad_norm": 2.515770673751831, "learning_rate": 4.071480144404332e-05, "loss": 1.6506, "step": 16422 }, { "epoch": 11.861371841155234, "grad_norm": 2.3975555896759033, "learning_rate": 4.0693140794223826e-05, "loss": 1.5987, "step": 16428 }, { "epoch": 11.865703971119133, "grad_norm": 2.3005967140197754, "learning_rate": 4.0671480144404336e-05, "loss": 1.5701, "step": 16434 }, { "epoch": 11.870036101083032, "grad_norm": 2.421525716781616, "learning_rate": 4.064981949458484e-05, "loss": 1.5406, "step": 16440 }, { "epoch": 11.874368231046931, "grad_norm": 2.1835169792175293, "learning_rate": 4.062815884476535e-05, "loss": 1.5655, "step": 16446 }, { "epoch": 11.87870036101083, "grad_norm": 2.12741756439209, "learning_rate": 4.060649819494585e-05, "loss": 1.5241, "step": 16452 }, { "epoch": 11.88303249097473, "grad_norm": 2.320248603820801, "learning_rate": 4.0584837545126355e-05, "loss": 1.5671, "step": 16458 }, { "epoch": 11.887364620938628, "grad_norm": 2.3053953647613525, "learning_rate": 4.056317689530686e-05, "loss": 1.593, "step": 16464 }, { "epoch": 11.891696750902527, "grad_norm": 2.5872857570648193, "learning_rate": 4.054151624548737e-05, "loss": 1.4963, "step": 16470 }, { "epoch": 11.896028880866426, "grad_norm": 2.4326798915863037, "learning_rate": 4.051985559566787e-05, "loss": 1.5248, "step": 16476 }, { "epoch": 11.900361010830325, "grad_norm": 2.3197991847991943, "learning_rate": 4.0498194945848375e-05, "loss": 1.5839, "step": 16482 }, { "epoch": 11.904693140794224, "grad_norm": 2.468214988708496, "learning_rate": 4.0476534296028885e-05, "loss": 1.5538, "step": 16488 }, { "epoch": 11.909025270758123, "grad_norm": 2.2990100383758545, "learning_rate": 4.045487364620939e-05, "loss": 1.5118, "step": 16494 }, { "epoch": 11.913357400722022, "grad_norm": 2.3069663047790527, "learning_rate": 4.043321299638989e-05, "loss": 1.4328, "step": 16500 }, { "epoch": 11.917689530685921, "grad_norm": 2.304384231567383, "learning_rate": 4.0411552346570395e-05, "loss": 1.4429, "step": 16506 }, { "epoch": 11.92202166064982, "grad_norm": 2.2518324851989746, "learning_rate": 4.0389891696750905e-05, "loss": 1.4411, "step": 16512 }, { "epoch": 11.92635379061372, "grad_norm": 2.35833477973938, "learning_rate": 4.036823104693141e-05, "loss": 1.5125, "step": 16518 }, { "epoch": 11.930685920577618, "grad_norm": 2.3284249305725098, "learning_rate": 4.034657039711192e-05, "loss": 1.4843, "step": 16524 }, { "epoch": 11.935018050541515, "grad_norm": 2.5053110122680664, "learning_rate": 4.032490974729242e-05, "loss": 1.5219, "step": 16530 }, { "epoch": 11.939350180505414, "grad_norm": 2.4999496936798096, "learning_rate": 4.030324909747293e-05, "loss": 1.5272, "step": 16536 }, { "epoch": 11.943682310469313, "grad_norm": 2.1994786262512207, "learning_rate": 4.028158844765343e-05, "loss": 1.5197, "step": 16542 }, { "epoch": 11.948014440433212, "grad_norm": 2.2162044048309326, "learning_rate": 4.025992779783394e-05, "loss": 1.5034, "step": 16548 }, { "epoch": 11.952346570397111, "grad_norm": 2.2570700645446777, "learning_rate": 4.023826714801444e-05, "loss": 1.4575, "step": 16554 }, { "epoch": 11.95667870036101, "grad_norm": 2.0523617267608643, "learning_rate": 4.021660649819495e-05, "loss": 1.378, "step": 16560 }, { "epoch": 11.96101083032491, "grad_norm": 2.099937915802002, "learning_rate": 4.0194945848375454e-05, "loss": 1.4337, "step": 16566 }, { "epoch": 11.965342960288808, "grad_norm": 2.1047236919403076, "learning_rate": 4.017328519855596e-05, "loss": 1.5949, "step": 16572 }, { "epoch": 11.969675090252707, "grad_norm": 2.042451858520508, "learning_rate": 4.015162454873647e-05, "loss": 1.4881, "step": 16578 }, { "epoch": 11.974007220216606, "grad_norm": 1.8301827907562256, "learning_rate": 4.0129963898916964e-05, "loss": 1.5267, "step": 16584 }, { "epoch": 11.978339350180505, "grad_norm": 2.0298404693603516, "learning_rate": 4.0108303249097474e-05, "loss": 1.5422, "step": 16590 }, { "epoch": 11.982671480144404, "grad_norm": 2.0936288833618164, "learning_rate": 4.008664259927798e-05, "loss": 1.5696, "step": 16596 }, { "epoch": 11.987003610108303, "grad_norm": 2.268059015274048, "learning_rate": 4.006498194945849e-05, "loss": 1.4912, "step": 16602 }, { "epoch": 11.991335740072202, "grad_norm": 2.2934491634368896, "learning_rate": 4.004332129963899e-05, "loss": 1.3747, "step": 16608 }, { "epoch": 11.995667870036101, "grad_norm": 2.2937822341918945, "learning_rate": 4.00216606498195e-05, "loss": 1.5532, "step": 16614 }, { "epoch": 12.0, "grad_norm": 2.6109139919281006, "learning_rate": 4e-05, "loss": 1.4478, "step": 16620 }, { "epoch": 12.004332129963899, "grad_norm": 2.0192694664001465, "learning_rate": 3.9978339350180506e-05, "loss": 1.3967, "step": 16626 }, { "epoch": 12.008664259927798, "grad_norm": 2.3439557552337646, "learning_rate": 3.995667870036101e-05, "loss": 1.5799, "step": 16632 }, { "epoch": 12.012996389891697, "grad_norm": 2.1722288131713867, "learning_rate": 3.993501805054152e-05, "loss": 1.4153, "step": 16638 }, { "epoch": 12.017328519855596, "grad_norm": 2.380666971206665, "learning_rate": 3.991335740072202e-05, "loss": 1.5892, "step": 16644 }, { "epoch": 12.021660649819495, "grad_norm": 2.2260618209838867, "learning_rate": 3.989169675090253e-05, "loss": 1.3687, "step": 16650 }, { "epoch": 12.025992779783394, "grad_norm": 2.1932132244110107, "learning_rate": 3.9870036101083036e-05, "loss": 1.4985, "step": 16656 }, { "epoch": 12.030324909747293, "grad_norm": 2.4223523139953613, "learning_rate": 3.984837545126354e-05, "loss": 1.4062, "step": 16662 }, { "epoch": 12.034657039711192, "grad_norm": 2.375209331512451, "learning_rate": 3.982671480144404e-05, "loss": 1.5108, "step": 16668 }, { "epoch": 12.03898916967509, "grad_norm": 2.1999199390411377, "learning_rate": 3.9805054151624546e-05, "loss": 1.5058, "step": 16674 }, { "epoch": 12.04332129963899, "grad_norm": 2.186201810836792, "learning_rate": 3.9783393501805056e-05, "loss": 1.4157, "step": 16680 }, { "epoch": 12.047653429602889, "grad_norm": 2.270533323287964, "learning_rate": 3.976173285198556e-05, "loss": 1.4856, "step": 16686 }, { "epoch": 12.051985559566788, "grad_norm": 1.8714853525161743, "learning_rate": 3.974007220216607e-05, "loss": 1.4668, "step": 16692 }, { "epoch": 12.056317689530687, "grad_norm": 1.9005995988845825, "learning_rate": 3.971841155234657e-05, "loss": 1.5134, "step": 16698 }, { "epoch": 12.060649819494586, "grad_norm": 2.283104658126831, "learning_rate": 3.969675090252708e-05, "loss": 1.4875, "step": 16704 }, { "epoch": 12.064981949458483, "grad_norm": 2.2414984703063965, "learning_rate": 3.967509025270758e-05, "loss": 1.4403, "step": 16710 }, { "epoch": 12.069314079422382, "grad_norm": 2.0708913803100586, "learning_rate": 3.965342960288809e-05, "loss": 1.5379, "step": 16716 }, { "epoch": 12.07364620938628, "grad_norm": 2.194012403488159, "learning_rate": 3.963176895306859e-05, "loss": 1.3849, "step": 16722 }, { "epoch": 12.07797833935018, "grad_norm": 2.3993148803710938, "learning_rate": 3.96101083032491e-05, "loss": 1.5912, "step": 16728 }, { "epoch": 12.082310469314079, "grad_norm": 2.260986328125, "learning_rate": 3.9588447653429605e-05, "loss": 1.3773, "step": 16734 }, { "epoch": 12.086642599277978, "grad_norm": 2.5972652435302734, "learning_rate": 3.956678700361011e-05, "loss": 1.4895, "step": 16740 }, { "epoch": 12.090974729241877, "grad_norm": 2.2936391830444336, "learning_rate": 3.954512635379062e-05, "loss": 1.4033, "step": 16746 }, { "epoch": 12.095306859205776, "grad_norm": 2.1608150005340576, "learning_rate": 3.952346570397112e-05, "loss": 1.527, "step": 16752 }, { "epoch": 12.099638989169675, "grad_norm": 2.279505968093872, "learning_rate": 3.9501805054151625e-05, "loss": 1.4269, "step": 16758 }, { "epoch": 12.103971119133574, "grad_norm": 2.2708849906921387, "learning_rate": 3.948014440433213e-05, "loss": 1.4498, "step": 16764 }, { "epoch": 12.108303249097473, "grad_norm": 2.2592616081237793, "learning_rate": 3.945848375451264e-05, "loss": 1.4601, "step": 16770 }, { "epoch": 12.112635379061372, "grad_norm": 2.276075839996338, "learning_rate": 3.943682310469314e-05, "loss": 1.4698, "step": 16776 }, { "epoch": 12.11696750902527, "grad_norm": 2.537707567214966, "learning_rate": 3.941516245487365e-05, "loss": 1.524, "step": 16782 }, { "epoch": 12.12129963898917, "grad_norm": 2.2678728103637695, "learning_rate": 3.9393501805054155e-05, "loss": 1.5236, "step": 16788 }, { "epoch": 12.125631768953069, "grad_norm": 2.0749077796936035, "learning_rate": 3.937184115523466e-05, "loss": 1.4142, "step": 16794 }, { "epoch": 12.129963898916968, "grad_norm": 2.3958353996276855, "learning_rate": 3.935018050541516e-05, "loss": 1.4387, "step": 16800 }, { "epoch": 12.134296028880867, "grad_norm": 2.419823169708252, "learning_rate": 3.932851985559567e-05, "loss": 1.5654, "step": 16806 }, { "epoch": 12.138628158844766, "grad_norm": 2.3017020225524902, "learning_rate": 3.9306859205776174e-05, "loss": 1.4914, "step": 16812 }, { "epoch": 12.142960288808665, "grad_norm": 2.3053274154663086, "learning_rate": 3.9285198555956684e-05, "loss": 1.5038, "step": 16818 }, { "epoch": 12.147292418772563, "grad_norm": 2.321998357772827, "learning_rate": 3.926353790613719e-05, "loss": 1.5044, "step": 16824 }, { "epoch": 12.151624548736462, "grad_norm": 2.22884202003479, "learning_rate": 3.924187725631769e-05, "loss": 1.4771, "step": 16830 }, { "epoch": 12.155956678700361, "grad_norm": 2.1675922870635986, "learning_rate": 3.9220216606498194e-05, "loss": 1.4688, "step": 16836 }, { "epoch": 12.16028880866426, "grad_norm": 2.3583219051361084, "learning_rate": 3.9198555956678704e-05, "loss": 1.5044, "step": 16842 }, { "epoch": 12.16462093862816, "grad_norm": 2.60506272315979, "learning_rate": 3.917689530685921e-05, "loss": 1.5114, "step": 16848 }, { "epoch": 12.168953068592058, "grad_norm": 2.248325824737549, "learning_rate": 3.915523465703971e-05, "loss": 1.5552, "step": 16854 }, { "epoch": 12.173285198555957, "grad_norm": 2.4585416316986084, "learning_rate": 3.913357400722022e-05, "loss": 1.463, "step": 16860 }, { "epoch": 12.177617328519856, "grad_norm": 2.275524854660034, "learning_rate": 3.9111913357400724e-05, "loss": 1.4841, "step": 16866 }, { "epoch": 12.181949458483755, "grad_norm": 2.2274420261383057, "learning_rate": 3.9090252707581234e-05, "loss": 1.3361, "step": 16872 }, { "epoch": 12.186281588447654, "grad_norm": 2.251668930053711, "learning_rate": 3.906859205776173e-05, "loss": 1.4422, "step": 16878 }, { "epoch": 12.190613718411552, "grad_norm": 2.4449918270111084, "learning_rate": 3.904693140794224e-05, "loss": 1.4593, "step": 16884 }, { "epoch": 12.19494584837545, "grad_norm": 2.5030899047851562, "learning_rate": 3.902527075812274e-05, "loss": 1.6453, "step": 16890 }, { "epoch": 12.19927797833935, "grad_norm": 2.2501237392425537, "learning_rate": 3.900361010830325e-05, "loss": 1.4504, "step": 16896 }, { "epoch": 12.203610108303248, "grad_norm": 2.2348971366882324, "learning_rate": 3.8981949458483756e-05, "loss": 1.5302, "step": 16902 }, { "epoch": 12.207942238267147, "grad_norm": 2.7858738899230957, "learning_rate": 3.8960288808664267e-05, "loss": 1.549, "step": 16908 }, { "epoch": 12.212274368231046, "grad_norm": 2.5331666469573975, "learning_rate": 3.893862815884477e-05, "loss": 1.5578, "step": 16914 }, { "epoch": 12.216606498194945, "grad_norm": 2.1402599811553955, "learning_rate": 3.891696750902527e-05, "loss": 1.4606, "step": 16920 }, { "epoch": 12.220938628158844, "grad_norm": 2.3635129928588867, "learning_rate": 3.8895306859205776e-05, "loss": 1.4624, "step": 16926 }, { "epoch": 12.225270758122743, "grad_norm": 2.1728463172912598, "learning_rate": 3.8873646209386286e-05, "loss": 1.5206, "step": 16932 }, { "epoch": 12.229602888086642, "grad_norm": 2.346597194671631, "learning_rate": 3.885198555956679e-05, "loss": 1.4773, "step": 16938 }, { "epoch": 12.233935018050541, "grad_norm": 2.318615674972534, "learning_rate": 3.883032490974729e-05, "loss": 1.4219, "step": 16944 }, { "epoch": 12.23826714801444, "grad_norm": 2.2937116622924805, "learning_rate": 3.88086642599278e-05, "loss": 1.5383, "step": 16950 }, { "epoch": 12.24259927797834, "grad_norm": 2.282334804534912, "learning_rate": 3.8787003610108306e-05, "loss": 1.5487, "step": 16956 }, { "epoch": 12.246931407942238, "grad_norm": 2.242652416229248, "learning_rate": 3.876534296028881e-05, "loss": 1.5051, "step": 16962 }, { "epoch": 12.251263537906137, "grad_norm": 2.4972829818725586, "learning_rate": 3.874368231046931e-05, "loss": 1.5523, "step": 16968 }, { "epoch": 12.255595667870036, "grad_norm": 2.414475202560425, "learning_rate": 3.872202166064982e-05, "loss": 1.4916, "step": 16974 }, { "epoch": 12.259927797833935, "grad_norm": 2.240234613418579, "learning_rate": 3.8700361010830326e-05, "loss": 1.5599, "step": 16980 }, { "epoch": 12.264259927797834, "grad_norm": 2.3836188316345215, "learning_rate": 3.8678700361010836e-05, "loss": 1.4981, "step": 16986 }, { "epoch": 12.268592057761733, "grad_norm": 2.2600574493408203, "learning_rate": 3.865703971119134e-05, "loss": 1.526, "step": 16992 }, { "epoch": 12.272924187725632, "grad_norm": 2.395291566848755, "learning_rate": 3.863537906137184e-05, "loss": 1.4187, "step": 16998 }, { "epoch": 12.277256317689531, "grad_norm": 2.147730827331543, "learning_rate": 3.8613718411552345e-05, "loss": 1.4449, "step": 17004 }, { "epoch": 12.28158844765343, "grad_norm": 2.5655853748321533, "learning_rate": 3.8592057761732855e-05, "loss": 1.611, "step": 17010 }, { "epoch": 12.285920577617329, "grad_norm": 2.5320088863372803, "learning_rate": 3.857039711191336e-05, "loss": 1.522, "step": 17016 }, { "epoch": 12.290252707581228, "grad_norm": 2.3203256130218506, "learning_rate": 3.854873646209387e-05, "loss": 1.5195, "step": 17022 }, { "epoch": 12.294584837545127, "grad_norm": 2.1262683868408203, "learning_rate": 3.852707581227437e-05, "loss": 1.3975, "step": 17028 }, { "epoch": 12.298916967509026, "grad_norm": 2.452780246734619, "learning_rate": 3.8505415162454875e-05, "loss": 1.5159, "step": 17034 }, { "epoch": 12.303249097472925, "grad_norm": 2.250030040740967, "learning_rate": 3.848375451263538e-05, "loss": 1.4065, "step": 17040 }, { "epoch": 12.307581227436824, "grad_norm": 2.4820809364318848, "learning_rate": 3.846209386281588e-05, "loss": 1.5035, "step": 17046 }, { "epoch": 12.311913357400723, "grad_norm": 2.492382526397705, "learning_rate": 3.844043321299639e-05, "loss": 1.5135, "step": 17052 }, { "epoch": 12.31624548736462, "grad_norm": 2.318692445755005, "learning_rate": 3.8418772563176895e-05, "loss": 1.5356, "step": 17058 }, { "epoch": 12.320577617328519, "grad_norm": 2.2669308185577393, "learning_rate": 3.8397111913357405e-05, "loss": 1.5192, "step": 17064 }, { "epoch": 12.324909747292418, "grad_norm": 2.236903429031372, "learning_rate": 3.837545126353791e-05, "loss": 1.6137, "step": 17070 }, { "epoch": 12.329241877256317, "grad_norm": 2.2819299697875977, "learning_rate": 3.835379061371842e-05, "loss": 1.442, "step": 17076 }, { "epoch": 12.333574007220216, "grad_norm": 2.537736177444458, "learning_rate": 3.8332129963898914e-05, "loss": 1.3962, "step": 17082 }, { "epoch": 12.337906137184115, "grad_norm": 2.03959321975708, "learning_rate": 3.8310469314079424e-05, "loss": 1.4729, "step": 17088 }, { "epoch": 12.342238267148014, "grad_norm": 2.4640324115753174, "learning_rate": 3.828880866425993e-05, "loss": 1.5813, "step": 17094 }, { "epoch": 12.346570397111913, "grad_norm": 2.194430351257324, "learning_rate": 3.826714801444044e-05, "loss": 1.4798, "step": 17100 }, { "epoch": 12.350902527075812, "grad_norm": 2.153595447540283, "learning_rate": 3.824548736462094e-05, "loss": 1.4679, "step": 17106 }, { "epoch": 12.355234657039711, "grad_norm": 2.348609447479248, "learning_rate": 3.8223826714801444e-05, "loss": 1.4988, "step": 17112 }, { "epoch": 12.35956678700361, "grad_norm": 2.2559471130371094, "learning_rate": 3.8202166064981954e-05, "loss": 1.407, "step": 17118 }, { "epoch": 12.363898916967509, "grad_norm": 2.3384580612182617, "learning_rate": 3.818050541516246e-05, "loss": 1.5049, "step": 17124 }, { "epoch": 12.368231046931408, "grad_norm": 2.3775136470794678, "learning_rate": 3.815884476534296e-05, "loss": 1.4559, "step": 17130 }, { "epoch": 12.372563176895307, "grad_norm": 2.206834316253662, "learning_rate": 3.8137184115523464e-05, "loss": 1.6483, "step": 17136 }, { "epoch": 12.376895306859206, "grad_norm": 2.4667086601257324, "learning_rate": 3.8115523465703974e-05, "loss": 1.5118, "step": 17142 }, { "epoch": 12.381227436823105, "grad_norm": 2.4737181663513184, "learning_rate": 3.809386281588448e-05, "loss": 1.5335, "step": 17148 }, { "epoch": 12.385559566787004, "grad_norm": 2.4044978618621826, "learning_rate": 3.807220216606499e-05, "loss": 1.5421, "step": 17154 }, { "epoch": 12.389891696750903, "grad_norm": 2.588010549545288, "learning_rate": 3.805054151624549e-05, "loss": 1.4256, "step": 17160 }, { "epoch": 12.394223826714802, "grad_norm": 2.2427098751068115, "learning_rate": 3.802888086642599e-05, "loss": 1.516, "step": 17166 }, { "epoch": 12.3985559566787, "grad_norm": 2.4134037494659424, "learning_rate": 3.8007220216606496e-05, "loss": 1.7107, "step": 17172 }, { "epoch": 12.4028880866426, "grad_norm": 2.5335350036621094, "learning_rate": 3.7985559566787007e-05, "loss": 1.4955, "step": 17178 }, { "epoch": 12.407220216606499, "grad_norm": 2.1225152015686035, "learning_rate": 3.796389891696751e-05, "loss": 1.3999, "step": 17184 }, { "epoch": 12.411552346570398, "grad_norm": 2.342261791229248, "learning_rate": 3.794223826714802e-05, "loss": 1.5821, "step": 17190 }, { "epoch": 12.415884476534297, "grad_norm": 2.433962106704712, "learning_rate": 3.792057761732852e-05, "loss": 1.5375, "step": 17196 }, { "epoch": 12.420216606498196, "grad_norm": 1.994956612586975, "learning_rate": 3.7898916967509026e-05, "loss": 1.4319, "step": 17202 }, { "epoch": 12.424548736462095, "grad_norm": 2.4866678714752197, "learning_rate": 3.787725631768953e-05, "loss": 1.4055, "step": 17208 }, { "epoch": 12.428880866425994, "grad_norm": 2.549171209335327, "learning_rate": 3.785559566787004e-05, "loss": 1.6065, "step": 17214 }, { "epoch": 12.433212996389893, "grad_norm": 2.444087028503418, "learning_rate": 3.783393501805054e-05, "loss": 1.4719, "step": 17220 }, { "epoch": 12.43754512635379, "grad_norm": 2.3872203826904297, "learning_rate": 3.7812274368231046e-05, "loss": 1.4447, "step": 17226 }, { "epoch": 12.441877256317689, "grad_norm": 2.4547035694122314, "learning_rate": 3.7790613718411556e-05, "loss": 1.4668, "step": 17232 }, { "epoch": 12.446209386281588, "grad_norm": 2.352126121520996, "learning_rate": 3.776895306859206e-05, "loss": 1.6, "step": 17238 }, { "epoch": 12.450541516245487, "grad_norm": 2.3293519020080566, "learning_rate": 3.774729241877257e-05, "loss": 1.5036, "step": 17244 }, { "epoch": 12.454873646209386, "grad_norm": 2.5735654830932617, "learning_rate": 3.7725631768953066e-05, "loss": 1.4684, "step": 17250 }, { "epoch": 12.459205776173285, "grad_norm": 2.253568649291992, "learning_rate": 3.7703971119133576e-05, "loss": 1.4779, "step": 17256 }, { "epoch": 12.463537906137184, "grad_norm": 2.359865665435791, "learning_rate": 3.768231046931408e-05, "loss": 1.5157, "step": 17262 }, { "epoch": 12.467870036101083, "grad_norm": 2.258496046066284, "learning_rate": 3.766064981949459e-05, "loss": 1.52, "step": 17268 }, { "epoch": 12.472202166064982, "grad_norm": 2.257800817489624, "learning_rate": 3.763898916967509e-05, "loss": 1.4637, "step": 17274 }, { "epoch": 12.47653429602888, "grad_norm": 2.30025053024292, "learning_rate": 3.76173285198556e-05, "loss": 1.4453, "step": 17280 }, { "epoch": 12.48086642599278, "grad_norm": 2.196463108062744, "learning_rate": 3.7595667870036105e-05, "loss": 1.5506, "step": 17286 }, { "epoch": 12.485198555956678, "grad_norm": 2.2742984294891357, "learning_rate": 3.757400722021661e-05, "loss": 1.4959, "step": 17292 }, { "epoch": 12.489530685920577, "grad_norm": 2.304678201675415, "learning_rate": 3.755234657039711e-05, "loss": 1.5164, "step": 17298 }, { "epoch": 12.493862815884476, "grad_norm": 2.301745653152466, "learning_rate": 3.753068592057762e-05, "loss": 1.5144, "step": 17304 }, { "epoch": 12.498194945848375, "grad_norm": 2.396341562271118, "learning_rate": 3.7509025270758125e-05, "loss": 1.5701, "step": 17310 }, { "epoch": 12.502527075812274, "grad_norm": 2.183166742324829, "learning_rate": 3.748736462093863e-05, "loss": 1.6367, "step": 17316 }, { "epoch": 12.506859205776173, "grad_norm": 2.5042848587036133, "learning_rate": 3.746570397111914e-05, "loss": 1.4294, "step": 17322 }, { "epoch": 12.511191335740072, "grad_norm": 2.152460813522339, "learning_rate": 3.744404332129964e-05, "loss": 1.5067, "step": 17328 }, { "epoch": 12.515523465703971, "grad_norm": 2.708660125732422, "learning_rate": 3.7422382671480145e-05, "loss": 1.5286, "step": 17334 }, { "epoch": 12.51985559566787, "grad_norm": 2.224719524383545, "learning_rate": 3.740072202166065e-05, "loss": 1.4671, "step": 17340 }, { "epoch": 12.52418772563177, "grad_norm": 2.266981840133667, "learning_rate": 3.737906137184116e-05, "loss": 1.5242, "step": 17346 }, { "epoch": 12.528519855595668, "grad_norm": 2.5527312755584717, "learning_rate": 3.735740072202166e-05, "loss": 1.4739, "step": 17352 }, { "epoch": 12.532851985559567, "grad_norm": 2.4082605838775635, "learning_rate": 3.733574007220217e-05, "loss": 1.5709, "step": 17358 }, { "epoch": 12.537184115523466, "grad_norm": 2.2685296535491943, "learning_rate": 3.7314079422382674e-05, "loss": 1.4384, "step": 17364 }, { "epoch": 12.541516245487365, "grad_norm": 2.4036061763763428, "learning_rate": 3.7292418772563184e-05, "loss": 1.4637, "step": 17370 }, { "epoch": 12.545848375451264, "grad_norm": 2.610372543334961, "learning_rate": 3.727075812274368e-05, "loss": 1.5638, "step": 17376 }, { "epoch": 12.550180505415163, "grad_norm": 2.55112361907959, "learning_rate": 3.724909747292419e-05, "loss": 1.5639, "step": 17382 }, { "epoch": 12.554512635379062, "grad_norm": 2.220541000366211, "learning_rate": 3.7227436823104694e-05, "loss": 1.5273, "step": 17388 }, { "epoch": 12.558844765342961, "grad_norm": 2.4928078651428223, "learning_rate": 3.7205776173285204e-05, "loss": 1.52, "step": 17394 }, { "epoch": 12.56317689530686, "grad_norm": 2.375063896179199, "learning_rate": 3.718411552346571e-05, "loss": 1.3846, "step": 17400 }, { "epoch": 12.567509025270757, "grad_norm": 2.3351051807403564, "learning_rate": 3.716245487364621e-05, "loss": 1.5402, "step": 17406 }, { "epoch": 12.571841155234656, "grad_norm": 2.392690658569336, "learning_rate": 3.714079422382672e-05, "loss": 1.5159, "step": 17412 }, { "epoch": 12.576173285198555, "grad_norm": 2.238818407058716, "learning_rate": 3.711913357400722e-05, "loss": 1.5009, "step": 17418 }, { "epoch": 12.580505415162454, "grad_norm": 2.4636850357055664, "learning_rate": 3.709747292418773e-05, "loss": 1.5541, "step": 17424 }, { "epoch": 12.584837545126353, "grad_norm": 2.4320149421691895, "learning_rate": 3.707581227436823e-05, "loss": 1.553, "step": 17430 }, { "epoch": 12.589169675090252, "grad_norm": 2.327601194381714, "learning_rate": 3.705415162454874e-05, "loss": 1.503, "step": 17436 }, { "epoch": 12.593501805054151, "grad_norm": 2.437288522720337, "learning_rate": 3.703249097472924e-05, "loss": 1.5548, "step": 17442 }, { "epoch": 12.59783393501805, "grad_norm": 2.2094645500183105, "learning_rate": 3.701083032490975e-05, "loss": 1.4137, "step": 17448 }, { "epoch": 12.60216606498195, "grad_norm": 2.1708738803863525, "learning_rate": 3.6989169675090257e-05, "loss": 1.4157, "step": 17454 }, { "epoch": 12.606498194945848, "grad_norm": 2.046398162841797, "learning_rate": 3.696750902527076e-05, "loss": 1.3963, "step": 17460 }, { "epoch": 12.610830324909747, "grad_norm": 2.213010787963867, "learning_rate": 3.694584837545126e-05, "loss": 1.4991, "step": 17466 }, { "epoch": 12.615162454873646, "grad_norm": 2.229412794113159, "learning_rate": 3.692418772563177e-05, "loss": 1.5003, "step": 17472 }, { "epoch": 12.619494584837545, "grad_norm": 2.1690070629119873, "learning_rate": 3.6902527075812276e-05, "loss": 1.5791, "step": 17478 }, { "epoch": 12.623826714801444, "grad_norm": 2.5734190940856934, "learning_rate": 3.688086642599278e-05, "loss": 1.5551, "step": 17484 }, { "epoch": 12.628158844765343, "grad_norm": 2.2498667240142822, "learning_rate": 3.685920577617329e-05, "loss": 1.5421, "step": 17490 }, { "epoch": 12.632490974729242, "grad_norm": 2.0538721084594727, "learning_rate": 3.683754512635379e-05, "loss": 1.5441, "step": 17496 }, { "epoch": 12.636823104693141, "grad_norm": 2.2457377910614014, "learning_rate": 3.6815884476534296e-05, "loss": 1.5591, "step": 17502 }, { "epoch": 12.64115523465704, "grad_norm": 2.3194501399993896, "learning_rate": 3.67942238267148e-05, "loss": 1.3896, "step": 17508 }, { "epoch": 12.645487364620939, "grad_norm": 2.3579676151275635, "learning_rate": 3.677256317689531e-05, "loss": 1.5282, "step": 17514 }, { "epoch": 12.649819494584838, "grad_norm": 2.3968262672424316, "learning_rate": 3.675090252707581e-05, "loss": 1.5005, "step": 17520 }, { "epoch": 12.654151624548737, "grad_norm": 2.2977466583251953, "learning_rate": 3.672924187725632e-05, "loss": 1.5016, "step": 17526 }, { "epoch": 12.658483754512636, "grad_norm": 2.4458224773406982, "learning_rate": 3.6707581227436826e-05, "loss": 1.4761, "step": 17532 }, { "epoch": 12.662815884476535, "grad_norm": 2.5550875663757324, "learning_rate": 3.668592057761733e-05, "loss": 1.4422, "step": 17538 }, { "epoch": 12.667148014440434, "grad_norm": 2.49090313911438, "learning_rate": 3.666425992779783e-05, "loss": 1.6106, "step": 17544 }, { "epoch": 12.671480144404333, "grad_norm": 2.2808785438537598, "learning_rate": 3.664259927797834e-05, "loss": 1.5649, "step": 17550 }, { "epoch": 12.675812274368232, "grad_norm": 2.174480438232422, "learning_rate": 3.6620938628158845e-05, "loss": 1.4152, "step": 17556 }, { "epoch": 12.68014440433213, "grad_norm": 2.257552146911621, "learning_rate": 3.6599277978339355e-05, "loss": 1.4152, "step": 17562 }, { "epoch": 12.684476534296028, "grad_norm": 2.5686330795288086, "learning_rate": 3.657761732851986e-05, "loss": 1.4804, "step": 17568 }, { "epoch": 12.688808664259927, "grad_norm": 2.3417282104492188, "learning_rate": 3.655595667870036e-05, "loss": 1.5046, "step": 17574 }, { "epoch": 12.693140794223826, "grad_norm": 2.3491597175598145, "learning_rate": 3.6534296028880865e-05, "loss": 1.4727, "step": 17580 }, { "epoch": 12.697472924187725, "grad_norm": 2.2331085205078125, "learning_rate": 3.6512635379061375e-05, "loss": 1.5567, "step": 17586 }, { "epoch": 12.701805054151624, "grad_norm": 2.290557861328125, "learning_rate": 3.649097472924188e-05, "loss": 1.3915, "step": 17592 }, { "epoch": 12.706137184115523, "grad_norm": 2.3298158645629883, "learning_rate": 3.646931407942238e-05, "loss": 1.4928, "step": 17598 }, { "epoch": 12.710469314079422, "grad_norm": 2.293748378753662, "learning_rate": 3.644765342960289e-05, "loss": 1.5832, "step": 17604 }, { "epoch": 12.71480144404332, "grad_norm": 2.3185017108917236, "learning_rate": 3.6425992779783395e-05, "loss": 1.498, "step": 17610 }, { "epoch": 12.71913357400722, "grad_norm": 2.2559006214141846, "learning_rate": 3.6404332129963905e-05, "loss": 1.4413, "step": 17616 }, { "epoch": 12.723465703971119, "grad_norm": 2.480546712875366, "learning_rate": 3.63826714801444e-05, "loss": 1.4971, "step": 17622 }, { "epoch": 12.727797833935018, "grad_norm": 2.1618759632110596, "learning_rate": 3.636101083032491e-05, "loss": 1.4462, "step": 17628 }, { "epoch": 12.732129963898917, "grad_norm": 2.5159764289855957, "learning_rate": 3.6339350180505414e-05, "loss": 1.5562, "step": 17634 }, { "epoch": 12.736462093862816, "grad_norm": 2.5831587314605713, "learning_rate": 3.6317689530685924e-05, "loss": 1.4341, "step": 17640 }, { "epoch": 12.740794223826715, "grad_norm": 2.332818031311035, "learning_rate": 3.629602888086643e-05, "loss": 1.4952, "step": 17646 }, { "epoch": 12.745126353790614, "grad_norm": 2.387537956237793, "learning_rate": 3.627436823104694e-05, "loss": 1.5315, "step": 17652 }, { "epoch": 12.749458483754513, "grad_norm": 2.372143030166626, "learning_rate": 3.625270758122744e-05, "loss": 1.4858, "step": 17658 }, { "epoch": 12.753790613718412, "grad_norm": 2.6558310985565186, "learning_rate": 3.6231046931407944e-05, "loss": 1.3461, "step": 17664 }, { "epoch": 12.75812274368231, "grad_norm": 2.349839210510254, "learning_rate": 3.620938628158845e-05, "loss": 1.4068, "step": 17670 }, { "epoch": 12.76245487364621, "grad_norm": 2.193542242050171, "learning_rate": 3.618772563176896e-05, "loss": 1.4463, "step": 17676 }, { "epoch": 12.766787003610109, "grad_norm": 2.1853740215301514, "learning_rate": 3.616606498194946e-05, "loss": 1.4603, "step": 17682 }, { "epoch": 12.771119133574008, "grad_norm": 2.2640743255615234, "learning_rate": 3.6144404332129964e-05, "loss": 1.474, "step": 17688 }, { "epoch": 12.775451263537906, "grad_norm": 2.3613524436950684, "learning_rate": 3.6122743682310474e-05, "loss": 1.4195, "step": 17694 }, { "epoch": 12.779783393501805, "grad_norm": 2.401092290878296, "learning_rate": 3.610108303249098e-05, "loss": 1.5768, "step": 17700 }, { "epoch": 12.784115523465704, "grad_norm": 2.184077262878418, "learning_rate": 3.607942238267148e-05, "loss": 1.4406, "step": 17706 }, { "epoch": 12.788447653429603, "grad_norm": 2.4494833946228027, "learning_rate": 3.605776173285198e-05, "loss": 1.5438, "step": 17712 }, { "epoch": 12.792779783393502, "grad_norm": 2.2727792263031006, "learning_rate": 3.603610108303249e-05, "loss": 1.4658, "step": 17718 }, { "epoch": 12.797111913357401, "grad_norm": 2.2577919960021973, "learning_rate": 3.6014440433212997e-05, "loss": 1.4604, "step": 17724 }, { "epoch": 12.8014440433213, "grad_norm": 2.579435110092163, "learning_rate": 3.5992779783393507e-05, "loss": 1.6334, "step": 17730 }, { "epoch": 12.8057761732852, "grad_norm": 2.24682879447937, "learning_rate": 3.597111913357401e-05, "loss": 1.4351, "step": 17736 }, { "epoch": 12.810108303249098, "grad_norm": 2.3698501586914062, "learning_rate": 3.594945848375452e-05, "loss": 1.5556, "step": 17742 }, { "epoch": 12.814440433212997, "grad_norm": 2.453380823135376, "learning_rate": 3.5927797833935016e-05, "loss": 1.5486, "step": 17748 }, { "epoch": 12.818772563176895, "grad_norm": 2.449760675430298, "learning_rate": 3.5906137184115526e-05, "loss": 1.5645, "step": 17754 }, { "epoch": 12.823104693140793, "grad_norm": 2.3518009185791016, "learning_rate": 3.588447653429603e-05, "loss": 1.4816, "step": 17760 }, { "epoch": 12.827436823104692, "grad_norm": 2.469226121902466, "learning_rate": 3.586281588447654e-05, "loss": 1.52, "step": 17766 }, { "epoch": 12.831768953068591, "grad_norm": 2.520305871963501, "learning_rate": 3.584115523465704e-05, "loss": 1.4768, "step": 17772 }, { "epoch": 12.83610108303249, "grad_norm": 2.379380464553833, "learning_rate": 3.5819494584837546e-05, "loss": 1.4904, "step": 17778 }, { "epoch": 12.84043321299639, "grad_norm": 2.3030216693878174, "learning_rate": 3.5797833935018056e-05, "loss": 1.4549, "step": 17784 }, { "epoch": 12.844765342960288, "grad_norm": 2.353506326675415, "learning_rate": 3.577617328519855e-05, "loss": 1.5469, "step": 17790 }, { "epoch": 12.849097472924187, "grad_norm": 2.2974421977996826, "learning_rate": 3.575451263537906e-05, "loss": 1.3424, "step": 17796 }, { "epoch": 12.853429602888086, "grad_norm": 2.431941032409668, "learning_rate": 3.5732851985559566e-05, "loss": 1.5974, "step": 17802 }, { "epoch": 12.857761732851985, "grad_norm": 2.397289991378784, "learning_rate": 3.5711191335740076e-05, "loss": 1.5521, "step": 17808 }, { "epoch": 12.862093862815884, "grad_norm": 2.269667863845825, "learning_rate": 3.568953068592058e-05, "loss": 1.3712, "step": 17814 }, { "epoch": 12.866425992779783, "grad_norm": 2.036743640899658, "learning_rate": 3.566787003610109e-05, "loss": 1.5555, "step": 17820 }, { "epoch": 12.870758122743682, "grad_norm": 2.360935688018799, "learning_rate": 3.564620938628159e-05, "loss": 1.5137, "step": 17826 }, { "epoch": 12.875090252707581, "grad_norm": 2.236589193344116, "learning_rate": 3.5624548736462095e-05, "loss": 1.527, "step": 17832 }, { "epoch": 12.87942238267148, "grad_norm": 2.6480045318603516, "learning_rate": 3.56028880866426e-05, "loss": 1.6259, "step": 17838 }, { "epoch": 12.88375451263538, "grad_norm": 2.298877477645874, "learning_rate": 3.558122743682311e-05, "loss": 1.6336, "step": 17844 }, { "epoch": 12.888086642599278, "grad_norm": 2.324199676513672, "learning_rate": 3.555956678700361e-05, "loss": 1.4795, "step": 17850 }, { "epoch": 12.892418772563177, "grad_norm": 2.4183850288391113, "learning_rate": 3.5537906137184115e-05, "loss": 1.5147, "step": 17856 }, { "epoch": 12.896750902527076, "grad_norm": 2.333462715148926, "learning_rate": 3.5516245487364625e-05, "loss": 1.4485, "step": 17862 }, { "epoch": 12.901083032490975, "grad_norm": 2.3888120651245117, "learning_rate": 3.549458483754513e-05, "loss": 1.4259, "step": 17868 }, { "epoch": 12.905415162454874, "grad_norm": 2.2608587741851807, "learning_rate": 3.547292418772563e-05, "loss": 1.5546, "step": 17874 }, { "epoch": 12.909747292418773, "grad_norm": 2.2936878204345703, "learning_rate": 3.5451263537906135e-05, "loss": 1.5678, "step": 17880 }, { "epoch": 12.914079422382672, "grad_norm": 2.2470672130584717, "learning_rate": 3.5429602888086645e-05, "loss": 1.5158, "step": 17886 }, { "epoch": 12.918411552346571, "grad_norm": 2.17495059967041, "learning_rate": 3.540794223826715e-05, "loss": 1.4693, "step": 17892 }, { "epoch": 12.92274368231047, "grad_norm": 2.3425323963165283, "learning_rate": 3.538628158844766e-05, "loss": 1.5855, "step": 17898 }, { "epoch": 12.927075812274369, "grad_norm": 2.2176942825317383, "learning_rate": 3.536462093862816e-05, "loss": 1.5546, "step": 17904 }, { "epoch": 12.931407942238268, "grad_norm": 2.3174469470977783, "learning_rate": 3.534296028880867e-05, "loss": 1.648, "step": 17910 }, { "epoch": 12.935740072202165, "grad_norm": 2.0531599521636963, "learning_rate": 3.532129963898917e-05, "loss": 1.4675, "step": 17916 }, { "epoch": 12.940072202166064, "grad_norm": 2.1208434104919434, "learning_rate": 3.529963898916968e-05, "loss": 1.5109, "step": 17922 }, { "epoch": 12.944404332129963, "grad_norm": 2.217986822128296, "learning_rate": 3.527797833935018e-05, "loss": 1.4595, "step": 17928 }, { "epoch": 12.948736462093862, "grad_norm": 2.5315189361572266, "learning_rate": 3.525631768953069e-05, "loss": 1.4921, "step": 17934 }, { "epoch": 12.953068592057761, "grad_norm": 2.253772020339966, "learning_rate": 3.5234657039711194e-05, "loss": 1.4773, "step": 17940 }, { "epoch": 12.95740072202166, "grad_norm": 2.3777897357940674, "learning_rate": 3.52129963898917e-05, "loss": 1.5553, "step": 17946 }, { "epoch": 12.961732851985559, "grad_norm": 2.3947982788085938, "learning_rate": 3.519133574007221e-05, "loss": 1.499, "step": 17952 }, { "epoch": 12.966064981949458, "grad_norm": 2.390638589859009, "learning_rate": 3.516967509025271e-05, "loss": 1.4481, "step": 17958 }, { "epoch": 12.970397111913357, "grad_norm": 2.3408362865448, "learning_rate": 3.5148014440433214e-05, "loss": 1.5248, "step": 17964 }, { "epoch": 12.974729241877256, "grad_norm": 2.2928483486175537, "learning_rate": 3.512635379061372e-05, "loss": 1.4474, "step": 17970 }, { "epoch": 12.979061371841155, "grad_norm": 2.2032880783081055, "learning_rate": 3.510469314079423e-05, "loss": 1.3839, "step": 17976 }, { "epoch": 12.983393501805054, "grad_norm": 2.2581377029418945, "learning_rate": 3.508303249097473e-05, "loss": 1.5581, "step": 17982 }, { "epoch": 12.987725631768953, "grad_norm": 2.254276990890503, "learning_rate": 3.506137184115524e-05, "loss": 1.4975, "step": 17988 }, { "epoch": 12.992057761732852, "grad_norm": 2.494955539703369, "learning_rate": 3.5039711191335737e-05, "loss": 1.6222, "step": 17994 }, { "epoch": 12.99638989169675, "grad_norm": 2.472038507461548, "learning_rate": 3.5018050541516247e-05, "loss": 1.4021, "step": 18000 }, { "epoch": 13.00072202166065, "grad_norm": 2.0454769134521484, "learning_rate": 3.499638989169675e-05, "loss": 1.5295, "step": 18006 }, { "epoch": 13.005054151624549, "grad_norm": 2.3485987186431885, "learning_rate": 3.497472924187726e-05, "loss": 1.4666, "step": 18012 }, { "epoch": 13.009386281588448, "grad_norm": 2.3682126998901367, "learning_rate": 3.495306859205776e-05, "loss": 1.4628, "step": 18018 }, { "epoch": 13.013718411552347, "grad_norm": 2.2539398670196533, "learning_rate": 3.493140794223827e-05, "loss": 1.3563, "step": 18024 }, { "epoch": 13.018050541516246, "grad_norm": 2.129157066345215, "learning_rate": 3.4909747292418776e-05, "loss": 1.501, "step": 18030 }, { "epoch": 13.022382671480145, "grad_norm": 2.1347594261169434, "learning_rate": 3.488808664259928e-05, "loss": 1.3819, "step": 18036 }, { "epoch": 13.026714801444044, "grad_norm": 2.2430317401885986, "learning_rate": 3.486642599277978e-05, "loss": 1.3782, "step": 18042 }, { "epoch": 13.031046931407943, "grad_norm": 2.217660427093506, "learning_rate": 3.484476534296029e-05, "loss": 1.3426, "step": 18048 }, { "epoch": 13.035379061371842, "grad_norm": 2.037261724472046, "learning_rate": 3.4823104693140796e-05, "loss": 1.3662, "step": 18054 }, { "epoch": 13.03971119133574, "grad_norm": 2.3976593017578125, "learning_rate": 3.48014440433213e-05, "loss": 1.4253, "step": 18060 }, { "epoch": 13.04404332129964, "grad_norm": 2.1239874362945557, "learning_rate": 3.477978339350181e-05, "loss": 1.4689, "step": 18066 }, { "epoch": 13.048375451263539, "grad_norm": 2.53798770904541, "learning_rate": 3.475812274368231e-05, "loss": 1.4662, "step": 18072 }, { "epoch": 13.052707581227438, "grad_norm": 2.0276594161987305, "learning_rate": 3.4736462093862816e-05, "loss": 1.5331, "step": 18078 }, { "epoch": 13.057039711191337, "grad_norm": 2.2765681743621826, "learning_rate": 3.471480144404332e-05, "loss": 1.5458, "step": 18084 }, { "epoch": 13.061371841155236, "grad_norm": 2.2108044624328613, "learning_rate": 3.469314079422383e-05, "loss": 1.5549, "step": 18090 }, { "epoch": 13.065703971119133, "grad_norm": 2.505842447280884, "learning_rate": 3.467148014440433e-05, "loss": 1.4891, "step": 18096 }, { "epoch": 13.070036101083032, "grad_norm": 2.2509279251098633, "learning_rate": 3.464981949458484e-05, "loss": 1.6006, "step": 18102 }, { "epoch": 13.07436823104693, "grad_norm": 2.229961633682251, "learning_rate": 3.4628158844765345e-05, "loss": 1.4471, "step": 18108 }, { "epoch": 13.07870036101083, "grad_norm": 2.454150676727295, "learning_rate": 3.4606498194945855e-05, "loss": 1.4924, "step": 18114 }, { "epoch": 13.083032490974729, "grad_norm": 2.3791708946228027, "learning_rate": 3.458483754512635e-05, "loss": 1.4269, "step": 18120 }, { "epoch": 13.087364620938628, "grad_norm": 2.44331431388855, "learning_rate": 3.456317689530686e-05, "loss": 1.357, "step": 18126 }, { "epoch": 13.091696750902527, "grad_norm": 2.35540771484375, "learning_rate": 3.4541516245487365e-05, "loss": 1.5556, "step": 18132 }, { "epoch": 13.096028880866426, "grad_norm": 2.423229455947876, "learning_rate": 3.4519855595667875e-05, "loss": 1.479, "step": 18138 }, { "epoch": 13.100361010830325, "grad_norm": 2.105195999145508, "learning_rate": 3.449819494584838e-05, "loss": 1.3951, "step": 18144 }, { "epoch": 13.104693140794224, "grad_norm": 2.608354330062866, "learning_rate": 3.447653429602888e-05, "loss": 1.5314, "step": 18150 }, { "epoch": 13.109025270758123, "grad_norm": 2.6511595249176025, "learning_rate": 3.445487364620939e-05, "loss": 1.5505, "step": 18156 }, { "epoch": 13.113357400722021, "grad_norm": 2.1882312297821045, "learning_rate": 3.443321299638989e-05, "loss": 1.3348, "step": 18162 }, { "epoch": 13.11768953068592, "grad_norm": 2.2576560974121094, "learning_rate": 3.44115523465704e-05, "loss": 1.4716, "step": 18168 }, { "epoch": 13.12202166064982, "grad_norm": 2.5099472999572754, "learning_rate": 3.43898916967509e-05, "loss": 1.4917, "step": 18174 }, { "epoch": 13.126353790613718, "grad_norm": 2.4268481731414795, "learning_rate": 3.436823104693141e-05, "loss": 1.5678, "step": 18180 }, { "epoch": 13.130685920577617, "grad_norm": 2.210664749145508, "learning_rate": 3.4346570397111914e-05, "loss": 1.4915, "step": 18186 }, { "epoch": 13.135018050541516, "grad_norm": 2.574079990386963, "learning_rate": 3.4324909747292424e-05, "loss": 1.5402, "step": 18192 }, { "epoch": 13.139350180505415, "grad_norm": 2.4915037155151367, "learning_rate": 3.430324909747293e-05, "loss": 1.4986, "step": 18198 }, { "epoch": 13.143682310469314, "grad_norm": 2.3626999855041504, "learning_rate": 3.428158844765343e-05, "loss": 1.4153, "step": 18204 }, { "epoch": 13.148014440433213, "grad_norm": 2.265549659729004, "learning_rate": 3.4259927797833934e-05, "loss": 1.538, "step": 18210 }, { "epoch": 13.152346570397112, "grad_norm": 2.4915616512298584, "learning_rate": 3.4238267148014444e-05, "loss": 1.4324, "step": 18216 }, { "epoch": 13.156678700361011, "grad_norm": 2.4273557662963867, "learning_rate": 3.421660649819495e-05, "loss": 1.3927, "step": 18222 }, { "epoch": 13.16101083032491, "grad_norm": 2.359760046005249, "learning_rate": 3.419494584837545e-05, "loss": 1.4737, "step": 18228 }, { "epoch": 13.16534296028881, "grad_norm": 2.5690207481384277, "learning_rate": 3.417328519855596e-05, "loss": 1.4586, "step": 18234 }, { "epoch": 13.169675090252708, "grad_norm": 2.3828256130218506, "learning_rate": 3.4151624548736464e-05, "loss": 1.5048, "step": 18240 }, { "epoch": 13.174007220216607, "grad_norm": 2.4413363933563232, "learning_rate": 3.412996389891697e-05, "loss": 1.5302, "step": 18246 }, { "epoch": 13.178339350180506, "grad_norm": 2.4356749057769775, "learning_rate": 3.410830324909747e-05, "loss": 1.4112, "step": 18252 }, { "epoch": 13.182671480144405, "grad_norm": 2.378981113433838, "learning_rate": 3.408664259927798e-05, "loss": 1.4359, "step": 18258 }, { "epoch": 13.187003610108302, "grad_norm": 2.088243007659912, "learning_rate": 3.406498194945848e-05, "loss": 1.3483, "step": 18264 }, { "epoch": 13.191335740072201, "grad_norm": 2.406777858734131, "learning_rate": 3.404332129963899e-05, "loss": 1.4632, "step": 18270 }, { "epoch": 13.1956678700361, "grad_norm": 2.2773871421813965, "learning_rate": 3.4021660649819497e-05, "loss": 1.5398, "step": 18276 }, { "epoch": 13.2, "grad_norm": 2.538294792175293, "learning_rate": 3.4000000000000007e-05, "loss": 1.505, "step": 18282 }, { "epoch": 13.204332129963898, "grad_norm": 2.3377368450164795, "learning_rate": 3.39783393501805e-05, "loss": 1.364, "step": 18288 }, { "epoch": 13.208664259927797, "grad_norm": 2.2938642501831055, "learning_rate": 3.395667870036101e-05, "loss": 1.5144, "step": 18294 }, { "epoch": 13.212996389891696, "grad_norm": 2.24360990524292, "learning_rate": 3.3935018050541516e-05, "loss": 1.5779, "step": 18300 }, { "epoch": 13.217328519855595, "grad_norm": 2.2587006092071533, "learning_rate": 3.3913357400722026e-05, "loss": 1.412, "step": 18306 }, { "epoch": 13.221660649819494, "grad_norm": 2.421426296234131, "learning_rate": 3.389169675090253e-05, "loss": 1.5417, "step": 18312 }, { "epoch": 13.225992779783393, "grad_norm": 2.297121524810791, "learning_rate": 3.387003610108303e-05, "loss": 1.4058, "step": 18318 }, { "epoch": 13.230324909747292, "grad_norm": 2.1894826889038086, "learning_rate": 3.384837545126354e-05, "loss": 1.4774, "step": 18324 }, { "epoch": 13.234657039711191, "grad_norm": 2.1453137397766113, "learning_rate": 3.3826714801444046e-05, "loss": 1.5079, "step": 18330 }, { "epoch": 13.23898916967509, "grad_norm": 2.347440481185913, "learning_rate": 3.380505415162455e-05, "loss": 1.5168, "step": 18336 }, { "epoch": 13.243321299638989, "grad_norm": 2.163522243499756, "learning_rate": 3.378339350180505e-05, "loss": 1.5589, "step": 18342 }, { "epoch": 13.247653429602888, "grad_norm": 2.525632381439209, "learning_rate": 3.376173285198556e-05, "loss": 1.4887, "step": 18348 }, { "epoch": 13.251985559566787, "grad_norm": 2.3489506244659424, "learning_rate": 3.3740072202166066e-05, "loss": 1.3861, "step": 18354 }, { "epoch": 13.256317689530686, "grad_norm": 2.707064151763916, "learning_rate": 3.3718411552346576e-05, "loss": 1.5653, "step": 18360 }, { "epoch": 13.260649819494585, "grad_norm": 2.3508431911468506, "learning_rate": 3.369675090252708e-05, "loss": 1.3759, "step": 18366 }, { "epoch": 13.264981949458484, "grad_norm": 2.4277920722961426, "learning_rate": 3.367509025270758e-05, "loss": 1.4213, "step": 18372 }, { "epoch": 13.269314079422383, "grad_norm": 2.6190896034240723, "learning_rate": 3.3653429602888085e-05, "loss": 1.5677, "step": 18378 }, { "epoch": 13.273646209386282, "grad_norm": 2.347515344619751, "learning_rate": 3.3631768953068595e-05, "loss": 1.4627, "step": 18384 }, { "epoch": 13.277978339350181, "grad_norm": 2.153172492980957, "learning_rate": 3.36101083032491e-05, "loss": 1.4981, "step": 18390 }, { "epoch": 13.28231046931408, "grad_norm": 2.356991767883301, "learning_rate": 3.358844765342961e-05, "loss": 1.3875, "step": 18396 }, { "epoch": 13.286642599277979, "grad_norm": 2.078807830810547, "learning_rate": 3.356678700361011e-05, "loss": 1.4185, "step": 18402 }, { "epoch": 13.290974729241878, "grad_norm": 2.3820762634277344, "learning_rate": 3.3545126353790615e-05, "loss": 1.4949, "step": 18408 }, { "epoch": 13.295306859205777, "grad_norm": 2.7049126625061035, "learning_rate": 3.352346570397112e-05, "loss": 1.6109, "step": 18414 }, { "epoch": 13.299638989169676, "grad_norm": 2.2845964431762695, "learning_rate": 3.350180505415163e-05, "loss": 1.4285, "step": 18420 }, { "epoch": 13.303971119133575, "grad_norm": 2.4654502868652344, "learning_rate": 3.348014440433213e-05, "loss": 1.4709, "step": 18426 }, { "epoch": 13.308303249097474, "grad_norm": 2.184924364089966, "learning_rate": 3.3458483754512635e-05, "loss": 1.5098, "step": 18432 }, { "epoch": 13.312635379061373, "grad_norm": 2.2818968296051025, "learning_rate": 3.3436823104693145e-05, "loss": 1.543, "step": 18438 }, { "epoch": 13.31696750902527, "grad_norm": 2.3762295246124268, "learning_rate": 3.341516245487365e-05, "loss": 1.5677, "step": 18444 }, { "epoch": 13.321299638989169, "grad_norm": 2.5106191635131836, "learning_rate": 3.339350180505416e-05, "loss": 1.4659, "step": 18450 }, { "epoch": 13.325631768953068, "grad_norm": 2.191404104232788, "learning_rate": 3.3371841155234654e-05, "loss": 1.4638, "step": 18456 }, { "epoch": 13.329963898916967, "grad_norm": 2.2654147148132324, "learning_rate": 3.3350180505415164e-05, "loss": 1.4303, "step": 18462 }, { "epoch": 13.334296028880866, "grad_norm": 2.2884631156921387, "learning_rate": 3.332851985559567e-05, "loss": 1.5533, "step": 18468 }, { "epoch": 13.338628158844765, "grad_norm": 2.344435214996338, "learning_rate": 3.330685920577618e-05, "loss": 1.6082, "step": 18474 }, { "epoch": 13.342960288808664, "grad_norm": 2.5276455879211426, "learning_rate": 3.328519855595668e-05, "loss": 1.4612, "step": 18480 }, { "epoch": 13.347292418772563, "grad_norm": 2.492220401763916, "learning_rate": 3.326353790613719e-05, "loss": 1.3443, "step": 18486 }, { "epoch": 13.351624548736462, "grad_norm": 2.569293737411499, "learning_rate": 3.3241877256317694e-05, "loss": 1.441, "step": 18492 }, { "epoch": 13.35595667870036, "grad_norm": 2.2464404106140137, "learning_rate": 3.32202166064982e-05, "loss": 1.5671, "step": 18498 }, { "epoch": 13.36028880866426, "grad_norm": 2.6170642375946045, "learning_rate": 3.31985559566787e-05, "loss": 1.4337, "step": 18504 }, { "epoch": 13.364620938628159, "grad_norm": 2.3511648178100586, "learning_rate": 3.317689530685921e-05, "loss": 1.4878, "step": 18510 }, { "epoch": 13.368953068592058, "grad_norm": 2.524991035461426, "learning_rate": 3.3155234657039714e-05, "loss": 1.4666, "step": 18516 }, { "epoch": 13.373285198555957, "grad_norm": 2.2442431449890137, "learning_rate": 3.313357400722022e-05, "loss": 1.3914, "step": 18522 }, { "epoch": 13.377617328519856, "grad_norm": 2.2966060638427734, "learning_rate": 3.311191335740073e-05, "loss": 1.459, "step": 18528 }, { "epoch": 13.381949458483755, "grad_norm": 2.2863237857818604, "learning_rate": 3.309025270758122e-05, "loss": 1.4399, "step": 18534 }, { "epoch": 13.386281588447654, "grad_norm": 2.218045234680176, "learning_rate": 3.306859205776173e-05, "loss": 1.3803, "step": 18540 }, { "epoch": 13.390613718411553, "grad_norm": 2.5849504470825195, "learning_rate": 3.3046931407942237e-05, "loss": 1.6365, "step": 18546 }, { "epoch": 13.394945848375452, "grad_norm": 2.509207010269165, "learning_rate": 3.3025270758122747e-05, "loss": 1.4431, "step": 18552 }, { "epoch": 13.39927797833935, "grad_norm": 2.5162408351898193, "learning_rate": 3.300361010830325e-05, "loss": 1.4687, "step": 18558 }, { "epoch": 13.40361010830325, "grad_norm": 2.2658798694610596, "learning_rate": 3.298194945848376e-05, "loss": 1.4603, "step": 18564 }, { "epoch": 13.407942238267148, "grad_norm": 2.1892261505126953, "learning_rate": 3.296028880866426e-05, "loss": 1.4449, "step": 18570 }, { "epoch": 13.412274368231047, "grad_norm": 2.48235821723938, "learning_rate": 3.2938628158844766e-05, "loss": 1.3517, "step": 18576 }, { "epoch": 13.416606498194946, "grad_norm": 2.277315139770508, "learning_rate": 3.291696750902527e-05, "loss": 1.5088, "step": 18582 }, { "epoch": 13.420938628158845, "grad_norm": 2.48370623588562, "learning_rate": 3.289530685920578e-05, "loss": 1.521, "step": 18588 }, { "epoch": 13.425270758122744, "grad_norm": 1.9667757749557495, "learning_rate": 3.287364620938628e-05, "loss": 1.5242, "step": 18594 }, { "epoch": 13.429602888086643, "grad_norm": 2.423464775085449, "learning_rate": 3.2851985559566786e-05, "loss": 1.5136, "step": 18600 }, { "epoch": 13.433935018050542, "grad_norm": 2.145099401473999, "learning_rate": 3.2830324909747296e-05, "loss": 1.4774, "step": 18606 }, { "epoch": 13.43826714801444, "grad_norm": 2.241647481918335, "learning_rate": 3.28086642599278e-05, "loss": 1.4733, "step": 18612 }, { "epoch": 13.442599277978339, "grad_norm": 2.4423835277557373, "learning_rate": 3.27870036101083e-05, "loss": 1.5227, "step": 18618 }, { "epoch": 13.446931407942238, "grad_norm": 2.451143980026245, "learning_rate": 3.2765342960288806e-05, "loss": 1.5281, "step": 18624 }, { "epoch": 13.451263537906136, "grad_norm": 2.3334543704986572, "learning_rate": 3.2743682310469316e-05, "loss": 1.4828, "step": 18630 }, { "epoch": 13.455595667870035, "grad_norm": 2.365832805633545, "learning_rate": 3.272202166064982e-05, "loss": 1.2962, "step": 18636 }, { "epoch": 13.459927797833934, "grad_norm": 2.4316341876983643, "learning_rate": 3.270036101083033e-05, "loss": 1.5035, "step": 18642 }, { "epoch": 13.464259927797833, "grad_norm": 2.530724048614502, "learning_rate": 3.267870036101083e-05, "loss": 1.4849, "step": 18648 }, { "epoch": 13.468592057761732, "grad_norm": 2.2115042209625244, "learning_rate": 3.265703971119134e-05, "loss": 1.3748, "step": 18654 }, { "epoch": 13.472924187725631, "grad_norm": 2.5443456172943115, "learning_rate": 3.263537906137184e-05, "loss": 1.4385, "step": 18660 }, { "epoch": 13.47725631768953, "grad_norm": 2.5096232891082764, "learning_rate": 3.261371841155235e-05, "loss": 1.4724, "step": 18666 }, { "epoch": 13.48158844765343, "grad_norm": 2.1951346397399902, "learning_rate": 3.259205776173285e-05, "loss": 1.5259, "step": 18672 }, { "epoch": 13.485920577617328, "grad_norm": 2.763258218765259, "learning_rate": 3.257039711191336e-05, "loss": 1.389, "step": 18678 }, { "epoch": 13.490252707581227, "grad_norm": 2.4510257244110107, "learning_rate": 3.2548736462093865e-05, "loss": 1.5284, "step": 18684 }, { "epoch": 13.494584837545126, "grad_norm": 2.375736713409424, "learning_rate": 3.252707581227437e-05, "loss": 1.3641, "step": 18690 }, { "epoch": 13.498916967509025, "grad_norm": 2.393035650253296, "learning_rate": 3.250541516245488e-05, "loss": 1.4932, "step": 18696 }, { "epoch": 13.503249097472924, "grad_norm": 2.3083295822143555, "learning_rate": 3.248375451263538e-05, "loss": 1.4214, "step": 18702 }, { "epoch": 13.507581227436823, "grad_norm": 2.233177661895752, "learning_rate": 3.2462093862815885e-05, "loss": 1.5546, "step": 18708 }, { "epoch": 13.511913357400722, "grad_norm": 2.4633982181549072, "learning_rate": 3.244043321299639e-05, "loss": 1.4983, "step": 18714 }, { "epoch": 13.516245487364621, "grad_norm": 2.234452724456787, "learning_rate": 3.24187725631769e-05, "loss": 1.4947, "step": 18720 }, { "epoch": 13.52057761732852, "grad_norm": 2.270704984664917, "learning_rate": 3.23971119133574e-05, "loss": 1.5021, "step": 18726 }, { "epoch": 13.52490974729242, "grad_norm": 2.3039908409118652, "learning_rate": 3.237545126353791e-05, "loss": 1.5019, "step": 18732 }, { "epoch": 13.529241877256318, "grad_norm": 2.336836099624634, "learning_rate": 3.2353790613718414e-05, "loss": 1.4755, "step": 18738 }, { "epoch": 13.533574007220217, "grad_norm": 2.514202117919922, "learning_rate": 3.233212996389892e-05, "loss": 1.4574, "step": 18744 }, { "epoch": 13.537906137184116, "grad_norm": 2.2948591709136963, "learning_rate": 3.231046931407942e-05, "loss": 1.4302, "step": 18750 }, { "epoch": 13.542238267148015, "grad_norm": 2.345107078552246, "learning_rate": 3.228880866425993e-05, "loss": 1.4747, "step": 18756 }, { "epoch": 13.546570397111914, "grad_norm": 2.18506121635437, "learning_rate": 3.2267148014440434e-05, "loss": 1.4418, "step": 18762 }, { "epoch": 13.550902527075813, "grad_norm": 2.2980399131774902, "learning_rate": 3.2245487364620944e-05, "loss": 1.5107, "step": 18768 }, { "epoch": 13.555234657039712, "grad_norm": 2.3648006916046143, "learning_rate": 3.222382671480145e-05, "loss": 1.5538, "step": 18774 }, { "epoch": 13.559566787003611, "grad_norm": 2.5538294315338135, "learning_rate": 3.220216606498195e-05, "loss": 1.4523, "step": 18780 }, { "epoch": 13.56389891696751, "grad_norm": 2.404728651046753, "learning_rate": 3.2180505415162454e-05, "loss": 1.5172, "step": 18786 }, { "epoch": 13.568231046931407, "grad_norm": 2.5776402950286865, "learning_rate": 3.2158844765342964e-05, "loss": 1.3649, "step": 18792 }, { "epoch": 13.572563176895306, "grad_norm": 2.398364543914795, "learning_rate": 3.213718411552347e-05, "loss": 1.3563, "step": 18798 }, { "epoch": 13.576895306859205, "grad_norm": 2.3687894344329834, "learning_rate": 3.211552346570397e-05, "loss": 1.5466, "step": 18804 }, { "epoch": 13.581227436823104, "grad_norm": 2.3752589225769043, "learning_rate": 3.209386281588448e-05, "loss": 1.4941, "step": 18810 }, { "epoch": 13.585559566787003, "grad_norm": 2.262075901031494, "learning_rate": 3.207220216606498e-05, "loss": 1.6197, "step": 18816 }, { "epoch": 13.589891696750902, "grad_norm": 2.31484317779541, "learning_rate": 3.205054151624549e-05, "loss": 1.488, "step": 18822 }, { "epoch": 13.594223826714801, "grad_norm": 2.3276734352111816, "learning_rate": 3.202888086642599e-05, "loss": 1.3686, "step": 18828 }, { "epoch": 13.5985559566787, "grad_norm": 2.5848398208618164, "learning_rate": 3.20072202166065e-05, "loss": 1.4361, "step": 18834 }, { "epoch": 13.602888086642599, "grad_norm": 2.093142032623291, "learning_rate": 3.1985559566787e-05, "loss": 1.5321, "step": 18840 }, { "epoch": 13.607220216606498, "grad_norm": 2.3628313541412354, "learning_rate": 3.196389891696751e-05, "loss": 1.4696, "step": 18846 }, { "epoch": 13.611552346570397, "grad_norm": 2.450303316116333, "learning_rate": 3.1942238267148016e-05, "loss": 1.3527, "step": 18852 }, { "epoch": 13.615884476534296, "grad_norm": 2.386425256729126, "learning_rate": 3.1920577617328526e-05, "loss": 1.4949, "step": 18858 }, { "epoch": 13.620216606498195, "grad_norm": 2.4949283599853516, "learning_rate": 3.189891696750903e-05, "loss": 1.4462, "step": 18864 }, { "epoch": 13.624548736462094, "grad_norm": 2.528071641921997, "learning_rate": 3.187725631768953e-05, "loss": 1.4923, "step": 18870 }, { "epoch": 13.628880866425993, "grad_norm": 2.1987948417663574, "learning_rate": 3.1855595667870036e-05, "loss": 1.5145, "step": 18876 }, { "epoch": 13.633212996389892, "grad_norm": 2.6242713928222656, "learning_rate": 3.183393501805054e-05, "loss": 1.5335, "step": 18882 }, { "epoch": 13.63754512635379, "grad_norm": 2.411984443664551, "learning_rate": 3.181227436823105e-05, "loss": 1.5959, "step": 18888 }, { "epoch": 13.64187725631769, "grad_norm": 2.155118942260742, "learning_rate": 3.179061371841155e-05, "loss": 1.4719, "step": 18894 }, { "epoch": 13.646209386281589, "grad_norm": 2.51487398147583, "learning_rate": 3.176895306859206e-05, "loss": 1.5689, "step": 18900 }, { "epoch": 13.650541516245488, "grad_norm": 2.1515471935272217, "learning_rate": 3.1747292418772566e-05, "loss": 1.5366, "step": 18906 }, { "epoch": 13.654873646209387, "grad_norm": 2.484970808029175, "learning_rate": 3.172563176895307e-05, "loss": 1.4226, "step": 18912 }, { "epoch": 13.659205776173286, "grad_norm": 2.398064136505127, "learning_rate": 3.170397111913357e-05, "loss": 1.5992, "step": 18918 }, { "epoch": 13.663537906137185, "grad_norm": 2.372020721435547, "learning_rate": 3.168231046931408e-05, "loss": 1.3938, "step": 18924 }, { "epoch": 13.667870036101084, "grad_norm": 2.5632858276367188, "learning_rate": 3.1660649819494585e-05, "loss": 1.5454, "step": 18930 }, { "epoch": 13.672202166064983, "grad_norm": 2.4690768718719482, "learning_rate": 3.1638989169675095e-05, "loss": 1.5539, "step": 18936 }, { "epoch": 13.676534296028882, "grad_norm": 2.2488656044006348, "learning_rate": 3.16173285198556e-05, "loss": 1.4747, "step": 18942 }, { "epoch": 13.68086642599278, "grad_norm": 2.1845052242279053, "learning_rate": 3.159566787003611e-05, "loss": 1.5058, "step": 18948 }, { "epoch": 13.685198555956678, "grad_norm": 2.4021811485290527, "learning_rate": 3.1574007220216605e-05, "loss": 1.4261, "step": 18954 }, { "epoch": 13.689530685920577, "grad_norm": 2.1812586784362793, "learning_rate": 3.1552346570397115e-05, "loss": 1.4011, "step": 18960 }, { "epoch": 13.693862815884476, "grad_norm": 2.587625503540039, "learning_rate": 3.153068592057762e-05, "loss": 1.3952, "step": 18966 }, { "epoch": 13.698194945848375, "grad_norm": 2.299161195755005, "learning_rate": 3.150902527075812e-05, "loss": 1.5411, "step": 18972 }, { "epoch": 13.702527075812274, "grad_norm": 2.459404468536377, "learning_rate": 3.148736462093863e-05, "loss": 1.5799, "step": 18978 }, { "epoch": 13.706859205776173, "grad_norm": 2.4458231925964355, "learning_rate": 3.1465703971119135e-05, "loss": 1.5072, "step": 18984 }, { "epoch": 13.711191335740072, "grad_norm": 2.551158905029297, "learning_rate": 3.1444043321299645e-05, "loss": 1.5412, "step": 18990 }, { "epoch": 13.71552346570397, "grad_norm": 2.1956348419189453, "learning_rate": 3.142238267148014e-05, "loss": 1.4637, "step": 18996 }, { "epoch": 13.71985559566787, "grad_norm": 2.3660218715667725, "learning_rate": 3.140072202166065e-05, "loss": 1.5089, "step": 19002 }, { "epoch": 13.724187725631769, "grad_norm": 2.332493543624878, "learning_rate": 3.1379061371841154e-05, "loss": 1.4577, "step": 19008 }, { "epoch": 13.728519855595668, "grad_norm": 2.3256678581237793, "learning_rate": 3.1357400722021664e-05, "loss": 1.4183, "step": 19014 }, { "epoch": 13.732851985559567, "grad_norm": 2.354607105255127, "learning_rate": 3.133574007220217e-05, "loss": 1.5023, "step": 19020 }, { "epoch": 13.737184115523466, "grad_norm": 2.5320186614990234, "learning_rate": 3.131407942238268e-05, "loss": 1.4872, "step": 19026 }, { "epoch": 13.741516245487365, "grad_norm": 2.318357467651367, "learning_rate": 3.1292418772563174e-05, "loss": 1.5935, "step": 19032 }, { "epoch": 13.745848375451263, "grad_norm": 1.928601861000061, "learning_rate": 3.1270758122743684e-05, "loss": 1.3972, "step": 19038 }, { "epoch": 13.750180505415162, "grad_norm": 2.3618698120117188, "learning_rate": 3.124909747292419e-05, "loss": 1.4362, "step": 19044 }, { "epoch": 13.754512635379061, "grad_norm": 2.2745556831359863, "learning_rate": 3.12274368231047e-05, "loss": 1.5476, "step": 19050 }, { "epoch": 13.75884476534296, "grad_norm": 2.4522690773010254, "learning_rate": 3.12057761732852e-05, "loss": 1.5356, "step": 19056 }, { "epoch": 13.76317689530686, "grad_norm": 2.202099084854126, "learning_rate": 3.1184115523465704e-05, "loss": 1.4974, "step": 19062 }, { "epoch": 13.767509025270758, "grad_norm": 2.2513270378112793, "learning_rate": 3.1162454873646214e-05, "loss": 1.4335, "step": 19068 }, { "epoch": 13.771841155234657, "grad_norm": 2.4167439937591553, "learning_rate": 3.114079422382672e-05, "loss": 1.5026, "step": 19074 }, { "epoch": 13.776173285198556, "grad_norm": 2.310234308242798, "learning_rate": 3.111913357400722e-05, "loss": 1.5535, "step": 19080 }, { "epoch": 13.780505415162455, "grad_norm": 2.470089912414551, "learning_rate": 3.109747292418772e-05, "loss": 1.5566, "step": 19086 }, { "epoch": 13.784837545126354, "grad_norm": 2.462695360183716, "learning_rate": 3.107581227436823e-05, "loss": 1.4656, "step": 19092 }, { "epoch": 13.789169675090253, "grad_norm": 2.2631962299346924, "learning_rate": 3.1054151624548737e-05, "loss": 1.4944, "step": 19098 }, { "epoch": 13.793501805054152, "grad_norm": 2.5170180797576904, "learning_rate": 3.1032490974729247e-05, "loss": 1.5272, "step": 19104 }, { "epoch": 13.797833935018051, "grad_norm": 2.249520778656006, "learning_rate": 3.101083032490975e-05, "loss": 1.4921, "step": 19110 }, { "epoch": 13.80216606498195, "grad_norm": 2.3539414405822754, "learning_rate": 3.098916967509025e-05, "loss": 1.4591, "step": 19116 }, { "epoch": 13.80649819494585, "grad_norm": 2.673633098602295, "learning_rate": 3.0967509025270756e-05, "loss": 1.47, "step": 19122 }, { "epoch": 13.810830324909748, "grad_norm": 2.4399871826171875, "learning_rate": 3.0945848375451266e-05, "loss": 1.4977, "step": 19128 }, { "epoch": 13.815162454873647, "grad_norm": 2.220182180404663, "learning_rate": 3.092418772563177e-05, "loss": 1.5725, "step": 19134 }, { "epoch": 13.819494584837544, "grad_norm": 2.5701284408569336, "learning_rate": 3.090252707581228e-05, "loss": 1.4756, "step": 19140 }, { "epoch": 13.823826714801443, "grad_norm": 2.430147647857666, "learning_rate": 3.088086642599278e-05, "loss": 1.5555, "step": 19146 }, { "epoch": 13.828158844765342, "grad_norm": 2.3403923511505127, "learning_rate": 3.0859205776173286e-05, "loss": 1.3735, "step": 19152 }, { "epoch": 13.832490974729241, "grad_norm": 2.5123207569122314, "learning_rate": 3.083754512635379e-05, "loss": 1.4953, "step": 19158 }, { "epoch": 13.83682310469314, "grad_norm": 2.043949604034424, "learning_rate": 3.08158844765343e-05, "loss": 1.5579, "step": 19164 }, { "epoch": 13.84115523465704, "grad_norm": 2.332061767578125, "learning_rate": 3.07942238267148e-05, "loss": 1.5279, "step": 19170 }, { "epoch": 13.845487364620938, "grad_norm": 2.5744845867156982, "learning_rate": 3.0772563176895306e-05, "loss": 1.4795, "step": 19176 }, { "epoch": 13.849819494584837, "grad_norm": 2.3096444606781006, "learning_rate": 3.0750902527075816e-05, "loss": 1.504, "step": 19182 }, { "epoch": 13.854151624548736, "grad_norm": 2.19189453125, "learning_rate": 3.072924187725632e-05, "loss": 1.5318, "step": 19188 }, { "epoch": 13.858483754512635, "grad_norm": 2.3442087173461914, "learning_rate": 3.070758122743683e-05, "loss": 1.4664, "step": 19194 }, { "epoch": 13.862815884476534, "grad_norm": 2.5518274307250977, "learning_rate": 3.0685920577617325e-05, "loss": 1.5412, "step": 19200 }, { "epoch": 13.867148014440433, "grad_norm": 2.4279301166534424, "learning_rate": 3.0664259927797835e-05, "loss": 1.576, "step": 19206 }, { "epoch": 13.871480144404332, "grad_norm": 2.5395307540893555, "learning_rate": 3.064259927797834e-05, "loss": 1.5183, "step": 19212 }, { "epoch": 13.875812274368231, "grad_norm": 2.236992359161377, "learning_rate": 3.062093862815885e-05, "loss": 1.4192, "step": 19218 }, { "epoch": 13.88014440433213, "grad_norm": 2.276650905609131, "learning_rate": 3.059927797833935e-05, "loss": 1.6493, "step": 19224 }, { "epoch": 13.884476534296029, "grad_norm": 2.498889684677124, "learning_rate": 3.057761732851986e-05, "loss": 1.4913, "step": 19230 }, { "epoch": 13.888808664259928, "grad_norm": 2.695711135864258, "learning_rate": 3.0555956678700365e-05, "loss": 1.4318, "step": 19236 }, { "epoch": 13.893140794223827, "grad_norm": 2.438796281814575, "learning_rate": 3.053429602888087e-05, "loss": 1.5854, "step": 19242 }, { "epoch": 13.897472924187726, "grad_norm": 2.551981210708618, "learning_rate": 3.051263537906137e-05, "loss": 1.4811, "step": 19248 }, { "epoch": 13.901805054151625, "grad_norm": 2.418126106262207, "learning_rate": 3.0490974729241878e-05, "loss": 1.5647, "step": 19254 }, { "epoch": 13.906137184115524, "grad_norm": 2.3526108264923096, "learning_rate": 3.0469314079422385e-05, "loss": 1.4001, "step": 19260 }, { "epoch": 13.910469314079423, "grad_norm": 2.2977399826049805, "learning_rate": 3.044765342960289e-05, "loss": 1.5354, "step": 19266 }, { "epoch": 13.914801444043322, "grad_norm": 2.324551582336426, "learning_rate": 3.0425992779783398e-05, "loss": 1.5015, "step": 19272 }, { "epoch": 13.91913357400722, "grad_norm": 2.4229979515075684, "learning_rate": 3.0404332129963904e-05, "loss": 1.4828, "step": 19278 }, { "epoch": 13.92346570397112, "grad_norm": 2.386613130569458, "learning_rate": 3.0382671480144404e-05, "loss": 1.5343, "step": 19284 }, { "epoch": 13.927797833935019, "grad_norm": 2.233022689819336, "learning_rate": 3.036101083032491e-05, "loss": 1.5422, "step": 19290 }, { "epoch": 13.932129963898918, "grad_norm": 2.228231430053711, "learning_rate": 3.0339350180505417e-05, "loss": 1.4214, "step": 19296 }, { "epoch": 13.936462093862815, "grad_norm": 2.3326590061187744, "learning_rate": 3.031768953068592e-05, "loss": 1.4283, "step": 19302 }, { "epoch": 13.940794223826714, "grad_norm": 2.46234130859375, "learning_rate": 3.0296028880866427e-05, "loss": 1.5212, "step": 19308 }, { "epoch": 13.945126353790613, "grad_norm": 2.300708293914795, "learning_rate": 3.0274368231046934e-05, "loss": 1.5044, "step": 19314 }, { "epoch": 13.949458483754512, "grad_norm": 2.418933629989624, "learning_rate": 3.025270758122744e-05, "loss": 1.4087, "step": 19320 }, { "epoch": 13.953790613718411, "grad_norm": 2.5231997966766357, "learning_rate": 3.023104693140794e-05, "loss": 1.5239, "step": 19326 }, { "epoch": 13.95812274368231, "grad_norm": 2.102337121963501, "learning_rate": 3.0209386281588447e-05, "loss": 1.4315, "step": 19332 }, { "epoch": 13.962454873646209, "grad_norm": 2.250791549682617, "learning_rate": 3.0187725631768954e-05, "loss": 1.5168, "step": 19338 }, { "epoch": 13.966787003610108, "grad_norm": 2.3754684925079346, "learning_rate": 3.016606498194946e-05, "loss": 1.4457, "step": 19344 }, { "epoch": 13.971119133574007, "grad_norm": 2.05521297454834, "learning_rate": 3.0144404332129967e-05, "loss": 1.5602, "step": 19350 }, { "epoch": 13.975451263537906, "grad_norm": 2.600750684738159, "learning_rate": 3.0122743682310473e-05, "loss": 1.4749, "step": 19356 }, { "epoch": 13.979783393501805, "grad_norm": 2.292548656463623, "learning_rate": 3.010108303249098e-05, "loss": 1.5698, "step": 19362 }, { "epoch": 13.984115523465704, "grad_norm": 2.300550937652588, "learning_rate": 3.007942238267148e-05, "loss": 1.4534, "step": 19368 }, { "epoch": 13.988447653429603, "grad_norm": 2.419433116912842, "learning_rate": 3.0057761732851987e-05, "loss": 1.5361, "step": 19374 }, { "epoch": 13.992779783393502, "grad_norm": 2.191162347793579, "learning_rate": 3.0036101083032493e-05, "loss": 1.5427, "step": 19380 }, { "epoch": 13.9971119133574, "grad_norm": 2.4308078289031982, "learning_rate": 3.0014440433212996e-05, "loss": 1.5377, "step": 19386 }, { "epoch": 14.0014440433213, "grad_norm": 2.6629650592803955, "learning_rate": 2.9992779783393503e-05, "loss": 1.5798, "step": 19392 }, { "epoch": 14.005776173285199, "grad_norm": 2.187110185623169, "learning_rate": 2.997111913357401e-05, "loss": 1.5018, "step": 19398 }, { "epoch": 14.010108303249098, "grad_norm": 2.499845027923584, "learning_rate": 2.9949458483754516e-05, "loss": 1.5367, "step": 19404 }, { "epoch": 14.014440433212997, "grad_norm": 2.5792362689971924, "learning_rate": 2.9927797833935016e-05, "loss": 1.2945, "step": 19410 }, { "epoch": 14.018772563176896, "grad_norm": 2.292320966720581, "learning_rate": 2.9906137184115523e-05, "loss": 1.3294, "step": 19416 }, { "epoch": 14.023104693140795, "grad_norm": 2.313807725906372, "learning_rate": 2.988447653429603e-05, "loss": 1.3745, "step": 19422 }, { "epoch": 14.027436823104694, "grad_norm": 2.4258663654327393, "learning_rate": 2.9862815884476536e-05, "loss": 1.3938, "step": 19428 }, { "epoch": 14.031768953068593, "grad_norm": 2.704780340194702, "learning_rate": 2.9841155234657042e-05, "loss": 1.4824, "step": 19434 }, { "epoch": 14.036101083032491, "grad_norm": 2.596135139465332, "learning_rate": 2.981949458483755e-05, "loss": 1.5102, "step": 19440 }, { "epoch": 14.04043321299639, "grad_norm": 2.122591733932495, "learning_rate": 2.9797833935018056e-05, "loss": 1.4224, "step": 19446 }, { "epoch": 14.04476534296029, "grad_norm": 2.3419442176818848, "learning_rate": 2.9776173285198556e-05, "loss": 1.4202, "step": 19452 }, { "epoch": 14.049097472924188, "grad_norm": 2.1176674365997314, "learning_rate": 2.9754512635379062e-05, "loss": 1.5044, "step": 19458 }, { "epoch": 14.053429602888087, "grad_norm": 2.3635165691375732, "learning_rate": 2.973285198555957e-05, "loss": 1.542, "step": 19464 }, { "epoch": 14.057761732851986, "grad_norm": 2.3322675228118896, "learning_rate": 2.9711191335740075e-05, "loss": 1.5983, "step": 19470 }, { "epoch": 14.062093862815885, "grad_norm": 2.1735637187957764, "learning_rate": 2.968953068592058e-05, "loss": 1.4962, "step": 19476 }, { "epoch": 14.066425992779783, "grad_norm": 2.510464668273926, "learning_rate": 2.9667870036101085e-05, "loss": 1.4673, "step": 19482 }, { "epoch": 14.070758122743682, "grad_norm": 2.2673418521881104, "learning_rate": 2.9646209386281592e-05, "loss": 1.4986, "step": 19488 }, { "epoch": 14.07509025270758, "grad_norm": 2.3771560192108154, "learning_rate": 2.962454873646209e-05, "loss": 1.4067, "step": 19494 }, { "epoch": 14.07942238267148, "grad_norm": 2.2006163597106934, "learning_rate": 2.9602888086642598e-05, "loss": 1.4202, "step": 19500 }, { "epoch": 14.083754512635378, "grad_norm": 2.2613768577575684, "learning_rate": 2.9581227436823105e-05, "loss": 1.4666, "step": 19506 }, { "epoch": 14.088086642599277, "grad_norm": 2.6549415588378906, "learning_rate": 2.955956678700361e-05, "loss": 1.4362, "step": 19512 }, { "epoch": 14.092418772563176, "grad_norm": 2.323676347732544, "learning_rate": 2.9537906137184118e-05, "loss": 1.4404, "step": 19518 }, { "epoch": 14.096750902527075, "grad_norm": 2.33064603805542, "learning_rate": 2.9516245487364625e-05, "loss": 1.4333, "step": 19524 }, { "epoch": 14.101083032490974, "grad_norm": 2.2084312438964844, "learning_rate": 2.9494584837545125e-05, "loss": 1.3955, "step": 19530 }, { "epoch": 14.105415162454873, "grad_norm": 2.090564250946045, "learning_rate": 2.947292418772563e-05, "loss": 1.372, "step": 19536 }, { "epoch": 14.109747292418772, "grad_norm": 2.5013561248779297, "learning_rate": 2.9451263537906138e-05, "loss": 1.5169, "step": 19542 }, { "epoch": 14.114079422382671, "grad_norm": 2.348332643508911, "learning_rate": 2.9429602888086644e-05, "loss": 1.4772, "step": 19548 }, { "epoch": 14.11841155234657, "grad_norm": 2.3805110454559326, "learning_rate": 2.940794223826715e-05, "loss": 1.4323, "step": 19554 }, { "epoch": 14.12274368231047, "grad_norm": 2.354691982269287, "learning_rate": 2.9386281588447658e-05, "loss": 1.5031, "step": 19560 }, { "epoch": 14.127075812274368, "grad_norm": 2.4196670055389404, "learning_rate": 2.936462093862816e-05, "loss": 1.536, "step": 19566 }, { "epoch": 14.131407942238267, "grad_norm": 2.176624059677124, "learning_rate": 2.9342960288808664e-05, "loss": 1.5365, "step": 19572 }, { "epoch": 14.135740072202166, "grad_norm": 2.0556905269622803, "learning_rate": 2.932129963898917e-05, "loss": 1.4404, "step": 19578 }, { "epoch": 14.140072202166065, "grad_norm": 2.163966178894043, "learning_rate": 2.9299638989169674e-05, "loss": 1.4502, "step": 19584 }, { "epoch": 14.144404332129964, "grad_norm": 2.241837978363037, "learning_rate": 2.927797833935018e-05, "loss": 1.4315, "step": 19590 }, { "epoch": 14.148736462093863, "grad_norm": 2.7219719886779785, "learning_rate": 2.9256317689530687e-05, "loss": 1.3432, "step": 19596 }, { "epoch": 14.153068592057762, "grad_norm": 2.421816110610962, "learning_rate": 2.9234657039711194e-05, "loss": 1.4081, "step": 19602 }, { "epoch": 14.157400722021661, "grad_norm": 2.411367177963257, "learning_rate": 2.92129963898917e-05, "loss": 1.5173, "step": 19608 }, { "epoch": 14.16173285198556, "grad_norm": 2.116854429244995, "learning_rate": 2.91913357400722e-05, "loss": 1.4394, "step": 19614 }, { "epoch": 14.166064981949459, "grad_norm": 2.6103620529174805, "learning_rate": 2.9169675090252707e-05, "loss": 1.4986, "step": 19620 }, { "epoch": 14.170397111913358, "grad_norm": 2.233997344970703, "learning_rate": 2.9148014440433213e-05, "loss": 1.3658, "step": 19626 }, { "epoch": 14.174729241877257, "grad_norm": 2.3356101512908936, "learning_rate": 2.912635379061372e-05, "loss": 1.5644, "step": 19632 }, { "epoch": 14.179061371841156, "grad_norm": 2.588231325149536, "learning_rate": 2.9104693140794227e-05, "loss": 1.5595, "step": 19638 }, { "epoch": 14.183393501805055, "grad_norm": 2.083523750305176, "learning_rate": 2.9083032490974733e-05, "loss": 1.4286, "step": 19644 }, { "epoch": 14.187725631768952, "grad_norm": 2.4975109100341797, "learning_rate": 2.906137184115524e-05, "loss": 1.427, "step": 19650 }, { "epoch": 14.192057761732851, "grad_norm": 2.4734725952148438, "learning_rate": 2.903971119133574e-05, "loss": 1.4764, "step": 19656 }, { "epoch": 14.19638989169675, "grad_norm": 2.4082367420196533, "learning_rate": 2.9018050541516246e-05, "loss": 1.5224, "step": 19662 }, { "epoch": 14.20072202166065, "grad_norm": 2.208632230758667, "learning_rate": 2.8996389891696753e-05, "loss": 1.2798, "step": 19668 }, { "epoch": 14.205054151624548, "grad_norm": 2.3908939361572266, "learning_rate": 2.8974729241877256e-05, "loss": 1.4739, "step": 19674 }, { "epoch": 14.209386281588447, "grad_norm": 2.347486972808838, "learning_rate": 2.8953068592057763e-05, "loss": 1.5098, "step": 19680 }, { "epoch": 14.213718411552346, "grad_norm": 2.4909915924072266, "learning_rate": 2.893140794223827e-05, "loss": 1.3757, "step": 19686 }, { "epoch": 14.218050541516245, "grad_norm": 2.3680546283721924, "learning_rate": 2.8909747292418776e-05, "loss": 1.4327, "step": 19692 }, { "epoch": 14.222382671480144, "grad_norm": 2.495722770690918, "learning_rate": 2.8888086642599276e-05, "loss": 1.4841, "step": 19698 }, { "epoch": 14.226714801444043, "grad_norm": 2.8317084312438965, "learning_rate": 2.8866425992779782e-05, "loss": 1.5106, "step": 19704 }, { "epoch": 14.231046931407942, "grad_norm": 2.4486451148986816, "learning_rate": 2.884476534296029e-05, "loss": 1.579, "step": 19710 }, { "epoch": 14.235379061371841, "grad_norm": 2.3912971019744873, "learning_rate": 2.8823104693140796e-05, "loss": 1.4577, "step": 19716 }, { "epoch": 14.23971119133574, "grad_norm": 2.230178117752075, "learning_rate": 2.8801444043321302e-05, "loss": 1.3326, "step": 19722 }, { "epoch": 14.244043321299639, "grad_norm": 2.393622398376465, "learning_rate": 2.877978339350181e-05, "loss": 1.4802, "step": 19728 }, { "epoch": 14.248375451263538, "grad_norm": 2.491215467453003, "learning_rate": 2.8758122743682316e-05, "loss": 1.3903, "step": 19734 }, { "epoch": 14.252707581227437, "grad_norm": 2.465266227722168, "learning_rate": 2.8736462093862815e-05, "loss": 1.4212, "step": 19740 }, { "epoch": 14.257039711191336, "grad_norm": 2.347666025161743, "learning_rate": 2.8714801444043322e-05, "loss": 1.3816, "step": 19746 }, { "epoch": 14.261371841155235, "grad_norm": 2.740046977996826, "learning_rate": 2.869314079422383e-05, "loss": 1.6825, "step": 19752 }, { "epoch": 14.265703971119134, "grad_norm": 2.2589528560638428, "learning_rate": 2.8671480144404332e-05, "loss": 1.3857, "step": 19758 }, { "epoch": 14.270036101083033, "grad_norm": 2.498661756515503, "learning_rate": 2.864981949458484e-05, "loss": 1.5341, "step": 19764 }, { "epoch": 14.274368231046932, "grad_norm": 2.5064573287963867, "learning_rate": 2.8628158844765345e-05, "loss": 1.5471, "step": 19770 }, { "epoch": 14.27870036101083, "grad_norm": 2.2615277767181396, "learning_rate": 2.860649819494585e-05, "loss": 1.4269, "step": 19776 }, { "epoch": 14.28303249097473, "grad_norm": 2.5111160278320312, "learning_rate": 2.858483754512635e-05, "loss": 1.3764, "step": 19782 }, { "epoch": 14.287364620938629, "grad_norm": 2.230556011199951, "learning_rate": 2.8563176895306858e-05, "loss": 1.5299, "step": 19788 }, { "epoch": 14.291696750902528, "grad_norm": 2.447625160217285, "learning_rate": 2.8541516245487365e-05, "loss": 1.467, "step": 19794 }, { "epoch": 14.296028880866427, "grad_norm": 2.7095797061920166, "learning_rate": 2.851985559566787e-05, "loss": 1.4185, "step": 19800 }, { "epoch": 14.300361010830326, "grad_norm": 2.1797595024108887, "learning_rate": 2.8498194945848378e-05, "loss": 1.4862, "step": 19806 }, { "epoch": 14.304693140794225, "grad_norm": 2.2840168476104736, "learning_rate": 2.8476534296028885e-05, "loss": 1.5487, "step": 19812 }, { "epoch": 14.309025270758124, "grad_norm": 2.433520555496216, "learning_rate": 2.845487364620939e-05, "loss": 1.5597, "step": 19818 }, { "epoch": 14.313357400722023, "grad_norm": 2.496788740158081, "learning_rate": 2.843321299638989e-05, "loss": 1.4138, "step": 19824 }, { "epoch": 14.31768953068592, "grad_norm": 2.6495628356933594, "learning_rate": 2.8411552346570398e-05, "loss": 1.4385, "step": 19830 }, { "epoch": 14.322021660649819, "grad_norm": 2.487288475036621, "learning_rate": 2.8389891696750904e-05, "loss": 1.4718, "step": 19836 }, { "epoch": 14.326353790613718, "grad_norm": 2.554415464401245, "learning_rate": 2.836823104693141e-05, "loss": 1.4213, "step": 19842 }, { "epoch": 14.330685920577617, "grad_norm": 2.5619688034057617, "learning_rate": 2.8346570397111914e-05, "loss": 1.486, "step": 19848 }, { "epoch": 14.335018050541516, "grad_norm": 2.381714105606079, "learning_rate": 2.832490974729242e-05, "loss": 1.4622, "step": 19854 }, { "epoch": 14.339350180505415, "grad_norm": 2.3318538665771484, "learning_rate": 2.8303249097472927e-05, "loss": 1.3977, "step": 19860 }, { "epoch": 14.343682310469314, "grad_norm": 2.527156114578247, "learning_rate": 2.8281588447653427e-05, "loss": 1.512, "step": 19866 }, { "epoch": 14.348014440433213, "grad_norm": 2.327470064163208, "learning_rate": 2.8259927797833934e-05, "loss": 1.482, "step": 19872 }, { "epoch": 14.352346570397112, "grad_norm": 2.367640733718872, "learning_rate": 2.823826714801444e-05, "loss": 1.5918, "step": 19878 }, { "epoch": 14.35667870036101, "grad_norm": 2.578782081604004, "learning_rate": 2.8216606498194947e-05, "loss": 1.5786, "step": 19884 }, { "epoch": 14.36101083032491, "grad_norm": 2.2925124168395996, "learning_rate": 2.8194945848375454e-05, "loss": 1.4802, "step": 19890 }, { "epoch": 14.365342960288809, "grad_norm": 1.8996590375900269, "learning_rate": 2.817328519855596e-05, "loss": 1.3185, "step": 19896 }, { "epoch": 14.369675090252708, "grad_norm": 2.48936128616333, "learning_rate": 2.8151624548736467e-05, "loss": 1.5684, "step": 19902 }, { "epoch": 14.374007220216606, "grad_norm": 2.298769235610962, "learning_rate": 2.8129963898916967e-05, "loss": 1.4119, "step": 19908 }, { "epoch": 14.378339350180505, "grad_norm": 2.6371541023254395, "learning_rate": 2.8108303249097473e-05, "loss": 1.4673, "step": 19914 }, { "epoch": 14.382671480144404, "grad_norm": 2.0818123817443848, "learning_rate": 2.808664259927798e-05, "loss": 1.4261, "step": 19920 }, { "epoch": 14.387003610108303, "grad_norm": 2.4026570320129395, "learning_rate": 2.8064981949458487e-05, "loss": 1.5253, "step": 19926 }, { "epoch": 14.391335740072202, "grad_norm": 2.2946457862854004, "learning_rate": 2.8043321299638993e-05, "loss": 1.4669, "step": 19932 }, { "epoch": 14.395667870036101, "grad_norm": 2.3421497344970703, "learning_rate": 2.8021660649819496e-05, "loss": 1.4572, "step": 19938 }, { "epoch": 14.4, "grad_norm": 2.603607416152954, "learning_rate": 2.8000000000000003e-05, "loss": 1.5008, "step": 19944 }, { "epoch": 14.4043321299639, "grad_norm": 2.506265878677368, "learning_rate": 2.7978339350180506e-05, "loss": 1.482, "step": 19950 }, { "epoch": 14.408664259927798, "grad_norm": 2.4200077056884766, "learning_rate": 2.795667870036101e-05, "loss": 1.3937, "step": 19956 }, { "epoch": 14.412996389891697, "grad_norm": 2.350249767303467, "learning_rate": 2.7935018050541516e-05, "loss": 1.4874, "step": 19962 }, { "epoch": 14.417328519855596, "grad_norm": 2.1460132598876953, "learning_rate": 2.7913357400722023e-05, "loss": 1.3677, "step": 19968 }, { "epoch": 14.421660649819495, "grad_norm": 2.4269189834594727, "learning_rate": 2.789169675090253e-05, "loss": 1.3991, "step": 19974 }, { "epoch": 14.425992779783394, "grad_norm": 2.383573532104492, "learning_rate": 2.7870036101083036e-05, "loss": 1.4792, "step": 19980 }, { "epoch": 14.430324909747293, "grad_norm": 2.432532548904419, "learning_rate": 2.7848375451263542e-05, "loss": 1.5154, "step": 19986 }, { "epoch": 14.434657039711192, "grad_norm": 2.6265552043914795, "learning_rate": 2.7826714801444042e-05, "loss": 1.4742, "step": 19992 }, { "epoch": 14.43898916967509, "grad_norm": 2.6067464351654053, "learning_rate": 2.780505415162455e-05, "loss": 1.3913, "step": 19998 }, { "epoch": 14.443321299638988, "grad_norm": 2.5310518741607666, "learning_rate": 2.7783393501805056e-05, "loss": 1.3859, "step": 20004 }, { "epoch": 14.447653429602887, "grad_norm": 2.118581771850586, "learning_rate": 2.7761732851985562e-05, "loss": 1.5728, "step": 20010 }, { "epoch": 14.451985559566786, "grad_norm": 2.6203253269195557, "learning_rate": 2.774007220216607e-05, "loss": 1.3332, "step": 20016 }, { "epoch": 14.456317689530685, "grad_norm": 2.477724552154541, "learning_rate": 2.7718411552346575e-05, "loss": 1.5344, "step": 20022 }, { "epoch": 14.460649819494584, "grad_norm": 2.4663374423980713, "learning_rate": 2.7696750902527075e-05, "loss": 1.4563, "step": 20028 }, { "epoch": 14.464981949458483, "grad_norm": 2.5534591674804688, "learning_rate": 2.7675090252707582e-05, "loss": 1.5151, "step": 20034 }, { "epoch": 14.469314079422382, "grad_norm": 2.5771102905273438, "learning_rate": 2.765342960288809e-05, "loss": 1.5463, "step": 20040 }, { "epoch": 14.473646209386281, "grad_norm": 2.46846866607666, "learning_rate": 2.763176895306859e-05, "loss": 1.5604, "step": 20046 }, { "epoch": 14.47797833935018, "grad_norm": 2.0965490341186523, "learning_rate": 2.7610108303249098e-05, "loss": 1.3719, "step": 20052 }, { "epoch": 14.48231046931408, "grad_norm": 2.612306833267212, "learning_rate": 2.7588447653429605e-05, "loss": 1.4059, "step": 20058 }, { "epoch": 14.486642599277978, "grad_norm": 2.3727149963378906, "learning_rate": 2.756678700361011e-05, "loss": 1.535, "step": 20064 }, { "epoch": 14.490974729241877, "grad_norm": 2.3062777519226074, "learning_rate": 2.754512635379061e-05, "loss": 1.3422, "step": 20070 }, { "epoch": 14.495306859205776, "grad_norm": 2.3639566898345947, "learning_rate": 2.7523465703971118e-05, "loss": 1.4212, "step": 20076 }, { "epoch": 14.499638989169675, "grad_norm": 2.326611280441284, "learning_rate": 2.7501805054151625e-05, "loss": 1.4924, "step": 20082 }, { "epoch": 14.503971119133574, "grad_norm": 2.462536096572876, "learning_rate": 2.748014440433213e-05, "loss": 1.4204, "step": 20088 }, { "epoch": 14.508303249097473, "grad_norm": 2.326033353805542, "learning_rate": 2.7458483754512638e-05, "loss": 1.4514, "step": 20094 }, { "epoch": 14.512635379061372, "grad_norm": 2.3501315116882324, "learning_rate": 2.7436823104693144e-05, "loss": 1.3817, "step": 20100 }, { "epoch": 14.516967509025271, "grad_norm": 2.7027487754821777, "learning_rate": 2.741516245487365e-05, "loss": 1.4775, "step": 20106 }, { "epoch": 14.52129963898917, "grad_norm": 2.4221246242523193, "learning_rate": 2.739350180505415e-05, "loss": 1.4247, "step": 20112 }, { "epoch": 14.525631768953069, "grad_norm": 2.230044364929199, "learning_rate": 2.7371841155234657e-05, "loss": 1.4644, "step": 20118 }, { "epoch": 14.529963898916968, "grad_norm": 2.35213303565979, "learning_rate": 2.7350180505415164e-05, "loss": 1.3503, "step": 20124 }, { "epoch": 14.534296028880867, "grad_norm": 2.533254623413086, "learning_rate": 2.7328519855595667e-05, "loss": 1.5201, "step": 20130 }, { "epoch": 14.538628158844766, "grad_norm": 2.5212409496307373, "learning_rate": 2.7306859205776174e-05, "loss": 1.4809, "step": 20136 }, { "epoch": 14.542960288808665, "grad_norm": 2.559194564819336, "learning_rate": 2.728519855595668e-05, "loss": 1.4807, "step": 20142 }, { "epoch": 14.547292418772564, "grad_norm": 2.322899580001831, "learning_rate": 2.7263537906137187e-05, "loss": 1.471, "step": 20148 }, { "epoch": 14.551624548736463, "grad_norm": 2.533825635910034, "learning_rate": 2.7241877256317687e-05, "loss": 1.3781, "step": 20154 }, { "epoch": 14.555956678700362, "grad_norm": 2.2780542373657227, "learning_rate": 2.7220216606498194e-05, "loss": 1.5167, "step": 20160 }, { "epoch": 14.56028880866426, "grad_norm": 2.676945447921753, "learning_rate": 2.71985559566787e-05, "loss": 1.5025, "step": 20166 }, { "epoch": 14.56462093862816, "grad_norm": 2.4278581142425537, "learning_rate": 2.7176895306859207e-05, "loss": 1.359, "step": 20172 }, { "epoch": 14.568953068592057, "grad_norm": 2.097811222076416, "learning_rate": 2.7155234657039713e-05, "loss": 1.4855, "step": 20178 }, { "epoch": 14.573285198555956, "grad_norm": 2.542976140975952, "learning_rate": 2.713357400722022e-05, "loss": 1.3904, "step": 20184 }, { "epoch": 14.577617328519855, "grad_norm": 2.2826428413391113, "learning_rate": 2.7111913357400727e-05, "loss": 1.5297, "step": 20190 }, { "epoch": 14.581949458483754, "grad_norm": 2.3883354663848877, "learning_rate": 2.7090252707581227e-05, "loss": 1.4719, "step": 20196 }, { "epoch": 14.586281588447653, "grad_norm": 2.5217013359069824, "learning_rate": 2.7068592057761733e-05, "loss": 1.5632, "step": 20202 }, { "epoch": 14.590613718411552, "grad_norm": 2.389286518096924, "learning_rate": 2.704693140794224e-05, "loss": 1.4736, "step": 20208 }, { "epoch": 14.59494584837545, "grad_norm": 2.3863117694854736, "learning_rate": 2.7025270758122746e-05, "loss": 1.5589, "step": 20214 }, { "epoch": 14.59927797833935, "grad_norm": 2.0176620483398438, "learning_rate": 2.700361010830325e-05, "loss": 1.358, "step": 20220 }, { "epoch": 14.603610108303249, "grad_norm": 2.337836265563965, "learning_rate": 2.6981949458483756e-05, "loss": 1.5611, "step": 20226 }, { "epoch": 14.607942238267148, "grad_norm": 2.4872279167175293, "learning_rate": 2.6960288808664263e-05, "loss": 1.542, "step": 20232 }, { "epoch": 14.612274368231047, "grad_norm": 2.2997806072235107, "learning_rate": 2.6938628158844763e-05, "loss": 1.4033, "step": 20238 }, { "epoch": 14.616606498194946, "grad_norm": 2.301957607269287, "learning_rate": 2.691696750902527e-05, "loss": 1.5195, "step": 20244 }, { "epoch": 14.620938628158845, "grad_norm": 2.4475646018981934, "learning_rate": 2.6895306859205776e-05, "loss": 1.4679, "step": 20250 }, { "epoch": 14.625270758122744, "grad_norm": 2.5275607109069824, "learning_rate": 2.6873646209386282e-05, "loss": 1.5577, "step": 20256 }, { "epoch": 14.629602888086643, "grad_norm": 2.4547243118286133, "learning_rate": 2.685198555956679e-05, "loss": 1.5388, "step": 20262 }, { "epoch": 14.633935018050542, "grad_norm": 2.3144004344940186, "learning_rate": 2.6830324909747296e-05, "loss": 1.4496, "step": 20268 }, { "epoch": 14.63826714801444, "grad_norm": 2.3593826293945312, "learning_rate": 2.6808664259927802e-05, "loss": 1.4373, "step": 20274 }, { "epoch": 14.64259927797834, "grad_norm": 2.3157219886779785, "learning_rate": 2.6787003610108302e-05, "loss": 1.4767, "step": 20280 }, { "epoch": 14.646931407942239, "grad_norm": 2.551145076751709, "learning_rate": 2.676534296028881e-05, "loss": 1.4675, "step": 20286 }, { "epoch": 14.651263537906138, "grad_norm": 2.232426404953003, "learning_rate": 2.6743682310469315e-05, "loss": 1.4189, "step": 20292 }, { "epoch": 14.655595667870037, "grad_norm": 2.672626256942749, "learning_rate": 2.6722021660649822e-05, "loss": 1.4378, "step": 20298 }, { "epoch": 14.659927797833936, "grad_norm": 2.488738775253296, "learning_rate": 2.670036101083033e-05, "loss": 1.466, "step": 20304 }, { "epoch": 14.664259927797834, "grad_norm": 2.2258689403533936, "learning_rate": 2.6678700361010832e-05, "loss": 1.485, "step": 20310 }, { "epoch": 14.668592057761733, "grad_norm": 2.2094521522521973, "learning_rate": 2.665703971119134e-05, "loss": 1.3268, "step": 20316 }, { "epoch": 14.672924187725632, "grad_norm": 2.7531397342681885, "learning_rate": 2.663537906137184e-05, "loss": 1.4569, "step": 20322 }, { "epoch": 14.677256317689531, "grad_norm": 2.424379825592041, "learning_rate": 2.6613718411552345e-05, "loss": 1.5053, "step": 20328 }, { "epoch": 14.68158844765343, "grad_norm": 2.2807297706604004, "learning_rate": 2.659205776173285e-05, "loss": 1.4843, "step": 20334 }, { "epoch": 14.685920577617328, "grad_norm": 2.4714319705963135, "learning_rate": 2.6570397111913358e-05, "loss": 1.4613, "step": 20340 }, { "epoch": 14.690252707581227, "grad_norm": 2.3918392658233643, "learning_rate": 2.6548736462093865e-05, "loss": 1.4736, "step": 20346 }, { "epoch": 14.694584837545126, "grad_norm": 2.4898056983947754, "learning_rate": 2.652707581227437e-05, "loss": 1.4044, "step": 20352 }, { "epoch": 14.698916967509025, "grad_norm": 2.619478464126587, "learning_rate": 2.6505415162454878e-05, "loss": 1.4201, "step": 20358 }, { "epoch": 14.703249097472924, "grad_norm": 1.9342107772827148, "learning_rate": 2.6483754512635378e-05, "loss": 1.4668, "step": 20364 }, { "epoch": 14.707581227436823, "grad_norm": 2.285146951675415, "learning_rate": 2.6462093862815884e-05, "loss": 1.4136, "step": 20370 }, { "epoch": 14.711913357400721, "grad_norm": 2.849867820739746, "learning_rate": 2.644043321299639e-05, "loss": 1.6016, "step": 20376 }, { "epoch": 14.71624548736462, "grad_norm": 2.420999050140381, "learning_rate": 2.6418772563176898e-05, "loss": 1.4862, "step": 20382 }, { "epoch": 14.72057761732852, "grad_norm": 2.3421525955200195, "learning_rate": 2.6397111913357404e-05, "loss": 1.5012, "step": 20388 }, { "epoch": 14.724909747292418, "grad_norm": 2.563594102859497, "learning_rate": 2.637545126353791e-05, "loss": 1.4664, "step": 20394 }, { "epoch": 14.729241877256317, "grad_norm": 2.3085849285125732, "learning_rate": 2.6353790613718414e-05, "loss": 1.4723, "step": 20400 }, { "epoch": 14.733574007220216, "grad_norm": 2.2950477600097656, "learning_rate": 2.6332129963898917e-05, "loss": 1.4473, "step": 20406 }, { "epoch": 14.737906137184115, "grad_norm": 2.337110757827759, "learning_rate": 2.6310469314079424e-05, "loss": 1.4839, "step": 20412 }, { "epoch": 14.742238267148014, "grad_norm": 2.4385499954223633, "learning_rate": 2.6288808664259927e-05, "loss": 1.439, "step": 20418 }, { "epoch": 14.746570397111913, "grad_norm": 2.2965922355651855, "learning_rate": 2.6267148014440434e-05, "loss": 1.4939, "step": 20424 }, { "epoch": 14.750902527075812, "grad_norm": 2.4060497283935547, "learning_rate": 2.624548736462094e-05, "loss": 1.5242, "step": 20430 }, { "epoch": 14.755234657039711, "grad_norm": 2.174461841583252, "learning_rate": 2.6223826714801447e-05, "loss": 1.4402, "step": 20436 }, { "epoch": 14.75956678700361, "grad_norm": 2.579164743423462, "learning_rate": 2.6202166064981954e-05, "loss": 1.4992, "step": 20442 }, { "epoch": 14.76389891696751, "grad_norm": 2.29301118850708, "learning_rate": 2.6180505415162453e-05, "loss": 1.4323, "step": 20448 }, { "epoch": 14.768231046931408, "grad_norm": 2.0468311309814453, "learning_rate": 2.615884476534296e-05, "loss": 1.4049, "step": 20454 }, { "epoch": 14.772563176895307, "grad_norm": 2.237449884414673, "learning_rate": 2.6137184115523467e-05, "loss": 1.4659, "step": 20460 }, { "epoch": 14.776895306859206, "grad_norm": 2.3687667846679688, "learning_rate": 2.6115523465703973e-05, "loss": 1.3817, "step": 20466 }, { "epoch": 14.781227436823105, "grad_norm": 2.241265296936035, "learning_rate": 2.609386281588448e-05, "loss": 1.3955, "step": 20472 }, { "epoch": 14.785559566787004, "grad_norm": 2.494204521179199, "learning_rate": 2.6072202166064987e-05, "loss": 1.572, "step": 20478 }, { "epoch": 14.789891696750903, "grad_norm": 2.7442409992218018, "learning_rate": 2.605054151624549e-05, "loss": 1.4268, "step": 20484 }, { "epoch": 14.794223826714802, "grad_norm": 2.5623788833618164, "learning_rate": 2.6028880866425993e-05, "loss": 1.4426, "step": 20490 }, { "epoch": 14.798555956678701, "grad_norm": 2.470139265060425, "learning_rate": 2.60072202166065e-05, "loss": 1.5064, "step": 20496 }, { "epoch": 14.8028880866426, "grad_norm": 2.3419671058654785, "learning_rate": 2.5985559566787003e-05, "loss": 1.4539, "step": 20502 }, { "epoch": 14.807220216606499, "grad_norm": 2.9425809383392334, "learning_rate": 2.596389891696751e-05, "loss": 1.484, "step": 20508 }, { "epoch": 14.811552346570398, "grad_norm": 2.4418118000030518, "learning_rate": 2.5942238267148016e-05, "loss": 1.5434, "step": 20514 }, { "epoch": 14.815884476534297, "grad_norm": 2.4483065605163574, "learning_rate": 2.5920577617328523e-05, "loss": 1.4346, "step": 20520 }, { "epoch": 14.820216606498194, "grad_norm": 2.3040144443511963, "learning_rate": 2.5898916967509022e-05, "loss": 1.5244, "step": 20526 }, { "epoch": 14.824548736462093, "grad_norm": 2.3156545162200928, "learning_rate": 2.587725631768953e-05, "loss": 1.2948, "step": 20532 }, { "epoch": 14.828880866425992, "grad_norm": 2.1183063983917236, "learning_rate": 2.5855595667870036e-05, "loss": 1.4496, "step": 20538 }, { "epoch": 14.833212996389891, "grad_norm": 2.3689112663269043, "learning_rate": 2.5833935018050542e-05, "loss": 1.4821, "step": 20544 }, { "epoch": 14.83754512635379, "grad_norm": 2.2768301963806152, "learning_rate": 2.581227436823105e-05, "loss": 1.4489, "step": 20550 }, { "epoch": 14.841877256317689, "grad_norm": 2.467912435531616, "learning_rate": 2.5790613718411556e-05, "loss": 1.4906, "step": 20556 }, { "epoch": 14.846209386281588, "grad_norm": 2.498408555984497, "learning_rate": 2.5768953068592062e-05, "loss": 1.5762, "step": 20562 }, { "epoch": 14.850541516245487, "grad_norm": 2.512515068054199, "learning_rate": 2.5747292418772562e-05, "loss": 1.439, "step": 20568 }, { "epoch": 14.854873646209386, "grad_norm": 2.279371738433838, "learning_rate": 2.572563176895307e-05, "loss": 1.5066, "step": 20574 }, { "epoch": 14.859205776173285, "grad_norm": 2.596982955932617, "learning_rate": 2.5703971119133575e-05, "loss": 1.6238, "step": 20580 }, { "epoch": 14.863537906137184, "grad_norm": 2.534688711166382, "learning_rate": 2.5682310469314082e-05, "loss": 1.4172, "step": 20586 }, { "epoch": 14.867870036101083, "grad_norm": 2.247267723083496, "learning_rate": 2.5660649819494585e-05, "loss": 1.5402, "step": 20592 }, { "epoch": 14.872202166064982, "grad_norm": 2.646299123764038, "learning_rate": 2.563898916967509e-05, "loss": 1.5088, "step": 20598 }, { "epoch": 14.876534296028881, "grad_norm": 2.5214223861694336, "learning_rate": 2.5617328519855598e-05, "loss": 1.5851, "step": 20604 }, { "epoch": 14.88086642599278, "grad_norm": 2.347086191177368, "learning_rate": 2.5595667870036098e-05, "loss": 1.451, "step": 20610 }, { "epoch": 14.885198555956679, "grad_norm": 2.493540048599243, "learning_rate": 2.5574007220216605e-05, "loss": 1.5297, "step": 20616 }, { "epoch": 14.889530685920578, "grad_norm": 2.2349395751953125, "learning_rate": 2.555234657039711e-05, "loss": 1.4016, "step": 20622 }, { "epoch": 14.893862815884477, "grad_norm": 2.389979600906372, "learning_rate": 2.5530685920577618e-05, "loss": 1.5121, "step": 20628 }, { "epoch": 14.898194945848376, "grad_norm": 2.401282787322998, "learning_rate": 2.5509025270758125e-05, "loss": 1.5776, "step": 20634 }, { "epoch": 14.902527075812275, "grad_norm": 2.3065052032470703, "learning_rate": 2.548736462093863e-05, "loss": 1.4826, "step": 20640 }, { "epoch": 14.906859205776174, "grad_norm": 2.2392466068267822, "learning_rate": 2.5465703971119138e-05, "loss": 1.4623, "step": 20646 }, { "epoch": 14.911191335740073, "grad_norm": 2.556823968887329, "learning_rate": 2.5444043321299638e-05, "loss": 1.4979, "step": 20652 }, { "epoch": 14.915523465703972, "grad_norm": 2.363654851913452, "learning_rate": 2.5422382671480144e-05, "loss": 1.5319, "step": 20658 }, { "epoch": 14.91985559566787, "grad_norm": 2.782283306121826, "learning_rate": 2.540072202166065e-05, "loss": 1.481, "step": 20664 }, { "epoch": 14.92418772563177, "grad_norm": 2.363201379776001, "learning_rate": 2.5379061371841157e-05, "loss": 1.4812, "step": 20670 }, { "epoch": 14.928519855595669, "grad_norm": 2.646317958831787, "learning_rate": 2.5357400722021664e-05, "loss": 1.5332, "step": 20676 }, { "epoch": 14.932851985559568, "grad_norm": 2.3695170879364014, "learning_rate": 2.5335740072202167e-05, "loss": 1.5534, "step": 20682 }, { "epoch": 14.937184115523465, "grad_norm": 2.412493944168091, "learning_rate": 2.5314079422382674e-05, "loss": 1.4689, "step": 20688 }, { "epoch": 14.941516245487364, "grad_norm": 2.6086432933807373, "learning_rate": 2.5292418772563177e-05, "loss": 1.507, "step": 20694 }, { "epoch": 14.945848375451263, "grad_norm": 2.404106378555298, "learning_rate": 2.527075812274368e-05, "loss": 1.4484, "step": 20700 }, { "epoch": 14.950180505415162, "grad_norm": 2.3902645111083984, "learning_rate": 2.5249097472924187e-05, "loss": 1.593, "step": 20706 }, { "epoch": 14.95451263537906, "grad_norm": 2.4292590618133545, "learning_rate": 2.5227436823104694e-05, "loss": 1.5868, "step": 20712 }, { "epoch": 14.95884476534296, "grad_norm": 2.3460450172424316, "learning_rate": 2.52057761732852e-05, "loss": 1.3573, "step": 20718 }, { "epoch": 14.963176895306859, "grad_norm": 2.467513084411621, "learning_rate": 2.5184115523465707e-05, "loss": 1.575, "step": 20724 }, { "epoch": 14.967509025270758, "grad_norm": 2.3241560459136963, "learning_rate": 2.5162454873646213e-05, "loss": 1.4575, "step": 20730 }, { "epoch": 14.971841155234657, "grad_norm": 2.547457456588745, "learning_rate": 2.5140794223826713e-05, "loss": 1.6269, "step": 20736 }, { "epoch": 14.976173285198556, "grad_norm": 2.3266780376434326, "learning_rate": 2.511913357400722e-05, "loss": 1.4253, "step": 20742 }, { "epoch": 14.980505415162455, "grad_norm": 2.4629247188568115, "learning_rate": 2.5097472924187727e-05, "loss": 1.4454, "step": 20748 }, { "epoch": 14.984837545126354, "grad_norm": 2.7490103244781494, "learning_rate": 2.5075812274368233e-05, "loss": 1.5553, "step": 20754 }, { "epoch": 14.989169675090253, "grad_norm": 2.341132640838623, "learning_rate": 2.505415162454874e-05, "loss": 1.4963, "step": 20760 }, { "epoch": 14.993501805054152, "grad_norm": 2.510101318359375, "learning_rate": 2.5032490974729243e-05, "loss": 1.4499, "step": 20766 }, { "epoch": 14.99783393501805, "grad_norm": 2.2933409214019775, "learning_rate": 2.501083032490975e-05, "loss": 1.4913, "step": 20772 }, { "epoch": 15.00216606498195, "grad_norm": 2.37432599067688, "learning_rate": 2.4989169675090253e-05, "loss": 1.3873, "step": 20778 }, { "epoch": 15.006498194945848, "grad_norm": 2.3491921424865723, "learning_rate": 2.496750902527076e-05, "loss": 1.5159, "step": 20784 }, { "epoch": 15.010830324909747, "grad_norm": 2.500429153442383, "learning_rate": 2.4945848375451263e-05, "loss": 1.4026, "step": 20790 }, { "epoch": 15.015162454873646, "grad_norm": 2.2848761081695557, "learning_rate": 2.492418772563177e-05, "loss": 1.4246, "step": 20796 }, { "epoch": 15.019494584837545, "grad_norm": 2.3370933532714844, "learning_rate": 2.4902527075812276e-05, "loss": 1.4555, "step": 20802 }, { "epoch": 15.023826714801444, "grad_norm": 2.6070382595062256, "learning_rate": 2.488086642599278e-05, "loss": 1.4909, "step": 20808 }, { "epoch": 15.028158844765343, "grad_norm": 2.5907180309295654, "learning_rate": 2.4859205776173286e-05, "loss": 1.4408, "step": 20814 }, { "epoch": 15.032490974729242, "grad_norm": 2.368215560913086, "learning_rate": 2.4837545126353792e-05, "loss": 1.4495, "step": 20820 }, { "epoch": 15.036823104693141, "grad_norm": 2.2720651626586914, "learning_rate": 2.48158844765343e-05, "loss": 1.3662, "step": 20826 }, { "epoch": 15.04115523465704, "grad_norm": 2.360363245010376, "learning_rate": 2.4794223826714802e-05, "loss": 1.4381, "step": 20832 }, { "epoch": 15.04548736462094, "grad_norm": 2.3068385124206543, "learning_rate": 2.477256317689531e-05, "loss": 1.4305, "step": 20838 }, { "epoch": 15.049819494584838, "grad_norm": 2.425668716430664, "learning_rate": 2.4750902527075815e-05, "loss": 1.5043, "step": 20844 }, { "epoch": 15.054151624548737, "grad_norm": 2.4928407669067383, "learning_rate": 2.472924187725632e-05, "loss": 1.4451, "step": 20850 }, { "epoch": 15.058483754512636, "grad_norm": 2.2243785858154297, "learning_rate": 2.4707581227436825e-05, "loss": 1.3674, "step": 20856 }, { "epoch": 15.062815884476535, "grad_norm": 2.3209235668182373, "learning_rate": 2.4685920577617332e-05, "loss": 1.4425, "step": 20862 }, { "epoch": 15.067148014440432, "grad_norm": 2.6605377197265625, "learning_rate": 2.4664259927797835e-05, "loss": 1.4517, "step": 20868 }, { "epoch": 15.071480144404331, "grad_norm": 2.3036885261535645, "learning_rate": 2.4642599277978338e-05, "loss": 1.4023, "step": 20874 }, { "epoch": 15.07581227436823, "grad_norm": 2.6507034301757812, "learning_rate": 2.4620938628158845e-05, "loss": 1.5339, "step": 20880 }, { "epoch": 15.08014440433213, "grad_norm": 2.814096689224243, "learning_rate": 2.459927797833935e-05, "loss": 1.5179, "step": 20886 }, { "epoch": 15.084476534296028, "grad_norm": 2.455571174621582, "learning_rate": 2.4577617328519855e-05, "loss": 1.462, "step": 20892 }, { "epoch": 15.088808664259927, "grad_norm": 2.368435859680176, "learning_rate": 2.455595667870036e-05, "loss": 1.4553, "step": 20898 }, { "epoch": 15.093140794223826, "grad_norm": 2.2705211639404297, "learning_rate": 2.4534296028880868e-05, "loss": 1.433, "step": 20904 }, { "epoch": 15.097472924187725, "grad_norm": 2.333761215209961, "learning_rate": 2.451263537906137e-05, "loss": 1.4426, "step": 20910 }, { "epoch": 15.101805054151624, "grad_norm": 2.468801498413086, "learning_rate": 2.4490974729241878e-05, "loss": 1.4403, "step": 20916 }, { "epoch": 15.106137184115523, "grad_norm": 2.485943078994751, "learning_rate": 2.4469314079422384e-05, "loss": 1.5076, "step": 20922 }, { "epoch": 15.110469314079422, "grad_norm": 2.5631964206695557, "learning_rate": 2.444765342960289e-05, "loss": 1.5214, "step": 20928 }, { "epoch": 15.114801444043321, "grad_norm": 2.054682970046997, "learning_rate": 2.4425992779783394e-05, "loss": 1.4267, "step": 20934 }, { "epoch": 15.11913357400722, "grad_norm": 1.983132004737854, "learning_rate": 2.44043321299639e-05, "loss": 1.3551, "step": 20940 }, { "epoch": 15.123465703971119, "grad_norm": 2.5164952278137207, "learning_rate": 2.4382671480144407e-05, "loss": 1.5386, "step": 20946 }, { "epoch": 15.127797833935018, "grad_norm": 2.630277156829834, "learning_rate": 2.436101083032491e-05, "loss": 1.4217, "step": 20952 }, { "epoch": 15.132129963898917, "grad_norm": 2.4000792503356934, "learning_rate": 2.4339350180505417e-05, "loss": 1.4889, "step": 20958 }, { "epoch": 15.136462093862816, "grad_norm": 2.415268659591675, "learning_rate": 2.431768953068592e-05, "loss": 1.4094, "step": 20964 }, { "epoch": 15.140794223826715, "grad_norm": 2.3433926105499268, "learning_rate": 2.4296028880866427e-05, "loss": 1.5106, "step": 20970 }, { "epoch": 15.145126353790614, "grad_norm": 2.584596872329712, "learning_rate": 2.427436823104693e-05, "loss": 1.4524, "step": 20976 }, { "epoch": 15.149458483754513, "grad_norm": 2.344235897064209, "learning_rate": 2.4252707581227437e-05, "loss": 1.4813, "step": 20982 }, { "epoch": 15.153790613718412, "grad_norm": 2.4428486824035645, "learning_rate": 2.4231046931407944e-05, "loss": 1.4409, "step": 20988 }, { "epoch": 15.158122743682311, "grad_norm": 2.419879913330078, "learning_rate": 2.4209386281588447e-05, "loss": 1.3271, "step": 20994 }, { "epoch": 15.16245487364621, "grad_norm": 2.266355276107788, "learning_rate": 2.4187725631768953e-05, "loss": 1.3241, "step": 21000 }, { "epoch": 15.166787003610109, "grad_norm": 2.460784673690796, "learning_rate": 2.416606498194946e-05, "loss": 1.5171, "step": 21006 }, { "epoch": 15.171119133574008, "grad_norm": 2.3318030834198, "learning_rate": 2.4144404332129967e-05, "loss": 1.4875, "step": 21012 }, { "epoch": 15.175451263537907, "grad_norm": 2.44161057472229, "learning_rate": 2.412274368231047e-05, "loss": 1.4198, "step": 21018 }, { "epoch": 15.179783393501806, "grad_norm": 2.6255502700805664, "learning_rate": 2.4101083032490977e-05, "loss": 1.4704, "step": 21024 }, { "epoch": 15.184115523465705, "grad_norm": 2.499457597732544, "learning_rate": 2.4079422382671483e-05, "loss": 1.4301, "step": 21030 }, { "epoch": 15.188447653429602, "grad_norm": 2.730381488800049, "learning_rate": 2.4057761732851986e-05, "loss": 1.3899, "step": 21036 }, { "epoch": 15.192779783393501, "grad_norm": 2.859116792678833, "learning_rate": 2.4036101083032493e-05, "loss": 1.5751, "step": 21042 }, { "epoch": 15.1971119133574, "grad_norm": 2.2441718578338623, "learning_rate": 2.4014440433213e-05, "loss": 1.4731, "step": 21048 }, { "epoch": 15.201444043321299, "grad_norm": 2.5899322032928467, "learning_rate": 2.3992779783393503e-05, "loss": 1.4431, "step": 21054 }, { "epoch": 15.205776173285198, "grad_norm": 2.5521740913391113, "learning_rate": 2.3971119133574006e-05, "loss": 1.5299, "step": 21060 }, { "epoch": 15.210108303249097, "grad_norm": 2.30202054977417, "learning_rate": 2.3949458483754513e-05, "loss": 1.3612, "step": 21066 }, { "epoch": 15.214440433212996, "grad_norm": 2.4547271728515625, "learning_rate": 2.392779783393502e-05, "loss": 1.5076, "step": 21072 }, { "epoch": 15.218772563176895, "grad_norm": 2.2082297801971436, "learning_rate": 2.3906137184115522e-05, "loss": 1.4233, "step": 21078 }, { "epoch": 15.223104693140794, "grad_norm": 2.442094564437866, "learning_rate": 2.388447653429603e-05, "loss": 1.4745, "step": 21084 }, { "epoch": 15.227436823104693, "grad_norm": 2.4378111362457275, "learning_rate": 2.3862815884476536e-05, "loss": 1.4906, "step": 21090 }, { "epoch": 15.231768953068592, "grad_norm": 2.5675950050354004, "learning_rate": 2.3841155234657042e-05, "loss": 1.3959, "step": 21096 }, { "epoch": 15.23610108303249, "grad_norm": 2.4588239192962646, "learning_rate": 2.3819494584837546e-05, "loss": 1.5351, "step": 21102 }, { "epoch": 15.24043321299639, "grad_norm": 2.2585244178771973, "learning_rate": 2.3797833935018052e-05, "loss": 1.3886, "step": 21108 }, { "epoch": 15.244765342960289, "grad_norm": 2.2609691619873047, "learning_rate": 2.377617328519856e-05, "loss": 1.439, "step": 21114 }, { "epoch": 15.249097472924188, "grad_norm": 2.692185163497925, "learning_rate": 2.3754512635379062e-05, "loss": 1.527, "step": 21120 }, { "epoch": 15.253429602888087, "grad_norm": 2.567211389541626, "learning_rate": 2.373285198555957e-05, "loss": 1.3904, "step": 21126 }, { "epoch": 15.257761732851986, "grad_norm": 2.301530122756958, "learning_rate": 2.3711191335740075e-05, "loss": 1.4135, "step": 21132 }, { "epoch": 15.262093862815885, "grad_norm": 2.458345413208008, "learning_rate": 2.368953068592058e-05, "loss": 1.3398, "step": 21138 }, { "epoch": 15.266425992779784, "grad_norm": 2.59879207611084, "learning_rate": 2.3667870036101085e-05, "loss": 1.488, "step": 21144 }, { "epoch": 15.270758122743683, "grad_norm": 2.5571155548095703, "learning_rate": 2.3646209386281588e-05, "loss": 1.4123, "step": 21150 }, { "epoch": 15.275090252707582, "grad_norm": 2.2510006427764893, "learning_rate": 2.3624548736462095e-05, "loss": 1.4224, "step": 21156 }, { "epoch": 15.27942238267148, "grad_norm": 2.3526768684387207, "learning_rate": 2.3602888086642598e-05, "loss": 1.3415, "step": 21162 }, { "epoch": 15.28375451263538, "grad_norm": 2.0359437465667725, "learning_rate": 2.3581227436823105e-05, "loss": 1.312, "step": 21168 }, { "epoch": 15.288086642599279, "grad_norm": 2.3320908546447754, "learning_rate": 2.355956678700361e-05, "loss": 1.3831, "step": 21174 }, { "epoch": 15.292418772563177, "grad_norm": 2.6473052501678467, "learning_rate": 2.3537906137184115e-05, "loss": 1.4984, "step": 21180 }, { "epoch": 15.296750902527076, "grad_norm": 2.1508090496063232, "learning_rate": 2.351624548736462e-05, "loss": 1.5005, "step": 21186 }, { "epoch": 15.301083032490975, "grad_norm": 2.1663620471954346, "learning_rate": 2.3494584837545128e-05, "loss": 1.39, "step": 21192 }, { "epoch": 15.305415162454874, "grad_norm": 2.6543548107147217, "learning_rate": 2.3472924187725634e-05, "loss": 1.4888, "step": 21198 }, { "epoch": 15.309747292418773, "grad_norm": 2.658210039138794, "learning_rate": 2.3451263537906138e-05, "loss": 1.322, "step": 21204 }, { "epoch": 15.314079422382672, "grad_norm": 2.612617015838623, "learning_rate": 2.3429602888086644e-05, "loss": 1.5157, "step": 21210 }, { "epoch": 15.31841155234657, "grad_norm": 2.415379524230957, "learning_rate": 2.340794223826715e-05, "loss": 1.5432, "step": 21216 }, { "epoch": 15.322743682310469, "grad_norm": 2.448824405670166, "learning_rate": 2.3386281588447654e-05, "loss": 1.4491, "step": 21222 }, { "epoch": 15.327075812274368, "grad_norm": 2.5728936195373535, "learning_rate": 2.336462093862816e-05, "loss": 1.4556, "step": 21228 }, { "epoch": 15.331407942238267, "grad_norm": 2.3520984649658203, "learning_rate": 2.3342960288808667e-05, "loss": 1.4384, "step": 21234 }, { "epoch": 15.335740072202166, "grad_norm": 2.646409034729004, "learning_rate": 2.332129963898917e-05, "loss": 1.4499, "step": 21240 }, { "epoch": 15.340072202166064, "grad_norm": 2.4709365367889404, "learning_rate": 2.3299638989169674e-05, "loss": 1.3136, "step": 21246 }, { "epoch": 15.344404332129963, "grad_norm": 2.3474185466766357, "learning_rate": 2.327797833935018e-05, "loss": 1.4668, "step": 21252 }, { "epoch": 15.348736462093862, "grad_norm": 2.4774599075317383, "learning_rate": 2.3256317689530687e-05, "loss": 1.5087, "step": 21258 }, { "epoch": 15.353068592057761, "grad_norm": 2.3662021160125732, "learning_rate": 2.323465703971119e-05, "loss": 1.3942, "step": 21264 }, { "epoch": 15.35740072202166, "grad_norm": 2.5622165203094482, "learning_rate": 2.3212996389891697e-05, "loss": 1.4441, "step": 21270 }, { "epoch": 15.36173285198556, "grad_norm": 2.315605640411377, "learning_rate": 2.3191335740072203e-05, "loss": 1.4456, "step": 21276 }, { "epoch": 15.366064981949458, "grad_norm": 2.420135021209717, "learning_rate": 2.316967509025271e-05, "loss": 1.4425, "step": 21282 }, { "epoch": 15.370397111913357, "grad_norm": 2.339715003967285, "learning_rate": 2.3148014440433213e-05, "loss": 1.4387, "step": 21288 }, { "epoch": 15.374729241877256, "grad_norm": 2.6269314289093018, "learning_rate": 2.312635379061372e-05, "loss": 1.3128, "step": 21294 }, { "epoch": 15.379061371841155, "grad_norm": 2.3927323818206787, "learning_rate": 2.3104693140794227e-05, "loss": 1.4589, "step": 21300 }, { "epoch": 15.383393501805054, "grad_norm": 2.3103482723236084, "learning_rate": 2.308303249097473e-05, "loss": 1.3115, "step": 21306 }, { "epoch": 15.387725631768953, "grad_norm": 2.40619158744812, "learning_rate": 2.3061371841155236e-05, "loss": 1.5186, "step": 21312 }, { "epoch": 15.392057761732852, "grad_norm": 2.7397515773773193, "learning_rate": 2.3039711191335743e-05, "loss": 1.5115, "step": 21318 }, { "epoch": 15.396389891696751, "grad_norm": 2.3871679306030273, "learning_rate": 2.3018050541516246e-05, "loss": 1.4373, "step": 21324 }, { "epoch": 15.40072202166065, "grad_norm": 2.6682546138763428, "learning_rate": 2.2996389891696753e-05, "loss": 1.5462, "step": 21330 }, { "epoch": 15.40505415162455, "grad_norm": 2.5615651607513428, "learning_rate": 2.2974729241877256e-05, "loss": 1.4841, "step": 21336 }, { "epoch": 15.409386281588448, "grad_norm": 2.2586231231689453, "learning_rate": 2.2953068592057763e-05, "loss": 1.4194, "step": 21342 }, { "epoch": 15.413718411552347, "grad_norm": 2.2229907512664795, "learning_rate": 2.2931407942238266e-05, "loss": 1.5119, "step": 21348 }, { "epoch": 15.418050541516246, "grad_norm": 2.531101942062378, "learning_rate": 2.2909747292418772e-05, "loss": 1.5244, "step": 21354 }, { "epoch": 15.422382671480145, "grad_norm": 2.3473098278045654, "learning_rate": 2.288808664259928e-05, "loss": 1.512, "step": 21360 }, { "epoch": 15.426714801444044, "grad_norm": 2.618690252304077, "learning_rate": 2.2866425992779786e-05, "loss": 1.3553, "step": 21366 }, { "epoch": 15.431046931407943, "grad_norm": 2.4802141189575195, "learning_rate": 2.284476534296029e-05, "loss": 1.5116, "step": 21372 }, { "epoch": 15.435379061371842, "grad_norm": 2.2707085609436035, "learning_rate": 2.2823104693140796e-05, "loss": 1.4097, "step": 21378 }, { "epoch": 15.43971119133574, "grad_norm": 2.226470708847046, "learning_rate": 2.2801444043321302e-05, "loss": 1.3788, "step": 21384 }, { "epoch": 15.444043321299638, "grad_norm": 2.6945278644561768, "learning_rate": 2.2779783393501805e-05, "loss": 1.4436, "step": 21390 }, { "epoch": 15.448375451263537, "grad_norm": 2.6996874809265137, "learning_rate": 2.2758122743682312e-05, "loss": 1.4361, "step": 21396 }, { "epoch": 15.452707581227436, "grad_norm": 2.496864080429077, "learning_rate": 2.273646209386282e-05, "loss": 1.5648, "step": 21402 }, { "epoch": 15.457039711191335, "grad_norm": 2.433349847793579, "learning_rate": 2.2714801444043325e-05, "loss": 1.4952, "step": 21408 }, { "epoch": 15.461371841155234, "grad_norm": 2.351283073425293, "learning_rate": 2.269314079422383e-05, "loss": 1.4368, "step": 21414 }, { "epoch": 15.465703971119133, "grad_norm": 2.3528287410736084, "learning_rate": 2.2671480144404335e-05, "loss": 1.3953, "step": 21420 }, { "epoch": 15.470036101083032, "grad_norm": 2.561591148376465, "learning_rate": 2.2649819494584838e-05, "loss": 1.4949, "step": 21426 }, { "epoch": 15.474368231046931, "grad_norm": 2.5218875408172607, "learning_rate": 2.262815884476534e-05, "loss": 1.4443, "step": 21432 }, { "epoch": 15.47870036101083, "grad_norm": 2.2431640625, "learning_rate": 2.2606498194945848e-05, "loss": 1.4795, "step": 21438 }, { "epoch": 15.483032490974729, "grad_norm": 2.1522903442382812, "learning_rate": 2.2584837545126355e-05, "loss": 1.4284, "step": 21444 }, { "epoch": 15.487364620938628, "grad_norm": 2.7257139682769775, "learning_rate": 2.2563176895306858e-05, "loss": 1.4461, "step": 21450 }, { "epoch": 15.491696750902527, "grad_norm": 2.4164514541625977, "learning_rate": 2.2541516245487365e-05, "loss": 1.5115, "step": 21456 }, { "epoch": 15.496028880866426, "grad_norm": 2.453547239303589, "learning_rate": 2.251985559566787e-05, "loss": 1.5043, "step": 21462 }, { "epoch": 15.500361010830325, "grad_norm": 2.3847827911376953, "learning_rate": 2.2498194945848378e-05, "loss": 1.4827, "step": 21468 }, { "epoch": 15.504693140794224, "grad_norm": 2.3090479373931885, "learning_rate": 2.247653429602888e-05, "loss": 1.3992, "step": 21474 }, { "epoch": 15.509025270758123, "grad_norm": 2.3858256340026855, "learning_rate": 2.2454873646209388e-05, "loss": 1.4113, "step": 21480 }, { "epoch": 15.513357400722022, "grad_norm": 2.2448978424072266, "learning_rate": 2.2433212996389894e-05, "loss": 1.3862, "step": 21486 }, { "epoch": 15.51768953068592, "grad_norm": 2.5845837593078613, "learning_rate": 2.2411552346570397e-05, "loss": 1.547, "step": 21492 }, { "epoch": 15.52202166064982, "grad_norm": 2.5413589477539062, "learning_rate": 2.2389891696750904e-05, "loss": 1.5654, "step": 21498 }, { "epoch": 15.526353790613719, "grad_norm": 2.2964553833007812, "learning_rate": 2.236823104693141e-05, "loss": 1.4888, "step": 21504 }, { "epoch": 15.530685920577618, "grad_norm": 2.2725563049316406, "learning_rate": 2.2346570397111914e-05, "loss": 1.4638, "step": 21510 }, { "epoch": 15.535018050541517, "grad_norm": 2.4338760375976562, "learning_rate": 2.232490974729242e-05, "loss": 1.4597, "step": 21516 }, { "epoch": 15.539350180505416, "grad_norm": 2.6649816036224365, "learning_rate": 2.2303249097472924e-05, "loss": 1.5012, "step": 21522 }, { "epoch": 15.543682310469315, "grad_norm": 2.7758853435516357, "learning_rate": 2.228158844765343e-05, "loss": 1.6079, "step": 21528 }, { "epoch": 15.548014440433214, "grad_norm": 2.4110560417175293, "learning_rate": 2.2259927797833934e-05, "loss": 1.4206, "step": 21534 }, { "epoch": 15.552346570397113, "grad_norm": 2.5839104652404785, "learning_rate": 2.223826714801444e-05, "loss": 1.476, "step": 21540 }, { "epoch": 15.556678700361012, "grad_norm": 2.33750581741333, "learning_rate": 2.2216606498194947e-05, "loss": 1.4922, "step": 21546 }, { "epoch": 15.56101083032491, "grad_norm": 2.2084803581237793, "learning_rate": 2.2194945848375453e-05, "loss": 1.371, "step": 21552 }, { "epoch": 15.56534296028881, "grad_norm": 2.3926186561584473, "learning_rate": 2.2173285198555957e-05, "loss": 1.3989, "step": 21558 }, { "epoch": 15.569675090252707, "grad_norm": 2.6373724937438965, "learning_rate": 2.2151624548736463e-05, "loss": 1.5258, "step": 21564 }, { "epoch": 15.574007220216606, "grad_norm": 2.4520955085754395, "learning_rate": 2.212996389891697e-05, "loss": 1.3761, "step": 21570 }, { "epoch": 15.578339350180505, "grad_norm": 2.295356273651123, "learning_rate": 2.2108303249097473e-05, "loss": 1.4323, "step": 21576 }, { "epoch": 15.582671480144404, "grad_norm": 2.8464529514312744, "learning_rate": 2.208664259927798e-05, "loss": 1.497, "step": 21582 }, { "epoch": 15.587003610108303, "grad_norm": 2.28285551071167, "learning_rate": 2.2064981949458486e-05, "loss": 1.4561, "step": 21588 }, { "epoch": 15.591335740072202, "grad_norm": 2.144359827041626, "learning_rate": 2.2043321299638993e-05, "loss": 1.3979, "step": 21594 }, { "epoch": 15.5956678700361, "grad_norm": 2.4530045986175537, "learning_rate": 2.2021660649819496e-05, "loss": 1.5477, "step": 21600 }, { "epoch": 15.6, "grad_norm": 2.3871161937713623, "learning_rate": 2.2000000000000003e-05, "loss": 1.4102, "step": 21606 }, { "epoch": 15.604332129963899, "grad_norm": 2.4007186889648438, "learning_rate": 2.1978339350180506e-05, "loss": 1.2992, "step": 21612 }, { "epoch": 15.608664259927798, "grad_norm": 2.513746976852417, "learning_rate": 2.195667870036101e-05, "loss": 1.467, "step": 21618 }, { "epoch": 15.612996389891697, "grad_norm": 2.335702896118164, "learning_rate": 2.1935018050541516e-05, "loss": 1.4541, "step": 21624 }, { "epoch": 15.617328519855596, "grad_norm": 2.2593014240264893, "learning_rate": 2.1913357400722022e-05, "loss": 1.3578, "step": 21630 }, { "epoch": 15.621660649819495, "grad_norm": 2.6028261184692383, "learning_rate": 2.189169675090253e-05, "loss": 1.479, "step": 21636 }, { "epoch": 15.625992779783394, "grad_norm": 2.66113018989563, "learning_rate": 2.1870036101083032e-05, "loss": 1.4522, "step": 21642 }, { "epoch": 15.630324909747292, "grad_norm": 2.291564464569092, "learning_rate": 2.184837545126354e-05, "loss": 1.4161, "step": 21648 }, { "epoch": 15.634657039711191, "grad_norm": 2.615675926208496, "learning_rate": 2.1826714801444046e-05, "loss": 1.5819, "step": 21654 }, { "epoch": 15.63898916967509, "grad_norm": 2.6732938289642334, "learning_rate": 2.180505415162455e-05, "loss": 1.4639, "step": 21660 }, { "epoch": 15.64332129963899, "grad_norm": 2.4117679595947266, "learning_rate": 2.1783393501805055e-05, "loss": 1.6153, "step": 21666 }, { "epoch": 15.647653429602888, "grad_norm": 2.433469772338867, "learning_rate": 2.1761732851985562e-05, "loss": 1.3133, "step": 21672 }, { "epoch": 15.651985559566787, "grad_norm": 2.2948780059814453, "learning_rate": 2.1740072202166065e-05, "loss": 1.3789, "step": 21678 }, { "epoch": 15.656317689530686, "grad_norm": 2.7083287239074707, "learning_rate": 2.1718411552346572e-05, "loss": 1.6107, "step": 21684 }, { "epoch": 15.660649819494585, "grad_norm": 2.3814096450805664, "learning_rate": 2.169675090252708e-05, "loss": 1.3658, "step": 21690 }, { "epoch": 15.664981949458484, "grad_norm": 2.7786054611206055, "learning_rate": 2.167509025270758e-05, "loss": 1.4804, "step": 21696 }, { "epoch": 15.669314079422383, "grad_norm": 2.4612066745758057, "learning_rate": 2.1653429602888088e-05, "loss": 1.4513, "step": 21702 }, { "epoch": 15.673646209386282, "grad_norm": 2.239924669265747, "learning_rate": 2.163176895306859e-05, "loss": 1.3911, "step": 21708 }, { "epoch": 15.677978339350181, "grad_norm": 2.2340776920318604, "learning_rate": 2.1610108303249098e-05, "loss": 1.4825, "step": 21714 }, { "epoch": 15.68231046931408, "grad_norm": 2.319196939468384, "learning_rate": 2.15884476534296e-05, "loss": 1.3343, "step": 21720 }, { "epoch": 15.686642599277977, "grad_norm": 2.428248167037964, "learning_rate": 2.1566787003610108e-05, "loss": 1.4845, "step": 21726 }, { "epoch": 15.690974729241876, "grad_norm": 2.543729543685913, "learning_rate": 2.1545126353790615e-05, "loss": 1.4549, "step": 21732 }, { "epoch": 15.695306859205775, "grad_norm": 2.7283949851989746, "learning_rate": 2.152346570397112e-05, "loss": 1.4189, "step": 21738 }, { "epoch": 15.699638989169674, "grad_norm": 2.4666426181793213, "learning_rate": 2.1501805054151624e-05, "loss": 1.4752, "step": 21744 }, { "epoch": 15.703971119133573, "grad_norm": 2.301211357116699, "learning_rate": 2.148014440433213e-05, "loss": 1.3898, "step": 21750 }, { "epoch": 15.708303249097472, "grad_norm": 2.4630372524261475, "learning_rate": 2.1458483754512638e-05, "loss": 1.4622, "step": 21756 }, { "epoch": 15.712635379061371, "grad_norm": 2.263450860977173, "learning_rate": 2.143682310469314e-05, "loss": 1.4402, "step": 21762 }, { "epoch": 15.71696750902527, "grad_norm": 2.6140501499176025, "learning_rate": 2.1415162454873647e-05, "loss": 1.4911, "step": 21768 }, { "epoch": 15.72129963898917, "grad_norm": 2.336671829223633, "learning_rate": 2.1393501805054154e-05, "loss": 1.5128, "step": 21774 }, { "epoch": 15.725631768953068, "grad_norm": 2.3164799213409424, "learning_rate": 2.137184115523466e-05, "loss": 1.3798, "step": 21780 }, { "epoch": 15.729963898916967, "grad_norm": 2.3275599479675293, "learning_rate": 2.1350180505415164e-05, "loss": 1.5264, "step": 21786 }, { "epoch": 15.734296028880866, "grad_norm": 2.4169158935546875, "learning_rate": 2.132851985559567e-05, "loss": 1.3851, "step": 21792 }, { "epoch": 15.738628158844765, "grad_norm": 2.452897071838379, "learning_rate": 2.1306859205776174e-05, "loss": 1.4851, "step": 21798 }, { "epoch": 15.742960288808664, "grad_norm": 2.7476847171783447, "learning_rate": 2.1285198555956677e-05, "loss": 1.4458, "step": 21804 }, { "epoch": 15.747292418772563, "grad_norm": 2.5604796409606934, "learning_rate": 2.1263537906137184e-05, "loss": 1.4455, "step": 21810 }, { "epoch": 15.751624548736462, "grad_norm": 2.3711626529693604, "learning_rate": 2.124187725631769e-05, "loss": 1.4325, "step": 21816 }, { "epoch": 15.755956678700361, "grad_norm": 2.4719088077545166, "learning_rate": 2.1220216606498197e-05, "loss": 1.5107, "step": 21822 }, { "epoch": 15.76028880866426, "grad_norm": 2.4473888874053955, "learning_rate": 2.11985559566787e-05, "loss": 1.5942, "step": 21828 }, { "epoch": 15.764620938628159, "grad_norm": 2.4301464557647705, "learning_rate": 2.1176895306859207e-05, "loss": 1.5153, "step": 21834 }, { "epoch": 15.768953068592058, "grad_norm": 2.644648790359497, "learning_rate": 2.1155234657039713e-05, "loss": 1.5521, "step": 21840 }, { "epoch": 15.773285198555957, "grad_norm": 2.479128122329712, "learning_rate": 2.1133574007220217e-05, "loss": 1.5805, "step": 21846 }, { "epoch": 15.777617328519856, "grad_norm": 2.536679983139038, "learning_rate": 2.1111913357400723e-05, "loss": 1.4795, "step": 21852 }, { "epoch": 15.781949458483755, "grad_norm": 2.500384569168091, "learning_rate": 2.109025270758123e-05, "loss": 1.4829, "step": 21858 }, { "epoch": 15.786281588447654, "grad_norm": 2.254927396774292, "learning_rate": 2.1068592057761736e-05, "loss": 1.4967, "step": 21864 }, { "epoch": 15.790613718411553, "grad_norm": 2.753495931625366, "learning_rate": 2.104693140794224e-05, "loss": 1.5395, "step": 21870 }, { "epoch": 15.794945848375452, "grad_norm": 2.2743849754333496, "learning_rate": 2.1025270758122746e-05, "loss": 1.4267, "step": 21876 }, { "epoch": 15.79927797833935, "grad_norm": 2.330022096633911, "learning_rate": 2.100361010830325e-05, "loss": 1.364, "step": 21882 }, { "epoch": 15.80361010830325, "grad_norm": 2.65877103805542, "learning_rate": 2.0981949458483756e-05, "loss": 1.5566, "step": 21888 }, { "epoch": 15.807942238267149, "grad_norm": 2.8260722160339355, "learning_rate": 2.096028880866426e-05, "loss": 1.4891, "step": 21894 }, { "epoch": 15.812274368231048, "grad_norm": 2.5242667198181152, "learning_rate": 2.0938628158844766e-05, "loss": 1.5248, "step": 21900 }, { "epoch": 15.816606498194945, "grad_norm": 2.3497438430786133, "learning_rate": 2.0916967509025272e-05, "loss": 1.5735, "step": 21906 }, { "epoch": 15.820938628158844, "grad_norm": 2.334878444671631, "learning_rate": 2.0895306859205776e-05, "loss": 1.3594, "step": 21912 }, { "epoch": 15.825270758122743, "grad_norm": 2.5985066890716553, "learning_rate": 2.0873646209386282e-05, "loss": 1.5321, "step": 21918 }, { "epoch": 15.829602888086642, "grad_norm": 2.467478036880493, "learning_rate": 2.085198555956679e-05, "loss": 1.4956, "step": 21924 }, { "epoch": 15.833935018050541, "grad_norm": 2.2911810874938965, "learning_rate": 2.0830324909747292e-05, "loss": 1.51, "step": 21930 }, { "epoch": 15.83826714801444, "grad_norm": 2.675346851348877, "learning_rate": 2.08086642599278e-05, "loss": 1.5511, "step": 21936 }, { "epoch": 15.842599277978339, "grad_norm": 2.0758166313171387, "learning_rate": 2.0787003610108305e-05, "loss": 1.3913, "step": 21942 }, { "epoch": 15.846931407942238, "grad_norm": 2.6970834732055664, "learning_rate": 2.076534296028881e-05, "loss": 1.4531, "step": 21948 }, { "epoch": 15.851263537906137, "grad_norm": 2.5336906909942627, "learning_rate": 2.0743682310469315e-05, "loss": 1.4148, "step": 21954 }, { "epoch": 15.855595667870036, "grad_norm": 2.4006106853485107, "learning_rate": 2.0722021660649822e-05, "loss": 1.3931, "step": 21960 }, { "epoch": 15.859927797833935, "grad_norm": 2.2997703552246094, "learning_rate": 2.070036101083033e-05, "loss": 1.3942, "step": 21966 }, { "epoch": 15.864259927797834, "grad_norm": 2.7695164680480957, "learning_rate": 2.067870036101083e-05, "loss": 1.4299, "step": 21972 }, { "epoch": 15.868592057761733, "grad_norm": 2.559014081954956, "learning_rate": 2.0657039711191338e-05, "loss": 1.551, "step": 21978 }, { "epoch": 15.872924187725632, "grad_norm": 2.625687837600708, "learning_rate": 2.063537906137184e-05, "loss": 1.4398, "step": 21984 }, { "epoch": 15.87725631768953, "grad_norm": 2.4514822959899902, "learning_rate": 2.0613718411552345e-05, "loss": 1.4958, "step": 21990 }, { "epoch": 15.88158844765343, "grad_norm": 2.7448301315307617, "learning_rate": 2.059205776173285e-05, "loss": 1.4542, "step": 21996 }, { "epoch": 15.885920577617329, "grad_norm": 2.5343475341796875, "learning_rate": 2.0570397111913358e-05, "loss": 1.5133, "step": 22002 }, { "epoch": 15.890252707581228, "grad_norm": 2.1386804580688477, "learning_rate": 2.0548736462093865e-05, "loss": 1.3786, "step": 22008 }, { "epoch": 15.894584837545127, "grad_norm": 2.69053053855896, "learning_rate": 2.0527075812274368e-05, "loss": 1.5843, "step": 22014 }, { "epoch": 15.898916967509026, "grad_norm": 2.4576144218444824, "learning_rate": 2.0505415162454874e-05, "loss": 1.5024, "step": 22020 }, { "epoch": 15.903249097472925, "grad_norm": 2.520270586013794, "learning_rate": 2.048375451263538e-05, "loss": 1.527, "step": 22026 }, { "epoch": 15.907581227436824, "grad_norm": 2.5921177864074707, "learning_rate": 2.0462093862815884e-05, "loss": 1.3683, "step": 22032 }, { "epoch": 15.911913357400723, "grad_norm": 2.444836139678955, "learning_rate": 2.044043321299639e-05, "loss": 1.5243, "step": 22038 }, { "epoch": 15.916245487364622, "grad_norm": 2.535447120666504, "learning_rate": 2.0418772563176897e-05, "loss": 1.4506, "step": 22044 }, { "epoch": 15.92057761732852, "grad_norm": 2.4244251251220703, "learning_rate": 2.0397111913357404e-05, "loss": 1.524, "step": 22050 }, { "epoch": 15.92490974729242, "grad_norm": 2.316645860671997, "learning_rate": 2.0375451263537907e-05, "loss": 1.369, "step": 22056 }, { "epoch": 15.929241877256318, "grad_norm": 2.285092353820801, "learning_rate": 2.0353790613718414e-05, "loss": 1.563, "step": 22062 }, { "epoch": 15.933574007220217, "grad_norm": 2.466958522796631, "learning_rate": 2.0332129963898917e-05, "loss": 1.4173, "step": 22068 }, { "epoch": 15.937906137184115, "grad_norm": 2.649493932723999, "learning_rate": 2.0310469314079424e-05, "loss": 1.3996, "step": 22074 }, { "epoch": 15.942238267148014, "grad_norm": 2.6553938388824463, "learning_rate": 2.0288808664259927e-05, "loss": 1.4582, "step": 22080 }, { "epoch": 15.946570397111913, "grad_norm": 2.586082935333252, "learning_rate": 2.0267148014440434e-05, "loss": 1.5682, "step": 22086 }, { "epoch": 15.950902527075812, "grad_norm": 2.491062879562378, "learning_rate": 2.024548736462094e-05, "loss": 1.5182, "step": 22092 }, { "epoch": 15.95523465703971, "grad_norm": 2.308349370956421, "learning_rate": 2.0223826714801443e-05, "loss": 1.5673, "step": 22098 }, { "epoch": 15.95956678700361, "grad_norm": 2.8096041679382324, "learning_rate": 2.020216606498195e-05, "loss": 1.6407, "step": 22104 }, { "epoch": 15.963898916967509, "grad_norm": 2.650573968887329, "learning_rate": 2.0180505415162457e-05, "loss": 1.416, "step": 22110 }, { "epoch": 15.968231046931407, "grad_norm": 2.5497844219207764, "learning_rate": 2.015884476534296e-05, "loss": 1.4646, "step": 22116 }, { "epoch": 15.972563176895306, "grad_norm": 2.260164499282837, "learning_rate": 2.0137184115523467e-05, "loss": 1.496, "step": 22122 }, { "epoch": 15.976895306859205, "grad_norm": 2.3654677867889404, "learning_rate": 2.0115523465703973e-05, "loss": 1.3505, "step": 22128 }, { "epoch": 15.981227436823104, "grad_norm": 2.387450933456421, "learning_rate": 2.009386281588448e-05, "loss": 1.4552, "step": 22134 }, { "epoch": 15.985559566787003, "grad_norm": 2.2871177196502686, "learning_rate": 2.0072202166064983e-05, "loss": 1.4671, "step": 22140 }, { "epoch": 15.989891696750902, "grad_norm": 2.3247880935668945, "learning_rate": 2.005054151624549e-05, "loss": 1.412, "step": 22146 }, { "epoch": 15.994223826714801, "grad_norm": 2.5120139122009277, "learning_rate": 2.0028880866425996e-05, "loss": 1.4628, "step": 22152 }, { "epoch": 15.9985559566787, "grad_norm": 2.298092842102051, "learning_rate": 2.00072202166065e-05, "loss": 1.3874, "step": 22158 }, { "epoch": 16.0028880866426, "grad_norm": 2.1879162788391113, "learning_rate": 1.9985559566787006e-05, "loss": 1.4374, "step": 22164 }, { "epoch": 16.0072202166065, "grad_norm": 2.108832359313965, "learning_rate": 1.996389891696751e-05, "loss": 1.3607, "step": 22170 }, { "epoch": 16.011552346570397, "grad_norm": 2.190314292907715, "learning_rate": 1.9942238267148012e-05, "loss": 1.4699, "step": 22176 }, { "epoch": 16.015884476534296, "grad_norm": 2.4915482997894287, "learning_rate": 1.992057761732852e-05, "loss": 1.3904, "step": 22182 }, { "epoch": 16.020216606498195, "grad_norm": 2.3572866916656494, "learning_rate": 1.9898916967509026e-05, "loss": 1.4713, "step": 22188 }, { "epoch": 16.024548736462094, "grad_norm": 2.4496500492095947, "learning_rate": 1.9877256317689532e-05, "loss": 1.4793, "step": 22194 }, { "epoch": 16.028880866425993, "grad_norm": 2.29494571685791, "learning_rate": 1.9855595667870036e-05, "loss": 1.4838, "step": 22200 }, { "epoch": 16.033212996389892, "grad_norm": 2.24456787109375, "learning_rate": 1.9833935018050542e-05, "loss": 1.4587, "step": 22206 }, { "epoch": 16.03754512635379, "grad_norm": 2.4719274044036865, "learning_rate": 1.981227436823105e-05, "loss": 1.4623, "step": 22212 }, { "epoch": 16.04187725631769, "grad_norm": 2.436277151107788, "learning_rate": 1.9790613718411552e-05, "loss": 1.4279, "step": 22218 }, { "epoch": 16.04620938628159, "grad_norm": 2.4601383209228516, "learning_rate": 1.976895306859206e-05, "loss": 1.5087, "step": 22224 }, { "epoch": 16.050541516245488, "grad_norm": 2.4934940338134766, "learning_rate": 1.9747292418772565e-05, "loss": 1.5125, "step": 22230 }, { "epoch": 16.054873646209387, "grad_norm": 2.5237977504730225, "learning_rate": 1.9725631768953072e-05, "loss": 1.3762, "step": 22236 }, { "epoch": 16.059205776173286, "grad_norm": 2.6407382488250732, "learning_rate": 1.9703971119133575e-05, "loss": 1.3143, "step": 22242 }, { "epoch": 16.063537906137185, "grad_norm": 2.690683364868164, "learning_rate": 1.968231046931408e-05, "loss": 1.2936, "step": 22248 }, { "epoch": 16.067870036101084, "grad_norm": 2.5491368770599365, "learning_rate": 1.9660649819494585e-05, "loss": 1.4461, "step": 22254 }, { "epoch": 16.072202166064983, "grad_norm": 2.4106876850128174, "learning_rate": 1.963898916967509e-05, "loss": 1.4392, "step": 22260 }, { "epoch": 16.076534296028882, "grad_norm": 2.5596354007720947, "learning_rate": 1.9617328519855595e-05, "loss": 1.3342, "step": 22266 }, { "epoch": 16.08086642599278, "grad_norm": 2.3254222869873047, "learning_rate": 1.95956678700361e-05, "loss": 1.4028, "step": 22272 }, { "epoch": 16.08519855595668, "grad_norm": 2.4483883380889893, "learning_rate": 1.9574007220216608e-05, "loss": 1.4215, "step": 22278 }, { "epoch": 16.08953068592058, "grad_norm": 2.421447515487671, "learning_rate": 1.955234657039711e-05, "loss": 1.4359, "step": 22284 }, { "epoch": 16.093862815884478, "grad_norm": 2.5111734867095947, "learning_rate": 1.9530685920577618e-05, "loss": 1.4755, "step": 22290 }, { "epoch": 16.098194945848377, "grad_norm": 2.352947950363159, "learning_rate": 1.9509025270758124e-05, "loss": 1.3945, "step": 22296 }, { "epoch": 16.102527075812276, "grad_norm": 2.697084665298462, "learning_rate": 1.9487364620938628e-05, "loss": 1.518, "step": 22302 }, { "epoch": 16.106859205776175, "grad_norm": 2.3338406085968018, "learning_rate": 1.9465703971119134e-05, "loss": 1.3764, "step": 22308 }, { "epoch": 16.111191335740074, "grad_norm": 2.694875955581665, "learning_rate": 1.944404332129964e-05, "loss": 1.4688, "step": 22314 }, { "epoch": 16.115523465703973, "grad_norm": 2.3711559772491455, "learning_rate": 1.9422382671480147e-05, "loss": 1.4372, "step": 22320 }, { "epoch": 16.11985559566787, "grad_norm": 2.24665904045105, "learning_rate": 1.940072202166065e-05, "loss": 1.3979, "step": 22326 }, { "epoch": 16.12418772563177, "grad_norm": 2.3258914947509766, "learning_rate": 1.9379061371841157e-05, "loss": 1.4258, "step": 22332 }, { "epoch": 16.128519855595666, "grad_norm": 2.5800297260284424, "learning_rate": 1.9357400722021664e-05, "loss": 1.531, "step": 22338 }, { "epoch": 16.132851985559565, "grad_norm": 2.44989013671875, "learning_rate": 1.9335740072202167e-05, "loss": 1.3599, "step": 22344 }, { "epoch": 16.137184115523464, "grad_norm": 2.644665241241455, "learning_rate": 1.9314079422382674e-05, "loss": 1.4847, "step": 22350 }, { "epoch": 16.141516245487363, "grad_norm": 2.640627861022949, "learning_rate": 1.9292418772563177e-05, "loss": 1.4214, "step": 22356 }, { "epoch": 16.145848375451262, "grad_norm": 2.7194039821624756, "learning_rate": 1.9270758122743684e-05, "loss": 1.4915, "step": 22362 }, { "epoch": 16.15018050541516, "grad_norm": 2.7108006477355957, "learning_rate": 1.9249097472924187e-05, "loss": 1.513, "step": 22368 }, { "epoch": 16.15451263537906, "grad_norm": 2.465214252471924, "learning_rate": 1.9227436823104693e-05, "loss": 1.369, "step": 22374 }, { "epoch": 16.15884476534296, "grad_norm": 2.4447007179260254, "learning_rate": 1.92057761732852e-05, "loss": 1.4184, "step": 22380 }, { "epoch": 16.163176895306858, "grad_norm": 2.7173492908477783, "learning_rate": 1.9184115523465703e-05, "loss": 1.431, "step": 22386 }, { "epoch": 16.167509025270757, "grad_norm": 2.7023072242736816, "learning_rate": 1.916245487364621e-05, "loss": 1.4943, "step": 22392 }, { "epoch": 16.171841155234656, "grad_norm": 2.3769569396972656, "learning_rate": 1.9140794223826717e-05, "loss": 1.3081, "step": 22398 }, { "epoch": 16.176173285198555, "grad_norm": 2.488182544708252, "learning_rate": 1.9119133574007223e-05, "loss": 1.448, "step": 22404 }, { "epoch": 16.180505415162454, "grad_norm": 2.263234853744507, "learning_rate": 1.9097472924187726e-05, "loss": 1.3345, "step": 22410 }, { "epoch": 16.184837545126353, "grad_norm": 2.601886749267578, "learning_rate": 1.9075812274368233e-05, "loss": 1.5437, "step": 22416 }, { "epoch": 16.189169675090252, "grad_norm": 2.5542690753936768, "learning_rate": 1.905415162454874e-05, "loss": 1.5446, "step": 22422 }, { "epoch": 16.19350180505415, "grad_norm": 2.5623586177825928, "learning_rate": 1.9032490974729243e-05, "loss": 1.5364, "step": 22428 }, { "epoch": 16.19783393501805, "grad_norm": 2.523991107940674, "learning_rate": 1.901083032490975e-05, "loss": 1.4507, "step": 22434 }, { "epoch": 16.20216606498195, "grad_norm": 2.25490403175354, "learning_rate": 1.8989169675090253e-05, "loss": 1.543, "step": 22440 }, { "epoch": 16.206498194945848, "grad_norm": 2.283578634262085, "learning_rate": 1.896750902527076e-05, "loss": 1.2827, "step": 22446 }, { "epoch": 16.210830324909747, "grad_norm": 2.7080016136169434, "learning_rate": 1.8945848375451262e-05, "loss": 1.5396, "step": 22452 }, { "epoch": 16.215162454873646, "grad_norm": 2.4866256713867188, "learning_rate": 1.892418772563177e-05, "loss": 1.4432, "step": 22458 }, { "epoch": 16.219494584837545, "grad_norm": 2.4831676483154297, "learning_rate": 1.8902527075812276e-05, "loss": 1.4797, "step": 22464 }, { "epoch": 16.223826714801444, "grad_norm": 2.4664578437805176, "learning_rate": 1.888086642599278e-05, "loss": 1.4514, "step": 22470 }, { "epoch": 16.228158844765343, "grad_norm": 2.443009376525879, "learning_rate": 1.8859205776173286e-05, "loss": 1.4767, "step": 22476 }, { "epoch": 16.23249097472924, "grad_norm": 2.5542845726013184, "learning_rate": 1.8837545126353792e-05, "loss": 1.4428, "step": 22482 }, { "epoch": 16.23682310469314, "grad_norm": 2.2960281372070312, "learning_rate": 1.8815884476534295e-05, "loss": 1.4472, "step": 22488 }, { "epoch": 16.24115523465704, "grad_norm": 2.5170767307281494, "learning_rate": 1.8794223826714802e-05, "loss": 1.3849, "step": 22494 }, { "epoch": 16.24548736462094, "grad_norm": 2.686244010925293, "learning_rate": 1.877256317689531e-05, "loss": 1.5414, "step": 22500 }, { "epoch": 16.249819494584838, "grad_norm": 2.344813823699951, "learning_rate": 1.8750902527075815e-05, "loss": 1.4145, "step": 22506 }, { "epoch": 16.254151624548737, "grad_norm": 2.4555039405822754, "learning_rate": 1.872924187725632e-05, "loss": 1.4948, "step": 22512 }, { "epoch": 16.258483754512635, "grad_norm": 2.622847080230713, "learning_rate": 1.8707581227436825e-05, "loss": 1.4771, "step": 22518 }, { "epoch": 16.262815884476534, "grad_norm": 2.370102882385254, "learning_rate": 1.868592057761733e-05, "loss": 1.4101, "step": 22524 }, { "epoch": 16.267148014440433, "grad_norm": 2.507901430130005, "learning_rate": 1.8664259927797835e-05, "loss": 1.5514, "step": 22530 }, { "epoch": 16.271480144404332, "grad_norm": 2.3603711128234863, "learning_rate": 1.8642599277978338e-05, "loss": 1.4508, "step": 22536 }, { "epoch": 16.27581227436823, "grad_norm": 2.5072951316833496, "learning_rate": 1.8620938628158845e-05, "loss": 1.4043, "step": 22542 }, { "epoch": 16.28014440433213, "grad_norm": 2.5982418060302734, "learning_rate": 1.859927797833935e-05, "loss": 1.3524, "step": 22548 }, { "epoch": 16.28447653429603, "grad_norm": 2.555490016937256, "learning_rate": 1.8577617328519855e-05, "loss": 1.5309, "step": 22554 }, { "epoch": 16.28880866425993, "grad_norm": 2.316880226135254, "learning_rate": 1.855595667870036e-05, "loss": 1.4364, "step": 22560 }, { "epoch": 16.293140794223827, "grad_norm": 2.1638331413269043, "learning_rate": 1.8534296028880868e-05, "loss": 1.4304, "step": 22566 }, { "epoch": 16.297472924187726, "grad_norm": 2.351278781890869, "learning_rate": 1.851263537906137e-05, "loss": 1.386, "step": 22572 }, { "epoch": 16.301805054151625, "grad_norm": 2.1275277137756348, "learning_rate": 1.8490974729241878e-05, "loss": 1.3999, "step": 22578 }, { "epoch": 16.306137184115524, "grad_norm": 2.442199468612671, "learning_rate": 1.8469314079422384e-05, "loss": 1.459, "step": 22584 }, { "epoch": 16.310469314079423, "grad_norm": 2.5874338150024414, "learning_rate": 1.844765342960289e-05, "loss": 1.3007, "step": 22590 }, { "epoch": 16.314801444043322, "grad_norm": 2.440951108932495, "learning_rate": 1.8425992779783394e-05, "loss": 1.4519, "step": 22596 }, { "epoch": 16.31913357400722, "grad_norm": 2.4968299865722656, "learning_rate": 1.84043321299639e-05, "loss": 1.3441, "step": 22602 }, { "epoch": 16.32346570397112, "grad_norm": 2.722332239151001, "learning_rate": 1.8382671480144407e-05, "loss": 1.5074, "step": 22608 }, { "epoch": 16.32779783393502, "grad_norm": 2.623183488845825, "learning_rate": 1.836101083032491e-05, "loss": 1.4786, "step": 22614 }, { "epoch": 16.332129963898918, "grad_norm": 2.307605504989624, "learning_rate": 1.8339350180505417e-05, "loss": 1.3472, "step": 22620 }, { "epoch": 16.336462093862817, "grad_norm": 2.2241427898406982, "learning_rate": 1.831768953068592e-05, "loss": 1.4196, "step": 22626 }, { "epoch": 16.340794223826716, "grad_norm": 2.8500490188598633, "learning_rate": 1.8296028880866427e-05, "loss": 1.4518, "step": 22632 }, { "epoch": 16.345126353790615, "grad_norm": 2.4180245399475098, "learning_rate": 1.827436823104693e-05, "loss": 1.4154, "step": 22638 }, { "epoch": 16.349458483754514, "grad_norm": 2.7415292263031006, "learning_rate": 1.8252707581227437e-05, "loss": 1.497, "step": 22644 }, { "epoch": 16.353790613718413, "grad_norm": 2.5366780757904053, "learning_rate": 1.8231046931407943e-05, "loss": 1.5862, "step": 22650 }, { "epoch": 16.358122743682312, "grad_norm": 2.277479648590088, "learning_rate": 1.8209386281588447e-05, "loss": 1.513, "step": 22656 }, { "epoch": 16.36245487364621, "grad_norm": 2.528768301010132, "learning_rate": 1.8187725631768953e-05, "loss": 1.4533, "step": 22662 }, { "epoch": 16.36678700361011, "grad_norm": 2.231139659881592, "learning_rate": 1.816606498194946e-05, "loss": 1.3343, "step": 22668 }, { "epoch": 16.37111913357401, "grad_norm": 2.7182822227478027, "learning_rate": 1.8144404332129967e-05, "loss": 1.4356, "step": 22674 }, { "epoch": 16.375451263537904, "grad_norm": 2.2431693077087402, "learning_rate": 1.812274368231047e-05, "loss": 1.3749, "step": 22680 }, { "epoch": 16.379783393501803, "grad_norm": 2.2644782066345215, "learning_rate": 1.8101083032490976e-05, "loss": 1.3571, "step": 22686 }, { "epoch": 16.384115523465702, "grad_norm": 2.585537910461426, "learning_rate": 1.8079422382671483e-05, "loss": 1.556, "step": 22692 }, { "epoch": 16.3884476534296, "grad_norm": 2.557805299758911, "learning_rate": 1.8057761732851986e-05, "loss": 1.4567, "step": 22698 }, { "epoch": 16.3927797833935, "grad_norm": 2.662822961807251, "learning_rate": 1.8036101083032493e-05, "loss": 1.4885, "step": 22704 }, { "epoch": 16.3971119133574, "grad_norm": 2.3569087982177734, "learning_rate": 1.8014440433213e-05, "loss": 1.4939, "step": 22710 }, { "epoch": 16.4014440433213, "grad_norm": 2.165031671524048, "learning_rate": 1.7992779783393503e-05, "loss": 1.3488, "step": 22716 }, { "epoch": 16.405776173285197, "grad_norm": 2.470813512802124, "learning_rate": 1.7971119133574006e-05, "loss": 1.4104, "step": 22722 }, { "epoch": 16.410108303249096, "grad_norm": 2.3419392108917236, "learning_rate": 1.7949458483754512e-05, "loss": 1.3486, "step": 22728 }, { "epoch": 16.414440433212995, "grad_norm": 2.5044291019439697, "learning_rate": 1.792779783393502e-05, "loss": 1.42, "step": 22734 }, { "epoch": 16.418772563176894, "grad_norm": 2.4098172187805176, "learning_rate": 1.7906137184115522e-05, "loss": 1.467, "step": 22740 }, { "epoch": 16.423104693140793, "grad_norm": 2.518244504928589, "learning_rate": 1.788447653429603e-05, "loss": 1.5236, "step": 22746 }, { "epoch": 16.427436823104692, "grad_norm": 2.4946787357330322, "learning_rate": 1.7862815884476536e-05, "loss": 1.4168, "step": 22752 }, { "epoch": 16.43176895306859, "grad_norm": 2.5525853633880615, "learning_rate": 1.784115523465704e-05, "loss": 1.3639, "step": 22758 }, { "epoch": 16.43610108303249, "grad_norm": 2.316025495529175, "learning_rate": 1.7819494584837545e-05, "loss": 1.5194, "step": 22764 }, { "epoch": 16.44043321299639, "grad_norm": 2.3559882640838623, "learning_rate": 1.7797833935018052e-05, "loss": 1.3632, "step": 22770 }, { "epoch": 16.444765342960288, "grad_norm": 2.555443048477173, "learning_rate": 1.777617328519856e-05, "loss": 1.4648, "step": 22776 }, { "epoch": 16.449097472924187, "grad_norm": 2.5540757179260254, "learning_rate": 1.7754512635379062e-05, "loss": 1.4642, "step": 22782 }, { "epoch": 16.453429602888086, "grad_norm": 2.6443629264831543, "learning_rate": 1.773285198555957e-05, "loss": 1.4452, "step": 22788 }, { "epoch": 16.457761732851985, "grad_norm": 2.4811880588531494, "learning_rate": 1.7711191335740075e-05, "loss": 1.5749, "step": 22794 }, { "epoch": 16.462093862815884, "grad_norm": 2.2384231090545654, "learning_rate": 1.768953068592058e-05, "loss": 1.348, "step": 22800 }, { "epoch": 16.466425992779783, "grad_norm": 2.51796555519104, "learning_rate": 1.7667870036101085e-05, "loss": 1.5423, "step": 22806 }, { "epoch": 16.470758122743682, "grad_norm": 2.452350616455078, "learning_rate": 1.7646209386281588e-05, "loss": 1.5114, "step": 22812 }, { "epoch": 16.47509025270758, "grad_norm": 2.3028147220611572, "learning_rate": 1.7624548736462095e-05, "loss": 1.4882, "step": 22818 }, { "epoch": 16.47942238267148, "grad_norm": 2.316208600997925, "learning_rate": 1.7602888086642598e-05, "loss": 1.4822, "step": 22824 }, { "epoch": 16.48375451263538, "grad_norm": 2.498546600341797, "learning_rate": 1.7581227436823105e-05, "loss": 1.493, "step": 22830 }, { "epoch": 16.488086642599278, "grad_norm": 2.5225257873535156, "learning_rate": 1.755956678700361e-05, "loss": 1.5044, "step": 22836 }, { "epoch": 16.492418772563177, "grad_norm": 2.4086999893188477, "learning_rate": 1.7537906137184114e-05, "loss": 1.4648, "step": 22842 }, { "epoch": 16.496750902527076, "grad_norm": 2.7733840942382812, "learning_rate": 1.751624548736462e-05, "loss": 1.4865, "step": 22848 }, { "epoch": 16.501083032490975, "grad_norm": 2.175537109375, "learning_rate": 1.7494584837545128e-05, "loss": 1.3571, "step": 22854 }, { "epoch": 16.505415162454874, "grad_norm": 2.6582541465759277, "learning_rate": 1.7472924187725634e-05, "loss": 1.5443, "step": 22860 }, { "epoch": 16.509747292418773, "grad_norm": 2.55281662940979, "learning_rate": 1.7451263537906137e-05, "loss": 1.4079, "step": 22866 }, { "epoch": 16.51407942238267, "grad_norm": 2.3408420085906982, "learning_rate": 1.7429602888086644e-05, "loss": 1.5346, "step": 22872 }, { "epoch": 16.51841155234657, "grad_norm": 2.494788885116577, "learning_rate": 1.740794223826715e-05, "loss": 1.4726, "step": 22878 }, { "epoch": 16.52274368231047, "grad_norm": 2.5037355422973633, "learning_rate": 1.7386281588447654e-05, "loss": 1.3863, "step": 22884 }, { "epoch": 16.52707581227437, "grad_norm": 2.5348050594329834, "learning_rate": 1.736462093862816e-05, "loss": 1.5323, "step": 22890 }, { "epoch": 16.531407942238268, "grad_norm": 2.672790288925171, "learning_rate": 1.7342960288808667e-05, "loss": 1.4563, "step": 22896 }, { "epoch": 16.535740072202167, "grad_norm": 2.71803879737854, "learning_rate": 1.732129963898917e-05, "loss": 1.4346, "step": 22902 }, { "epoch": 16.540072202166066, "grad_norm": 2.688337564468384, "learning_rate": 1.7299638989169674e-05, "loss": 1.395, "step": 22908 }, { "epoch": 16.544404332129965, "grad_norm": 2.433440923690796, "learning_rate": 1.727797833935018e-05, "loss": 1.5151, "step": 22914 }, { "epoch": 16.548736462093864, "grad_norm": 2.4048264026641846, "learning_rate": 1.7256317689530687e-05, "loss": 1.4035, "step": 22920 }, { "epoch": 16.553068592057762, "grad_norm": 2.410795211791992, "learning_rate": 1.723465703971119e-05, "loss": 1.3782, "step": 22926 }, { "epoch": 16.55740072202166, "grad_norm": 2.6115918159484863, "learning_rate": 1.7212996389891697e-05, "loss": 1.4617, "step": 22932 }, { "epoch": 16.56173285198556, "grad_norm": 2.3363890647888184, "learning_rate": 1.7191335740072203e-05, "loss": 1.4579, "step": 22938 }, { "epoch": 16.56606498194946, "grad_norm": 2.2710204124450684, "learning_rate": 1.7169675090252707e-05, "loss": 1.3388, "step": 22944 }, { "epoch": 16.57039711191336, "grad_norm": 2.3936610221862793, "learning_rate": 1.7148014440433213e-05, "loss": 1.3898, "step": 22950 }, { "epoch": 16.574729241877257, "grad_norm": 2.5087857246398926, "learning_rate": 1.712635379061372e-05, "loss": 1.4828, "step": 22956 }, { "epoch": 16.579061371841156, "grad_norm": 2.3738107681274414, "learning_rate": 1.7104693140794226e-05, "loss": 1.3853, "step": 22962 }, { "epoch": 16.583393501805055, "grad_norm": 2.4679672718048096, "learning_rate": 1.708303249097473e-05, "loss": 1.4615, "step": 22968 }, { "epoch": 16.587725631768954, "grad_norm": 2.5768115520477295, "learning_rate": 1.7061371841155236e-05, "loss": 1.3918, "step": 22974 }, { "epoch": 16.592057761732853, "grad_norm": 2.7615976333618164, "learning_rate": 1.7039711191335743e-05, "loss": 1.4507, "step": 22980 }, { "epoch": 16.596389891696752, "grad_norm": 2.372880697250366, "learning_rate": 1.7018050541516246e-05, "loss": 1.4592, "step": 22986 }, { "epoch": 16.60072202166065, "grad_norm": 2.6551382541656494, "learning_rate": 1.6996389891696753e-05, "loss": 1.5053, "step": 22992 }, { "epoch": 16.60505415162455, "grad_norm": 2.6750540733337402, "learning_rate": 1.6974729241877256e-05, "loss": 1.4798, "step": 22998 }, { "epoch": 16.60938628158845, "grad_norm": 2.399225950241089, "learning_rate": 1.6953068592057763e-05, "loss": 1.4142, "step": 23004 }, { "epoch": 16.613718411552348, "grad_norm": 2.401569128036499, "learning_rate": 1.6931407942238266e-05, "loss": 1.4267, "step": 23010 }, { "epoch": 16.618050541516247, "grad_norm": 2.489203929901123, "learning_rate": 1.6909747292418772e-05, "loss": 1.4724, "step": 23016 }, { "epoch": 16.622382671480146, "grad_norm": 2.435969829559326, "learning_rate": 1.688808664259928e-05, "loss": 1.5232, "step": 23022 }, { "epoch": 16.626714801444045, "grad_norm": 2.4107506275177, "learning_rate": 1.6866425992779782e-05, "loss": 1.4158, "step": 23028 }, { "epoch": 16.63104693140794, "grad_norm": 2.573716402053833, "learning_rate": 1.684476534296029e-05, "loss": 1.3711, "step": 23034 }, { "epoch": 16.63537906137184, "grad_norm": 2.004575252532959, "learning_rate": 1.6823104693140795e-05, "loss": 1.358, "step": 23040 }, { "epoch": 16.63971119133574, "grad_norm": 2.6446595191955566, "learning_rate": 1.6801444043321302e-05, "loss": 1.3912, "step": 23046 }, { "epoch": 16.644043321299637, "grad_norm": 2.4614298343658447, "learning_rate": 1.6779783393501805e-05, "loss": 1.4852, "step": 23052 }, { "epoch": 16.648375451263536, "grad_norm": 2.393332004547119, "learning_rate": 1.6758122743682312e-05, "loss": 1.4543, "step": 23058 }, { "epoch": 16.652707581227435, "grad_norm": 2.272947072982788, "learning_rate": 1.673646209386282e-05, "loss": 1.4464, "step": 23064 }, { "epoch": 16.657039711191334, "grad_norm": 2.247331380844116, "learning_rate": 1.6714801444043322e-05, "loss": 1.4653, "step": 23070 }, { "epoch": 16.661371841155233, "grad_norm": 2.5861055850982666, "learning_rate": 1.669314079422383e-05, "loss": 1.465, "step": 23076 }, { "epoch": 16.665703971119132, "grad_norm": 2.5227479934692383, "learning_rate": 1.6671480144404335e-05, "loss": 1.5217, "step": 23082 }, { "epoch": 16.67003610108303, "grad_norm": 2.564108371734619, "learning_rate": 1.6649819494584838e-05, "loss": 1.4848, "step": 23088 }, { "epoch": 16.67436823104693, "grad_norm": 2.8517096042633057, "learning_rate": 1.662815884476534e-05, "loss": 1.3214, "step": 23094 }, { "epoch": 16.67870036101083, "grad_norm": 2.5853562355041504, "learning_rate": 1.6606498194945848e-05, "loss": 1.4934, "step": 23100 }, { "epoch": 16.68303249097473, "grad_norm": 2.6312859058380127, "learning_rate": 1.6584837545126355e-05, "loss": 1.4392, "step": 23106 }, { "epoch": 16.687364620938627, "grad_norm": 2.399897575378418, "learning_rate": 1.6563176895306858e-05, "loss": 1.5298, "step": 23112 }, { "epoch": 16.691696750902526, "grad_norm": 2.7244017124176025, "learning_rate": 1.6541516245487364e-05, "loss": 1.4956, "step": 23118 }, { "epoch": 16.696028880866425, "grad_norm": 2.3789844512939453, "learning_rate": 1.651985559566787e-05, "loss": 1.3185, "step": 23124 }, { "epoch": 16.700361010830324, "grad_norm": 2.421360492706299, "learning_rate": 1.6498194945848378e-05, "loss": 1.5378, "step": 23130 }, { "epoch": 16.704693140794223, "grad_norm": 2.483761787414551, "learning_rate": 1.647653429602888e-05, "loss": 1.5626, "step": 23136 }, { "epoch": 16.709025270758122, "grad_norm": 2.2735414505004883, "learning_rate": 1.6454873646209388e-05, "loss": 1.3781, "step": 23142 }, { "epoch": 16.71335740072202, "grad_norm": 2.300145149230957, "learning_rate": 1.6433212996389894e-05, "loss": 1.4193, "step": 23148 }, { "epoch": 16.71768953068592, "grad_norm": 2.340571880340576, "learning_rate": 1.6411552346570397e-05, "loss": 1.4196, "step": 23154 }, { "epoch": 16.72202166064982, "grad_norm": 2.871610164642334, "learning_rate": 1.6389891696750904e-05, "loss": 1.488, "step": 23160 }, { "epoch": 16.726353790613718, "grad_norm": 2.5283730030059814, "learning_rate": 1.636823104693141e-05, "loss": 1.4689, "step": 23166 }, { "epoch": 16.730685920577617, "grad_norm": 2.6540722846984863, "learning_rate": 1.6346570397111917e-05, "loss": 1.5364, "step": 23172 }, { "epoch": 16.735018050541516, "grad_norm": 2.639061450958252, "learning_rate": 1.632490974729242e-05, "loss": 1.4135, "step": 23178 }, { "epoch": 16.739350180505415, "grad_norm": 2.6172256469726562, "learning_rate": 1.6303249097472924e-05, "loss": 1.4596, "step": 23184 }, { "epoch": 16.743682310469314, "grad_norm": 2.2530887126922607, "learning_rate": 1.628158844765343e-05, "loss": 1.5854, "step": 23190 }, { "epoch": 16.748014440433213, "grad_norm": 2.523808240890503, "learning_rate": 1.6259927797833933e-05, "loss": 1.4117, "step": 23196 }, { "epoch": 16.752346570397112, "grad_norm": 2.4866883754730225, "learning_rate": 1.623826714801444e-05, "loss": 1.4367, "step": 23202 }, { "epoch": 16.75667870036101, "grad_norm": 2.3686861991882324, "learning_rate": 1.6216606498194947e-05, "loss": 1.4187, "step": 23208 }, { "epoch": 16.76101083032491, "grad_norm": 2.4417686462402344, "learning_rate": 1.619494584837545e-05, "loss": 1.3545, "step": 23214 }, { "epoch": 16.76534296028881, "grad_norm": 2.469270706176758, "learning_rate": 1.6173285198555957e-05, "loss": 1.4951, "step": 23220 }, { "epoch": 16.769675090252708, "grad_norm": 2.47448992729187, "learning_rate": 1.6151624548736463e-05, "loss": 1.472, "step": 23226 }, { "epoch": 16.774007220216607, "grad_norm": 2.3079230785369873, "learning_rate": 1.612996389891697e-05, "loss": 1.3115, "step": 23232 }, { "epoch": 16.778339350180506, "grad_norm": 2.229299783706665, "learning_rate": 1.6108303249097473e-05, "loss": 1.5361, "step": 23238 }, { "epoch": 16.782671480144405, "grad_norm": 2.2769365310668945, "learning_rate": 1.608664259927798e-05, "loss": 1.3152, "step": 23244 }, { "epoch": 16.787003610108304, "grad_norm": 2.4418256282806396, "learning_rate": 1.6064981949458486e-05, "loss": 1.4943, "step": 23250 }, { "epoch": 16.791335740072203, "grad_norm": 2.488654375076294, "learning_rate": 1.604332129963899e-05, "loss": 1.4265, "step": 23256 }, { "epoch": 16.7956678700361, "grad_norm": 2.7736313343048096, "learning_rate": 1.6021660649819496e-05, "loss": 1.5224, "step": 23262 }, { "epoch": 16.8, "grad_norm": 2.3912503719329834, "learning_rate": 1.6000000000000003e-05, "loss": 1.2998, "step": 23268 }, { "epoch": 16.8043321299639, "grad_norm": 2.394449472427368, "learning_rate": 1.5978339350180506e-05, "loss": 1.4648, "step": 23274 }, { "epoch": 16.8086642599278, "grad_norm": 2.6352081298828125, "learning_rate": 1.595667870036101e-05, "loss": 1.2788, "step": 23280 }, { "epoch": 16.812996389891698, "grad_norm": 2.7116611003875732, "learning_rate": 1.5935018050541516e-05, "loss": 1.4498, "step": 23286 }, { "epoch": 16.817328519855597, "grad_norm": 2.5774290561676025, "learning_rate": 1.5913357400722022e-05, "loss": 1.425, "step": 23292 }, { "epoch": 16.821660649819496, "grad_norm": 2.532975912094116, "learning_rate": 1.5891696750902526e-05, "loss": 1.4639, "step": 23298 }, { "epoch": 16.825992779783395, "grad_norm": 2.75736927986145, "learning_rate": 1.5870036101083032e-05, "loss": 1.4878, "step": 23304 }, { "epoch": 16.830324909747294, "grad_norm": 2.4092602729797363, "learning_rate": 1.584837545126354e-05, "loss": 1.4268, "step": 23310 }, { "epoch": 16.834657039711193, "grad_norm": 2.600078582763672, "learning_rate": 1.5826714801444045e-05, "loss": 1.3579, "step": 23316 }, { "epoch": 16.83898916967509, "grad_norm": 2.288705587387085, "learning_rate": 1.580505415162455e-05, "loss": 1.3765, "step": 23322 }, { "epoch": 16.84332129963899, "grad_norm": 2.597463846206665, "learning_rate": 1.5783393501805055e-05, "loss": 1.4543, "step": 23328 }, { "epoch": 16.84765342960289, "grad_norm": 2.6238198280334473, "learning_rate": 1.5761732851985562e-05, "loss": 1.3604, "step": 23334 }, { "epoch": 16.85198555956679, "grad_norm": 2.3731727600097656, "learning_rate": 1.5740072202166065e-05, "loss": 1.4625, "step": 23340 }, { "epoch": 16.856317689530687, "grad_norm": 2.3422799110412598, "learning_rate": 1.5718411552346572e-05, "loss": 1.5115, "step": 23346 }, { "epoch": 16.860649819494586, "grad_norm": 2.696096897125244, "learning_rate": 1.569675090252708e-05, "loss": 1.5405, "step": 23352 }, { "epoch": 16.864981949458485, "grad_norm": 2.503427505493164, "learning_rate": 1.5675090252707585e-05, "loss": 1.404, "step": 23358 }, { "epoch": 16.869314079422384, "grad_norm": 2.858323574066162, "learning_rate": 1.5653429602888088e-05, "loss": 1.4487, "step": 23364 }, { "epoch": 16.87364620938628, "grad_norm": 2.6496741771698, "learning_rate": 1.563176895306859e-05, "loss": 1.4743, "step": 23370 }, { "epoch": 16.87797833935018, "grad_norm": 2.524414539337158, "learning_rate": 1.5610108303249098e-05, "loss": 1.5435, "step": 23376 }, { "epoch": 16.882310469314078, "grad_norm": 2.412529468536377, "learning_rate": 1.55884476534296e-05, "loss": 1.4572, "step": 23382 }, { "epoch": 16.886642599277977, "grad_norm": 2.561990737915039, "learning_rate": 1.5566787003610108e-05, "loss": 1.514, "step": 23388 }, { "epoch": 16.890974729241876, "grad_norm": 2.4362306594848633, "learning_rate": 1.5545126353790614e-05, "loss": 1.5415, "step": 23394 }, { "epoch": 16.895306859205775, "grad_norm": 2.793246269226074, "learning_rate": 1.552346570397112e-05, "loss": 1.4331, "step": 23400 }, { "epoch": 16.899638989169674, "grad_norm": 2.4868383407592773, "learning_rate": 1.5501805054151624e-05, "loss": 1.4649, "step": 23406 }, { "epoch": 16.903971119133573, "grad_norm": 2.4425199031829834, "learning_rate": 1.548014440433213e-05, "loss": 1.4188, "step": 23412 }, { "epoch": 16.90830324909747, "grad_norm": 2.5915751457214355, "learning_rate": 1.5458483754512638e-05, "loss": 1.4117, "step": 23418 }, { "epoch": 16.91263537906137, "grad_norm": 2.706326723098755, "learning_rate": 1.543682310469314e-05, "loss": 1.4637, "step": 23424 }, { "epoch": 16.91696750902527, "grad_norm": 2.4495041370391846, "learning_rate": 1.5415162454873647e-05, "loss": 1.4349, "step": 23430 }, { "epoch": 16.92129963898917, "grad_norm": 2.2014851570129395, "learning_rate": 1.5393501805054154e-05, "loss": 1.3, "step": 23436 }, { "epoch": 16.925631768953068, "grad_norm": 2.5050792694091797, "learning_rate": 1.5371841155234657e-05, "loss": 1.4674, "step": 23442 }, { "epoch": 16.929963898916967, "grad_norm": 2.538947582244873, "learning_rate": 1.5350180505415164e-05, "loss": 1.5015, "step": 23448 }, { "epoch": 16.934296028880865, "grad_norm": 2.6577084064483643, "learning_rate": 1.532851985559567e-05, "loss": 1.4632, "step": 23454 }, { "epoch": 16.938628158844764, "grad_norm": 2.4561915397644043, "learning_rate": 1.5306859205776174e-05, "loss": 1.4468, "step": 23460 }, { "epoch": 16.942960288808663, "grad_norm": 2.5064072608947754, "learning_rate": 1.5285198555956677e-05, "loss": 1.5003, "step": 23466 }, { "epoch": 16.947292418772562, "grad_norm": 2.5761220455169678, "learning_rate": 1.5263537906137183e-05, "loss": 1.4073, "step": 23472 }, { "epoch": 16.95162454873646, "grad_norm": 2.4160232543945312, "learning_rate": 1.524187725631769e-05, "loss": 1.379, "step": 23478 }, { "epoch": 16.95595667870036, "grad_norm": 2.325554132461548, "learning_rate": 1.5220216606498195e-05, "loss": 1.5074, "step": 23484 }, { "epoch": 16.96028880866426, "grad_norm": 2.249159336090088, "learning_rate": 1.51985559566787e-05, "loss": 1.4449, "step": 23490 }, { "epoch": 16.96462093862816, "grad_norm": 2.448500394821167, "learning_rate": 1.5176895306859207e-05, "loss": 1.5727, "step": 23496 }, { "epoch": 16.968953068592057, "grad_norm": 2.455134153366089, "learning_rate": 1.5155234657039713e-05, "loss": 1.4126, "step": 23502 }, { "epoch": 16.973285198555956, "grad_norm": 2.4678142070770264, "learning_rate": 1.5133574007220216e-05, "loss": 1.4453, "step": 23508 }, { "epoch": 16.977617328519855, "grad_norm": 3.0327908992767334, "learning_rate": 1.5111913357400723e-05, "loss": 1.5842, "step": 23514 }, { "epoch": 16.981949458483754, "grad_norm": 2.7425310611724854, "learning_rate": 1.509025270758123e-05, "loss": 1.5046, "step": 23520 }, { "epoch": 16.986281588447653, "grad_norm": 2.561739444732666, "learning_rate": 1.5068592057761733e-05, "loss": 1.4805, "step": 23526 }, { "epoch": 16.990613718411552, "grad_norm": 2.485543966293335, "learning_rate": 1.5046931407942238e-05, "loss": 1.3486, "step": 23532 }, { "epoch": 16.99494584837545, "grad_norm": 2.3090429306030273, "learning_rate": 1.5025270758122744e-05, "loss": 1.3476, "step": 23538 }, { "epoch": 16.99927797833935, "grad_norm": 2.4579086303710938, "learning_rate": 1.5003610108303251e-05, "loss": 1.5035, "step": 23544 }, { "epoch": 17.00361010830325, "grad_norm": 2.477008581161499, "learning_rate": 1.4981949458483754e-05, "loss": 1.5333, "step": 23550 }, { "epoch": 17.007942238267148, "grad_norm": 2.608142375946045, "learning_rate": 1.496028880866426e-05, "loss": 1.4685, "step": 23556 }, { "epoch": 17.012274368231047, "grad_norm": 2.6482203006744385, "learning_rate": 1.4938628158844767e-05, "loss": 1.584, "step": 23562 }, { "epoch": 17.016606498194946, "grad_norm": 2.4517416954040527, "learning_rate": 1.491696750902527e-05, "loss": 1.4876, "step": 23568 }, { "epoch": 17.020938628158845, "grad_norm": 2.381174087524414, "learning_rate": 1.4895306859205777e-05, "loss": 1.425, "step": 23574 }, { "epoch": 17.025270758122744, "grad_norm": 2.6705000400543213, "learning_rate": 1.4873646209386282e-05, "loss": 1.4427, "step": 23580 }, { "epoch": 17.029602888086643, "grad_norm": 2.2959063053131104, "learning_rate": 1.4851985559566789e-05, "loss": 1.4459, "step": 23586 }, { "epoch": 17.033935018050542, "grad_norm": 2.41650390625, "learning_rate": 1.4830324909747292e-05, "loss": 1.4474, "step": 23592 }, { "epoch": 17.03826714801444, "grad_norm": 2.560232162475586, "learning_rate": 1.4808664259927799e-05, "loss": 1.3597, "step": 23598 }, { "epoch": 17.04259927797834, "grad_norm": 2.5356595516204834, "learning_rate": 1.4787003610108305e-05, "loss": 1.2892, "step": 23604 }, { "epoch": 17.04693140794224, "grad_norm": 2.510014295578003, "learning_rate": 1.4765342960288808e-05, "loss": 1.3829, "step": 23610 }, { "epoch": 17.051263537906138, "grad_norm": 2.318939685821533, "learning_rate": 1.4743682310469315e-05, "loss": 1.4701, "step": 23616 }, { "epoch": 17.055595667870037, "grad_norm": 2.6170434951782227, "learning_rate": 1.472202166064982e-05, "loss": 1.3824, "step": 23622 }, { "epoch": 17.059927797833936, "grad_norm": 2.506161689758301, "learning_rate": 1.4700361010830327e-05, "loss": 1.4884, "step": 23628 }, { "epoch": 17.064259927797835, "grad_norm": 2.478917121887207, "learning_rate": 1.467870036101083e-05, "loss": 1.4752, "step": 23634 }, { "epoch": 17.068592057761734, "grad_norm": 2.341660737991333, "learning_rate": 1.4657039711191336e-05, "loss": 1.4602, "step": 23640 }, { "epoch": 17.072924187725633, "grad_norm": 2.606437921524048, "learning_rate": 1.4635379061371843e-05, "loss": 1.3973, "step": 23646 }, { "epoch": 17.077256317689532, "grad_norm": 2.6434786319732666, "learning_rate": 1.4613718411552346e-05, "loss": 1.5355, "step": 23652 }, { "epoch": 17.08158844765343, "grad_norm": 2.7129669189453125, "learning_rate": 1.4592057761732853e-05, "loss": 1.4673, "step": 23658 }, { "epoch": 17.08592057761733, "grad_norm": 2.8241469860076904, "learning_rate": 1.4570397111913358e-05, "loss": 1.3978, "step": 23664 }, { "epoch": 17.09025270758123, "grad_norm": 2.363154411315918, "learning_rate": 1.4548736462093864e-05, "loss": 1.499, "step": 23670 }, { "epoch": 17.094584837545128, "grad_norm": 2.442112922668457, "learning_rate": 1.4527075812274368e-05, "loss": 1.4236, "step": 23676 }, { "epoch": 17.098916967509027, "grad_norm": 2.729447841644287, "learning_rate": 1.4505415162454874e-05, "loss": 1.4478, "step": 23682 }, { "epoch": 17.103249097472926, "grad_norm": 2.542825937271118, "learning_rate": 1.4483754512635381e-05, "loss": 1.3623, "step": 23688 }, { "epoch": 17.107581227436825, "grad_norm": 2.515653610229492, "learning_rate": 1.4462093862815884e-05, "loss": 1.3836, "step": 23694 }, { "epoch": 17.111913357400724, "grad_norm": 2.5753135681152344, "learning_rate": 1.444043321299639e-05, "loss": 1.4275, "step": 23700 }, { "epoch": 17.116245487364623, "grad_norm": 2.3360183238983154, "learning_rate": 1.4418772563176897e-05, "loss": 1.3178, "step": 23706 }, { "epoch": 17.12057761732852, "grad_norm": 2.353851795196533, "learning_rate": 1.43971119133574e-05, "loss": 1.3929, "step": 23712 }, { "epoch": 17.12490974729242, "grad_norm": 2.4803409576416016, "learning_rate": 1.4375451263537905e-05, "loss": 1.4446, "step": 23718 }, { "epoch": 17.129241877256316, "grad_norm": 2.5297374725341797, "learning_rate": 1.4353790613718412e-05, "loss": 1.4817, "step": 23724 }, { "epoch": 17.133574007220215, "grad_norm": 2.533616065979004, "learning_rate": 1.4332129963898919e-05, "loss": 1.5024, "step": 23730 }, { "epoch": 17.137906137184114, "grad_norm": 2.314608335494995, "learning_rate": 1.4310469314079422e-05, "loss": 1.364, "step": 23736 }, { "epoch": 17.142238267148013, "grad_norm": 2.4799365997314453, "learning_rate": 1.4288808664259929e-05, "loss": 1.4281, "step": 23742 }, { "epoch": 17.146570397111912, "grad_norm": 2.233015775680542, "learning_rate": 1.4267148014440435e-05, "loss": 1.389, "step": 23748 }, { "epoch": 17.15090252707581, "grad_norm": 2.681321620941162, "learning_rate": 1.4245487364620938e-05, "loss": 1.4129, "step": 23754 }, { "epoch": 17.15523465703971, "grad_norm": 2.20121693611145, "learning_rate": 1.4223826714801445e-05, "loss": 1.4271, "step": 23760 }, { "epoch": 17.15956678700361, "grad_norm": 2.5994303226470947, "learning_rate": 1.420216606498195e-05, "loss": 1.3973, "step": 23766 }, { "epoch": 17.163898916967508, "grad_norm": 2.618360757827759, "learning_rate": 1.4180505415162457e-05, "loss": 1.4946, "step": 23772 }, { "epoch": 17.168231046931407, "grad_norm": 2.3640692234039307, "learning_rate": 1.415884476534296e-05, "loss": 1.4128, "step": 23778 }, { "epoch": 17.172563176895306, "grad_norm": 2.3687450885772705, "learning_rate": 1.4137184115523466e-05, "loss": 1.4258, "step": 23784 }, { "epoch": 17.176895306859205, "grad_norm": 2.7463719844818115, "learning_rate": 1.4115523465703973e-05, "loss": 1.5371, "step": 23790 }, { "epoch": 17.181227436823104, "grad_norm": 2.3834078311920166, "learning_rate": 1.4093862815884476e-05, "loss": 1.4738, "step": 23796 }, { "epoch": 17.185559566787003, "grad_norm": 2.277010202407837, "learning_rate": 1.4072202166064983e-05, "loss": 1.3797, "step": 23802 }, { "epoch": 17.1898916967509, "grad_norm": 2.5202488899230957, "learning_rate": 1.4050541516245488e-05, "loss": 1.3969, "step": 23808 }, { "epoch": 17.1942238267148, "grad_norm": 2.424647808074951, "learning_rate": 1.4028880866425994e-05, "loss": 1.4479, "step": 23814 }, { "epoch": 17.1985559566787, "grad_norm": 2.567002534866333, "learning_rate": 1.4007220216606498e-05, "loss": 1.3683, "step": 23820 }, { "epoch": 17.2028880866426, "grad_norm": 2.569549798965454, "learning_rate": 1.3985559566787004e-05, "loss": 1.4406, "step": 23826 }, { "epoch": 17.207220216606498, "grad_norm": 2.498809814453125, "learning_rate": 1.396389891696751e-05, "loss": 1.4332, "step": 23832 }, { "epoch": 17.211552346570397, "grad_norm": 2.766265630722046, "learning_rate": 1.3942238267148014e-05, "loss": 1.5167, "step": 23838 }, { "epoch": 17.215884476534296, "grad_norm": 2.53617262840271, "learning_rate": 1.392057761732852e-05, "loss": 1.4672, "step": 23844 }, { "epoch": 17.220216606498195, "grad_norm": 2.3539557456970215, "learning_rate": 1.3898916967509026e-05, "loss": 1.6088, "step": 23850 }, { "epoch": 17.224548736462094, "grad_norm": 2.33740496635437, "learning_rate": 1.3877256317689532e-05, "loss": 1.4044, "step": 23856 }, { "epoch": 17.228880866425992, "grad_norm": 2.286471366882324, "learning_rate": 1.3855595667870035e-05, "loss": 1.4392, "step": 23862 }, { "epoch": 17.23321299638989, "grad_norm": 2.5656025409698486, "learning_rate": 1.3833935018050542e-05, "loss": 1.435, "step": 23868 }, { "epoch": 17.23754512635379, "grad_norm": 2.309931516647339, "learning_rate": 1.3812274368231049e-05, "loss": 1.4292, "step": 23874 }, { "epoch": 17.24187725631769, "grad_norm": 2.862494468688965, "learning_rate": 1.3790613718411552e-05, "loss": 1.3964, "step": 23880 }, { "epoch": 17.24620938628159, "grad_norm": 2.8147995471954346, "learning_rate": 1.3768953068592058e-05, "loss": 1.4962, "step": 23886 }, { "epoch": 17.250541516245487, "grad_norm": 2.527841806411743, "learning_rate": 1.3747292418772565e-05, "loss": 1.4858, "step": 23892 }, { "epoch": 17.254873646209386, "grad_norm": 2.740058183670044, "learning_rate": 1.372563176895307e-05, "loss": 1.532, "step": 23898 }, { "epoch": 17.259205776173285, "grad_norm": 2.6253626346588135, "learning_rate": 1.3703971119133573e-05, "loss": 1.3087, "step": 23904 }, { "epoch": 17.263537906137184, "grad_norm": 2.4151246547698975, "learning_rate": 1.368231046931408e-05, "loss": 1.5174, "step": 23910 }, { "epoch": 17.267870036101083, "grad_norm": 2.427682876586914, "learning_rate": 1.3660649819494586e-05, "loss": 1.4011, "step": 23916 }, { "epoch": 17.272202166064982, "grad_norm": 2.8744406700134277, "learning_rate": 1.363898916967509e-05, "loss": 1.3485, "step": 23922 }, { "epoch": 17.27653429602888, "grad_norm": 2.6172618865966797, "learning_rate": 1.3617328519855596e-05, "loss": 1.424, "step": 23928 }, { "epoch": 17.28086642599278, "grad_norm": 2.6119942665100098, "learning_rate": 1.3595667870036103e-05, "loss": 1.4184, "step": 23934 }, { "epoch": 17.28519855595668, "grad_norm": 2.4191207885742188, "learning_rate": 1.3574007220216608e-05, "loss": 1.5006, "step": 23940 }, { "epoch": 17.289530685920578, "grad_norm": 2.242152690887451, "learning_rate": 1.3552346570397113e-05, "loss": 1.423, "step": 23946 }, { "epoch": 17.293862815884477, "grad_norm": 2.566763401031494, "learning_rate": 1.3530685920577618e-05, "loss": 1.366, "step": 23952 }, { "epoch": 17.298194945848376, "grad_norm": 2.7793266773223877, "learning_rate": 1.3509025270758124e-05, "loss": 1.3405, "step": 23958 }, { "epoch": 17.302527075812275, "grad_norm": 2.5584468841552734, "learning_rate": 1.3487364620938628e-05, "loss": 1.5695, "step": 23964 }, { "epoch": 17.306859205776174, "grad_norm": 2.4625706672668457, "learning_rate": 1.3465703971119134e-05, "loss": 1.2869, "step": 23970 }, { "epoch": 17.311191335740073, "grad_norm": 2.2971532344818115, "learning_rate": 1.344404332129964e-05, "loss": 1.4768, "step": 23976 }, { "epoch": 17.315523465703972, "grad_norm": 2.4783596992492676, "learning_rate": 1.3422382671480144e-05, "loss": 1.5604, "step": 23982 }, { "epoch": 17.31985559566787, "grad_norm": 2.339940071105957, "learning_rate": 1.340072202166065e-05, "loss": 1.4752, "step": 23988 }, { "epoch": 17.32418772563177, "grad_norm": 2.6752119064331055, "learning_rate": 1.3379061371841155e-05, "loss": 1.4672, "step": 23994 }, { "epoch": 17.32851985559567, "grad_norm": 2.783220052719116, "learning_rate": 1.3357400722021662e-05, "loss": 1.4946, "step": 24000 }, { "epoch": 17.332851985559568, "grad_norm": 2.5144336223602295, "learning_rate": 1.3335740072202165e-05, "loss": 1.3705, "step": 24006 }, { "epoch": 17.337184115523467, "grad_norm": 2.536536455154419, "learning_rate": 1.3314079422382672e-05, "loss": 1.4713, "step": 24012 }, { "epoch": 17.341516245487366, "grad_norm": 2.5503783226013184, "learning_rate": 1.3292418772563179e-05, "loss": 1.4397, "step": 24018 }, { "epoch": 17.345848375451265, "grad_norm": 2.8503215312957764, "learning_rate": 1.3270758122743682e-05, "loss": 1.4911, "step": 24024 }, { "epoch": 17.350180505415164, "grad_norm": 2.40666127204895, "learning_rate": 1.3249097472924188e-05, "loss": 1.4467, "step": 24030 }, { "epoch": 17.354512635379063, "grad_norm": 2.413696765899658, "learning_rate": 1.3227436823104693e-05, "loss": 1.5287, "step": 24036 }, { "epoch": 17.358844765342962, "grad_norm": 2.7524938583374023, "learning_rate": 1.32057761732852e-05, "loss": 1.4347, "step": 24042 }, { "epoch": 17.36317689530686, "grad_norm": 2.599123477935791, "learning_rate": 1.3184115523465703e-05, "loss": 1.3856, "step": 24048 }, { "epoch": 17.36750902527076, "grad_norm": 2.5550944805145264, "learning_rate": 1.316245487364621e-05, "loss": 1.5387, "step": 24054 }, { "epoch": 17.37184115523466, "grad_norm": 2.4486656188964844, "learning_rate": 1.3140794223826716e-05, "loss": 1.45, "step": 24060 }, { "epoch": 17.376173285198554, "grad_norm": 2.596395492553711, "learning_rate": 1.311913357400722e-05, "loss": 1.474, "step": 24066 }, { "epoch": 17.380505415162453, "grad_norm": 2.645688772201538, "learning_rate": 1.3097472924187726e-05, "loss": 1.4237, "step": 24072 }, { "epoch": 17.384837545126352, "grad_norm": 2.434633731842041, "learning_rate": 1.3075812274368233e-05, "loss": 1.4618, "step": 24078 }, { "epoch": 17.38916967509025, "grad_norm": 2.5136797428131104, "learning_rate": 1.3054151624548738e-05, "loss": 1.4618, "step": 24084 }, { "epoch": 17.39350180505415, "grad_norm": 2.4203014373779297, "learning_rate": 1.3032490974729241e-05, "loss": 1.4701, "step": 24090 }, { "epoch": 17.39783393501805, "grad_norm": 2.57486891746521, "learning_rate": 1.3010830324909748e-05, "loss": 1.4808, "step": 24096 }, { "epoch": 17.402166064981948, "grad_norm": 2.357024908065796, "learning_rate": 1.2989169675090254e-05, "loss": 1.345, "step": 24102 }, { "epoch": 17.406498194945847, "grad_norm": 2.585263729095459, "learning_rate": 1.2967509025270757e-05, "loss": 1.4092, "step": 24108 }, { "epoch": 17.410830324909746, "grad_norm": 2.3996474742889404, "learning_rate": 1.2945848375451264e-05, "loss": 1.4054, "step": 24114 }, { "epoch": 17.415162454873645, "grad_norm": 2.7590889930725098, "learning_rate": 1.292418772563177e-05, "loss": 1.4151, "step": 24120 }, { "epoch": 17.419494584837544, "grad_norm": 2.5270371437072754, "learning_rate": 1.2902527075812276e-05, "loss": 1.3224, "step": 24126 }, { "epoch": 17.423826714801443, "grad_norm": 2.2922046184539795, "learning_rate": 1.288086642599278e-05, "loss": 1.5052, "step": 24132 }, { "epoch": 17.428158844765342, "grad_norm": 2.564640998840332, "learning_rate": 1.2859205776173285e-05, "loss": 1.5052, "step": 24138 }, { "epoch": 17.43249097472924, "grad_norm": 2.228468179702759, "learning_rate": 1.2837545126353792e-05, "loss": 1.3956, "step": 24144 }, { "epoch": 17.43682310469314, "grad_norm": 2.3548102378845215, "learning_rate": 1.2815884476534295e-05, "loss": 1.393, "step": 24150 }, { "epoch": 17.44115523465704, "grad_norm": 2.597839593887329, "learning_rate": 1.2794223826714802e-05, "loss": 1.3977, "step": 24156 }, { "epoch": 17.445487364620938, "grad_norm": 2.451107978820801, "learning_rate": 1.2772563176895308e-05, "loss": 1.4063, "step": 24162 }, { "epoch": 17.449819494584837, "grad_norm": 2.659700632095337, "learning_rate": 1.2750902527075813e-05, "loss": 1.4008, "step": 24168 }, { "epoch": 17.454151624548736, "grad_norm": 2.674116849899292, "learning_rate": 1.2729241877256318e-05, "loss": 1.4092, "step": 24174 }, { "epoch": 17.458483754512635, "grad_norm": 2.4195396900177, "learning_rate": 1.2707581227436823e-05, "loss": 1.4574, "step": 24180 }, { "epoch": 17.462815884476534, "grad_norm": 2.52720046043396, "learning_rate": 1.268592057761733e-05, "loss": 1.4359, "step": 24186 }, { "epoch": 17.467148014440433, "grad_norm": 2.282388925552368, "learning_rate": 1.2664259927797833e-05, "loss": 1.482, "step": 24192 }, { "epoch": 17.47148014440433, "grad_norm": 2.5502371788024902, "learning_rate": 1.264259927797834e-05, "loss": 1.528, "step": 24198 }, { "epoch": 17.47581227436823, "grad_norm": 2.3543570041656494, "learning_rate": 1.2620938628158846e-05, "loss": 1.4442, "step": 24204 }, { "epoch": 17.48014440433213, "grad_norm": 2.556776762008667, "learning_rate": 1.259927797833935e-05, "loss": 1.4167, "step": 24210 }, { "epoch": 17.48447653429603, "grad_norm": 2.343733549118042, "learning_rate": 1.2577617328519856e-05, "loss": 1.4564, "step": 24216 }, { "epoch": 17.488808664259928, "grad_norm": 2.372086763381958, "learning_rate": 1.2555956678700361e-05, "loss": 1.3878, "step": 24222 }, { "epoch": 17.493140794223827, "grad_norm": 2.4349844455718994, "learning_rate": 1.2534296028880868e-05, "loss": 1.3911, "step": 24228 }, { "epoch": 17.497472924187726, "grad_norm": 2.6096115112304688, "learning_rate": 1.2512635379061371e-05, "loss": 1.2784, "step": 24234 }, { "epoch": 17.501805054151625, "grad_norm": 2.519465684890747, "learning_rate": 1.2490974729241878e-05, "loss": 1.525, "step": 24240 }, { "epoch": 17.506137184115524, "grad_norm": 2.6416985988616943, "learning_rate": 1.2469314079422384e-05, "loss": 1.4437, "step": 24246 }, { "epoch": 17.510469314079423, "grad_norm": 2.6101033687591553, "learning_rate": 1.2447653429602889e-05, "loss": 1.4465, "step": 24252 }, { "epoch": 17.51480144404332, "grad_norm": 2.6192572116851807, "learning_rate": 1.2425992779783394e-05, "loss": 1.6578, "step": 24258 }, { "epoch": 17.51913357400722, "grad_norm": 2.5784831047058105, "learning_rate": 1.24043321299639e-05, "loss": 1.5331, "step": 24264 }, { "epoch": 17.52346570397112, "grad_norm": 2.4230122566223145, "learning_rate": 1.2382671480144405e-05, "loss": 1.3962, "step": 24270 }, { "epoch": 17.52779783393502, "grad_norm": 2.4095828533172607, "learning_rate": 1.236101083032491e-05, "loss": 1.4649, "step": 24276 }, { "epoch": 17.532129963898917, "grad_norm": 2.6333091259002686, "learning_rate": 1.2339350180505415e-05, "loss": 1.494, "step": 24282 }, { "epoch": 17.536462093862816, "grad_norm": 2.511115550994873, "learning_rate": 1.231768953068592e-05, "loss": 1.3787, "step": 24288 }, { "epoch": 17.540794223826715, "grad_norm": 2.405632972717285, "learning_rate": 1.2296028880866427e-05, "loss": 1.388, "step": 24294 }, { "epoch": 17.545126353790614, "grad_norm": 2.5104572772979736, "learning_rate": 1.2274368231046932e-05, "loss": 1.5068, "step": 24300 }, { "epoch": 17.549458483754513, "grad_norm": 2.4719927310943604, "learning_rate": 1.2252707581227438e-05, "loss": 1.436, "step": 24306 }, { "epoch": 17.553790613718412, "grad_norm": 2.165531873703003, "learning_rate": 1.2231046931407943e-05, "loss": 1.4473, "step": 24312 }, { "epoch": 17.55812274368231, "grad_norm": 2.100123405456543, "learning_rate": 1.2209386281588448e-05, "loss": 1.4097, "step": 24318 }, { "epoch": 17.56245487364621, "grad_norm": 2.701317310333252, "learning_rate": 1.2187725631768953e-05, "loss": 1.4949, "step": 24324 }, { "epoch": 17.56678700361011, "grad_norm": 2.419718027114868, "learning_rate": 1.2166064981949458e-05, "loss": 1.4724, "step": 24330 }, { "epoch": 17.57111913357401, "grad_norm": 2.3396341800689697, "learning_rate": 1.2144404332129965e-05, "loss": 1.3458, "step": 24336 }, { "epoch": 17.575451263537907, "grad_norm": 2.713294267654419, "learning_rate": 1.212274368231047e-05, "loss": 1.3933, "step": 24342 }, { "epoch": 17.579783393501806, "grad_norm": 2.643739700317383, "learning_rate": 1.2101083032490976e-05, "loss": 1.495, "step": 24348 }, { "epoch": 17.584115523465705, "grad_norm": 2.4248507022857666, "learning_rate": 1.2079422382671481e-05, "loss": 1.453, "step": 24354 }, { "epoch": 17.588447653429604, "grad_norm": 2.263375997543335, "learning_rate": 1.2057761732851986e-05, "loss": 1.3233, "step": 24360 }, { "epoch": 17.592779783393503, "grad_norm": 2.6506307125091553, "learning_rate": 1.2036101083032491e-05, "loss": 1.6199, "step": 24366 }, { "epoch": 17.597111913357402, "grad_norm": 2.6690492630004883, "learning_rate": 1.2014440433212996e-05, "loss": 1.3807, "step": 24372 }, { "epoch": 17.6014440433213, "grad_norm": 2.548257350921631, "learning_rate": 1.1992779783393503e-05, "loss": 1.4452, "step": 24378 }, { "epoch": 17.6057761732852, "grad_norm": 2.3869338035583496, "learning_rate": 1.1971119133574007e-05, "loss": 1.3195, "step": 24384 }, { "epoch": 17.6101083032491, "grad_norm": 2.5770299434661865, "learning_rate": 1.1949458483754514e-05, "loss": 1.3879, "step": 24390 }, { "epoch": 17.614440433212998, "grad_norm": 2.444512367248535, "learning_rate": 1.1927797833935019e-05, "loss": 1.4034, "step": 24396 }, { "epoch": 17.618772563176897, "grad_norm": 2.668013334274292, "learning_rate": 1.1906137184115524e-05, "loss": 1.568, "step": 24402 }, { "epoch": 17.623104693140796, "grad_norm": 2.5982253551483154, "learning_rate": 1.1884476534296029e-05, "loss": 1.4059, "step": 24408 }, { "epoch": 17.627436823104695, "grad_norm": 2.690084218978882, "learning_rate": 1.1862815884476534e-05, "loss": 1.4372, "step": 24414 }, { "epoch": 17.63176895306859, "grad_norm": 2.4107401371002197, "learning_rate": 1.184115523465704e-05, "loss": 1.4567, "step": 24420 }, { "epoch": 17.63610108303249, "grad_norm": 2.4820656776428223, "learning_rate": 1.1819494584837545e-05, "loss": 1.3913, "step": 24426 }, { "epoch": 17.64043321299639, "grad_norm": 2.416689157485962, "learning_rate": 1.1797833935018052e-05, "loss": 1.493, "step": 24432 }, { "epoch": 17.644765342960287, "grad_norm": 2.5561718940734863, "learning_rate": 1.1776173285198557e-05, "loss": 1.476, "step": 24438 }, { "epoch": 17.649097472924186, "grad_norm": 2.3073267936706543, "learning_rate": 1.1754512635379062e-05, "loss": 1.3259, "step": 24444 }, { "epoch": 17.653429602888085, "grad_norm": 2.5704147815704346, "learning_rate": 1.1732851985559568e-05, "loss": 1.3931, "step": 24450 }, { "epoch": 17.657761732851984, "grad_norm": 2.536163568496704, "learning_rate": 1.1711191335740072e-05, "loss": 1.4309, "step": 24456 }, { "epoch": 17.662093862815883, "grad_norm": 2.4258830547332764, "learning_rate": 1.1689530685920578e-05, "loss": 1.4099, "step": 24462 }, { "epoch": 17.666425992779782, "grad_norm": 2.6853644847869873, "learning_rate": 1.1667870036101083e-05, "loss": 1.3528, "step": 24468 }, { "epoch": 17.67075812274368, "grad_norm": 2.362367630004883, "learning_rate": 1.164620938628159e-05, "loss": 1.3758, "step": 24474 }, { "epoch": 17.67509025270758, "grad_norm": 2.447282552719116, "learning_rate": 1.1624548736462095e-05, "loss": 1.5213, "step": 24480 }, { "epoch": 17.67942238267148, "grad_norm": 2.485158681869507, "learning_rate": 1.16028880866426e-05, "loss": 1.4445, "step": 24486 }, { "epoch": 17.683754512635378, "grad_norm": 2.6010279655456543, "learning_rate": 1.1581227436823106e-05, "loss": 1.4511, "step": 24492 }, { "epoch": 17.688086642599277, "grad_norm": 2.446956157684326, "learning_rate": 1.1559566787003611e-05, "loss": 1.4418, "step": 24498 }, { "epoch": 17.692418772563176, "grad_norm": 2.2845959663391113, "learning_rate": 1.1537906137184116e-05, "loss": 1.3531, "step": 24504 }, { "epoch": 17.696750902527075, "grad_norm": 2.681814432144165, "learning_rate": 1.1516245487364621e-05, "loss": 1.4159, "step": 24510 }, { "epoch": 17.701083032490974, "grad_norm": 2.969536781311035, "learning_rate": 1.1494584837545126e-05, "loss": 1.4441, "step": 24516 }, { "epoch": 17.705415162454873, "grad_norm": 2.601161003112793, "learning_rate": 1.1472924187725632e-05, "loss": 1.45, "step": 24522 }, { "epoch": 17.709747292418772, "grad_norm": 2.3238751888275146, "learning_rate": 1.1451263537906137e-05, "loss": 1.4487, "step": 24528 }, { "epoch": 17.71407942238267, "grad_norm": 2.899921417236328, "learning_rate": 1.1429602888086644e-05, "loss": 1.5158, "step": 24534 }, { "epoch": 17.71841155234657, "grad_norm": 2.506218910217285, "learning_rate": 1.1407942238267149e-05, "loss": 1.4798, "step": 24540 }, { "epoch": 17.72274368231047, "grad_norm": 2.3380446434020996, "learning_rate": 1.1386281588447654e-05, "loss": 1.3653, "step": 24546 }, { "epoch": 17.727075812274368, "grad_norm": 2.4882798194885254, "learning_rate": 1.1364620938628159e-05, "loss": 1.4395, "step": 24552 }, { "epoch": 17.731407942238267, "grad_norm": 2.6575896739959717, "learning_rate": 1.1342960288808664e-05, "loss": 1.4509, "step": 24558 }, { "epoch": 17.735740072202166, "grad_norm": 2.7704179286956787, "learning_rate": 1.132129963898917e-05, "loss": 1.4201, "step": 24564 }, { "epoch": 17.740072202166065, "grad_norm": 2.616309642791748, "learning_rate": 1.1299638989169675e-05, "loss": 1.557, "step": 24570 }, { "epoch": 17.744404332129964, "grad_norm": 2.3493287563323975, "learning_rate": 1.1277978339350182e-05, "loss": 1.356, "step": 24576 }, { "epoch": 17.748736462093863, "grad_norm": 2.9291059970855713, "learning_rate": 1.1256317689530687e-05, "loss": 1.3509, "step": 24582 }, { "epoch": 17.753068592057762, "grad_norm": 2.6678221225738525, "learning_rate": 1.1234657039711193e-05, "loss": 1.3467, "step": 24588 }, { "epoch": 17.75740072202166, "grad_norm": 2.3615150451660156, "learning_rate": 1.1212996389891697e-05, "loss": 1.4726, "step": 24594 }, { "epoch": 17.76173285198556, "grad_norm": 2.7362966537475586, "learning_rate": 1.1191335740072201e-05, "loss": 1.5893, "step": 24600 }, { "epoch": 17.76606498194946, "grad_norm": 2.416759729385376, "learning_rate": 1.1169675090252708e-05, "loss": 1.5081, "step": 24606 }, { "epoch": 17.770397111913358, "grad_norm": 2.634786367416382, "learning_rate": 1.1148014440433213e-05, "loss": 1.4855, "step": 24612 }, { "epoch": 17.774729241877257, "grad_norm": 2.411424160003662, "learning_rate": 1.112635379061372e-05, "loss": 1.4156, "step": 24618 }, { "epoch": 17.779061371841156, "grad_norm": 2.367830753326416, "learning_rate": 1.1104693140794225e-05, "loss": 1.3979, "step": 24624 }, { "epoch": 17.783393501805055, "grad_norm": 3.1134443283081055, "learning_rate": 1.1083032490974731e-05, "loss": 1.422, "step": 24630 }, { "epoch": 17.787725631768954, "grad_norm": 2.700331926345825, "learning_rate": 1.1061371841155236e-05, "loss": 1.4385, "step": 24636 }, { "epoch": 17.792057761732853, "grad_norm": 2.843167304992676, "learning_rate": 1.103971119133574e-05, "loss": 1.4038, "step": 24642 }, { "epoch": 17.79638989169675, "grad_norm": 2.671344518661499, "learning_rate": 1.1018050541516246e-05, "loss": 1.5012, "step": 24648 }, { "epoch": 17.80072202166065, "grad_norm": 2.565699815750122, "learning_rate": 1.099638989169675e-05, "loss": 1.5231, "step": 24654 }, { "epoch": 17.80505415162455, "grad_norm": 2.301448106765747, "learning_rate": 1.0974729241877257e-05, "loss": 1.3266, "step": 24660 }, { "epoch": 17.80938628158845, "grad_norm": 2.615147590637207, "learning_rate": 1.0953068592057762e-05, "loss": 1.4447, "step": 24666 }, { "epoch": 17.813718411552347, "grad_norm": 2.407475709915161, "learning_rate": 1.0931407942238267e-05, "loss": 1.3678, "step": 24672 }, { "epoch": 17.818050541516246, "grad_norm": 2.4803273677825928, "learning_rate": 1.0909747292418774e-05, "loss": 1.3992, "step": 24678 }, { "epoch": 17.822382671480145, "grad_norm": 2.5620779991149902, "learning_rate": 1.0888086642599279e-05, "loss": 1.3673, "step": 24684 }, { "epoch": 17.826714801444044, "grad_norm": 2.655454158782959, "learning_rate": 1.0866425992779784e-05, "loss": 1.4586, "step": 24690 }, { "epoch": 17.831046931407943, "grad_norm": 2.6031112670898438, "learning_rate": 1.0844765342960289e-05, "loss": 1.4538, "step": 24696 }, { "epoch": 17.835379061371842, "grad_norm": 2.419591188430786, "learning_rate": 1.0823104693140795e-05, "loss": 1.434, "step": 24702 }, { "epoch": 17.83971119133574, "grad_norm": 2.426605224609375, "learning_rate": 1.08014440433213e-05, "loss": 1.4565, "step": 24708 }, { "epoch": 17.84404332129964, "grad_norm": 2.5691397190093994, "learning_rate": 1.0779783393501805e-05, "loss": 1.3591, "step": 24714 }, { "epoch": 17.84837545126354, "grad_norm": 2.450993299484253, "learning_rate": 1.0758122743682312e-05, "loss": 1.5347, "step": 24720 }, { "epoch": 17.85270758122744, "grad_norm": 2.5806198120117188, "learning_rate": 1.0736462093862817e-05, "loss": 1.3469, "step": 24726 }, { "epoch": 17.857039711191337, "grad_norm": 2.40669322013855, "learning_rate": 1.0714801444043322e-05, "loss": 1.376, "step": 24732 }, { "epoch": 17.861371841155236, "grad_norm": 2.4576330184936523, "learning_rate": 1.0693140794223826e-05, "loss": 1.4686, "step": 24738 }, { "epoch": 17.865703971119135, "grad_norm": 2.591560125350952, "learning_rate": 1.0671480144404333e-05, "loss": 1.4431, "step": 24744 }, { "epoch": 17.870036101083034, "grad_norm": 2.7743990421295166, "learning_rate": 1.0649819494584838e-05, "loss": 1.5492, "step": 24750 }, { "epoch": 17.87436823104693, "grad_norm": 2.68503475189209, "learning_rate": 1.0628158844765343e-05, "loss": 1.4245, "step": 24756 }, { "epoch": 17.87870036101083, "grad_norm": 2.2060511112213135, "learning_rate": 1.060649819494585e-05, "loss": 1.3958, "step": 24762 }, { "epoch": 17.883032490974728, "grad_norm": 2.4060566425323486, "learning_rate": 1.0584837545126354e-05, "loss": 1.4373, "step": 24768 }, { "epoch": 17.887364620938627, "grad_norm": 2.221595048904419, "learning_rate": 1.0563176895306861e-05, "loss": 1.4344, "step": 24774 }, { "epoch": 17.891696750902526, "grad_norm": 2.3438751697540283, "learning_rate": 1.0541516245487364e-05, "loss": 1.3788, "step": 24780 }, { "epoch": 17.896028880866425, "grad_norm": 2.300288200378418, "learning_rate": 1.051985559566787e-05, "loss": 1.4139, "step": 24786 }, { "epoch": 17.900361010830323, "grad_norm": 2.5260822772979736, "learning_rate": 1.0498194945848376e-05, "loss": 1.3881, "step": 24792 }, { "epoch": 17.904693140794222, "grad_norm": 2.359837770462036, "learning_rate": 1.047653429602888e-05, "loss": 1.3752, "step": 24798 }, { "epoch": 17.90902527075812, "grad_norm": 2.4156579971313477, "learning_rate": 1.0454873646209387e-05, "loss": 1.4306, "step": 24804 }, { "epoch": 17.91335740072202, "grad_norm": 2.2337441444396973, "learning_rate": 1.0433212996389892e-05, "loss": 1.4645, "step": 24810 }, { "epoch": 17.91768953068592, "grad_norm": 2.137143611907959, "learning_rate": 1.0411552346570399e-05, "loss": 1.3899, "step": 24816 }, { "epoch": 17.92202166064982, "grad_norm": 2.414804220199585, "learning_rate": 1.0389891696750904e-05, "loss": 1.5094, "step": 24822 }, { "epoch": 17.926353790613717, "grad_norm": 2.8557143211364746, "learning_rate": 1.0368231046931407e-05, "loss": 1.4492, "step": 24828 }, { "epoch": 17.930685920577616, "grad_norm": 2.515172004699707, "learning_rate": 1.0346570397111914e-05, "loss": 1.3649, "step": 24834 }, { "epoch": 17.935018050541515, "grad_norm": 2.5495738983154297, "learning_rate": 1.0324909747292419e-05, "loss": 1.4501, "step": 24840 }, { "epoch": 17.939350180505414, "grad_norm": 2.4743454456329346, "learning_rate": 1.0303249097472925e-05, "loss": 1.3762, "step": 24846 }, { "epoch": 17.943682310469313, "grad_norm": 2.6269664764404297, "learning_rate": 1.028158844765343e-05, "loss": 1.5528, "step": 24852 }, { "epoch": 17.948014440433212, "grad_norm": 2.4391231536865234, "learning_rate": 1.0259927797833937e-05, "loss": 1.4572, "step": 24858 }, { "epoch": 17.95234657039711, "grad_norm": 2.698840618133545, "learning_rate": 1.0238267148014442e-05, "loss": 1.476, "step": 24864 }, { "epoch": 17.95667870036101, "grad_norm": 2.404500722885132, "learning_rate": 1.0216606498194947e-05, "loss": 1.4473, "step": 24870 }, { "epoch": 17.96101083032491, "grad_norm": 2.5185463428497314, "learning_rate": 1.0194945848375451e-05, "loss": 1.3807, "step": 24876 }, { "epoch": 17.965342960288808, "grad_norm": 2.5130105018615723, "learning_rate": 1.0173285198555956e-05, "loss": 1.3989, "step": 24882 }, { "epoch": 17.969675090252707, "grad_norm": 2.520437002182007, "learning_rate": 1.0151624548736463e-05, "loss": 1.3797, "step": 24888 }, { "epoch": 17.974007220216606, "grad_norm": 2.4897072315216064, "learning_rate": 1.0129963898916968e-05, "loss": 1.5221, "step": 24894 }, { "epoch": 17.978339350180505, "grad_norm": 2.5865542888641357, "learning_rate": 1.0108303249097473e-05, "loss": 1.4996, "step": 24900 }, { "epoch": 17.982671480144404, "grad_norm": 2.380913257598877, "learning_rate": 1.008664259927798e-05, "loss": 1.3934, "step": 24906 }, { "epoch": 17.987003610108303, "grad_norm": 2.472846508026123, "learning_rate": 1.0064981949458484e-05, "loss": 1.4048, "step": 24912 }, { "epoch": 17.991335740072202, "grad_norm": 2.4516441822052, "learning_rate": 1.004332129963899e-05, "loss": 1.4108, "step": 24918 }, { "epoch": 17.9956678700361, "grad_norm": 2.5429389476776123, "learning_rate": 1.0021660649819494e-05, "loss": 1.3793, "step": 24924 }, { "epoch": 18.0, "grad_norm": 2.906454086303711, "learning_rate": 1e-05, "loss": 1.3507, "step": 24930 }, { "epoch": 18.0043321299639, "grad_norm": 2.5277040004730225, "learning_rate": 9.978339350180506e-06, "loss": 1.5804, "step": 24936 }, { "epoch": 18.008664259927798, "grad_norm": 2.4094183444976807, "learning_rate": 9.95667870036101e-06, "loss": 1.5516, "step": 24942 }, { "epoch": 18.012996389891697, "grad_norm": 2.5471229553222656, "learning_rate": 9.935018050541517e-06, "loss": 1.3642, "step": 24948 }, { "epoch": 18.017328519855596, "grad_norm": 2.6470963954925537, "learning_rate": 9.913357400722022e-06, "loss": 1.4219, "step": 24954 }, { "epoch": 18.021660649819495, "grad_norm": 2.723896026611328, "learning_rate": 9.891696750902527e-06, "loss": 1.4514, "step": 24960 }, { "epoch": 18.025992779783394, "grad_norm": 2.5858280658721924, "learning_rate": 9.870036101083032e-06, "loss": 1.4526, "step": 24966 }, { "epoch": 18.030324909747293, "grad_norm": 2.4292101860046387, "learning_rate": 9.848375451263539e-06, "loss": 1.4624, "step": 24972 }, { "epoch": 18.034657039711192, "grad_norm": 2.4702391624450684, "learning_rate": 9.826714801444044e-06, "loss": 1.4521, "step": 24978 }, { "epoch": 18.03898916967509, "grad_norm": 2.471867799758911, "learning_rate": 9.805054151624548e-06, "loss": 1.4012, "step": 24984 }, { "epoch": 18.04332129963899, "grad_norm": 2.626677989959717, "learning_rate": 9.783393501805055e-06, "loss": 1.3224, "step": 24990 }, { "epoch": 18.04765342960289, "grad_norm": 2.4553751945495605, "learning_rate": 9.76173285198556e-06, "loss": 1.3902, "step": 24996 }, { "epoch": 18.051985559566788, "grad_norm": 2.58267879486084, "learning_rate": 9.740072202166067e-06, "loss": 1.4715, "step": 25002 }, { "epoch": 18.056317689530687, "grad_norm": 2.275757312774658, "learning_rate": 9.718411552346572e-06, "loss": 1.2954, "step": 25008 }, { "epoch": 18.060649819494586, "grad_norm": 2.454763889312744, "learning_rate": 9.696750902527076e-06, "loss": 1.4258, "step": 25014 }, { "epoch": 18.064981949458485, "grad_norm": 2.4178318977355957, "learning_rate": 9.675090252707581e-06, "loss": 1.2976, "step": 25020 }, { "epoch": 18.069314079422384, "grad_norm": 2.317509651184082, "learning_rate": 9.653429602888086e-06, "loss": 1.4413, "step": 25026 }, { "epoch": 18.073646209386283, "grad_norm": 2.6978750228881836, "learning_rate": 9.631768953068593e-06, "loss": 1.3543, "step": 25032 }, { "epoch": 18.07797833935018, "grad_norm": 2.6031436920166016, "learning_rate": 9.610108303249098e-06, "loss": 1.5698, "step": 25038 }, { "epoch": 18.08231046931408, "grad_norm": 2.5061745643615723, "learning_rate": 9.588447653429604e-06, "loss": 1.508, "step": 25044 }, { "epoch": 18.08664259927798, "grad_norm": 2.6833760738372803, "learning_rate": 9.56678700361011e-06, "loss": 1.435, "step": 25050 }, { "epoch": 18.09097472924188, "grad_norm": 2.3499927520751953, "learning_rate": 9.545126353790614e-06, "loss": 1.3329, "step": 25056 }, { "epoch": 18.095306859205778, "grad_norm": 2.45259165763855, "learning_rate": 9.52346570397112e-06, "loss": 1.4402, "step": 25062 }, { "epoch": 18.099638989169677, "grad_norm": 2.5327789783477783, "learning_rate": 9.501805054151624e-06, "loss": 1.4193, "step": 25068 }, { "epoch": 18.103971119133575, "grad_norm": 2.6370432376861572, "learning_rate": 9.48014440433213e-06, "loss": 1.2586, "step": 25074 }, { "epoch": 18.108303249097474, "grad_norm": 2.844822883605957, "learning_rate": 9.458483754512636e-06, "loss": 1.4641, "step": 25080 }, { "epoch": 18.112635379061373, "grad_norm": 2.51212215423584, "learning_rate": 9.436823104693142e-06, "loss": 1.3811, "step": 25086 }, { "epoch": 18.116967509025272, "grad_norm": 2.2442305088043213, "learning_rate": 9.415162454873647e-06, "loss": 1.4519, "step": 25092 }, { "epoch": 18.12129963898917, "grad_norm": 2.57635498046875, "learning_rate": 9.393501805054152e-06, "loss": 1.4425, "step": 25098 }, { "epoch": 18.12563176895307, "grad_norm": 2.269892930984497, "learning_rate": 9.371841155234657e-06, "loss": 1.2661, "step": 25104 }, { "epoch": 18.129963898916966, "grad_norm": 2.558344602584839, "learning_rate": 9.350180505415162e-06, "loss": 1.4127, "step": 25110 }, { "epoch": 18.134296028880865, "grad_norm": 2.3130810260772705, "learning_rate": 9.328519855595669e-06, "loss": 1.3878, "step": 25116 }, { "epoch": 18.138628158844764, "grad_norm": 2.649866819381714, "learning_rate": 9.306859205776173e-06, "loss": 1.3794, "step": 25122 }, { "epoch": 18.142960288808663, "grad_norm": 2.7611215114593506, "learning_rate": 9.28519855595668e-06, "loss": 1.4616, "step": 25128 }, { "epoch": 18.14729241877256, "grad_norm": 2.6186959743499756, "learning_rate": 9.263537906137185e-06, "loss": 1.5592, "step": 25134 }, { "epoch": 18.15162454873646, "grad_norm": 2.4330239295959473, "learning_rate": 9.24187725631769e-06, "loss": 1.419, "step": 25140 }, { "epoch": 18.15595667870036, "grad_norm": 2.4501123428344727, "learning_rate": 9.220216606498195e-06, "loss": 1.3869, "step": 25146 }, { "epoch": 18.16028880866426, "grad_norm": 2.7113375663757324, "learning_rate": 9.1985559566787e-06, "loss": 1.441, "step": 25152 }, { "epoch": 18.164620938628158, "grad_norm": 2.566588878631592, "learning_rate": 9.176895306859206e-06, "loss": 1.5429, "step": 25158 }, { "epoch": 18.168953068592057, "grad_norm": 2.8635799884796143, "learning_rate": 9.155234657039711e-06, "loss": 1.3675, "step": 25164 }, { "epoch": 18.173285198555956, "grad_norm": 2.455249309539795, "learning_rate": 9.133574007220216e-06, "loss": 1.4098, "step": 25170 }, { "epoch": 18.177617328519855, "grad_norm": 2.421525716781616, "learning_rate": 9.111913357400723e-06, "loss": 1.4486, "step": 25176 }, { "epoch": 18.181949458483754, "grad_norm": 2.5230355262756348, "learning_rate": 9.090252707581228e-06, "loss": 1.5061, "step": 25182 }, { "epoch": 18.186281588447653, "grad_norm": 2.4301769733428955, "learning_rate": 9.068592057761734e-06, "loss": 1.4103, "step": 25188 }, { "epoch": 18.19061371841155, "grad_norm": 2.19242525100708, "learning_rate": 9.04693140794224e-06, "loss": 1.341, "step": 25194 }, { "epoch": 18.19494584837545, "grad_norm": 2.4220283031463623, "learning_rate": 9.025270758122744e-06, "loss": 1.3997, "step": 25200 }, { "epoch": 18.19927797833935, "grad_norm": 2.59395694732666, "learning_rate": 9.003610108303249e-06, "loss": 1.4407, "step": 25206 }, { "epoch": 18.20361010830325, "grad_norm": 2.3178493976593018, "learning_rate": 8.981949458483754e-06, "loss": 1.4562, "step": 25212 }, { "epoch": 18.207942238267147, "grad_norm": 2.2953622341156006, "learning_rate": 8.96028880866426e-06, "loss": 1.2774, "step": 25218 }, { "epoch": 18.212274368231046, "grad_norm": 2.6727874279022217, "learning_rate": 8.938628158844766e-06, "loss": 1.5216, "step": 25224 }, { "epoch": 18.216606498194945, "grad_norm": 2.739229202270508, "learning_rate": 8.916967509025272e-06, "loss": 1.4298, "step": 25230 }, { "epoch": 18.220938628158844, "grad_norm": 2.363389492034912, "learning_rate": 8.895306859205777e-06, "loss": 1.4231, "step": 25236 }, { "epoch": 18.225270758122743, "grad_norm": 2.5746190547943115, "learning_rate": 8.873646209386282e-06, "loss": 1.5032, "step": 25242 }, { "epoch": 18.229602888086642, "grad_norm": 2.263359546661377, "learning_rate": 8.851985559566787e-06, "loss": 1.4368, "step": 25248 }, { "epoch": 18.23393501805054, "grad_norm": 2.488114595413208, "learning_rate": 8.830324909747292e-06, "loss": 1.3697, "step": 25254 }, { "epoch": 18.23826714801444, "grad_norm": 2.1320183277130127, "learning_rate": 8.808664259927798e-06, "loss": 1.3972, "step": 25260 }, { "epoch": 18.24259927797834, "grad_norm": 2.522862672805786, "learning_rate": 8.787003610108303e-06, "loss": 1.39, "step": 25266 }, { "epoch": 18.246931407942238, "grad_norm": 2.665755271911621, "learning_rate": 8.76534296028881e-06, "loss": 1.421, "step": 25272 }, { "epoch": 18.251263537906137, "grad_norm": 2.581174850463867, "learning_rate": 8.743682310469315e-06, "loss": 1.4421, "step": 25278 }, { "epoch": 18.255595667870036, "grad_norm": 2.457050323486328, "learning_rate": 8.72202166064982e-06, "loss": 1.4959, "step": 25284 }, { "epoch": 18.259927797833935, "grad_norm": 2.6380631923675537, "learning_rate": 8.700361010830325e-06, "loss": 1.4673, "step": 25290 }, { "epoch": 18.264259927797834, "grad_norm": 2.2918167114257812, "learning_rate": 8.67870036101083e-06, "loss": 1.2555, "step": 25296 }, { "epoch": 18.268592057761733, "grad_norm": 2.667854070663452, "learning_rate": 8.657039711191336e-06, "loss": 1.3689, "step": 25302 }, { "epoch": 18.272924187725632, "grad_norm": 2.5164520740509033, "learning_rate": 8.635379061371841e-06, "loss": 1.3935, "step": 25308 }, { "epoch": 18.27725631768953, "grad_norm": 2.4088337421417236, "learning_rate": 8.613718411552348e-06, "loss": 1.4123, "step": 25314 }, { "epoch": 18.28158844765343, "grad_norm": 2.6398489475250244, "learning_rate": 8.592057761732853e-06, "loss": 1.4355, "step": 25320 }, { "epoch": 18.28592057761733, "grad_norm": 2.859806537628174, "learning_rate": 8.570397111913358e-06, "loss": 1.5735, "step": 25326 }, { "epoch": 18.290252707581228, "grad_norm": 2.8486685752868652, "learning_rate": 8.548736462093863e-06, "loss": 1.5116, "step": 25332 }, { "epoch": 18.294584837545127, "grad_norm": 2.4447803497314453, "learning_rate": 8.527075812274368e-06, "loss": 1.4803, "step": 25338 }, { "epoch": 18.298916967509026, "grad_norm": 2.4816088676452637, "learning_rate": 8.505415162454874e-06, "loss": 1.4471, "step": 25344 }, { "epoch": 18.303249097472925, "grad_norm": 2.746464490890503, "learning_rate": 8.483754512635379e-06, "loss": 1.389, "step": 25350 }, { "epoch": 18.307581227436824, "grad_norm": 2.557114839553833, "learning_rate": 8.462093862815886e-06, "loss": 1.4354, "step": 25356 }, { "epoch": 18.311913357400723, "grad_norm": 2.5765645503997803, "learning_rate": 8.44043321299639e-06, "loss": 1.5177, "step": 25362 }, { "epoch": 18.316245487364622, "grad_norm": 2.582563638687134, "learning_rate": 8.418772563176895e-06, "loss": 1.4453, "step": 25368 }, { "epoch": 18.32057761732852, "grad_norm": 2.6340622901916504, "learning_rate": 8.397111913357402e-06, "loss": 1.5659, "step": 25374 }, { "epoch": 18.32490974729242, "grad_norm": 2.282095193862915, "learning_rate": 8.375451263537907e-06, "loss": 1.4066, "step": 25380 }, { "epoch": 18.32924187725632, "grad_norm": 2.34488844871521, "learning_rate": 8.353790613718412e-06, "loss": 1.4609, "step": 25386 }, { "epoch": 18.333574007220218, "grad_norm": 2.7516732215881348, "learning_rate": 8.332129963898917e-06, "loss": 1.4573, "step": 25392 }, { "epoch": 18.337906137184117, "grad_norm": 2.632214307785034, "learning_rate": 8.310469314079423e-06, "loss": 1.4596, "step": 25398 }, { "epoch": 18.342238267148016, "grad_norm": 2.662874698638916, "learning_rate": 8.288808664259928e-06, "loss": 1.4954, "step": 25404 }, { "epoch": 18.346570397111915, "grad_norm": 2.3402302265167236, "learning_rate": 8.267148014440433e-06, "loss": 1.4093, "step": 25410 }, { "epoch": 18.350902527075814, "grad_norm": 2.7105140686035156, "learning_rate": 8.24548736462094e-06, "loss": 1.4564, "step": 25416 }, { "epoch": 18.355234657039713, "grad_norm": 2.7485926151275635, "learning_rate": 8.223826714801445e-06, "loss": 1.4125, "step": 25422 }, { "epoch": 18.35956678700361, "grad_norm": 2.4180386066436768, "learning_rate": 8.20216606498195e-06, "loss": 1.3888, "step": 25428 }, { "epoch": 18.36389891696751, "grad_norm": 2.345008611679077, "learning_rate": 8.180505415162455e-06, "loss": 1.4184, "step": 25434 }, { "epoch": 18.36823104693141, "grad_norm": 2.5889105796813965, "learning_rate": 8.15884476534296e-06, "loss": 1.4303, "step": 25440 }, { "epoch": 18.37256317689531, "grad_norm": 2.630155563354492, "learning_rate": 8.137184115523466e-06, "loss": 1.3666, "step": 25446 }, { "epoch": 18.376895306859204, "grad_norm": 2.6034529209136963, "learning_rate": 8.115523465703971e-06, "loss": 1.5126, "step": 25452 }, { "epoch": 18.381227436823103, "grad_norm": 2.278834581375122, "learning_rate": 8.093862815884478e-06, "loss": 1.3929, "step": 25458 }, { "epoch": 18.385559566787002, "grad_norm": 2.4330008029937744, "learning_rate": 8.072202166064983e-06, "loss": 1.3333, "step": 25464 }, { "epoch": 18.3898916967509, "grad_norm": 2.434889554977417, "learning_rate": 8.050541516245488e-06, "loss": 1.3506, "step": 25470 }, { "epoch": 18.3942238267148, "grad_norm": 2.6452713012695312, "learning_rate": 8.028880866425993e-06, "loss": 1.4576, "step": 25476 }, { "epoch": 18.3985559566787, "grad_norm": 2.4605820178985596, "learning_rate": 8.007220216606497e-06, "loss": 1.4564, "step": 25482 }, { "epoch": 18.402888086642598, "grad_norm": 2.8054983615875244, "learning_rate": 7.985559566787004e-06, "loss": 1.3813, "step": 25488 }, { "epoch": 18.407220216606497, "grad_norm": 2.4830033779144287, "learning_rate": 7.963898916967509e-06, "loss": 1.453, "step": 25494 }, { "epoch": 18.411552346570396, "grad_norm": 2.3683512210845947, "learning_rate": 7.942238267148016e-06, "loss": 1.3642, "step": 25500 }, { "epoch": 18.415884476534295, "grad_norm": 2.724029779434204, "learning_rate": 7.92057761732852e-06, "loss": 1.4912, "step": 25506 }, { "epoch": 18.420216606498194, "grad_norm": 2.3643975257873535, "learning_rate": 7.898916967509027e-06, "loss": 1.5156, "step": 25512 }, { "epoch": 18.424548736462093, "grad_norm": 2.734856605529785, "learning_rate": 7.87725631768953e-06, "loss": 1.4844, "step": 25518 }, { "epoch": 18.42888086642599, "grad_norm": 2.514340400695801, "learning_rate": 7.855595667870035e-06, "loss": 1.5182, "step": 25524 }, { "epoch": 18.43321299638989, "grad_norm": 2.495861768722534, "learning_rate": 7.833935018050542e-06, "loss": 1.513, "step": 25530 }, { "epoch": 18.43754512635379, "grad_norm": 2.6837987899780273, "learning_rate": 7.812274368231047e-06, "loss": 1.4532, "step": 25536 }, { "epoch": 18.44187725631769, "grad_norm": 2.4121668338775635, "learning_rate": 7.790613718411553e-06, "loss": 1.4255, "step": 25542 }, { "epoch": 18.446209386281588, "grad_norm": 3.0792717933654785, "learning_rate": 7.768953068592058e-06, "loss": 1.4855, "step": 25548 }, { "epoch": 18.450541516245487, "grad_norm": 2.263551712036133, "learning_rate": 7.747292418772563e-06, "loss": 1.3997, "step": 25554 }, { "epoch": 18.454873646209386, "grad_norm": 2.4595577716827393, "learning_rate": 7.72563176895307e-06, "loss": 1.3507, "step": 25560 }, { "epoch": 18.459205776173285, "grad_norm": 2.6040632724761963, "learning_rate": 7.703971119133575e-06, "loss": 1.4843, "step": 25566 }, { "epoch": 18.463537906137184, "grad_norm": 2.5293450355529785, "learning_rate": 7.68231046931408e-06, "loss": 1.4533, "step": 25572 }, { "epoch": 18.467870036101083, "grad_norm": 2.595524311065674, "learning_rate": 7.660649819494585e-06, "loss": 1.3701, "step": 25578 }, { "epoch": 18.47220216606498, "grad_norm": 2.4639928340911865, "learning_rate": 7.638989169675091e-06, "loss": 1.2953, "step": 25584 }, { "epoch": 18.47653429602888, "grad_norm": 2.835838556289673, "learning_rate": 7.617328519855596e-06, "loss": 1.4473, "step": 25590 }, { "epoch": 18.48086642599278, "grad_norm": 2.505326747894287, "learning_rate": 7.595667870036101e-06, "loss": 1.4252, "step": 25596 }, { "epoch": 18.48519855595668, "grad_norm": 2.500735282897949, "learning_rate": 7.574007220216607e-06, "loss": 1.3765, "step": 25602 }, { "epoch": 18.489530685920577, "grad_norm": 2.452531099319458, "learning_rate": 7.552346570397112e-06, "loss": 1.3806, "step": 25608 }, { "epoch": 18.493862815884476, "grad_norm": 2.4538471698760986, "learning_rate": 7.530685920577618e-06, "loss": 1.409, "step": 25614 }, { "epoch": 18.498194945848375, "grad_norm": 2.617114543914795, "learning_rate": 7.509025270758123e-06, "loss": 1.4035, "step": 25620 }, { "epoch": 18.502527075812274, "grad_norm": 2.7706892490386963, "learning_rate": 7.487364620938629e-06, "loss": 1.335, "step": 25626 }, { "epoch": 18.506859205776173, "grad_norm": 2.4537551403045654, "learning_rate": 7.465703971119134e-06, "loss": 1.3239, "step": 25632 }, { "epoch": 18.511191335740072, "grad_norm": 2.622140884399414, "learning_rate": 7.444043321299639e-06, "loss": 1.4544, "step": 25638 }, { "epoch": 18.51552346570397, "grad_norm": 2.5863196849823, "learning_rate": 7.422382671480145e-06, "loss": 1.3935, "step": 25644 }, { "epoch": 18.51985559566787, "grad_norm": 2.4188992977142334, "learning_rate": 7.4007220216606496e-06, "loss": 1.3852, "step": 25650 }, { "epoch": 18.52418772563177, "grad_norm": 2.6401660442352295, "learning_rate": 7.379061371841156e-06, "loss": 1.374, "step": 25656 }, { "epoch": 18.52851985559567, "grad_norm": 2.7592341899871826, "learning_rate": 7.357400722021661e-06, "loss": 1.435, "step": 25662 }, { "epoch": 18.532851985559567, "grad_norm": 2.6172001361846924, "learning_rate": 7.335740072202166e-06, "loss": 1.3837, "step": 25668 }, { "epoch": 18.537184115523466, "grad_norm": 2.5642709732055664, "learning_rate": 7.314079422382672e-06, "loss": 1.2857, "step": 25674 }, { "epoch": 18.541516245487365, "grad_norm": 2.543064832687378, "learning_rate": 7.292418772563177e-06, "loss": 1.476, "step": 25680 }, { "epoch": 18.545848375451264, "grad_norm": 2.3022427558898926, "learning_rate": 7.270758122743683e-06, "loss": 1.4713, "step": 25686 }, { "epoch": 18.550180505415163, "grad_norm": 2.4300971031188965, "learning_rate": 7.249097472924188e-06, "loss": 1.4767, "step": 25692 }, { "epoch": 18.554512635379062, "grad_norm": 2.622804641723633, "learning_rate": 7.227436823104694e-06, "loss": 1.3954, "step": 25698 }, { "epoch": 18.55884476534296, "grad_norm": 2.3980512619018555, "learning_rate": 7.205776173285199e-06, "loss": 1.4235, "step": 25704 }, { "epoch": 18.56317689530686, "grad_norm": 2.5318496227264404, "learning_rate": 7.184115523465704e-06, "loss": 1.47, "step": 25710 }, { "epoch": 18.56750902527076, "grad_norm": 2.526195764541626, "learning_rate": 7.16245487364621e-06, "loss": 1.5145, "step": 25716 }, { "epoch": 18.571841155234658, "grad_norm": 2.823336362838745, "learning_rate": 7.1407942238267145e-06, "loss": 1.5293, "step": 25722 }, { "epoch": 18.576173285198557, "grad_norm": 2.1604864597320557, "learning_rate": 7.119133574007221e-06, "loss": 1.414, "step": 25728 }, { "epoch": 18.580505415162456, "grad_norm": 2.338991403579712, "learning_rate": 7.097472924187726e-06, "loss": 1.4266, "step": 25734 }, { "epoch": 18.584837545126355, "grad_norm": 2.6047961711883545, "learning_rate": 7.075812274368232e-06, "loss": 1.5564, "step": 25740 }, { "epoch": 18.589169675090254, "grad_norm": 2.3822877407073975, "learning_rate": 7.054151624548737e-06, "loss": 1.5247, "step": 25746 }, { "epoch": 18.593501805054153, "grad_norm": 2.6458706855773926, "learning_rate": 7.032490974729242e-06, "loss": 1.4668, "step": 25752 }, { "epoch": 18.597833935018052, "grad_norm": 2.421924591064453, "learning_rate": 7.010830324909748e-06, "loss": 1.3699, "step": 25758 }, { "epoch": 18.60216606498195, "grad_norm": 2.40505313873291, "learning_rate": 6.989169675090252e-06, "loss": 1.4034, "step": 25764 }, { "epoch": 18.60649819494585, "grad_norm": 2.533060312271118, "learning_rate": 6.967509025270759e-06, "loss": 1.4681, "step": 25770 }, { "epoch": 18.61083032490975, "grad_norm": 2.3343849182128906, "learning_rate": 6.945848375451264e-06, "loss": 1.3532, "step": 25776 }, { "epoch": 18.615162454873648, "grad_norm": 2.534320831298828, "learning_rate": 6.924187725631769e-06, "loss": 1.5437, "step": 25782 }, { "epoch": 18.619494584837547, "grad_norm": 2.587111711502075, "learning_rate": 6.9025270758122746e-06, "loss": 1.3592, "step": 25788 }, { "epoch": 18.623826714801446, "grad_norm": 2.5938796997070312, "learning_rate": 6.8808664259927795e-06, "loss": 1.4433, "step": 25794 }, { "epoch": 18.628158844765345, "grad_norm": 2.4212846755981445, "learning_rate": 6.859205776173286e-06, "loss": 1.385, "step": 25800 }, { "epoch": 18.63249097472924, "grad_norm": 2.5153894424438477, "learning_rate": 6.837545126353791e-06, "loss": 1.3128, "step": 25806 }, { "epoch": 18.63682310469314, "grad_norm": 2.516526222229004, "learning_rate": 6.815884476534297e-06, "loss": 1.4785, "step": 25812 }, { "epoch": 18.641155234657038, "grad_norm": 2.451327085494995, "learning_rate": 6.794223826714802e-06, "loss": 1.3684, "step": 25818 }, { "epoch": 18.645487364620937, "grad_norm": 2.3042256832122803, "learning_rate": 6.772563176895307e-06, "loss": 1.4178, "step": 25824 }, { "epoch": 18.649819494584836, "grad_norm": 2.6007509231567383, "learning_rate": 6.750902527075812e-06, "loss": 1.3491, "step": 25830 }, { "epoch": 18.654151624548735, "grad_norm": 2.5642452239990234, "learning_rate": 6.729241877256317e-06, "loss": 1.3604, "step": 25836 }, { "epoch": 18.658483754512634, "grad_norm": 2.5808424949645996, "learning_rate": 6.707581227436824e-06, "loss": 1.3972, "step": 25842 }, { "epoch": 18.662815884476533, "grad_norm": 2.6111204624176025, "learning_rate": 6.685920577617329e-06, "loss": 1.4189, "step": 25848 }, { "epoch": 18.667148014440432, "grad_norm": 2.467297315597534, "learning_rate": 6.664259927797835e-06, "loss": 1.4258, "step": 25854 }, { "epoch": 18.67148014440433, "grad_norm": 2.6957342624664307, "learning_rate": 6.6425992779783395e-06, "loss": 1.5529, "step": 25860 }, { "epoch": 18.67581227436823, "grad_norm": 2.714125394821167, "learning_rate": 6.6209386281588445e-06, "loss": 1.5508, "step": 25866 }, { "epoch": 18.68014440433213, "grad_norm": 2.7753748893737793, "learning_rate": 6.599277978339351e-06, "loss": 1.4563, "step": 25872 }, { "epoch": 18.684476534296028, "grad_norm": 2.4728777408599854, "learning_rate": 6.577617328519856e-06, "loss": 1.505, "step": 25878 }, { "epoch": 18.688808664259927, "grad_norm": 2.5424551963806152, "learning_rate": 6.555956678700362e-06, "loss": 1.4166, "step": 25884 }, { "epoch": 18.693140794223826, "grad_norm": 2.486319065093994, "learning_rate": 6.534296028880867e-06, "loss": 1.3994, "step": 25890 }, { "epoch": 18.697472924187725, "grad_norm": 2.3834550380706787, "learning_rate": 6.5126353790613724e-06, "loss": 1.3397, "step": 25896 }, { "epoch": 18.701805054151624, "grad_norm": 2.390401601791382, "learning_rate": 6.490974729241877e-06, "loss": 1.4405, "step": 25902 }, { "epoch": 18.706137184115523, "grad_norm": 2.7920989990234375, "learning_rate": 6.469314079422382e-06, "loss": 1.3516, "step": 25908 }, { "epoch": 18.710469314079422, "grad_norm": 2.8405086994171143, "learning_rate": 6.447653429602889e-06, "loss": 1.362, "step": 25914 }, { "epoch": 18.71480144404332, "grad_norm": 2.3563308715820312, "learning_rate": 6.425992779783394e-06, "loss": 1.4672, "step": 25920 }, { "epoch": 18.71913357400722, "grad_norm": 2.4227523803710938, "learning_rate": 6.4043321299638996e-06, "loss": 1.454, "step": 25926 }, { "epoch": 18.72346570397112, "grad_norm": 2.442182779312134, "learning_rate": 6.3826714801444045e-06, "loss": 1.4185, "step": 25932 }, { "epoch": 18.727797833935018, "grad_norm": 2.6726841926574707, "learning_rate": 6.361010830324909e-06, "loss": 1.461, "step": 25938 }, { "epoch": 18.732129963898917, "grad_norm": 2.404085874557495, "learning_rate": 6.339350180505416e-06, "loss": 1.4946, "step": 25944 }, { "epoch": 18.736462093862816, "grad_norm": 2.42073655128479, "learning_rate": 6.31768953068592e-06, "loss": 1.5104, "step": 25950 }, { "epoch": 18.740794223826715, "grad_norm": 2.55379319190979, "learning_rate": 6.296028880866427e-06, "loss": 1.3896, "step": 25956 }, { "epoch": 18.745126353790614, "grad_norm": 2.959498882293701, "learning_rate": 6.274368231046932e-06, "loss": 1.4219, "step": 25962 }, { "epoch": 18.749458483754513, "grad_norm": 2.6135993003845215, "learning_rate": 6.252707581227437e-06, "loss": 1.4361, "step": 25968 }, { "epoch": 18.75379061371841, "grad_norm": 2.2395131587982178, "learning_rate": 6.231046931407942e-06, "loss": 1.4057, "step": 25974 }, { "epoch": 18.75812274368231, "grad_norm": 2.5065739154815674, "learning_rate": 6.209386281588448e-06, "loss": 1.3478, "step": 25980 }, { "epoch": 18.76245487364621, "grad_norm": 2.7798080444335938, "learning_rate": 6.187725631768954e-06, "loss": 1.4834, "step": 25986 }, { "epoch": 18.76678700361011, "grad_norm": 2.3710777759552, "learning_rate": 6.166064981949459e-06, "loss": 1.4433, "step": 25992 }, { "epoch": 18.771119133574008, "grad_norm": 2.63191556930542, "learning_rate": 6.144404332129964e-06, "loss": 1.4258, "step": 25998 }, { "epoch": 18.775451263537906, "grad_norm": 2.6367692947387695, "learning_rate": 6.1227436823104695e-06, "loss": 1.4865, "step": 26004 }, { "epoch": 18.779783393501805, "grad_norm": 2.3158130645751953, "learning_rate": 6.101083032490975e-06, "loss": 1.4411, "step": 26010 }, { "epoch": 18.784115523465704, "grad_norm": 2.2477455139160156, "learning_rate": 6.07942238267148e-06, "loss": 1.3986, "step": 26016 }, { "epoch": 18.788447653429603, "grad_norm": 2.269463062286377, "learning_rate": 6.057761732851986e-06, "loss": 1.3615, "step": 26022 }, { "epoch": 18.792779783393502, "grad_norm": 2.857316493988037, "learning_rate": 6.036101083032492e-06, "loss": 1.5069, "step": 26028 }, { "epoch": 18.7971119133574, "grad_norm": 2.423489809036255, "learning_rate": 6.014440433212997e-06, "loss": 1.4173, "step": 26034 }, { "epoch": 18.8014440433213, "grad_norm": 2.847938299179077, "learning_rate": 5.9927797833935015e-06, "loss": 1.4264, "step": 26040 }, { "epoch": 18.8057761732852, "grad_norm": 2.3160452842712402, "learning_rate": 5.971119133574007e-06, "loss": 1.3396, "step": 26046 }, { "epoch": 18.8101083032491, "grad_norm": 2.568878412246704, "learning_rate": 5.949458483754513e-06, "loss": 1.404, "step": 26052 }, { "epoch": 18.814440433212997, "grad_norm": 2.714308977127075, "learning_rate": 5.927797833935019e-06, "loss": 1.3505, "step": 26058 }, { "epoch": 18.818772563176896, "grad_norm": 2.7437565326690674, "learning_rate": 5.906137184115524e-06, "loss": 1.4376, "step": 26064 }, { "epoch": 18.823104693140795, "grad_norm": 2.3828189373016357, "learning_rate": 5.884476534296029e-06, "loss": 1.2946, "step": 26070 }, { "epoch": 18.827436823104694, "grad_norm": 2.7261221408843994, "learning_rate": 5.862815884476534e-06, "loss": 1.3826, "step": 26076 }, { "epoch": 18.831768953068593, "grad_norm": 2.4781601428985596, "learning_rate": 5.84115523465704e-06, "loss": 1.3823, "step": 26082 }, { "epoch": 18.836101083032492, "grad_norm": 2.5709033012390137, "learning_rate": 5.819494584837545e-06, "loss": 1.4365, "step": 26088 }, { "epoch": 18.84043321299639, "grad_norm": 2.4698450565338135, "learning_rate": 5.797833935018051e-06, "loss": 1.4308, "step": 26094 }, { "epoch": 18.84476534296029, "grad_norm": 2.450120687484741, "learning_rate": 5.776173285198557e-06, "loss": 1.4374, "step": 26100 }, { "epoch": 18.84909747292419, "grad_norm": 2.8115227222442627, "learning_rate": 5.7545126353790616e-06, "loss": 1.3866, "step": 26106 }, { "epoch": 18.853429602888088, "grad_norm": 2.516477584838867, "learning_rate": 5.7328519855595665e-06, "loss": 1.3848, "step": 26112 }, { "epoch": 18.857761732851987, "grad_norm": 2.430712938308716, "learning_rate": 5.711191335740072e-06, "loss": 1.4297, "step": 26118 }, { "epoch": 18.862093862815886, "grad_norm": 2.2993967533111572, "learning_rate": 5.689530685920578e-06, "loss": 1.459, "step": 26124 }, { "epoch": 18.866425992779785, "grad_norm": 2.637568473815918, "learning_rate": 5.667870036101084e-06, "loss": 1.3631, "step": 26130 }, { "epoch": 18.870758122743684, "grad_norm": 2.5679028034210205, "learning_rate": 5.646209386281589e-06, "loss": 1.425, "step": 26136 }, { "epoch": 18.87509025270758, "grad_norm": 2.4209020137786865, "learning_rate": 5.6245487364620945e-06, "loss": 1.5026, "step": 26142 }, { "epoch": 18.87942238267148, "grad_norm": 2.6507515907287598, "learning_rate": 5.602888086642599e-06, "loss": 1.4603, "step": 26148 }, { "epoch": 18.883754512635377, "grad_norm": 2.5173587799072266, "learning_rate": 5.581227436823105e-06, "loss": 1.444, "step": 26154 }, { "epoch": 18.888086642599276, "grad_norm": 2.57255220413208, "learning_rate": 5.55956678700361e-06, "loss": 1.5246, "step": 26160 }, { "epoch": 18.892418772563175, "grad_norm": 2.5693416595458984, "learning_rate": 5.537906137184116e-06, "loss": 1.4643, "step": 26166 }, { "epoch": 18.896750902527074, "grad_norm": 2.5053489208221436, "learning_rate": 5.516245487364622e-06, "loss": 1.4161, "step": 26172 }, { "epoch": 18.901083032490973, "grad_norm": 2.7898833751678467, "learning_rate": 5.4945848375451265e-06, "loss": 1.4704, "step": 26178 }, { "epoch": 18.905415162454872, "grad_norm": 2.531447172164917, "learning_rate": 5.472924187725632e-06, "loss": 1.4272, "step": 26184 }, { "epoch": 18.90974729241877, "grad_norm": 2.5278000831604004, "learning_rate": 5.451263537906137e-06, "loss": 1.4279, "step": 26190 }, { "epoch": 18.91407942238267, "grad_norm": 2.412055730819702, "learning_rate": 5.429602888086643e-06, "loss": 1.4472, "step": 26196 }, { "epoch": 18.91841155234657, "grad_norm": 2.2881686687469482, "learning_rate": 5.407942238267148e-06, "loss": 1.368, "step": 26202 }, { "epoch": 18.922743682310468, "grad_norm": 2.6106719970703125, "learning_rate": 5.386281588447654e-06, "loss": 1.5105, "step": 26208 }, { "epoch": 18.927075812274367, "grad_norm": 2.4438135623931885, "learning_rate": 5.364620938628159e-06, "loss": 1.4374, "step": 26214 }, { "epoch": 18.931407942238266, "grad_norm": 2.57362699508667, "learning_rate": 5.342960288808665e-06, "loss": 1.504, "step": 26220 }, { "epoch": 18.935740072202165, "grad_norm": 2.759056806564331, "learning_rate": 5.321299638989169e-06, "loss": 1.3931, "step": 26226 }, { "epoch": 18.940072202166064, "grad_norm": 2.3733811378479004, "learning_rate": 5.299638989169675e-06, "loss": 1.4587, "step": 26232 }, { "epoch": 18.944404332129963, "grad_norm": 2.311753749847412, "learning_rate": 5.277978339350181e-06, "loss": 1.4811, "step": 26238 }, { "epoch": 18.948736462093862, "grad_norm": 2.5368638038635254, "learning_rate": 5.2563176895306866e-06, "loss": 1.3729, "step": 26244 }, { "epoch": 18.95306859205776, "grad_norm": 2.3048622608184814, "learning_rate": 5.2346570397111915e-06, "loss": 1.4428, "step": 26250 }, { "epoch": 18.95740072202166, "grad_norm": 2.4964938163757324, "learning_rate": 5.212996389891697e-06, "loss": 1.3917, "step": 26256 }, { "epoch": 18.96173285198556, "grad_norm": 2.3953073024749756, "learning_rate": 5.191335740072202e-06, "loss": 1.3986, "step": 26262 }, { "epoch": 18.966064981949458, "grad_norm": 2.5621039867401123, "learning_rate": 5.169675090252708e-06, "loss": 1.4925, "step": 26268 }, { "epoch": 18.970397111913357, "grad_norm": 2.8854527473449707, "learning_rate": 5.148014440433213e-06, "loss": 1.2769, "step": 26274 }, { "epoch": 18.974729241877256, "grad_norm": 2.3360116481781006, "learning_rate": 5.126353790613719e-06, "loss": 1.4047, "step": 26280 }, { "epoch": 18.979061371841155, "grad_norm": 2.683842897415161, "learning_rate": 5.104693140794224e-06, "loss": 1.4945, "step": 26286 }, { "epoch": 18.983393501805054, "grad_norm": 2.5305769443511963, "learning_rate": 5.083032490974729e-06, "loss": 1.443, "step": 26292 }, { "epoch": 18.987725631768953, "grad_norm": 2.30605149269104, "learning_rate": 5.061371841155235e-06, "loss": 1.3832, "step": 26298 }, { "epoch": 18.992057761732852, "grad_norm": 2.6903340816497803, "learning_rate": 5.03971119133574e-06, "loss": 1.5158, "step": 26304 }, { "epoch": 18.99638989169675, "grad_norm": 2.1341335773468018, "learning_rate": 5.018050541516246e-06, "loss": 1.4585, "step": 26310 }, { "epoch": 19.00072202166065, "grad_norm": 2.543792486190796, "learning_rate": 4.9963898916967515e-06, "loss": 1.3495, "step": 26316 }, { "epoch": 19.00505415162455, "grad_norm": 2.7721774578094482, "learning_rate": 4.9747292418772564e-06, "loss": 1.429, "step": 26322 }, { "epoch": 19.009386281588448, "grad_norm": 2.2218477725982666, "learning_rate": 4.953068592057762e-06, "loss": 1.4256, "step": 26328 }, { "epoch": 19.013718411552347, "grad_norm": 2.391784906387329, "learning_rate": 4.931407942238268e-06, "loss": 1.503, "step": 26334 }, { "epoch": 19.018050541516246, "grad_norm": 2.311460018157959, "learning_rate": 4.909747292418773e-06, "loss": 1.4203, "step": 26340 }, { "epoch": 19.022382671480145, "grad_norm": 2.315396785736084, "learning_rate": 4.888086642599278e-06, "loss": 1.3201, "step": 26346 }, { "epoch": 19.026714801444044, "grad_norm": 2.5649466514587402, "learning_rate": 4.8664259927797836e-06, "loss": 1.3047, "step": 26352 }, { "epoch": 19.031046931407943, "grad_norm": 2.32490873336792, "learning_rate": 4.844765342960289e-06, "loss": 1.339, "step": 26358 }, { "epoch": 19.03537906137184, "grad_norm": 2.6429243087768555, "learning_rate": 4.823104693140794e-06, "loss": 1.4281, "step": 26364 }, { "epoch": 19.03971119133574, "grad_norm": 2.825432538986206, "learning_rate": 4.8014440433213e-06, "loss": 1.3792, "step": 26370 }, { "epoch": 19.04404332129964, "grad_norm": 2.5132906436920166, "learning_rate": 4.779783393501806e-06, "loss": 1.3307, "step": 26376 }, { "epoch": 19.04837545126354, "grad_norm": 2.4590468406677246, "learning_rate": 4.758122743682311e-06, "loss": 1.3363, "step": 26382 }, { "epoch": 19.052707581227438, "grad_norm": 2.7493736743927, "learning_rate": 4.736462093862816e-06, "loss": 1.532, "step": 26388 }, { "epoch": 19.057039711191337, "grad_norm": 2.790750503540039, "learning_rate": 4.714801444043321e-06, "loss": 1.2904, "step": 26394 }, { "epoch": 19.061371841155236, "grad_norm": 2.703977584838867, "learning_rate": 4.693140794223827e-06, "loss": 1.4867, "step": 26400 }, { "epoch": 19.065703971119135, "grad_norm": 2.6767683029174805, "learning_rate": 4.671480144404333e-06, "loss": 1.4228, "step": 26406 }, { "epoch": 19.070036101083033, "grad_norm": 2.4717495441436768, "learning_rate": 4.649819494584838e-06, "loss": 1.4595, "step": 26412 }, { "epoch": 19.074368231046932, "grad_norm": 2.9438250064849854, "learning_rate": 4.628158844765343e-06, "loss": 1.5143, "step": 26418 }, { "epoch": 19.07870036101083, "grad_norm": 2.668248176574707, "learning_rate": 4.6064981949458485e-06, "loss": 1.4253, "step": 26424 }, { "epoch": 19.08303249097473, "grad_norm": 2.414027452468872, "learning_rate": 4.584837545126354e-06, "loss": 1.4153, "step": 26430 }, { "epoch": 19.08736462093863, "grad_norm": 2.5588250160217285, "learning_rate": 4.563176895306859e-06, "loss": 1.3529, "step": 26436 }, { "epoch": 19.09169675090253, "grad_norm": 2.5687947273254395, "learning_rate": 4.541516245487365e-06, "loss": 1.3289, "step": 26442 }, { "epoch": 19.096028880866427, "grad_norm": 2.7212233543395996, "learning_rate": 4.519855595667871e-06, "loss": 1.4976, "step": 26448 }, { "epoch": 19.100361010830326, "grad_norm": 2.616414785385132, "learning_rate": 4.498194945848376e-06, "loss": 1.4945, "step": 26454 }, { "epoch": 19.104693140794225, "grad_norm": 2.7173941135406494, "learning_rate": 4.476534296028881e-06, "loss": 1.4397, "step": 26460 }, { "epoch": 19.109025270758124, "grad_norm": 2.4771530628204346, "learning_rate": 4.454873646209386e-06, "loss": 1.3891, "step": 26466 }, { "epoch": 19.113357400722023, "grad_norm": 2.5390067100524902, "learning_rate": 4.433212996389892e-06, "loss": 1.3644, "step": 26472 }, { "epoch": 19.117689530685922, "grad_norm": 2.6154375076293945, "learning_rate": 4.411552346570397e-06, "loss": 1.3974, "step": 26478 }, { "epoch": 19.12202166064982, "grad_norm": 2.5204010009765625, "learning_rate": 4.389891696750903e-06, "loss": 1.4124, "step": 26484 }, { "epoch": 19.126353790613717, "grad_norm": 2.07788348197937, "learning_rate": 4.3682310469314086e-06, "loss": 1.4373, "step": 26490 }, { "epoch": 19.130685920577616, "grad_norm": 2.5384163856506348, "learning_rate": 4.3465703971119135e-06, "loss": 1.4046, "step": 26496 }, { "epoch": 19.135018050541515, "grad_norm": 2.5170512199401855, "learning_rate": 4.324909747292418e-06, "loss": 1.5336, "step": 26502 }, { "epoch": 19.139350180505414, "grad_norm": 2.40203857421875, "learning_rate": 4.303249097472924e-06, "loss": 1.422, "step": 26508 }, { "epoch": 19.143682310469313, "grad_norm": 2.4723963737487793, "learning_rate": 4.28158844765343e-06, "loss": 1.2931, "step": 26514 }, { "epoch": 19.14801444043321, "grad_norm": 2.869173765182495, "learning_rate": 4.259927797833936e-06, "loss": 1.4747, "step": 26520 }, { "epoch": 19.15234657039711, "grad_norm": 2.3800582885742188, "learning_rate": 4.238267148014441e-06, "loss": 1.4749, "step": 26526 }, { "epoch": 19.15667870036101, "grad_norm": 2.686772346496582, "learning_rate": 4.2166064981949455e-06, "loss": 1.4304, "step": 26532 }, { "epoch": 19.16101083032491, "grad_norm": 2.532423496246338, "learning_rate": 4.194945848375451e-06, "loss": 1.4537, "step": 26538 }, { "epoch": 19.165342960288807, "grad_norm": 2.4128901958465576, "learning_rate": 4.173285198555957e-06, "loss": 1.3454, "step": 26544 }, { "epoch": 19.169675090252706, "grad_norm": 2.5833377838134766, "learning_rate": 4.151624548736462e-06, "loss": 1.4124, "step": 26550 }, { "epoch": 19.174007220216605, "grad_norm": 2.3376660346984863, "learning_rate": 4.129963898916968e-06, "loss": 1.3902, "step": 26556 }, { "epoch": 19.178339350180504, "grad_norm": 2.699500560760498, "learning_rate": 4.1083032490974735e-06, "loss": 1.4147, "step": 26562 }, { "epoch": 19.182671480144403, "grad_norm": 2.286698818206787, "learning_rate": 4.086642599277979e-06, "loss": 1.4862, "step": 26568 }, { "epoch": 19.187003610108302, "grad_norm": 2.863367795944214, "learning_rate": 4.064981949458483e-06, "loss": 1.5246, "step": 26574 }, { "epoch": 19.1913357400722, "grad_norm": 2.628756046295166, "learning_rate": 4.043321299638989e-06, "loss": 1.495, "step": 26580 }, { "epoch": 19.1956678700361, "grad_norm": 2.5150563716888428, "learning_rate": 4.021660649819495e-06, "loss": 1.4093, "step": 26586 }, { "epoch": 19.2, "grad_norm": 2.4149670600891113, "learning_rate": 4.000000000000001e-06, "loss": 1.3253, "step": 26592 }, { "epoch": 19.2043321299639, "grad_norm": 2.8462131023406982, "learning_rate": 3.978339350180506e-06, "loss": 1.4254, "step": 26598 }, { "epoch": 19.208664259927797, "grad_norm": 2.389044761657715, "learning_rate": 3.956678700361011e-06, "loss": 1.4322, "step": 26604 }, { "epoch": 19.212996389891696, "grad_norm": 2.5180509090423584, "learning_rate": 3.935018050541516e-06, "loss": 1.353, "step": 26610 }, { "epoch": 19.217328519855595, "grad_norm": 2.5566768646240234, "learning_rate": 3.913357400722022e-06, "loss": 1.3928, "step": 26616 }, { "epoch": 19.221660649819494, "grad_norm": 2.4904301166534424, "learning_rate": 3.891696750902527e-06, "loss": 1.4347, "step": 26622 }, { "epoch": 19.225992779783393, "grad_norm": 2.272369146347046, "learning_rate": 3.870036101083033e-06, "loss": 1.3799, "step": 26628 }, { "epoch": 19.230324909747292, "grad_norm": 2.3054022789001465, "learning_rate": 3.8483754512635385e-06, "loss": 1.3261, "step": 26634 }, { "epoch": 19.23465703971119, "grad_norm": 2.373379707336426, "learning_rate": 3.826714801444043e-06, "loss": 1.4476, "step": 26640 }, { "epoch": 19.23898916967509, "grad_norm": 2.4987761974334717, "learning_rate": 3.8050541516245488e-06, "loss": 1.351, "step": 26646 }, { "epoch": 19.24332129963899, "grad_norm": 2.709127187728882, "learning_rate": 3.783393501805054e-06, "loss": 1.4849, "step": 26652 }, { "epoch": 19.247653429602888, "grad_norm": 2.7040655612945557, "learning_rate": 3.7617328519855594e-06, "loss": 1.3612, "step": 26658 }, { "epoch": 19.251985559566787, "grad_norm": 2.265495538711548, "learning_rate": 3.740072202166065e-06, "loss": 1.4226, "step": 26664 }, { "epoch": 19.256317689530686, "grad_norm": 2.5750174522399902, "learning_rate": 3.7184115523465705e-06, "loss": 1.4387, "step": 26670 }, { "epoch": 19.260649819494585, "grad_norm": 2.297348976135254, "learning_rate": 3.6967509025270763e-06, "loss": 1.3296, "step": 26676 }, { "epoch": 19.264981949458484, "grad_norm": 2.3359949588775635, "learning_rate": 3.6750902527075817e-06, "loss": 1.4071, "step": 26682 }, { "epoch": 19.269314079422383, "grad_norm": 2.4355897903442383, "learning_rate": 3.6534296028880866e-06, "loss": 1.405, "step": 26688 }, { "epoch": 19.273646209386282, "grad_norm": 2.5746710300445557, "learning_rate": 3.631768953068592e-06, "loss": 1.5064, "step": 26694 }, { "epoch": 19.27797833935018, "grad_norm": 2.653437614440918, "learning_rate": 3.6101083032490977e-06, "loss": 1.4504, "step": 26700 }, { "epoch": 19.28231046931408, "grad_norm": 2.547764539718628, "learning_rate": 3.588447653429603e-06, "loss": 1.4077, "step": 26706 }, { "epoch": 19.28664259927798, "grad_norm": 2.5042245388031006, "learning_rate": 3.566787003610109e-06, "loss": 1.3108, "step": 26712 }, { "epoch": 19.290974729241878, "grad_norm": 2.257692337036133, "learning_rate": 3.545126353790614e-06, "loss": 1.4113, "step": 26718 }, { "epoch": 19.295306859205777, "grad_norm": 2.292289972305298, "learning_rate": 3.523465703971119e-06, "loss": 1.4092, "step": 26724 }, { "epoch": 19.299638989169676, "grad_norm": 2.4948556423187256, "learning_rate": 3.5018050541516244e-06, "loss": 1.3149, "step": 26730 }, { "epoch": 19.303971119133575, "grad_norm": 2.4236228466033936, "learning_rate": 3.48014440433213e-06, "loss": 1.459, "step": 26736 }, { "epoch": 19.308303249097474, "grad_norm": 2.7799718379974365, "learning_rate": 3.4584837545126355e-06, "loss": 1.399, "step": 26742 }, { "epoch": 19.312635379061373, "grad_norm": 2.5745391845703125, "learning_rate": 3.4368231046931413e-06, "loss": 1.5041, "step": 26748 }, { "epoch": 19.31696750902527, "grad_norm": 2.389965772628784, "learning_rate": 3.4151624548736466e-06, "loss": 1.387, "step": 26754 }, { "epoch": 19.32129963898917, "grad_norm": 2.504967212677002, "learning_rate": 3.393501805054152e-06, "loss": 1.5512, "step": 26760 }, { "epoch": 19.32563176895307, "grad_norm": 2.6153242588043213, "learning_rate": 3.371841155234657e-06, "loss": 1.4492, "step": 26766 }, { "epoch": 19.32996389891697, "grad_norm": 2.5777933597564697, "learning_rate": 3.3501805054151626e-06, "loss": 1.4458, "step": 26772 }, { "epoch": 19.334296028880868, "grad_norm": 2.4357149600982666, "learning_rate": 3.328519855595668e-06, "loss": 1.4919, "step": 26778 }, { "epoch": 19.338628158844767, "grad_norm": 2.366792678833008, "learning_rate": 3.3068592057761733e-06, "loss": 1.3395, "step": 26784 }, { "epoch": 19.342960288808666, "grad_norm": 2.295520782470703, "learning_rate": 3.285198555956679e-06, "loss": 1.4399, "step": 26790 }, { "epoch": 19.347292418772565, "grad_norm": 2.621709108352661, "learning_rate": 3.2635379061371844e-06, "loss": 1.3895, "step": 26796 }, { "epoch": 19.351624548736464, "grad_norm": 2.288290500640869, "learning_rate": 3.2418772563176894e-06, "loss": 1.4258, "step": 26802 }, { "epoch": 19.355956678700363, "grad_norm": 2.455094575881958, "learning_rate": 3.220216606498195e-06, "loss": 1.3972, "step": 26808 }, { "epoch": 19.36028880866426, "grad_norm": 2.468846559524536, "learning_rate": 3.1985559566787005e-06, "loss": 1.4242, "step": 26814 }, { "epoch": 19.36462093862816, "grad_norm": 2.547663450241089, "learning_rate": 3.176895306859206e-06, "loss": 1.3722, "step": 26820 }, { "epoch": 19.36895306859206, "grad_norm": 2.4453258514404297, "learning_rate": 3.1552346570397116e-06, "loss": 1.3887, "step": 26826 }, { "epoch": 19.373285198555955, "grad_norm": 2.3692305088043213, "learning_rate": 3.133574007220217e-06, "loss": 1.465, "step": 26832 }, { "epoch": 19.377617328519854, "grad_norm": 2.528554677963257, "learning_rate": 3.1119133574007223e-06, "loss": 1.2897, "step": 26838 }, { "epoch": 19.381949458483753, "grad_norm": 2.4814436435699463, "learning_rate": 3.0902527075812276e-06, "loss": 1.4779, "step": 26844 }, { "epoch": 19.386281588447652, "grad_norm": 2.5663022994995117, "learning_rate": 3.068592057761733e-06, "loss": 1.4035, "step": 26850 }, { "epoch": 19.39061371841155, "grad_norm": 2.7176339626312256, "learning_rate": 3.0469314079422383e-06, "loss": 1.404, "step": 26856 }, { "epoch": 19.39494584837545, "grad_norm": 2.1547858715057373, "learning_rate": 3.025270758122744e-06, "loss": 1.3872, "step": 26862 }, { "epoch": 19.39927797833935, "grad_norm": 2.8005237579345703, "learning_rate": 3.003610108303249e-06, "loss": 1.4482, "step": 26868 }, { "epoch": 19.403610108303248, "grad_norm": 2.824673652648926, "learning_rate": 2.9819494584837547e-06, "loss": 1.4629, "step": 26874 }, { "epoch": 19.407942238267147, "grad_norm": 2.465768575668335, "learning_rate": 2.96028880866426e-06, "loss": 1.4644, "step": 26880 }, { "epoch": 19.412274368231046, "grad_norm": 2.738081216812134, "learning_rate": 2.9386281588447654e-06, "loss": 1.5108, "step": 26886 }, { "epoch": 19.416606498194945, "grad_norm": 2.6444203853607178, "learning_rate": 2.9169675090252708e-06, "loss": 1.4445, "step": 26892 }, { "epoch": 19.420938628158844, "grad_norm": 2.851893663406372, "learning_rate": 2.8953068592057765e-06, "loss": 1.4521, "step": 26898 }, { "epoch": 19.425270758122743, "grad_norm": 2.2797679901123047, "learning_rate": 2.8736462093862815e-06, "loss": 1.3997, "step": 26904 }, { "epoch": 19.42960288808664, "grad_norm": 2.4016995429992676, "learning_rate": 2.8519855595667872e-06, "loss": 1.4482, "step": 26910 }, { "epoch": 19.43393501805054, "grad_norm": 2.8234822750091553, "learning_rate": 2.8303249097472926e-06, "loss": 1.3483, "step": 26916 }, { "epoch": 19.43826714801444, "grad_norm": 2.358480215072632, "learning_rate": 2.8086642599277983e-06, "loss": 1.4852, "step": 26922 }, { "epoch": 19.44259927797834, "grad_norm": 2.2871172428131104, "learning_rate": 2.7870036101083032e-06, "loss": 1.3745, "step": 26928 }, { "epoch": 19.446931407942238, "grad_norm": 2.4724934101104736, "learning_rate": 2.765342960288809e-06, "loss": 1.4538, "step": 26934 }, { "epoch": 19.451263537906136, "grad_norm": 2.503798246383667, "learning_rate": 2.7436823104693144e-06, "loss": 1.3789, "step": 26940 }, { "epoch": 19.455595667870035, "grad_norm": 2.6082680225372314, "learning_rate": 2.7220216606498197e-06, "loss": 1.4682, "step": 26946 }, { "epoch": 19.459927797833934, "grad_norm": 2.5114083290100098, "learning_rate": 2.700361010830325e-06, "loss": 1.5219, "step": 26952 }, { "epoch": 19.464259927797833, "grad_norm": 2.4595327377319336, "learning_rate": 2.6787003610108304e-06, "loss": 1.4653, "step": 26958 }, { "epoch": 19.468592057761732, "grad_norm": 2.650552749633789, "learning_rate": 2.6570397111913357e-06, "loss": 1.4577, "step": 26964 }, { "epoch": 19.47292418772563, "grad_norm": 2.4266111850738525, "learning_rate": 2.635379061371841e-06, "loss": 1.4593, "step": 26970 }, { "epoch": 19.47725631768953, "grad_norm": 2.496300220489502, "learning_rate": 2.613718411552347e-06, "loss": 1.4886, "step": 26976 }, { "epoch": 19.48158844765343, "grad_norm": 2.228790521621704, "learning_rate": 2.5920577617328518e-06, "loss": 1.4043, "step": 26982 }, { "epoch": 19.48592057761733, "grad_norm": 2.5464909076690674, "learning_rate": 2.5703971119133575e-06, "loss": 1.4313, "step": 26988 }, { "epoch": 19.490252707581227, "grad_norm": 2.6035544872283936, "learning_rate": 2.548736462093863e-06, "loss": 1.3797, "step": 26994 }, { "epoch": 19.494584837545126, "grad_norm": 2.596024751663208, "learning_rate": 2.527075812274368e-06, "loss": 1.3802, "step": 27000 }, { "epoch": 19.498916967509025, "grad_norm": 2.6154043674468994, "learning_rate": 2.5054151624548736e-06, "loss": 1.4189, "step": 27006 }, { "epoch": 19.503249097472924, "grad_norm": 2.667776584625244, "learning_rate": 2.4837545126353793e-06, "loss": 1.4816, "step": 27012 }, { "epoch": 19.507581227436823, "grad_norm": 2.3155055046081543, "learning_rate": 2.4620938628158847e-06, "loss": 1.3149, "step": 27018 }, { "epoch": 19.511913357400722, "grad_norm": 2.588400363922119, "learning_rate": 2.44043321299639e-06, "loss": 1.4532, "step": 27024 }, { "epoch": 19.51624548736462, "grad_norm": 2.675983428955078, "learning_rate": 2.4187725631768953e-06, "loss": 1.4692, "step": 27030 }, { "epoch": 19.52057761732852, "grad_norm": 2.430321216583252, "learning_rate": 2.397111913357401e-06, "loss": 1.3855, "step": 27036 }, { "epoch": 19.52490974729242, "grad_norm": 2.4746577739715576, "learning_rate": 2.375451263537906e-06, "loss": 1.3449, "step": 27042 }, { "epoch": 19.529241877256318, "grad_norm": 2.518601894378662, "learning_rate": 2.353790613718412e-06, "loss": 1.4005, "step": 27048 }, { "epoch": 19.533574007220217, "grad_norm": 2.4738688468933105, "learning_rate": 2.332129963898917e-06, "loss": 1.383, "step": 27054 }, { "epoch": 19.537906137184116, "grad_norm": 2.7040364742279053, "learning_rate": 2.3104693140794225e-06, "loss": 1.4012, "step": 27060 }, { "epoch": 19.542238267148015, "grad_norm": 2.5072953701019287, "learning_rate": 2.288808664259928e-06, "loss": 1.4359, "step": 27066 }, { "epoch": 19.546570397111914, "grad_norm": 2.672487497329712, "learning_rate": 2.2671480144404336e-06, "loss": 1.4926, "step": 27072 }, { "epoch": 19.550902527075813, "grad_norm": 2.6927685737609863, "learning_rate": 2.2454873646209385e-06, "loss": 1.4559, "step": 27078 }, { "epoch": 19.555234657039712, "grad_norm": 2.6227478981018066, "learning_rate": 2.2238267148014443e-06, "loss": 1.36, "step": 27084 }, { "epoch": 19.55956678700361, "grad_norm": 2.3700733184814453, "learning_rate": 2.2021660649819496e-06, "loss": 1.4397, "step": 27090 }, { "epoch": 19.56389891696751, "grad_norm": 2.809624433517456, "learning_rate": 2.180505415162455e-06, "loss": 1.3626, "step": 27096 }, { "epoch": 19.56823104693141, "grad_norm": 2.511704921722412, "learning_rate": 2.1588447653429603e-06, "loss": 1.4265, "step": 27102 }, { "epoch": 19.572563176895308, "grad_norm": 2.456658124923706, "learning_rate": 2.1371841155234656e-06, "loss": 1.4159, "step": 27108 }, { "epoch": 19.576895306859207, "grad_norm": 2.133697509765625, "learning_rate": 2.1155234657039714e-06, "loss": 1.3702, "step": 27114 }, { "epoch": 19.581227436823106, "grad_norm": 2.385801076889038, "learning_rate": 2.0938628158844768e-06, "loss": 1.3959, "step": 27120 }, { "epoch": 19.585559566787005, "grad_norm": 2.1887307167053223, "learning_rate": 2.072202166064982e-06, "loss": 1.3012, "step": 27126 }, { "epoch": 19.589891696750904, "grad_norm": 2.5283753871917725, "learning_rate": 2.0505415162454874e-06, "loss": 1.3646, "step": 27132 }, { "epoch": 19.594223826714803, "grad_norm": 2.6152403354644775, "learning_rate": 2.0288808664259928e-06, "loss": 1.4737, "step": 27138 }, { "epoch": 19.5985559566787, "grad_norm": 2.4558279514312744, "learning_rate": 2.007220216606498e-06, "loss": 1.486, "step": 27144 }, { "epoch": 19.6028880866426, "grad_norm": 2.5573511123657227, "learning_rate": 1.985559566787004e-06, "loss": 1.3284, "step": 27150 }, { "epoch": 19.6072202166065, "grad_norm": 2.6110451221466064, "learning_rate": 1.963898916967509e-06, "loss": 1.3936, "step": 27156 }, { "epoch": 19.6115523465704, "grad_norm": 2.6064484119415283, "learning_rate": 1.9422382671480146e-06, "loss": 1.4597, "step": 27162 }, { "epoch": 19.615884476534298, "grad_norm": 2.6599624156951904, "learning_rate": 1.92057761732852e-06, "loss": 1.4366, "step": 27168 }, { "epoch": 19.620216606498197, "grad_norm": 2.677736759185791, "learning_rate": 1.8989169675090253e-06, "loss": 1.3743, "step": 27174 }, { "epoch": 19.624548736462096, "grad_norm": 2.3139147758483887, "learning_rate": 1.8772563176895308e-06, "loss": 1.4258, "step": 27180 }, { "epoch": 19.628880866425995, "grad_norm": 2.6534969806671143, "learning_rate": 1.8555956678700362e-06, "loss": 1.4507, "step": 27186 }, { "epoch": 19.63321299638989, "grad_norm": 2.35331130027771, "learning_rate": 1.8339350180505415e-06, "loss": 1.3476, "step": 27192 }, { "epoch": 19.63754512635379, "grad_norm": 2.266289234161377, "learning_rate": 1.812274368231047e-06, "loss": 1.4447, "step": 27198 }, { "epoch": 19.641877256317688, "grad_norm": 2.1479079723358154, "learning_rate": 1.7906137184115524e-06, "loss": 1.3785, "step": 27204 }, { "epoch": 19.646209386281587, "grad_norm": 2.61799693107605, "learning_rate": 1.768953068592058e-06, "loss": 1.4477, "step": 27210 }, { "epoch": 19.650541516245486, "grad_norm": 2.5010132789611816, "learning_rate": 1.747292418772563e-06, "loss": 1.3919, "step": 27216 }, { "epoch": 19.654873646209385, "grad_norm": 2.3431031703948975, "learning_rate": 1.7256317689530686e-06, "loss": 1.4039, "step": 27222 }, { "epoch": 19.659205776173284, "grad_norm": 2.684154748916626, "learning_rate": 1.7039711191335742e-06, "loss": 1.3555, "step": 27228 }, { "epoch": 19.663537906137183, "grad_norm": 2.5190656185150146, "learning_rate": 1.6823104693140793e-06, "loss": 1.2303, "step": 27234 }, { "epoch": 19.667870036101082, "grad_norm": 2.791242837905884, "learning_rate": 1.6606498194945849e-06, "loss": 1.5408, "step": 27240 }, { "epoch": 19.67220216606498, "grad_norm": 2.4511666297912598, "learning_rate": 1.6389891696750904e-06, "loss": 1.3261, "step": 27246 }, { "epoch": 19.67653429602888, "grad_norm": 2.506150245666504, "learning_rate": 1.6173285198555956e-06, "loss": 1.3038, "step": 27252 }, { "epoch": 19.68086642599278, "grad_norm": 2.75959849357605, "learning_rate": 1.5956678700361011e-06, "loss": 1.5282, "step": 27258 }, { "epoch": 19.685198555956678, "grad_norm": 2.546785593032837, "learning_rate": 1.5740072202166067e-06, "loss": 1.471, "step": 27264 }, { "epoch": 19.689530685920577, "grad_norm": 2.413780450820923, "learning_rate": 1.552346570397112e-06, "loss": 1.3562, "step": 27270 }, { "epoch": 19.693862815884476, "grad_norm": 2.970850706100464, "learning_rate": 1.5306859205776174e-06, "loss": 1.3897, "step": 27276 }, { "epoch": 19.698194945848375, "grad_norm": 2.539325714111328, "learning_rate": 1.509025270758123e-06, "loss": 1.3361, "step": 27282 }, { "epoch": 19.702527075812274, "grad_norm": 2.2217304706573486, "learning_rate": 1.4873646209386283e-06, "loss": 1.3811, "step": 27288 }, { "epoch": 19.706859205776173, "grad_norm": 2.393024206161499, "learning_rate": 1.4657039711191336e-06, "loss": 1.4106, "step": 27294 }, { "epoch": 19.71119133574007, "grad_norm": 2.579070806503296, "learning_rate": 1.4440433212996392e-06, "loss": 1.4253, "step": 27300 }, { "epoch": 19.71552346570397, "grad_norm": 2.695089101791382, "learning_rate": 1.4223826714801445e-06, "loss": 1.4565, "step": 27306 }, { "epoch": 19.71985559566787, "grad_norm": 2.961920738220215, "learning_rate": 1.4007220216606498e-06, "loss": 1.5995, "step": 27312 }, { "epoch": 19.72418772563177, "grad_norm": 2.4129157066345215, "learning_rate": 1.3790613718411554e-06, "loss": 1.4218, "step": 27318 }, { "epoch": 19.728519855595668, "grad_norm": 2.3928892612457275, "learning_rate": 1.3574007220216607e-06, "loss": 1.3613, "step": 27324 }, { "epoch": 19.732851985559567, "grad_norm": 2.354571580886841, "learning_rate": 1.3357400722021663e-06, "loss": 1.3663, "step": 27330 }, { "epoch": 19.737184115523466, "grad_norm": 2.403343439102173, "learning_rate": 1.3140794223826716e-06, "loss": 1.4901, "step": 27336 }, { "epoch": 19.741516245487365, "grad_norm": 2.5035006999969482, "learning_rate": 1.292418772563177e-06, "loss": 1.3878, "step": 27342 }, { "epoch": 19.745848375451263, "grad_norm": 2.760373592376709, "learning_rate": 1.2707581227436823e-06, "loss": 1.5393, "step": 27348 }, { "epoch": 19.750180505415162, "grad_norm": 2.7306740283966064, "learning_rate": 1.2490974729241879e-06, "loss": 1.4479, "step": 27354 }, { "epoch": 19.75451263537906, "grad_norm": 2.399951457977295, "learning_rate": 1.2274368231046932e-06, "loss": 1.5017, "step": 27360 }, { "epoch": 19.75884476534296, "grad_norm": 2.53074049949646, "learning_rate": 1.2057761732851986e-06, "loss": 1.5465, "step": 27366 }, { "epoch": 19.76317689530686, "grad_norm": 2.629042863845825, "learning_rate": 1.184115523465704e-06, "loss": 1.4235, "step": 27372 }, { "epoch": 19.76750902527076, "grad_norm": 2.398780107498169, "learning_rate": 1.1624548736462095e-06, "loss": 1.4533, "step": 27378 }, { "epoch": 19.771841155234657, "grad_norm": 2.6229801177978516, "learning_rate": 1.1407942238267148e-06, "loss": 1.525, "step": 27384 }, { "epoch": 19.776173285198556, "grad_norm": 2.495570659637451, "learning_rate": 1.1191335740072201e-06, "loss": 1.4153, "step": 27390 }, { "epoch": 19.780505415162455, "grad_norm": 2.455742359161377, "learning_rate": 1.0974729241877257e-06, "loss": 1.4688, "step": 27396 }, { "epoch": 19.784837545126354, "grad_norm": 2.500523805618286, "learning_rate": 1.075812274368231e-06, "loss": 1.4925, "step": 27402 }, { "epoch": 19.789169675090253, "grad_norm": 2.267819881439209, "learning_rate": 1.0541516245487364e-06, "loss": 1.3449, "step": 27408 }, { "epoch": 19.793501805054152, "grad_norm": 2.225870370864868, "learning_rate": 1.032490974729242e-06, "loss": 1.4212, "step": 27414 }, { "epoch": 19.79783393501805, "grad_norm": 2.4810354709625244, "learning_rate": 1.0108303249097473e-06, "loss": 1.3926, "step": 27420 }, { "epoch": 19.80216606498195, "grad_norm": 2.6158413887023926, "learning_rate": 9.891696750902528e-07, "loss": 1.399, "step": 27426 }, { "epoch": 19.80649819494585, "grad_norm": 2.4794883728027344, "learning_rate": 9.675090252707582e-07, "loss": 1.4483, "step": 27432 }, { "epoch": 19.810830324909748, "grad_norm": 2.4125781059265137, "learning_rate": 9.458483754512635e-07, "loss": 1.3558, "step": 27438 }, { "epoch": 19.815162454873647, "grad_norm": 2.6394312381744385, "learning_rate": 9.241877256317691e-07, "loss": 1.3952, "step": 27444 }, { "epoch": 19.819494584837546, "grad_norm": 2.5753896236419678, "learning_rate": 9.025270758122744e-07, "loss": 1.4376, "step": 27450 }, { "epoch": 19.823826714801445, "grad_norm": 2.5634255409240723, "learning_rate": 8.808664259927798e-07, "loss": 1.424, "step": 27456 }, { "epoch": 19.828158844765344, "grad_norm": 2.579447031021118, "learning_rate": 8.592057761732853e-07, "loss": 1.3721, "step": 27462 }, { "epoch": 19.832490974729243, "grad_norm": 2.7184503078460693, "learning_rate": 8.375451263537907e-07, "loss": 1.4407, "step": 27468 }, { "epoch": 19.836823104693142, "grad_norm": 2.609950304031372, "learning_rate": 8.158844765342961e-07, "loss": 1.4567, "step": 27474 }, { "epoch": 19.84115523465704, "grad_norm": 2.6822028160095215, "learning_rate": 7.942238267148015e-07, "loss": 1.4233, "step": 27480 }, { "epoch": 19.84548736462094, "grad_norm": 2.6408209800720215, "learning_rate": 7.725631768953069e-07, "loss": 1.4392, "step": 27486 }, { "epoch": 19.84981949458484, "grad_norm": 2.7046868801116943, "learning_rate": 7.509025270758122e-07, "loss": 1.492, "step": 27492 }, { "epoch": 19.854151624548738, "grad_norm": 2.6732356548309326, "learning_rate": 7.292418772563177e-07, "loss": 1.4952, "step": 27498 }, { "epoch": 19.858483754512637, "grad_norm": 2.652291774749756, "learning_rate": 7.075812274368231e-07, "loss": 1.4933, "step": 27504 }, { "epoch": 19.862815884476536, "grad_norm": 2.758049726486206, "learning_rate": 6.859205776173286e-07, "loss": 1.4485, "step": 27510 }, { "epoch": 19.867148014440435, "grad_norm": 2.6547536849975586, "learning_rate": 6.642599277978339e-07, "loss": 1.4149, "step": 27516 }, { "epoch": 19.87148014440433, "grad_norm": 2.6873414516448975, "learning_rate": 6.425992779783394e-07, "loss": 1.5007, "step": 27522 }, { "epoch": 19.87581227436823, "grad_norm": 2.487819194793701, "learning_rate": 6.209386281588448e-07, "loss": 1.3936, "step": 27528 }, { "epoch": 19.88014440433213, "grad_norm": 3.0470738410949707, "learning_rate": 5.992779783393503e-07, "loss": 1.3874, "step": 27534 }, { "epoch": 19.884476534296027, "grad_norm": 2.570284605026245, "learning_rate": 5.776173285198556e-07, "loss": 1.4398, "step": 27540 }, { "epoch": 19.888808664259926, "grad_norm": 2.5843591690063477, "learning_rate": 5.559566787003611e-07, "loss": 1.4238, "step": 27546 }, { "epoch": 19.893140794223825, "grad_norm": 2.487746000289917, "learning_rate": 5.342960288808664e-07, "loss": 1.4589, "step": 27552 }, { "epoch": 19.897472924187724, "grad_norm": 2.809028148651123, "learning_rate": 5.126353790613719e-07, "loss": 1.4187, "step": 27558 }, { "epoch": 19.901805054151623, "grad_norm": 2.680992841720581, "learning_rate": 4.909747292418772e-07, "loss": 1.4475, "step": 27564 }, { "epoch": 19.906137184115522, "grad_norm": 2.2696890830993652, "learning_rate": 4.693140794223827e-07, "loss": 1.3885, "step": 27570 }, { "epoch": 19.91046931407942, "grad_norm": 2.7030398845672607, "learning_rate": 4.476534296028881e-07, "loss": 1.2767, "step": 27576 }, { "epoch": 19.91480144404332, "grad_norm": 2.7544164657592773, "learning_rate": 4.2599277978339355e-07, "loss": 1.4808, "step": 27582 }, { "epoch": 19.91913357400722, "grad_norm": 2.286003589630127, "learning_rate": 4.043321299638989e-07, "loss": 1.3804, "step": 27588 }, { "epoch": 19.923465703971118, "grad_norm": 2.4203484058380127, "learning_rate": 3.8267148014440434e-07, "loss": 1.4605, "step": 27594 }, { "epoch": 19.927797833935017, "grad_norm": 2.4331867694854736, "learning_rate": 3.610108303249098e-07, "loss": 1.4271, "step": 27600 }, { "epoch": 19.932129963898916, "grad_norm": 2.377372980117798, "learning_rate": 3.393501805054152e-07, "loss": 1.3914, "step": 27606 }, { "epoch": 19.936462093862815, "grad_norm": 2.5173730850219727, "learning_rate": 3.176895306859206e-07, "loss": 1.3687, "step": 27612 }, { "epoch": 19.940794223826714, "grad_norm": 2.239994764328003, "learning_rate": 2.96028880866426e-07, "loss": 1.4661, "step": 27618 }, { "epoch": 19.945126353790613, "grad_norm": 2.7954928874969482, "learning_rate": 2.743682310469314e-07, "loss": 1.4402, "step": 27624 }, { "epoch": 19.949458483754512, "grad_norm": 2.4934043884277344, "learning_rate": 2.527075812274368e-07, "loss": 1.3296, "step": 27630 }, { "epoch": 19.95379061371841, "grad_norm": 2.3748035430908203, "learning_rate": 2.3104693140794227e-07, "loss": 1.4246, "step": 27636 }, { "epoch": 19.95812274368231, "grad_norm": 2.8449525833129883, "learning_rate": 2.0938628158844767e-07, "loss": 1.5278, "step": 27642 }, { "epoch": 19.96245487364621, "grad_norm": 2.4567933082580566, "learning_rate": 1.8772563176895306e-07, "loss": 1.4569, "step": 27648 }, { "epoch": 19.966787003610108, "grad_norm": 2.561180591583252, "learning_rate": 1.6606498194945848e-07, "loss": 1.3766, "step": 27654 }, { "epoch": 19.971119133574007, "grad_norm": 2.4242210388183594, "learning_rate": 1.444043321299639e-07, "loss": 1.3285, "step": 27660 }, { "epoch": 19.975451263537906, "grad_norm": 2.533121347427368, "learning_rate": 1.227436823104693e-07, "loss": 1.413, "step": 27666 }, { "epoch": 19.979783393501805, "grad_norm": 2.664203643798828, "learning_rate": 1.0108303249097472e-07, "loss": 1.4717, "step": 27672 }, { "epoch": 19.984115523465704, "grad_norm": 2.8825151920318604, "learning_rate": 7.942238267148015e-08, "loss": 1.5142, "step": 27678 }, { "epoch": 19.988447653429603, "grad_norm": 2.3971333503723145, "learning_rate": 5.776173285198557e-08, "loss": 1.3168, "step": 27684 }, { "epoch": 19.9927797833935, "grad_norm": 2.768261194229126, "learning_rate": 3.6101083032490976e-08, "loss": 1.4525, "step": 27690 }, { "epoch": 19.9971119133574, "grad_norm": 2.583491086959839, "learning_rate": 1.4440433212996392e-08, "loss": 1.4777, "step": 27696 } ], "logging_steps": 6, "max_steps": 27700, "num_input_tokens_seen": 0, "num_train_epochs": 20, "save_steps": 500, "total_flos": 9.685041142443633e+17, "train_batch_size": 24, "trial_name": null, "trial_params": null }