code-lora-hard / trainer_state.json
cterdam's picture
Upload 8 files
6ca82ee verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 5.0,
"eval_steps": 1000,
"global_step": 100000,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.01,
"grad_norm": 3.153648614883423,
"learning_rate": 5.9999999999999995e-05,
"loss": 1.9475,
"step": 100
},
{
"epoch": 0.01,
"grad_norm": 1.318032145500183,
"learning_rate": 0.00011999999999999999,
"loss": 1.7581,
"step": 200
},
{
"epoch": 0.01,
"grad_norm": 1.7753959894180298,
"learning_rate": 0.00017999999999999998,
"loss": 1.7353,
"step": 300
},
{
"epoch": 0.02,
"grad_norm": 1.262397289276123,
"learning_rate": 0.00023999999999999998,
"loss": 1.7634,
"step": 400
},
{
"epoch": 0.03,
"grad_norm": 1.793628454208374,
"learning_rate": 0.0003,
"loss": 1.7828,
"step": 500
},
{
"epoch": 0.03,
"grad_norm": 1.29915452003479,
"learning_rate": 0.00029969849246231153,
"loss": 1.7973,
"step": 600
},
{
"epoch": 0.04,
"grad_norm": 1.3753058910369873,
"learning_rate": 0.0002993969849246231,
"loss": 1.7677,
"step": 700
},
{
"epoch": 0.04,
"grad_norm": 3.032518148422241,
"learning_rate": 0.00029909547738693465,
"loss": 1.7775,
"step": 800
},
{
"epoch": 0.04,
"grad_norm": 1.4150569438934326,
"learning_rate": 0.0002987939698492462,
"loss": 1.7872,
"step": 900
},
{
"epoch": 0.05,
"grad_norm": 1.6522352695465088,
"learning_rate": 0.00029849246231155777,
"loss": 1.7433,
"step": 1000
},
{
"epoch": 0.05,
"eval_loss": 1.7443219423294067,
"eval_runtime": 37.3589,
"eval_samples_per_second": 26.767,
"eval_steps_per_second": 3.346,
"step": 1000
},
{
"epoch": 0.06,
"grad_norm": 1.8125498294830322,
"learning_rate": 0.00029819095477386933,
"loss": 1.7885,
"step": 1100
},
{
"epoch": 0.06,
"grad_norm": 1.4740030765533447,
"learning_rate": 0.0002978894472361809,
"loss": 1.7616,
"step": 1200
},
{
"epoch": 0.07,
"grad_norm": 1.670320749282837,
"learning_rate": 0.00029758793969849245,
"loss": 1.7545,
"step": 1300
},
{
"epoch": 0.07,
"grad_norm": 1.096781611442566,
"learning_rate": 0.000297286432160804,
"loss": 1.7072,
"step": 1400
},
{
"epoch": 0.07,
"grad_norm": 1.8927794694900513,
"learning_rate": 0.0002969849246231155,
"loss": 1.7243,
"step": 1500
},
{
"epoch": 0.08,
"grad_norm": 2.244074583053589,
"learning_rate": 0.00029668341708542713,
"loss": 1.7369,
"step": 1600
},
{
"epoch": 0.09,
"grad_norm": 1.6228864192962646,
"learning_rate": 0.0002963819095477387,
"loss": 1.7541,
"step": 1700
},
{
"epoch": 0.09,
"grad_norm": 0.9201287627220154,
"learning_rate": 0.00029608040201005025,
"loss": 1.7236,
"step": 1800
},
{
"epoch": 0.1,
"grad_norm": 1.308199405670166,
"learning_rate": 0.00029577889447236175,
"loss": 1.7345,
"step": 1900
},
{
"epoch": 0.1,
"grad_norm": 1.5356643199920654,
"learning_rate": 0.00029548040201005023,
"loss": 1.7256,
"step": 2000
},
{
"epoch": 0.1,
"eval_loss": 1.7340322732925415,
"eval_runtime": 37.5614,
"eval_samples_per_second": 26.623,
"eval_steps_per_second": 3.328,
"step": 2000
},
{
"epoch": 0.1,
"grad_norm": 2.5252068042755127,
"learning_rate": 0.0002951788944723618,
"loss": 1.7454,
"step": 2100
},
{
"epoch": 0.11,
"grad_norm": 1.5635749101638794,
"learning_rate": 0.00029487738693467335,
"loss": 1.7461,
"step": 2200
},
{
"epoch": 0.12,
"grad_norm": 1.3426711559295654,
"learning_rate": 0.0002945758793969849,
"loss": 1.75,
"step": 2300
},
{
"epoch": 0.12,
"grad_norm": 1.3829035758972168,
"learning_rate": 0.00029427437185929647,
"loss": 1.7554,
"step": 2400
},
{
"epoch": 0.12,
"grad_norm": 1.2779866456985474,
"learning_rate": 0.00029397286432160803,
"loss": 1.7262,
"step": 2500
},
{
"epoch": 0.13,
"grad_norm": 1.4913188219070435,
"learning_rate": 0.0002936713567839196,
"loss": 1.7333,
"step": 2600
},
{
"epoch": 0.14,
"grad_norm": 1.5596439838409424,
"learning_rate": 0.00029336984924623115,
"loss": 1.728,
"step": 2700
},
{
"epoch": 0.14,
"grad_norm": 1.1473088264465332,
"learning_rate": 0.0002930683417085427,
"loss": 1.7063,
"step": 2800
},
{
"epoch": 0.14,
"grad_norm": 1.4872281551361084,
"learning_rate": 0.0002927668341708542,
"loss": 1.7417,
"step": 2900
},
{
"epoch": 0.15,
"grad_norm": 1.3485779762268066,
"learning_rate": 0.0002924653266331658,
"loss": 1.7314,
"step": 3000
},
{
"epoch": 0.15,
"eval_loss": 1.680450677871704,
"eval_runtime": 38.1024,
"eval_samples_per_second": 26.245,
"eval_steps_per_second": 3.281,
"step": 3000
},
{
"epoch": 0.15,
"grad_norm": 1.3290046453475952,
"learning_rate": 0.0002921638190954774,
"loss": 1.7101,
"step": 3100
},
{
"epoch": 0.16,
"grad_norm": 0.9259174466133118,
"learning_rate": 0.00029186231155778895,
"loss": 1.75,
"step": 3200
},
{
"epoch": 0.17,
"grad_norm": 1.4094349145889282,
"learning_rate": 0.00029156080402010045,
"loss": 1.7219,
"step": 3300
},
{
"epoch": 0.17,
"grad_norm": 1.2119574546813965,
"learning_rate": 0.000291259296482412,
"loss": 1.741,
"step": 3400
},
{
"epoch": 0.17,
"grad_norm": 1.1941887140274048,
"learning_rate": 0.0002909577889447236,
"loss": 1.7005,
"step": 3500
},
{
"epoch": 0.18,
"grad_norm": 1.1339538097381592,
"learning_rate": 0.00029065628140703513,
"loss": 1.6965,
"step": 3600
},
{
"epoch": 0.18,
"grad_norm": 2.2265231609344482,
"learning_rate": 0.0002903547738693467,
"loss": 1.701,
"step": 3700
},
{
"epoch": 0.19,
"grad_norm": 2.011225938796997,
"learning_rate": 0.00029005326633165825,
"loss": 1.7241,
"step": 3800
},
{
"epoch": 0.2,
"grad_norm": 1.3544070720672607,
"learning_rate": 0.00028975175879396986,
"loss": 1.7039,
"step": 3900
},
{
"epoch": 0.2,
"grad_norm": 1.909501314163208,
"learning_rate": 0.00028945025125628137,
"loss": 1.676,
"step": 4000
},
{
"epoch": 0.2,
"eval_loss": 1.6920135021209717,
"eval_runtime": 37.5936,
"eval_samples_per_second": 26.6,
"eval_steps_per_second": 3.325,
"step": 4000
},
{
"epoch": 0.2,
"grad_norm": 1.7186845541000366,
"learning_rate": 0.0002891487437185929,
"loss": 1.7,
"step": 4100
},
{
"epoch": 0.21,
"grad_norm": 0.9834026098251343,
"learning_rate": 0.0002888502512562814,
"loss": 1.6639,
"step": 4200
},
{
"epoch": 0.21,
"grad_norm": 0.9373641014099121,
"learning_rate": 0.0002885487437185929,
"loss": 1.6786,
"step": 4300
},
{
"epoch": 0.22,
"grad_norm": 1.4471676349639893,
"learning_rate": 0.00028824723618090447,
"loss": 1.7199,
"step": 4400
},
{
"epoch": 0.23,
"grad_norm": 0.9259161949157715,
"learning_rate": 0.0002879457286432161,
"loss": 1.7013,
"step": 4500
},
{
"epoch": 0.23,
"grad_norm": 1.3199846744537354,
"learning_rate": 0.0002876442211055276,
"loss": 1.6865,
"step": 4600
},
{
"epoch": 0.23,
"grad_norm": 1.6591229438781738,
"learning_rate": 0.00028734271356783915,
"loss": 1.6976,
"step": 4700
},
{
"epoch": 0.24,
"grad_norm": 1.0676679611206055,
"learning_rate": 0.0002870412060301507,
"loss": 1.6921,
"step": 4800
},
{
"epoch": 0.24,
"grad_norm": 1.3471635580062866,
"learning_rate": 0.0002867396984924623,
"loss": 1.6951,
"step": 4900
},
{
"epoch": 0.25,
"grad_norm": 1.6372334957122803,
"learning_rate": 0.00028643819095477383,
"loss": 1.6951,
"step": 5000
},
{
"epoch": 0.25,
"eval_loss": 1.6992712020874023,
"eval_runtime": 37.5723,
"eval_samples_per_second": 26.615,
"eval_steps_per_second": 3.327,
"step": 5000
},
{
"epoch": 0.26,
"grad_norm": 2.9456429481506348,
"learning_rate": 0.0002861366834170854,
"loss": 1.6708,
"step": 5100
},
{
"epoch": 0.26,
"grad_norm": 1.8768843412399292,
"learning_rate": 0.00028583517587939695,
"loss": 1.701,
"step": 5200
},
{
"epoch": 0.27,
"grad_norm": 1.4709163904190063,
"learning_rate": 0.0002855336683417085,
"loss": 1.7119,
"step": 5300
},
{
"epoch": 0.27,
"grad_norm": 1.15412175655365,
"learning_rate": 0.00028523216080402007,
"loss": 1.6864,
"step": 5400
},
{
"epoch": 0.28,
"grad_norm": 1.9388611316680908,
"learning_rate": 0.0002849306532663316,
"loss": 1.6781,
"step": 5500
},
{
"epoch": 0.28,
"grad_norm": 2.746967315673828,
"learning_rate": 0.0002846291457286432,
"loss": 1.7134,
"step": 5600
},
{
"epoch": 0.28,
"grad_norm": 1.7656490802764893,
"learning_rate": 0.00028432763819095474,
"loss": 1.6441,
"step": 5700
},
{
"epoch": 0.29,
"grad_norm": 1.4275192022323608,
"learning_rate": 0.0002840261306532663,
"loss": 1.6893,
"step": 5800
},
{
"epoch": 0.29,
"grad_norm": 1.8908779621124268,
"learning_rate": 0.00028372462311557786,
"loss": 1.6814,
"step": 5900
},
{
"epoch": 0.3,
"grad_norm": 2.0805633068084717,
"learning_rate": 0.0002834231155778894,
"loss": 1.6765,
"step": 6000
},
{
"epoch": 0.3,
"eval_loss": 1.6724690198898315,
"eval_runtime": 37.9367,
"eval_samples_per_second": 26.36,
"eval_steps_per_second": 3.295,
"step": 6000
},
{
"epoch": 0.3,
"grad_norm": 1.8758082389831543,
"learning_rate": 0.000283121608040201,
"loss": 1.6462,
"step": 6100
},
{
"epoch": 0.31,
"grad_norm": 1.4686168432235718,
"learning_rate": 0.00028282010050251254,
"loss": 1.7117,
"step": 6200
},
{
"epoch": 0.32,
"grad_norm": 1.857920527458191,
"learning_rate": 0.0002825185929648241,
"loss": 1.648,
"step": 6300
},
{
"epoch": 0.32,
"grad_norm": 1.852232575416565,
"learning_rate": 0.00028221708542713566,
"loss": 1.7096,
"step": 6400
},
{
"epoch": 0.33,
"grad_norm": 1.6206820011138916,
"learning_rate": 0.0002819155778894472,
"loss": 1.6379,
"step": 6500
},
{
"epoch": 0.33,
"grad_norm": 0.9906002879142761,
"learning_rate": 0.0002816140703517588,
"loss": 1.6905,
"step": 6600
},
{
"epoch": 0.34,
"grad_norm": 1.8640550374984741,
"learning_rate": 0.00028131256281407034,
"loss": 1.6842,
"step": 6700
},
{
"epoch": 0.34,
"grad_norm": 1.5478594303131104,
"learning_rate": 0.00028101407035175876,
"loss": 1.6554,
"step": 6800
},
{
"epoch": 0.34,
"grad_norm": 1.2689837217330933,
"learning_rate": 0.0002807125628140703,
"loss": 1.6775,
"step": 6900
},
{
"epoch": 0.35,
"grad_norm": 1.8730539083480835,
"learning_rate": 0.0002804110552763819,
"loss": 1.6603,
"step": 7000
},
{
"epoch": 0.35,
"eval_loss": 1.6557646989822388,
"eval_runtime": 37.4609,
"eval_samples_per_second": 26.694,
"eval_steps_per_second": 3.337,
"step": 7000
},
{
"epoch": 0.35,
"grad_norm": 1.455672025680542,
"learning_rate": 0.00028010954773869344,
"loss": 1.6635,
"step": 7100
},
{
"epoch": 0.36,
"grad_norm": 1.6935358047485352,
"learning_rate": 0.000279808040201005,
"loss": 1.6796,
"step": 7200
},
{
"epoch": 0.36,
"grad_norm": 2.0117626190185547,
"learning_rate": 0.00027950653266331656,
"loss": 1.6972,
"step": 7300
},
{
"epoch": 0.37,
"grad_norm": 0.9937806129455566,
"learning_rate": 0.0002792050251256281,
"loss": 1.6666,
"step": 7400
},
{
"epoch": 0.38,
"grad_norm": 1.1981546878814697,
"learning_rate": 0.0002789035175879397,
"loss": 1.665,
"step": 7500
},
{
"epoch": 0.38,
"grad_norm": 2.0641427040100098,
"learning_rate": 0.00027860201005025124,
"loss": 1.6535,
"step": 7600
},
{
"epoch": 0.39,
"grad_norm": 1.7236799001693726,
"learning_rate": 0.0002783005025125628,
"loss": 1.676,
"step": 7700
},
{
"epoch": 0.39,
"grad_norm": 3.5143849849700928,
"learning_rate": 0.00027799899497487436,
"loss": 1.6521,
"step": 7800
},
{
"epoch": 0.4,
"grad_norm": 1.7787047624588013,
"learning_rate": 0.0002776974874371859,
"loss": 1.6637,
"step": 7900
},
{
"epoch": 0.4,
"grad_norm": 1.713392972946167,
"learning_rate": 0.0002773959798994975,
"loss": 1.6704,
"step": 8000
},
{
"epoch": 0.4,
"eval_loss": 1.656675100326538,
"eval_runtime": 37.9526,
"eval_samples_per_second": 26.349,
"eval_steps_per_second": 3.294,
"step": 8000
},
{
"epoch": 0.41,
"grad_norm": 1.1325870752334595,
"learning_rate": 0.00027709447236180904,
"loss": 1.6118,
"step": 8100
},
{
"epoch": 0.41,
"grad_norm": 1.746856689453125,
"learning_rate": 0.0002767929648241206,
"loss": 1.6718,
"step": 8200
},
{
"epoch": 0.41,
"grad_norm": 1.4181280136108398,
"learning_rate": 0.0002764914572864321,
"loss": 1.6957,
"step": 8300
},
{
"epoch": 0.42,
"grad_norm": 2.078387975692749,
"learning_rate": 0.0002761899497487437,
"loss": 1.6546,
"step": 8400
},
{
"epoch": 0.42,
"grad_norm": 2.694249153137207,
"learning_rate": 0.0002758884422110553,
"loss": 1.6875,
"step": 8500
},
{
"epoch": 0.43,
"grad_norm": 2.530956745147705,
"learning_rate": 0.00027558693467336683,
"loss": 1.6112,
"step": 8600
},
{
"epoch": 0.43,
"grad_norm": 1.4322137832641602,
"learning_rate": 0.00027528542713567834,
"loss": 1.6572,
"step": 8700
},
{
"epoch": 0.44,
"grad_norm": 1.996591329574585,
"learning_rate": 0.0002749839195979899,
"loss": 1.6442,
"step": 8800
},
{
"epoch": 0.45,
"grad_norm": 1.6497749090194702,
"learning_rate": 0.0002746824120603015,
"loss": 1.6404,
"step": 8900
},
{
"epoch": 0.45,
"grad_norm": 2.429196357727051,
"learning_rate": 0.000274380904522613,
"loss": 1.656,
"step": 9000
},
{
"epoch": 0.45,
"eval_loss": 1.6714575290679932,
"eval_runtime": 37.5615,
"eval_samples_per_second": 26.623,
"eval_steps_per_second": 3.328,
"step": 9000
},
{
"epoch": 0.46,
"grad_norm": 1.9687163829803467,
"learning_rate": 0.0002740824120603015,
"loss": 1.6644,
"step": 9100
},
{
"epoch": 0.46,
"grad_norm": 1.4858875274658203,
"learning_rate": 0.00027378090452261306,
"loss": 1.6434,
"step": 9200
},
{
"epoch": 0.47,
"grad_norm": 1.8358262777328491,
"learning_rate": 0.0002734793969849246,
"loss": 1.6565,
"step": 9300
},
{
"epoch": 0.47,
"grad_norm": 1.3822225332260132,
"learning_rate": 0.0002731778894472361,
"loss": 1.6506,
"step": 9400
},
{
"epoch": 0.47,
"grad_norm": 1.9319818019866943,
"learning_rate": 0.00027287638190954774,
"loss": 1.6256,
"step": 9500
},
{
"epoch": 0.48,
"grad_norm": 2.601515293121338,
"learning_rate": 0.0002725748743718593,
"loss": 1.6243,
"step": 9600
},
{
"epoch": 0.48,
"grad_norm": 1.610561490058899,
"learning_rate": 0.0002722733668341708,
"loss": 1.6259,
"step": 9700
},
{
"epoch": 0.49,
"grad_norm": 2.060863494873047,
"learning_rate": 0.00027197185929648236,
"loss": 1.655,
"step": 9800
},
{
"epoch": 0.49,
"grad_norm": 2.7591397762298584,
"learning_rate": 0.000271670351758794,
"loss": 1.6342,
"step": 9900
},
{
"epoch": 0.5,
"grad_norm": 1.2489566802978516,
"learning_rate": 0.00027136884422110553,
"loss": 1.6431,
"step": 10000
},
{
"epoch": 0.5,
"eval_loss": 1.6347763538360596,
"eval_runtime": 37.5252,
"eval_samples_per_second": 26.649,
"eval_steps_per_second": 3.331,
"step": 10000
},
{
"epoch": 0.51,
"grad_norm": 1.2953132390975952,
"learning_rate": 0.00027106733668341704,
"loss": 1.6872,
"step": 10100
},
{
"epoch": 0.51,
"grad_norm": 1.9919564723968506,
"learning_rate": 0.0002707658291457286,
"loss": 1.6576,
"step": 10200
},
{
"epoch": 0.52,
"grad_norm": 1.3081834316253662,
"learning_rate": 0.0002704643216080402,
"loss": 1.6354,
"step": 10300
},
{
"epoch": 0.52,
"grad_norm": 1.605245590209961,
"learning_rate": 0.0002701628140703517,
"loss": 1.6687,
"step": 10400
},
{
"epoch": 0.53,
"grad_norm": 2.1541988849639893,
"learning_rate": 0.0002698613065326633,
"loss": 1.6631,
"step": 10500
},
{
"epoch": 0.53,
"grad_norm": 1.8258408308029175,
"learning_rate": 0.00026955979899497484,
"loss": 1.6633,
"step": 10600
},
{
"epoch": 0.54,
"grad_norm": 1.3377407789230347,
"learning_rate": 0.00026925829145728645,
"loss": 1.6544,
"step": 10700
},
{
"epoch": 0.54,
"grad_norm": 0.998458206653595,
"learning_rate": 0.00026895678391959795,
"loss": 1.6126,
"step": 10800
},
{
"epoch": 0.55,
"grad_norm": 1.3561229705810547,
"learning_rate": 0.0002686552763819095,
"loss": 1.6614,
"step": 10900
},
{
"epoch": 0.55,
"grad_norm": 2.4729514122009277,
"learning_rate": 0.0002683537688442211,
"loss": 1.6439,
"step": 11000
},
{
"epoch": 0.55,
"eval_loss": 1.648992657661438,
"eval_runtime": 37.818,
"eval_samples_per_second": 26.442,
"eval_steps_per_second": 3.305,
"step": 11000
},
{
"epoch": 0.56,
"grad_norm": 1.2856806516647339,
"learning_rate": 0.00026805226130653263,
"loss": 1.605,
"step": 11100
},
{
"epoch": 0.56,
"grad_norm": 1.7708286046981812,
"learning_rate": 0.0002677507537688442,
"loss": 1.6257,
"step": 11200
},
{
"epoch": 0.56,
"grad_norm": 2.085149049758911,
"learning_rate": 0.00026744924623115575,
"loss": 1.6347,
"step": 11300
},
{
"epoch": 0.57,
"grad_norm": 0.9750702977180481,
"learning_rate": 0.0002671477386934673,
"loss": 1.6496,
"step": 11400
},
{
"epoch": 0.57,
"grad_norm": 1.9253658056259155,
"learning_rate": 0.00026684623115577887,
"loss": 1.6395,
"step": 11500
},
{
"epoch": 0.58,
"grad_norm": 2.441312313079834,
"learning_rate": 0.00026654472361809043,
"loss": 1.6444,
"step": 11600
},
{
"epoch": 0.58,
"grad_norm": 1.259020447731018,
"learning_rate": 0.000266243216080402,
"loss": 1.6114,
"step": 11700
},
{
"epoch": 0.59,
"grad_norm": 1.3337092399597168,
"learning_rate": 0.00026594170854271355,
"loss": 1.6243,
"step": 11800
},
{
"epoch": 0.59,
"grad_norm": 1.423687219619751,
"learning_rate": 0.0002656402010050251,
"loss": 1.623,
"step": 11900
},
{
"epoch": 0.6,
"grad_norm": 1.2547069787979126,
"learning_rate": 0.00026533869346733667,
"loss": 1.631,
"step": 12000
},
{
"epoch": 0.6,
"eval_loss": 1.6530547142028809,
"eval_runtime": 37.4898,
"eval_samples_per_second": 26.674,
"eval_steps_per_second": 3.334,
"step": 12000
},
{
"epoch": 0.6,
"grad_norm": 1.1078051328659058,
"learning_rate": 0.00026503718592964823,
"loss": 1.6229,
"step": 12100
},
{
"epoch": 0.61,
"grad_norm": 1.2142789363861084,
"learning_rate": 0.0002647356783919598,
"loss": 1.6251,
"step": 12200
},
{
"epoch": 0.61,
"grad_norm": 1.5933152437210083,
"learning_rate": 0.00026443417085427135,
"loss": 1.627,
"step": 12300
},
{
"epoch": 0.62,
"grad_norm": 1.6759315729141235,
"learning_rate": 0.0002641326633165829,
"loss": 1.6294,
"step": 12400
},
{
"epoch": 0.62,
"grad_norm": 1.487029790878296,
"learning_rate": 0.00026383115577889447,
"loss": 1.6406,
"step": 12500
},
{
"epoch": 0.63,
"grad_norm": 2.004643201828003,
"learning_rate": 0.000263529648241206,
"loss": 1.6342,
"step": 12600
},
{
"epoch": 0.64,
"grad_norm": 1.94633948802948,
"learning_rate": 0.00026322814070351753,
"loss": 1.6279,
"step": 12700
},
{
"epoch": 0.64,
"grad_norm": 1.5213886499404907,
"learning_rate": 0.00026292663316582914,
"loss": 1.6023,
"step": 12800
},
{
"epoch": 0.65,
"grad_norm": 1.4710832834243774,
"learning_rate": 0.0002626251256281407,
"loss": 1.6202,
"step": 12900
},
{
"epoch": 0.65,
"grad_norm": 2.367037296295166,
"learning_rate": 0.00026232361809045226,
"loss": 1.6463,
"step": 13000
},
{
"epoch": 0.65,
"eval_loss": 1.6226933002471924,
"eval_runtime": 37.4429,
"eval_samples_per_second": 26.707,
"eval_steps_per_second": 3.338,
"step": 13000
},
{
"epoch": 0.66,
"grad_norm": 2.803264856338501,
"learning_rate": 0.0002620251256281407,
"loss": 1.6219,
"step": 13100
},
{
"epoch": 0.66,
"grad_norm": 1.5915918350219727,
"learning_rate": 0.00026172361809045225,
"loss": 1.6252,
"step": 13200
},
{
"epoch": 0.67,
"grad_norm": 1.5102565288543701,
"learning_rate": 0.0002614221105527638,
"loss": 1.6346,
"step": 13300
},
{
"epoch": 0.67,
"grad_norm": 1.8278179168701172,
"learning_rate": 0.00026112060301507537,
"loss": 1.6386,
"step": 13400
},
{
"epoch": 0.68,
"grad_norm": 1.6756057739257812,
"learning_rate": 0.0002608190954773869,
"loss": 1.6135,
"step": 13500
},
{
"epoch": 0.68,
"grad_norm": 1.4984675645828247,
"learning_rate": 0.0002605175879396985,
"loss": 1.6316,
"step": 13600
},
{
"epoch": 0.69,
"grad_norm": 2.2536373138427734,
"learning_rate": 0.00026021608040201005,
"loss": 1.6317,
"step": 13700
},
{
"epoch": 0.69,
"grad_norm": 1.3149595260620117,
"learning_rate": 0.0002599145728643216,
"loss": 1.6311,
"step": 13800
},
{
"epoch": 0.69,
"grad_norm": 1.662287712097168,
"learning_rate": 0.00025961306532663316,
"loss": 1.6229,
"step": 13900
},
{
"epoch": 0.7,
"grad_norm": 1.2096275091171265,
"learning_rate": 0.0002593115577889447,
"loss": 1.6305,
"step": 14000
},
{
"epoch": 0.7,
"eval_loss": 1.632125735282898,
"eval_runtime": 37.4575,
"eval_samples_per_second": 26.697,
"eval_steps_per_second": 3.337,
"step": 14000
},
{
"epoch": 0.7,
"grad_norm": 2.01643705368042,
"learning_rate": 0.00025901005025125623,
"loss": 1.6271,
"step": 14100
},
{
"epoch": 0.71,
"grad_norm": 1.2067662477493286,
"learning_rate": 0.0002587085427135678,
"loss": 1.6082,
"step": 14200
},
{
"epoch": 0.71,
"grad_norm": 2.0471389293670654,
"learning_rate": 0.0002584070351758794,
"loss": 1.6346,
"step": 14300
},
{
"epoch": 0.72,
"grad_norm": 1.1723861694335938,
"learning_rate": 0.0002581055276381909,
"loss": 1.6449,
"step": 14400
},
{
"epoch": 0.72,
"grad_norm": 1.3024895191192627,
"learning_rate": 0.00025780402010050247,
"loss": 1.6159,
"step": 14500
},
{
"epoch": 0.73,
"grad_norm": 1.6715235710144043,
"learning_rate": 0.000257502512562814,
"loss": 1.6096,
"step": 14600
},
{
"epoch": 0.73,
"grad_norm": 2.116154432296753,
"learning_rate": 0.00025720100502512564,
"loss": 1.5984,
"step": 14700
},
{
"epoch": 0.74,
"grad_norm": 1.2046904563903809,
"learning_rate": 0.00025689949748743714,
"loss": 1.6125,
"step": 14800
},
{
"epoch": 0.74,
"grad_norm": 1.5058480501174927,
"learning_rate": 0.0002565979899497487,
"loss": 1.6164,
"step": 14900
},
{
"epoch": 0.75,
"grad_norm": 2.2814691066741943,
"learning_rate": 0.00025629648241206026,
"loss": 1.6312,
"step": 15000
},
{
"epoch": 0.75,
"eval_loss": 1.6935200691223145,
"eval_runtime": 37.4213,
"eval_samples_per_second": 26.723,
"eval_steps_per_second": 3.34,
"step": 15000
},
{
"epoch": 0.76,
"grad_norm": 1.5283032655715942,
"learning_rate": 0.0002559979899497487,
"loss": 1.6201,
"step": 15100
},
{
"epoch": 0.76,
"grad_norm": 2.6960535049438477,
"learning_rate": 0.00025569648241206025,
"loss": 1.6062,
"step": 15200
},
{
"epoch": 0.77,
"grad_norm": 1.080701470375061,
"learning_rate": 0.00025539798994974873,
"loss": 1.622,
"step": 15300
},
{
"epoch": 0.77,
"grad_norm": 1.6446688175201416,
"learning_rate": 0.0002550964824120603,
"loss": 1.6075,
"step": 15400
},
{
"epoch": 0.78,
"grad_norm": 1.9143463373184204,
"learning_rate": 0.00025479497487437185,
"loss": 1.6363,
"step": 15500
},
{
"epoch": 0.78,
"grad_norm": 1.6148111820220947,
"learning_rate": 0.0002544934673366834,
"loss": 1.6367,
"step": 15600
},
{
"epoch": 0.79,
"grad_norm": 1.7216590642929077,
"learning_rate": 0.00025419195979899497,
"loss": 1.5997,
"step": 15700
},
{
"epoch": 0.79,
"grad_norm": 1.442865014076233,
"learning_rate": 0.00025389045226130647,
"loss": 1.5776,
"step": 15800
},
{
"epoch": 0.8,
"grad_norm": 0.9140738844871521,
"learning_rate": 0.0002535889447236181,
"loss": 1.6275,
"step": 15900
},
{
"epoch": 0.8,
"grad_norm": 2.099306583404541,
"learning_rate": 0.00025328743718592965,
"loss": 1.6072,
"step": 16000
},
{
"epoch": 0.8,
"eval_loss": 1.5524722337722778,
"eval_runtime": 37.4031,
"eval_samples_per_second": 26.736,
"eval_steps_per_second": 3.342,
"step": 16000
},
{
"epoch": 0.81,
"grad_norm": 1.560035228729248,
"learning_rate": 0.0002529859296482412,
"loss": 1.5983,
"step": 16100
},
{
"epoch": 0.81,
"grad_norm": 1.8614755868911743,
"learning_rate": 0.0002526844221105527,
"loss": 1.6271,
"step": 16200
},
{
"epoch": 0.81,
"grad_norm": 1.7034022808074951,
"learning_rate": 0.0002523829145728643,
"loss": 1.6238,
"step": 16300
},
{
"epoch": 0.82,
"grad_norm": 1.086572527885437,
"learning_rate": 0.0002520814070351759,
"loss": 1.6241,
"step": 16400
},
{
"epoch": 0.82,
"grad_norm": 0.8860014081001282,
"learning_rate": 0.0002517798994974874,
"loss": 1.562,
"step": 16500
},
{
"epoch": 0.83,
"grad_norm": 1.7819429636001587,
"learning_rate": 0.00025147839195979895,
"loss": 1.5995,
"step": 16600
},
{
"epoch": 0.83,
"grad_norm": 3.512892246246338,
"learning_rate": 0.00025117688442211056,
"loss": 1.5866,
"step": 16700
},
{
"epoch": 0.84,
"grad_norm": 1.5962600708007812,
"learning_rate": 0.0002508753768844221,
"loss": 1.6358,
"step": 16800
},
{
"epoch": 0.84,
"grad_norm": 1.7714247703552246,
"learning_rate": 0.0002505738693467336,
"loss": 1.6043,
"step": 16900
},
{
"epoch": 0.85,
"grad_norm": 1.3199384212493896,
"learning_rate": 0.0002502723618090452,
"loss": 1.6102,
"step": 17000
},
{
"epoch": 0.85,
"eval_loss": 1.577386736869812,
"eval_runtime": 37.4173,
"eval_samples_per_second": 26.726,
"eval_steps_per_second": 3.341,
"step": 17000
},
{
"epoch": 0.85,
"grad_norm": 1.777269721031189,
"learning_rate": 0.00024997085427135675,
"loss": 1.6064,
"step": 17100
},
{
"epoch": 0.86,
"grad_norm": 2.2964231967926025,
"learning_rate": 0.0002496693467336683,
"loss": 1.5964,
"step": 17200
},
{
"epoch": 0.86,
"grad_norm": 1.7127012014389038,
"learning_rate": 0.00024936783919597986,
"loss": 1.6334,
"step": 17300
},
{
"epoch": 0.87,
"grad_norm": 1.442181944847107,
"learning_rate": 0.0002490663316582914,
"loss": 1.5679,
"step": 17400
},
{
"epoch": 0.88,
"grad_norm": 1.7092599868774414,
"learning_rate": 0.000248764824120603,
"loss": 1.6125,
"step": 17500
},
{
"epoch": 0.88,
"grad_norm": 1.8017587661743164,
"learning_rate": 0.00024846331658291454,
"loss": 1.6386,
"step": 17600
},
{
"epoch": 0.89,
"grad_norm": 3.3435380458831787,
"learning_rate": 0.0002481618090452261,
"loss": 1.5632,
"step": 17700
},
{
"epoch": 0.89,
"grad_norm": 1.2036772966384888,
"learning_rate": 0.00024786030150753766,
"loss": 1.5972,
"step": 17800
},
{
"epoch": 0.9,
"grad_norm": 2.212369203567505,
"learning_rate": 0.0002475587939698492,
"loss": 1.5777,
"step": 17900
},
{
"epoch": 0.9,
"grad_norm": 1.9149020910263062,
"learning_rate": 0.0002472572864321608,
"loss": 1.601,
"step": 18000
},
{
"epoch": 0.9,
"eval_loss": 1.5726821422576904,
"eval_runtime": 37.4487,
"eval_samples_per_second": 26.703,
"eval_steps_per_second": 3.338,
"step": 18000
},
{
"epoch": 0.91,
"grad_norm": 2.2173101902008057,
"learning_rate": 0.00024695577889447234,
"loss": 1.5975,
"step": 18100
},
{
"epoch": 0.91,
"grad_norm": 2.0151243209838867,
"learning_rate": 0.0002466542713567839,
"loss": 1.5975,
"step": 18200
},
{
"epoch": 0.92,
"grad_norm": 1.6275769472122192,
"learning_rate": 0.00024635276381909546,
"loss": 1.5909,
"step": 18300
},
{
"epoch": 0.92,
"grad_norm": 0.8248587846755981,
"learning_rate": 0.000246051256281407,
"loss": 1.5954,
"step": 18400
},
{
"epoch": 0.93,
"grad_norm": 2.119255542755127,
"learning_rate": 0.0002457497487437186,
"loss": 1.5651,
"step": 18500
},
{
"epoch": 0.93,
"grad_norm": 1.9526431560516357,
"learning_rate": 0.00024544824120603014,
"loss": 1.6184,
"step": 18600
},
{
"epoch": 0.94,
"grad_norm": 2.407723903656006,
"learning_rate": 0.0002451467336683417,
"loss": 1.6181,
"step": 18700
},
{
"epoch": 0.94,
"grad_norm": 1.526853084564209,
"learning_rate": 0.00024484522613065326,
"loss": 1.6172,
"step": 18800
},
{
"epoch": 0.94,
"grad_norm": 1.4161404371261597,
"learning_rate": 0.0002445437185929648,
"loss": 1.6154,
"step": 18900
},
{
"epoch": 0.95,
"grad_norm": 1.3028178215026855,
"learning_rate": 0.0002442422110552764,
"loss": 1.6053,
"step": 19000
},
{
"epoch": 0.95,
"eval_loss": 1.5818778276443481,
"eval_runtime": 37.4111,
"eval_samples_per_second": 26.73,
"eval_steps_per_second": 3.341,
"step": 19000
},
{
"epoch": 0.95,
"grad_norm": 1.6454105377197266,
"learning_rate": 0.0002439407035175879,
"loss": 1.6149,
"step": 19100
},
{
"epoch": 0.96,
"grad_norm": 1.664665699005127,
"learning_rate": 0.00024364221105527636,
"loss": 1.6178,
"step": 19200
},
{
"epoch": 0.96,
"grad_norm": 1.53481125831604,
"learning_rate": 0.00024334070351758792,
"loss": 1.5858,
"step": 19300
},
{
"epoch": 0.97,
"grad_norm": 6.695281982421875,
"learning_rate": 0.00024303919597989948,
"loss": 1.5717,
"step": 19400
},
{
"epoch": 0.97,
"grad_norm": 1.6920980215072632,
"learning_rate": 0.00024273768844221104,
"loss": 1.5754,
"step": 19500
},
{
"epoch": 0.98,
"grad_norm": 2.0135791301727295,
"learning_rate": 0.0002424361809045226,
"loss": 1.6219,
"step": 19600
},
{
"epoch": 0.98,
"grad_norm": 1.3980337381362915,
"learning_rate": 0.00024213467336683413,
"loss": 1.5913,
"step": 19700
},
{
"epoch": 0.99,
"grad_norm": 1.8416180610656738,
"learning_rate": 0.0002418331658291457,
"loss": 1.5691,
"step": 19800
},
{
"epoch": 0.99,
"grad_norm": 1.366356372833252,
"learning_rate": 0.00024153165829145728,
"loss": 1.5804,
"step": 19900
},
{
"epoch": 1.0,
"grad_norm": 1.3517309427261353,
"learning_rate": 0.00024123015075376884,
"loss": 1.5603,
"step": 20000
},
{
"epoch": 1.0,
"eval_loss": 1.6314821243286133,
"eval_runtime": 37.4553,
"eval_samples_per_second": 26.698,
"eval_steps_per_second": 3.337,
"step": 20000
},
{
"epoch": 1.0,
"grad_norm": 1.8194776773452759,
"learning_rate": 0.00024092864321608037,
"loss": 1.5964,
"step": 20100
},
{
"epoch": 1.01,
"grad_norm": 3.187936305999756,
"learning_rate": 0.00024062713567839193,
"loss": 1.5926,
"step": 20200
},
{
"epoch": 1.01,
"grad_norm": 1.2052127122879028,
"learning_rate": 0.00024032562814070351,
"loss": 1.5727,
"step": 20300
},
{
"epoch": 1.02,
"grad_norm": 1.6609078645706177,
"learning_rate": 0.00024002412060301505,
"loss": 1.5967,
"step": 20400
},
{
"epoch": 1.02,
"grad_norm": 1.502246379852295,
"learning_rate": 0.0002397226130653266,
"loss": 1.5708,
"step": 20500
},
{
"epoch": 1.03,
"grad_norm": 2.180826425552368,
"learning_rate": 0.00023942110552763817,
"loss": 1.5688,
"step": 20600
},
{
"epoch": 1.03,
"grad_norm": 2.257434606552124,
"learning_rate": 0.00023911959798994975,
"loss": 1.5263,
"step": 20700
},
{
"epoch": 1.04,
"grad_norm": 1.9630309343338013,
"learning_rate": 0.00023881809045226128,
"loss": 1.5776,
"step": 20800
},
{
"epoch": 1.04,
"grad_norm": 2.248621940612793,
"learning_rate": 0.00023851658291457284,
"loss": 1.6,
"step": 20900
},
{
"epoch": 1.05,
"grad_norm": 1.0489450693130493,
"learning_rate": 0.0002382150753768844,
"loss": 1.5699,
"step": 21000
},
{
"epoch": 1.05,
"eval_loss": 1.5522246360778809,
"eval_runtime": 37.4533,
"eval_samples_per_second": 26.7,
"eval_steps_per_second": 3.337,
"step": 21000
},
{
"epoch": 1.05,
"grad_norm": 1.457306981086731,
"learning_rate": 0.00023791356783919596,
"loss": 1.5953,
"step": 21100
},
{
"epoch": 1.06,
"grad_norm": 2.088200807571411,
"learning_rate": 0.00023761206030150752,
"loss": 1.5881,
"step": 21200
},
{
"epoch": 1.06,
"grad_norm": 1.4078480005264282,
"learning_rate": 0.00023731356783919598,
"loss": 1.5746,
"step": 21300
},
{
"epoch": 1.07,
"grad_norm": 1.5498270988464355,
"learning_rate": 0.00023701206030150753,
"loss": 1.5878,
"step": 21400
},
{
"epoch": 1.07,
"grad_norm": 1.8573285341262817,
"learning_rate": 0.00023671055276381907,
"loss": 1.574,
"step": 21500
},
{
"epoch": 1.08,
"grad_norm": 1.3603477478027344,
"learning_rate": 0.00023640904522613063,
"loss": 1.5994,
"step": 21600
},
{
"epoch": 1.08,
"grad_norm": 1.6878479719161987,
"learning_rate": 0.0002361075376884422,
"loss": 1.63,
"step": 21700
},
{
"epoch": 1.09,
"grad_norm": 1.879296064376831,
"learning_rate": 0.00023580603015075375,
"loss": 1.6313,
"step": 21800
},
{
"epoch": 1.09,
"grad_norm": 1.695983648300171,
"learning_rate": 0.0002355045226130653,
"loss": 1.5837,
"step": 21900
},
{
"epoch": 1.1,
"grad_norm": 2.6149425506591797,
"learning_rate": 0.00023520301507537686,
"loss": 1.5967,
"step": 22000
},
{
"epoch": 1.1,
"eval_loss": 1.6031874418258667,
"eval_runtime": 37.522,
"eval_samples_per_second": 26.651,
"eval_steps_per_second": 3.331,
"step": 22000
},
{
"epoch": 1.1,
"grad_norm": 2.0208663940429688,
"learning_rate": 0.00023490150753768845,
"loss": 1.5691,
"step": 22100
},
{
"epoch": 1.11,
"grad_norm": 2.202256202697754,
"learning_rate": 0.00023459999999999998,
"loss": 1.5815,
"step": 22200
},
{
"epoch": 1.11,
"grad_norm": 1.9692652225494385,
"learning_rate": 0.00023429849246231154,
"loss": 1.5854,
"step": 22300
},
{
"epoch": 1.12,
"grad_norm": 1.9030089378356934,
"learning_rate": 0.00023399698492462308,
"loss": 1.5937,
"step": 22400
},
{
"epoch": 1.12,
"grad_norm": 2.265805244445801,
"learning_rate": 0.00023369547738693463,
"loss": 1.5409,
"step": 22500
},
{
"epoch": 1.13,
"grad_norm": 1.9824811220169067,
"learning_rate": 0.00023339396984924622,
"loss": 1.5977,
"step": 22600
},
{
"epoch": 1.14,
"grad_norm": 1.3028334379196167,
"learning_rate": 0.00023309246231155778,
"loss": 1.573,
"step": 22700
},
{
"epoch": 1.14,
"grad_norm": 2.472731590270996,
"learning_rate": 0.0002327909547738693,
"loss": 1.5962,
"step": 22800
},
{
"epoch": 1.15,
"grad_norm": 1.1267619132995605,
"learning_rate": 0.00023248944723618087,
"loss": 1.5871,
"step": 22900
},
{
"epoch": 1.15,
"grad_norm": 1.7546107769012451,
"learning_rate": 0.00023218793969849246,
"loss": 1.6053,
"step": 23000
},
{
"epoch": 1.15,
"eval_loss": 1.5985630750656128,
"eval_runtime": 37.3462,
"eval_samples_per_second": 26.777,
"eval_steps_per_second": 3.347,
"step": 23000
},
{
"epoch": 1.16,
"grad_norm": 1.7325044870376587,
"learning_rate": 0.000231886432160804,
"loss": 1.557,
"step": 23100
},
{
"epoch": 1.16,
"grad_norm": 1.771462321281433,
"learning_rate": 0.00023158492462311555,
"loss": 1.6102,
"step": 23200
},
{
"epoch": 1.17,
"grad_norm": 1.077540397644043,
"learning_rate": 0.0002312834170854271,
"loss": 1.5607,
"step": 23300
},
{
"epoch": 1.17,
"grad_norm": 1.2537496089935303,
"learning_rate": 0.0002309819095477387,
"loss": 1.5801,
"step": 23400
},
{
"epoch": 1.18,
"grad_norm": 1.5025357007980347,
"learning_rate": 0.00023068040201005023,
"loss": 1.583,
"step": 23500
},
{
"epoch": 1.18,
"grad_norm": 1.2420939207077026,
"learning_rate": 0.0002303788944723618,
"loss": 1.569,
"step": 23600
},
{
"epoch": 1.19,
"grad_norm": 2.900120735168457,
"learning_rate": 0.00023007738693467335,
"loss": 1.5561,
"step": 23700
},
{
"epoch": 1.19,
"grad_norm": 1.6182540655136108,
"learning_rate": 0.0002297758793969849,
"loss": 1.5673,
"step": 23800
},
{
"epoch": 1.2,
"grad_norm": 1.5824536085128784,
"learning_rate": 0.00022947437185929647,
"loss": 1.5603,
"step": 23900
},
{
"epoch": 1.2,
"grad_norm": 1.1365340948104858,
"learning_rate": 0.00022917286432160803,
"loss": 1.571,
"step": 24000
},
{
"epoch": 1.2,
"eval_loss": 1.5782357454299927,
"eval_runtime": 37.4619,
"eval_samples_per_second": 26.694,
"eval_steps_per_second": 3.337,
"step": 24000
},
{
"epoch": 1.21,
"grad_norm": 1.8765733242034912,
"learning_rate": 0.00022887135678391956,
"loss": 1.5628,
"step": 24100
},
{
"epoch": 1.21,
"grad_norm": 2.7541840076446533,
"learning_rate": 0.00022856984924623115,
"loss": 1.5967,
"step": 24200
},
{
"epoch": 1.22,
"grad_norm": 1.2741538286209106,
"learning_rate": 0.0002282683417085427,
"loss": 1.5142,
"step": 24300
},
{
"epoch": 1.22,
"grad_norm": 1.5907307863235474,
"learning_rate": 0.00022796683417085426,
"loss": 1.5685,
"step": 24400
},
{
"epoch": 1.23,
"grad_norm": 2.187331438064575,
"learning_rate": 0.0002276653266331658,
"loss": 1.5788,
"step": 24500
},
{
"epoch": 1.23,
"grad_norm": 2.0679118633270264,
"learning_rate": 0.00022736381909547736,
"loss": 1.5585,
"step": 24600
},
{
"epoch": 1.23,
"grad_norm": 1.2467107772827148,
"learning_rate": 0.00022706231155778894,
"loss": 1.5531,
"step": 24700
},
{
"epoch": 1.24,
"grad_norm": 1.334625005722046,
"learning_rate": 0.00022676080402010047,
"loss": 1.5831,
"step": 24800
},
{
"epoch": 1.25,
"grad_norm": 1.466030240058899,
"learning_rate": 0.00022645929648241203,
"loss": 1.5753,
"step": 24900
},
{
"epoch": 1.25,
"grad_norm": 2.8223164081573486,
"learning_rate": 0.0002261577889447236,
"loss": 1.6131,
"step": 25000
},
{
"epoch": 1.25,
"eval_loss": 1.5768604278564453,
"eval_runtime": 37.5229,
"eval_samples_per_second": 26.65,
"eval_steps_per_second": 3.331,
"step": 25000
},
{
"epoch": 1.25,
"grad_norm": 1.1590194702148438,
"learning_rate": 0.00022585628140703518,
"loss": 1.5831,
"step": 25100
},
{
"epoch": 1.26,
"grad_norm": 1.1502777338027954,
"learning_rate": 0.00022555778894472358,
"loss": 1.5673,
"step": 25200
},
{
"epoch": 1.27,
"grad_norm": 1.8671566247940063,
"learning_rate": 0.00022525628140703517,
"loss": 1.5632,
"step": 25300
},
{
"epoch": 1.27,
"grad_norm": 2.4084360599517822,
"learning_rate": 0.00022495477386934673,
"loss": 1.5727,
"step": 25400
},
{
"epoch": 1.27,
"grad_norm": 1.3032163381576538,
"learning_rate": 0.00022465326633165826,
"loss": 1.5754,
"step": 25500
},
{
"epoch": 1.28,
"grad_norm": 2.2420716285705566,
"learning_rate": 0.00022435175879396982,
"loss": 1.5802,
"step": 25600
},
{
"epoch": 1.28,
"grad_norm": 1.3116205930709839,
"learning_rate": 0.0002240502512562814,
"loss": 1.5744,
"step": 25700
},
{
"epoch": 1.29,
"grad_norm": 1.0008848905563354,
"learning_rate": 0.00022374874371859294,
"loss": 1.5423,
"step": 25800
},
{
"epoch": 1.29,
"grad_norm": 2.07833194732666,
"learning_rate": 0.0002234472361809045,
"loss": 1.5657,
"step": 25900
},
{
"epoch": 1.3,
"grad_norm": 1.894939661026001,
"learning_rate": 0.00022314572864321605,
"loss": 1.5464,
"step": 26000
},
{
"epoch": 1.3,
"eval_loss": 1.568109154701233,
"eval_runtime": 37.5055,
"eval_samples_per_second": 26.663,
"eval_steps_per_second": 3.333,
"step": 26000
},
{
"epoch": 1.3,
"grad_norm": 2.3820385932922363,
"learning_rate": 0.00022284422110552764,
"loss": 1.562,
"step": 26100
},
{
"epoch": 1.31,
"grad_norm": 1.6219903230667114,
"learning_rate": 0.00022254271356783917,
"loss": 1.5683,
"step": 26200
},
{
"epoch": 1.31,
"grad_norm": 1.721933126449585,
"learning_rate": 0.00022224120603015073,
"loss": 1.5646,
"step": 26300
},
{
"epoch": 1.32,
"grad_norm": 1.8133726119995117,
"learning_rate": 0.0002219396984924623,
"loss": 1.5574,
"step": 26400
},
{
"epoch": 1.32,
"grad_norm": 1.871902585029602,
"learning_rate": 0.00022163819095477385,
"loss": 1.5728,
"step": 26500
},
{
"epoch": 1.33,
"grad_norm": 2.7765748500823975,
"learning_rate": 0.0002213366834170854,
"loss": 1.5768,
"step": 26600
},
{
"epoch": 1.33,
"grad_norm": 1.848578691482544,
"learning_rate": 0.00022103819095477386,
"loss": 1.5721,
"step": 26700
},
{
"epoch": 1.34,
"grad_norm": 1.944535732269287,
"learning_rate": 0.00022073668341708542,
"loss": 1.5722,
"step": 26800
},
{
"epoch": 1.34,
"grad_norm": 1.5262278318405151,
"learning_rate": 0.00022043517587939696,
"loss": 1.5642,
"step": 26900
},
{
"epoch": 1.35,
"grad_norm": 2.425851821899414,
"learning_rate": 0.00022013366834170852,
"loss": 1.5441,
"step": 27000
},
{
"epoch": 1.35,
"eval_loss": 1.5537927150726318,
"eval_runtime": 37.516,
"eval_samples_per_second": 26.655,
"eval_steps_per_second": 3.332,
"step": 27000
},
{
"epoch": 1.35,
"grad_norm": 1.2145837545394897,
"learning_rate": 0.0002198321608040201,
"loss": 1.5994,
"step": 27100
},
{
"epoch": 1.36,
"grad_norm": 2.361107349395752,
"learning_rate": 0.00021953065326633163,
"loss": 1.5551,
"step": 27200
},
{
"epoch": 1.36,
"grad_norm": 1.1575865745544434,
"learning_rate": 0.0002192291457286432,
"loss": 1.5749,
"step": 27300
},
{
"epoch": 1.37,
"grad_norm": 1.6274192333221436,
"learning_rate": 0.00021892763819095475,
"loss": 1.5729,
"step": 27400
},
{
"epoch": 1.38,
"grad_norm": 1.660609245300293,
"learning_rate": 0.00021862613065326629,
"loss": 1.577,
"step": 27500
},
{
"epoch": 1.38,
"grad_norm": 2.8537380695343018,
"learning_rate": 0.00021832462311557787,
"loss": 1.5962,
"step": 27600
},
{
"epoch": 1.39,
"grad_norm": 7.63838529586792,
"learning_rate": 0.00021802311557788943,
"loss": 1.5462,
"step": 27700
},
{
"epoch": 1.39,
"grad_norm": 3.0265047550201416,
"learning_rate": 0.000217721608040201,
"loss": 1.6263,
"step": 27800
},
{
"epoch": 1.4,
"grad_norm": 1.552452802658081,
"learning_rate": 0.00021742010050251252,
"loss": 1.5683,
"step": 27900
},
{
"epoch": 1.4,
"grad_norm": 1.2861206531524658,
"learning_rate": 0.0002171185929648241,
"loss": 1.5776,
"step": 28000
},
{
"epoch": 1.4,
"eval_loss": 1.5543726682662964,
"eval_runtime": 37.5463,
"eval_samples_per_second": 26.634,
"eval_steps_per_second": 3.329,
"step": 28000
},
{
"epoch": 1.41,
"grad_norm": 2.410630941390991,
"learning_rate": 0.00021681708542713567,
"loss": 1.5513,
"step": 28100
},
{
"epoch": 1.41,
"grad_norm": 1.8043034076690674,
"learning_rate": 0.0002165155778894472,
"loss": 1.594,
"step": 28200
},
{
"epoch": 1.42,
"grad_norm": 1.8414337635040283,
"learning_rate": 0.00021621407035175876,
"loss": 1.5547,
"step": 28300
},
{
"epoch": 1.42,
"grad_norm": 1.9315450191497803,
"learning_rate": 0.00021591256281407035,
"loss": 1.5387,
"step": 28400
},
{
"epoch": 1.43,
"grad_norm": 2.3810365200042725,
"learning_rate": 0.0002156110552763819,
"loss": 1.564,
"step": 28500
},
{
"epoch": 1.43,
"grad_norm": 1.4747709035873413,
"learning_rate": 0.00021530954773869344,
"loss": 1.5325,
"step": 28600
},
{
"epoch": 1.44,
"grad_norm": 1.264954924583435,
"learning_rate": 0.000215008040201005,
"loss": 1.5716,
"step": 28700
},
{
"epoch": 1.44,
"grad_norm": 2.2243235111236572,
"learning_rate": 0.00021470653266331659,
"loss": 1.5388,
"step": 28800
},
{
"epoch": 1.45,
"grad_norm": 4.794454097747803,
"learning_rate": 0.00021440502512562812,
"loss": 1.5404,
"step": 28900
},
{
"epoch": 1.45,
"grad_norm": 2.5027246475219727,
"learning_rate": 0.00021410351758793968,
"loss": 1.5592,
"step": 29000
},
{
"epoch": 1.45,
"eval_loss": 1.569115400314331,
"eval_runtime": 37.5972,
"eval_samples_per_second": 26.598,
"eval_steps_per_second": 3.325,
"step": 29000
},
{
"epoch": 1.46,
"grad_norm": 2.8168818950653076,
"learning_rate": 0.00021380201005025124,
"loss": 1.5781,
"step": 29100
},
{
"epoch": 1.46,
"grad_norm": 1.4402812719345093,
"learning_rate": 0.0002135005025125628,
"loss": 1.5626,
"step": 29200
},
{
"epoch": 1.47,
"grad_norm": 1.5837584733963013,
"learning_rate": 0.00021319899497487436,
"loss": 1.6064,
"step": 29300
},
{
"epoch": 1.47,
"grad_norm": 1.435542106628418,
"learning_rate": 0.00021289748743718592,
"loss": 1.5621,
"step": 29400
},
{
"epoch": 1.48,
"grad_norm": 1.8227218389511108,
"learning_rate": 0.00021259597989949745,
"loss": 1.5481,
"step": 29500
},
{
"epoch": 1.48,
"grad_norm": 1.9331644773483276,
"learning_rate": 0.00021229447236180903,
"loss": 1.5811,
"step": 29600
},
{
"epoch": 1.48,
"grad_norm": 2.971740484237671,
"learning_rate": 0.0002119929648241206,
"loss": 1.5412,
"step": 29700
},
{
"epoch": 1.49,
"grad_norm": 3.569145917892456,
"learning_rate": 0.00021169145728643215,
"loss": 1.5508,
"step": 29800
},
{
"epoch": 1.5,
"grad_norm": 2.1709325313568115,
"learning_rate": 0.00021138994974874369,
"loss": 1.5802,
"step": 29900
},
{
"epoch": 1.5,
"grad_norm": 1.8602783679962158,
"learning_rate": 0.00021108844221105524,
"loss": 1.5607,
"step": 30000
},
{
"epoch": 1.5,
"eval_loss": 1.5550180673599243,
"eval_runtime": 37.5792,
"eval_samples_per_second": 26.61,
"eval_steps_per_second": 3.326,
"step": 30000
},
{
"epoch": 1.5,
"grad_norm": 1.3990256786346436,
"learning_rate": 0.00021078693467336683,
"loss": 1.5611,
"step": 30100
},
{
"epoch": 1.51,
"grad_norm": 4.223874568939209,
"learning_rate": 0.00021048542713567836,
"loss": 1.6039,
"step": 30200
},
{
"epoch": 1.52,
"grad_norm": 1.763484001159668,
"learning_rate": 0.00021018391959798992,
"loss": 1.5938,
"step": 30300
},
{
"epoch": 1.52,
"grad_norm": 2.2013938426971436,
"learning_rate": 0.00020988241206030148,
"loss": 1.5883,
"step": 30400
},
{
"epoch": 1.52,
"grad_norm": 2.446477174758911,
"learning_rate": 0.00020958391959798994,
"loss": 1.5289,
"step": 30500
},
{
"epoch": 1.53,
"grad_norm": 1.6682789325714111,
"learning_rate": 0.00020928241206030147,
"loss": 1.5689,
"step": 30600
},
{
"epoch": 1.54,
"grad_norm": 1.8999930620193481,
"learning_rate": 0.00020898090452261305,
"loss": 1.568,
"step": 30700
},
{
"epoch": 1.54,
"grad_norm": 2.189378261566162,
"learning_rate": 0.00020867939698492461,
"loss": 1.5363,
"step": 30800
},
{
"epoch": 1.54,
"grad_norm": 2.609349489212036,
"learning_rate": 0.00020837788944723615,
"loss": 1.5523,
"step": 30900
},
{
"epoch": 1.55,
"grad_norm": 2.2627103328704834,
"learning_rate": 0.0002080763819095477,
"loss": 1.574,
"step": 31000
},
{
"epoch": 1.55,
"eval_loss": 1.5256458520889282,
"eval_runtime": 37.5514,
"eval_samples_per_second": 26.63,
"eval_steps_per_second": 3.329,
"step": 31000
},
{
"epoch": 1.56,
"grad_norm": 1.9266563653945923,
"learning_rate": 0.0002077748743718593,
"loss": 1.5387,
"step": 31100
},
{
"epoch": 1.56,
"grad_norm": 1.8447223901748657,
"learning_rate": 0.00020747336683417085,
"loss": 1.5632,
"step": 31200
},
{
"epoch": 1.56,
"grad_norm": 1.375827670097351,
"learning_rate": 0.00020717185929648238,
"loss": 1.5494,
"step": 31300
},
{
"epoch": 1.57,
"grad_norm": 2.164782762527466,
"learning_rate": 0.00020687035175879394,
"loss": 1.5468,
"step": 31400
},
{
"epoch": 1.57,
"grad_norm": 1.186018943786621,
"learning_rate": 0.00020656884422110553,
"loss": 1.584,
"step": 31500
},
{
"epoch": 1.58,
"grad_norm": 5.939981460571289,
"learning_rate": 0.00020626733668341706,
"loss": 1.5602,
"step": 31600
},
{
"epoch": 1.58,
"grad_norm": 2.185728073120117,
"learning_rate": 0.00020596582914572862,
"loss": 1.5301,
"step": 31700
},
{
"epoch": 1.59,
"grad_norm": 1.4321199655532837,
"learning_rate": 0.00020566432160804018,
"loss": 1.5308,
"step": 31800
},
{
"epoch": 1.59,
"grad_norm": 1.8272013664245605,
"learning_rate": 0.00020536281407035177,
"loss": 1.5828,
"step": 31900
},
{
"epoch": 1.6,
"grad_norm": 1.3199920654296875,
"learning_rate": 0.0002050613065326633,
"loss": 1.5704,
"step": 32000
},
{
"epoch": 1.6,
"eval_loss": 1.536350131034851,
"eval_runtime": 37.4489,
"eval_samples_per_second": 26.703,
"eval_steps_per_second": 3.338,
"step": 32000
},
{
"epoch": 1.6,
"grad_norm": 1.796789526939392,
"learning_rate": 0.00020475979899497486,
"loss": 1.56,
"step": 32100
},
{
"epoch": 1.61,
"grad_norm": 1.4835096597671509,
"learning_rate": 0.0002044582914572864,
"loss": 1.5409,
"step": 32200
},
{
"epoch": 1.61,
"grad_norm": 1.225799798965454,
"learning_rate": 0.00020415678391959798,
"loss": 1.5109,
"step": 32300
},
{
"epoch": 1.62,
"grad_norm": 1.5552102327346802,
"learning_rate": 0.00020385527638190954,
"loss": 1.5717,
"step": 32400
},
{
"epoch": 1.62,
"grad_norm": 1.3638191223144531,
"learning_rate": 0.0002035537688442211,
"loss": 1.5763,
"step": 32500
},
{
"epoch": 1.63,
"grad_norm": 2.177093505859375,
"learning_rate": 0.00020325226130653263,
"loss": 1.5173,
"step": 32600
},
{
"epoch": 1.64,
"grad_norm": 3.643524169921875,
"learning_rate": 0.0002029507537688442,
"loss": 1.5596,
"step": 32700
},
{
"epoch": 1.64,
"grad_norm": 1.2808345556259155,
"learning_rate": 0.00020264924623115578,
"loss": 1.5812,
"step": 32800
},
{
"epoch": 1.65,
"grad_norm": 2.262430191040039,
"learning_rate": 0.0002023477386934673,
"loss": 1.5424,
"step": 32900
},
{
"epoch": 1.65,
"grad_norm": 1.8705729246139526,
"learning_rate": 0.00020204623115577887,
"loss": 1.5552,
"step": 33000
},
{
"epoch": 1.65,
"eval_loss": 1.5897144079208374,
"eval_runtime": 37.4808,
"eval_samples_per_second": 26.68,
"eval_steps_per_second": 3.335,
"step": 33000
},
{
"epoch": 1.66,
"grad_norm": 1.52475106716156,
"learning_rate": 0.00020174472361809043,
"loss": 1.5737,
"step": 33100
},
{
"epoch": 1.66,
"grad_norm": 1.8177305459976196,
"learning_rate": 0.00020144321608040201,
"loss": 1.5638,
"step": 33200
},
{
"epoch": 1.67,
"grad_norm": 1.8988004922866821,
"learning_rate": 0.00020114170854271355,
"loss": 1.54,
"step": 33300
},
{
"epoch": 1.67,
"grad_norm": 1.61077880859375,
"learning_rate": 0.0002008402010050251,
"loss": 1.5463,
"step": 33400
},
{
"epoch": 1.68,
"grad_norm": 1.3590441942214966,
"learning_rate": 0.00020053869346733667,
"loss": 1.5646,
"step": 33500
},
{
"epoch": 1.68,
"grad_norm": 1.5618160963058472,
"learning_rate": 0.00020023718592964822,
"loss": 1.5316,
"step": 33600
},
{
"epoch": 1.69,
"grad_norm": 1.7087860107421875,
"learning_rate": 0.00019993567839195978,
"loss": 1.5429,
"step": 33700
},
{
"epoch": 1.69,
"grad_norm": 2.558692693710327,
"learning_rate": 0.00019963417085427134,
"loss": 1.558,
"step": 33800
},
{
"epoch": 1.69,
"grad_norm": 2.3594534397125244,
"learning_rate": 0.00019933266331658288,
"loss": 1.5453,
"step": 33900
},
{
"epoch": 1.7,
"grad_norm": 1.57821786403656,
"learning_rate": 0.00019903115577889446,
"loss": 1.5541,
"step": 34000
},
{
"epoch": 1.7,
"eval_loss": 1.5565224885940552,
"eval_runtime": 37.4415,
"eval_samples_per_second": 26.708,
"eval_steps_per_second": 3.339,
"step": 34000
},
{
"epoch": 1.71,
"grad_norm": 1.9368255138397217,
"learning_rate": 0.00019872964824120602,
"loss": 1.5808,
"step": 34100
},
{
"epoch": 1.71,
"grad_norm": 3.0767080783843994,
"learning_rate": 0.00019842814070351758,
"loss": 1.5613,
"step": 34200
},
{
"epoch": 1.71,
"grad_norm": 1.8630317449569702,
"learning_rate": 0.00019812663316582911,
"loss": 1.5638,
"step": 34300
},
{
"epoch": 1.72,
"grad_norm": 3.0488193035125732,
"learning_rate": 0.0001978251256281407,
"loss": 1.5615,
"step": 34400
},
{
"epoch": 1.73,
"grad_norm": 3.8991503715515137,
"learning_rate": 0.00019752361809045226,
"loss": 1.4857,
"step": 34500
},
{
"epoch": 1.73,
"grad_norm": 2.0849859714508057,
"learning_rate": 0.0001972221105527638,
"loss": 1.535,
"step": 34600
},
{
"epoch": 1.73,
"grad_norm": 1.825913906097412,
"learning_rate": 0.00019692060301507535,
"loss": 1.5664,
"step": 34700
},
{
"epoch": 1.74,
"grad_norm": 5.08195686340332,
"learning_rate": 0.0001966190954773869,
"loss": 1.5252,
"step": 34800
},
{
"epoch": 1.75,
"grad_norm": 1.785659909248352,
"learning_rate": 0.0001963175879396985,
"loss": 1.5391,
"step": 34900
},
{
"epoch": 1.75,
"grad_norm": 1.4543670415878296,
"learning_rate": 0.00019601909547738692,
"loss": 1.5484,
"step": 35000
},
{
"epoch": 1.75,
"eval_loss": 1.573486566543579,
"eval_runtime": 37.5549,
"eval_samples_per_second": 26.628,
"eval_steps_per_second": 3.328,
"step": 35000
},
{
"epoch": 1.75,
"grad_norm": 1.6768901348114014,
"learning_rate": 0.00019571758793969848,
"loss": 1.5479,
"step": 35100
},
{
"epoch": 1.76,
"grad_norm": 1.9926207065582275,
"learning_rate": 0.00019541608040201004,
"loss": 1.5712,
"step": 35200
},
{
"epoch": 1.77,
"grad_norm": 1.9285818338394165,
"learning_rate": 0.00019511457286432157,
"loss": 1.567,
"step": 35300
},
{
"epoch": 1.77,
"grad_norm": 4.1938252449035645,
"learning_rate": 0.00019481306532663313,
"loss": 1.5538,
"step": 35400
},
{
"epoch": 1.77,
"grad_norm": 3.3867177963256836,
"learning_rate": 0.00019451155778894472,
"loss": 1.532,
"step": 35500
},
{
"epoch": 1.78,
"grad_norm": 2.6039962768554688,
"learning_rate": 0.00019421005025125625,
"loss": 1.5914,
"step": 35600
},
{
"epoch": 1.79,
"grad_norm": 1.900150179862976,
"learning_rate": 0.0001939085427135678,
"loss": 1.5435,
"step": 35700
},
{
"epoch": 1.79,
"grad_norm": 6.110165596008301,
"learning_rate": 0.00019360703517587937,
"loss": 1.5532,
"step": 35800
},
{
"epoch": 1.79,
"grad_norm": 2.2953763008117676,
"learning_rate": 0.00019330552763819096,
"loss": 1.5573,
"step": 35900
},
{
"epoch": 1.8,
"grad_norm": 1.9452694654464722,
"learning_rate": 0.0001930040201005025,
"loss": 1.5602,
"step": 36000
},
{
"epoch": 1.8,
"eval_loss": 1.5185086727142334,
"eval_runtime": 37.4604,
"eval_samples_per_second": 26.695,
"eval_steps_per_second": 3.337,
"step": 36000
},
{
"epoch": 1.81,
"grad_norm": 1.729576587677002,
"learning_rate": 0.00019270251256281405,
"loss": 1.5778,
"step": 36100
},
{
"epoch": 1.81,
"grad_norm": 2.98600172996521,
"learning_rate": 0.0001924010050251256,
"loss": 1.5318,
"step": 36200
},
{
"epoch": 1.81,
"grad_norm": 2.2745935916900635,
"learning_rate": 0.00019209949748743717,
"loss": 1.5475,
"step": 36300
},
{
"epoch": 1.82,
"grad_norm": 1.7677953243255615,
"learning_rate": 0.00019179798994974873,
"loss": 1.5174,
"step": 36400
},
{
"epoch": 1.82,
"grad_norm": 3.6853749752044678,
"learning_rate": 0.0001914964824120603,
"loss": 1.5709,
"step": 36500
},
{
"epoch": 1.83,
"grad_norm": 2.8981404304504395,
"learning_rate": 0.00019119497487437182,
"loss": 1.5519,
"step": 36600
},
{
"epoch": 1.83,
"grad_norm": 1.9862598180770874,
"learning_rate": 0.0001908934673366834,
"loss": 1.5547,
"step": 36700
},
{
"epoch": 1.84,
"grad_norm": 2.030618667602539,
"learning_rate": 0.00019059195979899497,
"loss": 1.5388,
"step": 36800
},
{
"epoch": 1.84,
"grad_norm": 1.609573483467102,
"learning_rate": 0.00019029045226130653,
"loss": 1.5727,
"step": 36900
},
{
"epoch": 1.85,
"grad_norm": 2.0508673191070557,
"learning_rate": 0.00018998894472361806,
"loss": 1.5522,
"step": 37000
},
{
"epoch": 1.85,
"eval_loss": 1.4889146089553833,
"eval_runtime": 37.4776,
"eval_samples_per_second": 26.683,
"eval_steps_per_second": 3.335,
"step": 37000
},
{
"epoch": 1.85,
"grad_norm": 1.5507237911224365,
"learning_rate": 0.00018968743718592964,
"loss": 1.5196,
"step": 37100
},
{
"epoch": 1.86,
"grad_norm": 2.1241345405578613,
"learning_rate": 0.0001893859296482412,
"loss": 1.553,
"step": 37200
},
{
"epoch": 1.86,
"grad_norm": 1.6903210878372192,
"learning_rate": 0.00018908442211055274,
"loss": 1.5292,
"step": 37300
},
{
"epoch": 1.87,
"grad_norm": 1.684809923171997,
"learning_rate": 0.0001887829145728643,
"loss": 1.5597,
"step": 37400
},
{
"epoch": 1.88,
"grad_norm": 2.072158098220825,
"learning_rate": 0.00018848442211055275,
"loss": 1.4914,
"step": 37500
},
{
"epoch": 1.88,
"grad_norm": 2.545748472213745,
"learning_rate": 0.0001881829145728643,
"loss": 1.54,
"step": 37600
},
{
"epoch": 1.89,
"grad_norm": 2.4186930656433105,
"learning_rate": 0.00018788140703517587,
"loss": 1.5298,
"step": 37700
},
{
"epoch": 1.89,
"grad_norm": 1.8362512588500977,
"learning_rate": 0.00018757989949748743,
"loss": 1.5372,
"step": 37800
},
{
"epoch": 1.9,
"grad_norm": 2.376615047454834,
"learning_rate": 0.000187278391959799,
"loss": 1.5343,
"step": 37900
},
{
"epoch": 1.9,
"grad_norm": 2.72920823097229,
"learning_rate": 0.00018697688442211052,
"loss": 1.4955,
"step": 38000
},
{
"epoch": 1.9,
"eval_loss": 1.5501998662948608,
"eval_runtime": 37.4829,
"eval_samples_per_second": 26.679,
"eval_steps_per_second": 3.335,
"step": 38000
},
{
"epoch": 1.91,
"grad_norm": 2.2700889110565186,
"learning_rate": 0.00018667537688442208,
"loss": 1.5149,
"step": 38100
},
{
"epoch": 1.91,
"grad_norm": 2.9030048847198486,
"learning_rate": 0.00018637386934673367,
"loss": 1.5451,
"step": 38200
},
{
"epoch": 1.92,
"grad_norm": 3.4646873474121094,
"learning_rate": 0.00018607236180904522,
"loss": 1.5477,
"step": 38300
},
{
"epoch": 1.92,
"grad_norm": 2.5595543384552,
"learning_rate": 0.00018577085427135676,
"loss": 1.5013,
"step": 38400
},
{
"epoch": 1.93,
"grad_norm": 1.3651905059814453,
"learning_rate": 0.00018546934673366832,
"loss": 1.5199,
"step": 38500
},
{
"epoch": 1.93,
"grad_norm": 2.045830011367798,
"learning_rate": 0.0001851678391959799,
"loss": 1.5677,
"step": 38600
},
{
"epoch": 1.94,
"grad_norm": 1.8075040578842163,
"learning_rate": 0.00018486633165829144,
"loss": 1.5564,
"step": 38700
},
{
"epoch": 1.94,
"grad_norm": 1.4847893714904785,
"learning_rate": 0.000184564824120603,
"loss": 1.5178,
"step": 38800
},
{
"epoch": 1.94,
"grad_norm": 3.2269062995910645,
"learning_rate": 0.00018426331658291455,
"loss": 1.513,
"step": 38900
},
{
"epoch": 1.95,
"grad_norm": 1.8105851411819458,
"learning_rate": 0.00018396180904522614,
"loss": 1.5007,
"step": 39000
},
{
"epoch": 1.95,
"eval_loss": 1.5438071489334106,
"eval_runtime": 37.4427,
"eval_samples_per_second": 26.707,
"eval_steps_per_second": 3.338,
"step": 39000
},
{
"epoch": 1.96,
"grad_norm": 1.8713701963424683,
"learning_rate": 0.00018366030150753767,
"loss": 1.5266,
"step": 39100
},
{
"epoch": 1.96,
"grad_norm": 1.4177159070968628,
"learning_rate": 0.00018335879396984923,
"loss": 1.5345,
"step": 39200
},
{
"epoch": 1.96,
"grad_norm": 2.068962574005127,
"learning_rate": 0.00018305728643216076,
"loss": 1.5136,
"step": 39300
},
{
"epoch": 1.97,
"grad_norm": 2.0991291999816895,
"learning_rate": 0.00018275577889447235,
"loss": 1.5509,
"step": 39400
},
{
"epoch": 1.98,
"grad_norm": 2.322105884552002,
"learning_rate": 0.0001824542713567839,
"loss": 1.5222,
"step": 39500
},
{
"epoch": 1.98,
"grad_norm": 2.1995983123779297,
"learning_rate": 0.00018215276381909547,
"loss": 1.544,
"step": 39600
},
{
"epoch": 1.98,
"grad_norm": 1.4002470970153809,
"learning_rate": 0.000181851256281407,
"loss": 1.5184,
"step": 39700
},
{
"epoch": 1.99,
"grad_norm": 2.0827932357788086,
"learning_rate": 0.0001815497487437186,
"loss": 1.5388,
"step": 39800
},
{
"epoch": 2.0,
"grad_norm": 3.9671342372894287,
"learning_rate": 0.00018124824120603015,
"loss": 1.5331,
"step": 39900
},
{
"epoch": 2.0,
"grad_norm": 2.979598045349121,
"learning_rate": 0.00018094673366834168,
"loss": 1.5058,
"step": 40000
},
{
"epoch": 2.0,
"eval_loss": 1.5161113739013672,
"eval_runtime": 37.4278,
"eval_samples_per_second": 26.718,
"eval_steps_per_second": 3.34,
"step": 40000
},
{
"epoch": 2.0,
"grad_norm": 2.0998573303222656,
"learning_rate": 0.00018064824120603013,
"loss": 1.4831,
"step": 40100
},
{
"epoch": 2.01,
"grad_norm": 2.1450562477111816,
"learning_rate": 0.0001803467336683417,
"loss": 1.5567,
"step": 40200
},
{
"epoch": 2.02,
"grad_norm": 2.080376148223877,
"learning_rate": 0.00018004522613065325,
"loss": 1.5088,
"step": 40300
},
{
"epoch": 2.02,
"grad_norm": 1.682288408279419,
"learning_rate": 0.0001797437185929648,
"loss": 1.5461,
"step": 40400
},
{
"epoch": 2.02,
"grad_norm": 1.6313551664352417,
"learning_rate": 0.00017944221105527637,
"loss": 1.4776,
"step": 40500
},
{
"epoch": 2.03,
"grad_norm": 3.1746320724487305,
"learning_rate": 0.00017914070351758793,
"loss": 1.534,
"step": 40600
},
{
"epoch": 2.04,
"grad_norm": 1.7039170265197754,
"learning_rate": 0.00017883919597989946,
"loss": 1.565,
"step": 40700
},
{
"epoch": 2.04,
"grad_norm": 1.738771677017212,
"learning_rate": 0.00017853768844221102,
"loss": 1.5391,
"step": 40800
},
{
"epoch": 2.04,
"grad_norm": 4.0762715339660645,
"learning_rate": 0.0001782361809045226,
"loss": 1.5153,
"step": 40900
},
{
"epoch": 2.05,
"grad_norm": 3.6206607818603516,
"learning_rate": 0.00017793467336683417,
"loss": 1.5254,
"step": 41000
},
{
"epoch": 2.05,
"eval_loss": 1.5652326345443726,
"eval_runtime": 37.621,
"eval_samples_per_second": 26.581,
"eval_steps_per_second": 3.323,
"step": 41000
},
{
"epoch": 2.06,
"grad_norm": 1.2320189476013184,
"learning_rate": 0.0001776331658291457,
"loss": 1.5153,
"step": 41100
},
{
"epoch": 2.06,
"grad_norm": 2.275785446166992,
"learning_rate": 0.00017733165829145726,
"loss": 1.5192,
"step": 41200
},
{
"epoch": 2.06,
"grad_norm": 1.8659756183624268,
"learning_rate": 0.00017703015075376885,
"loss": 1.5563,
"step": 41300
},
{
"epoch": 2.07,
"grad_norm": 1.6176475286483765,
"learning_rate": 0.00017672864321608038,
"loss": 1.5251,
"step": 41400
},
{
"epoch": 2.08,
"grad_norm": 1.575243592262268,
"learning_rate": 0.00017642713567839194,
"loss": 1.5178,
"step": 41500
},
{
"epoch": 2.08,
"grad_norm": 1.6838304996490479,
"learning_rate": 0.0001761256281407035,
"loss": 1.5255,
"step": 41600
},
{
"epoch": 2.08,
"grad_norm": 5.561213493347168,
"learning_rate": 0.00017582412060301509,
"loss": 1.4885,
"step": 41700
},
{
"epoch": 2.09,
"grad_norm": 2.0810487270355225,
"learning_rate": 0.00017552261306532662,
"loss": 1.491,
"step": 41800
},
{
"epoch": 2.1,
"grad_norm": 1.894136667251587,
"learning_rate": 0.00017522110552763818,
"loss": 1.5072,
"step": 41900
},
{
"epoch": 2.1,
"grad_norm": 4.531317710876465,
"learning_rate": 0.00017492261306532663,
"loss": 1.5108,
"step": 42000
},
{
"epoch": 2.1,
"eval_loss": 1.5298963785171509,
"eval_runtime": 37.5318,
"eval_samples_per_second": 26.644,
"eval_steps_per_second": 3.331,
"step": 42000
},
{
"epoch": 2.1,
"grad_norm": 1.9595602750778198,
"learning_rate": 0.00017462110552763816,
"loss": 1.5178,
"step": 42100
},
{
"epoch": 2.11,
"grad_norm": 2.1969220638275146,
"learning_rate": 0.00017431959798994972,
"loss": 1.508,
"step": 42200
},
{
"epoch": 2.12,
"grad_norm": 3.0334370136260986,
"learning_rate": 0.0001740180904522613,
"loss": 1.5255,
"step": 42300
},
{
"epoch": 2.12,
"grad_norm": 1.8148131370544434,
"learning_rate": 0.00017371658291457287,
"loss": 1.4686,
"step": 42400
},
{
"epoch": 2.12,
"grad_norm": 5.726377964019775,
"learning_rate": 0.0001734150753768844,
"loss": 1.5355,
"step": 42500
},
{
"epoch": 2.13,
"grad_norm": 3.15856671333313,
"learning_rate": 0.00017311356783919596,
"loss": 1.534,
"step": 42600
},
{
"epoch": 2.13,
"grad_norm": 6.238559246063232,
"learning_rate": 0.00017281206030150755,
"loss": 1.5073,
"step": 42700
},
{
"epoch": 2.14,
"grad_norm": 2.281912326812744,
"learning_rate": 0.00017251055276381908,
"loss": 1.5341,
"step": 42800
},
{
"epoch": 2.15,
"grad_norm": 3.874361991882324,
"learning_rate": 0.00017220904522613064,
"loss": 1.5204,
"step": 42900
},
{
"epoch": 2.15,
"grad_norm": 2.5632574558258057,
"learning_rate": 0.0001719075376884422,
"loss": 1.5028,
"step": 43000
},
{
"epoch": 2.15,
"eval_loss": 1.5213427543640137,
"eval_runtime": 37.5103,
"eval_samples_per_second": 26.659,
"eval_steps_per_second": 3.332,
"step": 43000
},
{
"epoch": 2.15,
"grad_norm": 4.4505133628845215,
"learning_rate": 0.00017160603015075373,
"loss": 1.5318,
"step": 43100
},
{
"epoch": 2.16,
"grad_norm": 1.6099460124969482,
"learning_rate": 0.00017130452261306532,
"loss": 1.5093,
"step": 43200
},
{
"epoch": 2.17,
"grad_norm": 4.269103527069092,
"learning_rate": 0.00017100301507537688,
"loss": 1.4839,
"step": 43300
},
{
"epoch": 2.17,
"grad_norm": 10.070590019226074,
"learning_rate": 0.0001707015075376884,
"loss": 1.5407,
"step": 43400
},
{
"epoch": 2.17,
"grad_norm": 17.754796981811523,
"learning_rate": 0.00017039999999999997,
"loss": 1.5102,
"step": 43500
},
{
"epoch": 2.18,
"grad_norm": 2.8812096118927,
"learning_rate": 0.00017009849246231155,
"loss": 1.5227,
"step": 43600
},
{
"epoch": 2.19,
"grad_norm": 2.861893653869629,
"learning_rate": 0.00016979698492462311,
"loss": 1.544,
"step": 43700
},
{
"epoch": 2.19,
"grad_norm": 1.9935343265533447,
"learning_rate": 0.00016949547738693465,
"loss": 1.5209,
"step": 43800
},
{
"epoch": 2.19,
"grad_norm": 2.8261117935180664,
"learning_rate": 0.0001691939698492462,
"loss": 1.5404,
"step": 43900
},
{
"epoch": 2.2,
"grad_norm": 2.6541242599487305,
"learning_rate": 0.0001688924623115578,
"loss": 1.5048,
"step": 44000
},
{
"epoch": 2.2,
"eval_loss": 1.5309374332427979,
"eval_runtime": 65.7795,
"eval_samples_per_second": 15.202,
"eval_steps_per_second": 1.9,
"step": 44000
},
{
"epoch": 2.21,
"grad_norm": 1.1707357168197632,
"learning_rate": 0.00016859095477386932,
"loss": 1.5479,
"step": 44100
},
{
"epoch": 2.21,
"grad_norm": 2.058530807495117,
"learning_rate": 0.00016828944723618088,
"loss": 1.5423,
"step": 44200
},
{
"epoch": 2.21,
"grad_norm": 2.841702699661255,
"learning_rate": 0.00016798793969849244,
"loss": 1.5196,
"step": 44300
},
{
"epoch": 2.22,
"grad_norm": 1.6679179668426514,
"learning_rate": 0.00016768643216080403,
"loss": 1.5085,
"step": 44400
},
{
"epoch": 2.23,
"grad_norm": 2.1461362838745117,
"learning_rate": 0.00016738492462311556,
"loss": 1.538,
"step": 44500
},
{
"epoch": 2.23,
"grad_norm": 2.351161241531372,
"learning_rate": 0.00016708341708542712,
"loss": 1.4874,
"step": 44600
},
{
"epoch": 2.23,
"grad_norm": 2.474755048751831,
"learning_rate": 0.00016678190954773868,
"loss": 1.5646,
"step": 44700
},
{
"epoch": 2.24,
"grad_norm": 2.6611087322235107,
"learning_rate": 0.00016648040201005024,
"loss": 1.5373,
"step": 44800
},
{
"epoch": 2.25,
"grad_norm": 2.518184185028076,
"learning_rate": 0.0001661788944723618,
"loss": 1.4969,
"step": 44900
},
{
"epoch": 2.25,
"grad_norm": 4.304960250854492,
"learning_rate": 0.00016587738693467336,
"loss": 1.4827,
"step": 45000
},
{
"epoch": 2.25,
"eval_loss": 1.5653032064437866,
"eval_runtime": 39.6553,
"eval_samples_per_second": 25.217,
"eval_steps_per_second": 3.152,
"step": 45000
},
{
"epoch": 2.25,
"grad_norm": 1.1882766485214233,
"learning_rate": 0.0001655758793969849,
"loss": 1.5062,
"step": 45100
},
{
"epoch": 2.26,
"grad_norm": 2.196139335632324,
"learning_rate": 0.00016527437185929648,
"loss": 1.5026,
"step": 45200
},
{
"epoch": 2.27,
"grad_norm": 1.8797615766525269,
"learning_rate": 0.00016497286432160804,
"loss": 1.4947,
"step": 45300
},
{
"epoch": 2.27,
"grad_norm": 1.396849274635315,
"learning_rate": 0.00016467135678391957,
"loss": 1.5002,
"step": 45400
},
{
"epoch": 2.27,
"grad_norm": 2.432685375213623,
"learning_rate": 0.00016436984924623113,
"loss": 1.5021,
"step": 45500
},
{
"epoch": 2.28,
"grad_norm": 2.2086715698242188,
"learning_rate": 0.0001640683417085427,
"loss": 1.5067,
"step": 45600
},
{
"epoch": 2.29,
"grad_norm": 4.262127876281738,
"learning_rate": 0.00016376683417085428,
"loss": 1.519,
"step": 45700
},
{
"epoch": 2.29,
"grad_norm": 2.92459774017334,
"learning_rate": 0.0001634653266331658,
"loss": 1.5138,
"step": 45800
},
{
"epoch": 2.29,
"grad_norm": 1.3408390283584595,
"learning_rate": 0.00016316381909547737,
"loss": 1.5296,
"step": 45900
},
{
"epoch": 2.3,
"grad_norm": 1.9604805707931519,
"learning_rate": 0.00016286231155778893,
"loss": 1.5207,
"step": 46000
},
{
"epoch": 2.3,
"eval_loss": 1.5254641771316528,
"eval_runtime": 42.9693,
"eval_samples_per_second": 23.272,
"eval_steps_per_second": 2.909,
"step": 46000
},
{
"epoch": 2.31,
"grad_norm": 2.5174312591552734,
"learning_rate": 0.00016256080402010049,
"loss": 1.5487,
"step": 46100
},
{
"epoch": 2.31,
"grad_norm": 1.4481089115142822,
"learning_rate": 0.00016225929648241205,
"loss": 1.5167,
"step": 46200
},
{
"epoch": 2.31,
"grad_norm": 1.5994378328323364,
"learning_rate": 0.0001619577889447236,
"loss": 1.4549,
"step": 46300
},
{
"epoch": 2.32,
"grad_norm": 2.4599764347076416,
"learning_rate": 0.00016165628140703514,
"loss": 1.5128,
"step": 46400
},
{
"epoch": 2.33,
"grad_norm": 1.8390551805496216,
"learning_rate": 0.00016135477386934672,
"loss": 1.5248,
"step": 46500
},
{
"epoch": 2.33,
"grad_norm": 2.679804801940918,
"learning_rate": 0.00016105326633165828,
"loss": 1.5105,
"step": 46600
},
{
"epoch": 2.33,
"grad_norm": 1.7645354270935059,
"learning_rate": 0.00016075477386934674,
"loss": 1.4787,
"step": 46700
},
{
"epoch": 2.34,
"grad_norm": 1.035243034362793,
"learning_rate": 0.00016045326633165827,
"loss": 1.5257,
"step": 46800
},
{
"epoch": 2.34,
"grad_norm": 11.433294296264648,
"learning_rate": 0.00016015175879396983,
"loss": 1.4615,
"step": 46900
},
{
"epoch": 2.35,
"grad_norm": 1.600629448890686,
"learning_rate": 0.0001598502512562814,
"loss": 1.4973,
"step": 47000
},
{
"epoch": 2.35,
"eval_loss": 1.5015385150909424,
"eval_runtime": 41.9852,
"eval_samples_per_second": 23.818,
"eval_steps_per_second": 2.977,
"step": 47000
},
{
"epoch": 2.35,
"grad_norm": 2.1351780891418457,
"learning_rate": 0.00015954874371859297,
"loss": 1.5163,
"step": 47100
},
{
"epoch": 2.36,
"grad_norm": 2.6234447956085205,
"learning_rate": 0.0001592472361809045,
"loss": 1.5147,
"step": 47200
},
{
"epoch": 2.37,
"grad_norm": 2.0315120220184326,
"learning_rate": 0.00015894874371859296,
"loss": 1.5044,
"step": 47300
},
{
"epoch": 2.37,
"grad_norm": 2.7582693099975586,
"learning_rate": 0.00015864723618090452,
"loss": 1.5033,
"step": 47400
},
{
"epoch": 2.38,
"grad_norm": 2.5383968353271484,
"learning_rate": 0.00015834572864321605,
"loss": 1.5198,
"step": 47500
},
{
"epoch": 2.38,
"grad_norm": 3.1361851692199707,
"learning_rate": 0.0001580442211055276,
"loss": 1.5006,
"step": 47600
},
{
"epoch": 2.38,
"grad_norm": 1.669693946838379,
"learning_rate": 0.0001577427135678392,
"loss": 1.5203,
"step": 47700
},
{
"epoch": 2.39,
"grad_norm": 3.662080764770508,
"learning_rate": 0.00015744120603015076,
"loss": 1.5427,
"step": 47800
},
{
"epoch": 2.4,
"grad_norm": 2.101423740386963,
"learning_rate": 0.0001571396984924623,
"loss": 1.489,
"step": 47900
},
{
"epoch": 2.4,
"grad_norm": 8.889337539672852,
"learning_rate": 0.00015684120603015074,
"loss": 1.5198,
"step": 48000
},
{
"epoch": 2.4,
"eval_loss": 1.5436657667160034,
"eval_runtime": 42.227,
"eval_samples_per_second": 23.682,
"eval_steps_per_second": 2.96,
"step": 48000
},
{
"epoch": 2.41,
"grad_norm": 9.11814022064209,
"learning_rate": 0.0001565396984924623,
"loss": 1.5623,
"step": 48100
},
{
"epoch": 2.41,
"grad_norm": 11.104165077209473,
"learning_rate": 0.00015623819095477383,
"loss": 1.5019,
"step": 48200
},
{
"epoch": 2.42,
"grad_norm": 4.448848724365234,
"learning_rate": 0.00015593668341708542,
"loss": 1.5464,
"step": 48300
},
{
"epoch": 2.42,
"grad_norm": 2.207139253616333,
"learning_rate": 0.00015563517587939698,
"loss": 1.5105,
"step": 48400
},
{
"epoch": 2.42,
"grad_norm": 2.692350149154663,
"learning_rate": 0.00015533366834170854,
"loss": 1.5351,
"step": 48500
},
{
"epoch": 2.43,
"grad_norm": 1.8738429546356201,
"learning_rate": 0.00015503216080402007,
"loss": 1.4882,
"step": 48600
},
{
"epoch": 2.44,
"grad_norm": 2.8214309215545654,
"learning_rate": 0.00015473065326633166,
"loss": 1.4932,
"step": 48700
},
{
"epoch": 2.44,
"grad_norm": 1.4316879510879517,
"learning_rate": 0.00015442914572864322,
"loss": 1.498,
"step": 48800
},
{
"epoch": 2.44,
"grad_norm": 1.438391923904419,
"learning_rate": 0.00015412763819095475,
"loss": 1.4975,
"step": 48900
},
{
"epoch": 2.45,
"grad_norm": 2.7039265632629395,
"learning_rate": 0.0001538261306532663,
"loss": 1.511,
"step": 49000
},
{
"epoch": 2.45,
"eval_loss": 1.5066561698913574,
"eval_runtime": 42.052,
"eval_samples_per_second": 23.78,
"eval_steps_per_second": 2.973,
"step": 49000
},
{
"epoch": 2.46,
"grad_norm": 2.8025197982788086,
"learning_rate": 0.00015352462311557787,
"loss": 1.5082,
"step": 49100
},
{
"epoch": 2.46,
"grad_norm": 4.876307964324951,
"learning_rate": 0.00015322311557788946,
"loss": 1.4825,
"step": 49200
},
{
"epoch": 2.46,
"grad_norm": 4.34946346282959,
"learning_rate": 0.000152921608040201,
"loss": 1.5117,
"step": 49300
},
{
"epoch": 2.47,
"grad_norm": 2.5481321811676025,
"learning_rate": 0.00015262010050251255,
"loss": 1.5361,
"step": 49400
},
{
"epoch": 2.48,
"grad_norm": 1.5703433752059937,
"learning_rate": 0.00015231859296482408,
"loss": 1.5113,
"step": 49500
},
{
"epoch": 2.48,
"grad_norm": 1.8963671922683716,
"learning_rate": 0.00015201708542713567,
"loss": 1.5095,
"step": 49600
},
{
"epoch": 2.48,
"grad_norm": 2.296262264251709,
"learning_rate": 0.00015171557788944723,
"loss": 1.5479,
"step": 49700
},
{
"epoch": 2.49,
"grad_norm": 2.796741247177124,
"learning_rate": 0.00015141407035175879,
"loss": 1.4946,
"step": 49800
},
{
"epoch": 2.5,
"grad_norm": 2.633420705795288,
"learning_rate": 0.00015111256281407032,
"loss": 1.4917,
"step": 49900
},
{
"epoch": 2.5,
"grad_norm": 3.0930418968200684,
"learning_rate": 0.0001508110552763819,
"loss": 1.5058,
"step": 50000
},
{
"epoch": 2.5,
"eval_loss": 1.539115071296692,
"eval_runtime": 41.8844,
"eval_samples_per_second": 23.875,
"eval_steps_per_second": 2.984,
"step": 50000
},
{
"epoch": 2.5,
"grad_norm": 2.2942283153533936,
"learning_rate": 0.00015050954773869346,
"loss": 1.4981,
"step": 50100
},
{
"epoch": 2.51,
"grad_norm": 2.622481346130371,
"learning_rate": 0.000150208040201005,
"loss": 1.4829,
"step": 50200
},
{
"epoch": 2.52,
"grad_norm": 2.5901362895965576,
"learning_rate": 0.00014990653266331658,
"loss": 1.495,
"step": 50300
},
{
"epoch": 2.52,
"grad_norm": 3.8583192825317383,
"learning_rate": 0.00014960502512562812,
"loss": 1.494,
"step": 50400
},
{
"epoch": 2.52,
"grad_norm": 2.841306209564209,
"learning_rate": 0.00014930653266331657,
"loss": 1.4806,
"step": 50500
},
{
"epoch": 2.53,
"grad_norm": 2.646027088165283,
"learning_rate": 0.00014900502512562813,
"loss": 1.4907,
"step": 50600
},
{
"epoch": 2.54,
"grad_norm": 2.384547472000122,
"learning_rate": 0.0001487035175879397,
"loss": 1.4809,
"step": 50700
},
{
"epoch": 2.54,
"grad_norm": 2.698951005935669,
"learning_rate": 0.00014840201005025125,
"loss": 1.5145,
"step": 50800
},
{
"epoch": 2.54,
"grad_norm": 4.710977077484131,
"learning_rate": 0.0001481035175879397,
"loss": 1.4805,
"step": 50900
},
{
"epoch": 2.55,
"grad_norm": 5.228128433227539,
"learning_rate": 0.00014780201005025123,
"loss": 1.512,
"step": 51000
},
{
"epoch": 2.55,
"eval_loss": 1.5644181966781616,
"eval_runtime": 37.1723,
"eval_samples_per_second": 26.902,
"eval_steps_per_second": 3.363,
"step": 51000
},
{
"epoch": 2.56,
"grad_norm": 5.459705829620361,
"learning_rate": 0.00014750050251256282,
"loss": 1.4861,
"step": 51100
},
{
"epoch": 2.56,
"grad_norm": 1.9696797132492065,
"learning_rate": 0.00014719899497487435,
"loss": 1.5177,
"step": 51200
},
{
"epoch": 2.56,
"grad_norm": 2.262742757797241,
"learning_rate": 0.0001468974874371859,
"loss": 1.5074,
"step": 51300
},
{
"epoch": 2.57,
"grad_norm": 3.0286054611206055,
"learning_rate": 0.00014659597989949747,
"loss": 1.5233,
"step": 51400
},
{
"epoch": 2.58,
"grad_norm": 2.805699110031128,
"learning_rate": 0.00014629447236180903,
"loss": 1.5002,
"step": 51500
},
{
"epoch": 2.58,
"grad_norm": 7.009899616241455,
"learning_rate": 0.0001459929648241206,
"loss": 1.5224,
"step": 51600
},
{
"epoch": 2.58,
"grad_norm": 2.203697443008423,
"learning_rate": 0.00014569145728643215,
"loss": 1.5264,
"step": 51700
},
{
"epoch": 2.59,
"grad_norm": 38.17327117919922,
"learning_rate": 0.0001453899497487437,
"loss": 1.5233,
"step": 51800
},
{
"epoch": 2.59,
"grad_norm": 3.463594436645508,
"learning_rate": 0.00014508844221105527,
"loss": 1.5476,
"step": 51900
},
{
"epoch": 2.6,
"grad_norm": 4.695711135864258,
"learning_rate": 0.00014478693467336683,
"loss": 1.5092,
"step": 52000
},
{
"epoch": 2.6,
"eval_loss": 1.524036169052124,
"eval_runtime": 41.6949,
"eval_samples_per_second": 23.984,
"eval_steps_per_second": 2.998,
"step": 52000
},
{
"epoch": 2.6,
"grad_norm": 3.0102827548980713,
"learning_rate": 0.00014448542713567836,
"loss": 1.4978,
"step": 52100
},
{
"epoch": 2.61,
"grad_norm": 4.087903022766113,
"learning_rate": 0.00014418391959798995,
"loss": 1.4772,
"step": 52200
},
{
"epoch": 2.62,
"grad_norm": 2.3847885131835938,
"learning_rate": 0.00014388241206030148,
"loss": 1.5173,
"step": 52300
},
{
"epoch": 2.62,
"grad_norm": 2.907341241836548,
"learning_rate": 0.00014358090452261306,
"loss": 1.5287,
"step": 52400
},
{
"epoch": 2.62,
"grad_norm": 3.0708484649658203,
"learning_rate": 0.0001432793969849246,
"loss": 1.524,
"step": 52500
},
{
"epoch": 2.63,
"grad_norm": 2.70914888381958,
"learning_rate": 0.00014297788944723618,
"loss": 1.5115,
"step": 52600
},
{
"epoch": 2.63,
"grad_norm": 3.0721583366394043,
"learning_rate": 0.00014267638190954772,
"loss": 1.5273,
"step": 52700
},
{
"epoch": 2.64,
"grad_norm": 4.067818641662598,
"learning_rate": 0.00014237487437185927,
"loss": 1.5237,
"step": 52800
},
{
"epoch": 2.65,
"grad_norm": 22.84881591796875,
"learning_rate": 0.00014207336683417083,
"loss": 1.489,
"step": 52900
},
{
"epoch": 2.65,
"grad_norm": 4.021462440490723,
"learning_rate": 0.0001417718592964824,
"loss": 1.5284,
"step": 53000
},
{
"epoch": 2.65,
"eval_loss": 1.5171101093292236,
"eval_runtime": 41.2537,
"eval_samples_per_second": 24.24,
"eval_steps_per_second": 3.03,
"step": 53000
},
{
"epoch": 2.66,
"grad_norm": 14.866168975830078,
"learning_rate": 0.00014147035175879395,
"loss": 1.5114,
"step": 53100
},
{
"epoch": 2.66,
"grad_norm": 8.153103828430176,
"learning_rate": 0.0001411688442211055,
"loss": 1.4384,
"step": 53200
},
{
"epoch": 2.67,
"grad_norm": 3.6942877769470215,
"learning_rate": 0.00014086733668341707,
"loss": 1.5189,
"step": 53300
},
{
"epoch": 2.67,
"grad_norm": 1.7417048215866089,
"learning_rate": 0.00014056582914572863,
"loss": 1.5491,
"step": 53400
},
{
"epoch": 2.67,
"grad_norm": 3.291808605194092,
"learning_rate": 0.0001402643216080402,
"loss": 1.5085,
"step": 53500
},
{
"epoch": 2.68,
"grad_norm": 4.061951637268066,
"learning_rate": 0.00013996281407035175,
"loss": 1.5328,
"step": 53600
},
{
"epoch": 2.69,
"grad_norm": 2.119297742843628,
"learning_rate": 0.0001396613065326633,
"loss": 1.4872,
"step": 53700
},
{
"epoch": 2.69,
"grad_norm": 3.23767352104187,
"learning_rate": 0.00013935979899497487,
"loss": 1.5208,
"step": 53800
},
{
"epoch": 2.69,
"grad_norm": 4.745997428894043,
"learning_rate": 0.00013905829145728643,
"loss": 1.5251,
"step": 53900
},
{
"epoch": 2.7,
"grad_norm": 1.6231697797775269,
"learning_rate": 0.000138756783919598,
"loss": 1.5151,
"step": 54000
},
{
"epoch": 2.7,
"eval_loss": 1.4974777698516846,
"eval_runtime": 37.2945,
"eval_samples_per_second": 26.814,
"eval_steps_per_second": 3.352,
"step": 54000
},
{
"epoch": 2.71,
"grad_norm": 2.7751882076263428,
"learning_rate": 0.00013845527638190955,
"loss": 1.4891,
"step": 54100
},
{
"epoch": 2.71,
"grad_norm": 3.2454652786254883,
"learning_rate": 0.00013815376884422108,
"loss": 1.4929,
"step": 54200
},
{
"epoch": 2.71,
"grad_norm": 2.5808818340301514,
"learning_rate": 0.00013785226130653264,
"loss": 1.5728,
"step": 54300
},
{
"epoch": 2.72,
"grad_norm": 3.754495859146118,
"learning_rate": 0.0001375507537688442,
"loss": 1.5073,
"step": 54400
},
{
"epoch": 2.73,
"grad_norm": 2.158400535583496,
"learning_rate": 0.00013724924623115576,
"loss": 1.5014,
"step": 54500
},
{
"epoch": 2.73,
"grad_norm": 2.022975206375122,
"learning_rate": 0.00013694773869346732,
"loss": 1.5147,
"step": 54600
},
{
"epoch": 2.73,
"grad_norm": 2.061910629272461,
"learning_rate": 0.00013664623115577888,
"loss": 1.4701,
"step": 54700
},
{
"epoch": 2.74,
"grad_norm": 2.8262743949890137,
"learning_rate": 0.00013634472361809044,
"loss": 1.4988,
"step": 54800
},
{
"epoch": 2.75,
"grad_norm": 2.389618158340454,
"learning_rate": 0.000136043216080402,
"loss": 1.5138,
"step": 54900
},
{
"epoch": 2.75,
"grad_norm": 2.31430983543396,
"learning_rate": 0.00013574170854271356,
"loss": 1.5234,
"step": 55000
},
{
"epoch": 2.75,
"eval_loss": 1.4553519487380981,
"eval_runtime": 43.1448,
"eval_samples_per_second": 23.178,
"eval_steps_per_second": 2.897,
"step": 55000
},
{
"epoch": 2.75,
"grad_norm": 2.14939022064209,
"learning_rate": 0.00013544020100502512,
"loss": 1.4971,
"step": 55100
},
{
"epoch": 2.76,
"grad_norm": 2.3362553119659424,
"learning_rate": 0.00013513869346733667,
"loss": 1.4899,
"step": 55200
},
{
"epoch": 2.77,
"grad_norm": 1.8975802659988403,
"learning_rate": 0.00013483718592964823,
"loss": 1.4855,
"step": 55300
},
{
"epoch": 2.77,
"grad_norm": 1.6080349683761597,
"learning_rate": 0.0001345356783919598,
"loss": 1.4789,
"step": 55400
},
{
"epoch": 2.77,
"grad_norm": 2.4170632362365723,
"learning_rate": 0.00013423417085427135,
"loss": 1.5052,
"step": 55500
},
{
"epoch": 2.78,
"grad_norm": 3.0807037353515625,
"learning_rate": 0.0001339326633165829,
"loss": 1.4848,
"step": 55600
},
{
"epoch": 2.79,
"grad_norm": 2.7296719551086426,
"learning_rate": 0.00013363115577889447,
"loss": 1.4695,
"step": 55700
},
{
"epoch": 2.79,
"grad_norm": 2.392073154449463,
"learning_rate": 0.000133329648241206,
"loss": 1.4757,
"step": 55800
},
{
"epoch": 2.79,
"grad_norm": 1.3958042860031128,
"learning_rate": 0.0001330281407035176,
"loss": 1.5152,
"step": 55900
},
{
"epoch": 2.8,
"grad_norm": 5.350344657897949,
"learning_rate": 0.00013272663316582912,
"loss": 1.4918,
"step": 56000
},
{
"epoch": 2.8,
"eval_loss": 1.4755498170852661,
"eval_runtime": 55.3712,
"eval_samples_per_second": 18.06,
"eval_steps_per_second": 2.257,
"step": 56000
},
{
"epoch": 2.81,
"grad_norm": 2.9437155723571777,
"learning_rate": 0.0001324251256281407,
"loss": 1.4673,
"step": 56100
},
{
"epoch": 2.81,
"grad_norm": 2.0883982181549072,
"learning_rate": 0.00013212361809045224,
"loss": 1.5435,
"step": 56200
},
{
"epoch": 2.81,
"grad_norm": 1.475618600845337,
"learning_rate": 0.00013182211055276383,
"loss": 1.4895,
"step": 56300
},
{
"epoch": 2.82,
"grad_norm": 3.1873602867126465,
"learning_rate": 0.00013152060301507536,
"loss": 1.4753,
"step": 56400
},
{
"epoch": 2.83,
"grad_norm": 1.1994943618774414,
"learning_rate": 0.00013121909547738692,
"loss": 1.4627,
"step": 56500
},
{
"epoch": 2.83,
"grad_norm": 2.3503193855285645,
"learning_rate": 0.00013091758793969848,
"loss": 1.5312,
"step": 56600
},
{
"epoch": 2.83,
"grad_norm": 1.5025432109832764,
"learning_rate": 0.00013061608040201004,
"loss": 1.4981,
"step": 56700
},
{
"epoch": 2.84,
"grad_norm": 1.2746458053588867,
"learning_rate": 0.0001303145728643216,
"loss": 1.4644,
"step": 56800
},
{
"epoch": 2.84,
"grad_norm": 2.1974568367004395,
"learning_rate": 0.00013001306532663316,
"loss": 1.4958,
"step": 56900
},
{
"epoch": 2.85,
"grad_norm": 2.137925863265991,
"learning_rate": 0.00012971155778894472,
"loss": 1.4808,
"step": 57000
},
{
"epoch": 2.85,
"eval_loss": 1.4832957983016968,
"eval_runtime": 58.8442,
"eval_samples_per_second": 16.994,
"eval_steps_per_second": 2.124,
"step": 57000
},
{
"epoch": 2.85,
"grad_norm": 1.9320988655090332,
"learning_rate": 0.00012941005025125628,
"loss": 1.4718,
"step": 57100
},
{
"epoch": 2.86,
"grad_norm": 2.317089080810547,
"learning_rate": 0.00012910854271356784,
"loss": 1.436,
"step": 57200
},
{
"epoch": 2.87,
"grad_norm": 2.3655266761779785,
"learning_rate": 0.00012880703517587937,
"loss": 1.4967,
"step": 57300
},
{
"epoch": 2.87,
"grad_norm": 2.1567230224609375,
"learning_rate": 0.00012850552763819096,
"loss": 1.4319,
"step": 57400
},
{
"epoch": 2.88,
"grad_norm": 1.759560227394104,
"learning_rate": 0.0001282040201005025,
"loss": 1.4719,
"step": 57500
},
{
"epoch": 2.88,
"grad_norm": 3.5496578216552734,
"learning_rate": 0.00012790251256281407,
"loss": 1.513,
"step": 57600
},
{
"epoch": 2.88,
"grad_norm": 1.4334951639175415,
"learning_rate": 0.0001276010050251256,
"loss": 1.5037,
"step": 57700
},
{
"epoch": 2.89,
"grad_norm": 1.5903607606887817,
"learning_rate": 0.0001272994974874372,
"loss": 1.4386,
"step": 57800
},
{
"epoch": 2.9,
"grad_norm": 1.8693747520446777,
"learning_rate": 0.00012699798994974873,
"loss": 1.4691,
"step": 57900
},
{
"epoch": 2.9,
"grad_norm": 1.4653220176696777,
"learning_rate": 0.00012669648241206029,
"loss": 1.4531,
"step": 58000
},
{
"epoch": 2.9,
"eval_loss": 1.4672846794128418,
"eval_runtime": 70.1651,
"eval_samples_per_second": 14.252,
"eval_steps_per_second": 1.782,
"step": 58000
},
{
"epoch": 2.91,
"grad_norm": 1.9329131841659546,
"learning_rate": 0.00012639497487437184,
"loss": 1.4538,
"step": 58100
},
{
"epoch": 2.91,
"grad_norm": 1.952959418296814,
"learning_rate": 0.0001260934673366834,
"loss": 1.4872,
"step": 58200
},
{
"epoch": 2.92,
"grad_norm": 1.1974529027938843,
"learning_rate": 0.00012579195979899496,
"loss": 1.4487,
"step": 58300
},
{
"epoch": 2.92,
"grad_norm": 1.7099848985671997,
"learning_rate": 0.00012549045226130652,
"loss": 1.4708,
"step": 58400
},
{
"epoch": 2.92,
"grad_norm": 1.2982145547866821,
"learning_rate": 0.00012518894472361808,
"loss": 1.4524,
"step": 58500
},
{
"epoch": 2.93,
"grad_norm": 1.275212049484253,
"learning_rate": 0.00012488743718592964,
"loss": 1.4771,
"step": 58600
},
{
"epoch": 2.94,
"grad_norm": 1.4571611881256104,
"learning_rate": 0.0001245859296482412,
"loss": 1.4454,
"step": 58700
},
{
"epoch": 2.94,
"grad_norm": 1.9355239868164062,
"learning_rate": 0.00012428442211055273,
"loss": 1.4873,
"step": 58800
},
{
"epoch": 2.94,
"grad_norm": 1.2806047201156616,
"learning_rate": 0.00012398291457286432,
"loss": 1.4566,
"step": 58900
},
{
"epoch": 2.95,
"grad_norm": 3.054337739944458,
"learning_rate": 0.00012368442211055277,
"loss": 1.5002,
"step": 59000
},
{
"epoch": 2.95,
"eval_loss": 1.4360119104385376,
"eval_runtime": 40.7087,
"eval_samples_per_second": 24.565,
"eval_steps_per_second": 3.071,
"step": 59000
},
{
"epoch": 2.96,
"grad_norm": 3.0794198513031006,
"learning_rate": 0.0001233829145728643,
"loss": 1.4463,
"step": 59100
},
{
"epoch": 2.96,
"grad_norm": 2.48854923248291,
"learning_rate": 0.00012308140703517586,
"loss": 1.4211,
"step": 59200
},
{
"epoch": 2.96,
"grad_norm": 1.163191556930542,
"learning_rate": 0.00012277989949748742,
"loss": 1.4596,
"step": 59300
},
{
"epoch": 2.97,
"grad_norm": 2.0918381214141846,
"learning_rate": 0.00012247839195979898,
"loss": 1.4713,
"step": 59400
},
{
"epoch": 2.98,
"grad_norm": 2.207432746887207,
"learning_rate": 0.00012217688442211054,
"loss": 1.4747,
"step": 59500
},
{
"epoch": 2.98,
"grad_norm": 2.485342502593994,
"learning_rate": 0.00012187537688442209,
"loss": 1.4531,
"step": 59600
},
{
"epoch": 2.98,
"grad_norm": 1.9726412296295166,
"learning_rate": 0.00012157386934673366,
"loss": 1.4685,
"step": 59700
},
{
"epoch": 2.99,
"grad_norm": 2.211527109146118,
"learning_rate": 0.00012127236180904521,
"loss": 1.488,
"step": 59800
},
{
"epoch": 3.0,
"grad_norm": 2.2727925777435303,
"learning_rate": 0.00012097085427135678,
"loss": 1.4919,
"step": 59900
},
{
"epoch": 3.0,
"grad_norm": 1.9899802207946777,
"learning_rate": 0.00012067236180904521,
"loss": 1.496,
"step": 60000
},
{
"epoch": 3.0,
"eval_loss": 1.480151891708374,
"eval_runtime": 37.1123,
"eval_samples_per_second": 26.945,
"eval_steps_per_second": 3.368,
"step": 60000
},
{
"epoch": 3.0,
"grad_norm": 2.080214023590088,
"learning_rate": 0.00012037085427135678,
"loss": 1.4381,
"step": 60100
},
{
"epoch": 3.01,
"grad_norm": 2.6185152530670166,
"learning_rate": 0.00012006934673366833,
"loss": 1.4701,
"step": 60200
},
{
"epoch": 3.02,
"grad_norm": 2.33237361907959,
"learning_rate": 0.0001197678391959799,
"loss": 1.4754,
"step": 60300
},
{
"epoch": 3.02,
"grad_norm": 1.6651071310043335,
"learning_rate": 0.00011946633165829144,
"loss": 1.4264,
"step": 60400
},
{
"epoch": 3.02,
"grad_norm": 2.55290150642395,
"learning_rate": 0.000119164824120603,
"loss": 1.4643,
"step": 60500
},
{
"epoch": 3.03,
"grad_norm": 2.1360106468200684,
"learning_rate": 0.00011886331658291456,
"loss": 1.4485,
"step": 60600
},
{
"epoch": 3.04,
"grad_norm": 2.012885808944702,
"learning_rate": 0.00011856180904522612,
"loss": 1.4613,
"step": 60700
},
{
"epoch": 3.04,
"grad_norm": 2.1006858348846436,
"learning_rate": 0.00011826030150753767,
"loss": 1.4536,
"step": 60800
},
{
"epoch": 3.04,
"grad_norm": 1.3656415939331055,
"learning_rate": 0.00011795879396984924,
"loss": 1.452,
"step": 60900
},
{
"epoch": 3.05,
"grad_norm": 2.107713460922241,
"learning_rate": 0.00011765728643216079,
"loss": 1.4506,
"step": 61000
},
{
"epoch": 3.05,
"eval_loss": 1.4423640966415405,
"eval_runtime": 36.6897,
"eval_samples_per_second": 27.256,
"eval_steps_per_second": 3.407,
"step": 61000
},
{
"epoch": 3.06,
"grad_norm": 1.7318499088287354,
"learning_rate": 0.00011735577889447236,
"loss": 1.4439,
"step": 61100
},
{
"epoch": 3.06,
"grad_norm": 2.1133487224578857,
"learning_rate": 0.00011705427135678391,
"loss": 1.4401,
"step": 61200
},
{
"epoch": 3.06,
"grad_norm": 1.763225793838501,
"learning_rate": 0.00011675276381909548,
"loss": 1.4196,
"step": 61300
},
{
"epoch": 3.07,
"grad_norm": 3.682509660720825,
"learning_rate": 0.00011645125628140703,
"loss": 1.4355,
"step": 61400
},
{
"epoch": 3.08,
"grad_norm": 1.6236958503723145,
"learning_rate": 0.00011614974874371859,
"loss": 1.4288,
"step": 61500
},
{
"epoch": 3.08,
"grad_norm": 2.428751230239868,
"learning_rate": 0.00011584824120603014,
"loss": 1.4445,
"step": 61600
},
{
"epoch": 3.08,
"grad_norm": 1.9624316692352295,
"learning_rate": 0.0001155467336683417,
"loss": 1.4509,
"step": 61700
},
{
"epoch": 3.09,
"grad_norm": 1.2408591508865356,
"learning_rate": 0.00011524522613065325,
"loss": 1.4604,
"step": 61800
},
{
"epoch": 3.1,
"grad_norm": 1.9674237966537476,
"learning_rate": 0.00011494371859296481,
"loss": 1.4187,
"step": 61900
},
{
"epoch": 3.1,
"grad_norm": 1.1599769592285156,
"learning_rate": 0.00011464221105527637,
"loss": 1.4393,
"step": 62000
},
{
"epoch": 3.1,
"eval_loss": 1.4237370491027832,
"eval_runtime": 36.8952,
"eval_samples_per_second": 27.104,
"eval_steps_per_second": 3.388,
"step": 62000
},
{
"epoch": 3.1,
"grad_norm": 1.2391512393951416,
"learning_rate": 0.00011434070351758793,
"loss": 1.4673,
"step": 62100
},
{
"epoch": 3.11,
"grad_norm": 1.8557316064834595,
"learning_rate": 0.00011403919597989949,
"loss": 1.423,
"step": 62200
},
{
"epoch": 3.12,
"grad_norm": 2.8055520057678223,
"learning_rate": 0.00011373768844221103,
"loss": 1.4722,
"step": 62300
},
{
"epoch": 3.12,
"grad_norm": 3.723037004470825,
"learning_rate": 0.0001134361809045226,
"loss": 1.4313,
"step": 62400
},
{
"epoch": 3.12,
"grad_norm": 1.5366125106811523,
"learning_rate": 0.00011313467336683415,
"loss": 1.4678,
"step": 62500
},
{
"epoch": 3.13,
"grad_norm": 1.5289653539657593,
"learning_rate": 0.00011283316582914573,
"loss": 1.4313,
"step": 62600
},
{
"epoch": 3.13,
"grad_norm": 1.496334195137024,
"learning_rate": 0.00011253165829145727,
"loss": 1.4249,
"step": 62700
},
{
"epoch": 3.14,
"grad_norm": 2.485732078552246,
"learning_rate": 0.00011223015075376884,
"loss": 1.4071,
"step": 62800
},
{
"epoch": 3.15,
"grad_norm": 1.784316897392273,
"learning_rate": 0.00011192864321608039,
"loss": 1.458,
"step": 62900
},
{
"epoch": 3.15,
"grad_norm": 1.126514196395874,
"learning_rate": 0.00011162713567839195,
"loss": 1.4606,
"step": 63000
},
{
"epoch": 3.15,
"eval_loss": 1.4603298902511597,
"eval_runtime": 36.816,
"eval_samples_per_second": 27.162,
"eval_steps_per_second": 3.395,
"step": 63000
},
{
"epoch": 3.15,
"grad_norm": 1.4732190370559692,
"learning_rate": 0.00011132562814070351,
"loss": 1.4118,
"step": 63100
},
{
"epoch": 3.16,
"grad_norm": 2.9530584812164307,
"learning_rate": 0.00011102412060301507,
"loss": 1.4413,
"step": 63200
},
{
"epoch": 3.17,
"grad_norm": 1.5375559329986572,
"learning_rate": 0.00011072261306532661,
"loss": 1.4301,
"step": 63300
},
{
"epoch": 3.17,
"grad_norm": 3.584829568862915,
"learning_rate": 0.00011042110552763819,
"loss": 1.4563,
"step": 63400
},
{
"epoch": 3.17,
"grad_norm": 3.209752082824707,
"learning_rate": 0.00011011959798994973,
"loss": 1.4456,
"step": 63500
},
{
"epoch": 3.18,
"grad_norm": 1.9197559356689453,
"learning_rate": 0.0001098180904522613,
"loss": 1.4324,
"step": 63600
},
{
"epoch": 3.19,
"grad_norm": 1.6928168535232544,
"learning_rate": 0.00010951658291457285,
"loss": 1.4237,
"step": 63700
},
{
"epoch": 3.19,
"grad_norm": 2.0929553508758545,
"learning_rate": 0.00010921507537688443,
"loss": 1.4348,
"step": 63800
},
{
"epoch": 3.19,
"grad_norm": 1.1685271263122559,
"learning_rate": 0.00010891356783919597,
"loss": 1.4297,
"step": 63900
},
{
"epoch": 3.2,
"grad_norm": 1.703434944152832,
"learning_rate": 0.00010861206030150753,
"loss": 1.4324,
"step": 64000
},
{
"epoch": 3.2,
"eval_loss": 1.458383560180664,
"eval_runtime": 37.3628,
"eval_samples_per_second": 26.765,
"eval_steps_per_second": 3.346,
"step": 64000
},
{
"epoch": 3.21,
"grad_norm": 2.0976409912109375,
"learning_rate": 0.00010831055276381909,
"loss": 1.4132,
"step": 64100
},
{
"epoch": 3.21,
"grad_norm": 2.9562056064605713,
"learning_rate": 0.00010801206030150753,
"loss": 1.4216,
"step": 64200
},
{
"epoch": 3.21,
"grad_norm": 3.074629545211792,
"learning_rate": 0.00010771055276381909,
"loss": 1.4099,
"step": 64300
},
{
"epoch": 3.22,
"grad_norm": 2.264169454574585,
"learning_rate": 0.00010740904522613064,
"loss": 1.4065,
"step": 64400
},
{
"epoch": 3.23,
"grad_norm": 1.6872237920761108,
"learning_rate": 0.00010710753768844221,
"loss": 1.4264,
"step": 64500
},
{
"epoch": 3.23,
"grad_norm": 1.8734939098358154,
"learning_rate": 0.00010680603015075375,
"loss": 1.4346,
"step": 64600
},
{
"epoch": 3.23,
"grad_norm": 2.0195159912109375,
"learning_rate": 0.00010650452261306531,
"loss": 1.4557,
"step": 64700
},
{
"epoch": 3.24,
"grad_norm": 1.7375565767288208,
"learning_rate": 0.00010620301507537687,
"loss": 1.4237,
"step": 64800
},
{
"epoch": 3.25,
"grad_norm": 3.346041202545166,
"learning_rate": 0.00010590452261306531,
"loss": 1.4393,
"step": 64900
},
{
"epoch": 3.25,
"grad_norm": 1.8216383457183838,
"learning_rate": 0.00010560301507537687,
"loss": 1.4029,
"step": 65000
},
{
"epoch": 3.25,
"eval_loss": 1.4565457105636597,
"eval_runtime": 36.7819,
"eval_samples_per_second": 27.187,
"eval_steps_per_second": 3.398,
"step": 65000
},
{
"epoch": 3.25,
"grad_norm": 2.628929853439331,
"learning_rate": 0.00010530150753768843,
"loss": 1.3928,
"step": 65100
},
{
"epoch": 3.26,
"grad_norm": 2.476813554763794,
"learning_rate": 0.00010499999999999999,
"loss": 1.4377,
"step": 65200
},
{
"epoch": 3.27,
"grad_norm": 4.745122909545898,
"learning_rate": 0.00010469849246231155,
"loss": 1.4599,
"step": 65300
},
{
"epoch": 3.27,
"grad_norm": 4.166439533233643,
"learning_rate": 0.0001043969849246231,
"loss": 1.4135,
"step": 65400
},
{
"epoch": 3.27,
"grad_norm": 1.3662863969802856,
"learning_rate": 0.00010409547738693467,
"loss": 1.4398,
"step": 65500
},
{
"epoch": 3.28,
"grad_norm": 1.9009519815444946,
"learning_rate": 0.00010379396984924622,
"loss": 1.4354,
"step": 65600
},
{
"epoch": 3.29,
"grad_norm": 1.9361233711242676,
"learning_rate": 0.00010349246231155779,
"loss": 1.382,
"step": 65700
},
{
"epoch": 3.29,
"grad_norm": 2.0850670337677,
"learning_rate": 0.00010319095477386933,
"loss": 1.3983,
"step": 65800
},
{
"epoch": 3.29,
"grad_norm": 3.42114520072937,
"learning_rate": 0.0001028894472361809,
"loss": 1.414,
"step": 65900
},
{
"epoch": 3.3,
"grad_norm": 1.9612314701080322,
"learning_rate": 0.00010258793969849245,
"loss": 1.4494,
"step": 66000
},
{
"epoch": 3.3,
"eval_loss": 1.4201833009719849,
"eval_runtime": 36.8374,
"eval_samples_per_second": 27.146,
"eval_steps_per_second": 3.393,
"step": 66000
},
{
"epoch": 3.31,
"grad_norm": 2.3793907165527344,
"learning_rate": 0.00010228643216080401,
"loss": 1.4567,
"step": 66100
},
{
"epoch": 3.31,
"grad_norm": 2.0631420612335205,
"learning_rate": 0.00010198492462311557,
"loss": 1.4246,
"step": 66200
},
{
"epoch": 3.31,
"grad_norm": 1.973575234413147,
"learning_rate": 0.00010168643216080401,
"loss": 1.4315,
"step": 66300
},
{
"epoch": 3.32,
"grad_norm": 4.41493034362793,
"learning_rate": 0.00010138492462311557,
"loss": 1.3869,
"step": 66400
},
{
"epoch": 3.33,
"grad_norm": 4.229668617248535,
"learning_rate": 0.00010108341708542713,
"loss": 1.4064,
"step": 66500
},
{
"epoch": 3.33,
"grad_norm": 2.8293817043304443,
"learning_rate": 0.00010078190954773868,
"loss": 1.4204,
"step": 66600
},
{
"epoch": 3.33,
"grad_norm": 2.0787456035614014,
"learning_rate": 0.00010048040201005025,
"loss": 1.4011,
"step": 66700
},
{
"epoch": 3.34,
"grad_norm": 2.429809093475342,
"learning_rate": 0.0001001788944723618,
"loss": 1.4205,
"step": 66800
},
{
"epoch": 3.34,
"grad_norm": 1.428712248802185,
"learning_rate": 9.987738693467337e-05,
"loss": 1.4363,
"step": 66900
},
{
"epoch": 3.35,
"grad_norm": 1.6250827312469482,
"learning_rate": 9.957587939698491e-05,
"loss": 1.4225,
"step": 67000
},
{
"epoch": 3.35,
"eval_loss": 1.4025810956954956,
"eval_runtime": 37.6269,
"eval_samples_per_second": 26.577,
"eval_steps_per_second": 3.322,
"step": 67000
},
{
"epoch": 3.35,
"grad_norm": 16.908023834228516,
"learning_rate": 9.927437185929649e-05,
"loss": 1.4019,
"step": 67100
},
{
"epoch": 3.36,
"grad_norm": 2.529090166091919,
"learning_rate": 9.897286432160803e-05,
"loss": 1.4536,
"step": 67200
},
{
"epoch": 3.37,
"grad_norm": 2.6016106605529785,
"learning_rate": 9.867135678391959e-05,
"loss": 1.4144,
"step": 67300
},
{
"epoch": 3.37,
"grad_norm": 6.646603584289551,
"learning_rate": 9.836984924623115e-05,
"loss": 1.4585,
"step": 67400
},
{
"epoch": 3.38,
"grad_norm": 2.8519158363342285,
"learning_rate": 9.80683417085427e-05,
"loss": 1.4281,
"step": 67500
},
{
"epoch": 3.38,
"grad_norm": 2.310377836227417,
"learning_rate": 9.776683417085426e-05,
"loss": 1.42,
"step": 67600
},
{
"epoch": 3.38,
"grad_norm": 1.6958341598510742,
"learning_rate": 9.746532663316582e-05,
"loss": 1.4229,
"step": 67700
},
{
"epoch": 3.39,
"grad_norm": 3.763411045074463,
"learning_rate": 9.716381909547738e-05,
"loss": 1.4212,
"step": 67800
},
{
"epoch": 3.4,
"grad_norm": 3.6792852878570557,
"learning_rate": 9.686231155778894e-05,
"loss": 1.4438,
"step": 67900
},
{
"epoch": 3.4,
"grad_norm": 1.9338295459747314,
"learning_rate": 9.65608040201005e-05,
"loss": 1.4285,
"step": 68000
},
{
"epoch": 3.4,
"eval_loss": 1.4275307655334473,
"eval_runtime": 36.7735,
"eval_samples_per_second": 27.193,
"eval_steps_per_second": 3.399,
"step": 68000
},
{
"epoch": 3.41,
"grad_norm": 2.4784669876098633,
"learning_rate": 9.625929648241204e-05,
"loss": 1.426,
"step": 68100
},
{
"epoch": 3.41,
"grad_norm": 2.1132450103759766,
"learning_rate": 9.595778894472361e-05,
"loss": 1.4243,
"step": 68200
},
{
"epoch": 3.42,
"grad_norm": 1.0781810283660889,
"learning_rate": 9.565628140703516e-05,
"loss": 1.3787,
"step": 68300
},
{
"epoch": 3.42,
"grad_norm": 1.9404023885726929,
"learning_rate": 9.535477386934673e-05,
"loss": 1.4554,
"step": 68400
},
{
"epoch": 3.42,
"grad_norm": 2.5888288021087646,
"learning_rate": 9.505326633165828e-05,
"loss": 1.4224,
"step": 68500
},
{
"epoch": 3.43,
"grad_norm": 1.2692792415618896,
"learning_rate": 9.475175879396985e-05,
"loss": 1.4262,
"step": 68600
},
{
"epoch": 3.44,
"grad_norm": 3.69474196434021,
"learning_rate": 9.44502512562814e-05,
"loss": 1.4348,
"step": 68700
},
{
"epoch": 3.44,
"grad_norm": 2.908108949661255,
"learning_rate": 9.414874371859296e-05,
"loss": 1.3915,
"step": 68800
},
{
"epoch": 3.44,
"grad_norm": 2.764848470687866,
"learning_rate": 9.384723618090452e-05,
"loss": 1.4007,
"step": 68900
},
{
"epoch": 3.45,
"grad_norm": 2.2583723068237305,
"learning_rate": 9.354572864321608e-05,
"loss": 1.4274,
"step": 69000
},
{
"epoch": 3.45,
"eval_loss": 1.4281996488571167,
"eval_runtime": 36.7261,
"eval_samples_per_second": 27.229,
"eval_steps_per_second": 3.404,
"step": 69000
},
{
"epoch": 3.46,
"grad_norm": 2.0189828872680664,
"learning_rate": 9.324422110552762e-05,
"loss": 1.3972,
"step": 69100
},
{
"epoch": 3.46,
"grad_norm": 2.173034191131592,
"learning_rate": 9.29427135678392e-05,
"loss": 1.431,
"step": 69200
},
{
"epoch": 3.46,
"grad_norm": 6.977257251739502,
"learning_rate": 9.264120603015074e-05,
"loss": 1.422,
"step": 69300
},
{
"epoch": 3.47,
"grad_norm": 1.9481549263000488,
"learning_rate": 9.233969849246231e-05,
"loss": 1.424,
"step": 69400
},
{
"epoch": 3.48,
"grad_norm": 1.5347398519515991,
"learning_rate": 9.203819095477386e-05,
"loss": 1.4017,
"step": 69500
},
{
"epoch": 3.48,
"grad_norm": 4.210050582885742,
"learning_rate": 9.173668341708543e-05,
"loss": 1.4088,
"step": 69600
},
{
"epoch": 3.48,
"grad_norm": 1.4653583765029907,
"learning_rate": 9.143517587939698e-05,
"loss": 1.4136,
"step": 69700
},
{
"epoch": 3.49,
"grad_norm": 1.9992530345916748,
"learning_rate": 9.113366834170852e-05,
"loss": 1.4389,
"step": 69800
},
{
"epoch": 3.5,
"grad_norm": 2.986983060836792,
"learning_rate": 9.08321608040201e-05,
"loss": 1.4207,
"step": 69900
},
{
"epoch": 3.5,
"grad_norm": 3.8553383350372314,
"learning_rate": 9.053065326633164e-05,
"loss": 1.4158,
"step": 70000
},
{
"epoch": 3.5,
"eval_loss": 1.4086616039276123,
"eval_runtime": 44.1292,
"eval_samples_per_second": 22.661,
"eval_steps_per_second": 2.833,
"step": 70000
},
{
"epoch": 3.5,
"grad_norm": 1.801664113998413,
"learning_rate": 9.022914572864322e-05,
"loss": 1.4129,
"step": 70100
},
{
"epoch": 3.51,
"grad_norm": 4.6173529624938965,
"learning_rate": 8.992763819095476e-05,
"loss": 1.4052,
"step": 70200
},
{
"epoch": 3.52,
"grad_norm": 2.9409828186035156,
"learning_rate": 8.962914572864322e-05,
"loss": 1.3987,
"step": 70300
},
{
"epoch": 3.52,
"grad_norm": 1.4580490589141846,
"learning_rate": 8.932763819095476e-05,
"loss": 1.4278,
"step": 70400
},
{
"epoch": 3.52,
"grad_norm": 3.432305335998535,
"learning_rate": 8.902613065326632e-05,
"loss": 1.4045,
"step": 70500
},
{
"epoch": 3.53,
"grad_norm": 1.5630279779434204,
"learning_rate": 8.872763819095476e-05,
"loss": 1.4617,
"step": 70600
},
{
"epoch": 3.54,
"grad_norm": 2.1527063846588135,
"learning_rate": 8.842613065326632e-05,
"loss": 1.4445,
"step": 70700
},
{
"epoch": 3.54,
"grad_norm": 4.876186847686768,
"learning_rate": 8.812462311557788e-05,
"loss": 1.4025,
"step": 70800
},
{
"epoch": 3.54,
"grad_norm": 2.359768867492676,
"learning_rate": 8.782311557788944e-05,
"loss": 1.3849,
"step": 70900
},
{
"epoch": 3.55,
"grad_norm": 1.0750428438186646,
"learning_rate": 8.7521608040201e-05,
"loss": 1.3715,
"step": 71000
},
{
"epoch": 3.55,
"eval_loss": 1.4002715349197388,
"eval_runtime": 44.6209,
"eval_samples_per_second": 22.411,
"eval_steps_per_second": 2.801,
"step": 71000
},
{
"epoch": 3.56,
"grad_norm": 3.729515552520752,
"learning_rate": 8.722010050251256e-05,
"loss": 1.4194,
"step": 71100
},
{
"epoch": 3.56,
"grad_norm": 3.288299560546875,
"learning_rate": 8.69185929648241e-05,
"loss": 1.4096,
"step": 71200
},
{
"epoch": 3.56,
"grad_norm": 1.2813684940338135,
"learning_rate": 8.661708542713568e-05,
"loss": 1.4256,
"step": 71300
},
{
"epoch": 3.57,
"grad_norm": 1.7768347263336182,
"learning_rate": 8.631557788944722e-05,
"loss": 1.4021,
"step": 71400
},
{
"epoch": 3.58,
"grad_norm": 1.5083990097045898,
"learning_rate": 8.60140703517588e-05,
"loss": 1.3925,
"step": 71500
},
{
"epoch": 3.58,
"grad_norm": 2.056110382080078,
"learning_rate": 8.571256281407034e-05,
"loss": 1.3997,
"step": 71600
},
{
"epoch": 3.58,
"grad_norm": 1.591378927230835,
"learning_rate": 8.54110552763819e-05,
"loss": 1.382,
"step": 71700
},
{
"epoch": 3.59,
"grad_norm": 2.8046460151672363,
"learning_rate": 8.510954773869346e-05,
"loss": 1.4041,
"step": 71800
},
{
"epoch": 3.59,
"grad_norm": 3.3818118572235107,
"learning_rate": 8.480804020100502e-05,
"loss": 1.3774,
"step": 71900
},
{
"epoch": 3.6,
"grad_norm": 1.9918705224990845,
"learning_rate": 8.450653266331658e-05,
"loss": 1.4071,
"step": 72000
},
{
"epoch": 3.6,
"eval_loss": 1.3878278732299805,
"eval_runtime": 44.5957,
"eval_samples_per_second": 22.424,
"eval_steps_per_second": 2.803,
"step": 72000
},
{
"epoch": 3.6,
"grad_norm": 2.911811351776123,
"learning_rate": 8.420502512562814e-05,
"loss": 1.3809,
"step": 72100
},
{
"epoch": 3.61,
"grad_norm": 1.6014336347579956,
"learning_rate": 8.390351758793968e-05,
"loss": 1.434,
"step": 72200
},
{
"epoch": 3.62,
"grad_norm": 2.2394659519195557,
"learning_rate": 8.360201005025126e-05,
"loss": 1.4076,
"step": 72300
},
{
"epoch": 3.62,
"grad_norm": 2.3014962673187256,
"learning_rate": 8.33005025125628e-05,
"loss": 1.4367,
"step": 72400
},
{
"epoch": 3.62,
"grad_norm": 3.4550154209136963,
"learning_rate": 8.299899497487438e-05,
"loss": 1.3938,
"step": 72500
},
{
"epoch": 3.63,
"grad_norm": 2.599987268447876,
"learning_rate": 8.269748743718592e-05,
"loss": 1.3873,
"step": 72600
},
{
"epoch": 3.63,
"grad_norm": 2.6119203567504883,
"learning_rate": 8.23959798994975e-05,
"loss": 1.4069,
"step": 72700
},
{
"epoch": 3.64,
"grad_norm": 1.702276349067688,
"learning_rate": 8.209447236180904e-05,
"loss": 1.3923,
"step": 72800
},
{
"epoch": 3.65,
"grad_norm": 2.805104970932007,
"learning_rate": 8.179296482412059e-05,
"loss": 1.386,
"step": 72900
},
{
"epoch": 3.65,
"grad_norm": 2.306410312652588,
"learning_rate": 8.149145728643216e-05,
"loss": 1.4118,
"step": 73000
},
{
"epoch": 3.65,
"eval_loss": 1.387115716934204,
"eval_runtime": 37.4562,
"eval_samples_per_second": 26.698,
"eval_steps_per_second": 3.337,
"step": 73000
},
{
"epoch": 3.66,
"grad_norm": 2.5508077144622803,
"learning_rate": 8.11899497487437e-05,
"loss": 1.3963,
"step": 73100
},
{
"epoch": 3.66,
"grad_norm": 1.1488244533538818,
"learning_rate": 8.088844221105527e-05,
"loss": 1.4107,
"step": 73200
},
{
"epoch": 3.67,
"grad_norm": 3.1272361278533936,
"learning_rate": 8.058693467336682e-05,
"loss": 1.3444,
"step": 73300
},
{
"epoch": 3.67,
"grad_norm": 2.3045852184295654,
"learning_rate": 8.028542713567838e-05,
"loss": 1.3874,
"step": 73400
},
{
"epoch": 3.67,
"grad_norm": 2.2844760417938232,
"learning_rate": 7.998391959798994e-05,
"loss": 1.3748,
"step": 73500
},
{
"epoch": 3.68,
"grad_norm": 2.9295897483825684,
"learning_rate": 7.96824120603015e-05,
"loss": 1.3786,
"step": 73600
},
{
"epoch": 3.69,
"grad_norm": 1.5064588785171509,
"learning_rate": 7.938090452261305e-05,
"loss": 1.4079,
"step": 73700
},
{
"epoch": 3.69,
"grad_norm": 1.8818271160125732,
"learning_rate": 7.907939698492462e-05,
"loss": 1.3954,
"step": 73800
},
{
"epoch": 3.69,
"grad_norm": 1.1885104179382324,
"learning_rate": 7.877788944723617e-05,
"loss": 1.3836,
"step": 73900
},
{
"epoch": 3.7,
"grad_norm": 3.0773730278015137,
"learning_rate": 7.847638190954774e-05,
"loss": 1.3724,
"step": 74000
},
{
"epoch": 3.7,
"eval_loss": 1.3931760787963867,
"eval_runtime": 37.2623,
"eval_samples_per_second": 26.837,
"eval_steps_per_second": 3.355,
"step": 74000
},
{
"epoch": 3.71,
"grad_norm": 1.4827876091003418,
"learning_rate": 7.817487437185929e-05,
"loss": 1.4122,
"step": 74100
},
{
"epoch": 3.71,
"grad_norm": 1.8637442588806152,
"learning_rate": 7.787336683417086e-05,
"loss": 1.3466,
"step": 74200
},
{
"epoch": 3.71,
"grad_norm": 2.286088228225708,
"learning_rate": 7.75718592964824e-05,
"loss": 1.3815,
"step": 74300
},
{
"epoch": 3.72,
"grad_norm": 1.8894226551055908,
"learning_rate": 7.727035175879396e-05,
"loss": 1.4096,
"step": 74400
},
{
"epoch": 3.73,
"grad_norm": 1.9993146657943726,
"learning_rate": 7.696884422110552e-05,
"loss": 1.4196,
"step": 74500
},
{
"epoch": 3.73,
"grad_norm": 2.170727491378784,
"learning_rate": 7.666733668341708e-05,
"loss": 1.4034,
"step": 74600
},
{
"epoch": 3.73,
"grad_norm": 1.9965826272964478,
"learning_rate": 7.636582914572863e-05,
"loss": 1.3971,
"step": 74700
},
{
"epoch": 3.74,
"grad_norm": 2.5607786178588867,
"learning_rate": 7.60643216080402e-05,
"loss": 1.4116,
"step": 74800
},
{
"epoch": 3.75,
"grad_norm": 1.5363885164260864,
"learning_rate": 7.576281407035175e-05,
"loss": 1.3932,
"step": 74900
},
{
"epoch": 3.75,
"grad_norm": 1.9062122106552124,
"learning_rate": 7.546130653266332e-05,
"loss": 1.3625,
"step": 75000
},
{
"epoch": 3.75,
"eval_loss": 1.3875294923782349,
"eval_runtime": 37.1595,
"eval_samples_per_second": 26.911,
"eval_steps_per_second": 3.364,
"step": 75000
},
{
"epoch": 3.75,
"grad_norm": 2.8305203914642334,
"learning_rate": 7.516281407035175e-05,
"loss": 1.3907,
"step": 75100
},
{
"epoch": 3.76,
"grad_norm": 1.6526367664337158,
"learning_rate": 7.486130653266331e-05,
"loss": 1.3605,
"step": 75200
},
{
"epoch": 3.77,
"grad_norm": 3.5784194469451904,
"learning_rate": 7.455979899497487e-05,
"loss": 1.3833,
"step": 75300
},
{
"epoch": 3.77,
"grad_norm": 0.9500262141227722,
"learning_rate": 7.425829145728643e-05,
"loss": 1.3548,
"step": 75400
},
{
"epoch": 3.77,
"grad_norm": 2.8620994091033936,
"learning_rate": 7.395678391959799e-05,
"loss": 1.3829,
"step": 75500
},
{
"epoch": 3.78,
"grad_norm": 1.9316253662109375,
"learning_rate": 7.365527638190954e-05,
"loss": 1.3801,
"step": 75600
},
{
"epoch": 3.79,
"grad_norm": 1.5958226919174194,
"learning_rate": 7.33537688442211e-05,
"loss": 1.3837,
"step": 75700
},
{
"epoch": 3.79,
"grad_norm": 1.995013952255249,
"learning_rate": 7.305226130653266e-05,
"loss": 1.3933,
"step": 75800
},
{
"epoch": 3.79,
"grad_norm": 1.4954237937927246,
"learning_rate": 7.275075376884422e-05,
"loss": 1.4223,
"step": 75900
},
{
"epoch": 3.8,
"grad_norm": 1.363682508468628,
"learning_rate": 7.244924623115577e-05,
"loss": 1.3811,
"step": 76000
},
{
"epoch": 3.8,
"eval_loss": 1.3870151042938232,
"eval_runtime": 37.1537,
"eval_samples_per_second": 26.915,
"eval_steps_per_second": 3.364,
"step": 76000
},
{
"epoch": 3.81,
"grad_norm": 1.6989458799362183,
"learning_rate": 7.214773869346733e-05,
"loss": 1.3755,
"step": 76100
},
{
"epoch": 3.81,
"grad_norm": 2.042025327682495,
"learning_rate": 7.184623115577889e-05,
"loss": 1.3641,
"step": 76200
},
{
"epoch": 3.81,
"grad_norm": 1.4355963468551636,
"learning_rate": 7.154472361809045e-05,
"loss": 1.3675,
"step": 76300
},
{
"epoch": 3.82,
"grad_norm": 1.0404947996139526,
"learning_rate": 7.124321608040201e-05,
"loss": 1.3755,
"step": 76400
},
{
"epoch": 3.83,
"grad_norm": 2.588839530944824,
"learning_rate": 7.094170854271357e-05,
"loss": 1.369,
"step": 76500
},
{
"epoch": 3.83,
"grad_norm": 2.2656760215759277,
"learning_rate": 7.064020100502511e-05,
"loss": 1.4254,
"step": 76600
},
{
"epoch": 3.83,
"grad_norm": 2.1572184562683105,
"learning_rate": 7.033869346733667e-05,
"loss": 1.3928,
"step": 76700
},
{
"epoch": 3.84,
"grad_norm": 1.9283751249313354,
"learning_rate": 7.003718592964823e-05,
"loss": 1.3502,
"step": 76800
},
{
"epoch": 3.84,
"grad_norm": 2.215160846710205,
"learning_rate": 6.973567839195979e-05,
"loss": 1.3701,
"step": 76900
},
{
"epoch": 3.85,
"grad_norm": 1.9962304830551147,
"learning_rate": 6.943417085427135e-05,
"loss": 1.3353,
"step": 77000
},
{
"epoch": 3.85,
"eval_loss": 1.3705061674118042,
"eval_runtime": 36.8621,
"eval_samples_per_second": 27.128,
"eval_steps_per_second": 3.391,
"step": 77000
},
{
"epoch": 3.85,
"grad_norm": 1.818708062171936,
"learning_rate": 6.913266331658291e-05,
"loss": 1.4337,
"step": 77100
},
{
"epoch": 3.86,
"grad_norm": 2.5512075424194336,
"learning_rate": 6.883115577889447e-05,
"loss": 1.3923,
"step": 77200
},
{
"epoch": 3.87,
"grad_norm": 2.3595049381256104,
"learning_rate": 6.852964824120603e-05,
"loss": 1.3626,
"step": 77300
},
{
"epoch": 3.87,
"grad_norm": 2.033395767211914,
"learning_rate": 6.822814070351757e-05,
"loss": 1.4061,
"step": 77400
},
{
"epoch": 3.88,
"grad_norm": 1.8597971200942993,
"learning_rate": 6.792663316582913e-05,
"loss": 1.3672,
"step": 77500
},
{
"epoch": 3.88,
"grad_norm": 6.214807987213135,
"learning_rate": 6.762512562814069e-05,
"loss": 1.3817,
"step": 77600
},
{
"epoch": 3.88,
"grad_norm": 2.384474515914917,
"learning_rate": 6.732361809045225e-05,
"loss": 1.3811,
"step": 77700
},
{
"epoch": 3.89,
"grad_norm": 2.6242575645446777,
"learning_rate": 6.702211055276381e-05,
"loss": 1.3883,
"step": 77800
},
{
"epoch": 3.9,
"grad_norm": 1.4827545881271362,
"learning_rate": 6.672060301507537e-05,
"loss": 1.3755,
"step": 77900
},
{
"epoch": 3.9,
"grad_norm": 2.150613307952881,
"learning_rate": 6.641909547738693e-05,
"loss": 1.3937,
"step": 78000
},
{
"epoch": 3.9,
"eval_loss": 1.379770278930664,
"eval_runtime": 37.0122,
"eval_samples_per_second": 27.018,
"eval_steps_per_second": 3.377,
"step": 78000
},
{
"epoch": 3.91,
"grad_norm": 1.4805549383163452,
"learning_rate": 6.611758793969849e-05,
"loss": 1.365,
"step": 78100
},
{
"epoch": 3.91,
"grad_norm": 1.699484944343567,
"learning_rate": 6.581608040201005e-05,
"loss": 1.3893,
"step": 78200
},
{
"epoch": 3.92,
"grad_norm": 2.039008140563965,
"learning_rate": 6.551758793969849e-05,
"loss": 1.3601,
"step": 78300
},
{
"epoch": 3.92,
"grad_norm": 1.3309062719345093,
"learning_rate": 6.521608040201005e-05,
"loss": 1.4056,
"step": 78400
},
{
"epoch": 3.92,
"grad_norm": 2.382688045501709,
"learning_rate": 6.491758793969849e-05,
"loss": 1.3758,
"step": 78500
},
{
"epoch": 3.93,
"grad_norm": 3.1065239906311035,
"learning_rate": 6.461608040201005e-05,
"loss": 1.3725,
"step": 78600
},
{
"epoch": 3.94,
"grad_norm": 2.193957805633545,
"learning_rate": 6.431457286432161e-05,
"loss": 1.3347,
"step": 78700
},
{
"epoch": 3.94,
"grad_norm": 1.8022537231445312,
"learning_rate": 6.401306532663317e-05,
"loss": 1.3568,
"step": 78800
},
{
"epoch": 3.94,
"grad_norm": 1.7433044910430908,
"learning_rate": 6.371155778894473e-05,
"loss": 1.3643,
"step": 78900
},
{
"epoch": 3.95,
"grad_norm": 3.883768081665039,
"learning_rate": 6.341005025125627e-05,
"loss": 1.3158,
"step": 79000
},
{
"epoch": 3.95,
"eval_loss": 1.37774658203125,
"eval_runtime": 36.9703,
"eval_samples_per_second": 27.049,
"eval_steps_per_second": 3.381,
"step": 79000
},
{
"epoch": 3.96,
"grad_norm": 2.9996466636657715,
"learning_rate": 6.310854271356783e-05,
"loss": 1.3295,
"step": 79100
},
{
"epoch": 3.96,
"grad_norm": 1.2524511814117432,
"learning_rate": 6.280703517587939e-05,
"loss": 1.3362,
"step": 79200
},
{
"epoch": 3.96,
"grad_norm": 1.8024784326553345,
"learning_rate": 6.250552763819095e-05,
"loss": 1.3857,
"step": 79300
},
{
"epoch": 3.97,
"grad_norm": 1.3988703489303589,
"learning_rate": 6.220402010050251e-05,
"loss": 1.3463,
"step": 79400
},
{
"epoch": 3.98,
"grad_norm": 2.499194622039795,
"learning_rate": 6.190251256281407e-05,
"loss": 1.3761,
"step": 79500
},
{
"epoch": 3.98,
"grad_norm": 2.799959897994995,
"learning_rate": 6.160100502512562e-05,
"loss": 1.3995,
"step": 79600
},
{
"epoch": 3.98,
"grad_norm": 1.2367918491363525,
"learning_rate": 6.129949748743717e-05,
"loss": 1.3789,
"step": 79700
},
{
"epoch": 3.99,
"grad_norm": 1.7651324272155762,
"learning_rate": 6.0997989949748734e-05,
"loss": 1.3691,
"step": 79800
},
{
"epoch": 4.0,
"grad_norm": 1.8847932815551758,
"learning_rate": 6.0696482412060293e-05,
"loss": 1.353,
"step": 79900
},
{
"epoch": 4.0,
"grad_norm": 1.8860230445861816,
"learning_rate": 6.039497487437185e-05,
"loss": 1.3168,
"step": 80000
},
{
"epoch": 4.0,
"eval_loss": 1.3387900590896606,
"eval_runtime": 37.0136,
"eval_samples_per_second": 27.017,
"eval_steps_per_second": 3.377,
"step": 80000
},
{
"epoch": 4.0,
"grad_norm": 2.109632968902588,
"learning_rate": 6.009346733668341e-05,
"loss": 1.3518,
"step": 80100
},
{
"epoch": 4.01,
"grad_norm": 3.0571937561035156,
"learning_rate": 5.9791959798994965e-05,
"loss": 1.3664,
"step": 80200
},
{
"epoch": 4.01,
"grad_norm": 1.8600945472717285,
"learning_rate": 5.9490452261306525e-05,
"loss": 1.3252,
"step": 80300
},
{
"epoch": 4.02,
"grad_norm": 2.461089611053467,
"learning_rate": 5.9188944723618084e-05,
"loss": 1.3507,
"step": 80400
},
{
"epoch": 4.03,
"grad_norm": 3.1424946784973145,
"learning_rate": 5.8887437185929643e-05,
"loss": 1.3414,
"step": 80500
},
{
"epoch": 4.03,
"grad_norm": 1.701357126235962,
"learning_rate": 5.85859296482412e-05,
"loss": 1.3047,
"step": 80600
},
{
"epoch": 4.04,
"grad_norm": 2.331779956817627,
"learning_rate": 5.8284422110552756e-05,
"loss": 1.3309,
"step": 80700
},
{
"epoch": 4.04,
"grad_norm": 5.718899726867676,
"learning_rate": 5.7982914572864315e-05,
"loss": 1.3796,
"step": 80800
},
{
"epoch": 4.04,
"grad_norm": 1.1299536228179932,
"learning_rate": 5.7681407035175874e-05,
"loss": 1.3517,
"step": 80900
},
{
"epoch": 4.05,
"grad_norm": 2.256490468978882,
"learning_rate": 5.7379899497487434e-05,
"loss": 1.35,
"step": 81000
},
{
"epoch": 4.05,
"eval_loss": 1.3527089357376099,
"eval_runtime": 37.0846,
"eval_samples_per_second": 26.965,
"eval_steps_per_second": 3.371,
"step": 81000
},
{
"epoch": 4.05,
"grad_norm": 2.3526551723480225,
"learning_rate": 5.707839195979899e-05,
"loss": 1.3327,
"step": 81100
},
{
"epoch": 4.06,
"grad_norm": 2.987255573272705,
"learning_rate": 5.677688442211055e-05,
"loss": 1.3369,
"step": 81200
},
{
"epoch": 4.07,
"grad_norm": 3.35562801361084,
"learning_rate": 5.6475376884422105e-05,
"loss": 1.3583,
"step": 81300
},
{
"epoch": 4.07,
"grad_norm": 2.192720651626587,
"learning_rate": 5.6173869346733665e-05,
"loss": 1.3405,
"step": 81400
},
{
"epoch": 4.08,
"grad_norm": 1.91267728805542,
"learning_rate": 5.5872361809045224e-05,
"loss": 1.3281,
"step": 81500
},
{
"epoch": 4.08,
"grad_norm": 2.5109171867370605,
"learning_rate": 5.5570854271356784e-05,
"loss": 1.3126,
"step": 81600
},
{
"epoch": 4.08,
"grad_norm": 1.4430352449417114,
"learning_rate": 5.526934673366834e-05,
"loss": 1.2983,
"step": 81700
},
{
"epoch": 4.09,
"grad_norm": 1.873565912246704,
"learning_rate": 5.4967839195979896e-05,
"loss": 1.292,
"step": 81800
},
{
"epoch": 4.09,
"grad_norm": 1.6436150074005127,
"learning_rate": 5.466633165829145e-05,
"loss": 1.3268,
"step": 81900
},
{
"epoch": 4.1,
"grad_norm": 3.9100093841552734,
"learning_rate": 5.436482412060301e-05,
"loss": 1.3388,
"step": 82000
},
{
"epoch": 4.1,
"eval_loss": 1.3313664197921753,
"eval_runtime": 37.2377,
"eval_samples_per_second": 26.855,
"eval_steps_per_second": 3.357,
"step": 82000
},
{
"epoch": 4.11,
"grad_norm": 3.0072224140167236,
"learning_rate": 5.406331658291457e-05,
"loss": 1.2984,
"step": 82100
},
{
"epoch": 4.11,
"grad_norm": 1.3772695064544678,
"learning_rate": 5.376180904522612e-05,
"loss": 1.3182,
"step": 82200
},
{
"epoch": 4.12,
"grad_norm": 3.444744110107422,
"learning_rate": 5.346030150753768e-05,
"loss": 1.3237,
"step": 82300
},
{
"epoch": 4.12,
"grad_norm": 2.719045400619507,
"learning_rate": 5.315879396984924e-05,
"loss": 1.349,
"step": 82400
},
{
"epoch": 4.12,
"grad_norm": 4.052174091339111,
"learning_rate": 5.28572864321608e-05,
"loss": 1.3504,
"step": 82500
},
{
"epoch": 4.13,
"grad_norm": 1.3267349004745483,
"learning_rate": 5.255577889447236e-05,
"loss": 1.3038,
"step": 82600
},
{
"epoch": 4.13,
"grad_norm": 2.1242218017578125,
"learning_rate": 5.225427135678391e-05,
"loss": 1.3087,
"step": 82700
},
{
"epoch": 4.14,
"grad_norm": 3.2896034717559814,
"learning_rate": 5.195276381909547e-05,
"loss": 1.31,
"step": 82800
},
{
"epoch": 4.14,
"grad_norm": 2.3824546337127686,
"learning_rate": 5.165125628140703e-05,
"loss": 1.2972,
"step": 82900
},
{
"epoch": 4.15,
"grad_norm": 2.0106704235076904,
"learning_rate": 5.134974874371859e-05,
"loss": 1.3578,
"step": 83000
},
{
"epoch": 4.15,
"eval_loss": 1.3431867361068726,
"eval_runtime": 36.9495,
"eval_samples_per_second": 27.064,
"eval_steps_per_second": 3.383,
"step": 83000
},
{
"epoch": 4.16,
"grad_norm": 2.094095230102539,
"learning_rate": 5.104824120603015e-05,
"loss": 1.3172,
"step": 83100
},
{
"epoch": 4.16,
"grad_norm": 2.0634231567382812,
"learning_rate": 5.074673366834171e-05,
"loss": 1.3381,
"step": 83200
},
{
"epoch": 4.17,
"grad_norm": 4.8768486976623535,
"learning_rate": 5.044522613065326e-05,
"loss": 1.3148,
"step": 83300
},
{
"epoch": 4.17,
"grad_norm": 4.0618486404418945,
"learning_rate": 5.014371859296482e-05,
"loss": 1.3635,
"step": 83400
},
{
"epoch": 4.17,
"grad_norm": 1.6911975145339966,
"learning_rate": 4.984221105527638e-05,
"loss": 1.3333,
"step": 83500
},
{
"epoch": 4.18,
"grad_norm": 2.164687156677246,
"learning_rate": 4.954070351758794e-05,
"loss": 1.3391,
"step": 83600
},
{
"epoch": 4.18,
"grad_norm": 1.3353967666625977,
"learning_rate": 4.92391959798995e-05,
"loss": 1.2989,
"step": 83700
},
{
"epoch": 4.19,
"grad_norm": 3.4308252334594727,
"learning_rate": 4.893768844221105e-05,
"loss": 1.2966,
"step": 83800
},
{
"epoch": 4.2,
"grad_norm": 2.58341121673584,
"learning_rate": 4.863618090452261e-05,
"loss": 1.3036,
"step": 83900
},
{
"epoch": 4.2,
"grad_norm": 2.0035691261291504,
"learning_rate": 4.833467336683417e-05,
"loss": 1.3028,
"step": 84000
},
{
"epoch": 4.2,
"eval_loss": 1.3128948211669922,
"eval_runtime": 36.9125,
"eval_samples_per_second": 27.091,
"eval_steps_per_second": 3.386,
"step": 84000
},
{
"epoch": 4.21,
"grad_norm": 3.858201503753662,
"learning_rate": 4.803316582914573e-05,
"loss": 1.3459,
"step": 84100
},
{
"epoch": 4.21,
"grad_norm": 1.9417260885238647,
"learning_rate": 4.7731658291457275e-05,
"loss": 1.3243,
"step": 84200
},
{
"epoch": 4.21,
"grad_norm": 3.000342607498169,
"learning_rate": 4.7430150753768835e-05,
"loss": 1.3175,
"step": 84300
},
{
"epoch": 4.22,
"grad_norm": 5.205733776092529,
"learning_rate": 4.7128643216080394e-05,
"loss": 1.3486,
"step": 84400
},
{
"epoch": 4.22,
"grad_norm": 3.2263059616088867,
"learning_rate": 4.683015075376885e-05,
"loss": 1.3103,
"step": 84500
},
{
"epoch": 4.23,
"grad_norm": 2.210460662841797,
"learning_rate": 4.652864321608039e-05,
"loss": 1.339,
"step": 84600
},
{
"epoch": 4.24,
"grad_norm": 3.195037364959717,
"learning_rate": 4.622713567839195e-05,
"loss": 1.3472,
"step": 84700
},
{
"epoch": 4.24,
"grad_norm": 2.5944504737854004,
"learning_rate": 4.592562814070351e-05,
"loss": 1.3159,
"step": 84800
},
{
"epoch": 4.25,
"grad_norm": 2.6246960163116455,
"learning_rate": 4.562412060301507e-05,
"loss": 1.2896,
"step": 84900
},
{
"epoch": 4.25,
"grad_norm": 2.556218385696411,
"learning_rate": 4.5322613065326624e-05,
"loss": 1.3187,
"step": 85000
},
{
"epoch": 4.25,
"eval_loss": 1.324548363685608,
"eval_runtime": 36.9422,
"eval_samples_per_second": 27.069,
"eval_steps_per_second": 3.384,
"step": 85000
},
{
"epoch": 4.25,
"grad_norm": 2.348661184310913,
"learning_rate": 4.5021105527638184e-05,
"loss": 1.3004,
"step": 85100
},
{
"epoch": 4.26,
"grad_norm": 2.0715219974517822,
"learning_rate": 4.471959798994974e-05,
"loss": 1.3265,
"step": 85200
},
{
"epoch": 4.26,
"grad_norm": 2.5301461219787598,
"learning_rate": 4.44180904522613e-05,
"loss": 1.3187,
"step": 85300
},
{
"epoch": 4.27,
"grad_norm": 5.115157604217529,
"learning_rate": 4.411658291457286e-05,
"loss": 1.3179,
"step": 85400
},
{
"epoch": 4.28,
"grad_norm": 2.671520471572876,
"learning_rate": 4.3815075376884415e-05,
"loss": 1.2988,
"step": 85500
},
{
"epoch": 4.28,
"grad_norm": 2.0337090492248535,
"learning_rate": 4.3513567839195974e-05,
"loss": 1.2781,
"step": 85600
},
{
"epoch": 4.29,
"grad_norm": 2.735685348510742,
"learning_rate": 4.3212060301507534e-05,
"loss": 1.3167,
"step": 85700
},
{
"epoch": 4.29,
"grad_norm": 2.573694944381714,
"learning_rate": 4.291055276381909e-05,
"loss": 1.3057,
"step": 85800
},
{
"epoch": 4.29,
"grad_norm": 1.5959097146987915,
"learning_rate": 4.260904522613065e-05,
"loss": 1.327,
"step": 85900
},
{
"epoch": 4.3,
"grad_norm": 2.1494078636169434,
"learning_rate": 4.2307537688442205e-05,
"loss": 1.3291,
"step": 86000
},
{
"epoch": 4.3,
"eval_loss": 1.3199845552444458,
"eval_runtime": 36.8676,
"eval_samples_per_second": 27.124,
"eval_steps_per_second": 3.391,
"step": 86000
},
{
"epoch": 4.3,
"grad_norm": 1.7684657573699951,
"learning_rate": 4.2006030150753765e-05,
"loss": 1.3341,
"step": 86100
},
{
"epoch": 4.31,
"grad_norm": 2.656754732131958,
"learning_rate": 4.1704522613065324e-05,
"loss": 1.3275,
"step": 86200
},
{
"epoch": 4.32,
"grad_norm": 1.330156683921814,
"learning_rate": 4.1403015075376883e-05,
"loss": 1.3365,
"step": 86300
},
{
"epoch": 4.32,
"grad_norm": 3.0051920413970947,
"learning_rate": 4.110150753768844e-05,
"loss": 1.303,
"step": 86400
},
{
"epoch": 4.33,
"grad_norm": 1.494343638420105,
"learning_rate": 4.08e-05,
"loss": 1.2839,
"step": 86500
},
{
"epoch": 4.33,
"grad_norm": 3.456125020980835,
"learning_rate": 4.0498492462311555e-05,
"loss": 1.2964,
"step": 86600
},
{
"epoch": 4.33,
"grad_norm": 2.114023447036743,
"learning_rate": 4.0196984924623115e-05,
"loss": 1.3397,
"step": 86700
},
{
"epoch": 4.34,
"grad_norm": 2.5307486057281494,
"learning_rate": 3.9895477386934674e-05,
"loss": 1.2841,
"step": 86800
},
{
"epoch": 4.34,
"grad_norm": 1.866807222366333,
"learning_rate": 3.9596984924623113e-05,
"loss": 1.3218,
"step": 86900
},
{
"epoch": 4.35,
"grad_norm": 1.9067096710205078,
"learning_rate": 3.929547738693467e-05,
"loss": 1.3257,
"step": 87000
},
{
"epoch": 4.35,
"eval_loss": 1.3586480617523193,
"eval_runtime": 37.2768,
"eval_samples_per_second": 26.826,
"eval_steps_per_second": 3.353,
"step": 87000
},
{
"epoch": 4.36,
"grad_norm": 3.3964035511016846,
"learning_rate": 3.899396984924623e-05,
"loss": 1.3385,
"step": 87100
},
{
"epoch": 4.36,
"grad_norm": 3.567638874053955,
"learning_rate": 3.869246231155778e-05,
"loss": 1.3281,
"step": 87200
},
{
"epoch": 4.37,
"grad_norm": 2.170928955078125,
"learning_rate": 3.839095477386934e-05,
"loss": 1.3159,
"step": 87300
},
{
"epoch": 4.37,
"grad_norm": 2.326664924621582,
"learning_rate": 3.80894472361809e-05,
"loss": 1.3499,
"step": 87400
},
{
"epoch": 4.38,
"grad_norm": 2.8397111892700195,
"learning_rate": 3.778793969849246e-05,
"loss": 1.349,
"step": 87500
},
{
"epoch": 4.38,
"grad_norm": 2.2665247917175293,
"learning_rate": 3.7486432160804016e-05,
"loss": 1.3172,
"step": 87600
},
{
"epoch": 4.38,
"grad_norm": 4.748266696929932,
"learning_rate": 3.718793969849246e-05,
"loss": 1.2857,
"step": 87700
},
{
"epoch": 4.39,
"grad_norm": 1.4008738994598389,
"learning_rate": 3.6886432160804015e-05,
"loss": 1.276,
"step": 87800
},
{
"epoch": 4.39,
"grad_norm": 3.0103044509887695,
"learning_rate": 3.6584924623115574e-05,
"loss": 1.3114,
"step": 87900
},
{
"epoch": 4.4,
"grad_norm": 1.6098419427871704,
"learning_rate": 3.6283417085427134e-05,
"loss": 1.2995,
"step": 88000
},
{
"epoch": 4.4,
"eval_loss": 1.2803822755813599,
"eval_runtime": 37.2497,
"eval_samples_per_second": 26.846,
"eval_steps_per_second": 3.356,
"step": 88000
},
{
"epoch": 4.41,
"grad_norm": 1.6589020490646362,
"learning_rate": 3.5981909547738693e-05,
"loss": 1.3214,
"step": 88100
},
{
"epoch": 4.41,
"grad_norm": 3.2466797828674316,
"learning_rate": 3.5680402010050246e-05,
"loss": 1.2685,
"step": 88200
},
{
"epoch": 4.42,
"grad_norm": 3.273249626159668,
"learning_rate": 3.5378894472361806e-05,
"loss": 1.2773,
"step": 88300
},
{
"epoch": 4.42,
"grad_norm": 2.712141513824463,
"learning_rate": 3.5077386934673365e-05,
"loss": 1.3107,
"step": 88400
},
{
"epoch": 4.42,
"grad_norm": 2.671769857406616,
"learning_rate": 3.477587939698492e-05,
"loss": 1.3149,
"step": 88500
},
{
"epoch": 4.43,
"grad_norm": 1.6911038160324097,
"learning_rate": 3.447437185929648e-05,
"loss": 1.312,
"step": 88600
},
{
"epoch": 4.43,
"grad_norm": 2.3690972328186035,
"learning_rate": 3.4172864321608037e-05,
"loss": 1.2826,
"step": 88700
},
{
"epoch": 4.44,
"grad_norm": 2.1747915744781494,
"learning_rate": 3.3871356783919596e-05,
"loss": 1.3187,
"step": 88800
},
{
"epoch": 4.45,
"grad_norm": 1.7691452503204346,
"learning_rate": 3.3569849246231155e-05,
"loss": 1.3356,
"step": 88900
},
{
"epoch": 4.45,
"grad_norm": 2.9473211765289307,
"learning_rate": 3.3268341708542715e-05,
"loss": 1.2721,
"step": 89000
},
{
"epoch": 4.45,
"eval_loss": 1.3034113645553589,
"eval_runtime": 37.3017,
"eval_samples_per_second": 26.808,
"eval_steps_per_second": 3.351,
"step": 89000
},
{
"epoch": 4.46,
"grad_norm": 2.840188503265381,
"learning_rate": 3.296683417085427e-05,
"loss": 1.2842,
"step": 89100
},
{
"epoch": 4.46,
"grad_norm": 1.5584784746170044,
"learning_rate": 3.266532663316583e-05,
"loss": 1.2961,
"step": 89200
},
{
"epoch": 4.46,
"grad_norm": 1.889328956604004,
"learning_rate": 3.236381909547738e-05,
"loss": 1.3121,
"step": 89300
},
{
"epoch": 4.47,
"grad_norm": 1.627302646636963,
"learning_rate": 3.206231155778894e-05,
"loss": 1.3169,
"step": 89400
},
{
"epoch": 4.47,
"grad_norm": 3.445068359375,
"learning_rate": 3.17608040201005e-05,
"loss": 1.2691,
"step": 89500
},
{
"epoch": 4.48,
"grad_norm": 3.2616968154907227,
"learning_rate": 3.145929648241206e-05,
"loss": 1.2818,
"step": 89600
},
{
"epoch": 4.49,
"grad_norm": 2.2444956302642822,
"learning_rate": 3.115778894472362e-05,
"loss": 1.2836,
"step": 89700
},
{
"epoch": 4.49,
"grad_norm": 2.341099500656128,
"learning_rate": 3.085628140703517e-05,
"loss": 1.2712,
"step": 89800
},
{
"epoch": 4.5,
"grad_norm": 1.882042407989502,
"learning_rate": 3.055477386934673e-05,
"loss": 1.3316,
"step": 89900
},
{
"epoch": 4.5,
"grad_norm": 3.1667628288269043,
"learning_rate": 3.025326633165829e-05,
"loss": 1.312,
"step": 90000
},
{
"epoch": 4.5,
"eval_loss": 1.309124231338501,
"eval_runtime": 37.1306,
"eval_samples_per_second": 26.932,
"eval_steps_per_second": 3.366,
"step": 90000
},
{
"epoch": 4.5,
"grad_norm": 2.1243770122528076,
"learning_rate": 2.9954773869346732e-05,
"loss": 1.2924,
"step": 90100
},
{
"epoch": 4.51,
"grad_norm": 10.278544425964355,
"learning_rate": 2.9653266331658288e-05,
"loss": 1.3081,
"step": 90200
},
{
"epoch": 4.51,
"grad_norm": 8.060081481933594,
"learning_rate": 2.9351758793969847e-05,
"loss": 1.3317,
"step": 90300
},
{
"epoch": 4.52,
"grad_norm": 1.364973783493042,
"learning_rate": 2.9050251256281404e-05,
"loss": 1.2882,
"step": 90400
},
{
"epoch": 4.53,
"grad_norm": 1.1993753910064697,
"learning_rate": 2.8748743718592963e-05,
"loss": 1.292,
"step": 90500
},
{
"epoch": 4.53,
"grad_norm": 2.897918462753296,
"learning_rate": 2.8447236180904522e-05,
"loss": 1.2804,
"step": 90600
},
{
"epoch": 4.54,
"grad_norm": 2.7222065925598145,
"learning_rate": 2.814572864321608e-05,
"loss": 1.2985,
"step": 90700
},
{
"epoch": 4.54,
"grad_norm": 5.0719099044799805,
"learning_rate": 2.7844221105527635e-05,
"loss": 1.2768,
"step": 90800
},
{
"epoch": 4.54,
"grad_norm": 2.1225016117095947,
"learning_rate": 2.754271356783919e-05,
"loss": 1.2774,
"step": 90900
},
{
"epoch": 4.55,
"grad_norm": 1.314208984375,
"learning_rate": 2.724120603015075e-05,
"loss": 1.3064,
"step": 91000
},
{
"epoch": 4.55,
"eval_loss": 1.2666804790496826,
"eval_runtime": 37.0231,
"eval_samples_per_second": 27.01,
"eval_steps_per_second": 3.376,
"step": 91000
},
{
"epoch": 4.55,
"grad_norm": 1.5543466806411743,
"learning_rate": 2.693969849246231e-05,
"loss": 1.2905,
"step": 91100
},
{
"epoch": 4.56,
"grad_norm": 1.3213764429092407,
"learning_rate": 2.6638190954773866e-05,
"loss": 1.2541,
"step": 91200
},
{
"epoch": 4.56,
"grad_norm": 2.523181200027466,
"learning_rate": 2.6336683417085425e-05,
"loss": 1.2743,
"step": 91300
},
{
"epoch": 4.57,
"grad_norm": 3.6385650634765625,
"learning_rate": 2.6035175879396984e-05,
"loss": 1.2936,
"step": 91400
},
{
"epoch": 4.58,
"grad_norm": 2.2603909969329834,
"learning_rate": 2.573366834170854e-05,
"loss": 1.252,
"step": 91500
},
{
"epoch": 4.58,
"grad_norm": 3.178767442703247,
"learning_rate": 2.54321608040201e-05,
"loss": 1.2637,
"step": 91600
},
{
"epoch": 4.58,
"grad_norm": 1.8341214656829834,
"learning_rate": 2.5130653266331656e-05,
"loss": 1.2692,
"step": 91700
},
{
"epoch": 4.59,
"grad_norm": 3.1009461879730225,
"learning_rate": 2.4829145728643216e-05,
"loss": 1.2821,
"step": 91800
},
{
"epoch": 4.59,
"grad_norm": 2.75738263130188,
"learning_rate": 2.4527638190954775e-05,
"loss": 1.284,
"step": 91900
},
{
"epoch": 4.6,
"grad_norm": 3.137352466583252,
"learning_rate": 2.4226130653266328e-05,
"loss": 1.2846,
"step": 92000
},
{
"epoch": 4.6,
"eval_loss": 1.244607925415039,
"eval_runtime": 37.127,
"eval_samples_per_second": 26.935,
"eval_steps_per_second": 3.367,
"step": 92000
},
{
"epoch": 4.61,
"grad_norm": 2.5030505657196045,
"learning_rate": 2.3924623115577887e-05,
"loss": 1.2793,
"step": 92100
},
{
"epoch": 4.61,
"grad_norm": 2.625671625137329,
"learning_rate": 2.3623115577889443e-05,
"loss": 1.2575,
"step": 92200
},
{
"epoch": 4.62,
"grad_norm": 1.8129239082336426,
"learning_rate": 2.3321608040201003e-05,
"loss": 1.265,
"step": 92300
},
{
"epoch": 4.62,
"grad_norm": 2.747807502746582,
"learning_rate": 2.3020100502512562e-05,
"loss": 1.2781,
"step": 92400
},
{
"epoch": 4.62,
"grad_norm": 3.083634853363037,
"learning_rate": 2.2718592964824118e-05,
"loss": 1.2667,
"step": 92500
},
{
"epoch": 4.63,
"grad_norm": 3.475771427154541,
"learning_rate": 2.2417085427135678e-05,
"loss": 1.2278,
"step": 92600
},
{
"epoch": 4.63,
"grad_norm": 4.056103706359863,
"learning_rate": 2.2115577889447234e-05,
"loss": 1.276,
"step": 92700
},
{
"epoch": 4.64,
"grad_norm": 3.8437623977661133,
"learning_rate": 2.1814070351758793e-05,
"loss": 1.2799,
"step": 92800
},
{
"epoch": 4.64,
"grad_norm": 2.0102827548980713,
"learning_rate": 2.1512562814070353e-05,
"loss": 1.2589,
"step": 92900
},
{
"epoch": 4.65,
"grad_norm": 3.917171001434326,
"learning_rate": 2.121105527638191e-05,
"loss": 1.245,
"step": 93000
},
{
"epoch": 4.65,
"eval_loss": 1.2793523073196411,
"eval_runtime": 37.5859,
"eval_samples_per_second": 26.606,
"eval_steps_per_second": 3.326,
"step": 93000
},
{
"epoch": 4.66,
"grad_norm": 1.847579836845398,
"learning_rate": 2.0909547738693465e-05,
"loss": 1.2172,
"step": 93100
},
{
"epoch": 4.66,
"grad_norm": 1.8271080255508423,
"learning_rate": 2.060804020100502e-05,
"loss": 1.2781,
"step": 93200
},
{
"epoch": 4.67,
"grad_norm": 3.5897600650787354,
"learning_rate": 2.030653266331658e-05,
"loss": 1.2587,
"step": 93300
},
{
"epoch": 4.67,
"grad_norm": 2.1579065322875977,
"learning_rate": 2.0008040201005026e-05,
"loss": 1.2742,
"step": 93400
},
{
"epoch": 4.67,
"grad_norm": 2.0196428298950195,
"learning_rate": 1.970653266331658e-05,
"loss": 1.24,
"step": 93500
},
{
"epoch": 4.68,
"grad_norm": 2.4959633350372314,
"learning_rate": 1.940502512562814e-05,
"loss": 1.28,
"step": 93600
},
{
"epoch": 4.69,
"grad_norm": 2.4107792377471924,
"learning_rate": 1.9103517587939695e-05,
"loss": 1.326,
"step": 93700
},
{
"epoch": 4.69,
"grad_norm": 3.1433868408203125,
"learning_rate": 1.8802010050251254e-05,
"loss": 1.2423,
"step": 93800
},
{
"epoch": 4.7,
"grad_norm": 3.2713465690612793,
"learning_rate": 1.8500502512562814e-05,
"loss": 1.3136,
"step": 93900
},
{
"epoch": 4.7,
"grad_norm": 2.5864298343658447,
"learning_rate": 1.819899497487437e-05,
"loss": 1.3029,
"step": 94000
},
{
"epoch": 4.7,
"eval_loss": 1.3026132583618164,
"eval_runtime": 37.3025,
"eval_samples_per_second": 26.808,
"eval_steps_per_second": 3.351,
"step": 94000
},
{
"epoch": 4.71,
"grad_norm": 3.219913959503174,
"learning_rate": 1.789748743718593e-05,
"loss": 1.3279,
"step": 94100
},
{
"epoch": 4.71,
"grad_norm": 2.7392513751983643,
"learning_rate": 1.7595979899497485e-05,
"loss": 1.2637,
"step": 94200
},
{
"epoch": 4.71,
"grad_norm": 3.506613254547119,
"learning_rate": 1.7294472361809045e-05,
"loss": 1.2762,
"step": 94300
},
{
"epoch": 4.72,
"grad_norm": 1.6417380571365356,
"learning_rate": 1.69929648241206e-05,
"loss": 1.2733,
"step": 94400
},
{
"epoch": 4.72,
"grad_norm": 3.572312355041504,
"learning_rate": 1.669145728643216e-05,
"loss": 1.2524,
"step": 94500
},
{
"epoch": 4.73,
"grad_norm": 2.3081557750701904,
"learning_rate": 1.6389949748743716e-05,
"loss": 1.2388,
"step": 94600
},
{
"epoch": 4.74,
"grad_norm": 3.257410764694214,
"learning_rate": 1.6088442211055276e-05,
"loss": 1.2395,
"step": 94700
},
{
"epoch": 4.74,
"grad_norm": 3.3409955501556396,
"learning_rate": 1.5786934673366835e-05,
"loss": 1.2464,
"step": 94800
},
{
"epoch": 4.75,
"grad_norm": 2.011337995529175,
"learning_rate": 1.548542713567839e-05,
"loss": 1.2799,
"step": 94900
},
{
"epoch": 4.75,
"grad_norm": 3.929819107055664,
"learning_rate": 1.5183919597989947e-05,
"loss": 1.2408,
"step": 95000
},
{
"epoch": 4.75,
"eval_loss": 1.2593971490859985,
"eval_runtime": 37.2434,
"eval_samples_per_second": 26.85,
"eval_steps_per_second": 3.356,
"step": 95000
},
{
"epoch": 4.75,
"grad_norm": 3.053884983062744,
"learning_rate": 1.4882412060301507e-05,
"loss": 1.2509,
"step": 95100
},
{
"epoch": 4.76,
"grad_norm": 2.4242303371429443,
"learning_rate": 1.4580904522613064e-05,
"loss": 1.2419,
"step": 95200
},
{
"epoch": 4.76,
"grad_norm": 5.195686340332031,
"learning_rate": 1.4279396984924622e-05,
"loss": 1.2542,
"step": 95300
},
{
"epoch": 4.77,
"grad_norm": 2.4507782459259033,
"learning_rate": 1.397788944723618e-05,
"loss": 1.2388,
"step": 95400
},
{
"epoch": 4.78,
"grad_norm": 3.998619318008423,
"learning_rate": 1.3676381909547736e-05,
"loss": 1.225,
"step": 95500
},
{
"epoch": 4.78,
"grad_norm": 6.8137664794921875,
"learning_rate": 1.3374874371859295e-05,
"loss": 1.2379,
"step": 95600
},
{
"epoch": 4.79,
"grad_norm": 2.8627781867980957,
"learning_rate": 1.3073366834170853e-05,
"loss": 1.2752,
"step": 95700
},
{
"epoch": 4.79,
"grad_norm": 2.2454240322113037,
"learning_rate": 1.2771859296482411e-05,
"loss": 1.243,
"step": 95800
},
{
"epoch": 4.79,
"grad_norm": 1.6306633949279785,
"learning_rate": 1.2470351758793969e-05,
"loss": 1.2593,
"step": 95900
},
{
"epoch": 4.8,
"grad_norm": 2.288721799850464,
"learning_rate": 1.217185929648241e-05,
"loss": 1.246,
"step": 96000
},
{
"epoch": 4.8,
"eval_loss": 1.245192527770996,
"eval_runtime": 37.2658,
"eval_samples_per_second": 26.834,
"eval_steps_per_second": 3.354,
"step": 96000
},
{
"epoch": 4.8,
"grad_norm": 3.1359307765960693,
"learning_rate": 1.187035175879397e-05,
"loss": 1.2531,
"step": 96100
},
{
"epoch": 4.81,
"grad_norm": 3.468085527420044,
"learning_rate": 1.1568844221105527e-05,
"loss": 1.2602,
"step": 96200
},
{
"epoch": 4.81,
"grad_norm": 2.3539419174194336,
"learning_rate": 1.1267336683417085e-05,
"loss": 1.236,
"step": 96300
},
{
"epoch": 4.82,
"grad_norm": 2.901660442352295,
"learning_rate": 1.0965829145728641e-05,
"loss": 1.2401,
"step": 96400
},
{
"epoch": 4.83,
"grad_norm": 2.2228598594665527,
"learning_rate": 1.0664321608040199e-05,
"loss": 1.2718,
"step": 96500
},
{
"epoch": 4.83,
"grad_norm": 2.1923272609710693,
"learning_rate": 1.0362814070351758e-05,
"loss": 1.2642,
"step": 96600
},
{
"epoch": 4.83,
"grad_norm": 3.3871023654937744,
"learning_rate": 1.0061306532663316e-05,
"loss": 1.2247,
"step": 96700
},
{
"epoch": 4.84,
"grad_norm": 4.223761081695557,
"learning_rate": 9.759798994974874e-06,
"loss": 1.2761,
"step": 96800
},
{
"epoch": 4.84,
"grad_norm": 4.008758068084717,
"learning_rate": 9.458291457286431e-06,
"loss": 1.2895,
"step": 96900
},
{
"epoch": 4.85,
"grad_norm": 3.956162214279175,
"learning_rate": 9.156783919597989e-06,
"loss": 1.2573,
"step": 97000
},
{
"epoch": 4.85,
"eval_loss": 1.2709555625915527,
"eval_runtime": 37.1777,
"eval_samples_per_second": 26.898,
"eval_steps_per_second": 3.362,
"step": 97000
},
{
"epoch": 4.86,
"grad_norm": 1.0692745447158813,
"learning_rate": 8.855276381909547e-06,
"loss": 1.242,
"step": 97100
},
{
"epoch": 4.86,
"grad_norm": 2.2347793579101562,
"learning_rate": 8.553768844221105e-06,
"loss": 1.2211,
"step": 97200
},
{
"epoch": 4.87,
"grad_norm": 3.8911612033843994,
"learning_rate": 8.252261306532662e-06,
"loss": 1.2532,
"step": 97300
},
{
"epoch": 4.87,
"grad_norm": 3.3509294986724854,
"learning_rate": 7.95075376884422e-06,
"loss": 1.2599,
"step": 97400
},
{
"epoch": 4.88,
"grad_norm": 2.3662168979644775,
"learning_rate": 7.649246231155778e-06,
"loss": 1.2768,
"step": 97500
},
{
"epoch": 4.88,
"grad_norm": 3.0100605487823486,
"learning_rate": 7.3477386934673365e-06,
"loss": 1.266,
"step": 97600
},
{
"epoch": 4.88,
"grad_norm": 3.673358201980591,
"learning_rate": 7.046231155778893e-06,
"loss": 1.2288,
"step": 97700
},
{
"epoch": 4.89,
"grad_norm": 3.1406137943267822,
"learning_rate": 6.744723618090451e-06,
"loss": 1.2193,
"step": 97800
},
{
"epoch": 4.89,
"grad_norm": 2.9519011974334717,
"learning_rate": 6.44321608040201e-06,
"loss": 1.2347,
"step": 97900
},
{
"epoch": 4.9,
"grad_norm": 2.01814866065979,
"learning_rate": 6.141708542713567e-06,
"loss": 1.242,
"step": 98000
},
{
"epoch": 4.9,
"eval_loss": 1.2466384172439575,
"eval_runtime": 37.1499,
"eval_samples_per_second": 26.918,
"eval_steps_per_second": 3.365,
"step": 98000
},
{
"epoch": 4.91,
"grad_norm": 6.0040059089660645,
"learning_rate": 5.840201005025125e-06,
"loss": 1.2014,
"step": 98100
},
{
"epoch": 4.91,
"grad_norm": 5.865699291229248,
"learning_rate": 5.538693467336683e-06,
"loss": 1.2544,
"step": 98200
},
{
"epoch": 4.92,
"grad_norm": 3.115818500518799,
"learning_rate": 5.23718592964824e-06,
"loss": 1.2369,
"step": 98300
},
{
"epoch": 4.92,
"grad_norm": 2.521693468093872,
"learning_rate": 4.9356783919597985e-06,
"loss": 1.2166,
"step": 98400
},
{
"epoch": 4.92,
"grad_norm": 3.7152583599090576,
"learning_rate": 4.634170854271356e-06,
"loss": 1.2458,
"step": 98500
},
{
"epoch": 4.93,
"grad_norm": 8.62978458404541,
"learning_rate": 4.332663316582914e-06,
"loss": 1.2525,
"step": 98600
},
{
"epoch": 4.94,
"grad_norm": 3.633636951446533,
"learning_rate": 4.031155778894472e-06,
"loss": 1.2316,
"step": 98700
},
{
"epoch": 4.94,
"grad_norm": 4.357114791870117,
"learning_rate": 3.7296482412060296e-06,
"loss": 1.2195,
"step": 98800
},
{
"epoch": 4.95,
"grad_norm": 1.2212806940078735,
"learning_rate": 3.4281407035175877e-06,
"loss": 1.1938,
"step": 98900
},
{
"epoch": 4.95,
"grad_norm": 3.4345273971557617,
"learning_rate": 3.1266331658291455e-06,
"loss": 1.2312,
"step": 99000
},
{
"epoch": 4.95,
"eval_loss": 1.2566660642623901,
"eval_runtime": 37.0254,
"eval_samples_per_second": 27.009,
"eval_steps_per_second": 3.376,
"step": 99000
},
{
"epoch": 4.96,
"grad_norm": 4.7800188064575195,
"learning_rate": 2.8251256281407033e-06,
"loss": 1.1972,
"step": 99100
},
{
"epoch": 4.96,
"grad_norm": 2.3637685775756836,
"learning_rate": 2.5236180904522614e-06,
"loss": 1.2475,
"step": 99200
},
{
"epoch": 4.96,
"grad_norm": 3.591388463973999,
"learning_rate": 2.2221105527638188e-06,
"loss": 1.2505,
"step": 99300
},
{
"epoch": 4.97,
"grad_norm": 3.078840970993042,
"learning_rate": 1.9236180904522612e-06,
"loss": 1.2421,
"step": 99400
},
{
"epoch": 4.97,
"grad_norm": 3.289644479751587,
"learning_rate": 1.622110552763819e-06,
"loss": 1.257,
"step": 99500
},
{
"epoch": 4.98,
"grad_norm": 4.775394439697266,
"learning_rate": 1.3206030150753765e-06,
"loss": 1.256,
"step": 99600
},
{
"epoch": 4.99,
"grad_norm": 4.110093116760254,
"learning_rate": 1.0190954773869345e-06,
"loss": 1.1982,
"step": 99700
},
{
"epoch": 4.99,
"grad_norm": 1.86680006980896,
"learning_rate": 7.175879396984924e-07,
"loss": 1.2571,
"step": 99800
},
{
"epoch": 5.0,
"grad_norm": 3.5120298862457275,
"learning_rate": 4.160804020100502e-07,
"loss": 1.2196,
"step": 99900
},
{
"epoch": 5.0,
"grad_norm": 2.356171131134033,
"learning_rate": 1.1457286432160803e-07,
"loss": 1.2164,
"step": 100000
},
{
"epoch": 5.0,
"eval_loss": 1.2736848592758179,
"eval_runtime": 37.4852,
"eval_samples_per_second": 26.677,
"eval_steps_per_second": 3.335,
"step": 100000
}
],
"logging_steps": 100,
"max_steps": 100000,
"num_input_tokens_seen": 0,
"num_train_epochs": 5,
"save_steps": 1000,
"total_flos": 1.2076594495488e+18,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}