phi3nedtuned-ner / trainer_state.json
shujatoor's picture
shujatoor/phi3nedtuned-ner
8b2cbed verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.0,
"eval_steps": 500,
"global_step": 8415,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0023767082590612004,
"grad_norm": 3.921875,
"learning_rate": 2.3767082590612003e-06,
"loss": 3.6948,
"step": 20
},
{
"epoch": 0.004753416518122401,
"grad_norm": 5.28125,
"learning_rate": 4.753416518122401e-06,
"loss": 3.8478,
"step": 40
},
{
"epoch": 0.0071301247771836,
"grad_norm": 7.21875,
"learning_rate": 7.130124777183601e-06,
"loss": 3.6042,
"step": 60
},
{
"epoch": 0.009506833036244802,
"grad_norm": 2.625,
"learning_rate": 9.506833036244801e-06,
"loss": 3.5338,
"step": 80
},
{
"epoch": 0.011883541295306001,
"grad_norm": 1.515625,
"learning_rate": 1.1883541295306002e-05,
"loss": 3.2504,
"step": 100
},
{
"epoch": 0.0142602495543672,
"grad_norm": 1.3203125,
"learning_rate": 1.4260249554367203e-05,
"loss": 2.956,
"step": 120
},
{
"epoch": 0.016636957813428402,
"grad_norm": 1.6015625,
"learning_rate": 1.66369578134284e-05,
"loss": 2.9073,
"step": 140
},
{
"epoch": 0.019013666072489603,
"grad_norm": 1.0703125,
"learning_rate": 1.9013666072489602e-05,
"loss": 2.6436,
"step": 160
},
{
"epoch": 0.0213903743315508,
"grad_norm": 1.0234375,
"learning_rate": 2.1390374331550803e-05,
"loss": 2.6018,
"step": 180
},
{
"epoch": 0.023767082590612002,
"grad_norm": 1.0234375,
"learning_rate": 2.3767082590612004e-05,
"loss": 2.4289,
"step": 200
},
{
"epoch": 0.026143790849673203,
"grad_norm": 1.109375,
"learning_rate": 2.6143790849673204e-05,
"loss": 2.5239,
"step": 220
},
{
"epoch": 0.0285204991087344,
"grad_norm": 1.5625,
"learning_rate": 2.8520499108734405e-05,
"loss": 2.4617,
"step": 240
},
{
"epoch": 0.030897207367795602,
"grad_norm": 1.1171875,
"learning_rate": 3.08972073677956e-05,
"loss": 2.396,
"step": 260
},
{
"epoch": 0.033273915626856804,
"grad_norm": 1.03125,
"learning_rate": 3.32739156268568e-05,
"loss": 2.2682,
"step": 280
},
{
"epoch": 0.035650623885918005,
"grad_norm": 1.265625,
"learning_rate": 3.5650623885918004e-05,
"loss": 2.3627,
"step": 300
},
{
"epoch": 0.038027332144979206,
"grad_norm": 1.1484375,
"learning_rate": 3.8027332144979205e-05,
"loss": 2.3072,
"step": 320
},
{
"epoch": 0.04040404040404041,
"grad_norm": 1.5859375,
"learning_rate": 4.0404040404040405e-05,
"loss": 2.2124,
"step": 340
},
{
"epoch": 0.0427807486631016,
"grad_norm": 1.0703125,
"learning_rate": 4.2780748663101606e-05,
"loss": 2.2619,
"step": 360
},
{
"epoch": 0.0451574569221628,
"grad_norm": 1.46875,
"learning_rate": 4.515745692216281e-05,
"loss": 2.1928,
"step": 380
},
{
"epoch": 0.047534165181224004,
"grad_norm": 1.3515625,
"learning_rate": 4.753416518122401e-05,
"loss": 2.0428,
"step": 400
},
{
"epoch": 0.049910873440285206,
"grad_norm": 1.90625,
"learning_rate": 4.991087344028521e-05,
"loss": 1.9929,
"step": 420
},
{
"epoch": 0.05228758169934641,
"grad_norm": 1.3828125,
"learning_rate": 5.228758169934641e-05,
"loss": 2.146,
"step": 440
},
{
"epoch": 0.05466428995840761,
"grad_norm": 1.5859375,
"learning_rate": 5.466428995840761e-05,
"loss": 2.0289,
"step": 460
},
{
"epoch": 0.0570409982174688,
"grad_norm": 1.3125,
"learning_rate": 5.704099821746881e-05,
"loss": 1.9554,
"step": 480
},
{
"epoch": 0.059417706476530004,
"grad_norm": 1.6484375,
"learning_rate": 5.941770647653001e-05,
"loss": 1.9871,
"step": 500
},
{
"epoch": 0.061794414735591205,
"grad_norm": 1.9140625,
"learning_rate": 6.17944147355912e-05,
"loss": 2.0099,
"step": 520
},
{
"epoch": 0.06417112299465241,
"grad_norm": 2.140625,
"learning_rate": 6.41711229946524e-05,
"loss": 1.9199,
"step": 540
},
{
"epoch": 0.06654783125371361,
"grad_norm": 2.203125,
"learning_rate": 6.65478312537136e-05,
"loss": 2.0074,
"step": 560
},
{
"epoch": 0.0689245395127748,
"grad_norm": 1.2421875,
"learning_rate": 6.892453951277481e-05,
"loss": 1.8052,
"step": 580
},
{
"epoch": 0.07130124777183601,
"grad_norm": 2.25,
"learning_rate": 7.130124777183601e-05,
"loss": 2.0366,
"step": 600
},
{
"epoch": 0.0736779560308972,
"grad_norm": 2.4375,
"learning_rate": 7.367795603089721e-05,
"loss": 1.8426,
"step": 620
},
{
"epoch": 0.07605466428995841,
"grad_norm": 3.75,
"learning_rate": 7.605466428995841e-05,
"loss": 1.6836,
"step": 640
},
{
"epoch": 0.0784313725490196,
"grad_norm": 3.09375,
"learning_rate": 7.843137254901961e-05,
"loss": 2.0483,
"step": 660
},
{
"epoch": 0.08080808080808081,
"grad_norm": 1.6953125,
"learning_rate": 8.080808080808081e-05,
"loss": 1.902,
"step": 680
},
{
"epoch": 0.08318478906714201,
"grad_norm": 2.109375,
"learning_rate": 8.318478906714201e-05,
"loss": 1.9887,
"step": 700
},
{
"epoch": 0.0855614973262032,
"grad_norm": 2.984375,
"learning_rate": 8.556149732620321e-05,
"loss": 1.6542,
"step": 720
},
{
"epoch": 0.08793820558526441,
"grad_norm": 3.3125,
"learning_rate": 8.793820558526441e-05,
"loss": 1.6718,
"step": 740
},
{
"epoch": 0.0903149138443256,
"grad_norm": 2.9375,
"learning_rate": 9.031491384432561e-05,
"loss": 1.706,
"step": 760
},
{
"epoch": 0.09269162210338681,
"grad_norm": 3.59375,
"learning_rate": 9.269162210338681e-05,
"loss": 1.629,
"step": 780
},
{
"epoch": 0.09506833036244801,
"grad_norm": 2.40625,
"learning_rate": 9.506833036244802e-05,
"loss": 1.6748,
"step": 800
},
{
"epoch": 0.0974450386215092,
"grad_norm": 2.578125,
"learning_rate": 9.744503862150922e-05,
"loss": 1.8269,
"step": 820
},
{
"epoch": 0.09982174688057041,
"grad_norm": 2.328125,
"learning_rate": 9.982174688057042e-05,
"loss": 1.8727,
"step": 840
},
{
"epoch": 0.1021984551396316,
"grad_norm": 1.8671875,
"learning_rate": 0.00010219845513963162,
"loss": 1.7455,
"step": 860
},
{
"epoch": 0.10457516339869281,
"grad_norm": 2.140625,
"learning_rate": 0.00010457516339869282,
"loss": 1.7395,
"step": 880
},
{
"epoch": 0.10695187165775401,
"grad_norm": 1.609375,
"learning_rate": 0.00010695187165775402,
"loss": 1.7935,
"step": 900
},
{
"epoch": 0.10932857991681522,
"grad_norm": 1.6796875,
"learning_rate": 0.00010932857991681522,
"loss": 1.8009,
"step": 920
},
{
"epoch": 0.11170528817587641,
"grad_norm": 1.8515625,
"learning_rate": 0.00011170528817587642,
"loss": 1.5582,
"step": 940
},
{
"epoch": 0.1140819964349376,
"grad_norm": 2.0,
"learning_rate": 0.00011408199643493762,
"loss": 1.6402,
"step": 960
},
{
"epoch": 0.11645870469399881,
"grad_norm": 1.9921875,
"learning_rate": 0.00011645870469399882,
"loss": 1.665,
"step": 980
},
{
"epoch": 0.11883541295306001,
"grad_norm": 1.78125,
"learning_rate": 0.00011883541295306002,
"loss": 1.8407,
"step": 1000
},
{
"epoch": 0.12121212121212122,
"grad_norm": 1.9296875,
"learning_rate": 0.00012121212121212122,
"loss": 1.7665,
"step": 1020
},
{
"epoch": 0.12358882947118241,
"grad_norm": 2.5625,
"learning_rate": 0.0001235888294711824,
"loss": 1.8593,
"step": 1040
},
{
"epoch": 0.1259655377302436,
"grad_norm": 2.109375,
"learning_rate": 0.00012596553773024362,
"loss": 1.7738,
"step": 1060
},
{
"epoch": 0.12834224598930483,
"grad_norm": 2.078125,
"learning_rate": 0.0001283422459893048,
"loss": 1.8227,
"step": 1080
},
{
"epoch": 0.13071895424836602,
"grad_norm": 3.53125,
"learning_rate": 0.00013071895424836603,
"loss": 1.7159,
"step": 1100
},
{
"epoch": 0.13309566250742721,
"grad_norm": 1.890625,
"learning_rate": 0.0001330956625074272,
"loss": 1.6953,
"step": 1120
},
{
"epoch": 0.1354723707664884,
"grad_norm": 1.8671875,
"learning_rate": 0.00013547237076648843,
"loss": 1.5567,
"step": 1140
},
{
"epoch": 0.1378490790255496,
"grad_norm": 1.515625,
"learning_rate": 0.00013784907902554961,
"loss": 1.7637,
"step": 1160
},
{
"epoch": 0.14022578728461083,
"grad_norm": 2.375,
"learning_rate": 0.00014022578728461083,
"loss": 1.5884,
"step": 1180
},
{
"epoch": 0.14260249554367202,
"grad_norm": 2.046875,
"learning_rate": 0.00014260249554367202,
"loss": 1.5542,
"step": 1200
},
{
"epoch": 0.14497920380273321,
"grad_norm": 2.515625,
"learning_rate": 0.00014497920380273323,
"loss": 1.5124,
"step": 1220
},
{
"epoch": 0.1473559120617944,
"grad_norm": 1.609375,
"learning_rate": 0.00014735591206179442,
"loss": 1.5822,
"step": 1240
},
{
"epoch": 0.1497326203208556,
"grad_norm": 1.7890625,
"learning_rate": 0.00014973262032085563,
"loss": 1.6216,
"step": 1260
},
{
"epoch": 0.15210932857991682,
"grad_norm": 2.34375,
"learning_rate": 0.00015210932857991682,
"loss": 1.7366,
"step": 1280
},
{
"epoch": 0.15448603683897802,
"grad_norm": 1.8828125,
"learning_rate": 0.00015448603683897803,
"loss": 1.3673,
"step": 1300
},
{
"epoch": 0.1568627450980392,
"grad_norm": 1.8046875,
"learning_rate": 0.00015686274509803922,
"loss": 1.4071,
"step": 1320
},
{
"epoch": 0.1592394533571004,
"grad_norm": 2.59375,
"learning_rate": 0.00015923945335710043,
"loss": 1.6515,
"step": 1340
},
{
"epoch": 0.16161616161616163,
"grad_norm": 2.28125,
"learning_rate": 0.00016161616161616162,
"loss": 1.7604,
"step": 1360
},
{
"epoch": 0.16399286987522282,
"grad_norm": 1.9140625,
"learning_rate": 0.0001639928698752228,
"loss": 1.5925,
"step": 1380
},
{
"epoch": 0.16636957813428402,
"grad_norm": 3.40625,
"learning_rate": 0.00016636957813428402,
"loss": 1.5874,
"step": 1400
},
{
"epoch": 0.1687462863933452,
"grad_norm": 2.65625,
"learning_rate": 0.0001687462863933452,
"loss": 1.6116,
"step": 1420
},
{
"epoch": 0.1711229946524064,
"grad_norm": 1.8671875,
"learning_rate": 0.00017112299465240642,
"loss": 1.544,
"step": 1440
},
{
"epoch": 0.17349970291146763,
"grad_norm": 2.15625,
"learning_rate": 0.0001734997029114676,
"loss": 1.5015,
"step": 1460
},
{
"epoch": 0.17587641117052882,
"grad_norm": 1.8515625,
"learning_rate": 0.00017587641117052883,
"loss": 1.5858,
"step": 1480
},
{
"epoch": 0.17825311942959002,
"grad_norm": 1.90625,
"learning_rate": 0.00017825311942959,
"loss": 1.5469,
"step": 1500
},
{
"epoch": 0.1806298276886512,
"grad_norm": 2.421875,
"learning_rate": 0.00018062982768865123,
"loss": 1.5092,
"step": 1520
},
{
"epoch": 0.1830065359477124,
"grad_norm": 1.9921875,
"learning_rate": 0.00018300653594771241,
"loss": 1.3362,
"step": 1540
},
{
"epoch": 0.18538324420677363,
"grad_norm": 1.6015625,
"learning_rate": 0.00018538324420677363,
"loss": 1.5088,
"step": 1560
},
{
"epoch": 0.18775995246583482,
"grad_norm": 1.9765625,
"learning_rate": 0.00018775995246583482,
"loss": 1.5518,
"step": 1580
},
{
"epoch": 0.19013666072489602,
"grad_norm": 1.8046875,
"learning_rate": 0.00019013666072489603,
"loss": 1.4344,
"step": 1600
},
{
"epoch": 0.1925133689839572,
"grad_norm": 1.71875,
"learning_rate": 0.00019251336898395722,
"loss": 1.5411,
"step": 1620
},
{
"epoch": 0.1948900772430184,
"grad_norm": 1.515625,
"learning_rate": 0.00019489007724301843,
"loss": 1.5832,
"step": 1640
},
{
"epoch": 0.19726678550207963,
"grad_norm": 1.71875,
"learning_rate": 0.00019726678550207962,
"loss": 1.5468,
"step": 1660
},
{
"epoch": 0.19964349376114082,
"grad_norm": 2.078125,
"learning_rate": 0.00019964349376114083,
"loss": 1.3053,
"step": 1680
},
{
"epoch": 0.20202020202020202,
"grad_norm": 2.0625,
"learning_rate": 0.0001999968531423333,
"loss": 1.5303,
"step": 1700
},
{
"epoch": 0.2043969102792632,
"grad_norm": 1.4921875,
"learning_rate": 0.0001999850935511802,
"loss": 1.6331,
"step": 1720
},
{
"epoch": 0.20677361853832443,
"grad_norm": 1.8203125,
"learning_rate": 0.0001999646242513938,
"loss": 1.5621,
"step": 1740
},
{
"epoch": 0.20915032679738563,
"grad_norm": 1.578125,
"learning_rate": 0.00019993544702605638,
"loss": 1.4932,
"step": 1760
},
{
"epoch": 0.21152703505644682,
"grad_norm": 2.09375,
"learning_rate": 0.00019989756441679795,
"loss": 1.3936,
"step": 1780
},
{
"epoch": 0.21390374331550802,
"grad_norm": 1.4453125,
"learning_rate": 0.00019985097972357547,
"loss": 1.3227,
"step": 1800
},
{
"epoch": 0.2162804515745692,
"grad_norm": 1.671875,
"learning_rate": 0.0001997956970043848,
"loss": 1.4253,
"step": 1820
},
{
"epoch": 0.21865715983363043,
"grad_norm": 1.5078125,
"learning_rate": 0.00019973172107490758,
"loss": 1.3946,
"step": 1840
},
{
"epoch": 0.22103386809269163,
"grad_norm": 1.625,
"learning_rate": 0.00019965905750809158,
"loss": 1.2779,
"step": 1860
},
{
"epoch": 0.22341057635175282,
"grad_norm": 1.4296875,
"learning_rate": 0.00019957771263366526,
"loss": 1.5051,
"step": 1880
},
{
"epoch": 0.22578728461081402,
"grad_norm": 1.9765625,
"learning_rate": 0.00019948769353758653,
"loss": 1.3954,
"step": 1900
},
{
"epoch": 0.2281639928698752,
"grad_norm": 2.25,
"learning_rate": 0.00019938900806142518,
"loss": 1.2049,
"step": 1920
},
{
"epoch": 0.23054070112893643,
"grad_norm": 2.5625,
"learning_rate": 0.00019928166480168011,
"loss": 1.5335,
"step": 1940
},
{
"epoch": 0.23291740938799763,
"grad_norm": 2.03125,
"learning_rate": 0.00019916567310903035,
"loss": 1.3868,
"step": 1960
},
{
"epoch": 0.23529411764705882,
"grad_norm": 1.46875,
"learning_rate": 0.0001990410430875205,
"loss": 1.3677,
"step": 1980
},
{
"epoch": 0.23767082590612001,
"grad_norm": 2.484375,
"learning_rate": 0.0001989077855936807,
"loss": 1.2757,
"step": 2000
},
{
"epoch": 0.24004753416518124,
"grad_norm": 2.21875,
"learning_rate": 0.0001987659122355806,
"loss": 1.4893,
"step": 2020
},
{
"epoch": 0.24242424242424243,
"grad_norm": 1.28125,
"learning_rate": 0.00019861543537181867,
"loss": 1.2123,
"step": 2040
},
{
"epoch": 0.24480095068330363,
"grad_norm": 2.078125,
"learning_rate": 0.00019845636811044521,
"loss": 1.4714,
"step": 2060
},
{
"epoch": 0.24717765894236482,
"grad_norm": 2.453125,
"learning_rate": 0.0001982887243078207,
"loss": 1.463,
"step": 2080
},
{
"epoch": 0.24955436720142601,
"grad_norm": 1.953125,
"learning_rate": 0.00019811251856740873,
"loss": 1.5086,
"step": 2100
},
{
"epoch": 0.2519310754604872,
"grad_norm": 1.734375,
"learning_rate": 0.000197927766238504,
"loss": 1.3604,
"step": 2120
},
{
"epoch": 0.25430778371954843,
"grad_norm": 1.96875,
"learning_rate": 0.00019773448341489495,
"loss": 1.2226,
"step": 2140
},
{
"epoch": 0.25668449197860965,
"grad_norm": 1.609375,
"learning_rate": 0.00019753268693346225,
"loss": 1.2003,
"step": 2160
},
{
"epoch": 0.2590612002376708,
"grad_norm": 1.9296875,
"learning_rate": 0.0001973223943727117,
"loss": 1.192,
"step": 2180
},
{
"epoch": 0.26143790849673204,
"grad_norm": 1.9765625,
"learning_rate": 0.00019710362405124334,
"loss": 1.2616,
"step": 2200
},
{
"epoch": 0.2638146167557932,
"grad_norm": 1.2265625,
"learning_rate": 0.0001968763950261554,
"loss": 1.2493,
"step": 2220
},
{
"epoch": 0.26619132501485443,
"grad_norm": 2.234375,
"learning_rate": 0.0001966407270913846,
"loss": 1.278,
"step": 2240
},
{
"epoch": 0.26856803327391565,
"grad_norm": 1.5703125,
"learning_rate": 0.00019639664077598142,
"loss": 1.3797,
"step": 2260
},
{
"epoch": 0.2709447415329768,
"grad_norm": 2.078125,
"learning_rate": 0.0001961441573423223,
"loss": 1.1243,
"step": 2280
},
{
"epoch": 0.27332144979203804,
"grad_norm": 2.0,
"learning_rate": 0.0001958832987842571,
"loss": 1.324,
"step": 2300
},
{
"epoch": 0.2756981580510992,
"grad_norm": 2.375,
"learning_rate": 0.00019561408782519345,
"loss": 1.3572,
"step": 2320
},
{
"epoch": 0.27807486631016043,
"grad_norm": 2.265625,
"learning_rate": 0.0001953365479161172,
"loss": 1.3404,
"step": 2340
},
{
"epoch": 0.28045157456922165,
"grad_norm": 2.984375,
"learning_rate": 0.00019505070323354965,
"loss": 1.3237,
"step": 2360
},
{
"epoch": 0.2828282828282828,
"grad_norm": 1.8203125,
"learning_rate": 0.0001947565786774415,
"loss": 1.2894,
"step": 2380
},
{
"epoch": 0.28520499108734404,
"grad_norm": 1.7578125,
"learning_rate": 0.0001944541998690038,
"loss": 1.3445,
"step": 2400
},
{
"epoch": 0.2875816993464052,
"grad_norm": 2.015625,
"learning_rate": 0.0001941435931484761,
"loss": 1.3859,
"step": 2420
},
{
"epoch": 0.28995840760546643,
"grad_norm": 1.7890625,
"learning_rate": 0.00019382478557283204,
"loss": 1.2185,
"step": 2440
},
{
"epoch": 0.29233511586452765,
"grad_norm": 1.546875,
"learning_rate": 0.00019349780491342223,
"loss": 1.3545,
"step": 2460
},
{
"epoch": 0.2947118241235888,
"grad_norm": 2.125,
"learning_rate": 0.00019316267965355528,
"loss": 1.2694,
"step": 2480
},
{
"epoch": 0.29708853238265004,
"grad_norm": 2.03125,
"learning_rate": 0.00019281943898601645,
"loss": 1.0126,
"step": 2500
},
{
"epoch": 0.2994652406417112,
"grad_norm": 1.6328125,
"learning_rate": 0.00019246811281052487,
"loss": 1.2593,
"step": 2520
},
{
"epoch": 0.3018419489007724,
"grad_norm": 1.578125,
"learning_rate": 0.00019210873173112865,
"loss": 1.3609,
"step": 2540
},
{
"epoch": 0.30421865715983365,
"grad_norm": 1.375,
"learning_rate": 0.0001917413270535393,
"loss": 0.9368,
"step": 2560
},
{
"epoch": 0.3065953654188948,
"grad_norm": 2.25,
"learning_rate": 0.0001913659307824045,
"loss": 1.1893,
"step": 2580
},
{
"epoch": 0.30897207367795604,
"grad_norm": 2.21875,
"learning_rate": 0.0001909825756185202,
"loss": 1.1345,
"step": 2600
},
{
"epoch": 0.3113487819370172,
"grad_norm": 1.515625,
"learning_rate": 0.0001905912949559821,
"loss": 1.2925,
"step": 2620
},
{
"epoch": 0.3137254901960784,
"grad_norm": 1.296875,
"learning_rate": 0.00019019212287927663,
"loss": 1.0842,
"step": 2640
},
{
"epoch": 0.31610219845513965,
"grad_norm": 2.40625,
"learning_rate": 0.00018978509416031186,
"loss": 1.2722,
"step": 2660
},
{
"epoch": 0.3184789067142008,
"grad_norm": 1.796875,
"learning_rate": 0.00018937024425538855,
"loss": 1.3566,
"step": 2680
},
{
"epoch": 0.32085561497326204,
"grad_norm": 2.09375,
"learning_rate": 0.0001889476093021115,
"loss": 0.9899,
"step": 2700
},
{
"epoch": 0.32323232323232326,
"grad_norm": 1.8046875,
"learning_rate": 0.00018851722611624164,
"loss": 1.0713,
"step": 2720
},
{
"epoch": 0.3256090314913844,
"grad_norm": 2.25,
"learning_rate": 0.00018807913218848906,
"loss": 1.1957,
"step": 2740
},
{
"epoch": 0.32798573975044565,
"grad_norm": 1.453125,
"learning_rate": 0.000187633365681247,
"loss": 1.3258,
"step": 2760
},
{
"epoch": 0.3303624480095068,
"grad_norm": 1.5625,
"learning_rate": 0.00018717996542526777,
"loss": 1.1146,
"step": 2780
},
{
"epoch": 0.33273915626856804,
"grad_norm": 1.984375,
"learning_rate": 0.00018671897091627993,
"loss": 1.0215,
"step": 2800
},
{
"epoch": 0.33511586452762926,
"grad_norm": 1.5703125,
"learning_rate": 0.00018625042231154817,
"loss": 1.1107,
"step": 2820
},
{
"epoch": 0.3374925727866904,
"grad_norm": 1.8828125,
"learning_rate": 0.00018577436042637477,
"loss": 1.1655,
"step": 2840
},
{
"epoch": 0.33986928104575165,
"grad_norm": 1.765625,
"learning_rate": 0.00018529082673054457,
"loss": 1.1197,
"step": 2860
},
{
"epoch": 0.3422459893048128,
"grad_norm": 1.8828125,
"learning_rate": 0.0001847998633447123,
"loss": 1.1792,
"step": 2880
},
{
"epoch": 0.34462269756387404,
"grad_norm": 1.9765625,
"learning_rate": 0.0001843015130367335,
"loss": 1.1933,
"step": 2900
},
{
"epoch": 0.34699940582293526,
"grad_norm": 2.15625,
"learning_rate": 0.00018379581921793914,
"loss": 1.1603,
"step": 2920
},
{
"epoch": 0.3493761140819964,
"grad_norm": 1.09375,
"learning_rate": 0.00018328282593935377,
"loss": 1.1653,
"step": 2940
},
{
"epoch": 0.35175282234105765,
"grad_norm": 1.7578125,
"learning_rate": 0.00018276257788785855,
"loss": 1.0807,
"step": 2960
},
{
"epoch": 0.3541295306001188,
"grad_norm": 2.296875,
"learning_rate": 0.00018223512038229833,
"loss": 1.0923,
"step": 2980
},
{
"epoch": 0.35650623885918004,
"grad_norm": 1.7109375,
"learning_rate": 0.00018170049936953406,
"loss": 1.3022,
"step": 3000
},
{
"epoch": 0.35888294711824126,
"grad_norm": 1.625,
"learning_rate": 0.00018115876142044032,
"loss": 1.2415,
"step": 3020
},
{
"epoch": 0.3612596553773024,
"grad_norm": 1.4453125,
"learning_rate": 0.0001806099537258485,
"loss": 1.0849,
"step": 3040
},
{
"epoch": 0.36363636363636365,
"grad_norm": 1.8203125,
"learning_rate": 0.00018005412409243606,
"loss": 1.1712,
"step": 3060
},
{
"epoch": 0.3660130718954248,
"grad_norm": 1.4921875,
"learning_rate": 0.000179491320938562,
"loss": 1.2795,
"step": 3080
},
{
"epoch": 0.36838978015448604,
"grad_norm": 1.734375,
"learning_rate": 0.00017892159329004916,
"loss": 1.1304,
"step": 3100
},
{
"epoch": 0.37076648841354726,
"grad_norm": 2.265625,
"learning_rate": 0.00017834499077591374,
"loss": 1.277,
"step": 3120
},
{
"epoch": 0.3731431966726084,
"grad_norm": 1.5546875,
"learning_rate": 0.00017776156362404186,
"loss": 1.2222,
"step": 3140
},
{
"epoch": 0.37551990493166965,
"grad_norm": 2.4375,
"learning_rate": 0.0001771713626568143,
"loss": 1.2639,
"step": 3160
},
{
"epoch": 0.3778966131907308,
"grad_norm": 1.7109375,
"learning_rate": 0.0001765744392866795,
"loss": 1.1679,
"step": 3180
},
{
"epoch": 0.38027332144979203,
"grad_norm": 1.9921875,
"learning_rate": 0.00017597084551167476,
"loss": 1.0657,
"step": 3200
},
{
"epoch": 0.38265002970885326,
"grad_norm": 1.2109375,
"learning_rate": 0.00017536063391089697,
"loss": 1.1886,
"step": 3220
},
{
"epoch": 0.3850267379679144,
"grad_norm": 1.1875,
"learning_rate": 0.00017474385763992212,
"loss": 1.1526,
"step": 3240
},
{
"epoch": 0.38740344622697565,
"grad_norm": 1.6171875,
"learning_rate": 0.00017412057042617525,
"loss": 1.0648,
"step": 3260
},
{
"epoch": 0.3897801544860368,
"grad_norm": 1.8515625,
"learning_rate": 0.00017349082656424995,
"loss": 1.2938,
"step": 3280
},
{
"epoch": 0.39215686274509803,
"grad_norm": 1.34375,
"learning_rate": 0.00017285468091117904,
"loss": 1.1246,
"step": 3300
},
{
"epoch": 0.39453357100415926,
"grad_norm": 2.03125,
"learning_rate": 0.00017221218888165572,
"loss": 1.0179,
"step": 3320
},
{
"epoch": 0.3969102792632204,
"grad_norm": 1.8671875,
"learning_rate": 0.0001715634064432065,
"loss": 0.914,
"step": 3340
},
{
"epoch": 0.39928698752228164,
"grad_norm": 1.4296875,
"learning_rate": 0.0001709083901113159,
"loss": 1.1192,
"step": 3360
},
{
"epoch": 0.40166369578134287,
"grad_norm": 3.203125,
"learning_rate": 0.00017024719694450337,
"loss": 1.1462,
"step": 3380
},
{
"epoch": 0.40404040404040403,
"grad_norm": 1.3515625,
"learning_rate": 0.00016957988453935276,
"loss": 0.9452,
"step": 3400
},
{
"epoch": 0.40641711229946526,
"grad_norm": 2.28125,
"learning_rate": 0.00016890651102549538,
"loss": 1.1837,
"step": 3420
},
{
"epoch": 0.4087938205585264,
"grad_norm": 2.515625,
"learning_rate": 0.00016822713506054604,
"loss": 0.8889,
"step": 3440
},
{
"epoch": 0.41117052881758764,
"grad_norm": 1.515625,
"learning_rate": 0.0001675418158249935,
"loss": 1.0873,
"step": 3460
},
{
"epoch": 0.41354723707664887,
"grad_norm": 1.3515625,
"learning_rate": 0.0001668506130170453,
"loss": 0.995,
"step": 3480
},
{
"epoch": 0.41592394533571003,
"grad_norm": 1.7421875,
"learning_rate": 0.0001661535868474273,
"loss": 1.1595,
"step": 3500
},
{
"epoch": 0.41830065359477125,
"grad_norm": 2.15625,
"learning_rate": 0.00016545079803413892,
"loss": 1.1134,
"step": 3520
},
{
"epoch": 0.4206773618538324,
"grad_norm": 0.92578125,
"learning_rate": 0.00016474230779716384,
"loss": 0.7802,
"step": 3540
},
{
"epoch": 0.42305407011289364,
"grad_norm": 1.671875,
"learning_rate": 0.00016402817785313712,
"loss": 0.9085,
"step": 3560
},
{
"epoch": 0.42543077837195487,
"grad_norm": 2.390625,
"learning_rate": 0.00016330847040996915,
"loss": 1.0042,
"step": 3580
},
{
"epoch": 0.42780748663101603,
"grad_norm": 2.46875,
"learning_rate": 0.00016258324816142668,
"loss": 1.0318,
"step": 3600
},
{
"epoch": 0.43018419489007725,
"grad_norm": 3.375,
"learning_rate": 0.00016185257428167143,
"loss": 1.0833,
"step": 3620
},
{
"epoch": 0.4325609031491384,
"grad_norm": 2.078125,
"learning_rate": 0.00016111651241975734,
"loss": 1.0208,
"step": 3640
},
{
"epoch": 0.43493761140819964,
"grad_norm": 2.125,
"learning_rate": 0.00016037512669408565,
"loss": 1.0074,
"step": 3660
},
{
"epoch": 0.43731431966726086,
"grad_norm": 1.5703125,
"learning_rate": 0.0001596284816868198,
"loss": 1.2403,
"step": 3680
},
{
"epoch": 0.43969102792632203,
"grad_norm": 2.59375,
"learning_rate": 0.00015887664243825967,
"loss": 1.0694,
"step": 3700
},
{
"epoch": 0.44206773618538325,
"grad_norm": 1.3203125,
"learning_rate": 0.0001581196744411759,
"loss": 1.0033,
"step": 3720
},
{
"epoch": 0.4444444444444444,
"grad_norm": 2.359375,
"learning_rate": 0.0001573576436351046,
"loss": 1.0574,
"step": 3740
},
{
"epoch": 0.44682115270350564,
"grad_norm": 2.109375,
"learning_rate": 0.00015659061640060378,
"loss": 1.2204,
"step": 3760
},
{
"epoch": 0.44919786096256686,
"grad_norm": 1.8046875,
"learning_rate": 0.0001558186595534705,
"loss": 1.157,
"step": 3780
},
{
"epoch": 0.45157456922162803,
"grad_norm": 1.6171875,
"learning_rate": 0.0001550418403389208,
"loss": 0.9542,
"step": 3800
},
{
"epoch": 0.45395127748068925,
"grad_norm": 2.390625,
"learning_rate": 0.00015426022642573193,
"loss": 0.9281,
"step": 3820
},
{
"epoch": 0.4563279857397504,
"grad_norm": 1.4453125,
"learning_rate": 0.00015347388590034757,
"loss": 0.9331,
"step": 3840
},
{
"epoch": 0.45870469399881164,
"grad_norm": 1.3828125,
"learning_rate": 0.00015268288726094705,
"loss": 0.7978,
"step": 3860
},
{
"epoch": 0.46108140225787286,
"grad_norm": 2.421875,
"learning_rate": 0.00015188729941147824,
"loss": 0.9807,
"step": 3880
},
{
"epoch": 0.46345811051693403,
"grad_norm": 2.640625,
"learning_rate": 0.0001510871916556555,
"loss": 1.0974,
"step": 3900
},
{
"epoch": 0.46583481877599525,
"grad_norm": 1.171875,
"learning_rate": 0.00015028263369092253,
"loss": 0.9807,
"step": 3920
},
{
"epoch": 0.4682115270350565,
"grad_norm": 1.6484375,
"learning_rate": 0.00014947369560238104,
"loss": 1.0588,
"step": 3940
},
{
"epoch": 0.47058823529411764,
"grad_norm": 2.546875,
"learning_rate": 0.00014866044785668563,
"loss": 1.0442,
"step": 3960
},
{
"epoch": 0.47296494355317886,
"grad_norm": 2.9375,
"learning_rate": 0.00014784296129590548,
"loss": 0.8837,
"step": 3980
},
{
"epoch": 0.47534165181224003,
"grad_norm": 1.7890625,
"learning_rate": 0.00014702130713135317,
"loss": 1.129,
"step": 4000
},
{
"epoch": 0.47771836007130125,
"grad_norm": 2.171875,
"learning_rate": 0.00014619555693738166,
"loss": 1.0496,
"step": 4020
},
{
"epoch": 0.4800950683303625,
"grad_norm": 2.328125,
"learning_rate": 0.0001453657826451493,
"loss": 1.0136,
"step": 4040
},
{
"epoch": 0.48247177658942364,
"grad_norm": 2.203125,
"learning_rate": 0.00014453205653635376,
"loss": 0.9794,
"step": 4060
},
{
"epoch": 0.48484848484848486,
"grad_norm": 2.734375,
"learning_rate": 0.00014369445123693596,
"loss": 0.7651,
"step": 4080
},
{
"epoch": 0.48722519310754603,
"grad_norm": 2.15625,
"learning_rate": 0.0001428530397107533,
"loss": 0.9667,
"step": 4100
},
{
"epoch": 0.48960190136660725,
"grad_norm": 2.578125,
"learning_rate": 0.0001420078952532238,
"loss": 0.8897,
"step": 4120
},
{
"epoch": 0.4919786096256685,
"grad_norm": 4.28125,
"learning_rate": 0.0001411590914849415,
"loss": 1.0208,
"step": 4140
},
{
"epoch": 0.49435531788472964,
"grad_norm": 2.34375,
"learning_rate": 0.00014030670234526323,
"loss": 1.0138,
"step": 4160
},
{
"epoch": 0.49673202614379086,
"grad_norm": 1.234375,
"learning_rate": 0.00013945080208586775,
"loss": 0.8435,
"step": 4180
},
{
"epoch": 0.49910873440285203,
"grad_norm": 2.828125,
"learning_rate": 0.0001385914652642877,
"loss": 0.7405,
"step": 4200
},
{
"epoch": 0.5014854426619133,
"grad_norm": 2.59375,
"learning_rate": 0.00013772876673741498,
"loss": 0.8999,
"step": 4220
},
{
"epoch": 0.5038621509209744,
"grad_norm": 2.171875,
"learning_rate": 0.00013686278165497977,
"loss": 0.9063,
"step": 4240
},
{
"epoch": 0.5062388591800356,
"grad_norm": 3.578125,
"learning_rate": 0.00013599358545300438,
"loss": 0.8485,
"step": 4260
},
{
"epoch": 0.5086155674390969,
"grad_norm": 2.5625,
"learning_rate": 0.00013512125384723204,
"loss": 0.7843,
"step": 4280
},
{
"epoch": 0.5109922756981581,
"grad_norm": 1.5546875,
"learning_rate": 0.00013424586282653116,
"loss": 0.7548,
"step": 4300
},
{
"epoch": 0.5133689839572193,
"grad_norm": 2.8125,
"learning_rate": 0.00013336748864627593,
"loss": 0.9609,
"step": 4320
},
{
"epoch": 0.5157456922162804,
"grad_norm": 2.453125,
"learning_rate": 0.00013248620782170396,
"loss": 0.9358,
"step": 4340
},
{
"epoch": 0.5181224004753416,
"grad_norm": 2.21875,
"learning_rate": 0.00013160209712125074,
"loss": 1.0639,
"step": 4360
},
{
"epoch": 0.5204991087344029,
"grad_norm": 1.515625,
"learning_rate": 0.0001307152335598624,
"loss": 1.0147,
"step": 4380
},
{
"epoch": 0.5228758169934641,
"grad_norm": 2.4375,
"learning_rate": 0.00012982569439228713,
"loss": 0.7749,
"step": 4400
},
{
"epoch": 0.5252525252525253,
"grad_norm": 2.21875,
"learning_rate": 0.0001289335571063453,
"loss": 0.642,
"step": 4420
},
{
"epoch": 0.5276292335115864,
"grad_norm": 2.984375,
"learning_rate": 0.00012803889941617944,
"loss": 0.863,
"step": 4440
},
{
"epoch": 0.5300059417706476,
"grad_norm": 2.265625,
"learning_rate": 0.0001271417992554849,
"loss": 0.8817,
"step": 4460
},
{
"epoch": 0.5323826500297089,
"grad_norm": 2.3125,
"learning_rate": 0.00012624233477072057,
"loss": 0.9788,
"step": 4480
},
{
"epoch": 0.5347593582887701,
"grad_norm": 2.484375,
"learning_rate": 0.00012534058431430198,
"loss": 0.8273,
"step": 4500
},
{
"epoch": 0.5371360665478313,
"grad_norm": 1.2265625,
"learning_rate": 0.0001244366264377757,
"loss": 0.848,
"step": 4520
},
{
"epoch": 0.5395127748068924,
"grad_norm": 2.625,
"learning_rate": 0.00012353053988497684,
"loss": 0.8556,
"step": 4540
},
{
"epoch": 0.5418894830659536,
"grad_norm": 2.375,
"learning_rate": 0.00012262240358516967,
"loss": 0.8137,
"step": 4560
},
{
"epoch": 0.5442661913250149,
"grad_norm": 1.1015625,
"learning_rate": 0.00012171229664617208,
"loss": 0.7362,
"step": 4580
},
{
"epoch": 0.5466428995840761,
"grad_norm": 2.90625,
"learning_rate": 0.00012080029834746447,
"loss": 0.7504,
"step": 4600
},
{
"epoch": 0.5490196078431373,
"grad_norm": 3.109375,
"learning_rate": 0.00011988648813328367,
"loss": 1.0486,
"step": 4620
},
{
"epoch": 0.5513963161021984,
"grad_norm": 1.875,
"learning_rate": 0.00011897094560570265,
"loss": 0.7487,
"step": 4640
},
{
"epoch": 0.5537730243612596,
"grad_norm": 2.171875,
"learning_rate": 0.00011805375051769636,
"loss": 0.9325,
"step": 4660
},
{
"epoch": 0.5561497326203209,
"grad_norm": 2.28125,
"learning_rate": 0.00011713498276619432,
"loss": 0.906,
"step": 4680
},
{
"epoch": 0.5585264408793821,
"grad_norm": 1.953125,
"learning_rate": 0.0001162147223851209,
"loss": 0.7214,
"step": 4700
},
{
"epoch": 0.5609031491384433,
"grad_norm": 2.34375,
"learning_rate": 0.0001152930495384236,
"loss": 1.01,
"step": 4720
},
{
"epoch": 0.5632798573975044,
"grad_norm": 2.53125,
"learning_rate": 0.00011437004451308983,
"loss": 1.0897,
"step": 4740
},
{
"epoch": 0.5656565656565656,
"grad_norm": 3.296875,
"learning_rate": 0.00011344578771215319,
"loss": 0.8222,
"step": 4760
},
{
"epoch": 0.5680332739156269,
"grad_norm": 2.921875,
"learning_rate": 0.00011252035964768961,
"loss": 0.8411,
"step": 4780
},
{
"epoch": 0.5704099821746881,
"grad_norm": 1.9375,
"learning_rate": 0.00011159384093380377,
"loss": 0.747,
"step": 4800
},
{
"epoch": 0.5727866904337493,
"grad_norm": 2.375,
"learning_rate": 0.00011066631227960693,
"loss": 0.7352,
"step": 4820
},
{
"epoch": 0.5751633986928104,
"grad_norm": 2.484375,
"learning_rate": 0.00010973785448218639,
"loss": 0.901,
"step": 4840
},
{
"epoch": 0.5775401069518716,
"grad_norm": 1.7265625,
"learning_rate": 0.00010880854841956712,
"loss": 0.6489,
"step": 4860
},
{
"epoch": 0.5799168152109329,
"grad_norm": 1.3359375,
"learning_rate": 0.00010787847504366649,
"loss": 0.7034,
"step": 4880
},
{
"epoch": 0.5822935234699941,
"grad_norm": 3.046875,
"learning_rate": 0.00010694771537324269,
"loss": 0.7427,
"step": 4900
},
{
"epoch": 0.5846702317290553,
"grad_norm": 3.40625,
"learning_rate": 0.00010601635048683698,
"loss": 0.7655,
"step": 4920
},
{
"epoch": 0.5870469399881164,
"grad_norm": 2.21875,
"learning_rate": 0.00010508446151571109,
"loss": 0.7467,
"step": 4940
},
{
"epoch": 0.5894236482471776,
"grad_norm": 1.9375,
"learning_rate": 0.0001041521296367798,
"loss": 0.7942,
"step": 4960
},
{
"epoch": 0.5918003565062389,
"grad_norm": 3.34375,
"learning_rate": 0.00010321943606553961,
"loss": 1.033,
"step": 4980
},
{
"epoch": 0.5941770647653001,
"grad_norm": 2.765625,
"learning_rate": 0.00010228646204899401,
"loss": 0.7049,
"step": 5000
},
{
"epoch": 0.5965537730243613,
"grad_norm": 1.59375,
"learning_rate": 0.0001013532888585762,
"loss": 0.7016,
"step": 5020
},
{
"epoch": 0.5989304812834224,
"grad_norm": 2.875,
"learning_rate": 0.00010041999778306936,
"loss": 0.9511,
"step": 5040
},
{
"epoch": 0.6013071895424836,
"grad_norm": 3.640625,
"learning_rate": 9.948667012152566e-05,
"loss": 0.6895,
"step": 5060
},
{
"epoch": 0.6036838978015449,
"grad_norm": 3.3125,
"learning_rate": 9.855338717618432e-05,
"loss": 0.8696,
"step": 5080
},
{
"epoch": 0.6060606060606061,
"grad_norm": 2.625,
"learning_rate": 9.762023024538926e-05,
"loss": 0.8621,
"step": 5100
},
{
"epoch": 0.6084373143196673,
"grad_norm": 1.984375,
"learning_rate": 9.668728061650733e-05,
"loss": 0.9548,
"step": 5120
},
{
"epoch": 0.6108140225787284,
"grad_norm": 3.40625,
"learning_rate": 9.575461955884726e-05,
"loss": 0.8048,
"step": 5140
},
{
"epoch": 0.6131907308377896,
"grad_norm": 2.265625,
"learning_rate": 9.482232831658034e-05,
"loss": 0.8866,
"step": 5160
},
{
"epoch": 0.6155674390968509,
"grad_norm": 1.5390625,
"learning_rate": 9.389048810166317e-05,
"loss": 0.7404,
"step": 5180
},
{
"epoch": 0.6179441473559121,
"grad_norm": 1.359375,
"learning_rate": 9.29591800867634e-05,
"loss": 0.6817,
"step": 5200
},
{
"epoch": 0.6203208556149733,
"grad_norm": 5.0,
"learning_rate": 9.202848539818865e-05,
"loss": 0.7504,
"step": 5220
},
{
"epoch": 0.6226975638740344,
"grad_norm": 2.484375,
"learning_rate": 9.109848510881961e-05,
"loss": 0.7767,
"step": 5240
},
{
"epoch": 0.6250742721330956,
"grad_norm": 2.53125,
"learning_rate": 9.016926023104789e-05,
"loss": 0.7846,
"step": 5260
},
{
"epoch": 0.6274509803921569,
"grad_norm": 3.421875,
"learning_rate": 8.924089170971887e-05,
"loss": 0.8327,
"step": 5280
},
{
"epoch": 0.6298276886512181,
"grad_norm": 3.765625,
"learning_rate": 8.831346041508069e-05,
"loss": 0.8035,
"step": 5300
},
{
"epoch": 0.6322043969102793,
"grad_norm": 2.25,
"learning_rate": 8.738704713573959e-05,
"loss": 0.9569,
"step": 5320
},
{
"epoch": 0.6345811051693404,
"grad_norm": 2.34375,
"learning_rate": 8.646173257162245e-05,
"loss": 0.6629,
"step": 5340
},
{
"epoch": 0.6369578134284016,
"grad_norm": 2.875,
"learning_rate": 8.553759732694696e-05,
"loss": 0.6597,
"step": 5360
},
{
"epoch": 0.6393345216874629,
"grad_norm": 2.859375,
"learning_rate": 8.461472190320021e-05,
"loss": 0.6857,
"step": 5380
},
{
"epoch": 0.6417112299465241,
"grad_norm": 1.5234375,
"learning_rate": 8.369318669212625e-05,
"loss": 0.7234,
"step": 5400
},
{
"epoch": 0.6440879382055853,
"grad_norm": 3.359375,
"learning_rate": 8.277307196872303e-05,
"loss": 0.7483,
"step": 5420
},
{
"epoch": 0.6464646464646465,
"grad_norm": 1.21875,
"learning_rate": 8.185445788424974e-05,
"loss": 0.7538,
"step": 5440
},
{
"epoch": 0.6488413547237076,
"grad_norm": 3.234375,
"learning_rate": 8.093742445924491e-05,
"loss": 0.7882,
"step": 5460
},
{
"epoch": 0.6512180629827689,
"grad_norm": 3.546875,
"learning_rate": 8.002205157655554e-05,
"loss": 0.6984,
"step": 5480
},
{
"epoch": 0.6535947712418301,
"grad_norm": 2.28125,
"learning_rate": 7.910841897437875e-05,
"loss": 0.7044,
"step": 5500
},
{
"epoch": 0.6559714795008913,
"grad_norm": 3.390625,
"learning_rate": 7.819660623931575e-05,
"loss": 0.6569,
"step": 5520
},
{
"epoch": 0.6583481877599525,
"grad_norm": 1.5859375,
"learning_rate": 7.728669279943897e-05,
"loss": 0.8069,
"step": 5540
},
{
"epoch": 0.6607248960190136,
"grad_norm": 3.125,
"learning_rate": 7.637875791737299e-05,
"loss": 0.759,
"step": 5560
},
{
"epoch": 0.6631016042780749,
"grad_norm": 1.4921875,
"learning_rate": 7.547288068339025e-05,
"loss": 0.6806,
"step": 5580
},
{
"epoch": 0.6654783125371361,
"grad_norm": 2.875,
"learning_rate": 7.456914000852123e-05,
"loss": 0.7777,
"step": 5600
},
{
"epoch": 0.6678550207961973,
"grad_norm": 1.453125,
"learning_rate": 7.366761461768052e-05,
"loss": 0.7375,
"step": 5620
},
{
"epoch": 0.6702317290552585,
"grad_norm": 1.734375,
"learning_rate": 7.276838304280935e-05,
"loss": 0.6862,
"step": 5640
},
{
"epoch": 0.6726084373143196,
"grad_norm": 3.171875,
"learning_rate": 7.187152361603432e-05,
"loss": 0.6195,
"step": 5660
},
{
"epoch": 0.6749851455733809,
"grad_norm": 1.8046875,
"learning_rate": 7.097711446284405e-05,
"loss": 0.5486,
"step": 5680
},
{
"epoch": 0.6773618538324421,
"grad_norm": 2.203125,
"learning_rate": 7.008523349528377e-05,
"loss": 0.7452,
"step": 5700
},
{
"epoch": 0.6797385620915033,
"grad_norm": 3.03125,
"learning_rate": 6.919595840516815e-05,
"loss": 0.7417,
"step": 5720
},
{
"epoch": 0.6821152703505645,
"grad_norm": 3.1875,
"learning_rate": 6.830936665731371e-05,
"loss": 0.6798,
"step": 5740
},
{
"epoch": 0.6844919786096256,
"grad_norm": 2.328125,
"learning_rate": 6.742553548279095e-05,
"loss": 0.7222,
"step": 5760
},
{
"epoch": 0.6868686868686869,
"grad_norm": 1.0625,
"learning_rate": 6.654454187219649e-05,
"loss": 0.6629,
"step": 5780
},
{
"epoch": 0.6892453951277481,
"grad_norm": 1.6015625,
"learning_rate": 6.56664625689466e-05,
"loss": 0.7808,
"step": 5800
},
{
"epoch": 0.6916221033868093,
"grad_norm": 2.421875,
"learning_rate": 6.479137406259206e-05,
"loss": 0.7898,
"step": 5820
},
{
"epoch": 0.6939988116458705,
"grad_norm": 1.53125,
"learning_rate": 6.39193525821551e-05,
"loss": 0.8595,
"step": 5840
},
{
"epoch": 0.6963755199049316,
"grad_norm": 3.21875,
"learning_rate": 6.305047408948908e-05,
"loss": 0.8169,
"step": 5860
},
{
"epoch": 0.6987522281639929,
"grad_norm": 2.796875,
"learning_rate": 6.21848142726615e-05,
"loss": 0.7892,
"step": 5880
},
{
"epoch": 0.7011289364230541,
"grad_norm": 2.671875,
"learning_rate": 6.13224485393608e-05,
"loss": 0.8179,
"step": 5900
},
{
"epoch": 0.7035056446821153,
"grad_norm": 1.8046875,
"learning_rate": 6.046345201032748e-05,
"loss": 0.6881,
"step": 5920
},
{
"epoch": 0.7058823529411765,
"grad_norm": 1.515625,
"learning_rate": 5.960789951281052e-05,
"loss": 0.9263,
"step": 5940
},
{
"epoch": 0.7082590612002376,
"grad_norm": 2.21875,
"learning_rate": 5.8755865574049016e-05,
"loss": 0.7661,
"step": 5960
},
{
"epoch": 0.7106357694592988,
"grad_norm": 2.5625,
"learning_rate": 5.7907424414780135e-05,
"loss": 0.7932,
"step": 5980
},
{
"epoch": 0.7130124777183601,
"grad_norm": 1.8984375,
"learning_rate": 5.706264994277386e-05,
"loss": 0.8045,
"step": 6000
},
{
"epoch": 0.7153891859774213,
"grad_norm": 1.09375,
"learning_rate": 5.6221615746394644e-05,
"loss": 0.6996,
"step": 6020
},
{
"epoch": 0.7177658942364825,
"grad_norm": 1.796875,
"learning_rate": 5.538439508819139e-05,
"loss": 0.7662,
"step": 6040
},
{
"epoch": 0.7201426024955436,
"grad_norm": 1.890625,
"learning_rate": 5.4551060898515404e-05,
"loss": 0.6049,
"step": 6060
},
{
"epoch": 0.7225193107546048,
"grad_norm": 2.15625,
"learning_rate": 5.372168576916732e-05,
"loss": 0.6502,
"step": 6080
},
{
"epoch": 0.7248960190136661,
"grad_norm": 2.6875,
"learning_rate": 5.289634194707387e-05,
"loss": 0.6785,
"step": 6100
},
{
"epoch": 0.7272727272727273,
"grad_norm": 2.890625,
"learning_rate": 5.207510132799436e-05,
"loss": 0.7902,
"step": 6120
},
{
"epoch": 0.7296494355317885,
"grad_norm": 2.703125,
"learning_rate": 5.125803545025758e-05,
"loss": 0.6948,
"step": 6140
},
{
"epoch": 0.7320261437908496,
"grad_norm": 2.84375,
"learning_rate": 5.0445215488530525e-05,
"loss": 0.7113,
"step": 6160
},
{
"epoch": 0.7344028520499108,
"grad_norm": 3.484375,
"learning_rate": 4.963671224761808e-05,
"loss": 0.6249,
"step": 6180
},
{
"epoch": 0.7367795603089721,
"grad_norm": 1.59375,
"learning_rate": 4.883259615629515e-05,
"loss": 0.7295,
"step": 6200
},
{
"epoch": 0.7391562685680333,
"grad_norm": 1.359375,
"learning_rate": 4.8032937261171896e-05,
"loss": 0.8041,
"step": 6220
},
{
"epoch": 0.7415329768270945,
"grad_norm": 1.8046875,
"learning_rate": 4.7237805220591744e-05,
"loss": 0.6296,
"step": 6240
},
{
"epoch": 0.7439096850861556,
"grad_norm": 1.609375,
"learning_rate": 4.644726929856342e-05,
"loss": 0.8887,
"step": 6260
},
{
"epoch": 0.7462863933452168,
"grad_norm": 2.421875,
"learning_rate": 4.5661398358727524e-05,
"loss": 0.6501,
"step": 6280
},
{
"epoch": 0.7486631016042781,
"grad_norm": 1.7109375,
"learning_rate": 4.4880260858357746e-05,
"loss": 0.5981,
"step": 6300
},
{
"epoch": 0.7510398098633393,
"grad_norm": 1.765625,
"learning_rate": 4.4103924842397395e-05,
"loss": 0.5993,
"step": 6320
},
{
"epoch": 0.7534165181224005,
"grad_norm": 3.484375,
"learning_rate": 4.3332457937532246e-05,
"loss": 0.8187,
"step": 6340
},
{
"epoch": 0.7557932263814616,
"grad_norm": 7.09375,
"learning_rate": 4.256592734629947e-05,
"loss": 0.826,
"step": 6360
},
{
"epoch": 0.7581699346405228,
"grad_norm": 1.578125,
"learning_rate": 4.18043998412335e-05,
"loss": 0.5091,
"step": 6380
},
{
"epoch": 0.7605466428995841,
"grad_norm": 0.78125,
"learning_rate": 4.104794175904966e-05,
"loss": 0.609,
"step": 6400
},
{
"epoch": 0.7629233511586453,
"grad_norm": 3.171875,
"learning_rate": 4.02966189948655e-05,
"loss": 0.7031,
"step": 6420
},
{
"epoch": 0.7653000594177065,
"grad_norm": 1.6015625,
"learning_rate": 3.955049699646054e-05,
"loss": 0.5454,
"step": 6440
},
{
"epoch": 0.7676767676767676,
"grad_norm": 3.578125,
"learning_rate": 3.880964075857535e-05,
"loss": 0.629,
"step": 6460
},
{
"epoch": 0.7700534759358288,
"grad_norm": 2.609375,
"learning_rate": 3.80741148172497e-05,
"loss": 0.6376,
"step": 6480
},
{
"epoch": 0.7724301841948901,
"grad_norm": 2.015625,
"learning_rate": 3.734398324420073e-05,
"loss": 0.6744,
"step": 6500
},
{
"epoch": 0.7748068924539513,
"grad_norm": 3.140625,
"learning_rate": 3.661930964124193e-05,
"loss": 0.6133,
"step": 6520
},
{
"epoch": 0.7771836007130125,
"grad_norm": 1.734375,
"learning_rate": 3.5900157134742574e-05,
"loss": 0.5723,
"step": 6540
},
{
"epoch": 0.7795603089720736,
"grad_norm": 3.921875,
"learning_rate": 3.5186588370128746e-05,
"loss": 0.6248,
"step": 6560
},
{
"epoch": 0.7819370172311348,
"grad_norm": 3.203125,
"learning_rate": 3.447866550642649e-05,
"loss": 0.6323,
"step": 6580
},
{
"epoch": 0.7843137254901961,
"grad_norm": 3.1875,
"learning_rate": 3.377645021084701e-05,
"loss": 0.7705,
"step": 6600
},
{
"epoch": 0.7866904337492573,
"grad_norm": 2.546875,
"learning_rate": 3.3080003653414724e-05,
"loss": 0.5787,
"step": 6620
},
{
"epoch": 0.7890671420083185,
"grad_norm": 1.2109375,
"learning_rate": 3.238938650163899e-05,
"loss": 0.7117,
"step": 6640
},
{
"epoch": 0.7914438502673797,
"grad_norm": 1.640625,
"learning_rate": 3.17046589152292e-05,
"loss": 0.5979,
"step": 6660
},
{
"epoch": 0.7938205585264408,
"grad_norm": 4.53125,
"learning_rate": 3.10258805408542e-05,
"loss": 0.6202,
"step": 6680
},
{
"epoch": 0.7961972667855021,
"grad_norm": 3.484375,
"learning_rate": 3.0353110506946647e-05,
"loss": 0.6462,
"step": 6700
},
{
"epoch": 0.7985739750445633,
"grad_norm": 2.640625,
"learning_rate": 2.968640741855223e-05,
"loss": 0.576,
"step": 6720
},
{
"epoch": 0.8009506833036245,
"grad_norm": 1.875,
"learning_rate": 2.9025829352224477e-05,
"loss": 0.7645,
"step": 6740
},
{
"epoch": 0.8033273915626857,
"grad_norm": 2.453125,
"learning_rate": 2.8371433850965922e-05,
"loss": 0.6348,
"step": 6760
},
{
"epoch": 0.8057040998217468,
"grad_norm": 2.875,
"learning_rate": 2.7723277919215397e-05,
"loss": 0.7165,
"step": 6780
},
{
"epoch": 0.8080808080808081,
"grad_norm": 1.7890625,
"learning_rate": 2.70814180178823e-05,
"loss": 0.6772,
"step": 6800
},
{
"epoch": 0.8104575163398693,
"grad_norm": 3.25,
"learning_rate": 2.644591005942846e-05,
"loss": 0.6995,
"step": 6820
},
{
"epoch": 0.8128342245989305,
"grad_norm": 3.546875,
"learning_rate": 2.5816809402997522e-05,
"loss": 0.6519,
"step": 6840
},
{
"epoch": 0.8152109328579917,
"grad_norm": 4.375,
"learning_rate": 2.5194170849592492e-05,
"loss": 0.612,
"step": 6860
},
{
"epoch": 0.8175876411170528,
"grad_norm": 3.40625,
"learning_rate": 2.4578048637302208e-05,
"loss": 0.8347,
"step": 6880
},
{
"epoch": 0.8199643493761141,
"grad_norm": 3.953125,
"learning_rate": 2.396849643657657e-05,
"loss": 0.7846,
"step": 6900
},
{
"epoch": 0.8223410576351753,
"grad_norm": 3.0,
"learning_rate": 2.3365567345551233e-05,
"loss": 0.6799,
"step": 6920
},
{
"epoch": 0.8247177658942365,
"grad_norm": 2.15625,
"learning_rate": 2.276931388542235e-05,
"loss": 0.704,
"step": 6940
},
{
"epoch": 0.8270944741532977,
"grad_norm": 3.3125,
"learning_rate": 2.2179787995871403e-05,
"loss": 0.5832,
"step": 6960
},
{
"epoch": 0.8294711824123588,
"grad_norm": 1.8046875,
"learning_rate": 2.1597041030540643e-05,
"loss": 0.6464,
"step": 6980
},
{
"epoch": 0.8318478906714201,
"grad_norm": 1.0859375,
"learning_rate": 2.1021123752559836e-05,
"loss": 0.6596,
"step": 7000
},
{
"epoch": 0.8342245989304813,
"grad_norm": 2.4375,
"learning_rate": 2.0452086330124164e-05,
"loss": 0.6124,
"step": 7020
},
{
"epoch": 0.8366013071895425,
"grad_norm": 1.953125,
"learning_rate": 1.988997833212406e-05,
"loss": 0.8721,
"step": 7040
},
{
"epoch": 0.8389780154486037,
"grad_norm": 1.0234375,
"learning_rate": 1.933484872382737e-05,
"loss": 0.6729,
"step": 7060
},
{
"epoch": 0.8413547237076648,
"grad_norm": 3.859375,
"learning_rate": 1.8786745862613885e-05,
"loss": 0.7747,
"step": 7080
},
{
"epoch": 0.8437314319667261,
"grad_norm": 2.703125,
"learning_rate": 1.8245717493762925e-05,
"loss": 0.6807,
"step": 7100
},
{
"epoch": 0.8461081402257873,
"grad_norm": 2.4375,
"learning_rate": 1.7711810746294312e-05,
"loss": 0.851,
"step": 7120
},
{
"epoch": 0.8484848484848485,
"grad_norm": 1.4765625,
"learning_rate": 1.7185072128862933e-05,
"loss": 0.6449,
"step": 7140
},
{
"epoch": 0.8508615567439097,
"grad_norm": 2.140625,
"learning_rate": 1.6665547525707316e-05,
"loss": 0.6683,
"step": 7160
},
{
"epoch": 0.8532382650029708,
"grad_norm": 2.25,
"learning_rate": 1.6153282192652698e-05,
"loss": 0.5818,
"step": 7180
},
{
"epoch": 0.8556149732620321,
"grad_norm": 2.1875,
"learning_rate": 1.5648320753168844e-05,
"loss": 0.6614,
"step": 7200
},
{
"epoch": 0.8579916815210933,
"grad_norm": 2.5625,
"learning_rate": 1.5150707194482695e-05,
"loss": 0.7086,
"step": 7220
},
{
"epoch": 0.8603683897801545,
"grad_norm": 2.734375,
"learning_rate": 1.4660484863746938e-05,
"loss": 0.778,
"step": 7240
},
{
"epoch": 0.8627450980392157,
"grad_norm": 2.109375,
"learning_rate": 1.4177696464263723e-05,
"loss": 0.7629,
"step": 7260
},
{
"epoch": 0.8651218062982768,
"grad_norm": 3.1875,
"learning_rate": 1.3702384051765005e-05,
"loss": 0.8923,
"step": 7280
},
{
"epoch": 0.8674985145573381,
"grad_norm": 4.125,
"learning_rate": 1.3234589030748956e-05,
"loss": 0.6005,
"step": 7300
},
{
"epoch": 0.8698752228163993,
"grad_norm": 2.9375,
"learning_rate": 1.2774352150873203e-05,
"loss": 0.741,
"step": 7320
},
{
"epoch": 0.8722519310754605,
"grad_norm": 2.78125,
"learning_rate": 1.2321713503405208e-05,
"loss": 0.5276,
"step": 7340
},
{
"epoch": 0.8746286393345217,
"grad_norm": 3.28125,
"learning_rate": 1.187671251772987e-05,
"loss": 0.5997,
"step": 7360
},
{
"epoch": 0.8770053475935828,
"grad_norm": 1.8125,
"learning_rate": 1.143938795791476e-05,
"loss": 0.6653,
"step": 7380
},
{
"epoch": 0.8793820558526441,
"grad_norm": 2.46875,
"learning_rate": 1.1009777919333507e-05,
"loss": 0.5658,
"step": 7400
},
{
"epoch": 0.8817587641117053,
"grad_norm": 2.609375,
"learning_rate": 1.0587919825347236e-05,
"loss": 0.6695,
"step": 7420
},
{
"epoch": 0.8841354723707665,
"grad_norm": 1.7734375,
"learning_rate": 1.0173850424044596e-05,
"loss": 0.68,
"step": 7440
},
{
"epoch": 0.8865121806298277,
"grad_norm": 3.640625,
"learning_rate": 9.76760578504068e-06,
"loss": 0.8137,
"step": 7460
},
{
"epoch": 0.8888888888888888,
"grad_norm": 1.8125,
"learning_rate": 9.369221296335006e-06,
"loss": 0.7129,
"step": 7480
},
{
"epoch": 0.8912655971479501,
"grad_norm": 4.625,
"learning_rate": 8.978731661228768e-06,
"loss": 0.6628,
"step": 7500
},
{
"epoch": 0.8936423054070113,
"grad_norm": 4.125,
"learning_rate": 8.596170895301959e-06,
"loss": 0.6621,
"step": 7520
},
{
"epoch": 0.8960190136660725,
"grad_norm": 2.953125,
"learning_rate": 8.221572323450222e-06,
"loss": 0.6171,
"step": 7540
},
{
"epoch": 0.8983957219251337,
"grad_norm": 3.078125,
"learning_rate": 7.854968576981824e-06,
"loss": 0.5361,
"step": 7560
},
{
"epoch": 0.9007724301841948,
"grad_norm": 3.8125,
"learning_rate": 7.49639159077532e-06,
"loss": 0.6402,
"step": 7580
},
{
"epoch": 0.9031491384432561,
"grad_norm": 3.25,
"learning_rate": 7.145872600497561e-06,
"loss": 0.5515,
"step": 7600
},
{
"epoch": 0.9055258467023173,
"grad_norm": 2.078125,
"learning_rate": 6.8034421398827765e-06,
"loss": 0.6165,
"step": 7620
},
{
"epoch": 0.9079025549613785,
"grad_norm": 3.78125,
"learning_rate": 6.469130038072835e-06,
"loss": 0.8606,
"step": 7640
},
{
"epoch": 0.9102792632204397,
"grad_norm": 2.875,
"learning_rate": 6.142965417018798e-06,
"loss": 0.6364,
"step": 7660
},
{
"epoch": 0.9126559714795008,
"grad_norm": 2.53125,
"learning_rate": 5.824976688944051e-06,
"loss": 0.578,
"step": 7680
},
{
"epoch": 0.9150326797385621,
"grad_norm": 2.15625,
"learning_rate": 5.515191553869381e-06,
"loss": 0.5698,
"step": 7700
},
{
"epoch": 0.9174093879976233,
"grad_norm": 2.96875,
"learning_rate": 5.213636997200044e-06,
"loss": 0.6748,
"step": 7720
},
{
"epoch": 0.9197860962566845,
"grad_norm": 4.0625,
"learning_rate": 4.920339287374942e-06,
"loss": 0.7089,
"step": 7740
},
{
"epoch": 0.9221628045157457,
"grad_norm": 3.578125,
"learning_rate": 4.635323973578543e-06,
"loss": 0.7648,
"step": 7760
},
{
"epoch": 0.9245395127748068,
"grad_norm": 1.0390625,
"learning_rate": 4.3586158835151495e-06,
"loss": 0.6778,
"step": 7780
},
{
"epoch": 0.9269162210338681,
"grad_norm": 3.578125,
"learning_rate": 4.090239121246231e-06,
"loss": 0.7377,
"step": 7800
},
{
"epoch": 0.9292929292929293,
"grad_norm": 3.5,
"learning_rate": 3.830217065090702e-06,
"loss": 0.6488,
"step": 7820
},
{
"epoch": 0.9316696375519905,
"grad_norm": 1.40625,
"learning_rate": 3.5785723655884287e-06,
"loss": 0.5778,
"step": 7840
},
{
"epoch": 0.9340463458110517,
"grad_norm": 3.3125,
"learning_rate": 3.335326943527117e-06,
"loss": 0.6642,
"step": 7860
},
{
"epoch": 0.936423054070113,
"grad_norm": 3.421875,
"learning_rate": 3.100501988032878e-06,
"loss": 0.5973,
"step": 7880
},
{
"epoch": 0.9387997623291741,
"grad_norm": 4.71875,
"learning_rate": 2.874117954724309e-06,
"loss": 0.618,
"step": 7900
},
{
"epoch": 0.9411764705882353,
"grad_norm": 3.921875,
"learning_rate": 2.656194563930714e-06,
"loss": 0.7053,
"step": 7920
},
{
"epoch": 0.9435531788472965,
"grad_norm": 2.953125,
"learning_rate": 2.446750798974229e-06,
"loss": 0.6074,
"step": 7940
},
{
"epoch": 0.9459298871063577,
"grad_norm": 2.46875,
"learning_rate": 2.2458049045161244e-06,
"loss": 0.5246,
"step": 7960
},
{
"epoch": 0.948306595365419,
"grad_norm": 3.484375,
"learning_rate": 2.0533743849676436e-06,
"loss": 0.648,
"step": 7980
},
{
"epoch": 0.9506833036244801,
"grad_norm": 2.953125,
"learning_rate": 1.869476002965065e-06,
"loss": 0.6625,
"step": 8000
},
{
"epoch": 0.9530600118835413,
"grad_norm": 1.5859375,
"learning_rate": 1.69412577790955e-06,
"loss": 0.7165,
"step": 8020
},
{
"epoch": 0.9554367201426025,
"grad_norm": 1.3046875,
"learning_rate": 1.5273389845717245e-06,
"loss": 0.6024,
"step": 8040
},
{
"epoch": 0.9578134284016637,
"grad_norm": 3.6875,
"learning_rate": 1.3691301517610554e-06,
"loss": 0.6466,
"step": 8060
},
{
"epoch": 0.960190136660725,
"grad_norm": 1.171875,
"learning_rate": 1.2195130610602623e-06,
"loss": 0.6671,
"step": 8080
},
{
"epoch": 0.9625668449197861,
"grad_norm": 1.046875,
"learning_rate": 1.0785007456247886e-06,
"loss": 0.628,
"step": 8100
},
{
"epoch": 0.9649435531788473,
"grad_norm": 2.9375,
"learning_rate": 9.461054890474996e-07,
"loss": 0.776,
"step": 8120
},
{
"epoch": 0.9673202614379085,
"grad_norm": 1.671875,
"learning_rate": 8.223388242886265e-07,
"loss": 0.6368,
"step": 8140
},
{
"epoch": 0.9696969696969697,
"grad_norm": 2.625,
"learning_rate": 7.072115326711704e-07,
"loss": 0.5938,
"step": 8160
},
{
"epoch": 0.972073677956031,
"grad_norm": 1.171875,
"learning_rate": 6.00733642941742e-07,
"loss": 0.6331,
"step": 8180
},
{
"epoch": 0.9744503862150921,
"grad_norm": 2.234375,
"learning_rate": 5.029144303968724e-07,
"loss": 0.566,
"step": 8200
},
{
"epoch": 0.9768270944741533,
"grad_norm": 3.5,
"learning_rate": 4.1376241607518074e-07,
"loss": 0.6529,
"step": 8220
},
{
"epoch": 0.9792038027332145,
"grad_norm": 0.91015625,
"learning_rate": 3.332853660149904e-07,
"loss": 0.7165,
"step": 8240
},
{
"epoch": 0.9815805109922757,
"grad_norm": 2.5625,
"learning_rate": 2.6149029057785936e-07,
"loss": 0.6302,
"step": 8260
},
{
"epoch": 0.983957219251337,
"grad_norm": 0.97265625,
"learning_rate": 1.9838344383793505e-07,
"loss": 0.6525,
"step": 8280
},
{
"epoch": 0.9863339275103981,
"grad_norm": 2.90625,
"learning_rate": 1.4397032303715697e-07,
"loss": 0.6024,
"step": 8300
},
{
"epoch": 0.9887106357694593,
"grad_norm": 2.421875,
"learning_rate": 9.825566810633958e-08,
"loss": 0.4994,
"step": 8320
},
{
"epoch": 0.9910873440285205,
"grad_norm": 2.71875,
"learning_rate": 6.124346125233604e-08,
"loss": 0.7707,
"step": 8340
},
{
"epoch": 0.9934640522875817,
"grad_norm": 1.796875,
"learning_rate": 3.2936926611149e-08,
"loss": 0.6724,
"step": 8360
},
{
"epoch": 0.995840760546643,
"grad_norm": 1.5390625,
"learning_rate": 1.3338529967010793e-08,
"loss": 0.6268,
"step": 8380
},
{
"epoch": 0.9982174688057041,
"grad_norm": 3.828125,
"learning_rate": 2.449978537655273e-09,
"loss": 0.8173,
"step": 8400
},
{
"epoch": 1.0,
"step": 8415,
"total_flos": 9.688250026819584e+16,
"train_loss": 1.1022938394574842,
"train_runtime": 7319.3616,
"train_samples_per_second": 1.15,
"train_steps_per_second": 1.15
}
],
"logging_steps": 20,
"max_steps": 8415,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 100,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 9.688250026819584e+16,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}