lesso's picture
Training in progress, epoch 0, checkpoint
f82807b verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.08410959480202704,
"eval_steps": 500,
"global_step": 1000,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 8.410959480202704e-05,
"grad_norm": 0.27725711464881897,
"learning_rate": 0.00019999950652018584,
"loss": 0.8848,
"step": 1
},
{
"epoch": 0.00016821918960405408,
"grad_norm": 0.2778919041156769,
"learning_rate": 0.0001999980260856137,
"loss": 0.5957,
"step": 2
},
{
"epoch": 0.0002523287844060811,
"grad_norm": 0.5110663771629333,
"learning_rate": 0.000199995558710895,
"loss": 1.2037,
"step": 3
},
{
"epoch": 0.00033643837920810816,
"grad_norm": 0.4022953510284424,
"learning_rate": 0.00019999210442038162,
"loss": 1.1604,
"step": 4
},
{
"epoch": 0.0004205479740101352,
"grad_norm": 0.5099120736122131,
"learning_rate": 0.00019998766324816607,
"loss": 1.1557,
"step": 5
},
{
"epoch": 0.0005046575688121622,
"grad_norm": 0.4215346872806549,
"learning_rate": 0.0001999822352380809,
"loss": 1.0749,
"step": 6
},
{
"epoch": 0.0005887671636141893,
"grad_norm": 0.5019777417182922,
"learning_rate": 0.00019997582044369843,
"loss": 0.8256,
"step": 7
},
{
"epoch": 0.0006728767584162163,
"grad_norm": 0.39011430740356445,
"learning_rate": 0.00019996841892833,
"loss": 0.7847,
"step": 8
},
{
"epoch": 0.0007569863532182434,
"grad_norm": 0.49417048692703247,
"learning_rate": 0.00019996003076502565,
"loss": 1.0241,
"step": 9
},
{
"epoch": 0.0008410959480202704,
"grad_norm": 0.5206520557403564,
"learning_rate": 0.00019995065603657316,
"loss": 0.8424,
"step": 10
},
{
"epoch": 0.0009252055428222974,
"grad_norm": 0.4732626974582672,
"learning_rate": 0.0001999402948354973,
"loss": 0.7602,
"step": 11
},
{
"epoch": 0.0010093151376243245,
"grad_norm": 0.7119062542915344,
"learning_rate": 0.00019992894726405893,
"loss": 0.9964,
"step": 12
},
{
"epoch": 0.0010934247324263515,
"grad_norm": 0.6489611864089966,
"learning_rate": 0.000199916613434254,
"loss": 0.8901,
"step": 13
},
{
"epoch": 0.0011775343272283786,
"grad_norm": 0.6794676184654236,
"learning_rate": 0.0001999032934678125,
"loss": 0.8788,
"step": 14
},
{
"epoch": 0.0012616439220304056,
"grad_norm": 0.5384562015533447,
"learning_rate": 0.00019988898749619702,
"loss": 0.8936,
"step": 15
},
{
"epoch": 0.0013457535168324326,
"grad_norm": 0.6276252865791321,
"learning_rate": 0.00019987369566060176,
"loss": 0.7279,
"step": 16
},
{
"epoch": 0.0014298631116344597,
"grad_norm": 0.8388938903808594,
"learning_rate": 0.00019985741811195097,
"loss": 0.9929,
"step": 17
},
{
"epoch": 0.0015139727064364867,
"grad_norm": 0.5140799283981323,
"learning_rate": 0.00019984015501089752,
"loss": 0.6327,
"step": 18
},
{
"epoch": 0.0015980823012385138,
"grad_norm": 0.6707816123962402,
"learning_rate": 0.0001998219065278212,
"loss": 0.8234,
"step": 19
},
{
"epoch": 0.0016821918960405408,
"grad_norm": 0.721817672252655,
"learning_rate": 0.00019980267284282717,
"loss": 0.8038,
"step": 20
},
{
"epoch": 0.0017663014908425678,
"grad_norm": 0.6394887566566467,
"learning_rate": 0.00019978245414574417,
"loss": 0.7706,
"step": 21
},
{
"epoch": 0.0018504110856445949,
"grad_norm": 0.523406982421875,
"learning_rate": 0.00019976125063612252,
"loss": 0.6581,
"step": 22
},
{
"epoch": 0.001934520680446622,
"grad_norm": 0.5513215065002441,
"learning_rate": 0.00019973906252323238,
"loss": 0.9306,
"step": 23
},
{
"epoch": 0.002018630275248649,
"grad_norm": 0.5572922825813293,
"learning_rate": 0.0001997158900260614,
"loss": 0.7959,
"step": 24
},
{
"epoch": 0.002102739870050676,
"grad_norm": 0.567721426486969,
"learning_rate": 0.0001996917333733128,
"loss": 0.684,
"step": 25
},
{
"epoch": 0.002186849464852703,
"grad_norm": 0.8525415062904358,
"learning_rate": 0.00019966659280340297,
"loss": 0.6086,
"step": 26
},
{
"epoch": 0.0022709590596547303,
"grad_norm": 0.6331901550292969,
"learning_rate": 0.00019964046856445924,
"loss": 0.8775,
"step": 27
},
{
"epoch": 0.002355068654456757,
"grad_norm": 0.5875001549720764,
"learning_rate": 0.00019961336091431727,
"loss": 0.7884,
"step": 28
},
{
"epoch": 0.0024391782492587844,
"grad_norm": 0.4703105092048645,
"learning_rate": 0.00019958527012051857,
"loss": 0.8803,
"step": 29
},
{
"epoch": 0.002523287844060811,
"grad_norm": 0.46804890036582947,
"learning_rate": 0.00019955619646030802,
"loss": 0.7249,
"step": 30
},
{
"epoch": 0.0026073974388628384,
"grad_norm": 0.7693579792976379,
"learning_rate": 0.00019952614022063084,
"loss": 1.0132,
"step": 31
},
{
"epoch": 0.0026915070336648653,
"grad_norm": 0.7424771785736084,
"learning_rate": 0.00019949510169813003,
"loss": 0.9041,
"step": 32
},
{
"epoch": 0.0027756166284668925,
"grad_norm": 0.6431573033332825,
"learning_rate": 0.00019946308119914323,
"loss": 0.8694,
"step": 33
},
{
"epoch": 0.0028597262232689193,
"grad_norm": 0.5848168730735779,
"learning_rate": 0.0001994300790396999,
"loss": 0.7372,
"step": 34
},
{
"epoch": 0.0029438358180709466,
"grad_norm": 0.5710355043411255,
"learning_rate": 0.000199396095545518,
"loss": 0.7431,
"step": 35
},
{
"epoch": 0.0030279454128729734,
"grad_norm": 0.531509518623352,
"learning_rate": 0.00019936113105200085,
"loss": 0.7768,
"step": 36
},
{
"epoch": 0.0031120550076750007,
"grad_norm": 0.4804302752017975,
"learning_rate": 0.00019932518590423394,
"loss": 0.7301,
"step": 37
},
{
"epoch": 0.0031961646024770275,
"grad_norm": 0.5045036673545837,
"learning_rate": 0.00019928826045698136,
"loss": 0.6649,
"step": 38
},
{
"epoch": 0.0032802741972790548,
"grad_norm": 0.41377341747283936,
"learning_rate": 0.0001992503550746824,
"loss": 0.8951,
"step": 39
},
{
"epoch": 0.0033643837920810816,
"grad_norm": 0.534492552280426,
"learning_rate": 0.0001992114701314478,
"loss": 0.6731,
"step": 40
},
{
"epoch": 0.003448493386883109,
"grad_norm": 0.36469462513923645,
"learning_rate": 0.0001991716060110563,
"loss": 0.6212,
"step": 41
},
{
"epoch": 0.0035326029816851357,
"grad_norm": 0.44110405445098877,
"learning_rate": 0.00019913076310695068,
"loss": 0.8254,
"step": 42
},
{
"epoch": 0.003616712576487163,
"grad_norm": 0.4397999942302704,
"learning_rate": 0.00019908894182223388,
"loss": 0.7209,
"step": 43
},
{
"epoch": 0.0037008221712891897,
"grad_norm": 0.5739462971687317,
"learning_rate": 0.00019904614256966512,
"loss": 0.8095,
"step": 44
},
{
"epoch": 0.003784931766091217,
"grad_norm": 0.5561273097991943,
"learning_rate": 0.00019900236577165576,
"loss": 0.7944,
"step": 45
},
{
"epoch": 0.003869041360893244,
"grad_norm": 0.5051570534706116,
"learning_rate": 0.0001989576118602651,
"loss": 0.8223,
"step": 46
},
{
"epoch": 0.003953150955695271,
"grad_norm": 0.4156644642353058,
"learning_rate": 0.00019891188127719618,
"loss": 0.7488,
"step": 47
},
{
"epoch": 0.004037260550497298,
"grad_norm": 0.520811140537262,
"learning_rate": 0.0001988651744737914,
"loss": 0.8407,
"step": 48
},
{
"epoch": 0.004121370145299325,
"grad_norm": 0.7556874752044678,
"learning_rate": 0.00019881749191102808,
"loss": 0.7462,
"step": 49
},
{
"epoch": 0.004205479740101352,
"grad_norm": 0.692021369934082,
"learning_rate": 0.00019876883405951377,
"loss": 0.9717,
"step": 50
},
{
"epoch": 0.004289589334903379,
"grad_norm": 0.7206075191497803,
"learning_rate": 0.00019871920139948192,
"loss": 0.7717,
"step": 51
},
{
"epoch": 0.004373698929705406,
"grad_norm": 0.5266078114509583,
"learning_rate": 0.0001986685944207868,
"loss": 0.843,
"step": 52
},
{
"epoch": 0.004457808524507433,
"grad_norm": 0.4582868218421936,
"learning_rate": 0.0001986170136228989,
"loss": 0.709,
"step": 53
},
{
"epoch": 0.004541918119309461,
"grad_norm": 0.5274102687835693,
"learning_rate": 0.00019856445951489982,
"loss": 0.6895,
"step": 54
},
{
"epoch": 0.004626027714111487,
"grad_norm": 0.5545452237129211,
"learning_rate": 0.0001985109326154774,
"loss": 0.7461,
"step": 55
},
{
"epoch": 0.004710137308913514,
"grad_norm": 0.6029804944992065,
"learning_rate": 0.00019845643345292054,
"loss": 0.7165,
"step": 56
},
{
"epoch": 0.004794246903715541,
"grad_norm": 0.5293462872505188,
"learning_rate": 0.00019840096256511398,
"loss": 0.6772,
"step": 57
},
{
"epoch": 0.004878356498517569,
"grad_norm": 0.5571961402893066,
"learning_rate": 0.00019834452049953297,
"loss": 0.5852,
"step": 58
},
{
"epoch": 0.0049624660933195956,
"grad_norm": 0.5451956391334534,
"learning_rate": 0.00019828710781323792,
"loss": 0.7897,
"step": 59
},
{
"epoch": 0.005046575688121622,
"grad_norm": 0.6682764887809753,
"learning_rate": 0.0001982287250728689,
"loss": 0.8288,
"step": 60
},
{
"epoch": 0.005130685282923649,
"grad_norm": 0.6017076373100281,
"learning_rate": 0.0001981693728546399,
"loss": 0.8949,
"step": 61
},
{
"epoch": 0.005214794877725677,
"grad_norm": 0.5544684529304504,
"learning_rate": 0.0001981090517443334,
"loss": 0.6256,
"step": 62
},
{
"epoch": 0.005298904472527704,
"grad_norm": 0.5648895502090454,
"learning_rate": 0.00019804776233729444,
"loss": 0.6528,
"step": 63
},
{
"epoch": 0.0053830140673297305,
"grad_norm": 0.4875372350215912,
"learning_rate": 0.0001979855052384247,
"loss": 0.7416,
"step": 64
},
{
"epoch": 0.005467123662131757,
"grad_norm": 0.649088978767395,
"learning_rate": 0.00019792228106217658,
"loss": 0.7148,
"step": 65
},
{
"epoch": 0.005551233256933785,
"grad_norm": 0.39869123697280884,
"learning_rate": 0.00019785809043254722,
"loss": 0.5991,
"step": 66
},
{
"epoch": 0.005635342851735812,
"grad_norm": 0.6871259212493896,
"learning_rate": 0.0001977929339830722,
"loss": 0.6648,
"step": 67
},
{
"epoch": 0.005719452446537839,
"grad_norm": 0.5564553737640381,
"learning_rate": 0.00019772681235681936,
"loss": 0.842,
"step": 68
},
{
"epoch": 0.0058035620413398655,
"grad_norm": 0.6598843336105347,
"learning_rate": 0.00019765972620638248,
"loss": 1.0846,
"step": 69
},
{
"epoch": 0.005887671636141893,
"grad_norm": 0.5316952466964722,
"learning_rate": 0.00019759167619387476,
"loss": 0.8352,
"step": 70
},
{
"epoch": 0.00597178123094392,
"grad_norm": 0.923975944519043,
"learning_rate": 0.00019752266299092236,
"loss": 0.761,
"step": 71
},
{
"epoch": 0.006055890825745947,
"grad_norm": 0.6253910064697266,
"learning_rate": 0.00019745268727865774,
"loss": 0.8125,
"step": 72
},
{
"epoch": 0.006140000420547974,
"grad_norm": 0.5049079060554504,
"learning_rate": 0.0001973817497477129,
"loss": 0.7948,
"step": 73
},
{
"epoch": 0.006224110015350001,
"grad_norm": 0.5132893919944763,
"learning_rate": 0.00019730985109821266,
"loss": 0.7674,
"step": 74
},
{
"epoch": 0.006308219610152028,
"grad_norm": 0.5373795032501221,
"learning_rate": 0.00019723699203976766,
"loss": 0.9345,
"step": 75
},
{
"epoch": 0.006392329204954055,
"grad_norm": 0.658785343170166,
"learning_rate": 0.0001971631732914674,
"loss": 0.7366,
"step": 76
},
{
"epoch": 0.006476438799756082,
"grad_norm": 0.42672979831695557,
"learning_rate": 0.0001970883955818731,
"loss": 0.8569,
"step": 77
},
{
"epoch": 0.0065605483945581095,
"grad_norm": 0.6121531128883362,
"learning_rate": 0.0001970126596490106,
"loss": 0.7555,
"step": 78
},
{
"epoch": 0.006644657989360136,
"grad_norm": 0.5736799836158752,
"learning_rate": 0.00019693596624036292,
"loss": 0.7651,
"step": 79
},
{
"epoch": 0.006728767584162163,
"grad_norm": 0.5765815377235413,
"learning_rate": 0.0001968583161128631,
"loss": 0.7362,
"step": 80
},
{
"epoch": 0.00681287717896419,
"grad_norm": 0.5843616127967834,
"learning_rate": 0.00019677971003288655,
"loss": 0.8952,
"step": 81
},
{
"epoch": 0.006896986773766218,
"grad_norm": 0.5047760009765625,
"learning_rate": 0.00019670014877624353,
"loss": 0.6147,
"step": 82
},
{
"epoch": 0.0069810963685682445,
"grad_norm": 0.5199994444847107,
"learning_rate": 0.00019661963312817148,
"loss": 0.8876,
"step": 83
},
{
"epoch": 0.007065205963370271,
"grad_norm": 0.49524548649787903,
"learning_rate": 0.0001965381638833274,
"loss": 0.7281,
"step": 84
},
{
"epoch": 0.007149315558172298,
"grad_norm": 0.4999952018260956,
"learning_rate": 0.00019645574184577982,
"loss": 0.7967,
"step": 85
},
{
"epoch": 0.007233425152974326,
"grad_norm": 0.5383725762367249,
"learning_rate": 0.000196372367829001,
"loss": 0.7426,
"step": 86
},
{
"epoch": 0.007317534747776353,
"grad_norm": 0.4904189705848694,
"learning_rate": 0.00019628804265585877,
"loss": 0.8622,
"step": 87
},
{
"epoch": 0.0074016443425783795,
"grad_norm": 0.5760535001754761,
"learning_rate": 0.0001962027671586086,
"loss": 0.7718,
"step": 88
},
{
"epoch": 0.007485753937380406,
"grad_norm": 0.5867496728897095,
"learning_rate": 0.0001961165421788852,
"loss": 0.7467,
"step": 89
},
{
"epoch": 0.007569863532182434,
"grad_norm": 0.5310566425323486,
"learning_rate": 0.0001960293685676943,
"loss": 0.7393,
"step": 90
},
{
"epoch": 0.007653973126984461,
"grad_norm": 0.4217927157878876,
"learning_rate": 0.0001959412471854043,
"loss": 0.696,
"step": 91
},
{
"epoch": 0.007738082721786488,
"grad_norm": 0.6704983711242676,
"learning_rate": 0.0001958521789017376,
"loss": 0.8934,
"step": 92
},
{
"epoch": 0.007822192316588514,
"grad_norm": 0.543559193611145,
"learning_rate": 0.00019576216459576222,
"loss": 0.547,
"step": 93
},
{
"epoch": 0.007906301911390542,
"grad_norm": 0.47226864099502563,
"learning_rate": 0.00019567120515588308,
"loss": 0.6723,
"step": 94
},
{
"epoch": 0.007990411506192568,
"grad_norm": 0.41364991664886475,
"learning_rate": 0.00019557930147983302,
"loss": 0.6751,
"step": 95
},
{
"epoch": 0.008074521100994596,
"grad_norm": 0.595898449420929,
"learning_rate": 0.00019548645447466431,
"loss": 0.8247,
"step": 96
},
{
"epoch": 0.008158630695796623,
"grad_norm": 0.6959226131439209,
"learning_rate": 0.00019539266505673938,
"loss": 0.6472,
"step": 97
},
{
"epoch": 0.00824274029059865,
"grad_norm": 0.4702865779399872,
"learning_rate": 0.00019529793415172192,
"loss": 0.6569,
"step": 98
},
{
"epoch": 0.008326849885400677,
"grad_norm": 0.5658578872680664,
"learning_rate": 0.00019520226269456768,
"loss": 0.8219,
"step": 99
},
{
"epoch": 0.008410959480202705,
"grad_norm": 0.5993213057518005,
"learning_rate": 0.00019510565162951537,
"loss": 0.7014,
"step": 100
},
{
"epoch": 0.00849506907500473,
"grad_norm": 0.5463736057281494,
"learning_rate": 0.00019500810191007718,
"loss": 0.657,
"step": 101
},
{
"epoch": 0.008579178669806758,
"grad_norm": 0.5861630439758301,
"learning_rate": 0.00019490961449902946,
"loss": 0.7181,
"step": 102
},
{
"epoch": 0.008663288264608784,
"grad_norm": 0.5364252328872681,
"learning_rate": 0.0001948101903684032,
"loss": 0.7727,
"step": 103
},
{
"epoch": 0.008747397859410812,
"grad_norm": 0.6941062808036804,
"learning_rate": 0.00019470983049947444,
"loss": 0.7413,
"step": 104
},
{
"epoch": 0.00883150745421284,
"grad_norm": 0.7105836272239685,
"learning_rate": 0.00019460853588275454,
"loss": 0.8972,
"step": 105
},
{
"epoch": 0.008915617049014866,
"grad_norm": 0.4372139871120453,
"learning_rate": 0.00019450630751798048,
"loss": 0.5836,
"step": 106
},
{
"epoch": 0.008999726643816893,
"grad_norm": 0.6489446759223938,
"learning_rate": 0.000194403146414105,
"loss": 0.877,
"step": 107
},
{
"epoch": 0.009083836238618921,
"grad_norm": 0.5020645260810852,
"learning_rate": 0.00019429905358928646,
"loss": 0.6035,
"step": 108
},
{
"epoch": 0.009167945833420947,
"grad_norm": 0.7856727242469788,
"learning_rate": 0.00019419403007087907,
"loss": 0.715,
"step": 109
},
{
"epoch": 0.009252055428222975,
"grad_norm": 0.6128354072570801,
"learning_rate": 0.00019408807689542257,
"loss": 0.8376,
"step": 110
},
{
"epoch": 0.009336165023025,
"grad_norm": 0.5398645401000977,
"learning_rate": 0.00019398119510863197,
"loss": 0.7696,
"step": 111
},
{
"epoch": 0.009420274617827028,
"grad_norm": 0.575010359287262,
"learning_rate": 0.00019387338576538744,
"loss": 0.6847,
"step": 112
},
{
"epoch": 0.009504384212629056,
"grad_norm": 0.5950291156768799,
"learning_rate": 0.00019376464992972356,
"loss": 0.8757,
"step": 113
},
{
"epoch": 0.009588493807431082,
"grad_norm": 0.5944623947143555,
"learning_rate": 0.00019365498867481923,
"loss": 0.7609,
"step": 114
},
{
"epoch": 0.00967260340223311,
"grad_norm": 0.4505338668823242,
"learning_rate": 0.00019354440308298675,
"loss": 0.8388,
"step": 115
},
{
"epoch": 0.009756712997035137,
"grad_norm": 0.4795793294906616,
"learning_rate": 0.00019343289424566122,
"loss": 0.8994,
"step": 116
},
{
"epoch": 0.009840822591837163,
"grad_norm": 0.4607990086078644,
"learning_rate": 0.00019332046326338986,
"loss": 0.6956,
"step": 117
},
{
"epoch": 0.009924932186639191,
"grad_norm": 0.5157178640365601,
"learning_rate": 0.0001932071112458211,
"loss": 0.5357,
"step": 118
},
{
"epoch": 0.010009041781441217,
"grad_norm": 0.5431911945343018,
"learning_rate": 0.00019309283931169356,
"loss": 0.7704,
"step": 119
},
{
"epoch": 0.010093151376243245,
"grad_norm": 0.5151223540306091,
"learning_rate": 0.00019297764858882514,
"loss": 0.6825,
"step": 120
},
{
"epoch": 0.010177260971045272,
"grad_norm": 0.5834304094314575,
"learning_rate": 0.00019286154021410173,
"loss": 0.5308,
"step": 121
},
{
"epoch": 0.010261370565847298,
"grad_norm": 0.6810638904571533,
"learning_rate": 0.00019274451533346615,
"loss": 0.7112,
"step": 122
},
{
"epoch": 0.010345480160649326,
"grad_norm": 0.48631688952445984,
"learning_rate": 0.00019262657510190666,
"loss": 0.7143,
"step": 123
},
{
"epoch": 0.010429589755451354,
"grad_norm": 0.6056327819824219,
"learning_rate": 0.0001925077206834458,
"loss": 0.6615,
"step": 124
},
{
"epoch": 0.01051369935025338,
"grad_norm": 0.4841483235359192,
"learning_rate": 0.0001923879532511287,
"loss": 0.7337,
"step": 125
},
{
"epoch": 0.010597808945055407,
"grad_norm": 0.5701360702514648,
"learning_rate": 0.0001922672739870115,
"loss": 0.5952,
"step": 126
},
{
"epoch": 0.010681918539857433,
"grad_norm": 0.6007951498031616,
"learning_rate": 0.00019214568408214985,
"loss": 0.7754,
"step": 127
},
{
"epoch": 0.010766028134659461,
"grad_norm": 0.6508734822273254,
"learning_rate": 0.00019202318473658705,
"loss": 0.9641,
"step": 128
},
{
"epoch": 0.010850137729461489,
"grad_norm": 0.608518123626709,
"learning_rate": 0.00019189977715934213,
"loss": 0.8748,
"step": 129
},
{
"epoch": 0.010934247324263515,
"grad_norm": 0.43734872341156006,
"learning_rate": 0.00019177546256839812,
"loss": 0.742,
"step": 130
},
{
"epoch": 0.011018356919065542,
"grad_norm": 0.5520758628845215,
"learning_rate": 0.0001916502421906898,
"loss": 0.5509,
"step": 131
},
{
"epoch": 0.01110246651386757,
"grad_norm": 0.3889513909816742,
"learning_rate": 0.00019152411726209176,
"loss": 0.4892,
"step": 132
},
{
"epoch": 0.011186576108669596,
"grad_norm": 0.5818157196044922,
"learning_rate": 0.00019139708902740613,
"loss": 0.6986,
"step": 133
},
{
"epoch": 0.011270685703471624,
"grad_norm": 0.5319333076477051,
"learning_rate": 0.0001912691587403503,
"loss": 0.6992,
"step": 134
},
{
"epoch": 0.011354795298273651,
"grad_norm": 0.54970782995224,
"learning_rate": 0.00019114032766354453,
"loss": 0.7275,
"step": 135
},
{
"epoch": 0.011438904893075677,
"grad_norm": 0.4689725637435913,
"learning_rate": 0.00019101059706849957,
"loss": 0.7424,
"step": 136
},
{
"epoch": 0.011523014487877705,
"grad_norm": 0.625424861907959,
"learning_rate": 0.00019087996823560402,
"loss": 0.6456,
"step": 137
},
{
"epoch": 0.011607124082679731,
"grad_norm": 0.4590742290019989,
"learning_rate": 0.0001907484424541117,
"loss": 0.792,
"step": 138
},
{
"epoch": 0.011691233677481759,
"grad_norm": 0.48178377747535706,
"learning_rate": 0.00019061602102212898,
"loss": 0.8383,
"step": 139
},
{
"epoch": 0.011775343272283786,
"grad_norm": 0.49192577600479126,
"learning_rate": 0.00019048270524660196,
"loss": 0.7532,
"step": 140
},
{
"epoch": 0.011859452867085812,
"grad_norm": 0.5243375301361084,
"learning_rate": 0.0001903484964433035,
"loss": 0.8344,
"step": 141
},
{
"epoch": 0.01194356246188784,
"grad_norm": 0.4813317060470581,
"learning_rate": 0.00019021339593682028,
"loss": 0.9669,
"step": 142
},
{
"epoch": 0.012027672056689868,
"grad_norm": 0.49398431181907654,
"learning_rate": 0.00019007740506053983,
"loss": 0.6194,
"step": 143
},
{
"epoch": 0.012111781651491894,
"grad_norm": 0.5070907473564148,
"learning_rate": 0.0001899405251566371,
"loss": 0.4834,
"step": 144
},
{
"epoch": 0.012195891246293921,
"grad_norm": 0.5289168953895569,
"learning_rate": 0.00018980275757606157,
"loss": 0.6171,
"step": 145
},
{
"epoch": 0.012280000841095947,
"grad_norm": 0.6366864442825317,
"learning_rate": 0.00018966410367852362,
"loss": 0.8194,
"step": 146
},
{
"epoch": 0.012364110435897975,
"grad_norm": 0.7337288856506348,
"learning_rate": 0.00018952456483248119,
"loss": 0.8086,
"step": 147
},
{
"epoch": 0.012448220030700003,
"grad_norm": 0.683866560459137,
"learning_rate": 0.0001893841424151264,
"loss": 0.609,
"step": 148
},
{
"epoch": 0.012532329625502029,
"grad_norm": 0.5067821145057678,
"learning_rate": 0.0001892428378123718,
"loss": 0.8969,
"step": 149
},
{
"epoch": 0.012616439220304056,
"grad_norm": 0.48310670256614685,
"learning_rate": 0.0001891006524188368,
"loss": 0.6375,
"step": 150
},
{
"epoch": 0.012700548815106084,
"grad_norm": 0.5184181928634644,
"learning_rate": 0.00018895758763783383,
"loss": 0.7395,
"step": 151
},
{
"epoch": 0.01278465840990811,
"grad_norm": 0.4921519458293915,
"learning_rate": 0.00018881364488135448,
"loss": 0.8635,
"step": 152
},
{
"epoch": 0.012868768004710138,
"grad_norm": 0.42886897921562195,
"learning_rate": 0.00018866882557005567,
"loss": 0.6536,
"step": 153
},
{
"epoch": 0.012952877599512164,
"grad_norm": 0.5397732257843018,
"learning_rate": 0.00018852313113324552,
"loss": 0.698,
"step": 154
},
{
"epoch": 0.013036987194314191,
"grad_norm": 0.6344212889671326,
"learning_rate": 0.00018837656300886937,
"loss": 0.7285,
"step": 155
},
{
"epoch": 0.013121096789116219,
"grad_norm": 0.5104556679725647,
"learning_rate": 0.00018822912264349534,
"loss": 0.5943,
"step": 156
},
{
"epoch": 0.013205206383918245,
"grad_norm": 0.5612802505493164,
"learning_rate": 0.00018808081149230036,
"loss": 0.7107,
"step": 157
},
{
"epoch": 0.013289315978720273,
"grad_norm": 0.47731515765190125,
"learning_rate": 0.00018793163101905563,
"loss": 0.7419,
"step": 158
},
{
"epoch": 0.0133734255735223,
"grad_norm": 0.47317004203796387,
"learning_rate": 0.00018778158269611218,
"loss": 0.6444,
"step": 159
},
{
"epoch": 0.013457535168324326,
"grad_norm": 0.5630394816398621,
"learning_rate": 0.00018763066800438636,
"loss": 0.66,
"step": 160
},
{
"epoch": 0.013541644763126354,
"grad_norm": 0.4079178273677826,
"learning_rate": 0.0001874788884333453,
"loss": 0.6648,
"step": 161
},
{
"epoch": 0.01362575435792838,
"grad_norm": 0.4921400547027588,
"learning_rate": 0.00018732624548099204,
"loss": 0.6975,
"step": 162
},
{
"epoch": 0.013709863952730408,
"grad_norm": 0.5111605525016785,
"learning_rate": 0.0001871727406538509,
"loss": 0.7385,
"step": 163
},
{
"epoch": 0.013793973547532435,
"grad_norm": 0.6458659172058105,
"learning_rate": 0.0001870183754669526,
"loss": 0.7494,
"step": 164
},
{
"epoch": 0.013878083142334461,
"grad_norm": 0.5011044144630432,
"learning_rate": 0.00018686315144381913,
"loss": 0.5897,
"step": 165
},
{
"epoch": 0.013962192737136489,
"grad_norm": 0.524375319480896,
"learning_rate": 0.000186707070116449,
"loss": 0.721,
"step": 166
},
{
"epoch": 0.014046302331938517,
"grad_norm": 0.6293932795524597,
"learning_rate": 0.0001865501330253019,
"loss": 0.8962,
"step": 167
},
{
"epoch": 0.014130411926740543,
"grad_norm": 0.6843417882919312,
"learning_rate": 0.00018639234171928353,
"loss": 0.9065,
"step": 168
},
{
"epoch": 0.01421452152154257,
"grad_norm": 0.5260113477706909,
"learning_rate": 0.0001862336977557304,
"loss": 0.6079,
"step": 169
},
{
"epoch": 0.014298631116344596,
"grad_norm": 0.4797818064689636,
"learning_rate": 0.0001860742027003944,
"loss": 0.839,
"step": 170
},
{
"epoch": 0.014382740711146624,
"grad_norm": 0.5858710408210754,
"learning_rate": 0.00018591385812742725,
"loss": 0.7166,
"step": 171
},
{
"epoch": 0.014466850305948652,
"grad_norm": 0.47610747814178467,
"learning_rate": 0.00018575266561936523,
"loss": 0.5801,
"step": 172
},
{
"epoch": 0.014550959900750678,
"grad_norm": 0.7322710156440735,
"learning_rate": 0.00018559062676711332,
"loss": 0.7788,
"step": 173
},
{
"epoch": 0.014635069495552705,
"grad_norm": 0.5088084936141968,
"learning_rate": 0.0001854277431699295,
"loss": 0.6265,
"step": 174
},
{
"epoch": 0.014719179090354733,
"grad_norm": 0.5596153736114502,
"learning_rate": 0.00018526401643540922,
"loss": 0.7236,
"step": 175
},
{
"epoch": 0.014803288685156759,
"grad_norm": 0.7297357320785522,
"learning_rate": 0.00018509944817946922,
"loss": 0.6347,
"step": 176
},
{
"epoch": 0.014887398279958787,
"grad_norm": 0.45782700181007385,
"learning_rate": 0.00018493404002633166,
"loss": 0.6243,
"step": 177
},
{
"epoch": 0.014971507874760813,
"grad_norm": 0.46937909722328186,
"learning_rate": 0.00018476779360850832,
"loss": 0.7372,
"step": 178
},
{
"epoch": 0.01505561746956284,
"grad_norm": 0.4801153540611267,
"learning_rate": 0.00018460071056678422,
"loss": 0.6242,
"step": 179
},
{
"epoch": 0.015139727064364868,
"grad_norm": 0.42098918557167053,
"learning_rate": 0.00018443279255020152,
"loss": 0.823,
"step": 180
},
{
"epoch": 0.015223836659166894,
"grad_norm": 0.5428348183631897,
"learning_rate": 0.00018426404121604323,
"loss": 0.9499,
"step": 181
},
{
"epoch": 0.015307946253968922,
"grad_norm": 0.5740225315093994,
"learning_rate": 0.00018409445822981693,
"loss": 0.8347,
"step": 182
},
{
"epoch": 0.01539205584877095,
"grad_norm": 0.5669785141944885,
"learning_rate": 0.00018392404526523817,
"loss": 0.7831,
"step": 183
},
{
"epoch": 0.015476165443572975,
"grad_norm": 0.5455140471458435,
"learning_rate": 0.0001837528040042142,
"loss": 0.6875,
"step": 184
},
{
"epoch": 0.015560275038375003,
"grad_norm": 0.5965580344200134,
"learning_rate": 0.00018358073613682706,
"loss": 0.8344,
"step": 185
},
{
"epoch": 0.01564438463317703,
"grad_norm": 0.4952075779438019,
"learning_rate": 0.00018340784336131713,
"loss": 0.4143,
"step": 186
},
{
"epoch": 0.01572849422797906,
"grad_norm": 0.584922730922699,
"learning_rate": 0.00018323412738406635,
"loss": 0.6031,
"step": 187
},
{
"epoch": 0.015812603822781084,
"grad_norm": 0.5585885047912598,
"learning_rate": 0.00018305958991958127,
"loss": 0.7149,
"step": 188
},
{
"epoch": 0.01589671341758311,
"grad_norm": 0.6081206202507019,
"learning_rate": 0.0001828842326904762,
"loss": 0.7388,
"step": 189
},
{
"epoch": 0.015980823012385136,
"grad_norm": 0.5838598608970642,
"learning_rate": 0.00018270805742745617,
"loss": 0.7166,
"step": 190
},
{
"epoch": 0.016064932607187166,
"grad_norm": 0.5591777563095093,
"learning_rate": 0.00018253106586929997,
"loss": 0.6833,
"step": 191
},
{
"epoch": 0.01614904220198919,
"grad_norm": 0.49327778816223145,
"learning_rate": 0.00018235325976284275,
"loss": 0.6518,
"step": 192
},
{
"epoch": 0.016233151796791218,
"grad_norm": 0.5115340948104858,
"learning_rate": 0.00018217464086295904,
"loss": 0.9741,
"step": 193
},
{
"epoch": 0.016317261391593247,
"grad_norm": 0.44589102268218994,
"learning_rate": 0.00018199521093254523,
"loss": 0.7523,
"step": 194
},
{
"epoch": 0.016401370986395273,
"grad_norm": 0.4889369606971741,
"learning_rate": 0.00018181497174250236,
"loss": 0.5771,
"step": 195
},
{
"epoch": 0.0164854805811973,
"grad_norm": 0.5447149276733398,
"learning_rate": 0.00018163392507171842,
"loss": 0.7554,
"step": 196
},
{
"epoch": 0.01656959017599933,
"grad_norm": 0.4478766918182373,
"learning_rate": 0.00018145207270705096,
"loss": 0.7293,
"step": 197
},
{
"epoch": 0.016653699770801354,
"grad_norm": 0.4789084494113922,
"learning_rate": 0.0001812694164433094,
"loss": 0.7598,
"step": 198
},
{
"epoch": 0.01673780936560338,
"grad_norm": 0.5047253966331482,
"learning_rate": 0.00018108595808323736,
"loss": 0.703,
"step": 199
},
{
"epoch": 0.01682191896040541,
"grad_norm": 0.44823339581489563,
"learning_rate": 0.00018090169943749476,
"loss": 0.71,
"step": 200
},
{
"epoch": 0.016906028555207436,
"grad_norm": 0.5810935497283936,
"learning_rate": 0.00018071664232464002,
"loss": 0.7506,
"step": 201
},
{
"epoch": 0.01699013815000946,
"grad_norm": 0.44463852047920227,
"learning_rate": 0.0001805307885711122,
"loss": 0.7075,
"step": 202
},
{
"epoch": 0.01707424774481149,
"grad_norm": 0.5899863839149475,
"learning_rate": 0.00018034414001121278,
"loss": 0.7882,
"step": 203
},
{
"epoch": 0.017158357339613517,
"grad_norm": 0.5106207132339478,
"learning_rate": 0.00018015669848708767,
"loss": 0.8387,
"step": 204
},
{
"epoch": 0.017242466934415543,
"grad_norm": 0.5090427994728088,
"learning_rate": 0.00017996846584870908,
"loss": 0.8835,
"step": 205
},
{
"epoch": 0.01732657652921757,
"grad_norm": 0.5673806071281433,
"learning_rate": 0.0001797794439538571,
"loss": 0.6051,
"step": 206
},
{
"epoch": 0.0174106861240196,
"grad_norm": 0.4103122651576996,
"learning_rate": 0.0001795896346681016,
"loss": 0.6471,
"step": 207
},
{
"epoch": 0.017494795718821624,
"grad_norm": 0.5361846089363098,
"learning_rate": 0.00017939903986478355,
"loss": 0.6601,
"step": 208
},
{
"epoch": 0.01757890531362365,
"grad_norm": 0.6034552454948425,
"learning_rate": 0.00017920766142499672,
"loss": 0.7289,
"step": 209
},
{
"epoch": 0.01766301490842568,
"grad_norm": 0.5352544784545898,
"learning_rate": 0.00017901550123756906,
"loss": 0.6441,
"step": 210
},
{
"epoch": 0.017747124503227706,
"grad_norm": 0.48942509293556213,
"learning_rate": 0.00017882256119904403,
"loss": 0.7517,
"step": 211
},
{
"epoch": 0.01783123409802973,
"grad_norm": 0.6330734491348267,
"learning_rate": 0.00017862884321366188,
"loss": 0.8312,
"step": 212
},
{
"epoch": 0.01791534369283176,
"grad_norm": 0.5588996410369873,
"learning_rate": 0.000178434349193341,
"loss": 0.8497,
"step": 213
},
{
"epoch": 0.017999453287633787,
"grad_norm": 0.4862910807132721,
"learning_rate": 0.0001782390810576588,
"loss": 0.7367,
"step": 214
},
{
"epoch": 0.018083562882435813,
"grad_norm": 0.45474931597709656,
"learning_rate": 0.000178043040733833,
"loss": 0.6998,
"step": 215
},
{
"epoch": 0.018167672477237842,
"grad_norm": 0.4923236072063446,
"learning_rate": 0.00017784623015670238,
"loss": 0.8464,
"step": 216
},
{
"epoch": 0.018251782072039868,
"grad_norm": 0.6574183702468872,
"learning_rate": 0.00017764865126870786,
"loss": 0.7606,
"step": 217
},
{
"epoch": 0.018335891666841894,
"grad_norm": 0.4556383192539215,
"learning_rate": 0.00017745030601987337,
"loss": 0.5689,
"step": 218
},
{
"epoch": 0.018420001261643924,
"grad_norm": 0.49504584074020386,
"learning_rate": 0.00017725119636778644,
"loss": 0.6945,
"step": 219
},
{
"epoch": 0.01850411085644595,
"grad_norm": 0.48922199010849,
"learning_rate": 0.00017705132427757895,
"loss": 0.6024,
"step": 220
},
{
"epoch": 0.018588220451247976,
"grad_norm": 0.4516255259513855,
"learning_rate": 0.00017685069172190766,
"loss": 0.6825,
"step": 221
},
{
"epoch": 0.01867233004605,
"grad_norm": 0.5373181104660034,
"learning_rate": 0.00017664930068093498,
"loss": 0.7165,
"step": 222
},
{
"epoch": 0.01875643964085203,
"grad_norm": 0.45531630516052246,
"learning_rate": 0.00017644715314230918,
"loss": 0.614,
"step": 223
},
{
"epoch": 0.018840549235654057,
"grad_norm": 0.38286691904067993,
"learning_rate": 0.0001762442511011448,
"loss": 0.6237,
"step": 224
},
{
"epoch": 0.018924658830456083,
"grad_norm": 0.8253074288368225,
"learning_rate": 0.0001760405965600031,
"loss": 0.7572,
"step": 225
},
{
"epoch": 0.019008768425258112,
"grad_norm": 0.49788135290145874,
"learning_rate": 0.0001758361915288722,
"loss": 0.836,
"step": 226
},
{
"epoch": 0.019092878020060138,
"grad_norm": 0.7003273367881775,
"learning_rate": 0.0001756310380251472,
"loss": 0.6137,
"step": 227
},
{
"epoch": 0.019176987614862164,
"grad_norm": 0.5549869537353516,
"learning_rate": 0.00017542513807361037,
"loss": 0.7504,
"step": 228
},
{
"epoch": 0.019261097209664194,
"grad_norm": 0.5484214425086975,
"learning_rate": 0.00017521849370641114,
"loss": 0.8757,
"step": 229
},
{
"epoch": 0.01934520680446622,
"grad_norm": 0.48561275005340576,
"learning_rate": 0.00017501110696304596,
"loss": 0.6577,
"step": 230
},
{
"epoch": 0.019429316399268245,
"grad_norm": 0.5252694487571716,
"learning_rate": 0.00017480297989033825,
"loss": 0.7415,
"step": 231
},
{
"epoch": 0.019513425994070275,
"grad_norm": 0.5625510811805725,
"learning_rate": 0.00017459411454241822,
"loss": 0.7563,
"step": 232
},
{
"epoch": 0.0195975355888723,
"grad_norm": 0.5950868725776672,
"learning_rate": 0.00017438451298070252,
"loss": 0.5738,
"step": 233
},
{
"epoch": 0.019681645183674327,
"grad_norm": 0.3883379399776459,
"learning_rate": 0.00017417417727387394,
"loss": 0.6791,
"step": 234
},
{
"epoch": 0.019765754778476356,
"grad_norm": 0.3784724175930023,
"learning_rate": 0.000173963109497861,
"loss": 0.6673,
"step": 235
},
{
"epoch": 0.019849864373278382,
"grad_norm": 0.5025336146354675,
"learning_rate": 0.0001737513117358174,
"loss": 0.678,
"step": 236
},
{
"epoch": 0.019933973968080408,
"grad_norm": 0.5078927874565125,
"learning_rate": 0.0001735387860781016,
"loss": 0.7093,
"step": 237
},
{
"epoch": 0.020018083562882434,
"grad_norm": 0.67873215675354,
"learning_rate": 0.00017332553462225602,
"loss": 0.7362,
"step": 238
},
{
"epoch": 0.020102193157684464,
"grad_norm": 0.6332508325576782,
"learning_rate": 0.00017311155947298643,
"loss": 0.848,
"step": 239
},
{
"epoch": 0.02018630275248649,
"grad_norm": 0.6372016668319702,
"learning_rate": 0.00017289686274214118,
"loss": 0.7129,
"step": 240
},
{
"epoch": 0.020270412347288515,
"grad_norm": 0.6516194939613342,
"learning_rate": 0.0001726814465486903,
"loss": 0.7255,
"step": 241
},
{
"epoch": 0.020354521942090545,
"grad_norm": 0.5692350268363953,
"learning_rate": 0.0001724653130187047,
"loss": 0.6044,
"step": 242
},
{
"epoch": 0.02043863153689257,
"grad_norm": 0.6217871308326721,
"learning_rate": 0.00017224846428533499,
"loss": 0.8463,
"step": 243
},
{
"epoch": 0.020522741131694597,
"grad_norm": 0.4982471466064453,
"learning_rate": 0.0001720309024887907,
"loss": 0.7304,
"step": 244
},
{
"epoch": 0.020606850726496626,
"grad_norm": 0.46791961789131165,
"learning_rate": 0.00017181262977631888,
"loss": 0.6765,
"step": 245
},
{
"epoch": 0.020690960321298652,
"grad_norm": 0.54993736743927,
"learning_rate": 0.00017159364830218312,
"loss": 0.5422,
"step": 246
},
{
"epoch": 0.020775069916100678,
"grad_norm": 0.557114839553833,
"learning_rate": 0.00017137396022764214,
"loss": 0.8051,
"step": 247
},
{
"epoch": 0.020859179510902708,
"grad_norm": 0.4503937065601349,
"learning_rate": 0.00017115356772092857,
"loss": 0.6993,
"step": 248
},
{
"epoch": 0.020943289105704734,
"grad_norm": 0.5274853706359863,
"learning_rate": 0.0001709324729572274,
"loss": 0.5102,
"step": 249
},
{
"epoch": 0.02102739870050676,
"grad_norm": 0.5487212538719177,
"learning_rate": 0.00017071067811865476,
"loss": 0.7262,
"step": 250
},
{
"epoch": 0.02111150829530879,
"grad_norm": 0.4630710482597351,
"learning_rate": 0.00017048818539423615,
"loss": 0.6907,
"step": 251
},
{
"epoch": 0.021195617890110815,
"grad_norm": 0.4759877622127533,
"learning_rate": 0.00017026499697988493,
"loss": 0.6945,
"step": 252
},
{
"epoch": 0.02127972748491284,
"grad_norm": 0.44283753633499146,
"learning_rate": 0.00017004111507838064,
"loss": 0.491,
"step": 253
},
{
"epoch": 0.021363837079714867,
"grad_norm": 0.6089275479316711,
"learning_rate": 0.00016981654189934727,
"loss": 0.7725,
"step": 254
},
{
"epoch": 0.021447946674516896,
"grad_norm": 0.48466405272483826,
"learning_rate": 0.00016959127965923142,
"loss": 0.6735,
"step": 255
},
{
"epoch": 0.021532056269318922,
"grad_norm": 0.5322965383529663,
"learning_rate": 0.0001693653305812805,
"loss": 0.7648,
"step": 256
},
{
"epoch": 0.021616165864120948,
"grad_norm": 0.5427155494689941,
"learning_rate": 0.00016913869689552064,
"loss": 0.6212,
"step": 257
},
{
"epoch": 0.021700275458922978,
"grad_norm": 0.6860820651054382,
"learning_rate": 0.00016891138083873487,
"loss": 0.9867,
"step": 258
},
{
"epoch": 0.021784385053725003,
"grad_norm": 0.5238876342773438,
"learning_rate": 0.00016868338465444085,
"loss": 0.7087,
"step": 259
},
{
"epoch": 0.02186849464852703,
"grad_norm": 0.613838255405426,
"learning_rate": 0.00016845471059286887,
"loss": 0.6237,
"step": 260
},
{
"epoch": 0.02195260424332906,
"grad_norm": 0.6442722082138062,
"learning_rate": 0.00016822536091093965,
"loss": 0.6983,
"step": 261
},
{
"epoch": 0.022036713838131085,
"grad_norm": 0.5362198352813721,
"learning_rate": 0.00016799533787224192,
"loss": 0.6856,
"step": 262
},
{
"epoch": 0.02212082343293311,
"grad_norm": 0.5040967464447021,
"learning_rate": 0.00016776464374701025,
"loss": 0.5694,
"step": 263
},
{
"epoch": 0.02220493302773514,
"grad_norm": 0.6717795133590698,
"learning_rate": 0.00016753328081210245,
"loss": 0.7468,
"step": 264
},
{
"epoch": 0.022289042622537166,
"grad_norm": 0.5034369230270386,
"learning_rate": 0.00016730125135097735,
"loss": 0.7802,
"step": 265
},
{
"epoch": 0.022373152217339192,
"grad_norm": 0.49784526228904724,
"learning_rate": 0.000167068557653672,
"loss": 0.6088,
"step": 266
},
{
"epoch": 0.02245726181214122,
"grad_norm": 0.6437605619430542,
"learning_rate": 0.0001668352020167793,
"loss": 0.6883,
"step": 267
},
{
"epoch": 0.022541371406943247,
"grad_norm": 0.7116613984107971,
"learning_rate": 0.00016660118674342517,
"loss": 0.6185,
"step": 268
},
{
"epoch": 0.022625481001745273,
"grad_norm": 0.5817685723304749,
"learning_rate": 0.00016636651414324587,
"loss": 0.6651,
"step": 269
},
{
"epoch": 0.022709590596547303,
"grad_norm": 0.37707141041755676,
"learning_rate": 0.00016613118653236518,
"loss": 0.6047,
"step": 270
},
{
"epoch": 0.02279370019134933,
"grad_norm": 0.4310030937194824,
"learning_rate": 0.0001658952062333717,
"loss": 0.7464,
"step": 271
},
{
"epoch": 0.022877809786151355,
"grad_norm": 0.5687549710273743,
"learning_rate": 0.00016565857557529566,
"loss": 0.7018,
"step": 272
},
{
"epoch": 0.02296191938095338,
"grad_norm": 0.4826352894306183,
"learning_rate": 0.00016542129689358612,
"loss": 0.6731,
"step": 273
},
{
"epoch": 0.02304602897575541,
"grad_norm": 0.58961021900177,
"learning_rate": 0.0001651833725300879,
"loss": 0.7081,
"step": 274
},
{
"epoch": 0.023130138570557436,
"grad_norm": 0.3250476121902466,
"learning_rate": 0.00016494480483301836,
"loss": 0.3532,
"step": 275
},
{
"epoch": 0.023214248165359462,
"grad_norm": 0.6899309754371643,
"learning_rate": 0.00016470559615694446,
"loss": 0.7164,
"step": 276
},
{
"epoch": 0.02329835776016149,
"grad_norm": 0.5115370154380798,
"learning_rate": 0.00016446574886275913,
"loss": 0.8014,
"step": 277
},
{
"epoch": 0.023382467354963517,
"grad_norm": 0.4905194640159607,
"learning_rate": 0.00016422526531765846,
"loss": 0.5492,
"step": 278
},
{
"epoch": 0.023466576949765543,
"grad_norm": 0.5156686305999756,
"learning_rate": 0.00016398414789511786,
"loss": 0.5411,
"step": 279
},
{
"epoch": 0.023550686544567573,
"grad_norm": 0.6288424730300903,
"learning_rate": 0.000163742398974869,
"loss": 0.7018,
"step": 280
},
{
"epoch": 0.0236347961393696,
"grad_norm": 0.5037186145782471,
"learning_rate": 0.00016350002094287609,
"loss": 0.7827,
"step": 281
},
{
"epoch": 0.023718905734171625,
"grad_norm": 0.4114580750465393,
"learning_rate": 0.00016325701619131246,
"loss": 0.707,
"step": 282
},
{
"epoch": 0.023803015328973654,
"grad_norm": 0.4033624529838562,
"learning_rate": 0.00016301338711853693,
"loss": 0.7176,
"step": 283
},
{
"epoch": 0.02388712492377568,
"grad_norm": 0.7627992630004883,
"learning_rate": 0.00016276913612907007,
"loss": 0.7968,
"step": 284
},
{
"epoch": 0.023971234518577706,
"grad_norm": 0.5841469168663025,
"learning_rate": 0.00016252426563357055,
"loss": 0.7753,
"step": 285
},
{
"epoch": 0.024055344113379735,
"grad_norm": 0.7086002826690674,
"learning_rate": 0.00016227877804881127,
"loss": 0.701,
"step": 286
},
{
"epoch": 0.02413945370818176,
"grad_norm": 0.6601248979568481,
"learning_rate": 0.00016203267579765563,
"loss": 0.709,
"step": 287
},
{
"epoch": 0.024223563302983787,
"grad_norm": 0.6370972990989685,
"learning_rate": 0.00016178596130903344,
"loss": 0.7909,
"step": 288
},
{
"epoch": 0.024307672897785813,
"grad_norm": 0.5954873561859131,
"learning_rate": 0.00016153863701791717,
"loss": 0.5179,
"step": 289
},
{
"epoch": 0.024391782492587843,
"grad_norm": 0.6457350850105286,
"learning_rate": 0.00016129070536529766,
"loss": 0.8393,
"step": 290
},
{
"epoch": 0.02447589208738987,
"grad_norm": 0.6270333528518677,
"learning_rate": 0.00016104216879816026,
"loss": 0.6952,
"step": 291
},
{
"epoch": 0.024560001682191895,
"grad_norm": 0.4968658983707428,
"learning_rate": 0.00016079302976946055,
"loss": 0.8266,
"step": 292
},
{
"epoch": 0.024644111276993924,
"grad_norm": 0.6642244458198547,
"learning_rate": 0.00016054329073810015,
"loss": 0.7603,
"step": 293
},
{
"epoch": 0.02472822087179595,
"grad_norm": 0.4561289846897125,
"learning_rate": 0.00016029295416890248,
"loss": 0.5671,
"step": 294
},
{
"epoch": 0.024812330466597976,
"grad_norm": 0.5380001664161682,
"learning_rate": 0.00016004202253258842,
"loss": 0.7924,
"step": 295
},
{
"epoch": 0.024896440061400005,
"grad_norm": 0.48755741119384766,
"learning_rate": 0.0001597904983057519,
"loss": 0.5793,
"step": 296
},
{
"epoch": 0.02498054965620203,
"grad_norm": 0.4303657114505768,
"learning_rate": 0.00015953838397083552,
"loss": 0.5611,
"step": 297
},
{
"epoch": 0.025064659251004057,
"grad_norm": 0.5174483060836792,
"learning_rate": 0.00015928568201610595,
"loss": 0.8403,
"step": 298
},
{
"epoch": 0.025148768845806087,
"grad_norm": 0.4314514398574829,
"learning_rate": 0.00015903239493562948,
"loss": 0.6122,
"step": 299
},
{
"epoch": 0.025232878440608113,
"grad_norm": 0.5762454867362976,
"learning_rate": 0.00015877852522924732,
"loss": 0.6049,
"step": 300
},
{
"epoch": 0.02531698803541014,
"grad_norm": 0.40056511759757996,
"learning_rate": 0.00015852407540255104,
"loss": 0.6657,
"step": 301
},
{
"epoch": 0.025401097630212168,
"grad_norm": 0.3158378303050995,
"learning_rate": 0.00015826904796685762,
"loss": 0.4039,
"step": 302
},
{
"epoch": 0.025485207225014194,
"grad_norm": 0.4131661653518677,
"learning_rate": 0.00015801344543918495,
"loss": 0.6172,
"step": 303
},
{
"epoch": 0.02556931681981622,
"grad_norm": 0.46788647770881653,
"learning_rate": 0.00015775727034222675,
"loss": 0.6595,
"step": 304
},
{
"epoch": 0.025653426414618246,
"grad_norm": 0.8410905003547668,
"learning_rate": 0.00015750052520432787,
"loss": 0.6009,
"step": 305
},
{
"epoch": 0.025737536009420275,
"grad_norm": 0.4922102987766266,
"learning_rate": 0.0001572432125594591,
"loss": 0.737,
"step": 306
},
{
"epoch": 0.0258216456042223,
"grad_norm": 0.48740994930267334,
"learning_rate": 0.00015698533494719238,
"loss": 0.6165,
"step": 307
},
{
"epoch": 0.025905755199024327,
"grad_norm": 0.6041740775108337,
"learning_rate": 0.00015672689491267567,
"loss": 0.5473,
"step": 308
},
{
"epoch": 0.025989864793826357,
"grad_norm": 0.5123183727264404,
"learning_rate": 0.00015646789500660773,
"loss": 0.7032,
"step": 309
},
{
"epoch": 0.026073974388628383,
"grad_norm": 0.4975537359714508,
"learning_rate": 0.00015620833778521307,
"loss": 0.7373,
"step": 310
},
{
"epoch": 0.02615808398343041,
"grad_norm": 0.41992077231407166,
"learning_rate": 0.0001559482258102167,
"loss": 0.7797,
"step": 311
},
{
"epoch": 0.026242193578232438,
"grad_norm": 0.5822920203208923,
"learning_rate": 0.00015568756164881882,
"loss": 0.6739,
"step": 312
},
{
"epoch": 0.026326303173034464,
"grad_norm": 0.5167087316513062,
"learning_rate": 0.00015542634787366942,
"loss": 0.6933,
"step": 313
},
{
"epoch": 0.02641041276783649,
"grad_norm": 0.5934262871742249,
"learning_rate": 0.00015516458706284303,
"loss": 0.5772,
"step": 314
},
{
"epoch": 0.02649452236263852,
"grad_norm": 0.43231528997421265,
"learning_rate": 0.0001549022817998132,
"loss": 0.7363,
"step": 315
},
{
"epoch": 0.026578631957440545,
"grad_norm": 0.6113074421882629,
"learning_rate": 0.00015463943467342693,
"loss": 0.8104,
"step": 316
},
{
"epoch": 0.02666274155224257,
"grad_norm": 0.6448655724525452,
"learning_rate": 0.00015437604827787927,
"loss": 0.7304,
"step": 317
},
{
"epoch": 0.0267468511470446,
"grad_norm": 0.630864143371582,
"learning_rate": 0.00015411212521268758,
"loss": 0.7209,
"step": 318
},
{
"epoch": 0.026830960741846627,
"grad_norm": 0.5093964338302612,
"learning_rate": 0.00015384766808266602,
"loss": 0.7396,
"step": 319
},
{
"epoch": 0.026915070336648653,
"grad_norm": 0.5301403999328613,
"learning_rate": 0.00015358267949789966,
"loss": 0.8367,
"step": 320
},
{
"epoch": 0.02699917993145068,
"grad_norm": 0.5352559089660645,
"learning_rate": 0.00015331716207371888,
"loss": 0.6919,
"step": 321
},
{
"epoch": 0.027083289526252708,
"grad_norm": 0.4436148703098297,
"learning_rate": 0.0001530511184306734,
"loss": 0.6293,
"step": 322
},
{
"epoch": 0.027167399121054734,
"grad_norm": 0.5910261273384094,
"learning_rate": 0.00015278455119450664,
"loss": 0.7454,
"step": 323
},
{
"epoch": 0.02725150871585676,
"grad_norm": 0.5505343079566956,
"learning_rate": 0.0001525174629961296,
"loss": 0.7661,
"step": 324
},
{
"epoch": 0.02733561831065879,
"grad_norm": 0.5973290205001831,
"learning_rate": 0.0001522498564715949,
"loss": 0.754,
"step": 325
},
{
"epoch": 0.027419727905460815,
"grad_norm": 0.5219733715057373,
"learning_rate": 0.00015198173426207094,
"loss": 0.9089,
"step": 326
},
{
"epoch": 0.02750383750026284,
"grad_norm": 0.48295462131500244,
"learning_rate": 0.00015171309901381572,
"loss": 0.7115,
"step": 327
},
{
"epoch": 0.02758794709506487,
"grad_norm": 0.3980119228363037,
"learning_rate": 0.00015144395337815064,
"loss": 0.6346,
"step": 328
},
{
"epoch": 0.027672056689866897,
"grad_norm": 0.44824841618537903,
"learning_rate": 0.00015117430001143452,
"loss": 0.7356,
"step": 329
},
{
"epoch": 0.027756166284668923,
"grad_norm": 0.4944741129875183,
"learning_rate": 0.00015090414157503714,
"loss": 0.7034,
"step": 330
},
{
"epoch": 0.027840275879470952,
"grad_norm": 0.4512578845024109,
"learning_rate": 0.00015063348073531324,
"loss": 0.7027,
"step": 331
},
{
"epoch": 0.027924385474272978,
"grad_norm": 0.5666627287864685,
"learning_rate": 0.0001503623201635761,
"loss": 0.6587,
"step": 332
},
{
"epoch": 0.028008495069075004,
"grad_norm": 0.4888990521430969,
"learning_rate": 0.000150090662536071,
"loss": 0.5327,
"step": 333
},
{
"epoch": 0.028092604663877033,
"grad_norm": 0.5111712217330933,
"learning_rate": 0.0001498185105339491,
"loss": 0.6042,
"step": 334
},
{
"epoch": 0.02817671425867906,
"grad_norm": 0.4700281620025635,
"learning_rate": 0.00014954586684324078,
"loss": 0.583,
"step": 335
},
{
"epoch": 0.028260823853481085,
"grad_norm": 0.5448441505432129,
"learning_rate": 0.00014927273415482915,
"loss": 0.6319,
"step": 336
},
{
"epoch": 0.02834493344828311,
"grad_norm": 0.6051144599914551,
"learning_rate": 0.00014899911516442365,
"loss": 0.5532,
"step": 337
},
{
"epoch": 0.02842904304308514,
"grad_norm": 0.6583679914474487,
"learning_rate": 0.00014872501257253323,
"loss": 0.7324,
"step": 338
},
{
"epoch": 0.028513152637887167,
"grad_norm": 0.468326210975647,
"learning_rate": 0.0001484504290844398,
"loss": 0.6747,
"step": 339
},
{
"epoch": 0.028597262232689193,
"grad_norm": 0.5111216306686401,
"learning_rate": 0.00014817536741017152,
"loss": 0.8554,
"step": 340
},
{
"epoch": 0.028681371827491222,
"grad_norm": 0.42130234837532043,
"learning_rate": 0.00014789983026447612,
"loss": 0.5064,
"step": 341
},
{
"epoch": 0.028765481422293248,
"grad_norm": 0.6524533033370972,
"learning_rate": 0.0001476238203667939,
"loss": 0.7676,
"step": 342
},
{
"epoch": 0.028849591017095274,
"grad_norm": 0.5613992810249329,
"learning_rate": 0.0001473473404412312,
"loss": 0.563,
"step": 343
},
{
"epoch": 0.028933700611897303,
"grad_norm": 0.5069441795349121,
"learning_rate": 0.0001470703932165333,
"loss": 0.6787,
"step": 344
},
{
"epoch": 0.02901781020669933,
"grad_norm": 0.47367653250694275,
"learning_rate": 0.00014679298142605734,
"loss": 0.8746,
"step": 345
},
{
"epoch": 0.029101919801501355,
"grad_norm": 0.4816168546676636,
"learning_rate": 0.00014651510780774583,
"loss": 0.6131,
"step": 346
},
{
"epoch": 0.029186029396303385,
"grad_norm": 0.5362951159477234,
"learning_rate": 0.00014623677510409918,
"loss": 0.6675,
"step": 347
},
{
"epoch": 0.02927013899110541,
"grad_norm": 0.4960172772407532,
"learning_rate": 0.00014595798606214882,
"loss": 0.624,
"step": 348
},
{
"epoch": 0.029354248585907437,
"grad_norm": 0.3880994915962219,
"learning_rate": 0.00014567874343342997,
"loss": 0.4903,
"step": 349
},
{
"epoch": 0.029438358180709466,
"grad_norm": 0.5134441256523132,
"learning_rate": 0.00014539904997395468,
"loss": 0.7855,
"step": 350
},
{
"epoch": 0.029522467775511492,
"grad_norm": 0.5130655765533447,
"learning_rate": 0.00014511890844418453,
"loss": 0.723,
"step": 351
},
{
"epoch": 0.029606577370313518,
"grad_norm": 0.63141930103302,
"learning_rate": 0.00014483832160900326,
"loss": 0.8834,
"step": 352
},
{
"epoch": 0.029690686965115544,
"grad_norm": 0.5271434783935547,
"learning_rate": 0.00014455729223768966,
"loss": 0.5313,
"step": 353
},
{
"epoch": 0.029774796559917573,
"grad_norm": 0.44754379987716675,
"learning_rate": 0.0001442758231038902,
"loss": 0.6777,
"step": 354
},
{
"epoch": 0.0298589061547196,
"grad_norm": 0.5021183490753174,
"learning_rate": 0.00014399391698559152,
"loss": 0.476,
"step": 355
},
{
"epoch": 0.029943015749521625,
"grad_norm": 0.5433945059776306,
"learning_rate": 0.0001437115766650933,
"loss": 0.6773,
"step": 356
},
{
"epoch": 0.030027125344323655,
"grad_norm": 0.42916494607925415,
"learning_rate": 0.00014342880492898048,
"loss": 0.5919,
"step": 357
},
{
"epoch": 0.03011123493912568,
"grad_norm": 0.5809445381164551,
"learning_rate": 0.0001431456045680959,
"loss": 0.7717,
"step": 358
},
{
"epoch": 0.030195344533927707,
"grad_norm": 0.5388799905776978,
"learning_rate": 0.00014286197837751286,
"loss": 0.6828,
"step": 359
},
{
"epoch": 0.030279454128729736,
"grad_norm": 0.5621784329414368,
"learning_rate": 0.00014257792915650728,
"loss": 0.5758,
"step": 360
},
{
"epoch": 0.030363563723531762,
"grad_norm": 0.5903728604316711,
"learning_rate": 0.00014229345970853032,
"loss": 0.6692,
"step": 361
},
{
"epoch": 0.030447673318333788,
"grad_norm": 0.5247724652290344,
"learning_rate": 0.00014200857284118066,
"loss": 0.513,
"step": 362
},
{
"epoch": 0.030531782913135817,
"grad_norm": 0.6713283658027649,
"learning_rate": 0.00014172327136617656,
"loss": 0.6672,
"step": 363
},
{
"epoch": 0.030615892507937843,
"grad_norm": 0.4961901903152466,
"learning_rate": 0.00014143755809932845,
"loss": 0.5941,
"step": 364
},
{
"epoch": 0.03070000210273987,
"grad_norm": 0.5194029211997986,
"learning_rate": 0.00014115143586051088,
"loss": 0.7219,
"step": 365
},
{
"epoch": 0.0307841116975419,
"grad_norm": 0.5475531816482544,
"learning_rate": 0.00014086490747363493,
"loss": 0.7603,
"step": 366
},
{
"epoch": 0.030868221292343925,
"grad_norm": 0.6493870615959167,
"learning_rate": 0.00014057797576662,
"loss": 0.6935,
"step": 367
},
{
"epoch": 0.03095233088714595,
"grad_norm": 0.7089727520942688,
"learning_rate": 0.00014029064357136628,
"loss": 0.7187,
"step": 368
},
{
"epoch": 0.031036440481947977,
"grad_norm": 0.4623405337333679,
"learning_rate": 0.00014000291372372647,
"loss": 0.5853,
"step": 369
},
{
"epoch": 0.031120550076750006,
"grad_norm": 0.5741824507713318,
"learning_rate": 0.00013971478906347806,
"loss": 0.6912,
"step": 370
},
{
"epoch": 0.031204659671552032,
"grad_norm": 0.5672146081924438,
"learning_rate": 0.00013942627243429512,
"loss": 0.9029,
"step": 371
},
{
"epoch": 0.03128876926635406,
"grad_norm": 0.6760275959968567,
"learning_rate": 0.00013913736668372026,
"loss": 0.6691,
"step": 372
},
{
"epoch": 0.031372878861156084,
"grad_norm": 0.7026247978210449,
"learning_rate": 0.00013884807466313663,
"loss": 0.6567,
"step": 373
},
{
"epoch": 0.03145698845595812,
"grad_norm": 0.5599506497383118,
"learning_rate": 0.00013855839922773968,
"loss": 0.7411,
"step": 374
},
{
"epoch": 0.03154109805076014,
"grad_norm": 0.370534747838974,
"learning_rate": 0.000138268343236509,
"loss": 0.5276,
"step": 375
},
{
"epoch": 0.03162520764556217,
"grad_norm": 0.5061807036399841,
"learning_rate": 0.00013797790955218014,
"loss": 0.6831,
"step": 376
},
{
"epoch": 0.031709317240364195,
"grad_norm": 0.5783600807189941,
"learning_rate": 0.00013768710104121627,
"loss": 0.5448,
"step": 377
},
{
"epoch": 0.03179342683516622,
"grad_norm": 0.5063284039497375,
"learning_rate": 0.00013739592057378003,
"loss": 0.6389,
"step": 378
},
{
"epoch": 0.031877536429968246,
"grad_norm": 0.5093449950218201,
"learning_rate": 0.0001371043710237051,
"loss": 0.7035,
"step": 379
},
{
"epoch": 0.03196164602477027,
"grad_norm": 0.5934507250785828,
"learning_rate": 0.00013681245526846783,
"loss": 0.7793,
"step": 380
},
{
"epoch": 0.032045755619572305,
"grad_norm": 0.5645430088043213,
"learning_rate": 0.0001365201761891588,
"loss": 0.9189,
"step": 381
},
{
"epoch": 0.03212986521437433,
"grad_norm": 0.5648247599601746,
"learning_rate": 0.00013622753667045457,
"loss": 0.7319,
"step": 382
},
{
"epoch": 0.03221397480917636,
"grad_norm": 0.48350241780281067,
"learning_rate": 0.00013593453960058908,
"loss": 0.65,
"step": 383
},
{
"epoch": 0.03229808440397838,
"grad_norm": 0.4776511788368225,
"learning_rate": 0.00013564118787132506,
"loss": 0.5555,
"step": 384
},
{
"epoch": 0.03238219399878041,
"grad_norm": 0.339843213558197,
"learning_rate": 0.00013534748437792573,
"loss": 0.5663,
"step": 385
},
{
"epoch": 0.032466303593582435,
"grad_norm": 0.522289514541626,
"learning_rate": 0.0001350534320191259,
"loss": 0.7173,
"step": 386
},
{
"epoch": 0.03255041318838447,
"grad_norm": 0.6175903081893921,
"learning_rate": 0.0001347590336971037,
"loss": 0.6715,
"step": 387
},
{
"epoch": 0.032634522783186494,
"grad_norm": 0.5097800493240356,
"learning_rate": 0.0001344642923174517,
"loss": 0.7981,
"step": 388
},
{
"epoch": 0.03271863237798852,
"grad_norm": 0.45350587368011475,
"learning_rate": 0.00013416921078914835,
"loss": 0.7346,
"step": 389
},
{
"epoch": 0.032802741972790546,
"grad_norm": 0.7196130156517029,
"learning_rate": 0.00013387379202452917,
"loss": 0.6926,
"step": 390
},
{
"epoch": 0.03288685156759257,
"grad_norm": 0.44739460945129395,
"learning_rate": 0.00013357803893925807,
"loss": 0.6764,
"step": 391
},
{
"epoch": 0.0329709611623946,
"grad_norm": 0.42210137844085693,
"learning_rate": 0.00013328195445229868,
"loss": 0.5693,
"step": 392
},
{
"epoch": 0.033055070757196624,
"grad_norm": 0.25353190302848816,
"learning_rate": 0.00013298554148588528,
"loss": 0.2683,
"step": 393
},
{
"epoch": 0.03313918035199866,
"grad_norm": 0.5350023508071899,
"learning_rate": 0.00013268880296549425,
"loss": 0.6929,
"step": 394
},
{
"epoch": 0.03322328994680068,
"grad_norm": 0.5112295150756836,
"learning_rate": 0.00013239174181981495,
"loss": 0.6803,
"step": 395
},
{
"epoch": 0.03330739954160271,
"grad_norm": 0.5625605583190918,
"learning_rate": 0.00013209436098072095,
"loss": 0.7018,
"step": 396
},
{
"epoch": 0.033391509136404735,
"grad_norm": 0.6036564111709595,
"learning_rate": 0.00013179666338324108,
"loss": 0.6632,
"step": 397
},
{
"epoch": 0.03347561873120676,
"grad_norm": 0.6720706224441528,
"learning_rate": 0.0001314986519655305,
"loss": 0.6026,
"step": 398
},
{
"epoch": 0.033559728326008786,
"grad_norm": 0.5537442564964294,
"learning_rate": 0.0001312003296688415,
"loss": 0.7609,
"step": 399
},
{
"epoch": 0.03364383792081082,
"grad_norm": 0.567884087562561,
"learning_rate": 0.00013090169943749476,
"loss": 0.4756,
"step": 400
},
{
"epoch": 0.033727947515612845,
"grad_norm": 0.47909706830978394,
"learning_rate": 0.0001306027642188501,
"loss": 0.6392,
"step": 401
},
{
"epoch": 0.03381205711041487,
"grad_norm": 0.48807093501091003,
"learning_rate": 0.00013030352696327742,
"loss": 0.4313,
"step": 402
},
{
"epoch": 0.0338961667052169,
"grad_norm": 0.5797786712646484,
"learning_rate": 0.00013000399062412763,
"loss": 0.6879,
"step": 403
},
{
"epoch": 0.03398027630001892,
"grad_norm": 0.5691391229629517,
"learning_rate": 0.0001297041581577035,
"loss": 0.611,
"step": 404
},
{
"epoch": 0.03406438589482095,
"grad_norm": 0.491915762424469,
"learning_rate": 0.0001294040325232304,
"loss": 0.7652,
"step": 405
},
{
"epoch": 0.03414849548962298,
"grad_norm": 0.6337294578552246,
"learning_rate": 0.00012910361668282719,
"loss": 0.6599,
"step": 406
},
{
"epoch": 0.03423260508442501,
"grad_norm": 0.5192781686782837,
"learning_rate": 0.00012880291360147693,
"loss": 0.6136,
"step": 407
},
{
"epoch": 0.034316714679227034,
"grad_norm": 0.6465744376182556,
"learning_rate": 0.0001285019262469976,
"loss": 0.7522,
"step": 408
},
{
"epoch": 0.03440082427402906,
"grad_norm": 0.5028295516967773,
"learning_rate": 0.00012820065759001293,
"loss": 0.9119,
"step": 409
},
{
"epoch": 0.034484933868831086,
"grad_norm": 0.519190788269043,
"learning_rate": 0.00012789911060392294,
"loss": 0.4264,
"step": 410
},
{
"epoch": 0.03456904346363311,
"grad_norm": 0.6284279227256775,
"learning_rate": 0.0001275972882648746,
"loss": 0.7378,
"step": 411
},
{
"epoch": 0.03465315305843514,
"grad_norm": 0.679527997970581,
"learning_rate": 0.00012729519355173254,
"loss": 0.7436,
"step": 412
},
{
"epoch": 0.03473726265323717,
"grad_norm": 0.49145928025245667,
"learning_rate": 0.00012699282944604967,
"loss": 0.8936,
"step": 413
},
{
"epoch": 0.0348213722480392,
"grad_norm": 0.6400472521781921,
"learning_rate": 0.00012669019893203759,
"loss": 0.9139,
"step": 414
},
{
"epoch": 0.03490548184284122,
"grad_norm": 0.46638303995132446,
"learning_rate": 0.0001263873049965373,
"loss": 0.5398,
"step": 415
},
{
"epoch": 0.03498959143764325,
"grad_norm": 0.6132822632789612,
"learning_rate": 0.00012608415062898972,
"loss": 0.7835,
"step": 416
},
{
"epoch": 0.035073701032445274,
"grad_norm": 0.5918437838554382,
"learning_rate": 0.000125780738821406,
"loss": 0.6589,
"step": 417
},
{
"epoch": 0.0351578106272473,
"grad_norm": 0.45070680975914,
"learning_rate": 0.00012547707256833823,
"loss": 0.5931,
"step": 418
},
{
"epoch": 0.03524192022204933,
"grad_norm": 0.5677469372749329,
"learning_rate": 0.00012517315486684972,
"loss": 0.7679,
"step": 419
},
{
"epoch": 0.03532602981685136,
"grad_norm": 0.4486066997051239,
"learning_rate": 0.0001248689887164855,
"loss": 0.5029,
"step": 420
},
{
"epoch": 0.035410139411653385,
"grad_norm": 0.5143264532089233,
"learning_rate": 0.00012456457711924266,
"loss": 0.512,
"step": 421
},
{
"epoch": 0.03549424900645541,
"grad_norm": 0.5197947025299072,
"learning_rate": 0.00012425992307954075,
"loss": 0.6199,
"step": 422
},
{
"epoch": 0.03557835860125744,
"grad_norm": 0.5069385170936584,
"learning_rate": 0.0001239550296041922,
"loss": 0.5604,
"step": 423
},
{
"epoch": 0.03566246819605946,
"grad_norm": 0.5928438305854797,
"learning_rate": 0.00012364989970237248,
"loss": 0.8877,
"step": 424
},
{
"epoch": 0.035746577790861496,
"grad_norm": 0.5464084148406982,
"learning_rate": 0.00012334453638559057,
"loss": 0.7264,
"step": 425
},
{
"epoch": 0.03583068738566352,
"grad_norm": 0.5650874972343445,
"learning_rate": 0.00012303894266765908,
"loss": 0.6815,
"step": 426
},
{
"epoch": 0.03591479698046555,
"grad_norm": 0.5731722116470337,
"learning_rate": 0.00012273312156466464,
"loss": 0.9198,
"step": 427
},
{
"epoch": 0.035998906575267574,
"grad_norm": 0.3634185194969177,
"learning_rate": 0.00012242707609493814,
"loss": 0.6057,
"step": 428
},
{
"epoch": 0.0360830161700696,
"grad_norm": 0.5701162815093994,
"learning_rate": 0.00012212080927902474,
"loss": 0.6718,
"step": 429
},
{
"epoch": 0.036167125764871626,
"grad_norm": 0.5000714063644409,
"learning_rate": 0.00012181432413965428,
"loss": 0.7592,
"step": 430
},
{
"epoch": 0.03625123535967365,
"grad_norm": 0.5954747200012207,
"learning_rate": 0.00012150762370171136,
"loss": 0.7098,
"step": 431
},
{
"epoch": 0.036335344954475685,
"grad_norm": 0.5186465978622437,
"learning_rate": 0.00012120071099220549,
"loss": 0.5954,
"step": 432
},
{
"epoch": 0.03641945454927771,
"grad_norm": 0.5555379390716553,
"learning_rate": 0.00012089358904024117,
"loss": 0.7133,
"step": 433
},
{
"epoch": 0.036503564144079736,
"grad_norm": 0.46339845657348633,
"learning_rate": 0.00012058626087698814,
"loss": 0.7121,
"step": 434
},
{
"epoch": 0.03658767373888176,
"grad_norm": 0.45588961243629456,
"learning_rate": 0.00012027872953565125,
"loss": 0.705,
"step": 435
},
{
"epoch": 0.03667178333368379,
"grad_norm": 0.6387124061584473,
"learning_rate": 0.00011997099805144069,
"loss": 0.6375,
"step": 436
},
{
"epoch": 0.036755892928485814,
"grad_norm": 0.5656935572624207,
"learning_rate": 0.000119663069461542,
"loss": 0.7211,
"step": 437
},
{
"epoch": 0.03684000252328785,
"grad_norm": 0.4781017303466797,
"learning_rate": 0.00011935494680508606,
"loss": 0.6053,
"step": 438
},
{
"epoch": 0.03692411211808987,
"grad_norm": 0.5316019654273987,
"learning_rate": 0.00011904663312311901,
"loss": 0.6341,
"step": 439
},
{
"epoch": 0.0370082217128919,
"grad_norm": 0.535388171672821,
"learning_rate": 0.00011873813145857249,
"loss": 0.7128,
"step": 440
},
{
"epoch": 0.037092331307693925,
"grad_norm": 0.7049708366394043,
"learning_rate": 0.00011842944485623335,
"loss": 0.673,
"step": 441
},
{
"epoch": 0.03717644090249595,
"grad_norm": 0.5531265139579773,
"learning_rate": 0.00011812057636271374,
"loss": 0.7491,
"step": 442
},
{
"epoch": 0.03726055049729798,
"grad_norm": 0.5389276146888733,
"learning_rate": 0.000117811529026421,
"loss": 0.6979,
"step": 443
},
{
"epoch": 0.0373446600921,
"grad_norm": 0.3878541588783264,
"learning_rate": 0.00011750230589752762,
"loss": 0.4515,
"step": 444
},
{
"epoch": 0.037428769686902036,
"grad_norm": 0.7032809853553772,
"learning_rate": 0.00011719291002794096,
"loss": 0.6921,
"step": 445
},
{
"epoch": 0.03751287928170406,
"grad_norm": 0.5832718014717102,
"learning_rate": 0.00011688334447127338,
"loss": 0.508,
"step": 446
},
{
"epoch": 0.03759698887650609,
"grad_norm": 0.4945680499076843,
"learning_rate": 0.00011657361228281199,
"loss": 0.5809,
"step": 447
},
{
"epoch": 0.037681098471308114,
"grad_norm": 0.6034420728683472,
"learning_rate": 0.00011626371651948838,
"loss": 0.6125,
"step": 448
},
{
"epoch": 0.03776520806611014,
"grad_norm": 0.6533922553062439,
"learning_rate": 0.00011595366023984864,
"loss": 0.5984,
"step": 449
},
{
"epoch": 0.037849317660912166,
"grad_norm": 0.46035146713256836,
"learning_rate": 0.0001156434465040231,
"loss": 0.6094,
"step": 450
},
{
"epoch": 0.0379334272557142,
"grad_norm": 0.6323526501655579,
"learning_rate": 0.00011533307837369607,
"loss": 0.9116,
"step": 451
},
{
"epoch": 0.038017536850516225,
"grad_norm": 0.4355090260505676,
"learning_rate": 0.00011502255891207572,
"loss": 0.4778,
"step": 452
},
{
"epoch": 0.03810164644531825,
"grad_norm": 0.5190046429634094,
"learning_rate": 0.00011471189118386375,
"loss": 0.5997,
"step": 453
},
{
"epoch": 0.038185756040120276,
"grad_norm": 0.534466564655304,
"learning_rate": 0.00011440107825522521,
"loss": 0.7846,
"step": 454
},
{
"epoch": 0.0382698656349223,
"grad_norm": 0.6099439859390259,
"learning_rate": 0.00011409012319375827,
"loss": 0.5393,
"step": 455
},
{
"epoch": 0.03835397522972433,
"grad_norm": 0.4053705036640167,
"learning_rate": 0.0001137790290684638,
"loss": 0.547,
"step": 456
},
{
"epoch": 0.03843808482452636,
"grad_norm": 0.3879355788230896,
"learning_rate": 0.00011346779894971527,
"loss": 0.5772,
"step": 457
},
{
"epoch": 0.03852219441932839,
"grad_norm": 0.7428577542304993,
"learning_rate": 0.00011315643590922827,
"loss": 0.9059,
"step": 458
},
{
"epoch": 0.03860630401413041,
"grad_norm": 0.41162705421447754,
"learning_rate": 0.0001128449430200303,
"loss": 0.6041,
"step": 459
},
{
"epoch": 0.03869041360893244,
"grad_norm": 0.5828412175178528,
"learning_rate": 0.00011253332335643043,
"loss": 0.4467,
"step": 460
},
{
"epoch": 0.038774523203734465,
"grad_norm": 0.5897490978240967,
"learning_rate": 0.00011222157999398895,
"loss": 0.8586,
"step": 461
},
{
"epoch": 0.03885863279853649,
"grad_norm": 0.4602713882923126,
"learning_rate": 0.00011190971600948699,
"loss": 0.6802,
"step": 462
},
{
"epoch": 0.03894274239333852,
"grad_norm": 0.5070551633834839,
"learning_rate": 0.00011159773448089614,
"loss": 0.5585,
"step": 463
},
{
"epoch": 0.03902685198814055,
"grad_norm": 0.4254646897315979,
"learning_rate": 0.00011128563848734816,
"loss": 0.7244,
"step": 464
},
{
"epoch": 0.039110961582942576,
"grad_norm": 0.5656306147575378,
"learning_rate": 0.00011097343110910452,
"loss": 0.8105,
"step": 465
},
{
"epoch": 0.0391950711777446,
"grad_norm": 0.5550054311752319,
"learning_rate": 0.000110661115427526,
"loss": 0.4696,
"step": 466
},
{
"epoch": 0.03927918077254663,
"grad_norm": 0.47665438055992126,
"learning_rate": 0.00011034869452504226,
"loss": 0.4867,
"step": 467
},
{
"epoch": 0.039363290367348654,
"grad_norm": 0.6439712047576904,
"learning_rate": 0.00011003617148512149,
"loss": 0.6037,
"step": 468
},
{
"epoch": 0.03944739996215068,
"grad_norm": 0.49243077635765076,
"learning_rate": 0.00010972354939223996,
"loss": 0.646,
"step": 469
},
{
"epoch": 0.03953150955695271,
"grad_norm": 0.4704420268535614,
"learning_rate": 0.00010941083133185146,
"loss": 0.7801,
"step": 470
},
{
"epoch": 0.03961561915175474,
"grad_norm": 0.4875670075416565,
"learning_rate": 0.00010909802039035701,
"loss": 0.5314,
"step": 471
},
{
"epoch": 0.039699728746556764,
"grad_norm": 0.5254621505737305,
"learning_rate": 0.00010878511965507434,
"loss": 0.659,
"step": 472
},
{
"epoch": 0.03978383834135879,
"grad_norm": 0.5760181546211243,
"learning_rate": 0.00010847213221420736,
"loss": 0.9349,
"step": 473
},
{
"epoch": 0.039867947936160816,
"grad_norm": 0.4383930563926697,
"learning_rate": 0.00010815906115681578,
"loss": 0.8152,
"step": 474
},
{
"epoch": 0.03995205753096284,
"grad_norm": 0.5969526171684265,
"learning_rate": 0.0001078459095727845,
"loss": 0.7431,
"step": 475
},
{
"epoch": 0.04003616712576487,
"grad_norm": 0.5625801086425781,
"learning_rate": 0.00010753268055279329,
"loss": 0.7582,
"step": 476
},
{
"epoch": 0.0401202767205669,
"grad_norm": 0.5394108295440674,
"learning_rate": 0.0001072193771882861,
"loss": 0.6964,
"step": 477
},
{
"epoch": 0.04020438631536893,
"grad_norm": 0.8828211426734924,
"learning_rate": 0.00010690600257144061,
"loss": 0.6602,
"step": 478
},
{
"epoch": 0.04028849591017095,
"grad_norm": 0.6621891856193542,
"learning_rate": 0.0001065925597951378,
"loss": 0.6224,
"step": 479
},
{
"epoch": 0.04037260550497298,
"grad_norm": 0.5769515633583069,
"learning_rate": 0.00010627905195293135,
"loss": 0.7293,
"step": 480
},
{
"epoch": 0.040456715099775005,
"grad_norm": 0.47239354252815247,
"learning_rate": 0.00010596548213901708,
"loss": 0.6623,
"step": 481
},
{
"epoch": 0.04054082469457703,
"grad_norm": 0.5481030344963074,
"learning_rate": 0.00010565185344820247,
"loss": 0.8693,
"step": 482
},
{
"epoch": 0.040624934289379064,
"grad_norm": 0.5839731693267822,
"learning_rate": 0.00010533816897587606,
"loss": 0.7281,
"step": 483
},
{
"epoch": 0.04070904388418109,
"grad_norm": 0.5088500380516052,
"learning_rate": 0.00010502443181797697,
"loss": 0.6444,
"step": 484
},
{
"epoch": 0.040793153478983116,
"grad_norm": 0.514184832572937,
"learning_rate": 0.00010471064507096426,
"loss": 0.678,
"step": 485
},
{
"epoch": 0.04087726307378514,
"grad_norm": 0.48522424697875977,
"learning_rate": 0.0001043968118317865,
"loss": 0.6344,
"step": 486
},
{
"epoch": 0.04096137266858717,
"grad_norm": 0.5985432863235474,
"learning_rate": 0.00010408293519785101,
"loss": 0.6916,
"step": 487
},
{
"epoch": 0.041045482263389194,
"grad_norm": 0.5493951439857483,
"learning_rate": 0.00010376901826699348,
"loss": 0.6433,
"step": 488
},
{
"epoch": 0.041129591858191226,
"grad_norm": 0.5034871101379395,
"learning_rate": 0.00010345506413744726,
"loss": 0.6704,
"step": 489
},
{
"epoch": 0.04121370145299325,
"grad_norm": 0.5221031904220581,
"learning_rate": 0.00010314107590781284,
"loss": 0.4167,
"step": 490
},
{
"epoch": 0.04129781104779528,
"grad_norm": 0.5250301957130432,
"learning_rate": 0.00010282705667702734,
"loss": 0.5978,
"step": 491
},
{
"epoch": 0.041381920642597304,
"grad_norm": 0.5023308396339417,
"learning_rate": 0.00010251300954433376,
"loss": 0.448,
"step": 492
},
{
"epoch": 0.04146603023739933,
"grad_norm": 0.6943901181221008,
"learning_rate": 0.00010219893760925052,
"loss": 0.7664,
"step": 493
},
{
"epoch": 0.041550139832201356,
"grad_norm": 0.4787749946117401,
"learning_rate": 0.00010188484397154084,
"loss": 0.8114,
"step": 494
},
{
"epoch": 0.04163424942700338,
"grad_norm": 0.5617536902427673,
"learning_rate": 0.00010157073173118208,
"loss": 0.6003,
"step": 495
},
{
"epoch": 0.041718359021805415,
"grad_norm": 0.3669779300689697,
"learning_rate": 0.00010125660398833528,
"loss": 0.5983,
"step": 496
},
{
"epoch": 0.04180246861660744,
"grad_norm": 0.46090903878211975,
"learning_rate": 0.00010094246384331442,
"loss": 0.7085,
"step": 497
},
{
"epoch": 0.04188657821140947,
"grad_norm": 0.4907785952091217,
"learning_rate": 0.00010062831439655591,
"loss": 0.6197,
"step": 498
},
{
"epoch": 0.04197068780621149,
"grad_norm": 0.4900142252445221,
"learning_rate": 0.00010031415874858797,
"loss": 0.53,
"step": 499
},
{
"epoch": 0.04205479740101352,
"grad_norm": 0.5708168745040894,
"learning_rate": 0.0001,
"loss": 0.7168,
"step": 500
},
{
"epoch": 0.042138906995815545,
"grad_norm": 0.5526877045631409,
"learning_rate": 9.968584125141204e-05,
"loss": 0.7179,
"step": 501
},
{
"epoch": 0.04222301659061758,
"grad_norm": 0.4733210802078247,
"learning_rate": 9.937168560344412e-05,
"loss": 0.4766,
"step": 502
},
{
"epoch": 0.042307126185419604,
"grad_norm": 0.5687675476074219,
"learning_rate": 9.90575361566856e-05,
"loss": 0.7382,
"step": 503
},
{
"epoch": 0.04239123578022163,
"grad_norm": 0.5211085677146912,
"learning_rate": 9.874339601166473e-05,
"loss": 0.7914,
"step": 504
},
{
"epoch": 0.042475345375023656,
"grad_norm": 0.6340366005897522,
"learning_rate": 9.842926826881796e-05,
"loss": 0.7195,
"step": 505
},
{
"epoch": 0.04255945496982568,
"grad_norm": 0.5366642475128174,
"learning_rate": 9.81151560284592e-05,
"loss": 0.6782,
"step": 506
},
{
"epoch": 0.04264356456462771,
"grad_norm": 0.6311143040657043,
"learning_rate": 9.78010623907495e-05,
"loss": 0.7104,
"step": 507
},
{
"epoch": 0.042727674159429734,
"grad_norm": 0.5550001859664917,
"learning_rate": 9.748699045566626e-05,
"loss": 0.834,
"step": 508
},
{
"epoch": 0.042811783754231766,
"grad_norm": 0.601676881313324,
"learning_rate": 9.717294332297268e-05,
"loss": 0.7551,
"step": 509
},
{
"epoch": 0.04289589334903379,
"grad_norm": 0.4123936891555786,
"learning_rate": 9.685892409218717e-05,
"loss": 0.7682,
"step": 510
},
{
"epoch": 0.04298000294383582,
"grad_norm": 0.7674761414527893,
"learning_rate": 9.654493586255278e-05,
"loss": 0.6967,
"step": 511
},
{
"epoch": 0.043064112538637844,
"grad_norm": 0.6111290454864502,
"learning_rate": 9.623098173300654e-05,
"loss": 0.5819,
"step": 512
},
{
"epoch": 0.04314822213343987,
"grad_norm": 0.43844571709632874,
"learning_rate": 9.591706480214901e-05,
"loss": 0.6972,
"step": 513
},
{
"epoch": 0.043232331728241896,
"grad_norm": 0.538962185382843,
"learning_rate": 9.560318816821353e-05,
"loss": 0.7395,
"step": 514
},
{
"epoch": 0.04331644132304393,
"grad_norm": 0.5768121480941772,
"learning_rate": 9.528935492903575e-05,
"loss": 0.8873,
"step": 515
},
{
"epoch": 0.043400550917845955,
"grad_norm": 0.492720365524292,
"learning_rate": 9.497556818202306e-05,
"loss": 0.5948,
"step": 516
},
{
"epoch": 0.04348466051264798,
"grad_norm": 0.516925036907196,
"learning_rate": 9.466183102412395e-05,
"loss": 0.6319,
"step": 517
},
{
"epoch": 0.04356877010745001,
"grad_norm": 0.6040381193161011,
"learning_rate": 9.434814655179755e-05,
"loss": 0.6935,
"step": 518
},
{
"epoch": 0.04365287970225203,
"grad_norm": 0.48684102296829224,
"learning_rate": 9.403451786098294e-05,
"loss": 0.5688,
"step": 519
},
{
"epoch": 0.04373698929705406,
"grad_norm": 0.5389447212219238,
"learning_rate": 9.372094804706867e-05,
"loss": 0.8248,
"step": 520
},
{
"epoch": 0.04382109889185609,
"grad_norm": 0.4723369777202606,
"learning_rate": 9.340744020486222e-05,
"loss": 0.5912,
"step": 521
},
{
"epoch": 0.04390520848665812,
"grad_norm": 0.5267740488052368,
"learning_rate": 9.309399742855942e-05,
"loss": 0.5262,
"step": 522
},
{
"epoch": 0.043989318081460144,
"grad_norm": 0.52064049243927,
"learning_rate": 9.278062281171393e-05,
"loss": 0.6672,
"step": 523
},
{
"epoch": 0.04407342767626217,
"grad_norm": 0.46649792790412903,
"learning_rate": 9.246731944720675e-05,
"loss": 0.7222,
"step": 524
},
{
"epoch": 0.044157537271064196,
"grad_norm": 0.5226630568504333,
"learning_rate": 9.215409042721552e-05,
"loss": 0.6677,
"step": 525
},
{
"epoch": 0.04424164686586622,
"grad_norm": 0.5565252304077148,
"learning_rate": 9.184093884318425e-05,
"loss": 0.5528,
"step": 526
},
{
"epoch": 0.04432575646066825,
"grad_norm": 0.5086897611618042,
"learning_rate": 9.152786778579267e-05,
"loss": 0.6953,
"step": 527
},
{
"epoch": 0.04440986605547028,
"grad_norm": 0.5770063400268555,
"learning_rate": 9.121488034492569e-05,
"loss": 0.7933,
"step": 528
},
{
"epoch": 0.044493975650272306,
"grad_norm": 0.6236185431480408,
"learning_rate": 9.090197960964301e-05,
"loss": 0.7398,
"step": 529
},
{
"epoch": 0.04457808524507433,
"grad_norm": 0.605636477470398,
"learning_rate": 9.058916866814858e-05,
"loss": 0.7508,
"step": 530
},
{
"epoch": 0.04466219483987636,
"grad_norm": 0.5256229043006897,
"learning_rate": 9.027645060776006e-05,
"loss": 0.7415,
"step": 531
},
{
"epoch": 0.044746304434678384,
"grad_norm": 0.4349639117717743,
"learning_rate": 8.99638285148785e-05,
"loss": 0.7829,
"step": 532
},
{
"epoch": 0.04483041402948041,
"grad_norm": 0.5248279571533203,
"learning_rate": 8.965130547495776e-05,
"loss": 0.6153,
"step": 533
},
{
"epoch": 0.04491452362428244,
"grad_norm": 0.47936955094337463,
"learning_rate": 8.933888457247402e-05,
"loss": 0.5984,
"step": 534
},
{
"epoch": 0.04499863321908447,
"grad_norm": 0.4791763722896576,
"learning_rate": 8.902656889089548e-05,
"loss": 0.6953,
"step": 535
},
{
"epoch": 0.045082742813886495,
"grad_norm": 0.47786253690719604,
"learning_rate": 8.871436151265184e-05,
"loss": 0.6712,
"step": 536
},
{
"epoch": 0.04516685240868852,
"grad_norm": 0.580772340297699,
"learning_rate": 8.840226551910387e-05,
"loss": 0.6813,
"step": 537
},
{
"epoch": 0.04525096200349055,
"grad_norm": 0.6560893058776855,
"learning_rate": 8.809028399051302e-05,
"loss": 0.5794,
"step": 538
},
{
"epoch": 0.04533507159829257,
"grad_norm": 0.479247510433197,
"learning_rate": 8.777842000601105e-05,
"loss": 0.6182,
"step": 539
},
{
"epoch": 0.045419181193094606,
"grad_norm": 0.5315003395080566,
"learning_rate": 8.746667664356956e-05,
"loss": 0.6151,
"step": 540
},
{
"epoch": 0.04550329078789663,
"grad_norm": 0.4729808568954468,
"learning_rate": 8.715505697996971e-05,
"loss": 0.6079,
"step": 541
},
{
"epoch": 0.04558740038269866,
"grad_norm": 0.5379468202590942,
"learning_rate": 8.684356409077176e-05,
"loss": 0.6553,
"step": 542
},
{
"epoch": 0.045671509977500684,
"grad_norm": 0.6452252864837646,
"learning_rate": 8.653220105028474e-05,
"loss": 0.7077,
"step": 543
},
{
"epoch": 0.04575561957230271,
"grad_norm": 0.5322025418281555,
"learning_rate": 8.62209709315362e-05,
"loss": 0.5957,
"step": 544
},
{
"epoch": 0.045839729167104735,
"grad_norm": 0.6063228249549866,
"learning_rate": 8.590987680624174e-05,
"loss": 0.8442,
"step": 545
},
{
"epoch": 0.04592383876190676,
"grad_norm": 0.6377241015434265,
"learning_rate": 8.559892174477479e-05,
"loss": 0.6145,
"step": 546
},
{
"epoch": 0.046007948356708794,
"grad_norm": 0.49134746193885803,
"learning_rate": 8.528810881613626e-05,
"loss": 0.7311,
"step": 547
},
{
"epoch": 0.04609205795151082,
"grad_norm": 0.7310553193092346,
"learning_rate": 8.497744108792429e-05,
"loss": 0.7093,
"step": 548
},
{
"epoch": 0.046176167546312846,
"grad_norm": 0.44871386885643005,
"learning_rate": 8.466692162630392e-05,
"loss": 0.488,
"step": 549
},
{
"epoch": 0.04626027714111487,
"grad_norm": 0.604282021522522,
"learning_rate": 8.435655349597689e-05,
"loss": 0.5964,
"step": 550
},
{
"epoch": 0.0463443867359169,
"grad_norm": 0.46552160382270813,
"learning_rate": 8.404633976015134e-05,
"loss": 0.7442,
"step": 551
},
{
"epoch": 0.046428496330718924,
"grad_norm": 0.5520333051681519,
"learning_rate": 8.373628348051165e-05,
"loss": 0.5222,
"step": 552
},
{
"epoch": 0.04651260592552096,
"grad_norm": 0.5013942718505859,
"learning_rate": 8.342638771718802e-05,
"loss": 0.6123,
"step": 553
},
{
"epoch": 0.04659671552032298,
"grad_norm": 0.5522478222846985,
"learning_rate": 8.311665552872662e-05,
"loss": 0.6076,
"step": 554
},
{
"epoch": 0.04668082511512501,
"grad_norm": 0.43584609031677246,
"learning_rate": 8.280708997205904e-05,
"loss": 0.5835,
"step": 555
},
{
"epoch": 0.046764934709927035,
"grad_norm": 0.6730960011482239,
"learning_rate": 8.249769410247239e-05,
"loss": 0.5367,
"step": 556
},
{
"epoch": 0.04684904430472906,
"grad_norm": 0.4698852002620697,
"learning_rate": 8.218847097357898e-05,
"loss": 0.7015,
"step": 557
},
{
"epoch": 0.04693315389953109,
"grad_norm": 0.6545222401618958,
"learning_rate": 8.187942363728625e-05,
"loss": 0.7149,
"step": 558
},
{
"epoch": 0.04701726349433311,
"grad_norm": 0.5913302898406982,
"learning_rate": 8.157055514376666e-05,
"loss": 0.634,
"step": 559
},
{
"epoch": 0.047101373089135146,
"grad_norm": 0.6566689610481262,
"learning_rate": 8.126186854142752e-05,
"loss": 0.6646,
"step": 560
},
{
"epoch": 0.04718548268393717,
"grad_norm": 0.5921803116798401,
"learning_rate": 8.095336687688102e-05,
"loss": 0.5967,
"step": 561
},
{
"epoch": 0.0472695922787392,
"grad_norm": 0.5838126540184021,
"learning_rate": 8.064505319491398e-05,
"loss": 0.6807,
"step": 562
},
{
"epoch": 0.047353701873541224,
"grad_norm": 0.5889870524406433,
"learning_rate": 8.033693053845801e-05,
"loss": 0.6156,
"step": 563
},
{
"epoch": 0.04743781146834325,
"grad_norm": 0.5544804334640503,
"learning_rate": 8.002900194855932e-05,
"loss": 0.7279,
"step": 564
},
{
"epoch": 0.047521921063145275,
"grad_norm": 0.5051852464675903,
"learning_rate": 7.972127046434878e-05,
"loss": 0.6375,
"step": 565
},
{
"epoch": 0.04760603065794731,
"grad_norm": 0.6946410536766052,
"learning_rate": 7.941373912301189e-05,
"loss": 0.8067,
"step": 566
},
{
"epoch": 0.047690140252749334,
"grad_norm": 0.5952439904212952,
"learning_rate": 7.910641095975886e-05,
"loss": 0.8346,
"step": 567
},
{
"epoch": 0.04777424984755136,
"grad_norm": 0.5868372321128845,
"learning_rate": 7.879928900779456e-05,
"loss": 0.8255,
"step": 568
},
{
"epoch": 0.047858359442353386,
"grad_norm": 0.6287057399749756,
"learning_rate": 7.849237629828869e-05,
"loss": 0.5474,
"step": 569
},
{
"epoch": 0.04794246903715541,
"grad_norm": 0.5471776127815247,
"learning_rate": 7.818567586034577e-05,
"loss": 0.5467,
"step": 570
},
{
"epoch": 0.04802657863195744,
"grad_norm": 0.8107778429985046,
"learning_rate": 7.787919072097531e-05,
"loss": 0.6715,
"step": 571
},
{
"epoch": 0.04811068822675947,
"grad_norm": 0.6035006046295166,
"learning_rate": 7.75729239050619e-05,
"loss": 0.6016,
"step": 572
},
{
"epoch": 0.0481947978215615,
"grad_norm": 0.6550392508506775,
"learning_rate": 7.726687843533538e-05,
"loss": 0.5897,
"step": 573
},
{
"epoch": 0.04827890741636352,
"grad_norm": 0.5324556827545166,
"learning_rate": 7.696105733234098e-05,
"loss": 0.5424,
"step": 574
},
{
"epoch": 0.04836301701116555,
"grad_norm": 0.5881845355033875,
"learning_rate": 7.66554636144095e-05,
"loss": 0.6065,
"step": 575
},
{
"epoch": 0.048447126605967575,
"grad_norm": 0.537804901599884,
"learning_rate": 7.635010029762756e-05,
"loss": 0.5698,
"step": 576
},
{
"epoch": 0.0485312362007696,
"grad_norm": 0.5901293754577637,
"learning_rate": 7.604497039580785e-05,
"loss": 0.6954,
"step": 577
},
{
"epoch": 0.04861534579557163,
"grad_norm": 0.5681110620498657,
"learning_rate": 7.574007692045928e-05,
"loss": 0.6799,
"step": 578
},
{
"epoch": 0.04869945539037366,
"grad_norm": 0.5399741530418396,
"learning_rate": 7.543542288075739e-05,
"loss": 0.6881,
"step": 579
},
{
"epoch": 0.048783564985175686,
"grad_norm": 0.605559229850769,
"learning_rate": 7.513101128351454e-05,
"loss": 0.608,
"step": 580
},
{
"epoch": 0.04886767457997771,
"grad_norm": 0.5992718935012817,
"learning_rate": 7.48268451331503e-05,
"loss": 0.6869,
"step": 581
},
{
"epoch": 0.04895178417477974,
"grad_norm": 0.8363056182861328,
"learning_rate": 7.45229274316618e-05,
"loss": 0.7678,
"step": 582
},
{
"epoch": 0.04903589376958176,
"grad_norm": 0.421335905790329,
"learning_rate": 7.421926117859403e-05,
"loss": 0.508,
"step": 583
},
{
"epoch": 0.04912000336438379,
"grad_norm": 0.4553971588611603,
"learning_rate": 7.391584937101033e-05,
"loss": 0.6228,
"step": 584
},
{
"epoch": 0.04920411295918582,
"grad_norm": 0.5710154175758362,
"learning_rate": 7.361269500346274e-05,
"loss": 0.6005,
"step": 585
},
{
"epoch": 0.04928822255398785,
"grad_norm": 0.5714388489723206,
"learning_rate": 7.330980106796246e-05,
"loss": 0.7007,
"step": 586
},
{
"epoch": 0.049372332148789874,
"grad_norm": 0.5537533760070801,
"learning_rate": 7.300717055395039e-05,
"loss": 0.6398,
"step": 587
},
{
"epoch": 0.0494564417435919,
"grad_norm": 0.43929553031921387,
"learning_rate": 7.270480644826749e-05,
"loss": 0.6056,
"step": 588
},
{
"epoch": 0.049540551338393926,
"grad_norm": 0.485586017370224,
"learning_rate": 7.240271173512546e-05,
"loss": 0.5453,
"step": 589
},
{
"epoch": 0.04962466093319595,
"grad_norm": 0.43666794896125793,
"learning_rate": 7.210088939607708e-05,
"loss": 0.5794,
"step": 590
},
{
"epoch": 0.04970877052799798,
"grad_norm": 0.5430508852005005,
"learning_rate": 7.179934240998706e-05,
"loss": 0.8037,
"step": 591
},
{
"epoch": 0.04979288012280001,
"grad_norm": 0.6481785774230957,
"learning_rate": 7.149807375300239e-05,
"loss": 0.7141,
"step": 592
},
{
"epoch": 0.04987698971760204,
"grad_norm": 0.5810648202896118,
"learning_rate": 7.119708639852312e-05,
"loss": 0.5569,
"step": 593
},
{
"epoch": 0.04996109931240406,
"grad_norm": 0.6031057238578796,
"learning_rate": 7.089638331717284e-05,
"loss": 0.6682,
"step": 594
},
{
"epoch": 0.05004520890720609,
"grad_norm": 0.48759695887565613,
"learning_rate": 7.059596747676962e-05,
"loss": 0.634,
"step": 595
},
{
"epoch": 0.050129318502008115,
"grad_norm": 0.4522048532962799,
"learning_rate": 7.029584184229653e-05,
"loss": 0.6403,
"step": 596
},
{
"epoch": 0.05021342809681014,
"grad_norm": 0.4637600779533386,
"learning_rate": 6.999600937587239e-05,
"loss": 0.798,
"step": 597
},
{
"epoch": 0.050297537691612174,
"grad_norm": 0.5411447882652283,
"learning_rate": 6.969647303672262e-05,
"loss": 0.6316,
"step": 598
},
{
"epoch": 0.0503816472864142,
"grad_norm": 0.48814845085144043,
"learning_rate": 6.939723578114993e-05,
"loss": 0.775,
"step": 599
},
{
"epoch": 0.050465756881216225,
"grad_norm": 0.6729063391685486,
"learning_rate": 6.909830056250527e-05,
"loss": 0.6383,
"step": 600
},
{
"epoch": 0.05054986647601825,
"grad_norm": 0.5428520441055298,
"learning_rate": 6.879967033115853e-05,
"loss": 0.5884,
"step": 601
},
{
"epoch": 0.05063397607082028,
"grad_norm": 0.5439213514328003,
"learning_rate": 6.850134803446954e-05,
"loss": 0.6811,
"step": 602
},
{
"epoch": 0.0507180856656223,
"grad_norm": 0.5737605690956116,
"learning_rate": 6.820333661675893e-05,
"loss": 0.559,
"step": 603
},
{
"epoch": 0.050802195260424336,
"grad_norm": 0.610083281993866,
"learning_rate": 6.790563901927907e-05,
"loss": 0.6291,
"step": 604
},
{
"epoch": 0.05088630485522636,
"grad_norm": 0.6092396378517151,
"learning_rate": 6.760825818018508e-05,
"loss": 0.8043,
"step": 605
},
{
"epoch": 0.05097041445002839,
"grad_norm": 0.43380871415138245,
"learning_rate": 6.731119703450577e-05,
"loss": 0.8899,
"step": 606
},
{
"epoch": 0.051054524044830414,
"grad_norm": 0.6090589761734009,
"learning_rate": 6.701445851411472e-05,
"loss": 0.7548,
"step": 607
},
{
"epoch": 0.05113863363963244,
"grad_norm": 0.5451867580413818,
"learning_rate": 6.671804554770135e-05,
"loss": 0.5946,
"step": 608
},
{
"epoch": 0.051222743234434466,
"grad_norm": 0.6997014880180359,
"learning_rate": 6.642196106074194e-05,
"loss": 0.6974,
"step": 609
},
{
"epoch": 0.05130685282923649,
"grad_norm": 0.5238434672355652,
"learning_rate": 6.612620797547087e-05,
"loss": 0.6132,
"step": 610
},
{
"epoch": 0.051390962424038525,
"grad_norm": 0.5040790438652039,
"learning_rate": 6.583078921085167e-05,
"loss": 0.7068,
"step": 611
},
{
"epoch": 0.05147507201884055,
"grad_norm": 0.5203388333320618,
"learning_rate": 6.55357076825483e-05,
"loss": 0.7418,
"step": 612
},
{
"epoch": 0.05155918161364258,
"grad_norm": 0.5750290751457214,
"learning_rate": 6.52409663028963e-05,
"loss": 0.5788,
"step": 613
},
{
"epoch": 0.0516432912084446,
"grad_norm": 0.4557915925979614,
"learning_rate": 6.494656798087412e-05,
"loss": 0.6357,
"step": 614
},
{
"epoch": 0.05172740080324663,
"grad_norm": 0.48326605558395386,
"learning_rate": 6.465251562207431e-05,
"loss": 0.6878,
"step": 615
},
{
"epoch": 0.051811510398048655,
"grad_norm": 0.5541282296180725,
"learning_rate": 6.435881212867493e-05,
"loss": 0.5612,
"step": 616
},
{
"epoch": 0.05189561999285069,
"grad_norm": 0.5726374983787537,
"learning_rate": 6.406546039941094e-05,
"loss": 0.5127,
"step": 617
},
{
"epoch": 0.051979729587652714,
"grad_norm": 0.4574257731437683,
"learning_rate": 6.377246332954544e-05,
"loss": 0.6469,
"step": 618
},
{
"epoch": 0.05206383918245474,
"grad_norm": 0.5137540698051453,
"learning_rate": 6.347982381084123e-05,
"loss": 0.6729,
"step": 619
},
{
"epoch": 0.052147948777256765,
"grad_norm": 0.5800307393074036,
"learning_rate": 6.318754473153221e-05,
"loss": 0.7838,
"step": 620
},
{
"epoch": 0.05223205837205879,
"grad_norm": 0.45078638195991516,
"learning_rate": 6.289562897629492e-05,
"loss": 0.4851,
"step": 621
},
{
"epoch": 0.05231616796686082,
"grad_norm": 0.45273396372795105,
"learning_rate": 6.260407942621998e-05,
"loss": 0.4557,
"step": 622
},
{
"epoch": 0.05240027756166284,
"grad_norm": 0.46480339765548706,
"learning_rate": 6.231289895878375e-05,
"loss": 0.5967,
"step": 623
},
{
"epoch": 0.052484387156464876,
"grad_norm": 0.5039330124855042,
"learning_rate": 6.20220904478199e-05,
"loss": 0.6384,
"step": 624
},
{
"epoch": 0.0525684967512669,
"grad_norm": 0.637080192565918,
"learning_rate": 6.173165676349103e-05,
"loss": 0.8325,
"step": 625
},
{
"epoch": 0.05265260634606893,
"grad_norm": 0.5003658533096313,
"learning_rate": 6.144160077226036e-05,
"loss": 0.5859,
"step": 626
},
{
"epoch": 0.052736715940870954,
"grad_norm": 0.6767755150794983,
"learning_rate": 6.11519253368634e-05,
"loss": 0.6984,
"step": 627
},
{
"epoch": 0.05282082553567298,
"grad_norm": 0.5123332142829895,
"learning_rate": 6.086263331627976e-05,
"loss": 0.5831,
"step": 628
},
{
"epoch": 0.052904935130475006,
"grad_norm": 0.6226670742034912,
"learning_rate": 6.05737275657049e-05,
"loss": 0.7954,
"step": 629
},
{
"epoch": 0.05298904472527704,
"grad_norm": 0.584923267364502,
"learning_rate": 6.0285210936521955e-05,
"loss": 0.7863,
"step": 630
},
{
"epoch": 0.053073154320079065,
"grad_norm": 0.5787730813026428,
"learning_rate": 5.999708627627354e-05,
"loss": 0.7289,
"step": 631
},
{
"epoch": 0.05315726391488109,
"grad_norm": 0.5147082805633545,
"learning_rate": 5.9709356428633746e-05,
"loss": 0.5297,
"step": 632
},
{
"epoch": 0.05324137350968312,
"grad_norm": 0.6487715840339661,
"learning_rate": 5.9422024233380013e-05,
"loss": 0.6826,
"step": 633
},
{
"epoch": 0.05332548310448514,
"grad_norm": 0.5044452548027039,
"learning_rate": 5.913509252636511e-05,
"loss": 0.5771,
"step": 634
},
{
"epoch": 0.05340959269928717,
"grad_norm": 0.6998575329780579,
"learning_rate": 5.884856413948913e-05,
"loss": 0.7294,
"step": 635
},
{
"epoch": 0.0534937022940892,
"grad_norm": 0.4688400626182556,
"learning_rate": 5.856244190067159e-05,
"loss": 0.6579,
"step": 636
},
{
"epoch": 0.05357781188889123,
"grad_norm": 0.4600168466567993,
"learning_rate": 5.82767286338235e-05,
"loss": 0.5962,
"step": 637
},
{
"epoch": 0.05366192148369325,
"grad_norm": 0.6258283257484436,
"learning_rate": 5.799142715881938e-05,
"loss": 0.6653,
"step": 638
},
{
"epoch": 0.05374603107849528,
"grad_norm": 0.5476671457290649,
"learning_rate": 5.770654029146969e-05,
"loss": 0.6163,
"step": 639
},
{
"epoch": 0.053830140673297305,
"grad_norm": 0.4993649125099182,
"learning_rate": 5.7422070843492734e-05,
"loss": 0.6052,
"step": 640
},
{
"epoch": 0.05391425026809933,
"grad_norm": 0.5785011053085327,
"learning_rate": 5.713802162248718e-05,
"loss": 0.6716,
"step": 641
},
{
"epoch": 0.05399835986290136,
"grad_norm": 0.6870988607406616,
"learning_rate": 5.6854395431904094e-05,
"loss": 0.6964,
"step": 642
},
{
"epoch": 0.05408246945770339,
"grad_norm": 0.40551093220710754,
"learning_rate": 5.657119507101954e-05,
"loss": 0.6014,
"step": 643
},
{
"epoch": 0.054166579052505416,
"grad_norm": 0.4204919934272766,
"learning_rate": 5.6288423334906735e-05,
"loss": 0.6235,
"step": 644
},
{
"epoch": 0.05425068864730744,
"grad_norm": 0.6323750019073486,
"learning_rate": 5.6006083014408484e-05,
"loss": 0.7131,
"step": 645
},
{
"epoch": 0.05433479824210947,
"grad_norm": 0.4721733033657074,
"learning_rate": 5.572417689610987e-05,
"loss": 0.5972,
"step": 646
},
{
"epoch": 0.054418907836911494,
"grad_norm": 0.5242834687232971,
"learning_rate": 5.544270776231038e-05,
"loss": 0.7438,
"step": 647
},
{
"epoch": 0.05450301743171352,
"grad_norm": 0.5273514986038208,
"learning_rate": 5.5161678390996796e-05,
"loss": 0.753,
"step": 648
},
{
"epoch": 0.05458712702651555,
"grad_norm": 0.5042054653167725,
"learning_rate": 5.488109155581549e-05,
"loss": 0.5714,
"step": 649
},
{
"epoch": 0.05467123662131758,
"grad_norm": 0.6366623640060425,
"learning_rate": 5.4600950026045326e-05,
"loss": 0.6925,
"step": 650
},
{
"epoch": 0.054755346216119605,
"grad_norm": 0.4337083101272583,
"learning_rate": 5.4321256566570036e-05,
"loss": 0.49,
"step": 651
},
{
"epoch": 0.05483945581092163,
"grad_norm": 0.5946835875511169,
"learning_rate": 5.404201393785122e-05,
"loss": 0.6322,
"step": 652
},
{
"epoch": 0.05492356540572366,
"grad_norm": 0.37681615352630615,
"learning_rate": 5.3763224895900846e-05,
"loss": 0.6986,
"step": 653
},
{
"epoch": 0.05500767500052568,
"grad_norm": 0.4949415326118469,
"learning_rate": 5.348489219225416e-05,
"loss": 0.5161,
"step": 654
},
{
"epoch": 0.05509178459532771,
"grad_norm": 0.5567456483840942,
"learning_rate": 5.320701857394268e-05,
"loss": 0.6031,
"step": 655
},
{
"epoch": 0.05517589419012974,
"grad_norm": 0.6035876870155334,
"learning_rate": 5.292960678346675e-05,
"loss": 0.6714,
"step": 656
},
{
"epoch": 0.05526000378493177,
"grad_norm": 0.830230712890625,
"learning_rate": 5.265265955876879e-05,
"loss": 0.6222,
"step": 657
},
{
"epoch": 0.05534411337973379,
"grad_norm": 0.5250204801559448,
"learning_rate": 5.237617963320608e-05,
"loss": 0.7947,
"step": 658
},
{
"epoch": 0.05542822297453582,
"grad_norm": 0.5689926147460938,
"learning_rate": 5.210016973552391e-05,
"loss": 0.752,
"step": 659
},
{
"epoch": 0.055512332569337845,
"grad_norm": 0.5850744843482971,
"learning_rate": 5.182463258982846e-05,
"loss": 0.758,
"step": 660
},
{
"epoch": 0.05559644216413987,
"grad_norm": 0.8335689306259155,
"learning_rate": 5.1549570915560206e-05,
"loss": 0.8418,
"step": 661
},
{
"epoch": 0.055680551758941904,
"grad_norm": 0.5002387762069702,
"learning_rate": 5.127498742746675e-05,
"loss": 0.5669,
"step": 662
},
{
"epoch": 0.05576466135374393,
"grad_norm": 0.5952647924423218,
"learning_rate": 5.100088483557634e-05,
"loss": 0.6905,
"step": 663
},
{
"epoch": 0.055848770948545956,
"grad_norm": 0.6531146764755249,
"learning_rate": 5.072726584517086e-05,
"loss": 0.527,
"step": 664
},
{
"epoch": 0.05593288054334798,
"grad_norm": 0.48260536789894104,
"learning_rate": 5.045413315675924e-05,
"loss": 0.6368,
"step": 665
},
{
"epoch": 0.05601699013815001,
"grad_norm": 0.5707049369812012,
"learning_rate": 5.018148946605092e-05,
"loss": 0.676,
"step": 666
},
{
"epoch": 0.056101099732952034,
"grad_norm": 0.5223572850227356,
"learning_rate": 4.990933746392899e-05,
"loss": 0.6228,
"step": 667
},
{
"epoch": 0.05618520932775407,
"grad_norm": 0.6292992234230042,
"learning_rate": 4.9637679836423924e-05,
"loss": 0.5522,
"step": 668
},
{
"epoch": 0.05626931892255609,
"grad_norm": 0.6064843535423279,
"learning_rate": 4.9366519264686725e-05,
"loss": 0.6125,
"step": 669
},
{
"epoch": 0.05635342851735812,
"grad_norm": 0.6772509813308716,
"learning_rate": 4.909585842496287e-05,
"loss": 0.7981,
"step": 670
},
{
"epoch": 0.056437538112160145,
"grad_norm": 0.5350778698921204,
"learning_rate": 4.8825699988565485e-05,
"loss": 0.8081,
"step": 671
},
{
"epoch": 0.05652164770696217,
"grad_norm": 0.4920620620250702,
"learning_rate": 4.8556046621849346e-05,
"loss": 0.4888,
"step": 672
},
{
"epoch": 0.0566057573017642,
"grad_norm": 0.5023487210273743,
"learning_rate": 4.828690098618429e-05,
"loss": 0.7385,
"step": 673
},
{
"epoch": 0.05668986689656622,
"grad_norm": 0.7149183750152588,
"learning_rate": 4.8018265737929044e-05,
"loss": 0.6377,
"step": 674
},
{
"epoch": 0.056773976491368255,
"grad_norm": 0.5488125681877136,
"learning_rate": 4.7750143528405126e-05,
"loss": 0.6302,
"step": 675
},
{
"epoch": 0.05685808608617028,
"grad_norm": 0.6204077005386353,
"learning_rate": 4.748253700387042e-05,
"loss": 0.7237,
"step": 676
},
{
"epoch": 0.05694219568097231,
"grad_norm": 0.4989658296108246,
"learning_rate": 4.721544880549337e-05,
"loss": 0.6335,
"step": 677
},
{
"epoch": 0.05702630527577433,
"grad_norm": 0.582874059677124,
"learning_rate": 4.694888156932658e-05,
"loss": 0.6944,
"step": 678
},
{
"epoch": 0.05711041487057636,
"grad_norm": 0.5402207374572754,
"learning_rate": 4.668283792628114e-05,
"loss": 0.752,
"step": 679
},
{
"epoch": 0.057194524465378385,
"grad_norm": 0.46678709983825684,
"learning_rate": 4.6417320502100316e-05,
"loss": 0.625,
"step": 680
},
{
"epoch": 0.05727863406018042,
"grad_norm": 0.5934203267097473,
"learning_rate": 4.615233191733398e-05,
"loss": 0.7576,
"step": 681
},
{
"epoch": 0.057362743654982444,
"grad_norm": 0.6489574909210205,
"learning_rate": 4.588787478731242e-05,
"loss": 0.6303,
"step": 682
},
{
"epoch": 0.05744685324978447,
"grad_norm": 0.503392219543457,
"learning_rate": 4.5623951722120736e-05,
"loss": 0.6418,
"step": 683
},
{
"epoch": 0.057530962844586496,
"grad_norm": 0.7234579920768738,
"learning_rate": 4.5360565326573104e-05,
"loss": 0.8319,
"step": 684
},
{
"epoch": 0.05761507243938852,
"grad_norm": 0.6499959826469421,
"learning_rate": 4.5097718200186814e-05,
"loss": 0.6247,
"step": 685
},
{
"epoch": 0.05769918203419055,
"grad_norm": 0.5063481330871582,
"learning_rate": 4.483541293715698e-05,
"loss": 0.6584,
"step": 686
},
{
"epoch": 0.05778329162899258,
"grad_norm": 0.5390834808349609,
"learning_rate": 4.457365212633058e-05,
"loss": 0.7036,
"step": 687
},
{
"epoch": 0.05786740122379461,
"grad_norm": 0.4784635305404663,
"learning_rate": 4.431243835118124e-05,
"loss": 0.4651,
"step": 688
},
{
"epoch": 0.05795151081859663,
"grad_norm": 0.5334429740905762,
"learning_rate": 4.4051774189783315e-05,
"loss": 0.5427,
"step": 689
},
{
"epoch": 0.05803562041339866,
"grad_norm": 0.5312195420265198,
"learning_rate": 4.379166221478697e-05,
"loss": 0.6366,
"step": 690
},
{
"epoch": 0.058119730008200685,
"grad_norm": 0.6458830237388611,
"learning_rate": 4.3532104993392306e-05,
"loss": 0.7941,
"step": 691
},
{
"epoch": 0.05820383960300271,
"grad_norm": 0.5415295362472534,
"learning_rate": 4.327310508732437e-05,
"loss": 0.5574,
"step": 692
},
{
"epoch": 0.058287949197804736,
"grad_norm": 0.4901331663131714,
"learning_rate": 4.301466505280762e-05,
"loss": 0.6664,
"step": 693
},
{
"epoch": 0.05837205879260677,
"grad_norm": 0.5454357266426086,
"learning_rate": 4.2756787440540936e-05,
"loss": 0.7233,
"step": 694
},
{
"epoch": 0.058456168387408795,
"grad_norm": 0.6961559057235718,
"learning_rate": 4.249947479567218e-05,
"loss": 0.707,
"step": 695
},
{
"epoch": 0.05854027798221082,
"grad_norm": 0.6365110874176025,
"learning_rate": 4.224272965777326e-05,
"loss": 0.7353,
"step": 696
},
{
"epoch": 0.05862438757701285,
"grad_norm": 0.527192234992981,
"learning_rate": 4.1986554560815096e-05,
"loss": 0.7661,
"step": 697
},
{
"epoch": 0.05870849717181487,
"grad_norm": 0.48179832100868225,
"learning_rate": 4.173095203314241e-05,
"loss": 0.665,
"step": 698
},
{
"epoch": 0.0587926067666169,
"grad_norm": 0.6591465473175049,
"learning_rate": 4.1475924597449024e-05,
"loss": 0.7179,
"step": 699
},
{
"epoch": 0.05887671636141893,
"grad_norm": 0.4673934876918793,
"learning_rate": 4.12214747707527e-05,
"loss": 0.6477,
"step": 700
},
{
"epoch": 0.05896082595622096,
"grad_norm": 0.8560356497764587,
"learning_rate": 4.096760506437057e-05,
"loss": 0.4083,
"step": 701
},
{
"epoch": 0.059044935551022984,
"grad_norm": 0.5286902189254761,
"learning_rate": 4.071431798389408e-05,
"loss": 0.7202,
"step": 702
},
{
"epoch": 0.05912904514582501,
"grad_norm": 0.5945466160774231,
"learning_rate": 4.0461616029164526e-05,
"loss": 0.8511,
"step": 703
},
{
"epoch": 0.059213154740627036,
"grad_norm": 0.7009603381156921,
"learning_rate": 4.020950169424815e-05,
"loss": 0.8503,
"step": 704
},
{
"epoch": 0.05929726433542906,
"grad_norm": 0.5323594808578491,
"learning_rate": 3.9957977467411615e-05,
"loss": 0.6064,
"step": 705
},
{
"epoch": 0.05938137393023109,
"grad_norm": 0.4955034852027893,
"learning_rate": 3.9707045831097555e-05,
"loss": 0.6115,
"step": 706
},
{
"epoch": 0.05946548352503312,
"grad_norm": 0.5311620831489563,
"learning_rate": 3.945670926189987e-05,
"loss": 0.6827,
"step": 707
},
{
"epoch": 0.05954959311983515,
"grad_norm": 0.432479590177536,
"learning_rate": 3.920697023053949e-05,
"loss": 0.5194,
"step": 708
},
{
"epoch": 0.05963370271463717,
"grad_norm": 0.544047474861145,
"learning_rate": 3.895783120183976e-05,
"loss": 0.6654,
"step": 709
},
{
"epoch": 0.0597178123094392,
"grad_norm": 0.4839647114276886,
"learning_rate": 3.8709294634702376e-05,
"loss": 0.5999,
"step": 710
},
{
"epoch": 0.059801921904241225,
"grad_norm": 0.6779318451881409,
"learning_rate": 3.846136298208285e-05,
"loss": 0.655,
"step": 711
},
{
"epoch": 0.05988603149904325,
"grad_norm": 0.5102113485336304,
"learning_rate": 3.821403869096658e-05,
"loss": 0.6759,
"step": 712
},
{
"epoch": 0.05997014109384528,
"grad_norm": 0.5493488311767578,
"learning_rate": 3.796732420234443e-05,
"loss": 0.5339,
"step": 713
},
{
"epoch": 0.06005425068864731,
"grad_norm": 0.5925408005714417,
"learning_rate": 3.7721221951188765e-05,
"loss": 0.751,
"step": 714
},
{
"epoch": 0.060138360283449335,
"grad_norm": 0.6988538503646851,
"learning_rate": 3.747573436642951e-05,
"loss": 0.759,
"step": 715
},
{
"epoch": 0.06022246987825136,
"grad_norm": 0.5451468825340271,
"learning_rate": 3.7230863870929964e-05,
"loss": 0.5549,
"step": 716
},
{
"epoch": 0.06030657947305339,
"grad_norm": 0.5933196544647217,
"learning_rate": 3.698661288146311e-05,
"loss": 0.5338,
"step": 717
},
{
"epoch": 0.06039068906785541,
"grad_norm": 0.5368034839630127,
"learning_rate": 3.674298380868756e-05,
"loss": 0.6672,
"step": 718
},
{
"epoch": 0.060474798662657446,
"grad_norm": 0.6387892961502075,
"learning_rate": 3.649997905712396e-05,
"loss": 0.6483,
"step": 719
},
{
"epoch": 0.06055890825745947,
"grad_norm": 0.5074757933616638,
"learning_rate": 3.6257601025131026e-05,
"loss": 0.5061,
"step": 720
},
{
"epoch": 0.0606430178522615,
"grad_norm": 0.7502937912940979,
"learning_rate": 3.601585210488218e-05,
"loss": 0.7917,
"step": 721
},
{
"epoch": 0.060727127447063524,
"grad_norm": 0.5781522989273071,
"learning_rate": 3.577473468234156e-05,
"loss": 0.825,
"step": 722
},
{
"epoch": 0.06081123704186555,
"grad_norm": 0.559719443321228,
"learning_rate": 3.553425113724088e-05,
"loss": 0.5209,
"step": 723
},
{
"epoch": 0.060895346636667576,
"grad_norm": 0.714248538017273,
"learning_rate": 3.52944038430556e-05,
"loss": 0.5418,
"step": 724
},
{
"epoch": 0.0609794562314696,
"grad_norm": 0.5317071080207825,
"learning_rate": 3.5055195166981645e-05,
"loss": 0.6363,
"step": 725
},
{
"epoch": 0.061063565826271635,
"grad_norm": 0.5349583625793457,
"learning_rate": 3.481662746991214e-05,
"loss": 0.5828,
"step": 726
},
{
"epoch": 0.06114767542107366,
"grad_norm": 0.5363706350326538,
"learning_rate": 3.4578703106413904e-05,
"loss": 0.5334,
"step": 727
},
{
"epoch": 0.06123178501587569,
"grad_norm": 0.4438410699367523,
"learning_rate": 3.4341424424704375e-05,
"loss": 0.7108,
"step": 728
},
{
"epoch": 0.06131589461067771,
"grad_norm": 0.6724744439125061,
"learning_rate": 3.4104793766628304e-05,
"loss": 0.6905,
"step": 729
},
{
"epoch": 0.06140000420547974,
"grad_norm": 0.5650711059570312,
"learning_rate": 3.386881346763483e-05,
"loss": 0.5404,
"step": 730
},
{
"epoch": 0.061484113800281764,
"grad_norm": 0.6993920803070068,
"learning_rate": 3.363348585675414e-05,
"loss": 0.6725,
"step": 731
},
{
"epoch": 0.0615682233950838,
"grad_norm": 0.5407055616378784,
"learning_rate": 3.339881325657484e-05,
"loss": 0.6004,
"step": 732
},
{
"epoch": 0.06165233298988582,
"grad_norm": 0.5904746055603027,
"learning_rate": 3.316479798322072e-05,
"loss": 0.6763,
"step": 733
},
{
"epoch": 0.06173644258468785,
"grad_norm": 0.5810348987579346,
"learning_rate": 3.2931442346328004e-05,
"loss": 0.8605,
"step": 734
},
{
"epoch": 0.061820552179489875,
"grad_norm": 0.48206856846809387,
"learning_rate": 3.269874864902269e-05,
"loss": 0.6533,
"step": 735
},
{
"epoch": 0.0619046617742919,
"grad_norm": 0.5678583383560181,
"learning_rate": 3.246671918789755e-05,
"loss": 0.5872,
"step": 736
},
{
"epoch": 0.06198877136909393,
"grad_norm": 0.5930237174034119,
"learning_rate": 3.223535625298979e-05,
"loss": 0.7073,
"step": 737
},
{
"epoch": 0.06207288096389595,
"grad_norm": 0.4858033061027527,
"learning_rate": 3.200466212775808e-05,
"loss": 0.5829,
"step": 738
},
{
"epoch": 0.062156990558697986,
"grad_norm": 0.6299439668655396,
"learning_rate": 3.1774639089060363e-05,
"loss": 0.4702,
"step": 739
},
{
"epoch": 0.06224110015350001,
"grad_norm": 0.5381211042404175,
"learning_rate": 3.154528940713113e-05,
"loss": 0.7213,
"step": 740
},
{
"epoch": 0.06232520974830204,
"grad_norm": 0.4895922839641571,
"learning_rate": 3.1316615345559185e-05,
"loss": 0.571,
"step": 741
},
{
"epoch": 0.062409319343104064,
"grad_norm": 0.507371723651886,
"learning_rate": 3.108861916126518e-05,
"loss": 0.6783,
"step": 742
},
{
"epoch": 0.06249342893790609,
"grad_norm": 0.5576531291007996,
"learning_rate": 3.086130310447937e-05,
"loss": 0.6561,
"step": 743
},
{
"epoch": 0.06257753853270812,
"grad_norm": 0.42016541957855225,
"learning_rate": 3.063466941871952e-05,
"loss": 0.5153,
"step": 744
},
{
"epoch": 0.06266164812751014,
"grad_norm": 0.7200675010681152,
"learning_rate": 3.0408720340768572e-05,
"loss": 0.6269,
"step": 745
},
{
"epoch": 0.06274575772231217,
"grad_norm": 0.5375575423240662,
"learning_rate": 3.018345810065275e-05,
"loss": 0.5603,
"step": 746
},
{
"epoch": 0.0628298673171142,
"grad_norm": 0.45930182933807373,
"learning_rate": 2.9958884921619367e-05,
"loss": 0.7863,
"step": 747
},
{
"epoch": 0.06291397691191623,
"grad_norm": 0.5124622583389282,
"learning_rate": 2.9735003020115092e-05,
"loss": 0.7125,
"step": 748
},
{
"epoch": 0.06299808650671826,
"grad_norm": 0.5122153162956238,
"learning_rate": 2.9511814605763855e-05,
"loss": 0.7104,
"step": 749
},
{
"epoch": 0.06308219610152029,
"grad_norm": 0.56404048204422,
"learning_rate": 2.9289321881345254e-05,
"loss": 0.6569,
"step": 750
},
{
"epoch": 0.06316630569632231,
"grad_norm": 0.4476301968097687,
"learning_rate": 2.9067527042772636e-05,
"loss": 0.6703,
"step": 751
},
{
"epoch": 0.06325041529112434,
"grad_norm": 0.432229220867157,
"learning_rate": 2.8846432279071467e-05,
"loss": 0.7372,
"step": 752
},
{
"epoch": 0.06333452488592636,
"grad_norm": 0.5445795059204102,
"learning_rate": 2.8626039772357882e-05,
"loss": 0.5326,
"step": 753
},
{
"epoch": 0.06341863448072839,
"grad_norm": 0.645452618598938,
"learning_rate": 2.840635169781688e-05,
"loss": 0.3862,
"step": 754
},
{
"epoch": 0.06350274407553042,
"grad_norm": 0.4100976288318634,
"learning_rate": 2.8187370223681132e-05,
"loss": 0.8218,
"step": 755
},
{
"epoch": 0.06358685367033244,
"grad_norm": 0.7901608347892761,
"learning_rate": 2.7969097511209308e-05,
"loss": 0.6172,
"step": 756
},
{
"epoch": 0.06367096326513447,
"grad_norm": 0.4609740376472473,
"learning_rate": 2.775153571466502e-05,
"loss": 0.7415,
"step": 757
},
{
"epoch": 0.06375507285993649,
"grad_norm": 0.684585452079773,
"learning_rate": 2.753468698129533e-05,
"loss": 0.696,
"step": 758
},
{
"epoch": 0.06383918245473852,
"grad_norm": 0.5767757296562195,
"learning_rate": 2.7318553451309726e-05,
"loss": 0.6223,
"step": 759
},
{
"epoch": 0.06392329204954054,
"grad_norm": 0.5188784003257751,
"learning_rate": 2.7103137257858868e-05,
"loss": 0.7564,
"step": 760
},
{
"epoch": 0.06400740164434258,
"grad_norm": 0.6304595470428467,
"learning_rate": 2.688844052701359e-05,
"loss": 0.7426,
"step": 761
},
{
"epoch": 0.06409151123914461,
"grad_norm": 0.5373560190200806,
"learning_rate": 2.6674465377744017e-05,
"loss": 0.7388,
"step": 762
},
{
"epoch": 0.06417562083394664,
"grad_norm": 0.5632065534591675,
"learning_rate": 2.646121392189841e-05,
"loss": 0.5892,
"step": 763
},
{
"epoch": 0.06425973042874866,
"grad_norm": 0.5747659802436829,
"learning_rate": 2.624868826418262e-05,
"loss": 0.4149,
"step": 764
},
{
"epoch": 0.06434384002355069,
"grad_norm": 0.7094728946685791,
"learning_rate": 2.603689050213902e-05,
"loss": 0.6391,
"step": 765
},
{
"epoch": 0.06442794961835271,
"grad_norm": 0.6141682863235474,
"learning_rate": 2.582582272612609e-05,
"loss": 0.5478,
"step": 766
},
{
"epoch": 0.06451205921315474,
"grad_norm": 0.5336157083511353,
"learning_rate": 2.561548701929749e-05,
"loss": 0.6272,
"step": 767
},
{
"epoch": 0.06459616880795677,
"grad_norm": 0.5744621753692627,
"learning_rate": 2.540588545758179e-05,
"loss": 0.6357,
"step": 768
},
{
"epoch": 0.06468027840275879,
"grad_norm": 0.5280835628509521,
"learning_rate": 2.5197020109661772e-05,
"loss": 0.6744,
"step": 769
},
{
"epoch": 0.06476438799756082,
"grad_norm": 0.5798782110214233,
"learning_rate": 2.4988893036954043e-05,
"loss": 0.9299,
"step": 770
},
{
"epoch": 0.06484849759236284,
"grad_norm": 0.7400964498519897,
"learning_rate": 2.4781506293588873e-05,
"loss": 0.7047,
"step": 771
},
{
"epoch": 0.06493260718716487,
"grad_norm": 0.4643491208553314,
"learning_rate": 2.4574861926389615e-05,
"loss": 0.5174,
"step": 772
},
{
"epoch": 0.0650167167819669,
"grad_norm": 0.4251709580421448,
"learning_rate": 2.436896197485282e-05,
"loss": 0.6295,
"step": 773
},
{
"epoch": 0.06510082637676894,
"grad_norm": 0.48561739921569824,
"learning_rate": 2.4163808471127812e-05,
"loss": 0.6543,
"step": 774
},
{
"epoch": 0.06518493597157096,
"grad_norm": 0.6305271983146667,
"learning_rate": 2.3959403439996907e-05,
"loss": 0.5729,
"step": 775
},
{
"epoch": 0.06526904556637299,
"grad_norm": 0.6219698190689087,
"learning_rate": 2.37557488988552e-05,
"loss": 0.7075,
"step": 776
},
{
"epoch": 0.06535315516117501,
"grad_norm": 0.5188342332839966,
"learning_rate": 2.3552846857690846e-05,
"loss": 0.7433,
"step": 777
},
{
"epoch": 0.06543726475597704,
"grad_norm": 1.439207673072815,
"learning_rate": 2.3350699319065026e-05,
"loss": 0.8646,
"step": 778
},
{
"epoch": 0.06552137435077907,
"grad_norm": 0.5281020402908325,
"learning_rate": 2.3149308278092342e-05,
"loss": 0.7782,
"step": 779
},
{
"epoch": 0.06560548394558109,
"grad_norm": 0.5641273260116577,
"learning_rate": 2.2948675722421086e-05,
"loss": 0.8682,
"step": 780
},
{
"epoch": 0.06568959354038312,
"grad_norm": 0.6752006411552429,
"learning_rate": 2.2748803632213557e-05,
"loss": 0.652,
"step": 781
},
{
"epoch": 0.06577370313518514,
"grad_norm": 0.48423561453819275,
"learning_rate": 2.254969398012663e-05,
"loss": 0.3943,
"step": 782
},
{
"epoch": 0.06585781272998717,
"grad_norm": 0.5390269756317139,
"learning_rate": 2.235134873129213e-05,
"loss": 0.657,
"step": 783
},
{
"epoch": 0.0659419223247892,
"grad_norm": 0.469672828912735,
"learning_rate": 2.2153769843297667e-05,
"loss": 0.5305,
"step": 784
},
{
"epoch": 0.06602603191959122,
"grad_norm": 0.6228986978530884,
"learning_rate": 2.195695926616702e-05,
"loss": 0.546,
"step": 785
},
{
"epoch": 0.06611014151439325,
"grad_norm": 0.537236750125885,
"learning_rate": 2.1760918942341192e-05,
"loss": 0.7306,
"step": 786
},
{
"epoch": 0.06619425110919529,
"grad_norm": 0.48184770345687866,
"learning_rate": 2.1565650806658975e-05,
"loss": 0.6454,
"step": 787
},
{
"epoch": 0.06627836070399731,
"grad_norm": 0.6069773435592651,
"learning_rate": 2.137115678633811e-05,
"loss": 0.6599,
"step": 788
},
{
"epoch": 0.06636247029879934,
"grad_norm": 0.5169489979743958,
"learning_rate": 2.1177438800956007e-05,
"loss": 0.5624,
"step": 789
},
{
"epoch": 0.06644657989360137,
"grad_norm": 0.4128231704235077,
"learning_rate": 2.098449876243096e-05,
"loss": 0.5663,
"step": 790
},
{
"epoch": 0.06653068948840339,
"grad_norm": 0.715587854385376,
"learning_rate": 2.07923385750033e-05,
"loss": 0.7287,
"step": 791
},
{
"epoch": 0.06661479908320542,
"grad_norm": 0.44507458806037903,
"learning_rate": 2.0600960135216462e-05,
"loss": 0.6717,
"step": 792
},
{
"epoch": 0.06669890867800744,
"grad_norm": 0.5170130729675293,
"learning_rate": 2.0410365331898416e-05,
"loss": 0.6406,
"step": 793
},
{
"epoch": 0.06678301827280947,
"grad_norm": 0.46590495109558105,
"learning_rate": 2.0220556046142893e-05,
"loss": 0.6061,
"step": 794
},
{
"epoch": 0.0668671278676115,
"grad_norm": 0.5688216686248779,
"learning_rate": 2.0031534151290943e-05,
"loss": 0.7645,
"step": 795
},
{
"epoch": 0.06695123746241352,
"grad_norm": 0.6071348786354065,
"learning_rate": 1.9843301512912327e-05,
"loss": 0.8112,
"step": 796
},
{
"epoch": 0.06703534705721555,
"grad_norm": 0.639324963092804,
"learning_rate": 1.965585998878724e-05,
"loss": 0.683,
"step": 797
},
{
"epoch": 0.06711945665201757,
"grad_norm": 0.5158023238182068,
"learning_rate": 1.946921142888781e-05,
"loss": 0.6532,
"step": 798
},
{
"epoch": 0.06720356624681961,
"grad_norm": 0.5088067650794983,
"learning_rate": 1.928335767535997e-05,
"loss": 0.6117,
"step": 799
},
{
"epoch": 0.06728767584162164,
"grad_norm": 0.4564967155456543,
"learning_rate": 1.9098300562505266e-05,
"loss": 0.6276,
"step": 800
},
{
"epoch": 0.06737178543642366,
"grad_norm": 0.620216429233551,
"learning_rate": 1.891404191676265e-05,
"loss": 0.655,
"step": 801
},
{
"epoch": 0.06745589503122569,
"grad_norm": 0.698749303817749,
"learning_rate": 1.8730583556690605e-05,
"loss": 0.5483,
"step": 802
},
{
"epoch": 0.06754000462602772,
"grad_norm": 0.48406967520713806,
"learning_rate": 1.854792729294905e-05,
"loss": 0.7465,
"step": 803
},
{
"epoch": 0.06762411422082974,
"grad_norm": 0.5097486972808838,
"learning_rate": 1.8366074928281607e-05,
"loss": 0.4625,
"step": 804
},
{
"epoch": 0.06770822381563177,
"grad_norm": 0.5742330551147461,
"learning_rate": 1.818502825749764e-05,
"loss": 0.6649,
"step": 805
},
{
"epoch": 0.0677923334104338,
"grad_norm": 0.49662643671035767,
"learning_rate": 1.8004789067454764e-05,
"loss": 0.7763,
"step": 806
},
{
"epoch": 0.06787644300523582,
"grad_norm": 0.6889787912368774,
"learning_rate": 1.7825359137040988e-05,
"loss": 0.6646,
"step": 807
},
{
"epoch": 0.06796055260003785,
"grad_norm": 0.47545164823532104,
"learning_rate": 1.7646740237157256e-05,
"loss": 0.6404,
"step": 808
},
{
"epoch": 0.06804466219483987,
"grad_norm": 0.5143394470214844,
"learning_rate": 1.7468934130700044e-05,
"loss": 0.6727,
"step": 809
},
{
"epoch": 0.0681287717896419,
"grad_norm": 0.5231197476387024,
"learning_rate": 1.7291942572543807e-05,
"loss": 0.6471,
"step": 810
},
{
"epoch": 0.06821288138444392,
"grad_norm": 0.5731273293495178,
"learning_rate": 1.7115767309523812e-05,
"loss": 0.7455,
"step": 811
},
{
"epoch": 0.06829699097924596,
"grad_norm": 0.5894562005996704,
"learning_rate": 1.6940410080418723e-05,
"loss": 0.6111,
"step": 812
},
{
"epoch": 0.06838110057404799,
"grad_norm": 0.5060394406318665,
"learning_rate": 1.6765872615933677e-05,
"loss": 0.5798,
"step": 813
},
{
"epoch": 0.06846521016885002,
"grad_norm": 0.5075900554656982,
"learning_rate": 1.6592156638682886e-05,
"loss": 0.6851,
"step": 814
},
{
"epoch": 0.06854931976365204,
"grad_norm": 0.40963247418403625,
"learning_rate": 1.6419263863172997e-05,
"loss": 0.53,
"step": 815
},
{
"epoch": 0.06863342935845407,
"grad_norm": 0.5541537404060364,
"learning_rate": 1.6247195995785837e-05,
"loss": 0.8987,
"step": 816
},
{
"epoch": 0.0687175389532561,
"grad_norm": 0.6535544395446777,
"learning_rate": 1.6075954734761845e-05,
"loss": 0.5659,
"step": 817
},
{
"epoch": 0.06880164854805812,
"grad_norm": 0.4968286454677582,
"learning_rate": 1.5905541770183096e-05,
"loss": 0.6928,
"step": 818
},
{
"epoch": 0.06888575814286015,
"grad_norm": 0.852871835231781,
"learning_rate": 1.5735958783956794e-05,
"loss": 0.5483,
"step": 819
},
{
"epoch": 0.06896986773766217,
"grad_norm": 0.6427825689315796,
"learning_rate": 1.5567207449798515e-05,
"loss": 0.5945,
"step": 820
},
{
"epoch": 0.0690539773324642,
"grad_norm": 0.6467059850692749,
"learning_rate": 1.539928943321579e-05,
"loss": 0.5874,
"step": 821
},
{
"epoch": 0.06913808692726622,
"grad_norm": 0.5707730054855347,
"learning_rate": 1.5232206391491699e-05,
"loss": 0.616,
"step": 822
},
{
"epoch": 0.06922219652206825,
"grad_norm": 0.6659408211708069,
"learning_rate": 1.5065959973668353e-05,
"loss": 0.782,
"step": 823
},
{
"epoch": 0.06930630611687028,
"grad_norm": 0.5345143675804138,
"learning_rate": 1.4900551820530828e-05,
"loss": 0.7579,
"step": 824
},
{
"epoch": 0.06939041571167232,
"grad_norm": 0.5110843777656555,
"learning_rate": 1.4735983564590783e-05,
"loss": 0.778,
"step": 825
},
{
"epoch": 0.06947452530647434,
"grad_norm": 0.4799653887748718,
"learning_rate": 1.4572256830070497e-05,
"loss": 0.6927,
"step": 826
},
{
"epoch": 0.06955863490127637,
"grad_norm": 0.5991430282592773,
"learning_rate": 1.4409373232886702e-05,
"loss": 0.6772,
"step": 827
},
{
"epoch": 0.0696427444960784,
"grad_norm": 0.6813725233078003,
"learning_rate": 1.4247334380634792e-05,
"loss": 0.6461,
"step": 828
},
{
"epoch": 0.06972685409088042,
"grad_norm": 0.6396465301513672,
"learning_rate": 1.4086141872572789e-05,
"loss": 0.7156,
"step": 829
},
{
"epoch": 0.06981096368568245,
"grad_norm": 0.6053589582443237,
"learning_rate": 1.3925797299605647e-05,
"loss": 0.8174,
"step": 830
},
{
"epoch": 0.06989507328048447,
"grad_norm": 0.5992446541786194,
"learning_rate": 1.3766302244269624e-05,
"loss": 0.7147,
"step": 831
},
{
"epoch": 0.0699791828752865,
"grad_norm": 0.5334646701812744,
"learning_rate": 1.3607658280716473e-05,
"loss": 0.5372,
"step": 832
},
{
"epoch": 0.07006329247008852,
"grad_norm": 0.4751766622066498,
"learning_rate": 1.3449866974698122e-05,
"loss": 0.4352,
"step": 833
},
{
"epoch": 0.07014740206489055,
"grad_norm": 0.4926972985267639,
"learning_rate": 1.3292929883550998e-05,
"loss": 0.6821,
"step": 834
},
{
"epoch": 0.07023151165969257,
"grad_norm": 0.4819171726703644,
"learning_rate": 1.3136848556180892e-05,
"loss": 0.6051,
"step": 835
},
{
"epoch": 0.0703156212544946,
"grad_norm": 0.48462870717048645,
"learning_rate": 1.2981624533047432e-05,
"loss": 0.5676,
"step": 836
},
{
"epoch": 0.07039973084929663,
"grad_norm": 0.5962949395179749,
"learning_rate": 1.2827259346149122e-05,
"loss": 0.5559,
"step": 837
},
{
"epoch": 0.07048384044409867,
"grad_norm": 0.5228509306907654,
"learning_rate": 1.2673754519008008e-05,
"loss": 0.6072,
"step": 838
},
{
"epoch": 0.07056795003890069,
"grad_norm": 0.6193320751190186,
"learning_rate": 1.2521111566654731e-05,
"loss": 0.6156,
"step": 839
},
{
"epoch": 0.07065205963370272,
"grad_norm": 0.48941028118133545,
"learning_rate": 1.2369331995613665e-05,
"loss": 0.5834,
"step": 840
},
{
"epoch": 0.07073616922850474,
"grad_norm": 0.5763342380523682,
"learning_rate": 1.2218417303887842e-05,
"loss": 0.6142,
"step": 841
},
{
"epoch": 0.07082027882330677,
"grad_norm": 0.5619479417800903,
"learning_rate": 1.206836898094439e-05,
"loss": 0.5918,
"step": 842
},
{
"epoch": 0.0709043884181088,
"grad_norm": 0.4731215536594391,
"learning_rate": 1.191918850769964e-05,
"loss": 0.4884,
"step": 843
},
{
"epoch": 0.07098849801291082,
"grad_norm": 0.7118493318557739,
"learning_rate": 1.1770877356504683e-05,
"loss": 0.6499,
"step": 844
},
{
"epoch": 0.07107260760771285,
"grad_norm": 0.7185307741165161,
"learning_rate": 1.1623436991130654e-05,
"loss": 0.7735,
"step": 845
},
{
"epoch": 0.07115671720251487,
"grad_norm": 0.5481952428817749,
"learning_rate": 1.1476868866754486e-05,
"loss": 0.919,
"step": 846
},
{
"epoch": 0.0712408267973169,
"grad_norm": 0.596553385257721,
"learning_rate": 1.1331174429944347e-05,
"loss": 0.6963,
"step": 847
},
{
"epoch": 0.07132493639211893,
"grad_norm": 0.5250246524810791,
"learning_rate": 1.1186355118645554e-05,
"loss": 0.6193,
"step": 848
},
{
"epoch": 0.07140904598692095,
"grad_norm": 0.5260514616966248,
"learning_rate": 1.1042412362166222e-05,
"loss": 0.6361,
"step": 849
},
{
"epoch": 0.07149315558172299,
"grad_norm": 0.5367104411125183,
"learning_rate": 1.0899347581163221e-05,
"loss": 0.5347,
"step": 850
},
{
"epoch": 0.07157726517652502,
"grad_norm": 0.5701801776885986,
"learning_rate": 1.0757162187628222e-05,
"loss": 0.6477,
"step": 851
},
{
"epoch": 0.07166137477132704,
"grad_norm": 0.494419127702713,
"learning_rate": 1.0615857584873623e-05,
"loss": 0.6831,
"step": 852
},
{
"epoch": 0.07174548436612907,
"grad_norm": 0.6777026057243347,
"learning_rate": 1.0475435167518843e-05,
"loss": 0.6998,
"step": 853
},
{
"epoch": 0.0718295939609311,
"grad_norm": 0.6337087154388428,
"learning_rate": 1.0335896321476413e-05,
"loss": 0.6884,
"step": 854
},
{
"epoch": 0.07191370355573312,
"grad_norm": 0.508190393447876,
"learning_rate": 1.0197242423938446e-05,
"loss": 0.6033,
"step": 855
},
{
"epoch": 0.07199781315053515,
"grad_norm": 0.588848888874054,
"learning_rate": 1.0059474843362892e-05,
"loss": 0.6572,
"step": 856
},
{
"epoch": 0.07208192274533717,
"grad_norm": 0.6991647481918335,
"learning_rate": 9.922594939460194e-06,
"loss": 0.5044,
"step": 857
},
{
"epoch": 0.0721660323401392,
"grad_norm": 0.6183068752288818,
"learning_rate": 9.786604063179728e-06,
"loss": 0.5294,
"step": 858
},
{
"epoch": 0.07225014193494123,
"grad_norm": 0.6094670295715332,
"learning_rate": 9.651503556696516e-06,
"loss": 0.4282,
"step": 859
},
{
"epoch": 0.07233425152974325,
"grad_norm": 0.5989966988563538,
"learning_rate": 9.517294753398064e-06,
"loss": 0.6284,
"step": 860
},
{
"epoch": 0.07241836112454528,
"grad_norm": 0.5924322009086609,
"learning_rate": 9.383978977871021e-06,
"loss": 0.4946,
"step": 861
},
{
"epoch": 0.0725024707193473,
"grad_norm": 0.6521950364112854,
"learning_rate": 9.251557545888312e-06,
"loss": 0.6351,
"step": 862
},
{
"epoch": 0.07258658031414934,
"grad_norm": 0.5028837323188782,
"learning_rate": 9.120031764395987e-06,
"loss": 0.6651,
"step": 863
},
{
"epoch": 0.07267068990895137,
"grad_norm": 0.5665683746337891,
"learning_rate": 8.989402931500434e-06,
"loss": 0.6975,
"step": 864
},
{
"epoch": 0.0727547995037534,
"grad_norm": 0.6868077516555786,
"learning_rate": 8.85967233645547e-06,
"loss": 0.6654,
"step": 865
},
{
"epoch": 0.07283890909855542,
"grad_norm": 0.4886159300804138,
"learning_rate": 8.730841259649725e-06,
"loss": 0.6435,
"step": 866
},
{
"epoch": 0.07292301869335745,
"grad_norm": 0.5298961997032166,
"learning_rate": 8.602910972593892e-06,
"loss": 0.6013,
"step": 867
},
{
"epoch": 0.07300712828815947,
"grad_norm": 0.5650634765625,
"learning_rate": 8.475882737908248e-06,
"loss": 0.6252,
"step": 868
},
{
"epoch": 0.0730912378829615,
"grad_norm": 0.6334890723228455,
"learning_rate": 8.34975780931021e-06,
"loss": 0.5932,
"step": 869
},
{
"epoch": 0.07317534747776352,
"grad_norm": 0.42107874155044556,
"learning_rate": 8.224537431601886e-06,
"loss": 0.7897,
"step": 870
},
{
"epoch": 0.07325945707256555,
"grad_norm": 0.4216519594192505,
"learning_rate": 8.100222840657878e-06,
"loss": 0.671,
"step": 871
},
{
"epoch": 0.07334356666736758,
"grad_norm": 0.4145600497722626,
"learning_rate": 7.976815263412963e-06,
"loss": 0.5159,
"step": 872
},
{
"epoch": 0.0734276762621696,
"grad_norm": 0.6570865511894226,
"learning_rate": 7.854315917850163e-06,
"loss": 0.685,
"step": 873
},
{
"epoch": 0.07351178585697163,
"grad_norm": 0.42912426590919495,
"learning_rate": 7.73272601298851e-06,
"loss": 0.5689,
"step": 874
},
{
"epoch": 0.07359589545177365,
"grad_norm": 0.5340107679367065,
"learning_rate": 7.612046748871327e-06,
"loss": 0.6738,
"step": 875
},
{
"epoch": 0.0736800050465757,
"grad_norm": 0.6763095855712891,
"learning_rate": 7.492279316554207e-06,
"loss": 0.6679,
"step": 876
},
{
"epoch": 0.07376411464137772,
"grad_norm": 0.6157491207122803,
"learning_rate": 7.3734248980933395e-06,
"loss": 0.7077,
"step": 877
},
{
"epoch": 0.07384822423617975,
"grad_norm": 0.6420180797576904,
"learning_rate": 7.255484666533874e-06,
"loss": 0.6328,
"step": 878
},
{
"epoch": 0.07393233383098177,
"grad_norm": 0.7011188268661499,
"learning_rate": 7.138459785898266e-06,
"loss": 0.6339,
"step": 879
},
{
"epoch": 0.0740164434257838,
"grad_norm": 0.5735622048377991,
"learning_rate": 7.022351411174866e-06,
"loss": 0.6621,
"step": 880
},
{
"epoch": 0.07410055302058582,
"grad_norm": 0.49764999747276306,
"learning_rate": 6.907160688306425e-06,
"loss": 0.5328,
"step": 881
},
{
"epoch": 0.07418466261538785,
"grad_norm": 0.6401441097259521,
"learning_rate": 6.7928887541789055e-06,
"loss": 0.6566,
"step": 882
},
{
"epoch": 0.07426877221018988,
"grad_norm": 0.44652143120765686,
"learning_rate": 6.679536736610137e-06,
"loss": 0.5332,
"step": 883
},
{
"epoch": 0.0743528818049919,
"grad_norm": 0.6002373695373535,
"learning_rate": 6.5671057543387985e-06,
"loss": 0.7656,
"step": 884
},
{
"epoch": 0.07443699139979393,
"grad_norm": 0.6372143030166626,
"learning_rate": 6.455596917013273e-06,
"loss": 0.7022,
"step": 885
},
{
"epoch": 0.07452110099459595,
"grad_norm": 0.4614960253238678,
"learning_rate": 6.345011325180772e-06,
"loss": 0.62,
"step": 886
},
{
"epoch": 0.07460521058939798,
"grad_norm": 0.6100625395774841,
"learning_rate": 6.235350070276447e-06,
"loss": 0.7091,
"step": 887
},
{
"epoch": 0.0746893201842,
"grad_norm": 0.6311874985694885,
"learning_rate": 6.126614234612593e-06,
"loss": 0.6988,
"step": 888
},
{
"epoch": 0.07477342977900205,
"grad_norm": 0.6123420000076294,
"learning_rate": 6.018804891368035e-06,
"loss": 0.6147,
"step": 889
},
{
"epoch": 0.07485753937380407,
"grad_norm": 0.5420098304748535,
"learning_rate": 5.911923104577455e-06,
"loss": 0.6469,
"step": 890
},
{
"epoch": 0.0749416489686061,
"grad_norm": 0.5997559428215027,
"learning_rate": 5.805969929120947e-06,
"loss": 0.6006,
"step": 891
},
{
"epoch": 0.07502575856340812,
"grad_norm": 0.801465630531311,
"learning_rate": 5.700946410713548e-06,
"loss": 0.7753,
"step": 892
},
{
"epoch": 0.07510986815821015,
"grad_norm": 0.5048848986625671,
"learning_rate": 5.5968535858950345e-06,
"loss": 0.6048,
"step": 893
},
{
"epoch": 0.07519397775301218,
"grad_norm": 0.6630212664604187,
"learning_rate": 5.49369248201953e-06,
"loss": 0.8072,
"step": 894
},
{
"epoch": 0.0752780873478142,
"grad_norm": 0.7513829469680786,
"learning_rate": 5.39146411724547e-06,
"loss": 0.6736,
"step": 895
},
{
"epoch": 0.07536219694261623,
"grad_norm": 1.04941987991333,
"learning_rate": 5.290169500525577e-06,
"loss": 0.6241,
"step": 896
},
{
"epoch": 0.07544630653741825,
"grad_norm": 0.4719288647174835,
"learning_rate": 5.189809631596798e-06,
"loss": 0.5016,
"step": 897
},
{
"epoch": 0.07553041613222028,
"grad_norm": 0.5207799673080444,
"learning_rate": 5.0903855009705514e-06,
"loss": 0.6855,
"step": 898
},
{
"epoch": 0.0756145257270223,
"grad_norm": 0.5883520245552063,
"learning_rate": 4.991898089922819e-06,
"loss": 0.841,
"step": 899
},
{
"epoch": 0.07569863532182433,
"grad_norm": 0.48410382866859436,
"learning_rate": 4.8943483704846475e-06,
"loss": 0.7806,
"step": 900
},
{
"epoch": 0.07578274491662636,
"grad_norm": 0.5785539150238037,
"learning_rate": 4.797737305432337e-06,
"loss": 0.7894,
"step": 901
},
{
"epoch": 0.0758668545114284,
"grad_norm": 0.6412384510040283,
"learning_rate": 4.702065848278126e-06,
"loss": 0.5901,
"step": 902
},
{
"epoch": 0.07595096410623042,
"grad_norm": 0.6484128832817078,
"learning_rate": 4.607334943260655e-06,
"loss": 0.5787,
"step": 903
},
{
"epoch": 0.07603507370103245,
"grad_norm": 0.6749901175498962,
"learning_rate": 4.513545525335705e-06,
"loss": 0.6713,
"step": 904
},
{
"epoch": 0.07611918329583447,
"grad_norm": 0.6375152468681335,
"learning_rate": 4.420698520166988e-06,
"loss": 0.5882,
"step": 905
},
{
"epoch": 0.0762032928906365,
"grad_norm": 0.5309505462646484,
"learning_rate": 4.328794844116946e-06,
"loss": 0.6259,
"step": 906
},
{
"epoch": 0.07628740248543853,
"grad_norm": 0.5349912047386169,
"learning_rate": 4.237835404237778e-06,
"loss": 0.7206,
"step": 907
},
{
"epoch": 0.07637151208024055,
"grad_norm": 0.5662460327148438,
"learning_rate": 4.147821098262405e-06,
"loss": 0.8227,
"step": 908
},
{
"epoch": 0.07645562167504258,
"grad_norm": 0.5176487565040588,
"learning_rate": 4.0587528145957235e-06,
"loss": 0.6219,
"step": 909
},
{
"epoch": 0.0765397312698446,
"grad_norm": 0.680101752281189,
"learning_rate": 3.970631432305694e-06,
"loss": 0.822,
"step": 910
},
{
"epoch": 0.07662384086464663,
"grad_norm": 0.6776043176651001,
"learning_rate": 3.883457821114811e-06,
"loss": 0.5682,
"step": 911
},
{
"epoch": 0.07670795045944866,
"grad_norm": 0.5205222368240356,
"learning_rate": 3.797232841391407e-06,
"loss": 0.6134,
"step": 912
},
{
"epoch": 0.07679206005425068,
"grad_norm": 0.5915586948394775,
"learning_rate": 3.711957344141237e-06,
"loss": 0.748,
"step": 913
},
{
"epoch": 0.07687616964905272,
"grad_norm": 0.5967119336128235,
"learning_rate": 3.627632170999029e-06,
"loss": 0.6232,
"step": 914
},
{
"epoch": 0.07696027924385475,
"grad_norm": 0.48944488167762756,
"learning_rate": 3.5442581542201923e-06,
"loss": 0.5955,
"step": 915
},
{
"epoch": 0.07704438883865677,
"grad_norm": 0.6073352098464966,
"learning_rate": 3.461836116672612e-06,
"loss": 0.7206,
"step": 916
},
{
"epoch": 0.0771284984334588,
"grad_norm": 0.6056821346282959,
"learning_rate": 3.380366871828522e-06,
"loss": 0.6869,
"step": 917
},
{
"epoch": 0.07721260802826083,
"grad_norm": 0.50712651014328,
"learning_rate": 3.2998512237565005e-06,
"loss": 0.5963,
"step": 918
},
{
"epoch": 0.07729671762306285,
"grad_norm": 0.5057475566864014,
"learning_rate": 3.2202899671134546e-06,
"loss": 0.7265,
"step": 919
},
{
"epoch": 0.07738082721786488,
"grad_norm": 0.4675261378288269,
"learning_rate": 3.1416838871368924e-06,
"loss": 0.4619,
"step": 920
},
{
"epoch": 0.0774649368126669,
"grad_norm": 0.5350446105003357,
"learning_rate": 3.064033759637064e-06,
"loss": 0.6337,
"step": 921
},
{
"epoch": 0.07754904640746893,
"grad_norm": 0.653113067150116,
"learning_rate": 2.9873403509894203e-06,
"loss": 0.6033,
"step": 922
},
{
"epoch": 0.07763315600227096,
"grad_norm": 0.6161055564880371,
"learning_rate": 2.9116044181269007e-06,
"loss": 0.5878,
"step": 923
},
{
"epoch": 0.07771726559707298,
"grad_norm": 0.569428563117981,
"learning_rate": 2.836826708532603e-06,
"loss": 0.6588,
"step": 924
},
{
"epoch": 0.07780137519187501,
"grad_norm": 0.7410022020339966,
"learning_rate": 2.7630079602323442e-06,
"loss": 0.4855,
"step": 925
},
{
"epoch": 0.07788548478667703,
"grad_norm": 0.6758790016174316,
"learning_rate": 2.690148901787337e-06,
"loss": 0.6881,
"step": 926
},
{
"epoch": 0.07796959438147907,
"grad_norm": 0.49047234654426575,
"learning_rate": 2.618250252287113e-06,
"loss": 0.5011,
"step": 927
},
{
"epoch": 0.0780537039762811,
"grad_norm": 0.5375692248344421,
"learning_rate": 2.5473127213422763e-06,
"loss": 0.7307,
"step": 928
},
{
"epoch": 0.07813781357108313,
"grad_norm": 0.581413745880127,
"learning_rate": 2.4773370090776626e-06,
"loss": 0.6392,
"step": 929
},
{
"epoch": 0.07822192316588515,
"grad_norm": 0.6516633033752441,
"learning_rate": 2.4083238061252567e-06,
"loss": 0.6592,
"step": 930
},
{
"epoch": 0.07830603276068718,
"grad_norm": 0.4725818634033203,
"learning_rate": 2.3402737936175425e-06,
"loss": 0.6819,
"step": 931
},
{
"epoch": 0.0783901423554892,
"grad_norm": 0.6635422110557556,
"learning_rate": 2.273187643180652e-06,
"loss": 0.6987,
"step": 932
},
{
"epoch": 0.07847425195029123,
"grad_norm": 0.5256288647651672,
"learning_rate": 2.2070660169278166e-06,
"loss": 0.5697,
"step": 933
},
{
"epoch": 0.07855836154509326,
"grad_norm": 0.4704425036907196,
"learning_rate": 2.141909567452793e-06,
"loss": 0.4441,
"step": 934
},
{
"epoch": 0.07864247113989528,
"grad_norm": 0.4934440553188324,
"learning_rate": 2.0777189378234143e-06,
"loss": 0.6377,
"step": 935
},
{
"epoch": 0.07872658073469731,
"grad_norm": 0.5220550298690796,
"learning_rate": 2.014494761575314e-06,
"loss": 0.5492,
"step": 936
},
{
"epoch": 0.07881069032949933,
"grad_norm": 0.7263022065162659,
"learning_rate": 1.9522376627055583e-06,
"loss": 0.6481,
"step": 937
},
{
"epoch": 0.07889479992430136,
"grad_norm": 0.6368206739425659,
"learning_rate": 1.8909482556666024e-06,
"loss": 0.5843,
"step": 938
},
{
"epoch": 0.07897890951910339,
"grad_norm": 0.522922933101654,
"learning_rate": 1.8306271453601199e-06,
"loss": 0.7028,
"step": 939
},
{
"epoch": 0.07906301911390543,
"grad_norm": 0.5848362445831299,
"learning_rate": 1.771274927131139e-06,
"loss": 0.7176,
"step": 940
},
{
"epoch": 0.07914712870870745,
"grad_norm": 0.5615982413291931,
"learning_rate": 1.712892186762083e-06,
"loss": 0.5697,
"step": 941
},
{
"epoch": 0.07923123830350948,
"grad_norm": 0.4720976948738098,
"learning_rate": 1.6554795004670388e-06,
"loss": 0.4611,
"step": 942
},
{
"epoch": 0.0793153478983115,
"grad_norm": 0.5845123529434204,
"learning_rate": 1.5990374348860305e-06,
"loss": 0.6374,
"step": 943
},
{
"epoch": 0.07939945749311353,
"grad_norm": 0.5101219415664673,
"learning_rate": 1.543566547079467e-06,
"loss": 0.551,
"step": 944
},
{
"epoch": 0.07948356708791555,
"grad_norm": 0.6655003428459167,
"learning_rate": 1.4890673845226133e-06,
"loss": 0.7873,
"step": 945
},
{
"epoch": 0.07956767668271758,
"grad_norm": 0.552952766418457,
"learning_rate": 1.4355404851001952e-06,
"loss": 0.7553,
"step": 946
},
{
"epoch": 0.0796517862775196,
"grad_norm": 0.37656185030937195,
"learning_rate": 1.3829863771011253e-06,
"loss": 0.5656,
"step": 947
},
{
"epoch": 0.07973589587232163,
"grad_norm": 0.7558304667472839,
"learning_rate": 1.3314055792131964e-06,
"loss": 0.6068,
"step": 948
},
{
"epoch": 0.07982000546712366,
"grad_norm": 0.584898054599762,
"learning_rate": 1.280798600518085e-06,
"loss": 0.5412,
"step": 949
},
{
"epoch": 0.07990411506192568,
"grad_norm": 0.542145848274231,
"learning_rate": 1.231165940486234e-06,
"loss": 0.563,
"step": 950
},
{
"epoch": 0.07998822465672771,
"grad_norm": 0.6324697732925415,
"learning_rate": 1.1825080889719563e-06,
"loss": 0.5728,
"step": 951
},
{
"epoch": 0.08007233425152974,
"grad_norm": 0.5715826749801636,
"learning_rate": 1.134825526208605e-06,
"loss": 0.6769,
"step": 952
},
{
"epoch": 0.08015644384633178,
"grad_norm": 0.5874225497245789,
"learning_rate": 1.0881187228038215e-06,
"loss": 0.579,
"step": 953
},
{
"epoch": 0.0802405534411338,
"grad_norm": 0.5252524018287659,
"learning_rate": 1.0423881397349068e-06,
"loss": 0.6326,
"step": 954
},
{
"epoch": 0.08032466303593583,
"grad_norm": 0.6097508668899536,
"learning_rate": 9.976342283442463e-07,
"loss": 0.6524,
"step": 955
},
{
"epoch": 0.08040877263073785,
"grad_norm": 0.41464993357658386,
"learning_rate": 9.538574303348813e-07,
"loss": 0.4743,
"step": 956
},
{
"epoch": 0.08049288222553988,
"grad_norm": 0.46832600235939026,
"learning_rate": 9.110581777661331e-07,
"loss": 0.5314,
"step": 957
},
{
"epoch": 0.0805769918203419,
"grad_norm": 0.5717182159423828,
"learning_rate": 8.692368930493521e-07,
"loss": 0.7329,
"step": 958
},
{
"epoch": 0.08066110141514393,
"grad_norm": 0.5042733550071716,
"learning_rate": 8.283939889437209e-07,
"loss": 0.7913,
"step": 959
},
{
"epoch": 0.08074521100994596,
"grad_norm": 0.626025915145874,
"learning_rate": 7.885298685522235e-07,
"loss": 0.7545,
"step": 960
},
{
"epoch": 0.08082932060474798,
"grad_norm": 0.5646237730979919,
"learning_rate": 7.496449253176274e-07,
"loss": 0.5513,
"step": 961
},
{
"epoch": 0.08091343019955001,
"grad_norm": 0.6173761487007141,
"learning_rate": 7.117395430186414e-07,
"loss": 0.6928,
"step": 962
},
{
"epoch": 0.08099753979435204,
"grad_norm": 0.5799218416213989,
"learning_rate": 6.748140957660631e-07,
"loss": 0.8141,
"step": 963
},
{
"epoch": 0.08108164938915406,
"grad_norm": 0.5109354853630066,
"learning_rate": 6.388689479991605e-07,
"loss": 0.8059,
"step": 964
},
{
"epoch": 0.0811657589839561,
"grad_norm": 0.5159839391708374,
"learning_rate": 6.039044544820404e-07,
"loss": 0.5749,
"step": 965
},
{
"epoch": 0.08124986857875813,
"grad_norm": 0.6682405471801758,
"learning_rate": 5.699209603001076e-07,
"loss": 0.7188,
"step": 966
},
{
"epoch": 0.08133397817356015,
"grad_norm": 0.8375939130783081,
"learning_rate": 5.369188008567672e-07,
"loss": 0.9425,
"step": 967
},
{
"epoch": 0.08141808776836218,
"grad_norm": 0.6056286692619324,
"learning_rate": 5.048983018699827e-07,
"loss": 0.6488,
"step": 968
},
{
"epoch": 0.0815021973631642,
"grad_norm": 0.5580015778541565,
"learning_rate": 4.738597793691679e-07,
"loss": 0.5895,
"step": 969
},
{
"epoch": 0.08158630695796623,
"grad_norm": 0.5345575213432312,
"learning_rate": 4.438035396920004e-07,
"loss": 0.6993,
"step": 970
},
{
"epoch": 0.08167041655276826,
"grad_norm": 0.5608043074607849,
"learning_rate": 4.1472987948143473e-07,
"loss": 0.7225,
"step": 971
},
{
"epoch": 0.08175452614757028,
"grad_norm": 0.5627851486206055,
"learning_rate": 3.866390856827495e-07,
"loss": 0.6505,
"step": 972
},
{
"epoch": 0.08183863574237231,
"grad_norm": 0.7156445384025574,
"learning_rate": 3.595314355407609e-07,
"loss": 0.8306,
"step": 973
},
{
"epoch": 0.08192274533717434,
"grad_norm": 0.8838444352149963,
"learning_rate": 3.3340719659701313e-07,
"loss": 0.5709,
"step": 974
},
{
"epoch": 0.08200685493197636,
"grad_norm": 0.5641219615936279,
"learning_rate": 3.0826662668720364e-07,
"loss": 0.7521,
"step": 975
},
{
"epoch": 0.08209096452677839,
"grad_norm": 0.636716902256012,
"learning_rate": 2.841099739386066e-07,
"loss": 0.6791,
"step": 976
},
{
"epoch": 0.08217507412158041,
"grad_norm": 0.5187475681304932,
"learning_rate": 2.609374767676309e-07,
"loss": 0.5493,
"step": 977
},
{
"epoch": 0.08225918371638245,
"grad_norm": 0.555397629737854,
"learning_rate": 2.387493638774774e-07,
"loss": 0.8756,
"step": 978
},
{
"epoch": 0.08234329331118448,
"grad_norm": 0.4511655271053314,
"learning_rate": 2.175458542558517e-07,
"loss": 0.43,
"step": 979
},
{
"epoch": 0.0824274029059865,
"grad_norm": 0.7089412212371826,
"learning_rate": 1.973271571728441e-07,
"loss": 0.6912,
"step": 980
},
{
"epoch": 0.08251151250078853,
"grad_norm": 0.46385011076927185,
"learning_rate": 1.7809347217881966e-07,
"loss": 0.6673,
"step": 981
},
{
"epoch": 0.08259562209559056,
"grad_norm": 0.5081963539123535,
"learning_rate": 1.598449891024978e-07,
"loss": 0.6091,
"step": 982
},
{
"epoch": 0.08267973169039258,
"grad_norm": 0.4684099853038788,
"learning_rate": 1.425818880490315e-07,
"loss": 0.6555,
"step": 983
},
{
"epoch": 0.08276384128519461,
"grad_norm": 0.5321608185768127,
"learning_rate": 1.2630433939825327e-07,
"loss": 0.6384,
"step": 984
},
{
"epoch": 0.08284795087999663,
"grad_norm": 0.5801177620887756,
"learning_rate": 1.1101250380300965e-07,
"loss": 0.6731,
"step": 985
},
{
"epoch": 0.08293206047479866,
"grad_norm": 0.5255969166755676,
"learning_rate": 9.670653218752934e-08,
"loss": 0.6411,
"step": 986
},
{
"epoch": 0.08301617006960069,
"grad_norm": 0.5361884236335754,
"learning_rate": 8.33865657459909e-08,
"loss": 0.4983,
"step": 987
},
{
"epoch": 0.08310027966440271,
"grad_norm": 0.42637842893600464,
"learning_rate": 7.105273594107953e-08,
"loss": 0.501,
"step": 988
},
{
"epoch": 0.08318438925920474,
"grad_norm": 0.5724627375602722,
"learning_rate": 5.970516450271025e-08,
"loss": 0.7244,
"step": 989
},
{
"epoch": 0.08326849885400676,
"grad_norm": 0.5623000264167786,
"learning_rate": 4.934396342684e-08,
"loss": 0.6702,
"step": 990
},
{
"epoch": 0.0833526084488088,
"grad_norm": 0.505596399307251,
"learning_rate": 3.996923497434635e-08,
"loss": 0.7967,
"step": 991
},
{
"epoch": 0.08343671804361083,
"grad_norm": 0.7859603762626648,
"learning_rate": 3.1581071670006015e-08,
"loss": 0.7663,
"step": 992
},
{
"epoch": 0.08352082763841286,
"grad_norm": 0.5909059643745422,
"learning_rate": 2.417955630159563e-08,
"loss": 0.6006,
"step": 993
},
{
"epoch": 0.08360493723321488,
"grad_norm": 0.4647238552570343,
"learning_rate": 1.7764761919103477e-08,
"loss": 0.6949,
"step": 994
},
{
"epoch": 0.08368904682801691,
"grad_norm": 0.5446833372116089,
"learning_rate": 1.2336751833941229e-08,
"loss": 0.5849,
"step": 995
},
{
"epoch": 0.08377315642281893,
"grad_norm": 0.5875367522239685,
"learning_rate": 7.895579618388827e-09,
"loss": 0.769,
"step": 996
},
{
"epoch": 0.08385726601762096,
"grad_norm": 0.769709587097168,
"learning_rate": 4.4412891050171765e-09,
"loss": 0.8165,
"step": 997
},
{
"epoch": 0.08394137561242299,
"grad_norm": 0.5166672468185425,
"learning_rate": 1.973914386288467e-09,
"loss": 0.696,
"step": 998
},
{
"epoch": 0.08402548520722501,
"grad_norm": 0.6133189797401428,
"learning_rate": 4.934798141786879e-10,
"loss": 0.7572,
"step": 999
},
{
"epoch": 0.08410959480202704,
"grad_norm": 0.4853799045085907,
"learning_rate": 0.0,
"loss": 0.6055,
"step": 1000
},
{
"epoch": 0.08410959480202704,
"eval_loss": 0.6270027160644531,
"eval_runtime": 118.2303,
"eval_samples_per_second": 21.179,
"eval_steps_per_second": 21.179,
"step": 1000
}
],
"logging_steps": 1,
"max_steps": 1000,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 5.774858388504576e+16,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}