{
  "best_metric": null,
  "best_model_checkpoint": null,
  "epoch": 3.4606934431857193,
  "eval_steps": 91,
  "global_step": 1274,
  "is_hyper_param_search": false,
  "is_local_process_zero": true,
  "is_world_process_zero": true,
  "log_history": [
    {
      "epoch": 0.0027463096464126332,
      "grad_norm": 0.056396484375,
      "learning_rate": 3e-06,
      "loss": 0.6232,
      "step": 1
    },
    {
      "epoch": 0.0027463096464126332,
      "eval_loss": 0.6296440362930298,
      "eval_runtime": 599.6568,
      "eval_samples_per_second": 15.285,
      "eval_steps_per_second": 15.285,
      "step": 1
    },
    {
      "epoch": 0.0054926192928252664,
      "grad_norm": 0.060546875,
      "learning_rate": 6e-06,
      "loss": 0.5596,
      "step": 2
    },
    {
      "epoch": 0.008238928939237899,
      "grad_norm": 0.054443359375,
      "learning_rate": 9e-06,
      "loss": 0.7123,
      "step": 3
    },
    {
      "epoch": 0.010985238585650533,
      "grad_norm": 0.05224609375,
      "learning_rate": 1.2e-05,
      "loss": 0.6337,
      "step": 4
    },
    {
      "epoch": 0.013731548232063165,
      "grad_norm": 0.051025390625,
      "learning_rate": 1.5e-05,
      "loss": 0.5764,
      "step": 5
    },
    {
      "epoch": 0.016477857878475798,
      "grad_norm": 0.0546875,
      "learning_rate": 1.8e-05,
      "loss": 0.7453,
      "step": 6
    },
    {
      "epoch": 0.01922416752488843,
      "grad_norm": 0.057373046875,
      "learning_rate": 2.1e-05,
      "loss": 0.7076,
      "step": 7
    },
    {
      "epoch": 0.021970477171301066,
      "grad_norm": 0.06298828125,
      "learning_rate": 2.4e-05,
      "loss": 0.5094,
      "step": 8
    },
    {
      "epoch": 0.024716786817713696,
      "grad_norm": 0.05859375,
      "learning_rate": 2.7000000000000002e-05,
      "loss": 0.6062,
      "step": 9
    },
    {
      "epoch": 0.02746309646412633,
      "grad_norm": 0.056884765625,
      "learning_rate": 3e-05,
      "loss": 0.6501,
      "step": 10
    },
    {
      "epoch": 0.030209406110538965,
      "grad_norm": 0.060546875,
      "learning_rate": 2.9999964598289033e-05,
      "loss": 0.6403,
      "step": 11
    },
    {
      "epoch": 0.032955715756951595,
      "grad_norm": 0.061279296875,
      "learning_rate": 2.999985839332323e-05,
      "loss": 0.6464,
      "step": 12
    },
    {
      "epoch": 0.03570202540336423,
      "grad_norm": 0.05859375,
      "learning_rate": 2.9999681385603907e-05,
      "loss": 0.6528,
      "step": 13
    },
    {
      "epoch": 0.03844833504977686,
      "grad_norm": 0.062255859375,
      "learning_rate": 2.9999433575966585e-05,
      "loss": 0.7109,
      "step": 14
    },
    {
      "epoch": 0.0411946446961895,
      "grad_norm": 0.06689453125,
      "learning_rate": 2.999911496558097e-05,
      "loss": 0.4795,
      "step": 15
    },
    {
      "epoch": 0.04394095434260213,
      "grad_norm": 0.061767578125,
      "learning_rate": 2.9998725555950983e-05,
      "loss": 0.6743,
      "step": 16
    },
    {
      "epoch": 0.04668726398901476,
      "grad_norm": 0.0654296875,
      "learning_rate": 2.9998265348914726e-05,
      "loss": 0.6143,
      "step": 17
    },
    {
      "epoch": 0.04943357363542739,
      "grad_norm": 0.06494140625,
      "learning_rate": 2.9997734346644482e-05,
      "loss": 0.5521,
      "step": 18
    },
    {
      "epoch": 0.05217988328184003,
      "grad_norm": 0.06396484375,
      "learning_rate": 2.99971325516467e-05,
      "loss": 0.6291,
      "step": 19
    },
    {
      "epoch": 0.05492619292825266,
      "grad_norm": 0.06591796875,
      "learning_rate": 2.9996459966761994e-05,
      "loss": 0.5828,
      "step": 20
    },
    {
      "epoch": 0.057672502574665295,
      "grad_norm": 0.068359375,
      "learning_rate": 2.9995716595165114e-05,
      "loss": 0.5432,
      "step": 21
    },
    {
      "epoch": 0.06041881222107793,
      "grad_norm": 0.076171875,
      "learning_rate": 2.9994902440364943e-05,
      "loss": 0.5769,
      "step": 22
    },
    {
      "epoch": 0.06316512186749056,
      "grad_norm": 0.07568359375,
      "learning_rate": 2.999401750620448e-05,
      "loss": 0.469,
      "step": 23
    },
    {
      "epoch": 0.06591143151390319,
      "grad_norm": 0.06884765625,
      "learning_rate": 2.999306179686082e-05,
      "loss": 0.4414,
      "step": 24
    },
    {
      "epoch": 0.06865774116031582,
      "grad_norm": 0.05419921875,
      "learning_rate": 2.9992035316845125e-05,
      "loss": 0.7219,
      "step": 25
    },
    {
      "epoch": 0.07140405080672846,
      "grad_norm": 0.0654296875,
      "learning_rate": 2.9990938071002606e-05,
      "loss": 0.6742,
      "step": 26
    },
    {
      "epoch": 0.07415036045314109,
      "grad_norm": 0.04541015625,
      "learning_rate": 2.998977006451253e-05,
      "loss": 0.5633,
      "step": 27
    },
    {
      "epoch": 0.07689667009955373,
      "grad_norm": 0.043212890625,
      "learning_rate": 2.998853130288814e-05,
      "loss": 0.5486,
      "step": 28
    },
    {
      "epoch": 0.07964297974596636,
      "grad_norm": 0.04345703125,
      "learning_rate": 2.9987221791976687e-05,
      "loss": 0.4064,
      "step": 29
    },
    {
      "epoch": 0.082389289392379,
      "grad_norm": 0.04541015625,
      "learning_rate": 2.9985841537959345e-05,
      "loss": 0.5184,
      "step": 30
    },
    {
      "epoch": 0.08513559903879163,
      "grad_norm": 0.044921875,
      "learning_rate": 2.9984390547351244e-05,
      "loss": 0.5407,
      "step": 31
    },
    {
      "epoch": 0.08788190868520426,
      "grad_norm": 0.044677734375,
      "learning_rate": 2.998286882700138e-05,
      "loss": 0.5532,
      "step": 32
    },
    {
      "epoch": 0.09062821833161688,
      "grad_norm": 0.14453125,
      "learning_rate": 2.9981276384092628e-05,
      "loss": 1.2318,
      "step": 33
    },
    {
      "epoch": 0.09337452797802952,
      "grad_norm": 0.04052734375,
      "learning_rate": 2.9979613226141672e-05,
      "loss": 0.5457,
      "step": 34
    },
    {
      "epoch": 0.09612083762444215,
      "grad_norm": 0.03857421875,
      "learning_rate": 2.9977879360999007e-05,
      "loss": 0.5391,
      "step": 35
    },
    {
      "epoch": 0.09886714727085479,
      "grad_norm": 0.0400390625,
      "learning_rate": 2.9976074796848866e-05,
      "loss": 0.5919,
      "step": 36
    },
    {
      "epoch": 0.10161345691726742,
      "grad_norm": 0.037353515625,
      "learning_rate": 2.99741995422092e-05,
      "loss": 0.6211,
      "step": 37
    },
    {
      "epoch": 0.10435976656368005,
      "grad_norm": 0.036865234375,
      "learning_rate": 2.997225360593165e-05,
      "loss": 0.5296,
      "step": 38
    },
    {
      "epoch": 0.10710607621009269,
      "grad_norm": 0.037109375,
      "learning_rate": 2.997023699720147e-05,
      "loss": 0.5469,
      "step": 39
    },
    {
      "epoch": 0.10985238585650532,
      "grad_norm": 0.038818359375,
      "learning_rate": 2.9968149725537515e-05,
      "loss": 0.6447,
      "step": 40
    },
    {
      "epoch": 0.11259869550291796,
      "grad_norm": 0.03564453125,
      "learning_rate": 2.9965991800792185e-05,
      "loss": 0.5941,
      "step": 41
    },
    {
      "epoch": 0.11534500514933059,
      "grad_norm": 0.0341796875,
      "learning_rate": 2.9963763233151377e-05,
      "loss": 0.5228,
      "step": 42
    },
    {
      "epoch": 0.11809131479574322,
      "grad_norm": 0.035888671875,
      "learning_rate": 2.9961464033134444e-05,
      "loss": 0.6166,
      "step": 43
    },
    {
      "epoch": 0.12083762444215586,
      "grad_norm": 0.0380859375,
      "learning_rate": 2.9959094211594122e-05,
      "loss": 0.6156,
      "step": 44
    },
    {
      "epoch": 0.12358393408856849,
      "grad_norm": 0.034912109375,
      "learning_rate": 2.9956653779716517e-05,
      "loss": 0.5492,
      "step": 45
    },
    {
      "epoch": 0.1263302437349811,
      "grad_norm": 0.03564453125,
      "learning_rate": 2.9954142749021023e-05,
      "loss": 0.5405,
      "step": 46
    },
    {
      "epoch": 0.12907655338139376,
      "grad_norm": 0.035400390625,
      "learning_rate": 2.9951561131360278e-05,
      "loss": 0.584,
      "step": 47
    },
    {
      "epoch": 0.13182286302780638,
      "grad_norm": 0.0380859375,
      "learning_rate": 2.994890893892011e-05,
      "loss": 0.62,
      "step": 48
    },
    {
      "epoch": 0.13456917267421903,
      "grad_norm": 0.031982421875,
      "learning_rate": 2.994618618421946e-05,
      "loss": 0.4026,
      "step": 49
    },
    {
      "epoch": 0.13731548232063165,
      "grad_norm": 0.0341796875,
      "learning_rate": 2.994339288011037e-05,
      "loss": 0.6126,
      "step": 50
    },
    {
      "epoch": 0.1400617919670443,
      "grad_norm": 0.037109375,
      "learning_rate": 2.9940529039777855e-05,
      "loss": 0.6316,
      "step": 51
    },
    {
      "epoch": 0.14280810161345692,
      "grad_norm": 0.033935546875,
      "learning_rate": 2.9937594676739907e-05,
      "loss": 0.3887,
      "step": 52
    },
    {
      "epoch": 0.14555441125986954,
      "grad_norm": 0.03564453125,
      "learning_rate": 2.9934589804847382e-05,
      "loss": 0.5214,
      "step": 53
    },
    {
      "epoch": 0.14830072090628219,
      "grad_norm": 0.034423828125,
      "learning_rate": 2.9931514438283966e-05,
      "loss": 0.5914,
      "step": 54
    },
    {
      "epoch": 0.1510470305526948,
      "grad_norm": 0.036376953125,
      "learning_rate": 2.9928368591566085e-05,
      "loss": 0.5443,
      "step": 55
    },
    {
      "epoch": 0.15379334019910745,
      "grad_norm": 0.03466796875,
      "learning_rate": 2.9925152279542856e-05,
      "loss": 0.5522,
      "step": 56
    },
    {
      "epoch": 0.15653964984552007,
      "grad_norm": 0.031494140625,
      "learning_rate": 2.9921865517396008e-05,
      "loss": 0.5146,
      "step": 57
    },
    {
      "epoch": 0.15928595949193272,
      "grad_norm": 0.033447265625,
      "learning_rate": 2.9918508320639803e-05,
      "loss": 0.5396,
      "step": 58
    },
    {
      "epoch": 0.16203226913834534,
      "grad_norm": 0.035888671875,
      "learning_rate": 2.9915080705120976e-05,
      "loss": 0.5118,
      "step": 59
    },
    {
      "epoch": 0.164778578784758,
      "grad_norm": 0.03662109375,
      "learning_rate": 2.991158268701866e-05,
      "loss": 0.6652,
      "step": 60
    },
    {
      "epoch": 0.1675248884311706,
      "grad_norm": 0.0341796875,
      "learning_rate": 2.9908014282844295e-05,
      "loss": 0.4211,
      "step": 61
    },
    {
      "epoch": 0.17027119807758326,
      "grad_norm": 0.0361328125,
      "learning_rate": 2.9904375509441562e-05,
      "loss": 0.4445,
      "step": 62
    },
    {
      "epoch": 0.17301750772399588,
      "grad_norm": 0.03369140625,
      "learning_rate": 2.9900666383986303e-05,
      "loss": 0.588,
      "step": 63
    },
    {
      "epoch": 0.17576381737040853,
      "grad_norm": 0.033203125,
      "learning_rate": 2.9896886923986433e-05,
      "loss": 0.5705,
      "step": 64
    },
    {
      "epoch": 0.17851012701682115,
      "grad_norm": 0.0341796875,
      "learning_rate": 2.989303714728187e-05,
      "loss": 0.5068,
      "step": 65
    },
    {
      "epoch": 0.18125643666323377,
      "grad_norm": 0.033203125,
      "learning_rate": 2.9889117072044436e-05,
      "loss": 0.5196,
      "step": 66
    },
    {
      "epoch": 0.18400274630964641,
      "grad_norm": 0.03515625,
      "learning_rate": 2.9885126716777776e-05,
      "loss": 0.5952,
      "step": 67
    },
    {
      "epoch": 0.18674905595605903,
      "grad_norm": 0.0322265625,
      "learning_rate": 2.9881066100317288e-05,
      "loss": 0.6194,
      "step": 68
    },
    {
      "epoch": 0.18949536560247168,
      "grad_norm": 0.0341796875,
      "learning_rate": 2.987693524183e-05,
      "loss": 0.4453,
      "step": 69
    },
    {
      "epoch": 0.1922416752488843,
      "grad_norm": 0.03369140625,
      "learning_rate": 2.987273416081451e-05,
      "loss": 0.524,
      "step": 70
    },
    {
      "epoch": 0.19498798489529695,
      "grad_norm": 0.03857421875,
      "learning_rate": 2.9868462877100875e-05,
      "loss": 0.5899,
      "step": 71
    },
    {
      "epoch": 0.19773429454170957,
      "grad_norm": 0.036376953125,
      "learning_rate": 2.9864121410850527e-05,
      "loss": 0.4603,
      "step": 72
    },
    {
      "epoch": 0.20048060418812222,
      "grad_norm": 0.03759765625,
      "learning_rate": 2.9859709782556185e-05,
      "loss": 0.4829,
      "step": 73
    },
    {
      "epoch": 0.20322691383453484,
      "grad_norm": 0.034912109375,
      "learning_rate": 2.9855228013041737e-05,
      "loss": 0.5735,
      "step": 74
    },
    {
      "epoch": 0.2059732234809475,
      "grad_norm": 0.03369140625,
      "learning_rate": 2.9850676123462157e-05,
      "loss": 0.5104,
      "step": 75
    },
    {
      "epoch": 0.2087195331273601,
      "grad_norm": 0.03369140625,
      "learning_rate": 2.98460541353034e-05,
      "loss": 0.5501,
      "step": 76
    },
    {
      "epoch": 0.21146584277377276,
      "grad_norm": 0.032958984375,
      "learning_rate": 2.9841362070382307e-05,
      "loss": 0.5119,
      "step": 77
    },
    {
      "epoch": 0.21421215242018538,
      "grad_norm": 0.035400390625,
      "learning_rate": 2.9836599950846493e-05,
      "loss": 0.589,
      "step": 78
    },
    {
      "epoch": 0.216958462066598,
      "grad_norm": 0.03173828125,
      "learning_rate": 2.9831767799174255e-05,
      "loss": 0.4544,
      "step": 79
    },
    {
      "epoch": 0.21970477171301064,
      "grad_norm": 0.03564453125,
      "learning_rate": 2.9826865638174445e-05,
      "loss": 0.4294,
      "step": 80
    },
    {
      "epoch": 0.22245108135942326,
      "grad_norm": 0.03466796875,
      "learning_rate": 2.9821893490986382e-05,
      "loss": 0.5649,
      "step": 81
    },
    {
      "epoch": 0.2251973910058359,
      "grad_norm": 0.03759765625,
      "learning_rate": 2.981685138107974e-05,
      "loss": 0.532,
      "step": 82
    },
    {
      "epoch": 0.22794370065224853,
      "grad_norm": 0.03369140625,
      "learning_rate": 2.9811739332254418e-05,
      "loss": 0.6026,
      "step": 83
    },
    {
      "epoch": 0.23069001029866118,
      "grad_norm": 0.035400390625,
      "learning_rate": 2.9806557368640457e-05,
      "loss": 0.5516,
      "step": 84
    },
    {
      "epoch": 0.2334363199450738,
      "grad_norm": 0.033203125,
      "learning_rate": 2.9801305514697913e-05,
      "loss": 0.4544,
      "step": 85
    },
    {
      "epoch": 0.23618262959148645,
      "grad_norm": 0.03173828125,
      "learning_rate": 2.9795983795216727e-05,
      "loss": 0.5327,
      "step": 86
    },
    {
      "epoch": 0.23892893923789907,
      "grad_norm": 0.0380859375,
      "learning_rate": 2.979059223531664e-05,
      "loss": 0.5217,
      "step": 87
    },
    {
      "epoch": 0.24167524888431172,
      "grad_norm": 0.0361328125,
      "learning_rate": 2.978513086044703e-05,
      "loss": 0.4562,
      "step": 88
    },
    {
      "epoch": 0.24442155853072434,
      "grad_norm": 0.0380859375,
      "learning_rate": 2.9779599696386846e-05,
      "loss": 0.763,
      "step": 89
    },
    {
      "epoch": 0.24716786817713698,
      "grad_norm": 0.03369140625,
      "learning_rate": 2.9773998769244434e-05,
      "loss": 0.4698,
      "step": 90
    },
    {
      "epoch": 0.2499141778235496,
      "grad_norm": 0.03662109375,
      "learning_rate": 2.976832810545745e-05,
      "loss": 0.5602,
      "step": 91
    },
    {
      "epoch": 0.2499141778235496,
      "eval_loss": 0.5245938897132874,
      "eval_runtime": 620.8292,
      "eval_samples_per_second": 14.764,
      "eval_steps_per_second": 14.764,
      "step": 91
    },
    {
      "epoch": 0.2526604874699622,
      "grad_norm": 0.030517578125,
      "learning_rate": 2.9762587731792725e-05,
      "loss": 0.477,
      "step": 92
    },
    {
      "epoch": 0.2554067971163749,
      "grad_norm": 0.03271484375,
      "learning_rate": 2.9756777675346128e-05,
      "loss": 0.5536,
      "step": 93
    },
    {
      "epoch": 0.2581531067627875,
      "grad_norm": 0.036376953125,
      "learning_rate": 2.9750897963542453e-05,
      "loss": 0.5581,
      "step": 94
    },
    {
      "epoch": 0.2608994164092001,
      "grad_norm": 0.04345703125,
      "learning_rate": 2.974494862413528e-05,
      "loss": 0.5737,
      "step": 95
    },
    {
      "epoch": 0.26364572605561276,
      "grad_norm": 0.03662109375,
      "learning_rate": 2.973892968520685e-05,
      "loss": 0.5191,
      "step": 96
    },
    {
      "epoch": 0.2663920357020254,
      "grad_norm": 0.035888671875,
      "learning_rate": 2.9732841175167924e-05,
      "loss": 0.6794,
      "step": 97
    },
    {
      "epoch": 0.26913834534843806,
      "grad_norm": 0.036376953125,
      "learning_rate": 2.9726683122757664e-05,
      "loss": 0.5615,
      "step": 98
    },
    {
      "epoch": 0.27188465499485065,
      "grad_norm": 0.036376953125,
      "learning_rate": 2.972045555704348e-05,
      "loss": 0.521,
      "step": 99
    },
    {
      "epoch": 0.2746309646412633,
      "grad_norm": 0.036376953125,
      "learning_rate": 2.97141585074209e-05,
      "loss": 0.4473,
      "step": 100
    },
    {
      "epoch": 0.27737727428767595,
      "grad_norm": 0.033203125,
      "learning_rate": 2.9707792003613434e-05,
      "loss": 0.6017,
      "step": 101
    },
    {
      "epoch": 0.2801235839340886,
      "grad_norm": 0.042236328125,
      "learning_rate": 2.9701356075672442e-05,
      "loss": 0.5079,
      "step": 102
    },
    {
      "epoch": 0.2828698935805012,
      "grad_norm": 0.038818359375,
      "learning_rate": 2.969485075397696e-05,
      "loss": 0.5738,
      "step": 103
    },
    {
      "epoch": 0.28561620322691383,
      "grad_norm": 0.034423828125,
      "learning_rate": 2.9688276069233596e-05,
      "loss": 0.4251,
      "step": 104
    },
    {
      "epoch": 0.2883625128733265,
      "grad_norm": 0.03759765625,
      "learning_rate": 2.968163205247636e-05,
      "loss": 0.5902,
      "step": 105
    },
    {
      "epoch": 0.2911088225197391,
      "grad_norm": 0.043701171875,
      "learning_rate": 2.9674918735066534e-05,
      "loss": 0.4307,
      "step": 106
    },
    {
      "epoch": 0.2938551321661517,
      "grad_norm": 0.039306640625,
      "learning_rate": 2.9668136148692497e-05,
      "loss": 0.4018,
      "step": 107
    },
    {
      "epoch": 0.29660144181256437,
      "grad_norm": 0.039306640625,
      "learning_rate": 2.966128432536961e-05,
      "loss": 0.5109,
      "step": 108
    },
    {
      "epoch": 0.299347751458977,
      "grad_norm": 0.0361328125,
      "learning_rate": 2.9654363297440045e-05,
      "loss": 0.6136,
      "step": 109
    },
    {
      "epoch": 0.3020940611053896,
      "grad_norm": 0.038818359375,
      "learning_rate": 2.964737309757262e-05,
      "loss": 0.4161,
      "step": 110
    },
    {
      "epoch": 0.30484037075180226,
      "grad_norm": 0.0361328125,
      "learning_rate": 2.9640313758762692e-05,
      "loss": 0.4268,
      "step": 111
    },
    {
      "epoch": 0.3075866803982149,
      "grad_norm": 0.03466796875,
      "learning_rate": 2.9633185314331933e-05,
      "loss": 0.4809,
      "step": 112
    },
    {
      "epoch": 0.31033299004462755,
      "grad_norm": 0.03662109375,
      "learning_rate": 2.9625987797928237e-05,
      "loss": 0.4976,
      "step": 113
    },
    {
      "epoch": 0.31307929969104015,
      "grad_norm": 0.03955078125,
      "learning_rate": 2.9618721243525522e-05,
      "loss": 0.5508,
      "step": 114
    },
    {
      "epoch": 0.3158256093374528,
      "grad_norm": 0.045166015625,
      "learning_rate": 2.9611385685423582e-05,
      "loss": 0.4852,
      "step": 115
    },
    {
      "epoch": 0.31857191898386544,
      "grad_norm": 0.040771484375,
      "learning_rate": 2.9603981158247918e-05,
      "loss": 0.4301,
      "step": 116
    },
    {
      "epoch": 0.3213182286302781,
      "grad_norm": 0.0419921875,
      "learning_rate": 2.9596507696949598e-05,
      "loss": 0.4456,
      "step": 117
    },
    {
      "epoch": 0.3240645382766907,
      "grad_norm": 0.038330078125,
      "learning_rate": 2.9588965336805065e-05,
      "loss": 0.6092,
      "step": 118
    },
    {
      "epoch": 0.32681084792310333,
      "grad_norm": 0.033935546875,
      "learning_rate": 2.958135411341597e-05,
      "loss": 0.4823,
      "step": 119
    },
    {
      "epoch": 0.329557157569516,
      "grad_norm": 0.03662109375,
      "learning_rate": 2.9573674062709024e-05,
      "loss": 0.4666,
      "step": 120
    },
    {
      "epoch": 0.33230346721592857,
      "grad_norm": 0.035888671875,
      "learning_rate": 2.9565925220935828e-05,
      "loss": 0.4868,
      "step": 121
    },
    {
      "epoch": 0.3350497768623412,
      "grad_norm": 0.034423828125,
      "learning_rate": 2.9558107624672673e-05,
      "loss": 0.529,
      "step": 122
    },
    {
      "epoch": 0.33779608650875387,
      "grad_norm": 0.0341796875,
      "learning_rate": 2.9550221310820405e-05,
      "loss": 0.4308,
      "step": 123
    },
    {
      "epoch": 0.3405423961551665,
      "grad_norm": 0.0341796875,
      "learning_rate": 2.9542266316604213e-05,
      "loss": 0.4058,
      "step": 124
    },
    {
      "epoch": 0.3432887058015791,
      "grad_norm": 0.03466796875,
      "learning_rate": 2.95342426795735e-05,
      "loss": 0.4788,
      "step": 125
    },
    {
      "epoch": 0.34603501544799176,
      "grad_norm": 0.20703125,
      "learning_rate": 2.952615043760165e-05,
      "loss": 1.2963,
      "step": 126
    },
    {
      "epoch": 0.3487813250944044,
      "grad_norm": 0.0361328125,
      "learning_rate": 2.95179896288859e-05,
      "loss": 0.5734,
      "step": 127
    },
    {
      "epoch": 0.35152763474081705,
      "grad_norm": 0.03515625,
      "learning_rate": 2.9509760291947128e-05,
      "loss": 0.4352,
      "step": 128
    },
    {
      "epoch": 0.35427394438722964,
      "grad_norm": 0.03564453125,
      "learning_rate": 2.9501462465629672e-05,
      "loss": 0.6082,
      "step": 129
    },
    {
      "epoch": 0.3570202540336423,
      "grad_norm": 0.035400390625,
      "learning_rate": 2.949309618910118e-05,
      "loss": 0.4699,
      "step": 130
    },
    {
      "epoch": 0.35976656368005494,
      "grad_norm": 0.03759765625,
      "learning_rate": 2.9484661501852373e-05,
      "loss": 0.5504,
      "step": 131
    },
    {
      "epoch": 0.36251287332646753,
      "grad_norm": 0.037353515625,
      "learning_rate": 2.94761584436969e-05,
      "loss": 0.545,
      "step": 132
    },
    {
      "epoch": 0.3652591829728802,
      "grad_norm": 0.03515625,
      "learning_rate": 2.9467587054771146e-05,
      "loss": 0.445,
      "step": 133
    },
    {
      "epoch": 0.36800549261929283,
      "grad_norm": 0.1806640625,
      "learning_rate": 2.945894737553401e-05,
      "loss": 1.1891,
      "step": 134
    },
    {
      "epoch": 0.3707518022657055,
      "grad_norm": 0.037841796875,
      "learning_rate": 2.945023944676676e-05,
      "loss": 0.565,
      "step": 135
    },
    {
      "epoch": 0.37349811191211807,
      "grad_norm": 0.042236328125,
      "learning_rate": 2.9441463309572797e-05,
      "loss": 0.6599,
      "step": 136
    },
    {
      "epoch": 0.3762444215585307,
      "grad_norm": 0.0380859375,
      "learning_rate": 2.9432619005377496e-05,
      "loss": 0.4754,
      "step": 137
    },
    {
      "epoch": 0.37899073120494337,
      "grad_norm": 0.036376953125,
      "learning_rate": 2.9423706575927985e-05,
      "loss": 0.4966,
      "step": 138
    },
    {
      "epoch": 0.381737040851356,
      "grad_norm": 0.0390625,
      "learning_rate": 2.9414726063292974e-05,
      "loss": 0.4269,
      "step": 139
    },
    {
      "epoch": 0.3844833504977686,
      "grad_norm": 0.038818359375,
      "learning_rate": 2.940567750986252e-05,
      "loss": 0.5516,
      "step": 140
    },
    {
      "epoch": 0.38722966014418125,
      "grad_norm": 0.0361328125,
      "learning_rate": 2.9396560958347865e-05,
      "loss": 0.486,
      "step": 141
    },
    {
      "epoch": 0.3899759697905939,
      "grad_norm": 0.034912109375,
      "learning_rate": 2.9387376451781215e-05,
      "loss": 0.4506,
      "step": 142
    },
    {
      "epoch": 0.39272227943700655,
      "grad_norm": 0.039306640625,
      "learning_rate": 2.9378124033515533e-05,
      "loss": 0.6122,
      "step": 143
    },
    {
      "epoch": 0.39546858908341914,
      "grad_norm": 0.03466796875,
      "learning_rate": 2.936880374722434e-05,
      "loss": 0.4776,
      "step": 144
    },
    {
      "epoch": 0.3982148987298318,
      "grad_norm": 0.041015625,
      "learning_rate": 2.9359415636901522e-05,
      "loss": 0.5574,
      "step": 145
    },
    {
      "epoch": 0.40096120837624444,
      "grad_norm": 0.035888671875,
      "learning_rate": 2.9349959746861093e-05,
      "loss": 0.5289,
      "step": 146
    },
    {
      "epoch": 0.40370751802265703,
      "grad_norm": 0.037109375,
      "learning_rate": 2.9340436121737018e-05,
      "loss": 0.4664,
      "step": 147
    },
    {
      "epoch": 0.4064538276690697,
      "grad_norm": 0.04345703125,
      "learning_rate": 2.9330844806482974e-05,
      "loss": 0.5322,
      "step": 148
    },
    {
      "epoch": 0.4092001373154823,
      "grad_norm": 0.037109375,
      "learning_rate": 2.9321185846372162e-05,
      "loss": 0.4143,
      "step": 149
    },
    {
      "epoch": 0.411946446961895,
      "grad_norm": 0.0400390625,
      "learning_rate": 2.9311459286997073e-05,
      "loss": 0.5038,
      "step": 150
    },
    {
      "epoch": 0.41469275660830757,
      "grad_norm": 0.036376953125,
      "learning_rate": 2.930166517426929e-05,
      "loss": 0.4905,
      "step": 151
    },
    {
      "epoch": 0.4174390662547202,
      "grad_norm": 0.03955078125,
      "learning_rate": 2.929180355441926e-05,
      "loss": 0.5357,
      "step": 152
    },
    {
      "epoch": 0.42018537590113286,
      "grad_norm": 0.038818359375,
      "learning_rate": 2.9281874473996077e-05,
      "loss": 0.5449,
      "step": 153
    },
    {
      "epoch": 0.4229316855475455,
      "grad_norm": 0.1689453125,
      "learning_rate": 2.9271877979867263e-05,
      "loss": 1.3347,
      "step": 154
    },
    {
      "epoch": 0.4256779951939581,
      "grad_norm": 0.03759765625,
      "learning_rate": 2.926181411921855e-05,
      "loss": 0.4532,
      "step": 155
    },
    {
      "epoch": 0.42842430484037075,
      "grad_norm": 0.03857421875,
      "learning_rate": 2.9251682939553662e-05,
      "loss": 0.5425,
      "step": 156
    },
    {
      "epoch": 0.4311706144867834,
      "grad_norm": 0.047607421875,
      "learning_rate": 2.9241484488694074e-05,
      "loss": 0.4875,
      "step": 157
    },
    {
      "epoch": 0.433916924133196,
      "grad_norm": 0.040283203125,
      "learning_rate": 2.92312188147788e-05,
      "loss": 0.4574,
      "step": 158
    },
    {
      "epoch": 0.43666323377960864,
      "grad_norm": 0.035888671875,
      "learning_rate": 2.9220885966264174e-05,
      "loss": 0.5003,
      "step": 159
    },
    {
      "epoch": 0.4394095434260213,
      "grad_norm": 0.037841796875,
      "learning_rate": 2.9210485991923577e-05,
      "loss": 0.4766,
      "step": 160
    },
    {
      "epoch": 0.44215585307243394,
      "grad_norm": 0.03515625,
      "learning_rate": 2.9200018940847278e-05,
      "loss": 0.3866,
      "step": 161
    },
    {
      "epoch": 0.44490216271884653,
      "grad_norm": 0.035888671875,
      "learning_rate": 2.918948486244214e-05,
      "loss": 0.4401,
      "step": 162
    },
    {
      "epoch": 0.4476484723652592,
      "grad_norm": 0.037841796875,
      "learning_rate": 2.917888380643142e-05,
      "loss": 0.5193,
      "step": 163
    },
    {
      "epoch": 0.4503947820116718,
      "grad_norm": 0.0380859375,
      "learning_rate": 2.916821582285451e-05,
      "loss": 0.4802,
      "step": 164
    },
    {
      "epoch": 0.45314109165808447,
      "grad_norm": 0.037353515625,
      "learning_rate": 2.915748096206674e-05,
      "loss": 0.4693,
      "step": 165
    },
    {
      "epoch": 0.45588740130449706,
      "grad_norm": 0.04052734375,
      "learning_rate": 2.914667927473909e-05,
      "loss": 0.4949,
      "step": 166
    },
    {
      "epoch": 0.4586337109509097,
      "grad_norm": 0.036376953125,
      "learning_rate": 2.9135810811857994e-05,
      "loss": 0.5453,
      "step": 167
    },
    {
      "epoch": 0.46138002059732236,
      "grad_norm": 0.0498046875,
      "learning_rate": 2.912487562472508e-05,
      "loss": 0.4653,
      "step": 168
    },
    {
      "epoch": 0.464126330243735,
      "grad_norm": 0.0380859375,
      "learning_rate": 2.9113873764956917e-05,
      "loss": 0.5032,
      "step": 169
    },
    {
      "epoch": 0.4668726398901476,
      "grad_norm": 0.037109375,
      "learning_rate": 2.91028052844848e-05,
      "loss": 0.4736,
      "step": 170
    },
    {
      "epoch": 0.46961894953656025,
      "grad_norm": 0.038818359375,
      "learning_rate": 2.9091670235554478e-05,
      "loss": 0.4773,
      "step": 171
    },
    {
      "epoch": 0.4723652591829729,
      "grad_norm": 0.041748046875,
      "learning_rate": 2.9080468670725922e-05,
      "loss": 0.5689,
      "step": 172
    },
    {
      "epoch": 0.4751115688293855,
      "grad_norm": 0.04296875,
      "learning_rate": 2.906920064287308e-05,
      "loss": 0.51,
      "step": 173
    },
    {
      "epoch": 0.47785787847579814,
      "grad_norm": 0.03564453125,
      "learning_rate": 2.9057866205183606e-05,
      "loss": 0.446,
      "step": 174
    },
    {
      "epoch": 0.4806041881222108,
      "grad_norm": 0.036376953125,
      "learning_rate": 2.9046465411158634e-05,
      "loss": 0.4956,
      "step": 175
    },
    {
      "epoch": 0.48335049776862343,
      "grad_norm": 0.038818359375,
      "learning_rate": 2.9034998314612516e-05,
      "loss": 0.4963,
      "step": 176
    },
    {
      "epoch": 0.486096807415036,
      "grad_norm": 0.0419921875,
      "learning_rate": 2.902346496967256e-05,
      "loss": 0.4928,
      "step": 177
    },
    {
      "epoch": 0.4888431170614487,
      "grad_norm": 0.0439453125,
      "learning_rate": 2.9011865430778782e-05,
      "loss": 0.4731,
      "step": 178
    },
    {
      "epoch": 0.4915894267078613,
      "grad_norm": 0.03955078125,
      "learning_rate": 2.9000199752683663e-05,
      "loss": 0.5374,
      "step": 179
    },
    {
      "epoch": 0.49433573635427397,
      "grad_norm": 0.040771484375,
      "learning_rate": 2.8988467990451853e-05,
      "loss": 0.6108,
      "step": 180
    },
    {
      "epoch": 0.49708204600068656,
      "grad_norm": 0.037109375,
      "learning_rate": 2.8976670199459953e-05,
      "loss": 0.4189,
      "step": 181
    },
    {
      "epoch": 0.4998283556470992,
      "grad_norm": 0.0380859375,
      "learning_rate": 2.8964806435396227e-05,
      "loss": 0.4773,
      "step": 182
    },
    {
      "epoch": 0.4998283556470992,
      "eval_loss": 0.5154861807823181,
      "eval_runtime": 620.9342,
      "eval_samples_per_second": 14.762,
      "eval_steps_per_second": 14.762,
      "step": 182
    },
    {
      "epoch": 0.5025746652935118,
      "grad_norm": 0.043701171875,
      "learning_rate": 2.8952876754260342e-05,
      "loss": 0.5624,
      "step": 183
    },
    {
      "epoch": 0.5053209749399244,
      "grad_norm": 0.1904296875,
      "learning_rate": 2.8940881212363124e-05,
      "loss": 1.2595,
      "step": 184
    },
    {
      "epoch": 0.5080672845863371,
      "grad_norm": 0.0390625,
      "learning_rate": 2.8928819866326262e-05,
      "loss": 0.6287,
      "step": 185
    },
    {
      "epoch": 0.5108135942327497,
      "grad_norm": 0.035400390625,
      "learning_rate": 2.891669277308206e-05,
      "loss": 0.4508,
      "step": 186
    },
    {
      "epoch": 0.5135599038791624,
      "grad_norm": 0.03662109375,
      "learning_rate": 2.8904499989873166e-05,
      "loss": 0.5141,
      "step": 187
    },
    {
      "epoch": 0.516306213525575,
      "grad_norm": 0.0390625,
      "learning_rate": 2.88922415742523e-05,
      "loss": 0.4496,
      "step": 188
    },
    {
      "epoch": 0.5190525231719877,
      "grad_norm": 0.0361328125,
      "learning_rate": 2.8879917584081975e-05,
      "loss": 0.5467,
      "step": 189
    },
    {
      "epoch": 0.5217988328184002,
      "grad_norm": 0.046875,
      "learning_rate": 2.886752807753424e-05,
      "loss": 0.4188,
      "step": 190
    },
    {
      "epoch": 0.5245451424648129,
      "grad_norm": 0.039794921875,
      "learning_rate": 2.8855073113090395e-05,
      "loss": 0.5347,
      "step": 191
    },
    {
      "epoch": 0.5272914521112255,
      "grad_norm": 0.0400390625,
      "learning_rate": 2.8842552749540708e-05,
      "loss": 0.4117,
      "step": 192
    },
    {
      "epoch": 0.5300377617576382,
      "grad_norm": 0.037841796875,
      "learning_rate": 2.8829967045984155e-05,
      "loss": 0.5413,
      "step": 193
    },
    {
      "epoch": 0.5327840714040508,
      "grad_norm": 0.036865234375,
      "learning_rate": 2.8817316061828126e-05,
      "loss": 0.5683,
      "step": 194
    },
    {
      "epoch": 0.5355303810504635,
      "grad_norm": 0.03955078125,
      "learning_rate": 2.8804599856788154e-05,
      "loss": 0.3851,
      "step": 195
    },
    {
      "epoch": 0.5382766906968761,
      "grad_norm": 0.03662109375,
      "learning_rate": 2.8791818490887628e-05,
      "loss": 0.42,
      "step": 196
    },
    {
      "epoch": 0.5410230003432887,
      "grad_norm": 0.039306640625,
      "learning_rate": 2.8778972024457504e-05,
      "loss": 0.5491,
      "step": 197
    },
    {
      "epoch": 0.5437693099897013,
      "grad_norm": 0.043701171875,
      "learning_rate": 2.876606051813604e-05,
      "loss": 0.5299,
      "step": 198
    },
    {
      "epoch": 0.546515619636114,
      "grad_norm": 0.039794921875,
      "learning_rate": 2.8753084032868494e-05,
      "loss": 0.4881,
      "step": 199
    },
    {
      "epoch": 0.5492619292825266,
      "grad_norm": 0.03857421875,
      "learning_rate": 2.8740042629906833e-05,
      "loss": 0.4698,
      "step": 200
    },
    {
      "epoch": 0.5520082389289392,
      "grad_norm": 0.042724609375,
      "learning_rate": 2.8726936370809455e-05,
      "loss": 0.5685,
      "step": 201
    },
    {
      "epoch": 0.5547545485753519,
      "grad_norm": 0.040771484375,
      "learning_rate": 2.8713765317440895e-05,
      "loss": 0.5536,
      "step": 202
    },
    {
      "epoch": 0.5575008582217645,
      "grad_norm": 0.040283203125,
      "learning_rate": 2.870052953197152e-05,
      "loss": 0.4891,
      "step": 203
    },
    {
      "epoch": 0.5602471678681772,
      "grad_norm": 0.048828125,
      "learning_rate": 2.8687229076877274e-05,
      "loss": 0.4182,
      "step": 204
    },
    {
      "epoch": 0.5629934775145897,
      "grad_norm": 0.0400390625,
      "learning_rate": 2.867386401493932e-05,
      "loss": 0.507,
      "step": 205
    },
    {
      "epoch": 0.5657397871610024,
      "grad_norm": 0.03466796875,
      "learning_rate": 2.8660434409243817e-05,
      "loss": 0.4052,
      "step": 206
    },
    {
      "epoch": 0.568486096807415,
      "grad_norm": 0.044677734375,
      "learning_rate": 2.8646940323181553e-05,
      "loss": 0.4503,
      "step": 207
    },
    {
      "epoch": 0.5712324064538277,
      "grad_norm": 0.037841796875,
      "learning_rate": 2.86333818204477e-05,
      "loss": 0.4234,
      "step": 208
    },
    {
      "epoch": 0.5739787161002403,
      "grad_norm": 0.03857421875,
      "learning_rate": 2.8619758965041488e-05,
      "loss": 0.5319,
      "step": 209
    },
    {
      "epoch": 0.576725025746653,
      "grad_norm": 0.03662109375,
      "learning_rate": 2.8606071821265888e-05,
      "loss": 0.5282,
      "step": 210
    },
    {
      "epoch": 0.5794713353930656,
      "grad_norm": 0.04248046875,
      "learning_rate": 2.8592320453727356e-05,
      "loss": 0.4596,
      "step": 211
    },
    {
      "epoch": 0.5822176450394781,
      "grad_norm": 0.037841796875,
      "learning_rate": 2.857850492733548e-05,
      "loss": 0.5258,
      "step": 212
    },
    {
      "epoch": 0.5849639546858908,
      "grad_norm": 0.0380859375,
      "learning_rate": 2.856462530730269e-05,
      "loss": 0.4836,
      "step": 213
    },
    {
      "epoch": 0.5877102643323034,
      "grad_norm": 0.040771484375,
      "learning_rate": 2.855068165914397e-05,
      "loss": 0.4973,
      "step": 214
    },
    {
      "epoch": 0.5904565739787161,
      "grad_norm": 0.03857421875,
      "learning_rate": 2.8536674048676506e-05,
      "loss": 0.5643,
      "step": 215
    },
    {
      "epoch": 0.5932028836251287,
      "grad_norm": 0.04150390625,
      "learning_rate": 2.8522602542019425e-05,
      "loss": 0.476,
      "step": 216
    },
    {
      "epoch": 0.5959491932715414,
      "grad_norm": 0.0390625,
      "learning_rate": 2.850846720559345e-05,
      "loss": 0.4767,
      "step": 217
    },
    {
      "epoch": 0.598695502917954,
      "grad_norm": 0.044189453125,
      "learning_rate": 2.8494268106120586e-05,
      "loss": 0.5567,
      "step": 218
    },
    {
      "epoch": 0.6014418125643667,
      "grad_norm": 0.0439453125,
      "learning_rate": 2.8480005310623823e-05,
      "loss": 0.536,
      "step": 219
    },
    {
      "epoch": 0.6041881222107792,
      "grad_norm": 0.04248046875,
      "learning_rate": 2.8465678886426814e-05,
      "loss": 0.4813,
      "step": 220
    },
    {
      "epoch": 0.6069344318571919,
      "grad_norm": 0.038818359375,
      "learning_rate": 2.845128890115355e-05,
      "loss": 0.4215,
      "step": 221
    },
    {
      "epoch": 0.6096807415036045,
      "grad_norm": 0.03857421875,
      "learning_rate": 2.8436835422728036e-05,
      "loss": 0.547,
      "step": 222
    },
    {
      "epoch": 0.6124270511500172,
      "grad_norm": 0.0380859375,
      "learning_rate": 2.8422318519373996e-05,
      "loss": 0.4629,
      "step": 223
    },
    {
      "epoch": 0.6151733607964298,
      "grad_norm": 0.052490234375,
      "learning_rate": 2.8407738259614524e-05,
      "loss": 0.4823,
      "step": 224
    },
    {
      "epoch": 0.6179196704428425,
      "grad_norm": 0.04052734375,
      "learning_rate": 2.8393094712271772e-05,
      "loss": 0.5568,
      "step": 225
    },
    {
      "epoch": 0.6206659800892551,
      "grad_norm": 0.03857421875,
      "learning_rate": 2.8378387946466623e-05,
      "loss": 0.4709,
      "step": 226
    },
    {
      "epoch": 0.6234122897356676,
      "grad_norm": 0.0439453125,
      "learning_rate": 2.8363618031618364e-05,
      "loss": 0.4205,
      "step": 227
    },
    {
      "epoch": 0.6261585993820803,
      "grad_norm": 0.0380859375,
      "learning_rate": 2.8348785037444366e-05,
      "loss": 0.4985,
      "step": 228
    },
    {
      "epoch": 0.6289049090284929,
      "grad_norm": 0.039794921875,
      "learning_rate": 2.8333889033959746e-05,
      "loss": 0.4527,
      "step": 229
    },
    {
      "epoch": 0.6316512186749056,
      "grad_norm": 0.03857421875,
      "learning_rate": 2.8318930091477037e-05,
      "loss": 0.582,
      "step": 230
    },
    {
      "epoch": 0.6343975283213182,
      "grad_norm": 0.04296875,
      "learning_rate": 2.8303908280605854e-05,
      "loss": 0.5028,
      "step": 231
    },
    {
      "epoch": 0.6371438379677309,
      "grad_norm": 0.03857421875,
      "learning_rate": 2.8288823672252586e-05,
      "loss": 0.5349,
      "step": 232
    },
    {
      "epoch": 0.6398901476141435,
      "grad_norm": 0.038818359375,
      "learning_rate": 2.827367633762001e-05,
      "loss": 0.4251,
      "step": 233
    },
    {
      "epoch": 0.6426364572605562,
      "grad_norm": 0.041259765625,
      "learning_rate": 2.825846634820701e-05,
      "loss": 0.5079,
      "step": 234
    },
    {
      "epoch": 0.6453827669069687,
      "grad_norm": 0.353515625,
      "learning_rate": 2.824319377580821e-05,
      "loss": 1.2174,
      "step": 235
    },
    {
      "epoch": 0.6481290765533814,
      "grad_norm": 0.038330078125,
      "learning_rate": 2.8227858692513626e-05,
      "loss": 0.4188,
      "step": 236
    },
    {
      "epoch": 0.650875386199794,
      "grad_norm": 0.042236328125,
      "learning_rate": 2.821246117070835e-05,
      "loss": 0.4767,
      "step": 237
    },
    {
      "epoch": 0.6536216958462067,
      "grad_norm": 0.03662109375,
      "learning_rate": 2.8197001283072205e-05,
      "loss": 0.4736,
      "step": 238
    },
    {
      "epoch": 0.6563680054926193,
      "grad_norm": 0.035400390625,
      "learning_rate": 2.8181479102579383e-05,
      "loss": 0.388,
      "step": 239
    },
    {
      "epoch": 0.659114315139032,
      "grad_norm": 0.0439453125,
      "learning_rate": 2.8165894702498116e-05,
      "loss": 0.6023,
      "step": 240
    },
    {
      "epoch": 0.6618606247854446,
      "grad_norm": 0.03955078125,
      "learning_rate": 2.8150248156390327e-05,
      "loss": 0.5319,
      "step": 241
    },
    {
      "epoch": 0.6646069344318571,
      "grad_norm": 0.037109375,
      "learning_rate": 2.8134539538111286e-05,
      "loss": 0.5133,
      "step": 242
    },
    {
      "epoch": 0.6673532440782698,
      "grad_norm": 0.039794921875,
      "learning_rate": 2.8118768921809258e-05,
      "loss": 0.4813,
      "step": 243
    },
    {
      "epoch": 0.6700995537246824,
      "grad_norm": 0.03857421875,
      "learning_rate": 2.8102936381925143e-05,
      "loss": 0.5085,
      "step": 244
    },
    {
      "epoch": 0.6728458633710951,
      "grad_norm": 0.04248046875,
      "learning_rate": 2.8087041993192148e-05,
      "loss": 0.4245,
      "step": 245
    },
    {
      "epoch": 0.6755921730175077,
      "grad_norm": 0.045166015625,
      "learning_rate": 2.8071085830635404e-05,
      "loss": 0.5026,
      "step": 246
    },
    {
      "epoch": 0.6783384826639204,
      "grad_norm": 0.0439453125,
      "learning_rate": 2.8055067969571647e-05,
      "loss": 0.5615,
      "step": 247
    },
    {
      "epoch": 0.681084792310333,
      "grad_norm": 0.037841796875,
      "learning_rate": 2.803898848560883e-05,
      "loss": 0.4929,
      "step": 248
    },
    {
      "epoch": 0.6838311019567456,
      "grad_norm": 0.042724609375,
      "learning_rate": 2.802284745464579e-05,
      "loss": 0.5747,
      "step": 249
    },
    {
      "epoch": 0.6865774116031582,
      "grad_norm": 0.041015625,
      "learning_rate": 2.800664495287187e-05,
      "loss": 0.4181,
      "step": 250
    },
    {
      "epoch": 0.6893237212495709,
      "grad_norm": 0.040283203125,
      "learning_rate": 2.7990381056766583e-05,
      "loss": 0.548,
      "step": 251
    },
    {
      "epoch": 0.6920700308959835,
      "grad_norm": 0.043212890625,
      "learning_rate": 2.797405584309922e-05,
      "loss": 0.5344,
      "step": 252
    },
    {
      "epoch": 0.6948163405423962,
      "grad_norm": 0.03955078125,
      "learning_rate": 2.7957669388928517e-05,
      "loss": 0.4484,
      "step": 253
    },
    {
      "epoch": 0.6975626501888088,
      "grad_norm": 0.03857421875,
      "learning_rate": 2.7941221771602278e-05,
      "loss": 0.5194,
      "step": 254
    },
    {
      "epoch": 0.7003089598352215,
      "grad_norm": 0.0390625,
      "learning_rate": 2.7924713068757004e-05,
      "loss": 0.4297,
      "step": 255
    },
    {
      "epoch": 0.7030552694816341,
      "grad_norm": 0.0400390625,
      "learning_rate": 2.7908143358317545e-05,
      "loss": 0.4723,
      "step": 256
    },
    {
      "epoch": 0.7058015791280466,
      "grad_norm": 0.03662109375,
      "learning_rate": 2.7891512718496712e-05,
      "loss": 0.4401,
      "step": 257
    },
    {
      "epoch": 0.7085478887744593,
      "grad_norm": 0.041015625,
      "learning_rate": 2.7874821227794915e-05,
      "loss": 0.5961,
      "step": 258
    },
    {
      "epoch": 0.7112941984208719,
      "grad_norm": 0.039794921875,
      "learning_rate": 2.78580689649998e-05,
      "loss": 0.5483,
      "step": 259
    },
    {
      "epoch": 0.7140405080672846,
      "grad_norm": 0.05126953125,
      "learning_rate": 2.7841256009185876e-05,
      "loss": 0.493,
      "step": 260
    },
    {
      "epoch": 0.7167868177136972,
      "grad_norm": 0.042236328125,
      "learning_rate": 2.782438243971412e-05,
      "loss": 0.5366,
      "step": 261
    },
    {
      "epoch": 0.7195331273601099,
      "grad_norm": 0.0380859375,
      "learning_rate": 2.7807448336231635e-05,
      "loss": 0.3991,
      "step": 262
    },
    {
      "epoch": 0.7222794370065225,
      "grad_norm": 0.048095703125,
      "learning_rate": 2.7790453778671248e-05,
      "loss": 0.528,
      "step": 263
    },
    {
      "epoch": 0.7250257466529351,
      "grad_norm": 0.0458984375,
      "learning_rate": 2.7773398847251152e-05,
      "loss": 0.4221,
      "step": 264
    },
    {
      "epoch": 0.7277720562993477,
      "grad_norm": 0.03955078125,
      "learning_rate": 2.7756283622474515e-05,
      "loss": 0.4483,
      "step": 265
    },
    {
      "epoch": 0.7305183659457604,
      "grad_norm": 0.0400390625,
      "learning_rate": 2.77391081851291e-05,
      "loss": 0.4633,
      "step": 266
    },
    {
      "epoch": 0.733264675592173,
      "grad_norm": 0.04248046875,
      "learning_rate": 2.7721872616286888e-05,
      "loss": 0.5595,
      "step": 267
    },
    {
      "epoch": 0.7360109852385857,
      "grad_norm": 0.0380859375,
      "learning_rate": 2.7704576997303694e-05,
      "loss": 0.5091,
      "step": 268
    },
    {
      "epoch": 0.7387572948849983,
      "grad_norm": 0.0361328125,
      "learning_rate": 2.768722140981879e-05,
      "loss": 0.4357,
      "step": 269
    },
    {
      "epoch": 0.741503604531411,
      "grad_norm": 0.03955078125,
      "learning_rate": 2.766980593575451e-05,
      "loss": 0.4608,
      "step": 270
    },
    {
      "epoch": 0.7442499141778236,
      "grad_norm": 0.03662109375,
      "learning_rate": 2.765233065731586e-05,
      "loss": 0.4593,
      "step": 271
    },
    {
      "epoch": 0.7469962238242361,
      "grad_norm": 0.04150390625,
      "learning_rate": 2.7634795656990143e-05,
      "loss": 0.5097,
      "step": 272
    },
    {
      "epoch": 0.7497425334706488,
      "grad_norm": 0.039794921875,
      "learning_rate": 2.761720101754656e-05,
      "loss": 0.4375,
      "step": 273
    },
    {
      "epoch": 0.7497425334706488,
      "eval_loss": 0.5116191506385803,
      "eval_runtime": 620.1922,
      "eval_samples_per_second": 14.779,
      "eval_steps_per_second": 14.779,
      "step": 273
    },
    {
      "epoch": 0.7524888431170614,
      "grad_norm": 0.039306640625,
      "learning_rate": 2.7599546822035817e-05,
      "loss": 0.5089,
      "step": 274
    },
    {
      "epoch": 0.7552351527634741,
      "grad_norm": 0.04150390625,
      "learning_rate": 2.758183315378976e-05,
      "loss": 0.5961,
      "step": 275
    },
    {
      "epoch": 0.7579814624098867,
      "grad_norm": 0.041259765625,
      "learning_rate": 2.7564060096420925e-05,
      "loss": 0.4763,
      "step": 276
    },
    {
      "epoch": 0.7607277720562994,
      "grad_norm": 0.0419921875,
      "learning_rate": 2.754622773382221e-05,
      "loss": 0.5076,
      "step": 277
    },
    {
      "epoch": 0.763474081702712,
      "grad_norm": 0.0439453125,
      "learning_rate": 2.7528336150166436e-05,
      "loss": 0.4411,
      "step": 278
    },
    {
      "epoch": 0.7662203913491246,
      "grad_norm": 0.04736328125,
      "learning_rate": 2.751038542990595e-05,
      "loss": 0.5316,
      "step": 279
    },
    {
      "epoch": 0.7689667009955372,
      "grad_norm": 0.0439453125,
      "learning_rate": 2.7492375657772254e-05,
      "loss": 0.4153,
      "step": 280
    },
    {
      "epoch": 0.7717130106419499,
      "grad_norm": 0.039794921875,
      "learning_rate": 2.7474306918775576e-05,
      "loss": 0.5106,
      "step": 281
    },
    {
      "epoch": 0.7744593202883625,
      "grad_norm": 0.04638671875,
      "learning_rate": 2.745617929820449e-05,
      "loss": 0.474,
      "step": 282
    },
    {
      "epoch": 0.7772056299347752,
      "grad_norm": 0.142578125,
      "learning_rate": 2.74379928816255e-05,
      "loss": 1.2147,
      "step": 283
    },
    {
      "epoch": 0.7799519395811878,
      "grad_norm": 0.044677734375,
      "learning_rate": 2.7419747754882637e-05,
      "loss": 0.5727,
      "step": 284
    },
    {
      "epoch": 0.7826982492276005,
      "grad_norm": 0.046142578125,
      "learning_rate": 2.740144400409707e-05,
      "loss": 0.5203,
      "step": 285
    },
    {
      "epoch": 0.7854445588740131,
      "grad_norm": 0.041259765625,
      "learning_rate": 2.738308171566667e-05,
      "loss": 0.5998,
      "step": 286
    },
    {
      "epoch": 0.7881908685204256,
      "grad_norm": 0.043212890625,
      "learning_rate": 2.7364660976265624e-05,
      "loss": 0.5133,
      "step": 287
    },
    {
      "epoch": 0.7909371781668383,
      "grad_norm": 0.038818359375,
      "learning_rate": 2.7346181872844037e-05,
      "loss": 0.4711,
      "step": 288
    },
    {
      "epoch": 0.7936834878132509,
      "grad_norm": 0.041748046875,
      "learning_rate": 2.7327644492627487e-05,
      "loss": 0.5563,
      "step": 289
    },
    {
      "epoch": 0.7964297974596636,
      "grad_norm": 0.042236328125,
      "learning_rate": 2.7309048923116635e-05,
      "loss": 0.4684,
      "step": 290
    },
    {
      "epoch": 0.7991761071060762,
      "grad_norm": 0.037841796875,
      "learning_rate": 2.729039525208682e-05,
      "loss": 0.4581,
      "step": 291
    },
    {
      "epoch": 0.8019224167524889,
      "grad_norm": 0.03759765625,
      "learning_rate": 2.7271683567587608e-05,
      "loss": 0.4502,
      "step": 292
    },
    {
      "epoch": 0.8046687263989015,
      "grad_norm": 0.0390625,
      "learning_rate": 2.7252913957942435e-05,
      "loss": 0.564,
      "step": 293
    },
    {
      "epoch": 0.8074150360453141,
      "grad_norm": 0.041259765625,
      "learning_rate": 2.723408651174813e-05,
      "loss": 0.4386,
      "step": 294
    },
    {
      "epoch": 0.8101613456917267,
      "grad_norm": 0.039794921875,
      "learning_rate": 2.7215201317874537e-05,
      "loss": 0.5623,
      "step": 295
    },
    {
      "epoch": 0.8129076553381394,
      "grad_norm": 0.043701171875,
      "learning_rate": 2.7196258465464087e-05,
      "loss": 0.5303,
      "step": 296
    },
    {
      "epoch": 0.815653964984552,
      "grad_norm": 0.04248046875,
      "learning_rate": 2.7177258043931354e-05,
      "loss": 0.5094,
      "step": 297
    },
    {
      "epoch": 0.8184002746309647,
      "grad_norm": 0.038818359375,
      "learning_rate": 2.7158200142962665e-05,
      "loss": 0.502,
      "step": 298
    },
    {
      "epoch": 0.8211465842773773,
      "grad_norm": 0.044921875,
      "learning_rate": 2.7139084852515665e-05,
      "loss": 0.4744,
      "step": 299
    },
    {
      "epoch": 0.82389289392379,
      "grad_norm": 0.039306640625,
      "learning_rate": 2.7119912262818878e-05,
      "loss": 0.5895,
      "step": 300
    },
    {
      "epoch": 0.8266392035702025,
      "grad_norm": 0.040771484375,
      "learning_rate": 2.7100682464371306e-05,
      "loss": 0.3948,
      "step": 301
    },
    {
      "epoch": 0.8293855132166151,
      "grad_norm": 0.038330078125,
      "learning_rate": 2.7081395547941986e-05,
      "loss": 0.4514,
      "step": 302
    },
    {
      "epoch": 0.8321318228630278,
      "grad_norm": 0.04443359375,
      "learning_rate": 2.7062051604569562e-05,
      "loss": 0.4525,
      "step": 303
    },
    {
      "epoch": 0.8348781325094404,
      "grad_norm": 0.038818359375,
      "learning_rate": 2.7042650725561854e-05,
      "loss": 0.4161,
      "step": 304
    },
    {
      "epoch": 0.8376244421558531,
      "grad_norm": 0.042724609375,
      "learning_rate": 2.7023193002495447e-05,
      "loss": 0.5065,
      "step": 305
    },
    {
      "epoch": 0.8403707518022657,
      "grad_norm": 0.16796875,
      "learning_rate": 2.7003678527215224e-05,
      "loss": 1.3831,
      "step": 306
    },
    {
      "epoch": 0.8431170614486784,
      "grad_norm": 0.0419921875,
      "learning_rate": 2.6984107391833972e-05,
      "loss": 0.5368,
      "step": 307
    },
    {
      "epoch": 0.845863371095091,
      "grad_norm": 0.037353515625,
      "learning_rate": 2.6964479688731897e-05,
      "loss": 0.4434,
      "step": 308
    },
    {
      "epoch": 0.8486096807415036,
      "grad_norm": 0.04541015625,
      "learning_rate": 2.694479551055625e-05,
      "loss": 0.5286,
      "step": 309
    },
    {
      "epoch": 0.8513559903879162,
      "grad_norm": 0.03759765625,
      "learning_rate": 2.6925054950220834e-05,
      "loss": 0.4054,
      "step": 310
    },
    {
      "epoch": 0.8541023000343289,
      "grad_norm": 0.05419921875,
      "learning_rate": 2.69052581009056e-05,
      "loss": 0.3735,
      "step": 311
    },
    {
      "epoch": 0.8568486096807415,
      "grad_norm": 0.039794921875,
      "learning_rate": 2.68854050560562e-05,
      "loss": 0.5696,
      "step": 312
    },
    {
      "epoch": 0.8595949193271541,
      "grad_norm": 0.041259765625,
      "learning_rate": 2.6865495909383525e-05,
      "loss": 0.4851,
      "step": 313
    },
    {
      "epoch": 0.8623412289735668,
      "grad_norm": 0.042724609375,
      "learning_rate": 2.684553075486329e-05,
      "loss": 0.5755,
      "step": 314
    },
    {
      "epoch": 0.8650875386199794,
      "grad_norm": 0.04150390625,
      "learning_rate": 2.682550968673558e-05,
      "loss": 0.5376,
      "step": 315
    },
    {
      "epoch": 0.867833848266392,
      "grad_norm": 0.040283203125,
      "learning_rate": 2.6805432799504407e-05,
      "loss": 0.5374,
      "step": 316
    },
    {
      "epoch": 0.8705801579128046,
      "grad_norm": 0.037841796875,
      "learning_rate": 2.6785300187937264e-05,
      "loss": 0.421,
      "step": 317
    },
    {
      "epoch": 0.8733264675592173,
      "grad_norm": 0.035888671875,
      "learning_rate": 2.6765111947064654e-05,
      "loss": 0.4206,
      "step": 318
    },
    {
      "epoch": 0.8760727772056299,
      "grad_norm": 0.046630859375,
      "learning_rate": 2.6744868172179692e-05,
      "loss": 0.5895,
      "step": 319
    },
    {
      "epoch": 0.8788190868520426,
      "grad_norm": 0.041259765625,
      "learning_rate": 2.672456895883761e-05,
      "loss": 0.4784,
      "step": 320
    },
    {
      "epoch": 0.8815653964984552,
      "grad_norm": 0.039794921875,
      "learning_rate": 2.670421440285533e-05,
      "loss": 0.4898,
      "step": 321
    },
    {
      "epoch": 0.8843117061448679,
      "grad_norm": 0.045654296875,
      "learning_rate": 2.6683804600310997e-05,
      "loss": 0.6258,
      "step": 322
    },
    {
      "epoch": 0.8870580157912805,
      "grad_norm": 0.044677734375,
      "learning_rate": 2.6663339647543528e-05,
      "loss": 0.5587,
      "step": 323
    },
    {
      "epoch": 0.8898043254376931,
      "grad_norm": 0.038330078125,
      "learning_rate": 2.664281964115218e-05,
      "loss": 0.4539,
      "step": 324
    },
    {
      "epoch": 0.8925506350841057,
      "grad_norm": 0.037353515625,
      "learning_rate": 2.6622244677996058e-05,
      "loss": 0.4652,
      "step": 325
    },
    {
      "epoch": 0.8952969447305184,
      "grad_norm": 0.037841796875,
      "learning_rate": 2.660161485519368e-05,
      "loss": 0.4624,
      "step": 326
    },
    {
      "epoch": 0.898043254376931,
      "grad_norm": 0.039306640625,
      "learning_rate": 2.6580930270122524e-05,
      "loss": 0.5089,
      "step": 327
    },
    {
      "epoch": 0.9007895640233436,
      "grad_norm": 0.044921875,
      "learning_rate": 2.6560191020418545e-05,
      "loss": 0.4246,
      "step": 328
    },
    {
      "epoch": 0.9035358736697563,
      "grad_norm": 0.044189453125,
      "learning_rate": 2.6539397203975732e-05,
      "loss": 0.516,
      "step": 329
    },
    {
      "epoch": 0.9062821833161689,
      "grad_norm": 0.041259765625,
      "learning_rate": 2.6518548918945646e-05,
      "loss": 0.5008,
      "step": 330
    },
    {
      "epoch": 0.9090284929625815,
      "grad_norm": 0.04052734375,
      "learning_rate": 2.6497646263736943e-05,
      "loss": 0.5195,
      "step": 331
    },
    {
      "epoch": 0.9117748026089941,
      "grad_norm": 0.044189453125,
      "learning_rate": 2.6476689337014925e-05,
      "loss": 0.5701,
      "step": 332
    },
    {
      "epoch": 0.9145211122554068,
      "grad_norm": 0.043212890625,
      "learning_rate": 2.6455678237701072e-05,
      "loss": 0.5766,
      "step": 333
    },
    {
      "epoch": 0.9172674219018194,
      "grad_norm": 0.03955078125,
      "learning_rate": 2.643461306497256e-05,
      "loss": 0.4613,
      "step": 334
    },
    {
      "epoch": 0.9200137315482321,
      "grad_norm": 0.041748046875,
      "learning_rate": 2.641349391826182e-05,
      "loss": 0.4347,
      "step": 335
    },
    {
      "epoch": 0.9227600411946447,
      "grad_norm": 0.045166015625,
      "learning_rate": 2.6392320897256034e-05,
      "loss": 0.4371,
      "step": 336
    },
    {
      "epoch": 0.9255063508410574,
      "grad_norm": 0.04052734375,
      "learning_rate": 2.637109410189669e-05,
      "loss": 0.5219,
      "step": 337
    },
    {
      "epoch": 0.92825266048747,
      "grad_norm": 0.040283203125,
      "learning_rate": 2.6349813632379103e-05,
      "loss": 0.5435,
      "step": 338
    },
    {
      "epoch": 0.9309989701338826,
      "grad_norm": 0.04248046875,
      "learning_rate": 2.6328479589151953e-05,
      "loss": 0.4764,
      "step": 339
    },
    {
      "epoch": 0.9337452797802952,
      "grad_norm": 0.042236328125,
      "learning_rate": 2.6307092072916786e-05,
      "loss": 0.4664,
      "step": 340
    },
    {
      "epoch": 0.9364915894267078,
      "grad_norm": 0.045166015625,
      "learning_rate": 2.628565118462756e-05,
      "loss": 0.4723,
      "step": 341
    },
    {
      "epoch": 0.9392378990731205,
      "grad_norm": 0.041259765625,
      "learning_rate": 2.626415702549015e-05,
      "loss": 0.5179,
      "step": 342
    },
    {
      "epoch": 0.9419842087195331,
      "grad_norm": 0.0419921875,
      "learning_rate": 2.62426096969619e-05,
      "loss": 0.5736,
      "step": 343
    },
    {
      "epoch": 0.9447305183659458,
      "grad_norm": 0.04541015625,
      "learning_rate": 2.6221009300751113e-05,
      "loss": 0.5238,
      "step": 344
    },
    {
      "epoch": 0.9474768280123584,
      "grad_norm": 0.04248046875,
      "learning_rate": 2.6199355938816586e-05,
      "loss": 0.4591,
      "step": 345
    },
    {
      "epoch": 0.950223137658771,
      "grad_norm": 0.040771484375,
      "learning_rate": 2.6177649713367136e-05,
      "loss": 0.5288,
      "step": 346
    },
    {
      "epoch": 0.9529694473051836,
      "grad_norm": 0.044677734375,
      "learning_rate": 2.6155890726861084e-05,
      "loss": 0.5066,
      "step": 347
    },
    {
      "epoch": 0.9557157569515963,
      "grad_norm": 0.0673828125,
      "learning_rate": 2.613407908200582e-05,
      "loss": 0.4485,
      "step": 348
    },
    {
      "epoch": 0.9584620665980089,
      "grad_norm": 0.04736328125,
      "learning_rate": 2.6112214881757285e-05,
      "loss": 0.5076,
      "step": 349
    },
    {
      "epoch": 0.9612083762444216,
      "grad_norm": 0.044189453125,
      "learning_rate": 2.6090298229319477e-05,
      "loss": 0.5024,
      "step": 350
    },
    {
      "epoch": 0.9639546858908342,
      "grad_norm": 0.043212890625,
      "learning_rate": 2.6068329228144016e-05,
      "loss": 0.4839,
      "step": 351
    },
    {
      "epoch": 0.9667009955372469,
      "grad_norm": 0.04150390625,
      "learning_rate": 2.604630798192959e-05,
      "loss": 0.5425,
      "step": 352
    },
    {
      "epoch": 0.9694473051836594,
      "grad_norm": 0.04150390625,
      "learning_rate": 2.60242345946215e-05,
      "loss": 0.4468,
      "step": 353
    },
    {
      "epoch": 0.972193614830072,
      "grad_norm": 0.045166015625,
      "learning_rate": 2.6002109170411178e-05,
      "loss": 0.5624,
      "step": 354
    },
    {
      "epoch": 0.9749399244764847,
      "grad_norm": 0.03759765625,
      "learning_rate": 2.597993181373567e-05,
      "loss": 0.3949,
      "step": 355
    },
    {
      "epoch": 0.9776862341228973,
      "grad_norm": 0.0390625,
      "learning_rate": 2.5957702629277154e-05,
      "loss": 0.5243,
      "step": 356
    },
    {
      "epoch": 0.98043254376931,
      "grad_norm": 0.04296875,
      "learning_rate": 2.593542172196246e-05,
      "loss": 0.574,
      "step": 357
    },
    {
      "epoch": 0.9831788534157226,
      "grad_norm": 0.050048828125,
      "learning_rate": 2.5913089196962547e-05,
      "loss": 0.4708,
      "step": 358
    },
    {
      "epoch": 0.9859251630621353,
      "grad_norm": 0.044189453125,
      "learning_rate": 2.5890705159692036e-05,
      "loss": 0.4344,
      "step": 359
    },
    {
      "epoch": 0.9886714727085479,
      "grad_norm": 0.041748046875,
      "learning_rate": 2.5868269715808685e-05,
      "loss": 0.4977,
      "step": 360
    },
    {
      "epoch": 0.9914177823549605,
      "grad_norm": 0.04248046875,
      "learning_rate": 2.58457829712129e-05,
      "loss": 0.551,
      "step": 361
    },
    {
      "epoch": 0.9941640920013731,
      "grad_norm": 0.039794921875,
      "learning_rate": 2.5823245032047255e-05,
      "loss": 0.5069,
      "step": 362
    },
    {
      "epoch": 0.9969104016477858,
      "grad_norm": 0.0419921875,
      "learning_rate": 2.5800656004695962e-05,
      "loss": 0.5246,
      "step": 363
    },
    {
      "epoch": 0.9996567112941984,
      "grad_norm": 0.048583984375,
      "learning_rate": 2.5778015995784385e-05,
      "loss": 0.6325,
      "step": 364
    },
    {
      "epoch": 0.9996567112941984,
      "eval_loss": 0.509181559085846,
      "eval_runtime": 618.8303,
      "eval_samples_per_second": 14.812,
      "eval_steps_per_second": 14.812,
      "step": 364
    },
    {
      "epoch": 1.002403020940611,
      "grad_norm": 0.04248046875,
      "learning_rate": 2.575532511217852e-05,
      "loss": 0.607,
      "step": 365
    },
    {
      "epoch": 1.0051493305870236,
      "grad_norm": 0.043701171875,
      "learning_rate": 2.5732583460984527e-05,
      "loss": 0.5572,
      "step": 366
    },
    {
      "epoch": 1.0078956402334363,
      "grad_norm": 0.201171875,
      "learning_rate": 2.5709791149548184e-05,
      "loss": 1.256,
      "step": 367
    },
    {
      "epoch": 1.010641949879849,
      "grad_norm": 0.04248046875,
      "learning_rate": 2.56869482854544e-05,
      "loss": 0.4604,
      "step": 368
    },
    {
      "epoch": 1.0020597322348095,
      "grad_norm": 0.04345703125,
      "learning_rate": 2.5664054976526702e-05,
      "loss": 0.5396,
      "step": 369
    },
    {
      "epoch": 1.0048060418812221,
      "grad_norm": 0.04541015625,
      "learning_rate": 2.564111133082674e-05,
      "loss": 0.4803,
      "step": 370
    },
    {
      "epoch": 1.0075523515276348,
      "grad_norm": 0.05712890625,
      "learning_rate": 2.561811745665374e-05,
      "loss": 0.3781,
      "step": 371
    },
    {
      "epoch": 1.0102986611740474,
      "grad_norm": 0.041015625,
      "learning_rate": 2.5595073462544046e-05,
      "loss": 0.4143,
      "step": 372
    },
    {
      "epoch": 1.01304497082046,
      "grad_norm": 0.043212890625,
      "learning_rate": 2.5571979457270565e-05,
      "loss": 0.4698,
      "step": 373
    },
    {
      "epoch": 1.0157912804668727,
      "grad_norm": 0.0400390625,
      "learning_rate": 2.5548835549842274e-05,
      "loss": 0.5101,
      "step": 374
    },
    {
      "epoch": 1.0185375901132854,
      "grad_norm": 0.039794921875,
      "learning_rate": 2.5525641849503685e-05,
      "loss": 0.4252,
      "step": 375
    },
    {
      "epoch": 1.0212838997596978,
      "grad_norm": 0.0458984375,
      "learning_rate": 2.5502398465734357e-05,
      "loss": 0.5116,
      "step": 376
    },
    {
      "epoch": 1.0240302094061104,
      "grad_norm": 0.040283203125,
      "learning_rate": 2.5479105508248373e-05,
      "loss": 0.4816,
      "step": 377
    },
    {
      "epoch": 1.026776519052523,
      "grad_norm": 0.044921875,
      "learning_rate": 2.54557630869938e-05,
      "loss": 0.4521,
      "step": 378
    },
    {
      "epoch": 1.0295228286989357,
      "grad_norm": 0.041748046875,
      "learning_rate": 2.543237131215219e-05,
      "loss": 0.4769,
      "step": 379
    },
    {
      "epoch": 1.0322691383453484,
      "grad_norm": 0.044189453125,
      "learning_rate": 2.5408930294138065e-05,
      "loss": 0.5011,
      "step": 380
    },
    {
      "epoch": 1.035015447991761,
      "grad_norm": 0.0390625,
      "learning_rate": 2.538544014359837e-05,
      "loss": 0.407,
      "step": 381
    },
    {
      "epoch": 1.0377617576381737,
      "grad_norm": 0.038330078125,
      "learning_rate": 2.536190097141197e-05,
      "loss": 0.4991,
      "step": 382
    },
    {
      "epoch": 1.0405080672845863,
      "grad_norm": 0.04248046875,
      "learning_rate": 2.5338312888689137e-05,
      "loss": 0.5129,
      "step": 383
    },
    {
      "epoch": 1.043254376930999,
      "grad_norm": 0.043212890625,
      "learning_rate": 2.5314676006771e-05,
      "loss": 0.4409,
      "step": 384
    },
    {
      "epoch": 1.0460006865774116,
      "grad_norm": 0.038818359375,
      "learning_rate": 2.529099043722903e-05,
      "loss": 0.542,
      "step": 385
    },
    {
      "epoch": 1.0487469962238243,
      "grad_norm": 0.041748046875,
      "learning_rate": 2.526725629186452e-05,
      "loss": 0.5767,
      "step": 386
    },
    {
      "epoch": 1.051493305870237,
      "grad_norm": 0.04345703125,
      "learning_rate": 2.5243473682708057e-05,
      "loss": 0.5457,
      "step": 387
    },
    {
      "epoch": 1.0542396155166496,
      "grad_norm": 0.0380859375,
      "learning_rate": 2.5219642722018975e-05,
      "loss": 0.4768,
      "step": 388
    },
    {
      "epoch": 1.0569859251630622,
      "grad_norm": 0.04345703125,
      "learning_rate": 2.5195763522284848e-05,
      "loss": 0.58,
      "step": 389
    },
    {
      "epoch": 1.0597322348094749,
      "grad_norm": 0.041259765625,
      "learning_rate": 2.5171836196220946e-05,
      "loss": 0.5176,
      "step": 390
    },
    {
      "epoch": 1.0624785444558873,
      "grad_norm": 0.047607421875,
      "learning_rate": 2.51478608567697e-05,
      "loss": 0.4992,
      "step": 391
    },
    {
      "epoch": 1.0652248541023,
      "grad_norm": 0.036865234375,
      "learning_rate": 2.512383761710019e-05,
      "loss": 0.5167,
      "step": 392
    },
    {
      "epoch": 1.0679711637487126,
      "grad_norm": 0.162109375,
      "learning_rate": 2.5099766590607587e-05,
      "loss": 1.119,
      "step": 393
    },
    {
      "epoch": 1.0707174733951252,
      "grad_norm": 0.048828125,
      "learning_rate": 2.5075647890912628e-05,
      "loss": 0.4643,
      "step": 394
    },
    {
      "epoch": 1.0734637830415379,
      "grad_norm": 0.04052734375,
      "learning_rate": 2.505148163186107e-05,
      "loss": 0.5572,
      "step": 395
    },
    {
      "epoch": 1.0762100926879505,
      "grad_norm": 0.041748046875,
      "learning_rate": 2.5027267927523178e-05,
      "loss": 0.4685,
      "step": 396
    },
    {
      "epoch": 1.0789564023343632,
      "grad_norm": 0.040771484375,
      "learning_rate": 2.500300689219315e-05,
      "loss": 0.5597,
      "step": 397
    },
    {
      "epoch": 1.0817027119807758,
      "grad_norm": 0.04052734375,
      "learning_rate": 2.4978698640388617e-05,
      "loss": 0.47,
      "step": 398
    },
    {
      "epoch": 1.0844490216271885,
      "grad_norm": 0.04833984375,
      "learning_rate": 2.495434328685007e-05,
      "loss": 0.5364,
      "step": 399
    },
    {
      "epoch": 1.0871953312736011,
      "grad_norm": 0.041748046875,
      "learning_rate": 2.492994094654033e-05,
      "loss": 0.4303,
      "step": 400
    },
    {
      "epoch": 1.0899416409200138,
      "grad_norm": 0.1435546875,
      "learning_rate": 2.490549173464402e-05,
      "loss": 1.1982,
      "step": 401
    },
    {
      "epoch": 1.0926879505664264,
      "grad_norm": 0.0390625,
      "learning_rate": 2.4880995766566986e-05,
      "loss": 0.5137,
      "step": 402
    },
    {
      "epoch": 1.095434260212839,
      "grad_norm": 0.04248046875,
      "learning_rate": 2.4856453157935795e-05,
      "loss": 0.4997,
      "step": 403
    },
    {
      "epoch": 1.0981805698592517,
      "grad_norm": 0.040771484375,
      "learning_rate": 2.483186402459715e-05,
      "loss": 0.5209,
      "step": 404
    },
    {
      "epoch": 1.1009268795056641,
      "grad_norm": 0.04541015625,
      "learning_rate": 2.4807228482617376e-05,
      "loss": 0.483,
      "step": 405
    },
    {
      "epoch": 1.1036731891520768,
      "grad_norm": 0.04345703125,
      "learning_rate": 2.4782546648281848e-05,
      "loss": 0.5055,
      "step": 406
    },
    {
      "epoch": 1.1064194987984894,
      "grad_norm": 0.039306640625,
      "learning_rate": 2.4757818638094457e-05,
      "loss": 0.462,
      "step": 407
    },
    {
      "epoch": 1.109165808444902,
      "grad_norm": 0.04150390625,
      "learning_rate": 2.473304456877705e-05,
      "loss": 0.4663,
      "step": 408
    },
    {
      "epoch": 1.1119121180913147,
      "grad_norm": 0.04541015625,
      "learning_rate": 2.470822455726889e-05,
      "loss": 0.5343,
      "step": 409
    },
    {
      "epoch": 1.1146584277377274,
      "grad_norm": 0.039306640625,
      "learning_rate": 2.468335872072609e-05,
      "loss": 0.4854,
      "step": 410
    },
    {
      "epoch": 1.11740473738414,
      "grad_norm": 0.046875,
      "learning_rate": 2.4658447176521076e-05,
      "loss": 0.5206,
      "step": 411
    },
    {
      "epoch": 1.1201510470305527,
      "grad_norm": 0.04150390625,
      "learning_rate": 2.463349004224201e-05,
      "loss": 0.4738,
      "step": 412
    },
    {
      "epoch": 1.1228973566769653,
      "grad_norm": 0.04248046875,
      "learning_rate": 2.460848743569227e-05,
      "loss": 0.5632,
      "step": 413
    },
    {
      "epoch": 1.125643666323378,
      "grad_norm": 0.043212890625,
      "learning_rate": 2.458343947488985e-05,
      "loss": 0.6056,
      "step": 414
    },
    {
      "epoch": 1.1283899759697906,
      "grad_norm": 0.146484375,
      "learning_rate": 2.4558346278066853e-05,
      "loss": 1.1007,
      "step": 415
    },
    {
      "epoch": 1.1311362856162033,
      "grad_norm": 0.041259765625,
      "learning_rate": 2.4533207963668883e-05,
      "loss": 0.4747,
      "step": 416
    },
    {
      "epoch": 1.133882595262616,
      "grad_norm": 0.0419921875,
      "learning_rate": 2.4508024650354525e-05,
      "loss": 0.439,
      "step": 417
    },
    {
      "epoch": 1.1366289049090286,
      "grad_norm": 0.041748046875,
      "learning_rate": 2.4482796456994757e-05,
      "loss": 0.4913,
      "step": 418
    },
    {
      "epoch": 1.1393752145554412,
      "grad_norm": 0.049072265625,
      "learning_rate": 2.4457523502672415e-05,
      "loss": 0.5722,
      "step": 419
    },
    {
      "epoch": 1.1421215242018539,
      "grad_norm": 0.040283203125,
      "learning_rate": 2.44322059066816e-05,
      "loss": 0.3971,
      "step": 420
    },
    {
      "epoch": 1.1448678338482665,
      "grad_norm": 0.0419921875,
      "learning_rate": 2.440684378852714e-05,
      "loss": 0.4724,
      "step": 421
    },
    {
      "epoch": 1.147614143494679,
      "grad_norm": 0.052734375,
      "learning_rate": 2.438143726792403e-05,
      "loss": 0.5305,
      "step": 422
    },
    {
      "epoch": 1.1503604531410916,
      "grad_norm": 0.056396484375,
      "learning_rate": 2.435598646479683e-05,
      "loss": 0.4924,
      "step": 423
    },
    {
      "epoch": 1.1531067627875042,
      "grad_norm": 0.0390625,
      "learning_rate": 2.4330491499279148e-05,
      "loss": 0.4927,
      "step": 424
    },
    {
      "epoch": 1.1558530724339169,
      "grad_norm": 0.0390625,
      "learning_rate": 2.4304952491713035e-05,
      "loss": 0.45,
      "step": 425
    },
    {
      "epoch": 1.1585993820803295,
      "grad_norm": 0.048095703125,
      "learning_rate": 2.4279369562648424e-05,
      "loss": 0.5892,
      "step": 426
    },
    {
      "epoch": 1.1613456917267422,
      "grad_norm": 0.045654296875,
      "learning_rate": 2.4253742832842583e-05,
      "loss": 0.4727,
      "step": 427
    },
    {
      "epoch": 1.1640920013731548,
      "grad_norm": 0.04443359375,
      "learning_rate": 2.4228072423259527e-05,
      "loss": 0.5063,
      "step": 428
    },
    {
      "epoch": 1.1668383110195675,
      "grad_norm": 0.06201171875,
      "learning_rate": 2.420235845506944e-05,
      "loss": 0.4872,
      "step": 429
    },
    {
      "epoch": 1.1695846206659801,
      "grad_norm": 0.0390625,
      "learning_rate": 2.4176601049648116e-05,
      "loss": 0.3843,
      "step": 430
    },
    {
      "epoch": 1.1723309303123928,
      "grad_norm": 0.05224609375,
      "learning_rate": 2.415080032857639e-05,
      "loss": 0.4478,
      "step": 431
    },
    {
      "epoch": 1.1750772399588054,
      "grad_norm": 0.0419921875,
      "learning_rate": 2.4124956413639548e-05,
      "loss": 0.4964,
      "step": 432
    },
    {
      "epoch": 1.177823549605218,
      "grad_norm": 0.04248046875,
      "learning_rate": 2.4099069426826766e-05,
      "loss": 0.5176,
      "step": 433
    },
    {
      "epoch": 1.1805698592516307,
      "grad_norm": 0.04296875,
      "learning_rate": 2.4073139490330526e-05,
      "loss": 0.5596,
      "step": 434
    },
    {
      "epoch": 1.1833161688980431,
      "grad_norm": 0.04638671875,
      "learning_rate": 2.4047166726546047e-05,
      "loss": 0.485,
      "step": 435
    },
    {
      "epoch": 1.1860624785444558,
      "grad_norm": 0.04248046875,
      "learning_rate": 2.4021151258070694e-05,
      "loss": 0.4768,
      "step": 436
    },
    {
      "epoch": 1.1888087881908684,
      "grad_norm": 0.05810546875,
      "learning_rate": 2.3995093207703413e-05,
      "loss": 0.5097,
      "step": 437
    },
    {
      "epoch": 1.191555097837281,
      "grad_norm": 0.048095703125,
      "learning_rate": 2.3968992698444153e-05,
      "loss": 0.5401,
      "step": 438
    },
    {
      "epoch": 1.1943014074836937,
      "grad_norm": 0.04248046875,
      "learning_rate": 2.394284985349327e-05,
      "loss": 0.425,
      "step": 439
    },
    {
      "epoch": 1.1970477171301064,
      "grad_norm": 0.044677734375,
      "learning_rate": 2.3916664796250946e-05,
      "loss": 0.3752,
      "step": 440
    },
    {
      "epoch": 1.199794026776519,
      "grad_norm": 0.042236328125,
      "learning_rate": 2.389043765031664e-05,
      "loss": 0.4724,
      "step": 441
    },
    {
      "epoch": 1.2025403364229317,
      "grad_norm": 0.046142578125,
      "learning_rate": 2.386416853948845e-05,
      "loss": 0.5598,
      "step": 442
    },
    {
      "epoch": 1.2052866460693443,
      "grad_norm": 0.0458984375,
      "learning_rate": 2.3837857587762583e-05,
      "loss": 0.3885,
      "step": 443
    },
    {
      "epoch": 1.208032955715757,
      "grad_norm": 0.04931640625,
      "learning_rate": 2.3811504919332727e-05,
      "loss": 0.4608,
      "step": 444
    },
    {
      "epoch": 1.2107792653621696,
      "grad_norm": 0.0390625,
      "learning_rate": 2.378511065858949e-05,
      "loss": 0.4457,
      "step": 445
    },
    {
      "epoch": 1.2135255750085823,
      "grad_norm": 0.039306640625,
      "learning_rate": 2.3758674930119807e-05,
      "loss": 0.4162,
      "step": 446
    },
    {
      "epoch": 1.216271884654995,
      "grad_norm": 0.054443359375,
      "learning_rate": 2.3732197858706343e-05,
      "loss": 0.4656,
      "step": 447
    },
    {
      "epoch": 1.2190181943014076,
      "grad_norm": 0.0478515625,
      "learning_rate": 2.370567956932692e-05,
      "loss": 0.4525,
      "step": 448
    },
    {
      "epoch": 1.2217645039478202,
      "grad_norm": 0.044921875,
      "learning_rate": 2.367912018715391e-05,
      "loss": 0.498,
      "step": 449
    },
    {
      "epoch": 1.2245108135942329,
      "grad_norm": 0.047119140625,
      "learning_rate": 2.3652519837553655e-05,
      "loss": 0.3724,
      "step": 450
    },
    {
      "epoch": 1.2272571232406453,
      "grad_norm": 0.051025390625,
      "learning_rate": 2.3625878646085873e-05,
      "loss": 0.3611,
      "step": 451
    },
    {
      "epoch": 1.230003432887058,
      "grad_norm": 0.044189453125,
      "learning_rate": 2.3599196738503068e-05,
      "loss": 0.4002,
      "step": 452
    },
    {
      "epoch": 1.2327497425334706,
      "grad_norm": 0.045654296875,
      "learning_rate": 2.3572474240749932e-05,
      "loss": 0.5691,
      "step": 453
    },
    {
      "epoch": 1.2354960521798832,
      "grad_norm": 0.042236328125,
      "learning_rate": 2.354571127896275e-05,
      "loss": 0.536,
      "step": 454
    },
    {
      "epoch": 1.2382423618262959,
      "grad_norm": 0.045654296875,
      "learning_rate": 2.3518907979468807e-05,
      "loss": 0.4385,
      "step": 455
    },
    {
      "epoch": 1.2382423618262959,
      "eval_loss": 0.5073373913764954,
      "eval_runtime": 627.5271,
      "eval_samples_per_second": 14.607,
      "eval_steps_per_second": 14.607,
      "step": 455
    },
    {
      "epoch": 1.2409886714727085,
      "grad_norm": 0.04296875,
      "learning_rate": 2.349206446878578e-05,
      "loss": 0.5131,
      "step": 456
    },
    {
      "epoch": 1.2437349811191212,
      "grad_norm": 0.042236328125,
      "learning_rate": 2.346518087362118e-05,
      "loss": 0.4821,
      "step": 457
    },
    {
      "epoch": 1.2464812907655338,
      "grad_norm": 0.04541015625,
      "learning_rate": 2.3438257320871704e-05,
      "loss": 0.5344,
      "step": 458
    },
    {
      "epoch": 1.2492276004119465,
      "grad_norm": 0.0419921875,
      "learning_rate": 2.3411293937622658e-05,
      "loss": 0.4752,
      "step": 459
    },
    {
      "epoch": 1.2519739100583591,
      "grad_norm": 0.0400390625,
      "learning_rate": 2.338429085114737e-05,
      "loss": 0.4887,
      "step": 460
    },
    {
      "epoch": 1.2547202197047718,
      "grad_norm": 0.04248046875,
      "learning_rate": 2.335724818890656e-05,
      "loss": 0.4445,
      "step": 461
    },
    {
      "epoch": 1.2574665293511844,
      "grad_norm": 0.047119140625,
      "learning_rate": 2.3330166078547763e-05,
      "loss": 0.5841,
      "step": 462
    },
    {
      "epoch": 1.2602128389975968,
      "grad_norm": 0.045654296875,
      "learning_rate": 2.3303044647904725e-05,
      "loss": 0.519,
      "step": 463
    },
    {
      "epoch": 1.2629591486440095,
      "grad_norm": 0.044189453125,
      "learning_rate": 2.3275884024996784e-05,
      "loss": 0.5149,
      "step": 464
    },
    {
      "epoch": 1.2657054582904221,
      "grad_norm": 0.04248046875,
      "learning_rate": 2.324868433802827e-05,
      "loss": 0.4681,
      "step": 465
    },
    {
      "epoch": 1.2684517679368348,
      "grad_norm": 0.044921875,
      "learning_rate": 2.3221445715387917e-05,
      "loss": 0.5058,
      "step": 466
    },
    {
      "epoch": 1.2711980775832474,
      "grad_norm": 0.05126953125,
      "learning_rate": 2.319416828564824e-05,
      "loss": 0.5142,
      "step": 467
    },
    {
      "epoch": 1.27394438722966,
      "grad_norm": 0.042236328125,
      "learning_rate": 2.3166852177564925e-05,
      "loss": 0.4682,
      "step": 468
    },
    {
      "epoch": 1.2766906968760727,
      "grad_norm": 0.044677734375,
      "learning_rate": 2.3139497520076233e-05,
      "loss": 0.4361,
      "step": 469
    },
    {
      "epoch": 1.2794370065224854,
      "grad_norm": 0.043212890625,
      "learning_rate": 2.3112104442302393e-05,
      "loss": 0.5738,
      "step": 470
    },
    {
      "epoch": 1.282183316168898,
      "grad_norm": 0.0478515625,
      "learning_rate": 2.3084673073544976e-05,
      "loss": 0.4828,
      "step": 471
    },
    {
      "epoch": 1.2849296258153107,
      "grad_norm": 0.04248046875,
      "learning_rate": 2.3057203543286297e-05,
      "loss": 0.503,
      "step": 472
    },
    {
      "epoch": 1.2876759354617233,
      "grad_norm": 0.04541015625,
      "learning_rate": 2.3029695981188818e-05,
      "loss": 0.5526,
      "step": 473
    },
    {
      "epoch": 1.290422245108136,
      "grad_norm": 0.042236328125,
      "learning_rate": 2.3002150517094496e-05,
      "loss": 0.4757,
      "step": 474
    },
    {
      "epoch": 1.2931685547545486,
      "grad_norm": 0.044189453125,
      "learning_rate": 2.297456728102421e-05,
      "loss": 0.5773,
      "step": 475
    },
    {
      "epoch": 1.2959148644009613,
      "grad_norm": 0.041748046875,
      "learning_rate": 2.294694640317713e-05,
      "loss": 0.5248,
      "step": 476
    },
    {
      "epoch": 1.298661174047374,
      "grad_norm": 0.045166015625,
      "learning_rate": 2.2919288013930094e-05,
      "loss": 0.4915,
      "step": 477
    },
    {
      "epoch": 1.3014074836937866,
      "grad_norm": 0.0400390625,
      "learning_rate": 2.2891592243837015e-05,
      "loss": 0.5389,
      "step": 478
    },
    {
      "epoch": 1.3041537933401992,
      "grad_norm": 0.04248046875,
      "learning_rate": 2.286385922362824e-05,
      "loss": 0.4232,
      "step": 479
    },
    {
      "epoch": 1.3069001029866119,
      "grad_norm": 0.0439453125,
      "learning_rate": 2.2836089084209955e-05,
      "loss": 0.5072,
      "step": 480
    },
    {
      "epoch": 1.3096464126330245,
      "grad_norm": 0.044677734375,
      "learning_rate": 2.280828195666355e-05,
      "loss": 0.54,
      "step": 481
    },
    {
      "epoch": 1.312392722279437,
      "grad_norm": 0.048828125,
      "learning_rate": 2.2780437972245014e-05,
      "loss": 0.5446,
      "step": 482
    },
    {
      "epoch": 1.3151390319258496,
      "grad_norm": 0.046142578125,
      "learning_rate": 2.2752557262384307e-05,
      "loss": 0.4725,
      "step": 483
    },
    {
      "epoch": 1.3178853415722622,
      "grad_norm": 0.1650390625,
      "learning_rate": 2.2724639958684733e-05,
      "loss": 1.2587,
      "step": 484
    },
    {
      "epoch": 1.3206316512186749,
      "grad_norm": 0.043212890625,
      "learning_rate": 2.2696686192922342e-05,
      "loss": 0.4965,
      "step": 485
    },
    {
      "epoch": 1.3233779608650875,
      "grad_norm": 0.043701171875,
      "learning_rate": 2.2668696097045284e-05,
      "loss": 0.5382,
      "step": 486
    },
    {
      "epoch": 1.3261242705115002,
      "grad_norm": 0.048583984375,
      "learning_rate": 2.2640669803173195e-05,
      "loss": 0.4305,
      "step": 487
    },
    {
      "epoch": 1.3288705801579128,
      "grad_norm": 0.04150390625,
      "learning_rate": 2.2612607443596572e-05,
      "loss": 0.4622,
      "step": 488
    },
    {
      "epoch": 1.3316168898043255,
      "grad_norm": 0.05908203125,
      "learning_rate": 2.258450915077616e-05,
      "loss": 0.4975,
      "step": 489
    },
    {
      "epoch": 1.3343631994507381,
      "grad_norm": 0.04541015625,
      "learning_rate": 2.2556375057342306e-05,
      "loss": 0.6356,
      "step": 490
    },
    {
      "epoch": 1.3371095090971508,
      "grad_norm": 0.042724609375,
      "learning_rate": 2.2528205296094356e-05,
      "loss": 0.4422,
      "step": 491
    },
    {
      "epoch": 1.3398558187435634,
      "grad_norm": 0.04296875,
      "learning_rate": 2.25e-05,
      "loss": 0.446,
      "step": 492
    },
    {
      "epoch": 1.3426021283899758,
      "grad_norm": 0.051025390625,
      "learning_rate": 2.247175930219468e-05,
      "loss": 0.5996,
      "step": 493
    },
    {
      "epoch": 1.3453484380363885,
      "grad_norm": 0.04833984375,
      "learning_rate": 2.2443483335980924e-05,
      "loss": 0.5905,
      "step": 494
    },
    {
      "epoch": 1.3480947476828011,
      "grad_norm": 0.047607421875,
      "learning_rate": 2.2415172234827754e-05,
      "loss": 0.5824,
      "step": 495
    },
    {
      "epoch": 1.3508410573292138,
      "grad_norm": 0.041748046875,
      "learning_rate": 2.238682613237001e-05,
      "loss": 0.4885,
      "step": 496
    },
    {
      "epoch": 1.3535873669756264,
      "grad_norm": 0.048095703125,
      "learning_rate": 2.2358445162407775e-05,
      "loss": 0.587,
      "step": 497
    },
    {
      "epoch": 1.356333676622039,
      "grad_norm": 0.042724609375,
      "learning_rate": 2.2330029458905697e-05,
      "loss": 0.5453,
      "step": 498
    },
    {
      "epoch": 1.3590799862684517,
      "grad_norm": 0.04296875,
      "learning_rate": 2.230157915599238e-05,
      "loss": 0.4596,
      "step": 499
    },
    {
      "epoch": 1.3618262959148644,
      "grad_norm": 0.04736328125,
      "learning_rate": 2.2273094387959747e-05,
      "loss": 0.4349,
      "step": 500
    },
    {
      "epoch": 1.364572605561277,
      "grad_norm": 0.0458984375,
      "learning_rate": 2.2244575289262394e-05,
      "loss": 0.4613,
      "step": 501
    },
    {
      "epoch": 1.3673189152076897,
      "grad_norm": 0.0419921875,
      "learning_rate": 2.221602199451698e-05,
      "loss": 0.4176,
      "step": 502
    },
    {
      "epoch": 1.3700652248541023,
      "grad_norm": 0.049560546875,
      "learning_rate": 2.2187434638501564e-05,
      "loss": 0.4799,
      "step": 503
    },
    {
      "epoch": 1.372811534500515,
      "grad_norm": 0.03955078125,
      "learning_rate": 2.215881335615499e-05,
      "loss": 0.4335,
      "step": 504
    },
    {
      "epoch": 1.3755578441469276,
      "grad_norm": 0.0478515625,
      "learning_rate": 2.2130158282576245e-05,
      "loss": 0.5999,
      "step": 505
    },
    {
      "epoch": 1.3783041537933403,
      "grad_norm": 0.0478515625,
      "learning_rate": 2.2101469553023807e-05,
      "loss": 0.4654,
      "step": 506
    },
    {
      "epoch": 1.381050463439753,
      "grad_norm": 0.04150390625,
      "learning_rate": 2.2072747302915026e-05,
      "loss": 0.4423,
      "step": 507
    },
    {
      "epoch": 1.3837967730861656,
      "grad_norm": 0.04541015625,
      "learning_rate": 2.2043991667825478e-05,
      "loss": 0.5145,
      "step": 508
    },
    {
      "epoch": 1.3865430827325782,
      "grad_norm": 0.0400390625,
      "learning_rate": 2.2015202783488316e-05,
      "loss": 0.5894,
      "step": 509
    },
    {
      "epoch": 1.3892893923789909,
      "grad_norm": 0.042236328125,
      "learning_rate": 2.1986380785793646e-05,
      "loss": 0.5228,
      "step": 510
    },
    {
      "epoch": 1.3920357020254035,
      "grad_norm": 0.048828125,
      "learning_rate": 2.195752581078787e-05,
      "loss": 0.5529,
      "step": 511
    },
    {
      "epoch": 1.394782011671816,
      "grad_norm": 0.043701171875,
      "learning_rate": 2.1928637994673053e-05,
      "loss": 0.5783,
      "step": 512
    },
    {
      "epoch": 1.3975283213182286,
      "grad_norm": 0.04248046875,
      "learning_rate": 2.1899717473806273e-05,
      "loss": 0.418,
      "step": 513
    },
    {
      "epoch": 1.4002746309646412,
      "grad_norm": 0.04541015625,
      "learning_rate": 2.1870764384698992e-05,
      "loss": 0.4945,
      "step": 514
    },
    {
      "epoch": 1.4030209406110539,
      "grad_norm": 0.050048828125,
      "learning_rate": 2.1841778864016396e-05,
      "loss": 0.496,
      "step": 515
    },
    {
      "epoch": 1.4057672502574665,
      "grad_norm": 0.042724609375,
      "learning_rate": 2.1812761048576752e-05,
      "loss": 0.5087,
      "step": 516
    },
    {
      "epoch": 1.4085135599038792,
      "grad_norm": 0.04833984375,
      "learning_rate": 2.1783711075350766e-05,
      "loss": 0.4898,
      "step": 517
    },
    {
      "epoch": 1.4112598695502918,
      "grad_norm": 0.0419921875,
      "learning_rate": 2.1754629081460947e-05,
      "loss": 0.4379,
      "step": 518
    },
    {
      "epoch": 1.4140061791967045,
      "grad_norm": 0.04541015625,
      "learning_rate": 2.172551520418093e-05,
      "loss": 0.4827,
      "step": 519
    },
    {
      "epoch": 1.416752488843117,
      "grad_norm": 0.044677734375,
      "learning_rate": 2.169636958093487e-05,
      "loss": 0.5007,
      "step": 520
    },
    {
      "epoch": 1.4194987984895298,
      "grad_norm": 0.043212890625,
      "learning_rate": 2.1667192349296746e-05,
      "loss": 0.4651,
      "step": 521
    },
    {
      "epoch": 1.4222451081359424,
      "grad_norm": 0.041748046875,
      "learning_rate": 2.1637983646989758e-05,
      "loss": 0.4674,
      "step": 522
    },
    {
      "epoch": 1.4249914177823548,
      "grad_norm": 0.045166015625,
      "learning_rate": 2.1608743611885633e-05,
      "loss": 0.4794,
      "step": 523
    },
    {
      "epoch": 1.4277377274287675,
      "grad_norm": 0.045166015625,
      "learning_rate": 2.1579472382004015e-05,
      "loss": 0.5292,
      "step": 524
    },
    {
      "epoch": 1.4304840370751801,
      "grad_norm": 0.04443359375,
      "learning_rate": 2.1550170095511784e-05,
      "loss": 0.4964,
      "step": 525
    },
    {
      "epoch": 1.4332303467215928,
      "grad_norm": 0.0537109375,
      "learning_rate": 2.1520836890722416e-05,
      "loss": 0.4236,
      "step": 526
    },
    {
      "epoch": 1.4359766563680054,
      "grad_norm": 0.044921875,
      "learning_rate": 2.149147290609533e-05,
      "loss": 0.4859,
      "step": 527
    },
    {
      "epoch": 1.438722966014418,
      "grad_norm": 0.04638671875,
      "learning_rate": 2.146207828023524e-05,
      "loss": 0.4659,
      "step": 528
    },
    {
      "epoch": 1.4414692756608307,
      "grad_norm": 0.048828125,
      "learning_rate": 2.1432653151891473e-05,
      "loss": 0.4424,
      "step": 529
    },
    {
      "epoch": 1.4442155853072434,
      "grad_norm": 0.04345703125,
      "learning_rate": 2.1403197659957356e-05,
      "loss": 0.4515,
      "step": 530
    },
    {
      "epoch": 1.446961894953656,
      "grad_norm": 0.041015625,
      "learning_rate": 2.137371194346953e-05,
      "loss": 0.4618,
      "step": 531
    },
    {
      "epoch": 1.4497082046000687,
      "grad_norm": 0.042236328125,
      "learning_rate": 2.1344196141607297e-05,
      "loss": 0.3928,
      "step": 532
    },
    {
      "epoch": 1.4524545142464813,
      "grad_norm": 0.044189453125,
      "learning_rate": 2.1314650393691984e-05,
      "loss": 0.4598,
      "step": 533
    },
    {
      "epoch": 1.455200823892894,
      "grad_norm": 0.046630859375,
      "learning_rate": 2.1285074839186257e-05,
      "loss": 0.5646,
      "step": 534
    },
    {
      "epoch": 1.4579471335393066,
      "grad_norm": 0.0517578125,
      "learning_rate": 2.1255469617693476e-05,
      "loss": 0.5984,
      "step": 535
    },
    {
      "epoch": 1.4606934431857193,
      "grad_norm": 0.04296875,
      "learning_rate": 2.122583486895705e-05,
      "loss": 0.5419,
      "step": 536
    },
    {
      "epoch": 1.463439752832132,
      "grad_norm": 0.043212890625,
      "learning_rate": 2.119617073285974e-05,
      "loss": 0.5481,
      "step": 537
    },
    {
      "epoch": 1.4661860624785445,
      "grad_norm": 0.042236328125,
      "learning_rate": 2.116647734942305e-05,
      "loss": 0.5588,
      "step": 538
    },
    {
      "epoch": 1.4689323721249572,
      "grad_norm": 0.04736328125,
      "learning_rate": 2.113675485880652e-05,
      "loss": 0.5621,
      "step": 539
    },
    {
      "epoch": 1.4716786817713698,
      "grad_norm": 0.044189453125,
      "learning_rate": 2.110700340130708e-05,
      "loss": 0.5056,
      "step": 540
    },
    {
      "epoch": 1.4744249914177825,
      "grad_norm": 0.044921875,
      "learning_rate": 2.1077223117358395e-05,
      "loss": 0.5526,
      "step": 541
    },
    {
      "epoch": 1.477171301064195,
      "grad_norm": 0.048828125,
      "learning_rate": 2.104741414753021e-05,
      "loss": 0.5414,
      "step": 542
    },
    {
      "epoch": 1.4799176107106076,
      "grad_norm": 0.056640625,
      "learning_rate": 2.1017576632527662e-05,
      "loss": 0.5472,
      "step": 543
    },
    {
      "epoch": 1.4826639203570202,
      "grad_norm": 0.0517578125,
      "learning_rate": 2.098771071319062e-05,
      "loss": 0.4568,
      "step": 544
    },
    {
      "epoch": 1.4854102300034329,
      "grad_norm": 0.046142578125,
      "learning_rate": 2.0957816530493037e-05,
      "loss": 0.4277,
      "step": 545
    },
    {
      "epoch": 1.4881565396498455,
      "grad_norm": 0.048583984375,
      "learning_rate": 2.0927894225542282e-05,
      "loss": 0.4949,
      "step": 546
    },
    {
      "epoch": 1.4881565396498455,
      "eval_loss": 0.5060501098632812,
      "eval_runtime": 630.5882,
      "eval_samples_per_second": 14.536,
      "eval_steps_per_second": 14.536,
      "step": 546
    },
    {
      "epoch": 1.4909028492962582,
      "grad_norm": 0.04052734375,
      "learning_rate": 2.089794393957846e-05,
      "loss": 0.3558,
      "step": 547
    },
    {
      "epoch": 1.4936491589426708,
      "grad_norm": 0.043212890625,
      "learning_rate": 2.086796581397374e-05,
      "loss": 0.4622,
      "step": 548
    },
    {
      "epoch": 1.4963954685890835,
      "grad_norm": 0.04833984375,
      "learning_rate": 2.083795999023173e-05,
      "loss": 0.5402,
      "step": 549
    },
    {
      "epoch": 1.499141778235496,
      "grad_norm": 0.048583984375,
      "learning_rate": 2.080792660998676e-05,
      "loss": 0.5271,
      "step": 550
    },
    {
      "epoch": 1.5018880878819085,
      "grad_norm": 0.0439453125,
      "learning_rate": 2.0777865815003234e-05,
      "loss": 0.5152,
      "step": 551
    },
    {
      "epoch": 1.5046343975283212,
      "grad_norm": 0.04541015625,
      "learning_rate": 2.074777774717496e-05,
      "loss": 0.5099,
      "step": 552
    },
    {
      "epoch": 1.5073807071747338,
      "grad_norm": 0.041015625,
      "learning_rate": 2.0717662548524482e-05,
      "loss": 0.4075,
      "step": 553
    },
    {
      "epoch": 1.5101270168211465,
      "grad_norm": 0.048828125,
      "learning_rate": 2.068752036120241e-05,
      "loss": 0.5205,
      "step": 554
    },
    {
      "epoch": 1.5128733264675591,
      "grad_norm": 0.045166015625,
      "learning_rate": 2.0657351327486745e-05,
      "loss": 0.5127,
      "step": 555
    },
    {
      "epoch": 1.5156196361139718,
      "grad_norm": 0.044921875,
      "learning_rate": 2.0627155589782212e-05,
      "loss": 0.5399,
      "step": 556
    },
    {
      "epoch": 1.5183659457603844,
      "grad_norm": 0.04150390625,
      "learning_rate": 2.0596933290619572e-05,
      "loss": 0.4869,
      "step": 557
    },
    {
      "epoch": 1.521112255406797,
      "grad_norm": 0.0498046875,
      "learning_rate": 2.0566684572654978e-05,
      "loss": 0.6318,
      "step": 558
    },
    {
      "epoch": 1.5238585650532097,
      "grad_norm": 0.0576171875,
      "learning_rate": 2.0536409578669277e-05,
      "loss": 0.4729,
      "step": 559
    },
    {
      "epoch": 1.5266048746996224,
      "grad_norm": 0.046630859375,
      "learning_rate": 2.0506108451567347e-05,
      "loss": 0.5059,
      "step": 560
    },
    {
      "epoch": 1.529351184346035,
      "grad_norm": 0.043701171875,
      "learning_rate": 2.0475781334377426e-05,
      "loss": 0.3829,
      "step": 561
    },
    {
      "epoch": 1.5320974939924477,
      "grad_norm": 0.0478515625,
      "learning_rate": 2.044542837025042e-05,
      "loss": 0.4582,
      "step": 562
    },
    {
      "epoch": 1.5348438036388603,
      "grad_norm": 0.046630859375,
      "learning_rate": 2.0415049702459244e-05,
      "loss": 0.5344,
      "step": 563
    },
    {
      "epoch": 1.537590113285273,
      "grad_norm": 0.042236328125,
      "learning_rate": 2.0384645474398137e-05,
      "loss": 0.4508,
      "step": 564
    },
    {
      "epoch": 1.5403364229316856,
      "grad_norm": 0.040283203125,
      "learning_rate": 2.0354215829582005e-05,
      "loss": 0.4973,
      "step": 565
    },
    {
      "epoch": 1.5430827325780982,
      "grad_norm": 0.046630859375,
      "learning_rate": 2.03237609116457e-05,
      "loss": 0.5406,
      "step": 566
    },
    {
      "epoch": 1.545829042224511,
      "grad_norm": 0.041259765625,
      "learning_rate": 2.029328086434339e-05,
      "loss": 0.4956,
      "step": 567
    },
    {
      "epoch": 1.5485753518709235,
      "grad_norm": 0.06396484375,
      "learning_rate": 2.0262775831547847e-05,
      "loss": 0.5642,
      "step": 568
    },
    {
      "epoch": 1.5513216615173362,
      "grad_norm": 0.043701171875,
      "learning_rate": 2.0232245957249788e-05,
      "loss": 0.5424,
      "step": 569
    },
    {
      "epoch": 1.5540679711637488,
      "grad_norm": 0.044677734375,
      "learning_rate": 2.020169138555718e-05,
      "loss": 0.4972,
      "step": 570
    },
    {
      "epoch": 1.5568142808101615,
      "grad_norm": 0.05126953125,
      "learning_rate": 2.0171112260694576e-05,
      "loss": 0.4511,
      "step": 571
    },
    {
      "epoch": 1.5595605904565741,
      "grad_norm": 0.045654296875,
      "learning_rate": 2.0140508727002422e-05,
      "loss": 0.4669,
      "step": 572
    },
    {
      "epoch": 1.5623069001029866,
      "grad_norm": 0.046875,
      "learning_rate": 2.0109880928936375e-05,
      "loss": 0.5472,
      "step": 573
    },
    {
      "epoch": 1.5650532097493992,
      "grad_norm": 0.042236328125,
      "learning_rate": 2.007922901106663e-05,
      "loss": 0.5493,
      "step": 574
    },
    {
      "epoch": 1.5677995193958119,
      "grad_norm": 0.0439453125,
      "learning_rate": 2.0048553118077238e-05,
      "loss": 0.46,
      "step": 575
    },
    {
      "epoch": 1.5705458290422245,
      "grad_norm": 0.04931640625,
      "learning_rate": 2.0017853394765402e-05,
      "loss": 0.6062,
      "step": 576
    },
    {
      "epoch": 1.5732921386886372,
      "grad_norm": 0.044189453125,
      "learning_rate": 1.9987129986040825e-05,
      "loss": 0.5053,
      "step": 577
    },
    {
      "epoch": 1.5760384483350498,
      "grad_norm": 0.05029296875,
      "learning_rate": 1.9956383036925006e-05,
      "loss": 0.5205,
      "step": 578
    },
    {
      "epoch": 1.5787847579814624,
      "grad_norm": 0.045654296875,
      "learning_rate": 1.9925612692550554e-05,
      "loss": 0.5296,
      "step": 579
    },
    {
      "epoch": 1.581531067627875,
      "grad_norm": 0.042236328125,
      "learning_rate": 1.989481909816052e-05,
      "loss": 0.577,
      "step": 580
    },
    {
      "epoch": 1.5842773772742875,
      "grad_norm": 0.04833984375,
      "learning_rate": 1.986400239910769e-05,
      "loss": 0.5867,
      "step": 581
    },
    {
      "epoch": 1.5870236869207002,
      "grad_norm": 0.044189453125,
      "learning_rate": 1.9833162740853916e-05,
      "loss": 0.5371,
      "step": 582
    },
    {
      "epoch": 1.5897699965671128,
      "grad_norm": 0.044921875,
      "learning_rate": 1.980230026896942e-05,
      "loss": 0.4848,
      "step": 583
    },
    {
      "epoch": 1.5925163062135255,
      "grad_norm": 0.040283203125,
      "learning_rate": 1.977141512913211e-05,
      "loss": 0.4747,
      "step": 584
    },
    {
      "epoch": 1.5952626158599381,
      "grad_norm": 0.041748046875,
      "learning_rate": 1.974050746712689e-05,
      "loss": 0.4296,
      "step": 585
    },
    {
      "epoch": 1.5980089255063508,
      "grad_norm": 0.04296875,
      "learning_rate": 1.9709577428844984e-05,
      "loss": 0.4943,
      "step": 586
    },
    {
      "epoch": 1.6007552351527634,
      "grad_norm": 0.04296875,
      "learning_rate": 1.967862516028321e-05,
      "loss": 0.487,
      "step": 587
    },
    {
      "epoch": 1.603501544799176,
      "grad_norm": 0.046875,
      "learning_rate": 1.9647650807543358e-05,
      "loss": 0.5275,
      "step": 588
    },
    {
      "epoch": 1.6062478544455887,
      "grad_norm": 0.046630859375,
      "learning_rate": 1.961665451683143e-05,
      "loss": 0.557,
      "step": 589
    },
    {
      "epoch": 1.6089941640920014,
      "grad_norm": 0.0419921875,
      "learning_rate": 1.9585636434456988e-05,
      "loss": 0.4689,
      "step": 590
    },
    {
      "epoch": 1.611740473738414,
      "grad_norm": 0.0439453125,
      "learning_rate": 1.9554596706832457e-05,
      "loss": 0.5351,
      "step": 591
    },
    {
      "epoch": 1.6144867833848267,
      "grad_norm": 0.04931640625,
      "learning_rate": 1.952353548047243e-05,
      "loss": 0.5714,
      "step": 592
    },
    {
      "epoch": 1.6172330930312393,
      "grad_norm": 0.045654296875,
      "learning_rate": 1.9492452901992987e-05,
      "loss": 0.5468,
      "step": 593
    },
    {
      "epoch": 1.619979402677652,
      "grad_norm": 0.04931640625,
      "learning_rate": 1.946134911811099e-05,
      "loss": 0.5812,
      "step": 594
    },
    {
      "epoch": 1.6227257123240646,
      "grad_norm": 0.044189453125,
      "learning_rate": 1.9430224275643388e-05,
      "loss": 0.5367,
      "step": 595
    },
    {
      "epoch": 1.6254720219704772,
      "grad_norm": 0.046630859375,
      "learning_rate": 1.9399078521506546e-05,
      "loss": 0.5746,
      "step": 596
    },
    {
      "epoch": 1.62821833161689,
      "grad_norm": 0.051513671875,
      "learning_rate": 1.9367912002715524e-05,
      "loss": 0.4458,
      "step": 597
    },
    {
      "epoch": 1.6309646412633025,
      "grad_norm": 0.043212890625,
      "learning_rate": 1.93367248663834e-05,
      "loss": 0.4413,
      "step": 598
    },
    {
      "epoch": 1.6337109509097152,
      "grad_norm": 0.0498046875,
      "learning_rate": 1.9305517259720573e-05,
      "loss": 0.5666,
      "step": 599
    },
    {
      "epoch": 1.6364572605561278,
      "grad_norm": 0.042236328125,
      "learning_rate": 1.9274289330034068e-05,
      "loss": 0.5282,
      "step": 600
    },
    {
      "epoch": 1.6392035702025405,
      "grad_norm": 0.050537109375,
      "learning_rate": 1.924304122472683e-05,
      "loss": 0.5065,
      "step": 601
    },
    {
      "epoch": 1.6419498798489531,
      "grad_norm": 0.042724609375,
      "learning_rate": 1.9211773091297057e-05,
      "loss": 0.5519,
      "step": 602
    },
    {
      "epoch": 1.6446961894953656,
      "grad_norm": 0.044189453125,
      "learning_rate": 1.9180485077337462e-05,
      "loss": 0.5044,
      "step": 603
    },
    {
      "epoch": 1.6474424991417782,
      "grad_norm": 0.040283203125,
      "learning_rate": 1.9149177330534614e-05,
      "loss": 0.4895,
      "step": 604
    },
    {
      "epoch": 1.6501888087881909,
      "grad_norm": 0.044677734375,
      "learning_rate": 1.9117849998668212e-05,
      "loss": 0.4553,
      "step": 605
    },
    {
      "epoch": 1.6529351184346035,
      "grad_norm": 0.04638671875,
      "learning_rate": 1.9086503229610418e-05,
      "loss": 0.5583,
      "step": 606
    },
    {
      "epoch": 1.6556814280810161,
      "grad_norm": 0.03955078125,
      "learning_rate": 1.905513717132513e-05,
      "loss": 0.3757,
      "step": 607
    },
    {
      "epoch": 1.6584277377274288,
      "grad_norm": 0.04541015625,
      "learning_rate": 1.90237519718673e-05,
      "loss": 0.5956,
      "step": 608
    },
    {
      "epoch": 1.6611740473738414,
      "grad_norm": 0.14453125,
      "learning_rate": 1.899234777938222e-05,
      "loss": 1.1236,
      "step": 609
    },
    {
      "epoch": 1.6639203570202539,
      "grad_norm": 0.044677734375,
      "learning_rate": 1.8960924742104856e-05,
      "loss": 0.5466,
      "step": 610
    },
    {
      "epoch": 1.6666666666666665,
      "grad_norm": 0.042724609375,
      "learning_rate": 1.892948300835911e-05,
      "loss": 0.4874,
      "step": 611
    },
    {
      "epoch": 1.6694129763130792,
      "grad_norm": 0.0478515625,
      "learning_rate": 1.889802272655713e-05,
      "loss": 0.5116,
      "step": 612
    },
    {
      "epoch": 1.6721592859594918,
      "grad_norm": 0.043701171875,
      "learning_rate": 1.8866544045198634e-05,
      "loss": 0.587,
      "step": 613
    },
    {
      "epoch": 1.6749055956059045,
      "grad_norm": 0.04296875,
      "learning_rate": 1.8835047112870163e-05,
      "loss": 0.4174,
      "step": 614
    },
    {
      "epoch": 1.677651905252317,
      "grad_norm": 0.04150390625,
      "learning_rate": 1.880353207824444e-05,
      "loss": 0.4023,
      "step": 615
    },
    {
      "epoch": 1.6803982148987298,
      "grad_norm": 0.042724609375,
      "learning_rate": 1.8771999090079613e-05,
      "loss": 0.5134,
      "step": 616
    },
    {
      "epoch": 1.6831445245451424,
      "grad_norm": 0.044677734375,
      "learning_rate": 1.8740448297218575e-05,
      "loss": 0.4694,
      "step": 617
    },
    {
      "epoch": 1.685890834191555,
      "grad_norm": 0.0458984375,
      "learning_rate": 1.8708879848588268e-05,
      "loss": 0.5185,
      "step": 618
    },
    {
      "epoch": 1.6886371438379677,
      "grad_norm": 0.0458984375,
      "learning_rate": 1.8677293893198976e-05,
      "loss": 0.5077,
      "step": 619
    },
    {
      "epoch": 1.6913834534843804,
      "grad_norm": 0.045166015625,
      "learning_rate": 1.864569058014361e-05,
      "loss": 0.4517,
      "step": 620
    },
    {
      "epoch": 1.694129763130793,
      "grad_norm": 0.040771484375,
      "learning_rate": 1.8614070058597014e-05,
      "loss": 0.4703,
      "step": 621
    },
    {
      "epoch": 1.6968760727772056,
      "grad_norm": 0.04541015625,
      "learning_rate": 1.8582432477815268e-05,
      "loss": 0.5061,
      "step": 622
    },
    {
      "epoch": 1.6996223824236183,
      "grad_norm": 0.04248046875,
      "learning_rate": 1.855077798713497e-05,
      "loss": 0.5413,
      "step": 623
    },
    {
      "epoch": 1.702368692070031,
      "grad_norm": 0.05615234375,
      "learning_rate": 1.8519106735972535e-05,
      "loss": 0.4586,
      "step": 624
    },
    {
      "epoch": 1.7051150017164436,
      "grad_norm": 0.042236328125,
      "learning_rate": 1.84874188738235e-05,
      "loss": 0.5022,
      "step": 625
    },
    {
      "epoch": 1.7078613113628562,
      "grad_norm": 0.05224609375,
      "learning_rate": 1.8455714550261793e-05,
      "loss": 0.4945,
      "step": 626
    },
    {
      "epoch": 1.7106076210092689,
      "grad_norm": 0.0478515625,
      "learning_rate": 1.8423993914939063e-05,
      "loss": 0.5806,
      "step": 627
    },
    {
      "epoch": 1.7133539306556815,
      "grad_norm": 0.04296875,
      "learning_rate": 1.8392257117583944e-05,
      "loss": 0.462,
      "step": 628
    },
    {
      "epoch": 1.7161002403020942,
      "grad_norm": 0.042236328125,
      "learning_rate": 1.836050430800135e-05,
      "loss": 0.4944,
      "step": 629
    },
    {
      "epoch": 1.7188465499485068,
      "grad_norm": 0.1708984375,
      "learning_rate": 1.83287356360718e-05,
      "loss": 1.1722,
      "step": 630
    },
    {
      "epoch": 1.7215928595949195,
      "grad_norm": 0.04248046875,
      "learning_rate": 1.8296951251750667e-05,
      "loss": 0.3718,
      "step": 631
    },
    {
      "epoch": 1.7243391692413321,
      "grad_norm": 0.04345703125,
      "learning_rate": 1.8265151305067486e-05,
      "loss": 0.484,
      "step": 632
    },
    {
      "epoch": 1.7270854788877446,
      "grad_norm": 0.04443359375,
      "learning_rate": 1.8233335946125275e-05,
      "loss": 0.4783,
      "step": 633
    },
    {
      "epoch": 1.7298317885341572,
      "grad_norm": 0.045654296875,
      "learning_rate": 1.8201505325099782e-05,
      "loss": 0.5684,
      "step": 634
    },
    {
      "epoch": 1.7325780981805698,
      "grad_norm": 0.043212890625,
      "learning_rate": 1.8169659592238797e-05,
      "loss": 0.4518,
      "step": 635
    },
    {
      "epoch": 1.7353244078269825,
      "grad_norm": 0.0458984375,
      "learning_rate": 1.813779889786144e-05,
      "loss": 0.4535,
      "step": 636
    },
    {
      "epoch": 1.7380707174733951,
      "grad_norm": 0.0458984375,
      "learning_rate": 1.8105923392357464e-05,
      "loss": 0.503,
      "step": 637
    },
    {
      "epoch": 1.7380707174733951,
      "eval_loss": 0.5051947832107544,
      "eval_runtime": 630.1537,
      "eval_samples_per_second": 14.546,
      "eval_steps_per_second": 14.546,
      "step": 637
    },
    {
      "epoch": 1.7408170271198078,
      "grad_norm": 0.045166015625,
      "learning_rate": 1.807403322618653e-05,
      "loss": 0.4961,
      "step": 638
    },
    {
      "epoch": 1.7435633367662204,
      "grad_norm": 0.05224609375,
      "learning_rate": 1.8042128549877483e-05,
      "loss": 0.519,
      "step": 639
    },
    {
      "epoch": 1.7463096464126329,
      "grad_norm": 0.04541015625,
      "learning_rate": 1.8010209514027687e-05,
      "loss": 0.4011,
      "step": 640
    },
    {
      "epoch": 1.7490559560590455,
      "grad_norm": 0.045166015625,
      "learning_rate": 1.7978276269302275e-05,
      "loss": 0.3935,
      "step": 641
    },
    {
      "epoch": 1.7518022657054582,
      "grad_norm": 0.047607421875,
      "learning_rate": 1.794632896643343e-05,
      "loss": 0.6534,
      "step": 642
    },
    {
      "epoch": 1.7545485753518708,
      "grad_norm": 0.04638671875,
      "learning_rate": 1.7914367756219725e-05,
      "loss": 0.5715,
      "step": 643
    },
    {
      "epoch": 1.7572948849982835,
      "grad_norm": 0.04931640625,
      "learning_rate": 1.7882392789525358e-05,
      "loss": 0.5439,
      "step": 644
    },
    {
      "epoch": 1.760041194644696,
      "grad_norm": 0.047607421875,
      "learning_rate": 1.7850404217279467e-05,
      "loss": 0.5277,
      "step": 645
    },
    {
      "epoch": 1.7627875042911088,
      "grad_norm": 0.04736328125,
      "learning_rate": 1.781840219047541e-05,
      "loss": 0.586,
      "step": 646
    },
    {
      "epoch": 1.7655338139375214,
      "grad_norm": 0.045166015625,
      "learning_rate": 1.7786386860170054e-05,
      "loss": 0.5291,
      "step": 647
    },
    {
      "epoch": 1.768280123583934,
      "grad_norm": 0.04296875,
      "learning_rate": 1.775435837748306e-05,
      "loss": 0.3863,
      "step": 648
    },
    {
      "epoch": 1.7710264332303467,
      "grad_norm": 0.044677734375,
      "learning_rate": 1.7722316893596176e-05,
      "loss": 0.5247,
      "step": 649
    },
    {
      "epoch": 1.7737727428767593,
      "grad_norm": 0.04345703125,
      "learning_rate": 1.7690262559752516e-05,
      "loss": 0.4046,
      "step": 650
    },
    {
      "epoch": 1.776519052523172,
      "grad_norm": 0.04345703125,
      "learning_rate": 1.7658195527255847e-05,
      "loss": 0.4744,
      "step": 651
    },
    {
      "epoch": 1.7792653621695846,
      "grad_norm": 0.0625,
      "learning_rate": 1.7626115947469877e-05,
      "loss": 0.424,
      "step": 652
    },
    {
      "epoch": 1.7820116718159973,
      "grad_norm": 0.045166015625,
      "learning_rate": 1.759402397181754e-05,
      "loss": 0.4644,
      "step": 653
    },
    {
      "epoch": 1.78475798146241,
      "grad_norm": 0.04833984375,
      "learning_rate": 1.7561919751780278e-05,
      "loss": 0.6509,
      "step": 654
    },
    {
      "epoch": 1.7875042911088226,
      "grad_norm": 0.04345703125,
      "learning_rate": 1.7529803438897346e-05,
      "loss": 0.4544,
      "step": 655
    },
    {
      "epoch": 1.7902506007552352,
      "grad_norm": 0.04541015625,
      "learning_rate": 1.7497675184765064e-05,
      "loss": 0.4991,
      "step": 656
    },
    {
      "epoch": 1.7929969104016479,
      "grad_norm": 0.04296875,
      "learning_rate": 1.746553514103611e-05,
      "loss": 0.5494,
      "step": 657
    },
    {
      "epoch": 1.7957432200480605,
      "grad_norm": 0.04345703125,
      "learning_rate": 1.743338345941883e-05,
      "loss": 0.4772,
      "step": 658
    },
    {
      "epoch": 1.7984895296944732,
      "grad_norm": 0.04638671875,
      "learning_rate": 1.74012202916765e-05,
      "loss": 0.5995,
      "step": 659
    },
    {
      "epoch": 1.8012358393408858,
      "grad_norm": 0.0439453125,
      "learning_rate": 1.7369045789626603e-05,
      "loss": 0.5156,
      "step": 660
    },
    {
      "epoch": 1.8039821489872985,
      "grad_norm": 0.0498046875,
      "learning_rate": 1.7336860105140134e-05,
      "loss": 0.3329,
      "step": 661
    },
    {
      "epoch": 1.806728458633711,
      "grad_norm": 0.05078125,
      "learning_rate": 1.730466339014086e-05,
      "loss": 0.4797,
      "step": 662
    },
    {
      "epoch": 1.8094747682801235,
      "grad_norm": 0.043212890625,
      "learning_rate": 1.7272455796604622e-05,
      "loss": 0.4494,
      "step": 663
    },
    {
      "epoch": 1.8122210779265362,
      "grad_norm": 0.04345703125,
      "learning_rate": 1.7240237476558615e-05,
      "loss": 0.5881,
      "step": 664
    },
    {
      "epoch": 1.8149673875729488,
      "grad_norm": 0.046142578125,
      "learning_rate": 1.7208008582080652e-05,
      "loss": 0.451,
      "step": 665
    },
    {
      "epoch": 1.8177136972193615,
      "grad_norm": 0.05419921875,
      "learning_rate": 1.7175769265298472e-05,
      "loss": 0.3846,
      "step": 666
    },
    {
      "epoch": 1.8204600068657741,
      "grad_norm": 0.04443359375,
      "learning_rate": 1.7143519678389004e-05,
      "loss": 0.4766,
      "step": 667
    },
    {
      "epoch": 1.8232063165121868,
      "grad_norm": 0.04638671875,
      "learning_rate": 1.7111259973577655e-05,
      "loss": 0.4932,
      "step": 668
    },
    {
      "epoch": 1.8259526261585994,
      "grad_norm": 0.045654296875,
      "learning_rate": 1.7078990303137584e-05,
      "loss": 0.4978,
      "step": 669
    },
    {
      "epoch": 1.8286989358050119,
      "grad_norm": 0.043701171875,
      "learning_rate": 1.7046710819389012e-05,
      "loss": 0.5164,
      "step": 670
    },
    {
      "epoch": 1.8314452454514245,
      "grad_norm": 0.04248046875,
      "learning_rate": 1.7014421674698458e-05,
      "loss": 0.5542,
      "step": 671
    },
    {
      "epoch": 1.8341915550978372,
      "grad_norm": 0.041259765625,
      "learning_rate": 1.6982123021478046e-05,
      "loss": 0.3729,
      "step": 672
    },
    {
      "epoch": 1.8369378647442498,
      "grad_norm": 0.046142578125,
      "learning_rate": 1.6949815012184795e-05,
      "loss": 0.4723,
      "step": 673
    },
    {
      "epoch": 1.8396841743906625,
      "grad_norm": 0.044921875,
      "learning_rate": 1.6917497799319876e-05,
      "loss": 0.5643,
      "step": 674
    },
    {
      "epoch": 1.842430484037075,
      "grad_norm": 0.052978515625,
      "learning_rate": 1.6885171535427913e-05,
      "loss": 0.4695,
      "step": 675
    },
    {
      "epoch": 1.8451767936834877,
      "grad_norm": 0.048095703125,
      "learning_rate": 1.685283637309623e-05,
      "loss": 0.4316,
      "step": 676
    },
    {
      "epoch": 1.8479231033299004,
      "grad_norm": 0.041748046875,
      "learning_rate": 1.6820492464954187e-05,
      "loss": 0.4624,
      "step": 677
    },
    {
      "epoch": 1.850669412976313,
      "grad_norm": 0.0439453125,
      "learning_rate": 1.67881399636724e-05,
      "loss": 0.4515,
      "step": 678
    },
    {
      "epoch": 1.8534157226227257,
      "grad_norm": 0.046630859375,
      "learning_rate": 1.6755779021962056e-05,
      "loss": 0.5498,
      "step": 679
    },
    {
      "epoch": 1.8561620322691383,
      "grad_norm": 0.0419921875,
      "learning_rate": 1.6723409792574185e-05,
      "loss": 0.4184,
      "step": 680
    },
    {
      "epoch": 1.858908341915551,
      "grad_norm": 0.042236328125,
      "learning_rate": 1.6691032428298934e-05,
      "loss": 0.437,
      "step": 681
    },
    {
      "epoch": 1.8616546515619636,
      "grad_norm": 0.046142578125,
      "learning_rate": 1.665864708196485e-05,
      "loss": 0.5498,
      "step": 682
    },
    {
      "epoch": 1.8644009612083763,
      "grad_norm": 0.046630859375,
      "learning_rate": 1.6626253906438148e-05,
      "loss": 0.4403,
      "step": 683
    },
    {
      "epoch": 1.867147270854789,
      "grad_norm": 0.04736328125,
      "learning_rate": 1.6593853054622016e-05,
      "loss": 0.5116,
      "step": 684
    },
    {
      "epoch": 1.8698935805012016,
      "grad_norm": 0.04345703125,
      "learning_rate": 1.6561444679455858e-05,
      "loss": 0.4179,
      "step": 685
    },
    {
      "epoch": 1.8726398901476142,
      "grad_norm": 0.04638671875,
      "learning_rate": 1.6529028933914604e-05,
      "loss": 0.4291,
      "step": 686
    },
    {
      "epoch": 1.8753861997940269,
      "grad_norm": 0.047119140625,
      "learning_rate": 1.649660597100797e-05,
      "loss": 0.4856,
      "step": 687
    },
    {
      "epoch": 1.8781325094404395,
      "grad_norm": 0.042236328125,
      "learning_rate": 1.646417594377973e-05,
      "loss": 0.5419,
      "step": 688
    },
    {
      "epoch": 1.8808788190868522,
      "grad_norm": 0.045166015625,
      "learning_rate": 1.6431739005307014e-05,
      "loss": 0.4287,
      "step": 689
    },
    {
      "epoch": 1.8836251287332648,
      "grad_norm": 0.05078125,
      "learning_rate": 1.6399295308699572e-05,
      "loss": 0.4848,
      "step": 690
    },
    {
      "epoch": 1.8863714383796775,
      "grad_norm": 0.044677734375,
      "learning_rate": 1.636684500709905e-05,
      "loss": 0.3635,
      "step": 691
    },
    {
      "epoch": 1.88911774802609,
      "grad_norm": 0.04541015625,
      "learning_rate": 1.6334388253678285e-05,
      "loss": 0.5319,
      "step": 692
    },
    {
      "epoch": 1.8918640576725025,
      "grad_norm": 0.047607421875,
      "learning_rate": 1.6301925201640542e-05,
      "loss": 0.5852,
      "step": 693
    },
    {
      "epoch": 1.8946103673189152,
      "grad_norm": 0.0419921875,
      "learning_rate": 1.6269456004218844e-05,
      "loss": 0.5184,
      "step": 694
    },
    {
      "epoch": 1.8973566769653278,
      "grad_norm": 0.04638671875,
      "learning_rate": 1.6236980814675204e-05,
      "loss": 0.4528,
      "step": 695
    },
    {
      "epoch": 1.9001029866117405,
      "grad_norm": 0.04541015625,
      "learning_rate": 1.620449978629993e-05,
      "loss": 0.4608,
      "step": 696
    },
    {
      "epoch": 1.9028492962581531,
      "grad_norm": 0.048095703125,
      "learning_rate": 1.617201307241088e-05,
      "loss": 0.5007,
      "step": 697
    },
    {
      "epoch": 1.9055956059045658,
      "grad_norm": 0.0458984375,
      "learning_rate": 1.6139520826352765e-05,
      "loss": 0.5226,
      "step": 698
    },
    {
      "epoch": 1.9083419155509782,
      "grad_norm": 0.045166015625,
      "learning_rate": 1.6107023201496378e-05,
      "loss": 0.4345,
      "step": 699
    },
    {
      "epoch": 1.9110882251973909,
      "grad_norm": 0.048583984375,
      "learning_rate": 1.6074520351237947e-05,
      "loss": 0.4386,
      "step": 700
    },
    {
      "epoch": 1.9138345348438035,
      "grad_norm": 0.042724609375,
      "learning_rate": 1.6042012428998325e-05,
      "loss": 0.4791,
      "step": 701
    },
    {
      "epoch": 1.9165808444902162,
      "grad_norm": 0.04443359375,
      "learning_rate": 1.6009499588222325e-05,
      "loss": 0.3982,
      "step": 702
    },
    {
      "epoch": 1.9193271541366288,
      "grad_norm": 0.044189453125,
      "learning_rate": 1.597698198237797e-05,
      "loss": 0.3487,
      "step": 703
    },
    {
      "epoch": 1.9220734637830414,
      "grad_norm": 0.046630859375,
      "learning_rate": 1.5944459764955784e-05,
      "loss": 0.3082,
      "step": 704
    },
    {
      "epoch": 1.924819773429454,
      "grad_norm": 0.04443359375,
      "learning_rate": 1.5911933089468048e-05,
      "loss": 0.4835,
      "step": 705
    },
    {
      "epoch": 1.9275660830758667,
      "grad_norm": 0.047607421875,
      "learning_rate": 1.5879402109448093e-05,
      "loss": 0.503,
      "step": 706
    },
    {
      "epoch": 1.9303123927222794,
      "grad_norm": 0.04638671875,
      "learning_rate": 1.584686697844956e-05,
      "loss": 0.5597,
      "step": 707
    },
    {
      "epoch": 1.933058702368692,
      "grad_norm": 0.044189453125,
      "learning_rate": 1.5814327850045697e-05,
      "loss": 0.5074,
      "step": 708
    },
    {
      "epoch": 1.9358050120151047,
      "grad_norm": 0.05029296875,
      "learning_rate": 1.5781784877828607e-05,
      "loss": 0.6022,
      "step": 709
    },
    {
      "epoch": 1.9385513216615173,
      "grad_norm": 0.043701171875,
      "learning_rate": 1.5749238215408548e-05,
      "loss": 0.5197,
      "step": 710
    },
    {
      "epoch": 1.94129763130793,
      "grad_norm": 0.044677734375,
      "learning_rate": 1.571668801641319e-05,
      "loss": 0.5147,
      "step": 711
    },
    {
      "epoch": 1.9440439409543426,
      "grad_norm": 0.042724609375,
      "learning_rate": 1.5684134434486893e-05,
      "loss": 0.5506,
      "step": 712
    },
    {
      "epoch": 1.9467902506007553,
      "grad_norm": 0.044921875,
      "learning_rate": 1.565157762329e-05,
      "loss": 0.4598,
      "step": 713
    },
    {
      "epoch": 1.949536560247168,
      "grad_norm": 0.048828125,
      "learning_rate": 1.5619017736498076e-05,
      "loss": 0.5802,
      "step": 714
    },
    {
      "epoch": 1.9522828698935806,
      "grad_norm": 0.046630859375,
      "learning_rate": 1.5586454927801223e-05,
      "loss": 0.567,
      "step": 715
    },
    {
      "epoch": 1.9550291795399932,
      "grad_norm": 0.04931640625,
      "learning_rate": 1.555388935090332e-05,
      "loss": 0.5956,
      "step": 716
    },
    {
      "epoch": 1.9577754891864059,
      "grad_norm": 0.0400390625,
      "learning_rate": 1.5521321159521326e-05,
      "loss": 0.4019,
      "step": 717
    },
    {
      "epoch": 1.9605217988328185,
      "grad_norm": 0.045654296875,
      "learning_rate": 1.548875050738453e-05,
      "loss": 0.4996,
      "step": 718
    },
    {
      "epoch": 1.9632681084792312,
      "grad_norm": 0.0498046875,
      "learning_rate": 1.545617754823384e-05,
      "loss": 0.4999,
      "step": 719
    },
    {
      "epoch": 1.9660144181256438,
      "grad_norm": 0.047607421875,
      "learning_rate": 1.5423602435821055e-05,
      "loss": 0.6049,
      "step": 720
    },
    {
      "epoch": 1.9687607277720565,
      "grad_norm": 0.049560546875,
      "learning_rate": 1.5391025323908134e-05,
      "loss": 0.4799,
      "step": 721
    },
    {
      "epoch": 1.9715070374184689,
      "grad_norm": 0.043212890625,
      "learning_rate": 1.5358446366266483e-05,
      "loss": 0.4836,
      "step": 722
    },
    {
      "epoch": 1.9742533470648815,
      "grad_norm": 0.1767578125,
      "learning_rate": 1.532586571667621e-05,
      "loss": 1.1204,
      "step": 723
    },
    {
      "epoch": 1.9769996567112942,
      "grad_norm": 0.043701171875,
      "learning_rate": 1.5293283528925412e-05,
      "loss": 0.4647,
      "step": 724
    },
    {
      "epoch": 1.9797459663577068,
      "grad_norm": 0.042236328125,
      "learning_rate": 1.5260699956809456e-05,
      "loss": 0.4984,
      "step": 725
    },
    {
      "epoch": 1.9824922760041195,
      "grad_norm": 0.04052734375,
      "learning_rate": 1.522811515413023e-05,
      "loss": 0.4019,
      "step": 726
    },
    {
      "epoch": 1.9852385856505321,
      "grad_norm": 0.041748046875,
      "learning_rate": 1.5195529274695436e-05,
      "loss": 0.4028,
      "step": 727
    },
    {
      "epoch": 1.9879848952969448,
      "grad_norm": 0.05029296875,
      "learning_rate": 1.5162942472317858e-05,
      "loss": 0.5023,
      "step": 728
    },
    {
      "epoch": 1.9879848952969448,
      "eval_loss": 0.5045637488365173,
      "eval_runtime": 616.8097,
      "eval_samples_per_second": 14.86,
      "eval_steps_per_second": 14.86,
      "step": 728
    },
    {
      "epoch": 1.9907312049433572,
      "grad_norm": 0.056884765625,
      "learning_rate": 1.5130354900814643e-05,
      "loss": 0.4917,
      "step": 729
    },
    {
      "epoch": 1.9934775145897699,
      "grad_norm": 0.0498046875,
      "learning_rate": 1.5097766714006553e-05,
      "loss": 0.4892,
      "step": 730
    },
    {
      "epoch": 1.9962238242361825,
      "grad_norm": 0.04736328125,
      "learning_rate": 1.5065178065717274e-05,
      "loss": 0.4261,
      "step": 731
    },
    {
      "epoch": 1.9989701338825951,
      "grad_norm": 0.044921875,
      "learning_rate": 1.5032589109772655e-05,
      "loss": 0.5345,
      "step": 732
    },
    {
      "epoch": 2.001716443529008,
      "grad_norm": 0.048095703125,
      "learning_rate": 1.5e-05,
      "loss": 0.5285,
      "step": 733
    },
    {
      "epoch": 2.0044627531754204,
      "grad_norm": 0.044189453125,
      "learning_rate": 1.4967410890227347e-05,
      "loss": 0.4216,
      "step": 734
    },
    {
      "epoch": 2.007209062821833,
      "grad_norm": 0.044921875,
      "learning_rate": 1.4934821934282728e-05,
      "loss": 0.5005,
      "step": 735
    },
    {
      "epoch": 2.0099553724682457,
      "grad_norm": 0.0458984375,
      "learning_rate": 1.4902233285993447e-05,
      "loss": 0.495,
      "step": 736
    },
    {
      "epoch": 2.0127016821146584,
      "grad_norm": 0.04638671875,
      "learning_rate": 1.4869645099185361e-05,
      "loss": 0.4652,
      "step": 737
    },
    {
      "epoch": 2.0013731548232063,
      "grad_norm": 0.051025390625,
      "learning_rate": 1.4837057527682142e-05,
      "loss": 0.621,
      "step": 738
    },
    {
      "epoch": 2.004119464469619,
      "grad_norm": 0.0419921875,
      "learning_rate": 1.4804470725304567e-05,
      "loss": 0.353,
      "step": 739
    },
    {
      "epoch": 2.0068657741160316,
      "grad_norm": 0.042724609375,
      "learning_rate": 1.4771884845869772e-05,
      "loss": 0.431,
      "step": 740
    },
    {
      "epoch": 2.0096120837624443,
      "grad_norm": 0.04638671875,
      "learning_rate": 1.4739300043190547e-05,
      "loss": 0.5136,
      "step": 741
    },
    {
      "epoch": 2.012358393408857,
      "grad_norm": 0.040771484375,
      "learning_rate": 1.470671647107459e-05,
      "loss": 0.3657,
      "step": 742
    },
    {
      "epoch": 2.0151047030552696,
      "grad_norm": 0.043701171875,
      "learning_rate": 1.4674134283323792e-05,
      "loss": 0.5771,
      "step": 743
    },
    {
      "epoch": 2.017851012701682,
      "grad_norm": 0.039794921875,
      "learning_rate": 1.4641553633733519e-05,
      "loss": 0.3684,
      "step": 744
    },
    {
      "epoch": 2.020597322348095,
      "grad_norm": 0.047119140625,
      "learning_rate": 1.460897467609187e-05,
      "loss": 0.5238,
      "step": 745
    },
    {
      "epoch": 2.0233436319945075,
      "grad_norm": 0.0654296875,
      "learning_rate": 1.4576397564178951e-05,
      "loss": 0.451,
      "step": 746
    },
    {
      "epoch": 2.02608994164092,
      "grad_norm": 0.043701171875,
      "learning_rate": 1.4543822451766166e-05,
      "loss": 0.5708,
      "step": 747
    },
    {
      "epoch": 2.028836251287333,
      "grad_norm": 0.050537109375,
      "learning_rate": 1.4511249492615477e-05,
      "loss": 0.6172,
      "step": 748
    },
    {
      "epoch": 2.0315825609337455,
      "grad_norm": 0.056884765625,
      "learning_rate": 1.447867884047868e-05,
      "loss": 0.6446,
      "step": 749
    },
    {
      "epoch": 2.034328870580158,
      "grad_norm": 0.044677734375,
      "learning_rate": 1.4446110649096683e-05,
      "loss": 0.5011,
      "step": 750
    },
    {
      "epoch": 2.0370751802265707,
      "grad_norm": 0.054443359375,
      "learning_rate": 1.4413545072198783e-05,
      "loss": 0.4862,
      "step": 751
    },
    {
      "epoch": 2.0398214898729834,
      "grad_norm": 0.046875,
      "learning_rate": 1.438098226350193e-05,
      "loss": 0.4955,
      "step": 752
    },
    {
      "epoch": 2.0425677995193956,
      "grad_norm": 0.048828125,
      "learning_rate": 1.4348422376710009e-05,
      "loss": 0.3889,
      "step": 753
    },
    {
      "epoch": 2.0453141091658082,
      "grad_norm": 0.166015625,
      "learning_rate": 1.4315865565513111e-05,
      "loss": 1.1553,
      "step": 754
    },
    {
      "epoch": 2.048060418812221,
      "grad_norm": 0.044921875,
      "learning_rate": 1.4283311983586818e-05,
      "loss": 0.5192,
      "step": 755
    },
    {
      "epoch": 2.0508067284586335,
      "grad_norm": 0.048828125,
      "learning_rate": 1.4250761784591451e-05,
      "loss": 0.4926,
      "step": 756
    },
    {
      "epoch": 2.053553038105046,
      "grad_norm": 0.044921875,
      "learning_rate": 1.4218215122171392e-05,
      "loss": 0.5362,
      "step": 757
    },
    {
      "epoch": 2.056299347751459,
      "grad_norm": 0.0517578125,
      "learning_rate": 1.4185672149954304e-05,
      "loss": 0.4926,
      "step": 758
    },
    {
      "epoch": 2.0590456573978715,
      "grad_norm": 0.0478515625,
      "learning_rate": 1.4153133021550438e-05,
      "loss": 0.6137,
      "step": 759
    },
    {
      "epoch": 2.061791967044284,
      "grad_norm": 0.04296875,
      "learning_rate": 1.4120597890551908e-05,
      "loss": 0.4648,
      "step": 760
    },
    {
      "epoch": 2.064538276690697,
      "grad_norm": 0.045166015625,
      "learning_rate": 1.4088066910531951e-05,
      "loss": 0.6486,
      "step": 761
    },
    {
      "epoch": 2.0672845863371094,
      "grad_norm": 0.042236328125,
      "learning_rate": 1.4055540235044213e-05,
      "loss": 0.4291,
      "step": 762
    },
    {
      "epoch": 2.070030895983522,
      "grad_norm": 0.04736328125,
      "learning_rate": 1.402301801762203e-05,
      "loss": 0.515,
      "step": 763
    },
    {
      "epoch": 2.0727772056299347,
      "grad_norm": 0.05224609375,
      "learning_rate": 1.3990500411777677e-05,
      "loss": 0.6079,
      "step": 764
    },
    {
      "epoch": 2.0755235152763474,
      "grad_norm": 0.046875,
      "learning_rate": 1.3957987571001676e-05,
      "loss": 0.5589,
      "step": 765
    },
    {
      "epoch": 2.07826982492276,
      "grad_norm": 0.049560546875,
      "learning_rate": 1.3925479648762055e-05,
      "loss": 0.6439,
      "step": 766
    },
    {
      "epoch": 2.0810161345691727,
      "grad_norm": 0.053955078125,
      "learning_rate": 1.3892976798503621e-05,
      "loss": 0.4723,
      "step": 767
    },
    {
      "epoch": 2.0837624442155853,
      "grad_norm": 0.048828125,
      "learning_rate": 1.3860479173647241e-05,
      "loss": 0.5328,
      "step": 768
    },
    {
      "epoch": 2.086508753861998,
      "grad_norm": 0.046142578125,
      "learning_rate": 1.3827986927589118e-05,
      "loss": 0.5182,
      "step": 769
    },
    {
      "epoch": 2.0892550635084106,
      "grad_norm": 0.043701171875,
      "learning_rate": 1.3795500213700072e-05,
      "loss": 0.4433,
      "step": 770
    },
    {
      "epoch": 2.0920013731548233,
      "grad_norm": 0.044921875,
      "learning_rate": 1.3763019185324797e-05,
      "loss": 0.483,
      "step": 771
    },
    {
      "epoch": 2.094747682801236,
      "grad_norm": 0.046630859375,
      "learning_rate": 1.3730543995781158e-05,
      "loss": 0.4826,
      "step": 772
    },
    {
      "epoch": 2.0974939924476486,
      "grad_norm": 0.054931640625,
      "learning_rate": 1.3698074798359458e-05,
      "loss": 0.5313,
      "step": 773
    },
    {
      "epoch": 2.100240302094061,
      "grad_norm": 0.04833984375,
      "learning_rate": 1.3665611746321718e-05,
      "loss": 0.4303,
      "step": 774
    },
    {
      "epoch": 2.102986611740474,
      "grad_norm": 0.0546875,
      "learning_rate": 1.363315499290095e-05,
      "loss": 0.5252,
      "step": 775
    },
    {
      "epoch": 2.1057329213868865,
      "grad_norm": 0.044921875,
      "learning_rate": 1.360070469130043e-05,
      "loss": 0.4501,
      "step": 776
    },
    {
      "epoch": 2.108479231033299,
      "grad_norm": 0.044921875,
      "learning_rate": 1.3568260994692988e-05,
      "loss": 0.4423,
      "step": 777
    },
    {
      "epoch": 2.111225540679712,
      "grad_norm": 0.04638671875,
      "learning_rate": 1.3535824056220273e-05,
      "loss": 0.5341,
      "step": 778
    },
    {
      "epoch": 2.1139718503261244,
      "grad_norm": 0.042724609375,
      "learning_rate": 1.3503394028992032e-05,
      "loss": 0.4019,
      "step": 779
    },
    {
      "epoch": 2.116718159972537,
      "grad_norm": 0.051513671875,
      "learning_rate": 1.3470971066085395e-05,
      "loss": 0.5329,
      "step": 780
    },
    {
      "epoch": 2.1194644696189497,
      "grad_norm": 0.041748046875,
      "learning_rate": 1.3438555320544143e-05,
      "loss": 0.5412,
      "step": 781
    },
    {
      "epoch": 2.1222107792653624,
      "grad_norm": 0.0439453125,
      "learning_rate": 1.3406146945377987e-05,
      "loss": 0.4902,
      "step": 782
    },
    {
      "epoch": 2.1249570889117746,
      "grad_norm": 0.04736328125,
      "learning_rate": 1.3373746093561855e-05,
      "loss": 0.6356,
      "step": 783
    },
    {
      "epoch": 2.1277033985581872,
      "grad_norm": 0.047119140625,
      "learning_rate": 1.3341352918035156e-05,
      "loss": 0.4674,
      "step": 784
    },
    {
      "epoch": 2.1304497082046,
      "grad_norm": 0.05029296875,
      "learning_rate": 1.330896757170107e-05,
      "loss": 0.5155,
      "step": 785
    },
    {
      "epoch": 2.1331960178510125,
      "grad_norm": 0.046875,
      "learning_rate": 1.327659020742582e-05,
      "loss": 0.5342,
      "step": 786
    },
    {
      "epoch": 2.135942327497425,
      "grad_norm": 0.047607421875,
      "learning_rate": 1.3244220978037945e-05,
      "loss": 0.5219,
      "step": 787
    },
    {
      "epoch": 2.138688637143838,
      "grad_norm": 0.042724609375,
      "learning_rate": 1.3211860036327604e-05,
      "loss": 0.5404,
      "step": 788
    },
    {
      "epoch": 2.1414349467902505,
      "grad_norm": 0.05078125,
      "learning_rate": 1.3179507535045819e-05,
      "loss": 0.4683,
      "step": 789
    },
    {
      "epoch": 2.144181256436663,
      "grad_norm": 0.049560546875,
      "learning_rate": 1.3147163626903774e-05,
      "loss": 0.4784,
      "step": 790
    },
    {
      "epoch": 2.1469275660830758,
      "grad_norm": 0.0478515625,
      "learning_rate": 1.3114828464572096e-05,
      "loss": 0.5399,
      "step": 791
    },
    {
      "epoch": 2.1496738757294884,
      "grad_norm": 0.041259765625,
      "learning_rate": 1.3082502200680128e-05,
      "loss": 0.425,
      "step": 792
    },
    {
      "epoch": 2.152420185375901,
      "grad_norm": 0.044189453125,
      "learning_rate": 1.305018498781521e-05,
      "loss": 0.4779,
      "step": 793
    },
    {
      "epoch": 2.1551664950223137,
      "grad_norm": 0.04443359375,
      "learning_rate": 1.301787697852196e-05,
      "loss": 0.5025,
      "step": 794
    },
    {
      "epoch": 2.1579128046687264,
      "grad_norm": 0.043701171875,
      "learning_rate": 1.298557832530155e-05,
      "loss": 0.5144,
      "step": 795
    },
    {
      "epoch": 2.160659114315139,
      "grad_norm": 0.042724609375,
      "learning_rate": 1.2953289180610994e-05,
      "loss": 0.4723,
      "step": 796
    },
    {
      "epoch": 2.1634054239615517,
      "grad_norm": 0.04248046875,
      "learning_rate": 1.2921009696862419e-05,
      "loss": 0.5309,
      "step": 797
    },
    {
      "epoch": 2.1661517336079643,
      "grad_norm": 0.048828125,
      "learning_rate": 1.2888740026422354e-05,
      "loss": 0.3767,
      "step": 798
    },
    {
      "epoch": 2.168898043254377,
      "grad_norm": 0.044189453125,
      "learning_rate": 1.2856480321611004e-05,
      "loss": 0.559,
      "step": 799
    },
    {
      "epoch": 2.1716443529007896,
      "grad_norm": 0.048095703125,
      "learning_rate": 1.2824230734701535e-05,
      "loss": 0.565,
      "step": 800
    },
    {
      "epoch": 2.1743906625472023,
      "grad_norm": 0.0458984375,
      "learning_rate": 1.2791991417919347e-05,
      "loss": 0.5005,
      "step": 801
    },
    {
      "epoch": 2.177136972193615,
      "grad_norm": 0.04248046875,
      "learning_rate": 1.2759762523441386e-05,
      "loss": 0.4312,
      "step": 802
    },
    {
      "epoch": 2.1798832818400276,
      "grad_norm": 0.044189453125,
      "learning_rate": 1.2727544203395377e-05,
      "loss": 0.5526,
      "step": 803
    },
    {
      "epoch": 2.18262959148644,
      "grad_norm": 0.047119140625,
      "learning_rate": 1.269533660985914e-05,
      "loss": 0.4463,
      "step": 804
    },
    {
      "epoch": 2.185375901132853,
      "grad_norm": 0.043701171875,
      "learning_rate": 1.2663139894859867e-05,
      "loss": 0.4219,
      "step": 805
    },
    {
      "epoch": 2.1881222107792655,
      "grad_norm": 0.045166015625,
      "learning_rate": 1.2630954210373396e-05,
      "loss": 0.3865,
      "step": 806
    },
    {
      "epoch": 2.190868520425678,
      "grad_norm": 0.045654296875,
      "learning_rate": 1.2598779708323499e-05,
      "loss": 0.5792,
      "step": 807
    },
    {
      "epoch": 2.193614830072091,
      "grad_norm": 0.043701171875,
      "learning_rate": 1.2566616540581168e-05,
      "loss": 0.462,
      "step": 808
    },
    {
      "epoch": 2.1963611397185034,
      "grad_norm": 0.046142578125,
      "learning_rate": 1.2534464858963892e-05,
      "loss": 0.4869,
      "step": 809
    },
    {
      "epoch": 2.199107449364916,
      "grad_norm": 0.0458984375,
      "learning_rate": 1.2502324815234942e-05,
      "loss": 0.5559,
      "step": 810
    },
    {
      "epoch": 2.2018537590113283,
      "grad_norm": 0.041015625,
      "learning_rate": 1.2470196561102655e-05,
      "loss": 0.3752,
      "step": 811
    },
    {
      "epoch": 2.204600068657741,
      "grad_norm": 0.04443359375,
      "learning_rate": 1.2438080248219723e-05,
      "loss": 0.5054,
      "step": 812
    },
    {
      "epoch": 2.2073463783041536,
      "grad_norm": 0.041259765625,
      "learning_rate": 1.2405976028182464e-05,
      "loss": 0.3525,
      "step": 813
    },
    {
      "epoch": 2.2100926879505662,
      "grad_norm": 0.043701171875,
      "learning_rate": 1.2373884052530127e-05,
      "loss": 0.4951,
      "step": 814
    },
    {
      "epoch": 2.212838997596979,
      "grad_norm": 0.047119140625,
      "learning_rate": 1.2341804472744157e-05,
      "loss": 0.4484,
      "step": 815
    },
    {
      "epoch": 2.2155853072433915,
      "grad_norm": 0.046630859375,
      "learning_rate": 1.2309737440247486e-05,
      "loss": 0.5412,
      "step": 816
    },
    {
      "epoch": 2.218331616889804,
      "grad_norm": 0.050048828125,
      "learning_rate": 1.2277683106403826e-05,
      "loss": 0.6162,
      "step": 817
    },
    {
      "epoch": 2.221077926536217,
      "grad_norm": 0.04443359375,
      "learning_rate": 1.2245641622516943e-05,
      "loss": 0.4606,
      "step": 818
    },
    {
      "epoch": 2.2238242361826295,
      "grad_norm": 0.055908203125,
      "learning_rate": 1.2213613139829949e-05,
      "loss": 0.3737,
      "step": 819
    },
    {
      "epoch": 2.2238242361826295,
      "eval_loss": 0.504136323928833,
      "eval_runtime": 615.1173,
      "eval_samples_per_second": 14.901,
      "eval_steps_per_second": 14.901,
      "step": 819
    },
    {
      "epoch": 2.226570545829042,
      "grad_norm": 0.044921875,
      "learning_rate": 1.2181597809524594e-05,
      "loss": 0.3953,
      "step": 820
    },
    {
      "epoch": 2.2293168554754548,
      "grad_norm": 0.044189453125,
      "learning_rate": 1.2149595782720537e-05,
      "loss": 0.4174,
      "step": 821
    },
    {
      "epoch": 2.2320631651218674,
      "grad_norm": 0.04638671875,
      "learning_rate": 1.2117607210474645e-05,
      "loss": 0.5269,
      "step": 822
    },
    {
      "epoch": 2.23480947476828,
      "grad_norm": 0.042724609375,
      "learning_rate": 1.2085632243780278e-05,
      "loss": 0.4668,
      "step": 823
    },
    {
      "epoch": 2.2375557844146927,
      "grad_norm": 0.046630859375,
      "learning_rate": 1.205367103356657e-05,
      "loss": 0.4565,
      "step": 824
    },
    {
      "epoch": 2.2403020940611054,
      "grad_norm": 0.043701171875,
      "learning_rate": 1.202172373069773e-05,
      "loss": 0.4427,
      "step": 825
    },
    {
      "epoch": 2.243048403707518,
      "grad_norm": 0.06103515625,
      "learning_rate": 1.1989790485972312e-05,
      "loss": 0.4414,
      "step": 826
    },
    {
      "epoch": 2.2457947133539307,
      "grad_norm": 0.046142578125,
      "learning_rate": 1.1957871450122516e-05,
      "loss": 0.5547,
      "step": 827
    },
    {
      "epoch": 2.2485410230003433,
      "grad_norm": 0.044677734375,
      "learning_rate": 1.1925966773813476e-05,
      "loss": 0.5273,
      "step": 828
    },
    {
      "epoch": 2.251287332646756,
      "grad_norm": 0.0439453125,
      "learning_rate": 1.1894076607642537e-05,
      "loss": 0.5066,
      "step": 829
    },
    {
      "epoch": 2.2540336422931686,
      "grad_norm": 0.047119140625,
      "learning_rate": 1.1862201102138562e-05,
      "loss": 0.5397,
      "step": 830
    },
    {
      "epoch": 2.2567799519395813,
      "grad_norm": 0.0478515625,
      "learning_rate": 1.1830340407761207e-05,
      "loss": 0.4944,
      "step": 831
    },
    {
      "epoch": 2.259526261585994,
      "grad_norm": 0.042724609375,
      "learning_rate": 1.1798494674900222e-05,
      "loss": 0.4056,
      "step": 832
    },
    {
      "epoch": 2.2622725712324065,
      "grad_norm": 0.050537109375,
      "learning_rate": 1.1766664053874726e-05,
      "loss": 0.5453,
      "step": 833
    },
    {
      "epoch": 2.265018880878819,
      "grad_norm": 0.04638671875,
      "learning_rate": 1.1734848694932514e-05,
      "loss": 0.456,
      "step": 834
    },
    {
      "epoch": 2.267765190525232,
      "grad_norm": 0.050537109375,
      "learning_rate": 1.170304874824934e-05,
      "loss": 0.5696,
      "step": 835
    },
    {
      "epoch": 2.2705115001716445,
      "grad_norm": 0.04248046875,
      "learning_rate": 1.1671264363928205e-05,
      "loss": 0.4873,
      "step": 836
    },
    {
      "epoch": 2.273257809818057,
      "grad_norm": 0.046142578125,
      "learning_rate": 1.1639495691998653e-05,
      "loss": 0.5142,
      "step": 837
    },
    {
      "epoch": 2.27600411946447,
      "grad_norm": 0.044189453125,
      "learning_rate": 1.1607742882416064e-05,
      "loss": 0.4905,
      "step": 838
    },
    {
      "epoch": 2.2787504291108824,
      "grad_norm": 0.044677734375,
      "learning_rate": 1.1576006085060941e-05,
      "loss": 0.4352,
      "step": 839
    },
    {
      "epoch": 2.281496738757295,
      "grad_norm": 0.05078125,
      "learning_rate": 1.1544285449738211e-05,
      "loss": 0.5675,
      "step": 840
    },
    {
      "epoch": 2.2842430484037077,
      "grad_norm": 0.047119140625,
      "learning_rate": 1.1512581126176508e-05,
      "loss": 0.4553,
      "step": 841
    },
    {
      "epoch": 2.2869893580501204,
      "grad_norm": 0.052001953125,
      "learning_rate": 1.1480893264027469e-05,
      "loss": 0.5391,
      "step": 842
    },
    {
      "epoch": 2.289735667696533,
      "grad_norm": 0.04736328125,
      "learning_rate": 1.1449222012865037e-05,
      "loss": 0.5003,
      "step": 843
    },
    {
      "epoch": 2.2924819773429452,
      "grad_norm": 0.051513671875,
      "learning_rate": 1.1417567522184738e-05,
      "loss": 0.5302,
      "step": 844
    },
    {
      "epoch": 2.295228286989358,
      "grad_norm": 0.04931640625,
      "learning_rate": 1.1385929941402993e-05,
      "loss": 0.571,
      "step": 845
    },
    {
      "epoch": 2.2979745966357705,
      "grad_norm": 0.04638671875,
      "learning_rate": 1.1354309419856392e-05,
      "loss": 0.5726,
      "step": 846
    },
    {
      "epoch": 2.300720906282183,
      "grad_norm": 0.045166015625,
      "learning_rate": 1.1322706106801025e-05,
      "loss": 0.5884,
      "step": 847
    },
    {
      "epoch": 2.303467215928596,
      "grad_norm": 0.046142578125,
      "learning_rate": 1.1291120151411731e-05,
      "loss": 0.5926,
      "step": 848
    },
    {
      "epoch": 2.3062135255750085,
      "grad_norm": 0.0439453125,
      "learning_rate": 1.1259551702781426e-05,
      "loss": 0.4487,
      "step": 849
    },
    {
      "epoch": 2.308959835221421,
      "grad_norm": 0.040771484375,
      "learning_rate": 1.1228000909920388e-05,
      "loss": 0.3924,
      "step": 850
    },
    {
      "epoch": 2.3117061448678338,
      "grad_norm": 0.044189453125,
      "learning_rate": 1.119646792175556e-05,
      "loss": 0.4218,
      "step": 851
    },
    {
      "epoch": 2.3144524545142464,
      "grad_norm": 0.236328125,
      "learning_rate": 1.1164952887129836e-05,
      "loss": 1.1613,
      "step": 852
    },
    {
      "epoch": 2.317198764160659,
      "grad_norm": 0.05029296875,
      "learning_rate": 1.1133455954801372e-05,
      "loss": 0.4224,
      "step": 853
    },
    {
      "epoch": 2.3199450738070717,
      "grad_norm": 0.045166015625,
      "learning_rate": 1.1101977273442873e-05,
      "loss": 0.4405,
      "step": 854
    },
    {
      "epoch": 2.3226913834534844,
      "grad_norm": 0.046142578125,
      "learning_rate": 1.1070516991640894e-05,
      "loss": 0.4972,
      "step": 855
    },
    {
      "epoch": 2.325437693099897,
      "grad_norm": 0.044921875,
      "learning_rate": 1.1039075257895146e-05,
      "loss": 0.5403,
      "step": 856
    },
    {
      "epoch": 2.3281840027463097,
      "grad_norm": 0.046630859375,
      "learning_rate": 1.1007652220617778e-05,
      "loss": 0.5295,
      "step": 857
    },
    {
      "epoch": 2.3309303123927223,
      "grad_norm": 0.046630859375,
      "learning_rate": 1.0976248028132705e-05,
      "loss": 0.5899,
      "step": 858
    },
    {
      "epoch": 2.333676622039135,
      "grad_norm": 0.04541015625,
      "learning_rate": 1.0944862828674872e-05,
      "loss": 0.4907,
      "step": 859
    },
    {
      "epoch": 2.3364229316855476,
      "grad_norm": 0.047119140625,
      "learning_rate": 1.0913496770389585e-05,
      "loss": 0.5142,
      "step": 860
    },
    {
      "epoch": 2.3391692413319602,
      "grad_norm": 0.04150390625,
      "learning_rate": 1.088215000133179e-05,
      "loss": 0.5103,
      "step": 861
    },
    {
      "epoch": 2.341915550978373,
      "grad_norm": 0.04833984375,
      "learning_rate": 1.0850822669465392e-05,
      "loss": 0.5814,
      "step": 862
    },
    {
      "epoch": 2.3446618606247855,
      "grad_norm": 0.0517578125,
      "learning_rate": 1.081951492266254e-05,
      "loss": 0.5544,
      "step": 863
    },
    {
      "epoch": 2.347408170271198,
      "grad_norm": 0.1416015625,
      "learning_rate": 1.0788226908702945e-05,
      "loss": 1.1435,
      "step": 864
    },
    {
      "epoch": 2.350154479917611,
      "grad_norm": 0.043701171875,
      "learning_rate": 1.0756958775273169e-05,
      "loss": 0.4895,
      "step": 865
    },
    {
      "epoch": 2.3529007895640235,
      "grad_norm": 0.045654296875,
      "learning_rate": 1.0725710669965936e-05,
      "loss": 0.5886,
      "step": 866
    },
    {
      "epoch": 2.355647099210436,
      "grad_norm": 0.04248046875,
      "learning_rate": 1.0694482740279428e-05,
      "loss": 0.4469,
      "step": 867
    },
    {
      "epoch": 2.358393408856849,
      "grad_norm": 0.043212890625,
      "learning_rate": 1.0663275133616603e-05,
      "loss": 0.4049,
      "step": 868
    },
    {
      "epoch": 2.3611397185032614,
      "grad_norm": 0.04052734375,
      "learning_rate": 1.063208799728448e-05,
      "loss": 0.3659,
      "step": 869
    },
    {
      "epoch": 2.3638860281496736,
      "grad_norm": 0.0458984375,
      "learning_rate": 1.0600921478493455e-05,
      "loss": 0.5023,
      "step": 870
    },
    {
      "epoch": 2.3666323377960863,
      "grad_norm": 0.04296875,
      "learning_rate": 1.0569775724356611e-05,
      "loss": 0.4065,
      "step": 871
    },
    {
      "epoch": 2.369378647442499,
      "grad_norm": 0.043701171875,
      "learning_rate": 1.0538650881889013e-05,
      "loss": 0.4033,
      "step": 872
    },
    {
      "epoch": 2.3721249570889116,
      "grad_norm": 0.039794921875,
      "learning_rate": 1.0507547098007015e-05,
      "loss": 0.4139,
      "step": 873
    },
    {
      "epoch": 2.3748712667353242,
      "grad_norm": 0.048095703125,
      "learning_rate": 1.0476464519527574e-05,
      "loss": 0.5499,
      "step": 874
    },
    {
      "epoch": 2.377617576381737,
      "grad_norm": 0.047119140625,
      "learning_rate": 1.0445403293167547e-05,
      "loss": 0.4546,
      "step": 875
    },
    {
      "epoch": 2.3803638860281495,
      "grad_norm": 0.04150390625,
      "learning_rate": 1.0414363565543016e-05,
      "loss": 0.4263,
      "step": 876
    },
    {
      "epoch": 2.383110195674562,
      "grad_norm": 0.049072265625,
      "learning_rate": 1.0383345483168573e-05,
      "loss": 0.446,
      "step": 877
    },
    {
      "epoch": 2.385856505320975,
      "grad_norm": 0.044921875,
      "learning_rate": 1.0352349192456643e-05,
      "loss": 0.445,
      "step": 878
    },
    {
      "epoch": 2.3886028149673875,
      "grad_norm": 0.04833984375,
      "learning_rate": 1.032137483971679e-05,
      "loss": 0.5788,
      "step": 879
    },
    {
      "epoch": 2.3913491246138,
      "grad_norm": 0.04638671875,
      "learning_rate": 1.0290422571155024e-05,
      "loss": 0.4657,
      "step": 880
    },
    {
      "epoch": 2.3940954342602128,
      "grad_norm": 0.04345703125,
      "learning_rate": 1.0259492532873113e-05,
      "loss": 0.5917,
      "step": 881
    },
    {
      "epoch": 2.3968417439066254,
      "grad_norm": 0.05126953125,
      "learning_rate": 1.0228584870867896e-05,
      "loss": 0.4403,
      "step": 882
    },
    {
      "epoch": 2.399588053553038,
      "grad_norm": 0.041259765625,
      "learning_rate": 1.0197699731030584e-05,
      "loss": 0.4274,
      "step": 883
    },
    {
      "epoch": 2.4023343631994507,
      "grad_norm": 0.048095703125,
      "learning_rate": 1.016683725914609e-05,
      "loss": 0.5997,
      "step": 884
    },
    {
      "epoch": 2.4050806728458634,
      "grad_norm": 0.046630859375,
      "learning_rate": 1.0135997600892316e-05,
      "loss": 0.4612,
      "step": 885
    },
    {
      "epoch": 2.407826982492276,
      "grad_norm": 0.044677734375,
      "learning_rate": 1.0105180901839487e-05,
      "loss": 0.4969,
      "step": 886
    },
    {
      "epoch": 2.4105732921386887,
      "grad_norm": 0.045166015625,
      "learning_rate": 1.0074387307449452e-05,
      "loss": 0.4989,
      "step": 887
    },
    {
      "epoch": 2.4133196017851013,
      "grad_norm": 0.05029296875,
      "learning_rate": 1.0043616963075001e-05,
      "loss": 0.4879,
      "step": 888
    },
    {
      "epoch": 2.416065911431514,
      "grad_norm": 0.046630859375,
      "learning_rate": 1.0012870013959182e-05,
      "loss": 0.4521,
      "step": 889
    },
    {
      "epoch": 2.4188122210779266,
      "grad_norm": 0.0498046875,
      "learning_rate": 9.982146605234604e-06,
      "loss": 0.6295,
      "step": 890
    },
    {
      "epoch": 2.4215585307243392,
      "grad_norm": 0.044921875,
      "learning_rate": 9.95144688192277e-06,
      "loss": 0.4266,
      "step": 891
    },
    {
      "epoch": 2.424304840370752,
      "grad_norm": 0.044677734375,
      "learning_rate": 9.920770988933366e-06,
      "loss": 0.3733,
      "step": 892
    },
    {
      "epoch": 2.4270511500171645,
      "grad_norm": 0.046630859375,
      "learning_rate": 9.890119071063624e-06,
      "loss": 0.5311,
      "step": 893
    },
    {
      "epoch": 2.429797459663577,
      "grad_norm": 0.040283203125,
      "learning_rate": 9.859491272997579e-06,
      "loss": 0.3879,
      "step": 894
    },
    {
      "epoch": 2.43254376930999,
      "grad_norm": 0.048095703125,
      "learning_rate": 9.828887739305423e-06,
      "loss": 0.4734,
      "step": 895
    },
    {
      "epoch": 2.4352900789564025,
      "grad_norm": 0.05224609375,
      "learning_rate": 9.798308614442822e-06,
      "loss": 0.4965,
      "step": 896
    },
    {
      "epoch": 2.438036388602815,
      "grad_norm": 0.05224609375,
      "learning_rate": 9.767754042750214e-06,
      "loss": 0.502,
      "step": 897
    },
    {
      "epoch": 2.4407826982492278,
      "grad_norm": 0.046142578125,
      "learning_rate": 9.737224168452154e-06,
      "loss": 0.4924,
      "step": 898
    },
    {
      "epoch": 2.4435290078956404,
      "grad_norm": 0.045166015625,
      "learning_rate": 9.706719135656613e-06,
      "loss": 0.5116,
      "step": 899
    },
    {
      "epoch": 2.446275317542053,
      "grad_norm": 0.048095703125,
      "learning_rate": 9.676239088354302e-06,
      "loss": 0.5165,
      "step": 900
    },
    {
      "epoch": 2.4490216271884657,
      "grad_norm": 0.04296875,
      "learning_rate": 9.645784170417996e-06,
      "loss": 0.4293,
      "step": 901
    },
    {
      "epoch": 2.4517679368348784,
      "grad_norm": 0.04443359375,
      "learning_rate": 9.615354525601859e-06,
      "loss": 0.5352,
      "step": 902
    },
    {
      "epoch": 2.4545142464812906,
      "grad_norm": 0.042236328125,
      "learning_rate": 9.584950297540759e-06,
      "loss": 0.474,
      "step": 903
    },
    {
      "epoch": 2.457260556127703,
      "grad_norm": 0.047607421875,
      "learning_rate": 9.554571629749585e-06,
      "loss": 0.495,
      "step": 904
    },
    {
      "epoch": 2.460006865774116,
      "grad_norm": 0.049072265625,
      "learning_rate": 9.524218665622578e-06,
      "loss": 0.4432,
      "step": 905
    },
    {
      "epoch": 2.4627531754205285,
      "grad_norm": 0.050048828125,
      "learning_rate": 9.493891548432654e-06,
      "loss": 0.4222,
      "step": 906
    },
    {
      "epoch": 2.465499485066941,
      "grad_norm": 0.045166015625,
      "learning_rate": 9.463590421330727e-06,
      "loss": 0.5396,
      "step": 907
    },
    {
      "epoch": 2.468245794713354,
      "grad_norm": 0.048095703125,
      "learning_rate": 9.433315427345028e-06,
      "loss": 0.4899,
      "step": 908
    },
    {
      "epoch": 2.4709921043597665,
      "grad_norm": 0.047119140625,
      "learning_rate": 9.403066709380432e-06,
      "loss": 0.6021,
      "step": 909
    },
    {
      "epoch": 2.473738414006179,
      "grad_norm": 0.048095703125,
      "learning_rate": 9.372844410217792e-06,
      "loss": 0.505,
      "step": 910
    },
    {
      "epoch": 2.473738414006179,
      "eval_loss": 0.5039077997207642,
      "eval_runtime": 617.6957,
      "eval_samples_per_second": 14.839,
      "eval_steps_per_second": 14.839,
      "step": 910
    },
    {
      "epoch": 2.4764847236525918,
      "grad_norm": 0.051513671875,
      "learning_rate": 9.342648672513254e-06,
      "loss": 0.5927,
      "step": 911
    },
    {
      "epoch": 2.4792310332990044,
      "grad_norm": 0.05126953125,
      "learning_rate": 9.31247963879759e-06,
      "loss": 0.4284,
      "step": 912
    },
    {
      "epoch": 2.481977342945417,
      "grad_norm": 0.0458984375,
      "learning_rate": 9.28233745147552e-06,
      "loss": 0.4401,
      "step": 913
    },
    {
      "epoch": 2.4847236525918297,
      "grad_norm": 0.045166015625,
      "learning_rate": 9.252222252825043e-06,
      "loss": 0.5268,
      "step": 914
    },
    {
      "epoch": 2.4874699622382423,
      "grad_norm": 0.045654296875,
      "learning_rate": 9.222134184996769e-06,
      "loss": 0.5029,
      "step": 915
    },
    {
      "epoch": 2.490216271884655,
      "grad_norm": 0.044677734375,
      "learning_rate": 9.19207339001324e-06,
      "loss": 0.4947,
      "step": 916
    },
    {
      "epoch": 2.4929625815310676,
      "grad_norm": 0.048828125,
      "learning_rate": 9.16204000976827e-06,
      "loss": 0.5041,
      "step": 917
    },
    {
      "epoch": 2.4957088911774803,
      "grad_norm": 0.0439453125,
      "learning_rate": 9.13203418602626e-06,
      "loss": 0.4975,
      "step": 918
    },
    {
      "epoch": 2.498455200823893,
      "grad_norm": 0.045166015625,
      "learning_rate": 9.102056060421545e-06,
      "loss": 0.4445,
      "step": 919
    },
    {
      "epoch": 2.5012015104703056,
      "grad_norm": 0.049072265625,
      "learning_rate": 9.07210577445772e-06,
      "loss": 0.4441,
      "step": 920
    },
    {
      "epoch": 2.5039478201167182,
      "grad_norm": 0.045654296875,
      "learning_rate": 9.042183469506964e-06,
      "loss": 0.5294,
      "step": 921
    },
    {
      "epoch": 2.506694129763131,
      "grad_norm": 0.049072265625,
      "learning_rate": 9.012289286809384e-06,
      "loss": 0.5546,
      "step": 922
    },
    {
      "epoch": 2.5094404394095435,
      "grad_norm": 0.1748046875,
      "learning_rate": 8.982423367472344e-06,
      "loss": 1.2097,
      "step": 923
    },
    {
      "epoch": 2.512186749055956,
      "grad_norm": 0.046630859375,
      "learning_rate": 8.95258585246979e-06,
      "loss": 0.5349,
      "step": 924
    },
    {
      "epoch": 2.514933058702369,
      "grad_norm": 0.05078125,
      "learning_rate": 8.922776882641604e-06,
      "loss": 0.531,
      "step": 925
    },
    {
      "epoch": 2.5176793683487815,
      "grad_norm": 0.043701171875,
      "learning_rate": 8.892996598692928e-06,
      "loss": 0.4151,
      "step": 926
    },
    {
      "epoch": 2.5204256779951937,
      "grad_norm": 0.0478515625,
      "learning_rate": 8.863245141193487e-06,
      "loss": 0.4963,
      "step": 927
    },
    {
      "epoch": 2.5231719876416063,
      "grad_norm": 0.045654296875,
      "learning_rate": 8.833522650576955e-06,
      "loss": 0.5466,
      "step": 928
    },
    {
      "epoch": 2.525918297288019,
      "grad_norm": 0.043212890625,
      "learning_rate": 8.803829267140263e-06,
      "loss": 0.4034,
      "step": 929
    },
    {
      "epoch": 2.5286646069344316,
      "grad_norm": 0.048095703125,
      "learning_rate": 8.774165131042957e-06,
      "loss": 0.4094,
      "step": 930
    },
    {
      "epoch": 2.5314109165808443,
      "grad_norm": 0.0537109375,
      "learning_rate": 8.744530382306528e-06,
      "loss": 0.47,
      "step": 931
    },
    {
      "epoch": 2.534157226227257,
      "grad_norm": 0.04345703125,
      "learning_rate": 8.714925160813752e-06,
      "loss": 0.4783,
      "step": 932
    },
    {
      "epoch": 2.5369035358736696,
      "grad_norm": 0.0390625,
      "learning_rate": 8.68534960630802e-06,
      "loss": 0.4795,
      "step": 933
    },
    {
      "epoch": 2.539649845520082,
      "grad_norm": 0.04833984375,
      "learning_rate": 8.655803858392707e-06,
      "loss": 0.5637,
      "step": 934
    },
    {
      "epoch": 2.542396155166495,
      "grad_norm": 0.047607421875,
      "learning_rate": 8.626288056530474e-06,
      "loss": 0.5958,
      "step": 935
    },
    {
      "epoch": 2.5451424648129075,
      "grad_norm": 0.0546875,
      "learning_rate": 8.596802340042648e-06,
      "loss": 0.5443,
      "step": 936
    },
    {
      "epoch": 2.54788877445932,
      "grad_norm": 0.04443359375,
      "learning_rate": 8.567346848108523e-06,
      "loss": 0.5042,
      "step": 937
    },
    {
      "epoch": 2.550635084105733,
      "grad_norm": 0.04052734375,
      "learning_rate": 8.53792171976476e-06,
      "loss": 0.3745,
      "step": 938
    },
    {
      "epoch": 2.5533813937521455,
      "grad_norm": 0.051025390625,
      "learning_rate": 8.508527093904663e-06,
      "loss": 0.4595,
      "step": 939
    },
    {
      "epoch": 2.556127703398558,
      "grad_norm": 0.048583984375,
      "learning_rate": 8.479163109277583e-06,
      "loss": 0.5502,
      "step": 940
    },
    {
      "epoch": 2.5588740130449708,
      "grad_norm": 0.044189453125,
      "learning_rate": 8.449829904488216e-06,
      "loss": 0.4784,
      "step": 941
    },
    {
      "epoch": 2.5616203226913834,
      "grad_norm": 0.043212890625,
      "learning_rate": 8.42052761799599e-06,
      "loss": 0.5084,
      "step": 942
    },
    {
      "epoch": 2.564366632337796,
      "grad_norm": 0.04736328125,
      "learning_rate": 8.391256388114367e-06,
      "loss": 0.4844,
      "step": 943
    },
    {
      "epoch": 2.5671129419842087,
      "grad_norm": 0.047607421875,
      "learning_rate": 8.362016353010248e-06,
      "loss": 0.5863,
      "step": 944
    },
    {
      "epoch": 2.5698592516306213,
      "grad_norm": 0.0517578125,
      "learning_rate": 8.332807650703255e-06,
      "loss": 0.453,
      "step": 945
    },
    {
      "epoch": 2.572605561277034,
      "grad_norm": 0.051513671875,
      "learning_rate": 8.303630419065136e-06,
      "loss": 0.6364,
      "step": 946
    },
    {
      "epoch": 2.5753518709234466,
      "grad_norm": 0.047119140625,
      "learning_rate": 8.274484795819068e-06,
      "loss": 0.521,
      "step": 947
    },
    {
      "epoch": 2.5780981805698593,
      "grad_norm": 0.0478515625,
      "learning_rate": 8.245370918539057e-06,
      "loss": 0.47,
      "step": 948
    },
    {
      "epoch": 2.580844490216272,
      "grad_norm": 0.04931640625,
      "learning_rate": 8.216288924649233e-06,
      "loss": 0.516,
      "step": 949
    },
    {
      "epoch": 2.5835907998626846,
      "grad_norm": 0.043701171875,
      "learning_rate": 8.187238951423254e-06,
      "loss": 0.4951,
      "step": 950
    },
    {
      "epoch": 2.5863371095090972,
      "grad_norm": 0.048583984375,
      "learning_rate": 8.158221135983606e-06,
      "loss": 0.4366,
      "step": 951
    },
    {
      "epoch": 2.58908341915551,
      "grad_norm": 0.047607421875,
      "learning_rate": 8.129235615301012e-06,
      "loss": 0.5727,
      "step": 952
    },
    {
      "epoch": 2.5918297288019225,
      "grad_norm": 0.05078125,
      "learning_rate": 8.10028252619373e-06,
      "loss": 0.601,
      "step": 953
    },
    {
      "epoch": 2.594576038448335,
      "grad_norm": 0.0439453125,
      "learning_rate": 8.07136200532695e-06,
      "loss": 0.474,
      "step": 954
    },
    {
      "epoch": 2.597322348094748,
      "grad_norm": 0.04052734375,
      "learning_rate": 8.042474189212133e-06,
      "loss": 0.3888,
      "step": 955
    },
    {
      "epoch": 2.6000686577411605,
      "grad_norm": 0.043701171875,
      "learning_rate": 8.013619214206353e-06,
      "loss": 0.4508,
      "step": 956
    },
    {
      "epoch": 2.602814967387573,
      "grad_norm": 0.04541015625,
      "learning_rate": 7.984797216511686e-06,
      "loss": 0.45,
      "step": 957
    },
    {
      "epoch": 2.6055612770339858,
      "grad_norm": 0.048583984375,
      "learning_rate": 7.956008332174523e-06,
      "loss": 0.4348,
      "step": 958
    },
    {
      "epoch": 2.6083075866803984,
      "grad_norm": 0.043212890625,
      "learning_rate": 7.927252697084976e-06,
      "loss": 0.4279,
      "step": 959
    },
    {
      "epoch": 2.611053896326811,
      "grad_norm": 0.045654296875,
      "learning_rate": 7.898530446976194e-06,
      "loss": 0.5555,
      "step": 960
    },
    {
      "epoch": 2.6138002059732237,
      "grad_norm": 0.043701171875,
      "learning_rate": 7.86984171742376e-06,
      "loss": 0.5695,
      "step": 961
    },
    {
      "epoch": 2.6165465156196364,
      "grad_norm": 0.04541015625,
      "learning_rate": 7.841186643845009e-06,
      "loss": 0.4705,
      "step": 962
    },
    {
      "epoch": 2.619292825266049,
      "grad_norm": 0.048095703125,
      "learning_rate": 7.81256536149844e-06,
      "loss": 0.486,
      "step": 963
    },
    {
      "epoch": 2.6220391349124617,
      "grad_norm": 0.0546875,
      "learning_rate": 7.783978005483024e-06,
      "loss": 0.5018,
      "step": 964
    },
    {
      "epoch": 2.624785444558874,
      "grad_norm": 0.04736328125,
      "learning_rate": 7.75542471073761e-06,
      "loss": 0.4491,
      "step": 965
    },
    {
      "epoch": 2.6275317542052865,
      "grad_norm": 0.04345703125,
      "learning_rate": 7.726905612040257e-06,
      "loss": 0.4566,
      "step": 966
    },
    {
      "epoch": 2.630278063851699,
      "grad_norm": 0.04931640625,
      "learning_rate": 7.698420844007624e-06,
      "loss": 0.5227,
      "step": 967
    },
    {
      "epoch": 2.633024373498112,
      "grad_norm": 0.049072265625,
      "learning_rate": 7.669970541094304e-06,
      "loss": 0.4866,
      "step": 968
    },
    {
      "epoch": 2.6357706831445245,
      "grad_norm": 0.047607421875,
      "learning_rate": 7.64155483759223e-06,
      "loss": 0.4499,
      "step": 969
    },
    {
      "epoch": 2.638516992790937,
      "grad_norm": 0.1337890625,
      "learning_rate": 7.613173867629991e-06,
      "loss": 0.9577,
      "step": 970
    },
    {
      "epoch": 2.6412633024373497,
      "grad_norm": 0.0498046875,
      "learning_rate": 7.584827765172254e-06,
      "loss": 0.51,
      "step": 971
    },
    {
      "epoch": 2.6440096120837624,
      "grad_norm": 0.046142578125,
      "learning_rate": 7.5565166640190784e-06,
      "loss": 0.4697,
      "step": 972
    },
    {
      "epoch": 2.646755921730175,
      "grad_norm": 0.046875,
      "learning_rate": 7.528240697805321e-06,
      "loss": 0.4789,
      "step": 973
    },
    {
      "epoch": 2.6495022313765877,
      "grad_norm": 0.045654296875,
      "learning_rate": 7.500000000000004e-06,
      "loss": 0.5087,
      "step": 974
    },
    {
      "epoch": 2.6522485410230003,
      "grad_norm": 0.045654296875,
      "learning_rate": 7.471794703905647e-06,
      "loss": 0.5238,
      "step": 975
    },
    {
      "epoch": 2.654994850669413,
      "grad_norm": 0.045654296875,
      "learning_rate": 7.443624942657698e-06,
      "loss": 0.5521,
      "step": 976
    },
    {
      "epoch": 2.6577411603158256,
      "grad_norm": 0.04833984375,
      "learning_rate": 7.415490849223844e-06,
      "loss": 0.4471,
      "step": 977
    },
    {
      "epoch": 2.6604874699622383,
      "grad_norm": 0.045654296875,
      "learning_rate": 7.387392556403433e-06,
      "loss": 0.4795,
      "step": 978
    },
    {
      "epoch": 2.663233779608651,
      "grad_norm": 0.044921875,
      "learning_rate": 7.359330196826808e-06,
      "loss": 0.4769,
      "step": 979
    },
    {
      "epoch": 2.6659800892550636,
      "grad_norm": 0.0400390625,
      "learning_rate": 7.33130390295472e-06,
      "loss": 0.3953,
      "step": 980
    },
    {
      "epoch": 2.6687263989014762,
      "grad_norm": 0.0498046875,
      "learning_rate": 7.303313807077658e-06,
      "loss": 0.5334,
      "step": 981
    },
    {
      "epoch": 2.671472708547889,
      "grad_norm": 0.046630859375,
      "learning_rate": 7.275360041315263e-06,
      "loss": 0.512,
      "step": 982
    },
    {
      "epoch": 2.6742190181943015,
      "grad_norm": 0.042236328125,
      "learning_rate": 7.24744273761569e-06,
      "loss": 0.4317,
      "step": 983
    },
    {
      "epoch": 2.676965327840714,
      "grad_norm": 0.0419921875,
      "learning_rate": 7.219562027754985e-06,
      "loss": 0.5105,
      "step": 984
    },
    {
      "epoch": 2.679711637487127,
      "grad_norm": 0.0458984375,
      "learning_rate": 7.191718043336447e-06,
      "loss": 0.4319,
      "step": 985
    },
    {
      "epoch": 2.6824579471335395,
      "grad_norm": 0.04443359375,
      "learning_rate": 7.163910915790047e-06,
      "loss": 0.4596,
      "step": 986
    },
    {
      "epoch": 2.6852042567799517,
      "grad_norm": 0.0537109375,
      "learning_rate": 7.13614077637176e-06,
      "loss": 0.4915,
      "step": 987
    },
    {
      "epoch": 2.6879505664263643,
      "grad_norm": 0.04296875,
      "learning_rate": 7.108407756162988e-06,
      "loss": 0.4317,
      "step": 988
    },
    {
      "epoch": 2.690696876072777,
      "grad_norm": 0.04296875,
      "learning_rate": 7.080711986069905e-06,
      "loss": 0.5411,
      "step": 989
    },
    {
      "epoch": 2.6934431857191896,
      "grad_norm": 0.040283203125,
      "learning_rate": 7.053053596822872e-06,
      "loss": 0.3315,
      "step": 990
    },
    {
      "epoch": 2.6961894953656023,
      "grad_norm": 0.052978515625,
      "learning_rate": 7.025432718975787e-06,
      "loss": 0.417,
      "step": 991
    },
    {
      "epoch": 2.698935805012015,
      "grad_norm": 0.048828125,
      "learning_rate": 6.997849482905506e-06,
      "loss": 0.5751,
      "step": 992
    },
    {
      "epoch": 2.7016821146584276,
      "grad_norm": 0.04736328125,
      "learning_rate": 6.970304018811183e-06,
      "loss": 0.5515,
      "step": 993
    },
    {
      "epoch": 2.70442842430484,
      "grad_norm": 0.04638671875,
      "learning_rate": 6.942796456713706e-06,
      "loss": 0.553,
      "step": 994
    },
    {
      "epoch": 2.707174733951253,
      "grad_norm": 0.043212890625,
      "learning_rate": 6.915326926455029e-06,
      "loss": 0.4753,
      "step": 995
    },
    {
      "epoch": 2.7099210435976655,
      "grad_norm": 0.05322265625,
      "learning_rate": 6.887895557697614e-06,
      "loss": 0.4289,
      "step": 996
    },
    {
      "epoch": 2.712667353244078,
      "grad_norm": 0.044189453125,
      "learning_rate": 6.860502479923769e-06,
      "loss": 0.4171,
      "step": 997
    },
    {
      "epoch": 2.715413662890491,
      "grad_norm": 0.05029296875,
      "learning_rate": 6.833147822435075e-06,
      "loss": 0.4769,
      "step": 998
    },
    {
      "epoch": 2.7181599725369034,
      "grad_norm": 0.04248046875,
      "learning_rate": 6.8058317143517615e-06,
      "loss": 0.4042,
      "step": 999
    },
    {
      "epoch": 2.720906282183316,
      "grad_norm": 0.0439453125,
      "learning_rate": 6.778554284612078e-06,
      "loss": 0.5019,
      "step": 1000
    },
    {
      "epoch": 2.7236525918297287,
      "grad_norm": 0.0458984375,
      "learning_rate": 6.751315661971731e-06,
      "loss": 0.4833,
      "step": 1001
    },
    {
      "epoch": 2.7236525918297287,
      "eval_loss": 0.5037957429885864,
      "eval_runtime": 619.9243,
      "eval_samples_per_second": 14.786,
      "eval_steps_per_second": 14.786,
      "step": 1001
    },
    {
      "epoch": 2.7263989014761414,
      "grad_norm": 0.0419921875,
      "learning_rate": 6.724115975003217e-06,
      "loss": 0.4036,
      "step": 1002
    },
    {
      "epoch": 2.729145211122554,
      "grad_norm": 0.053955078125,
      "learning_rate": 6.696955352095277e-06,
      "loss": 0.4995,
      "step": 1003
    },
    {
      "epoch": 2.7318915207689667,
      "grad_norm": 0.04052734375,
      "learning_rate": 6.6698339214522374e-06,
      "loss": 0.39,
      "step": 1004
    },
    {
      "epoch": 2.7346378304153793,
      "grad_norm": 0.04541015625,
      "learning_rate": 6.642751811093446e-06,
      "loss": 0.4771,
      "step": 1005
    },
    {
      "epoch": 2.737384140061792,
      "grad_norm": 0.0439453125,
      "learning_rate": 6.6157091488526324e-06,
      "loss": 0.4343,
      "step": 1006
    },
    {
      "epoch": 2.7401304497082046,
      "grad_norm": 0.044677734375,
      "learning_rate": 6.588706062377344e-06,
      "loss": 0.4141,
      "step": 1007
    },
    {
      "epoch": 2.7428767593546173,
      "grad_norm": 0.043701171875,
      "learning_rate": 6.561742679128296e-06,
      "loss": 0.4756,
      "step": 1008
    },
    {
      "epoch": 2.74562306900103,
      "grad_norm": 0.0458984375,
      "learning_rate": 6.534819126378821e-06,
      "loss": 0.6022,
      "step": 1009
    },
    {
      "epoch": 2.7483693786474426,
      "grad_norm": 0.049560546875,
      "learning_rate": 6.507935531214218e-06,
      "loss": 0.5495,
      "step": 1010
    },
    {
      "epoch": 2.751115688293855,
      "grad_norm": 0.04833984375,
      "learning_rate": 6.4810920205312006e-06,
      "loss": 0.4997,
      "step": 1011
    },
    {
      "epoch": 2.753861997940268,
      "grad_norm": 0.044189453125,
      "learning_rate": 6.454288721037252e-06,
      "loss": 0.438,
      "step": 1012
    },
    {
      "epoch": 2.7566083075866805,
      "grad_norm": 0.047607421875,
      "learning_rate": 6.427525759250071e-06,
      "loss": 0.5343,
      "step": 1013
    },
    {
      "epoch": 2.759354617233093,
      "grad_norm": 0.16796875,
      "learning_rate": 6.400803261496933e-06,
      "loss": 1.0934,
      "step": 1014
    },
    {
      "epoch": 2.762100926879506,
      "grad_norm": 0.04541015625,
      "learning_rate": 6.374121353914132e-06,
      "loss": 0.4902,
      "step": 1015
    },
    {
      "epoch": 2.7648472365259185,
      "grad_norm": 0.047607421875,
      "learning_rate": 6.347480162446349e-06,
      "loss": 0.6164,
      "step": 1016
    },
    {
      "epoch": 2.767593546172331,
      "grad_norm": 0.047119140625,
      "learning_rate": 6.320879812846093e-06,
      "loss": 0.3764,
      "step": 1017
    },
    {
      "epoch": 2.7703398558187438,
      "grad_norm": 0.048095703125,
      "learning_rate": 6.294320430673085e-06,
      "loss": 0.5365,
      "step": 1018
    },
    {
      "epoch": 2.7730861654651564,
      "grad_norm": 0.0458984375,
      "learning_rate": 6.267802141293657e-06,
      "loss": 0.4324,
      "step": 1019
    },
    {
      "epoch": 2.775832475111569,
      "grad_norm": 0.04248046875,
      "learning_rate": 6.241325069880198e-06,
      "loss": 0.367,
      "step": 1020
    },
    {
      "epoch": 2.7785787847579817,
      "grad_norm": 0.0498046875,
      "learning_rate": 6.214889341410512e-06,
      "loss": 0.4586,
      "step": 1021
    },
    {
      "epoch": 2.7813250944043943,
      "grad_norm": 0.050048828125,
      "learning_rate": 6.188495080667278e-06,
      "loss": 0.5402,
      "step": 1022
    },
    {
      "epoch": 2.784071404050807,
      "grad_norm": 0.04736328125,
      "learning_rate": 6.162142412237421e-06,
      "loss": 0.5498,
      "step": 1023
    },
    {
      "epoch": 2.786817713697219,
      "grad_norm": 0.05029296875,
      "learning_rate": 6.135831460511555e-06,
      "loss": 0.4409,
      "step": 1024
    },
    {
      "epoch": 2.789564023343632,
      "grad_norm": 0.043212890625,
      "learning_rate": 6.109562349683366e-06,
      "loss": 0.4341,
      "step": 1025
    },
    {
      "epoch": 2.7923103329900445,
      "grad_norm": 0.053466796875,
      "learning_rate": 6.083335203749059e-06,
      "loss": 0.6233,
      "step": 1026
    },
    {
      "epoch": 2.795056642636457,
      "grad_norm": 0.04541015625,
      "learning_rate": 6.057150146506732e-06,
      "loss": 0.5764,
      "step": 1027
    },
    {
      "epoch": 2.79780295228287,
      "grad_norm": 0.055908203125,
      "learning_rate": 6.031007301555849e-06,
      "loss": 0.4758,
      "step": 1028
    },
    {
      "epoch": 2.8005492619292824,
      "grad_norm": 0.05224609375,
      "learning_rate": 6.004906792296584e-06,
      "loss": 0.4903,
      "step": 1029
    },
    {
      "epoch": 2.803295571575695,
      "grad_norm": 0.05029296875,
      "learning_rate": 5.978848741929308e-06,
      "loss": 0.5788,
      "step": 1030
    },
    {
      "epoch": 2.8060418812221077,
      "grad_norm": 0.0439453125,
      "learning_rate": 5.952833273453953e-06,
      "loss": 0.4795,
      "step": 1031
    },
    {
      "epoch": 2.8087881908685204,
      "grad_norm": 0.044677734375,
      "learning_rate": 5.926860509669474e-06,
      "loss": 0.4128,
      "step": 1032
    },
    {
      "epoch": 2.811534500514933,
      "grad_norm": 0.04638671875,
      "learning_rate": 5.900930573173232e-06,
      "loss": 0.5129,
      "step": 1033
    },
    {
      "epoch": 2.8142808101613457,
      "grad_norm": 0.046630859375,
      "learning_rate": 5.8750435863604515e-06,
      "loss": 0.5751,
      "step": 1034
    },
    {
      "epoch": 2.8170271198077583,
      "grad_norm": 0.045166015625,
      "learning_rate": 5.849199671423609e-06,
      "loss": 0.4868,
      "step": 1035
    },
    {
      "epoch": 2.819773429454171,
      "grad_norm": 0.047607421875,
      "learning_rate": 5.823398950351886e-06,
      "loss": 0.5558,
      "step": 1036
    },
    {
      "epoch": 2.8225197391005836,
      "grad_norm": 0.04638671875,
      "learning_rate": 5.797641544930561e-06,
      "loss": 0.497,
      "step": 1037
    },
    {
      "epoch": 2.8252660487469963,
      "grad_norm": 0.045654296875,
      "learning_rate": 5.771927576740476e-06,
      "loss": 0.4415,
      "step": 1038
    },
    {
      "epoch": 2.828012358393409,
      "grad_norm": 0.04736328125,
      "learning_rate": 5.746257167157416e-06,
      "loss": 0.5724,
      "step": 1039
    },
    {
      "epoch": 2.8307586680398216,
      "grad_norm": 0.048583984375,
      "learning_rate": 5.72063043735158e-06,
      "loss": 0.5275,
      "step": 1040
    },
    {
      "epoch": 2.833504977686234,
      "grad_norm": 0.049560546875,
      "learning_rate": 5.6950475082869685e-06,
      "loss": 0.4577,
      "step": 1041
    },
    {
      "epoch": 2.836251287332647,
      "grad_norm": 0.045166015625,
      "learning_rate": 5.669508500720849e-06,
      "loss": 0.5401,
      "step": 1042
    },
    {
      "epoch": 2.8389975969790595,
      "grad_norm": 0.0419921875,
      "learning_rate": 5.6440135352031695e-06,
      "loss": 0.4133,
      "step": 1043
    },
    {
      "epoch": 2.841743906625472,
      "grad_norm": 0.051025390625,
      "learning_rate": 5.618562732075969e-06,
      "loss": 0.4756,
      "step": 1044
    },
    {
      "epoch": 2.844490216271885,
      "grad_norm": 0.04833984375,
      "learning_rate": 5.593156211472861e-06,
      "loss": 0.5736,
      "step": 1045
    },
    {
      "epoch": 2.847236525918297,
      "grad_norm": 0.044677734375,
      "learning_rate": 5.567794093318403e-06,
      "loss": 0.5078,
      "step": 1046
    },
    {
      "epoch": 2.8499828355647097,
      "grad_norm": 0.050048828125,
      "learning_rate": 5.542476497327591e-06,
      "loss": 0.5637,
      "step": 1047
    },
    {
      "epoch": 2.8527291452111223,
      "grad_norm": 0.054931640625,
      "learning_rate": 5.517203543005242e-06,
      "loss": 0.4383,
      "step": 1048
    },
    {
      "epoch": 2.855475454857535,
      "grad_norm": 0.04541015625,
      "learning_rate": 5.491975349645479e-06,
      "loss": 0.5174,
      "step": 1049
    },
    {
      "epoch": 2.8582217645039476,
      "grad_norm": 0.04248046875,
      "learning_rate": 5.466792036331117e-06,
      "loss": 0.4554,
      "step": 1050
    },
    {
      "epoch": 2.8609680741503603,
      "grad_norm": 0.04736328125,
      "learning_rate": 5.44165372193315e-06,
      "loss": 0.537,
      "step": 1051
    },
    {
      "epoch": 2.863714383796773,
      "grad_norm": 0.042724609375,
      "learning_rate": 5.416560525110149e-06,
      "loss": 0.4111,
      "step": 1052
    },
    {
      "epoch": 2.8664606934431855,
      "grad_norm": 0.048828125,
      "learning_rate": 5.391512564307737e-06,
      "loss": 0.5282,
      "step": 1053
    },
    {
      "epoch": 2.869207003089598,
      "grad_norm": 0.049072265625,
      "learning_rate": 5.36650995775799e-06,
      "loss": 0.5688,
      "step": 1054
    },
    {
      "epoch": 2.871953312736011,
      "grad_norm": 0.04248046875,
      "learning_rate": 5.341552823478929e-06,
      "loss": 0.3545,
      "step": 1055
    },
    {
      "epoch": 2.8746996223824235,
      "grad_norm": 0.04150390625,
      "learning_rate": 5.316641279273909e-06,
      "loss": 0.3866,
      "step": 1056
    },
    {
      "epoch": 2.877445932028836,
      "grad_norm": 0.044189453125,
      "learning_rate": 5.291775442731112e-06,
      "loss": 0.4777,
      "step": 1057
    },
    {
      "epoch": 2.880192241675249,
      "grad_norm": 0.0458984375,
      "learning_rate": 5.266955431222949e-06,
      "loss": 0.498,
      "step": 1058
    },
    {
      "epoch": 2.8829385513216614,
      "grad_norm": 0.047119140625,
      "learning_rate": 5.242181361905548e-06,
      "loss": 0.4791,
      "step": 1059
    },
    {
      "epoch": 2.885684860968074,
      "grad_norm": 0.044677734375,
      "learning_rate": 5.217453351718155e-06,
      "loss": 0.435,
      "step": 1060
    },
    {
      "epoch": 2.8884311706144867,
      "grad_norm": 0.04150390625,
      "learning_rate": 5.192771517382627e-06,
      "loss": 0.4513,
      "step": 1061
    },
    {
      "epoch": 2.8911774802608994,
      "grad_norm": 0.049560546875,
      "learning_rate": 5.168135975402854e-06,
      "loss": 0.5548,
      "step": 1062
    },
    {
      "epoch": 2.893923789907312,
      "grad_norm": 0.04296875,
      "learning_rate": 5.143546842064209e-06,
      "loss": 0.4624,
      "step": 1063
    },
    {
      "epoch": 2.8966700995537247,
      "grad_norm": 0.046875,
      "learning_rate": 5.1190042334330185e-06,
      "loss": 0.5901,
      "step": 1064
    },
    {
      "epoch": 2.8994164092001373,
      "grad_norm": 0.054931640625,
      "learning_rate": 5.094508265355983e-06,
      "loss": 0.5007,
      "step": 1065
    },
    {
      "epoch": 2.90216271884655,
      "grad_norm": 0.056396484375,
      "learning_rate": 5.070059053459672e-06,
      "loss": 0.3924,
      "step": 1066
    },
    {
      "epoch": 2.9049090284929626,
      "grad_norm": 0.04638671875,
      "learning_rate": 5.045656713149932e-06,
      "loss": 0.5346,
      "step": 1067
    },
    {
      "epoch": 2.9076553381393753,
      "grad_norm": 0.046630859375,
      "learning_rate": 5.021301359611387e-06,
      "loss": 0.4761,
      "step": 1068
    },
    {
      "epoch": 2.910401647785788,
      "grad_norm": 0.04150390625,
      "learning_rate": 4.996993107806853e-06,
      "loss": 0.4432,
      "step": 1069
    },
    {
      "epoch": 2.9131479574322006,
      "grad_norm": 0.04736328125,
      "learning_rate": 4.972732072476831e-06,
      "loss": 0.4404,
      "step": 1070
    },
    {
      "epoch": 2.915894267078613,
      "grad_norm": 0.046875,
      "learning_rate": 4.948518368138933e-06,
      "loss": 0.5556,
      "step": 1071
    },
    {
      "epoch": 2.918640576725026,
      "grad_norm": 0.048583984375,
      "learning_rate": 4.9243521090873745e-06,
      "loss": 0.523,
      "step": 1072
    },
    {
      "epoch": 2.9213868863714385,
      "grad_norm": 0.046875,
      "learning_rate": 4.900233409392409e-06,
      "loss": 0.5381,
      "step": 1073
    },
    {
      "epoch": 2.924133196017851,
      "grad_norm": 0.046630859375,
      "learning_rate": 4.876162382899809e-06,
      "loss": 0.5505,
      "step": 1074
    },
    {
      "epoch": 2.926879505664264,
      "grad_norm": 0.197265625,
      "learning_rate": 4.852139143230296e-06,
      "loss": 1.176,
      "step": 1075
    },
    {
      "epoch": 2.9296258153106765,
      "grad_norm": 0.042724609375,
      "learning_rate": 4.828163803779057e-06,
      "loss": 0.4169,
      "step": 1076
    },
    {
      "epoch": 2.932372124957089,
      "grad_norm": 0.04638671875,
      "learning_rate": 4.804236477715152e-06,
      "loss": 0.6101,
      "step": 1077
    },
    {
      "epoch": 2.9351184346035017,
      "grad_norm": 0.044921875,
      "learning_rate": 4.780357277981027e-06,
      "loss": 0.4059,
      "step": 1078
    },
    {
      "epoch": 2.9378647442499144,
      "grad_norm": 0.0556640625,
      "learning_rate": 4.7565263172919415e-06,
      "loss": 0.5825,
      "step": 1079
    },
    {
      "epoch": 2.940611053896327,
      "grad_norm": 0.0458984375,
      "learning_rate": 4.732743708135479e-06,
      "loss": 0.45,
      "step": 1080
    },
    {
      "epoch": 2.9433573635427397,
      "grad_norm": 0.047119140625,
      "learning_rate": 4.709009562770971e-06,
      "loss": 0.4906,
      "step": 1081
    },
    {
      "epoch": 2.9461036731891523,
      "grad_norm": 0.047119140625,
      "learning_rate": 4.685323993229005e-06,
      "loss": 0.5843,
      "step": 1082
    },
    {
      "epoch": 2.948849982835565,
      "grad_norm": 0.047607421875,
      "learning_rate": 4.661687111310865e-06,
      "loss": 0.4679,
      "step": 1083
    },
    {
      "epoch": 2.951596292481977,
      "grad_norm": 0.047119140625,
      "learning_rate": 4.638099028588034e-06,
      "loss": 0.5253,
      "step": 1084
    },
    {
      "epoch": 2.95434260212839,
      "grad_norm": 0.0439453125,
      "learning_rate": 4.614559856401635e-06,
      "loss": 0.4255,
      "step": 1085
    },
    {
      "epoch": 2.9570889117748025,
      "grad_norm": 0.044189453125,
      "learning_rate": 4.591069705861935e-06,
      "loss": 0.4591,
      "step": 1086
    },
    {
      "epoch": 2.959835221421215,
      "grad_norm": 0.047119140625,
      "learning_rate": 4.567628687847808e-06,
      "loss": 0.4433,
      "step": 1087
    },
    {
      "epoch": 2.962581531067628,
      "grad_norm": 0.045166015625,
      "learning_rate": 4.544236913006199e-06,
      "loss": 0.4516,
      "step": 1088
    },
    {
      "epoch": 2.9653278407140404,
      "grad_norm": 0.047607421875,
      "learning_rate": 4.520894491751629e-06,
      "loss": 0.5292,
      "step": 1089
    },
    {
      "epoch": 2.968074150360453,
      "grad_norm": 0.04638671875,
      "learning_rate": 4.497601534265641e-06,
      "loss": 0.5397,
      "step": 1090
    },
    {
      "epoch": 2.9708204600068657,
      "grad_norm": 0.045166015625,
      "learning_rate": 4.4743581504963206e-06,
      "loss": 0.5584,
      "step": 1091
    },
    {
      "epoch": 2.9735667696532784,
      "grad_norm": 0.04541015625,
      "learning_rate": 4.451164450157729e-06,
      "loss": 0.4986,
      "step": 1092
    },
    {
      "epoch": 2.9735667696532784,
      "eval_loss": 0.5037021636962891,
      "eval_runtime": 614.9978,
      "eval_samples_per_second": 14.904,
      "eval_steps_per_second": 14.904,
      "step": 1092
    },
    {
      "epoch": 2.976313079299691,
      "grad_norm": 0.049560546875,
      "learning_rate": 4.428020542729436e-06,
      "loss": 0.5396,
      "step": 1093
    },
    {
      "epoch": 2.9790593889461037,
      "grad_norm": 0.043701171875,
      "learning_rate": 4.4049265374559536e-06,
      "loss": 0.4538,
      "step": 1094
    },
    {
      "epoch": 2.9818056985925163,
      "grad_norm": 0.046142578125,
      "learning_rate": 4.381882543346262e-06,
      "loss": 0.3633,
      "step": 1095
    },
    {
      "epoch": 2.984552008238929,
      "grad_norm": 0.046875,
      "learning_rate": 4.358888669173264e-06,
      "loss": 0.5483,
      "step": 1096
    },
    {
      "epoch": 2.9872983178853416,
      "grad_norm": 0.05029296875,
      "learning_rate": 4.3359450234733e-06,
      "loss": 0.3848,
      "step": 1097
    },
    {
      "epoch": 2.9900446275317543,
      "grad_norm": 0.049560546875,
      "learning_rate": 4.3130517145456e-06,
      "loss": 0.6011,
      "step": 1098
    },
    {
      "epoch": 2.992790937178167,
      "grad_norm": 0.04443359375,
      "learning_rate": 4.29020885045182e-06,
      "loss": 0.4609,
      "step": 1099
    },
    {
      "epoch": 2.9955372468245796,
      "grad_norm": 0.052001953125,
      "learning_rate": 4.267416539015474e-06,
      "loss": 0.3615,
      "step": 1100
    },
    {
      "epoch": 2.998283556470992,
      "grad_norm": 0.049072265625,
      "learning_rate": 4.244674887821483e-06,
      "loss": 0.4688,
      "step": 1101
    },
    {
      "epoch": 3.0010298661174044,
      "grad_norm": 0.040771484375,
      "learning_rate": 4.221984004215623e-06,
      "loss": 0.3021,
      "step": 1102
    },
    {
      "epoch": 3.003776175763817,
      "grad_norm": 0.05322265625,
      "learning_rate": 4.199343995304044e-06,
      "loss": 0.3841,
      "step": 1103
    },
    {
      "epoch": 3.0065224854102297,
      "grad_norm": 0.055908203125,
      "learning_rate": 4.176754967952749e-06,
      "loss": 0.4316,
      "step": 1104
    },
    {
      "epoch": 3.0092687950566424,
      "grad_norm": 0.046142578125,
      "learning_rate": 4.154217028787101e-06,
      "loss": 0.5092,
      "step": 1105
    },
    {
      "epoch": 3.012015104703055,
      "grad_norm": 0.050537109375,
      "learning_rate": 4.131730284191321e-06,
      "loss": 0.4633,
      "step": 1106
    },
    {
      "epoch": 3.0020597322348093,
      "grad_norm": 0.04541015625,
      "learning_rate": 4.109294840307966e-06,
      "loss": 0.4454,
      "step": 1107
    },
    {
      "epoch": 3.004806041881222,
      "grad_norm": 0.046142578125,
      "learning_rate": 4.086910803037453e-06,
      "loss": 0.4654,
      "step": 1108
    },
    {
      "epoch": 3.0075523515276346,
      "grad_norm": 0.045654296875,
      "learning_rate": 4.064578278037542e-06,
      "loss": 0.4323,
      "step": 1109
    },
    {
      "epoch": 3.010298661174047,
      "grad_norm": 0.04248046875,
      "learning_rate": 4.042297370722851e-06,
      "loss": 0.4796,
      "step": 1110
    },
    {
      "epoch": 3.01304497082046,
      "grad_norm": 0.04345703125,
      "learning_rate": 4.0200681862643355e-06,
      "loss": 0.4253,
      "step": 1111
    },
    {
      "epoch": 3.0157912804668725,
      "grad_norm": 0.04345703125,
      "learning_rate": 3.9978908295888285e-06,
      "loss": 0.4095,
      "step": 1112
    },
    {
      "epoch": 3.018537590113285,
      "grad_norm": 0.044921875,
      "learning_rate": 3.975765405378502e-06,
      "loss": 0.4575,
      "step": 1113
    },
    {
      "epoch": 3.021283899759698,
      "grad_norm": 0.04248046875,
      "learning_rate": 3.953692018070417e-06,
      "loss": 0.4556,
      "step": 1114
    },
    {
      "epoch": 3.0240302094061104,
      "grad_norm": 0.047119140625,
      "learning_rate": 3.931670771855986e-06,
      "loss": 0.5403,
      "step": 1115
    },
    {
      "epoch": 3.026776519052523,
      "grad_norm": 0.045166015625,
      "learning_rate": 3.909701770680524e-06,
      "loss": 0.4718,
      "step": 1116
    },
    {
      "epoch": 3.0295228286989357,
      "grad_norm": 0.0419921875,
      "learning_rate": 3.887785118242722e-06,
      "loss": 0.4633,
      "step": 1117
    },
    {
      "epoch": 3.0322691383453484,
      "grad_norm": 0.045654296875,
      "learning_rate": 3.8659209179941804e-06,
      "loss": 0.5703,
      "step": 1118
    },
    {
      "epoch": 3.035015447991761,
      "grad_norm": 0.046630859375,
      "learning_rate": 3.844109273138914e-06,
      "loss": 0.5709,
      "step": 1119
    },
    {
      "epoch": 3.0377617576381737,
      "grad_norm": 0.046875,
      "learning_rate": 3.822350286632867e-06,
      "loss": 0.4592,
      "step": 1120
    },
    {
      "epoch": 3.0405080672845863,
      "grad_norm": 0.044677734375,
      "learning_rate": 3.8006440611834103e-06,
      "loss": 0.4843,
      "step": 1121
    },
    {
      "epoch": 3.043254376930999,
      "grad_norm": 0.04833984375,
      "learning_rate": 3.7789906992488875e-06,
      "loss": 0.4962,
      "step": 1122
    },
    {
      "epoch": 3.0460006865774116,
      "grad_norm": 0.0478515625,
      "learning_rate": 3.7573903030381003e-06,
      "loss": 0.5629,
      "step": 1123
    },
    {
      "epoch": 3.0487469962238243,
      "grad_norm": 0.04833984375,
      "learning_rate": 3.7358429745098525e-06,
      "loss": 0.5103,
      "step": 1124
    },
    {
      "epoch": 3.051493305870237,
      "grad_norm": 0.044189453125,
      "learning_rate": 3.7143488153724454e-06,
      "loss": 0.4677,
      "step": 1125
    },
    {
      "epoch": 3.0542396155166496,
      "grad_norm": 0.042724609375,
      "learning_rate": 3.692907927083217e-06,
      "loss": 0.423,
      "step": 1126
    },
    {
      "epoch": 3.0569859251630622,
      "grad_norm": 0.048095703125,
      "learning_rate": 3.6715204108480473e-06,
      "loss": 0.4903,
      "step": 1127
    },
    {
      "epoch": 3.059732234809475,
      "grad_norm": 0.1494140625,
      "learning_rate": 3.6501863676208984e-06,
      "loss": 0.9496,
      "step": 1128
    },
    {
      "epoch": 3.0624785444558875,
      "grad_norm": 0.044677734375,
      "learning_rate": 3.6289058981033136e-06,
      "loss": 0.4253,
      "step": 1129
    },
    {
      "epoch": 3.0652248541023,
      "grad_norm": 0.04736328125,
      "learning_rate": 3.607679102743968e-06,
      "loss": 0.5686,
      "step": 1130
    },
    {
      "epoch": 3.067971163748713,
      "grad_norm": 0.04638671875,
      "learning_rate": 3.586506081738181e-06,
      "loss": 0.5278,
      "step": 1131
    },
    {
      "epoch": 3.0707174733951255,
      "grad_norm": 0.043212890625,
      "learning_rate": 3.5653869350274357e-06,
      "loss": 0.4348,
      "step": 1132
    },
    {
      "epoch": 3.073463783041538,
      "grad_norm": 0.050537109375,
      "learning_rate": 3.5443217622989294e-06,
      "loss": 0.5263,
      "step": 1133
    },
    {
      "epoch": 3.0762100926879508,
      "grad_norm": 0.046875,
      "learning_rate": 3.5233106629850736e-06,
      "loss": 0.5263,
      "step": 1134
    },
    {
      "epoch": 3.0789564023343634,
      "grad_norm": 0.04443359375,
      "learning_rate": 3.5023537362630605e-06,
      "loss": 0.4807,
      "step": 1135
    },
    {
      "epoch": 3.0817027119807756,
      "grad_norm": 0.04638671875,
      "learning_rate": 3.4814510810543553e-06,
      "loss": 0.6053,
      "step": 1136
    },
    {
      "epoch": 3.0844490216271883,
      "grad_norm": 0.05126953125,
      "learning_rate": 3.46060279602427e-06,
      "loss": 0.391,
      "step": 1137
    },
    {
      "epoch": 3.087195331273601,
      "grad_norm": 0.0439453125,
      "learning_rate": 3.439808979581455e-06,
      "loss": 0.4525,
      "step": 1138
    },
    {
      "epoch": 3.0899416409200136,
      "grad_norm": 0.04833984375,
      "learning_rate": 3.4190697298774772e-06,
      "loss": 0.532,
      "step": 1139
    },
    {
      "epoch": 3.092687950566426,
      "grad_norm": 0.045166015625,
      "learning_rate": 3.398385144806318e-06,
      "loss": 0.5811,
      "step": 1140
    },
    {
      "epoch": 3.095434260212839,
      "grad_norm": 0.04638671875,
      "learning_rate": 3.3777553220039455e-06,
      "loss": 0.5059,
      "step": 1141
    },
    {
      "epoch": 3.0981805698592515,
      "grad_norm": 0.04638671875,
      "learning_rate": 3.357180358847822e-06,
      "loss": 0.4974,
      "step": 1142
    },
    {
      "epoch": 3.100926879505664,
      "grad_norm": 0.047607421875,
      "learning_rate": 3.3366603524564736e-06,
      "loss": 0.52,
      "step": 1143
    },
    {
      "epoch": 3.103673189152077,
      "grad_norm": 0.04296875,
      "learning_rate": 3.316195399689007e-06,
      "loss": 0.4295,
      "step": 1144
    },
    {
      "epoch": 3.1064194987984894,
      "grad_norm": 0.0498046875,
      "learning_rate": 3.2957855971446737e-06,
      "loss": 0.4381,
      "step": 1145
    },
    {
      "epoch": 3.109165808444902,
      "grad_norm": 0.048095703125,
      "learning_rate": 3.2754310411623888e-06,
      "loss": 0.4879,
      "step": 1146
    },
    {
      "epoch": 3.1119121180913147,
      "grad_norm": 0.04296875,
      "learning_rate": 3.255131827820311e-06,
      "loss": 0.4444,
      "step": 1147
    },
    {
      "epoch": 3.1146584277377274,
      "grad_norm": 0.043212890625,
      "learning_rate": 3.2348880529353484e-06,
      "loss": 0.4969,
      "step": 1148
    },
    {
      "epoch": 3.11740473738414,
      "grad_norm": 0.047607421875,
      "learning_rate": 3.21469981206274e-06,
      "loss": 0.5399,
      "step": 1149
    },
    {
      "epoch": 3.1201510470305527,
      "grad_norm": 0.04638671875,
      "learning_rate": 3.194567200495593e-06,
      "loss": 0.3839,
      "step": 1150
    },
    {
      "epoch": 3.1228973566769653,
      "grad_norm": 0.046875,
      "learning_rate": 3.1744903132644197e-06,
      "loss": 0.5803,
      "step": 1151
    },
    {
      "epoch": 3.125643666323378,
      "grad_norm": 0.048095703125,
      "learning_rate": 3.1544692451367147e-06,
      "loss": 0.5422,
      "step": 1152
    },
    {
      "epoch": 3.1283899759697906,
      "grad_norm": 0.04443359375,
      "learning_rate": 3.1345040906164787e-06,
      "loss": 0.4212,
      "step": 1153
    },
    {
      "epoch": 3.1311362856162033,
      "grad_norm": 0.04638671875,
      "learning_rate": 3.1145949439438054e-06,
      "loss": 0.4019,
      "step": 1154
    },
    {
      "epoch": 3.133882595262616,
      "grad_norm": 0.04443359375,
      "learning_rate": 3.094741899094399e-06,
      "loss": 0.3445,
      "step": 1155
    },
    {
      "epoch": 3.1366289049090286,
      "grad_norm": 0.04248046875,
      "learning_rate": 3.0749450497791693e-06,
      "loss": 0.518,
      "step": 1156
    },
    {
      "epoch": 3.139375214555441,
      "grad_norm": 0.04833984375,
      "learning_rate": 3.055204489443753e-06,
      "loss": 0.4594,
      "step": 1157
    },
    {
      "epoch": 3.142121524201854,
      "grad_norm": 0.0458984375,
      "learning_rate": 3.0355203112681063e-06,
      "loss": 0.5042,
      "step": 1158
    },
    {
      "epoch": 3.1448678338482665,
      "grad_norm": 0.041015625,
      "learning_rate": 3.0158926081660338e-06,
      "loss": 0.4187,
      "step": 1159
    },
    {
      "epoch": 3.147614143494679,
      "grad_norm": 0.047119140625,
      "learning_rate": 2.9963214727847773e-06,
      "loss": 0.5166,
      "step": 1160
    },
    {
      "epoch": 3.150360453141092,
      "grad_norm": 0.04443359375,
      "learning_rate": 2.976806997504555e-06,
      "loss": 0.4656,
      "step": 1161
    },
    {
      "epoch": 3.1531067627875045,
      "grad_norm": 0.041748046875,
      "learning_rate": 2.9573492744381475e-06,
      "loss": 0.4555,
      "step": 1162
    },
    {
      "epoch": 3.155853072433917,
      "grad_norm": 0.051513671875,
      "learning_rate": 2.9379483954304386e-06,
      "loss": 0.5357,
      "step": 1163
    },
    {
      "epoch": 3.1585993820803298,
      "grad_norm": 0.1708984375,
      "learning_rate": 2.9186044520580145e-06,
      "loss": 1.2069,
      "step": 1164
    },
    {
      "epoch": 3.1613456917267424,
      "grad_norm": 0.04736328125,
      "learning_rate": 2.8993175356286934e-06,
      "loss": 0.5469,
      "step": 1165
    },
    {
      "epoch": 3.164092001373155,
      "grad_norm": 0.045654296875,
      "learning_rate": 2.8800877371811245e-06,
      "loss": 0.5138,
      "step": 1166
    },
    {
      "epoch": 3.1668383110195673,
      "grad_norm": 0.042724609375,
      "learning_rate": 2.8609151474843377e-06,
      "loss": 0.486,
      "step": 1167
    },
    {
      "epoch": 3.16958462066598,
      "grad_norm": 0.048095703125,
      "learning_rate": 2.841799857037337e-06,
      "loss": 0.5253,
      "step": 1168
    },
    {
      "epoch": 3.1723309303123925,
      "grad_norm": 0.04443359375,
      "learning_rate": 2.822741956068648e-06,
      "loss": 0.4386,
      "step": 1169
    },
    {
      "epoch": 3.175077239958805,
      "grad_norm": 0.04541015625,
      "learning_rate": 2.803741534535916e-06,
      "loss": 0.4587,
      "step": 1170
    },
    {
      "epoch": 3.177823549605218,
      "grad_norm": 0.047119140625,
      "learning_rate": 2.7847986821254605e-06,
      "loss": 0.445,
      "step": 1171
    },
    {
      "epoch": 3.1805698592516305,
      "grad_norm": 0.047119140625,
      "learning_rate": 2.7659134882518715e-06,
      "loss": 0.532,
      "step": 1172
    },
    {
      "epoch": 3.183316168898043,
      "grad_norm": 0.045654296875,
      "learning_rate": 2.747086042057566e-06,
      "loss": 0.5697,
      "step": 1173
    },
    {
      "epoch": 3.186062478544456,
      "grad_norm": 0.047119140625,
      "learning_rate": 2.7283164324123904e-06,
      "loss": 0.5351,
      "step": 1174
    },
    {
      "epoch": 3.1888087881908684,
      "grad_norm": 0.04833984375,
      "learning_rate": 2.7096047479131848e-06,
      "loss": 0.5825,
      "step": 1175
    },
    {
      "epoch": 3.191555097837281,
      "grad_norm": 0.044189453125,
      "learning_rate": 2.6909510768833606e-06,
      "loss": 0.4346,
      "step": 1176
    },
    {
      "epoch": 3.1943014074836937,
      "grad_norm": 0.04931640625,
      "learning_rate": 2.6723555073725125e-06,
      "loss": 0.4932,
      "step": 1177
    },
    {
      "epoch": 3.1970477171301064,
      "grad_norm": 0.0458984375,
      "learning_rate": 2.653818127155959e-06,
      "loss": 0.5453,
      "step": 1178
    },
    {
      "epoch": 3.199794026776519,
      "grad_norm": 0.049072265625,
      "learning_rate": 2.635339023734374e-06,
      "loss": 0.5351,
      "step": 1179
    },
    {
      "epoch": 3.2025403364229317,
      "grad_norm": 0.04833984375,
      "learning_rate": 2.6169182843333334e-06,
      "loss": 0.4668,
      "step": 1180
    },
    {
      "epoch": 3.2052866460693443,
      "grad_norm": 0.04833984375,
      "learning_rate": 2.5985559959029347e-06,
      "loss": 0.4176,
      "step": 1181
    },
    {
      "epoch": 3.208032955715757,
      "grad_norm": 0.04638671875,
      "learning_rate": 2.5802522451173627e-06,
      "loss": 0.4717,
      "step": 1182
    },
    {
      "epoch": 3.2107792653621696,
      "grad_norm": 0.047119140625,
      "learning_rate": 2.562007118374504e-06,
      "loss": 0.5227,
      "step": 1183
    },
    {
      "epoch": 3.2107792653621696,
      "eval_loss": 0.5037118196487427,
      "eval_runtime": 639.2982,
      "eval_samples_per_second": 14.338,
      "eval_steps_per_second": 14.338,
      "step": 1183
    },
    {
      "epoch": 3.2135255750085823,
      "grad_norm": 0.04248046875,
      "learning_rate": 2.543820701795511e-06,
      "loss": 0.4845,
      "step": 1184
    },
    {
      "epoch": 3.216271884654995,
      "grad_norm": 0.1611328125,
      "learning_rate": 2.5256930812244273e-06,
      "loss": 1.1749,
      "step": 1185
    },
    {
      "epoch": 3.2190181943014076,
      "grad_norm": 0.1923828125,
      "learning_rate": 2.507624342227748e-06,
      "loss": 1.1636,
      "step": 1186
    },
    {
      "epoch": 3.22176450394782,
      "grad_norm": 0.047607421875,
      "learning_rate": 2.4896145700940524e-06,
      "loss": 0.508,
      "step": 1187
    },
    {
      "epoch": 3.224510813594233,
      "grad_norm": 0.19140625,
      "learning_rate": 2.471663849833567e-06,
      "loss": 1.1039,
      "step": 1188
    },
    {
      "epoch": 3.2272571232406455,
      "grad_norm": 0.044921875,
      "learning_rate": 2.453772266177791e-06,
      "loss": 0.5217,
      "step": 1189
    },
    {
      "epoch": 3.230003432887058,
      "grad_norm": 0.04833984375,
      "learning_rate": 2.435939903579075e-06,
      "loss": 0.5168,
      "step": 1190
    },
    {
      "epoch": 3.232749742533471,
      "grad_norm": 0.05615234375,
      "learning_rate": 2.4181668462102478e-06,
      "loss": 0.6517,
      "step": 1191
    },
    {
      "epoch": 3.2354960521798835,
      "grad_norm": 0.043212890625,
      "learning_rate": 2.4004531779641835e-06,
      "loss": 0.439,
      "step": 1192
    },
    {
      "epoch": 3.238242361826296,
      "grad_norm": 0.047119140625,
      "learning_rate": 2.382798982453444e-06,
      "loss": 0.4339,
      "step": 1193
    },
    {
      "epoch": 3.2409886714727087,
      "grad_norm": 0.044677734375,
      "learning_rate": 2.3652043430098624e-06,
      "loss": 0.493,
      "step": 1194
    },
    {
      "epoch": 3.243734981119121,
      "grad_norm": 0.046142578125,
      "learning_rate": 2.3476693426841417e-06,
      "loss": 0.389,
      "step": 1195
    },
    {
      "epoch": 3.2464812907655336,
      "grad_norm": 0.0439453125,
      "learning_rate": 2.3301940642454932e-06,
      "loss": 0.5022,
      "step": 1196
    },
    {
      "epoch": 3.2492276004119462,
      "grad_norm": 0.047119140625,
      "learning_rate": 2.3127785901812093e-06,
      "loss": 0.5275,
      "step": 1197
    },
    {
      "epoch": 3.251973910058359,
      "grad_norm": 0.049560546875,
      "learning_rate": 2.2954230026963092e-06,
      "loss": 0.5681,
      "step": 1198
    },
    {
      "epoch": 3.2547202197047715,
      "grad_norm": 0.04150390625,
      "learning_rate": 2.278127383713117e-06,
      "loss": 0.4153,
      "step": 1199
    },
    {
      "epoch": 3.257466529351184,
      "grad_norm": 0.046142578125,
      "learning_rate": 2.2608918148709057e-06,
      "loss": 0.4188,
      "step": 1200
    },
    {
      "epoch": 3.260212838997597,
      "grad_norm": 0.0458984375,
      "learning_rate": 2.2437163775254863e-06,
      "loss": 0.4721,
      "step": 1201
    },
    {
      "epoch": 3.2629591486440095,
      "grad_norm": 0.048828125,
      "learning_rate": 2.22660115274885e-06,
      "loss": 0.443,
      "step": 1202
    },
    {
      "epoch": 3.265705458290422,
      "grad_norm": 0.0478515625,
      "learning_rate": 2.2095462213287526e-06,
      "loss": 0.4396,
      "step": 1203
    },
    {
      "epoch": 3.268451767936835,
      "grad_norm": 0.042724609375,
      "learning_rate": 2.19255166376837e-06,
      "loss": 0.4431,
      "step": 1204
    },
    {
      "epoch": 3.2711980775832474,
      "grad_norm": 0.0517578125,
      "learning_rate": 2.175617560285883e-06,
      "loss": 0.5235,
      "step": 1205
    },
    {
      "epoch": 3.27394438722966,
      "grad_norm": 0.03955078125,
      "learning_rate": 2.158743990814128e-06,
      "loss": 0.357,
      "step": 1206
    },
    {
      "epoch": 3.2766906968760727,
      "grad_norm": 0.1318359375,
      "learning_rate": 2.1419310350001997e-06,
      "loss": 1.1825,
      "step": 1207
    },
    {
      "epoch": 3.2794370065224854,
      "grad_norm": 0.049072265625,
      "learning_rate": 2.1251787722050854e-06,
      "loss": 0.5559,
      "step": 1208
    },
    {
      "epoch": 3.282183316168898,
      "grad_norm": 0.045654296875,
      "learning_rate": 2.1084872815032885e-06,
      "loss": 0.529,
      "step": 1209
    },
    {
      "epoch": 3.2849296258153107,
      "grad_norm": 0.047119140625,
      "learning_rate": 2.0918566416824557e-06,
      "loss": 0.5814,
      "step": 1210
    },
    {
      "epoch": 3.2876759354617233,
      "grad_norm": 0.04736328125,
      "learning_rate": 2.075286931242995e-06,
      "loss": 0.4434,
      "step": 1211
    },
    {
      "epoch": 3.290422245108136,
      "grad_norm": 0.044677734375,
      "learning_rate": 2.058778228397726e-06,
      "loss": 0.4634,
      "step": 1212
    },
    {
      "epoch": 3.2931685547545486,
      "grad_norm": 0.0478515625,
      "learning_rate": 2.0423306110714846e-06,
      "loss": 0.4865,
      "step": 1213
    },
    {
      "epoch": 3.2959148644009613,
      "grad_norm": 0.052978515625,
      "learning_rate": 2.0259441569007836e-06,
      "loss": 0.4826,
      "step": 1214
    },
    {
      "epoch": 3.298661174047374,
      "grad_norm": 0.054931640625,
      "learning_rate": 2.0096189432334194e-06,
      "loss": 0.399,
      "step": 1215
    },
    {
      "epoch": 3.3014074836937866,
      "grad_norm": 0.053466796875,
      "learning_rate": 1.9933550471281315e-06,
      "loss": 0.5276,
      "step": 1216
    },
    {
      "epoch": 3.304153793340199,
      "grad_norm": 0.04345703125,
      "learning_rate": 1.9771525453542123e-06,
      "loss": 0.4359,
      "step": 1217
    },
    {
      "epoch": 3.306900102986612,
      "grad_norm": 0.043212890625,
      "learning_rate": 1.9610115143911696e-06,
      "loss": 0.4383,
      "step": 1218
    },
    {
      "epoch": 3.3096464126330245,
      "grad_norm": 0.044189453125,
      "learning_rate": 1.9449320304283545e-06,
      "loss": 0.4844,
      "step": 1219
    },
    {
      "epoch": 3.312392722279437,
      "grad_norm": 0.0517578125,
      "learning_rate": 1.928914169364595e-06,
      "loss": 0.4703,
      "step": 1220
    },
    {
      "epoch": 3.31513903192585,
      "grad_norm": 0.044189453125,
      "learning_rate": 1.9129580068078556e-06,
      "loss": 0.5235,
      "step": 1221
    },
    {
      "epoch": 3.3178853415722624,
      "grad_norm": 0.046630859375,
      "learning_rate": 1.8970636180748542e-06,
      "loss": 0.4168,
      "step": 1222
    },
    {
      "epoch": 3.320631651218675,
      "grad_norm": 0.048095703125,
      "learning_rate": 1.8812310781907416e-06,
      "loss": 0.6157,
      "step": 1223
    },
    {
      "epoch": 3.3233779608650877,
      "grad_norm": 0.046630859375,
      "learning_rate": 1.8654604618887095e-06,
      "loss": 0.5189,
      "step": 1224
    },
    {
      "epoch": 3.3261242705115004,
      "grad_norm": 0.0439453125,
      "learning_rate": 1.8497518436096727e-06,
      "loss": 0.5767,
      "step": 1225
    },
    {
      "epoch": 3.328870580157913,
      "grad_norm": 0.04638671875,
      "learning_rate": 1.8341052975018856e-06,
      "loss": 0.4845,
      "step": 1226
    },
    {
      "epoch": 3.3316168898043257,
      "grad_norm": 0.047119140625,
      "learning_rate": 1.8185208974206202e-06,
      "loss": 0.5266,
      "step": 1227
    },
    {
      "epoch": 3.334363199450738,
      "grad_norm": 0.0439453125,
      "learning_rate": 1.8029987169277962e-06,
      "loss": 0.4485,
      "step": 1228
    },
    {
      "epoch": 3.3371095090971505,
      "grad_norm": 0.059814453125,
      "learning_rate": 1.7875388292916516e-06,
      "loss": 0.5584,
      "step": 1229
    },
    {
      "epoch": 3.339855818743563,
      "grad_norm": 0.0400390625,
      "learning_rate": 1.7721413074863769e-06,
      "loss": 0.4716,
      "step": 1230
    },
    {
      "epoch": 3.342602128389976,
      "grad_norm": 0.04736328125,
      "learning_rate": 1.7568062241917937e-06,
      "loss": 0.5874,
      "step": 1231
    },
    {
      "epoch": 3.3453484380363885,
      "grad_norm": 0.043701171875,
      "learning_rate": 1.7415336517929886e-06,
      "loss": 0.384,
      "step": 1232
    },
    {
      "epoch": 3.348094747682801,
      "grad_norm": 0.04541015625,
      "learning_rate": 1.726323662379992e-06,
      "loss": 0.5479,
      "step": 1233
    },
    {
      "epoch": 3.350841057329214,
      "grad_norm": 0.045654296875,
      "learning_rate": 1.7111763277474179e-06,
      "loss": 0.4631,
      "step": 1234
    },
    {
      "epoch": 3.3535873669756264,
      "grad_norm": 0.049072265625,
      "learning_rate": 1.6960917193941478e-06,
      "loss": 0.5318,
      "step": 1235
    },
    {
      "epoch": 3.356333676622039,
      "grad_norm": 0.04345703125,
      "learning_rate": 1.6810699085229685e-06,
      "loss": 0.4914,
      "step": 1236
    },
    {
      "epoch": 3.3590799862684517,
      "grad_norm": 0.049072265625,
      "learning_rate": 1.6661109660402563e-06,
      "loss": 0.5059,
      "step": 1237
    },
    {
      "epoch": 3.3618262959148644,
      "grad_norm": 0.039794921875,
      "learning_rate": 1.6512149625556371e-06,
      "loss": 0.3797,
      "step": 1238
    },
    {
      "epoch": 3.364572605561277,
      "grad_norm": 0.045166015625,
      "learning_rate": 1.6363819683816372e-06,
      "loss": 0.5895,
      "step": 1239
    },
    {
      "epoch": 3.3673189152076897,
      "grad_norm": 0.045654296875,
      "learning_rate": 1.6216120535333818e-06,
      "loss": 0.5115,
      "step": 1240
    },
    {
      "epoch": 3.3700652248541023,
      "grad_norm": 0.049560546875,
      "learning_rate": 1.6069052877282292e-06,
      "loss": 0.414,
      "step": 1241
    },
    {
      "epoch": 3.372811534500515,
      "grad_norm": 0.048583984375,
      "learning_rate": 1.5922617403854768e-06,
      "loss": 0.5862,
      "step": 1242
    },
    {
      "epoch": 3.3755578441469276,
      "grad_norm": 0.05078125,
      "learning_rate": 1.5776814806260032e-06,
      "loss": 0.5591,
      "step": 1243
    },
    {
      "epoch": 3.3783041537933403,
      "grad_norm": 0.048583984375,
      "learning_rate": 1.563164577271965e-06,
      "loss": 0.5291,
      "step": 1244
    },
    {
      "epoch": 3.381050463439753,
      "grad_norm": 0.047119140625,
      "learning_rate": 1.5487110988464526e-06,
      "loss": 0.5359,
      "step": 1245
    },
    {
      "epoch": 3.3837967730861656,
      "grad_norm": 0.04345703125,
      "learning_rate": 1.5343211135731894e-06,
      "loss": 0.4184,
      "step": 1246
    },
    {
      "epoch": 3.386543082732578,
      "grad_norm": 0.04736328125,
      "learning_rate": 1.5199946893761785e-06,
      "loss": 0.4923,
      "step": 1247
    },
    {
      "epoch": 3.389289392378991,
      "grad_norm": 0.046630859375,
      "learning_rate": 1.5057318938794195e-06,
      "loss": 0.4915,
      "step": 1248
    },
    {
      "epoch": 3.3920357020254035,
      "grad_norm": 0.050048828125,
      "learning_rate": 1.4915327944065544e-06,
      "loss": 0.5856,
      "step": 1249
    },
    {
      "epoch": 3.394782011671816,
      "grad_norm": 0.048583984375,
      "learning_rate": 1.4773974579805783e-06,
      "loss": 0.5155,
      "step": 1250
    },
    {
      "epoch": 3.397528321318229,
      "grad_norm": 0.046142578125,
      "learning_rate": 1.463325951323496e-06,
      "loss": 0.5409,
      "step": 1251
    },
    {
      "epoch": 3.4002746309646414,
      "grad_norm": 0.045166015625,
      "learning_rate": 1.4493183408560361e-06,
      "loss": 0.4235,
      "step": 1252
    },
    {
      "epoch": 3.403020940611054,
      "grad_norm": 0.0458984375,
      "learning_rate": 1.4353746926973077e-06,
      "loss": 0.4903,
      "step": 1253
    },
    {
      "epoch": 3.4057672502574663,
      "grad_norm": 0.046142578125,
      "learning_rate": 1.421495072664522e-06,
      "loss": 0.5838,
      "step": 1254
    },
    {
      "epoch": 3.408513559903879,
      "grad_norm": 0.041748046875,
      "learning_rate": 1.407679546272641e-06,
      "loss": 0.4681,
      "step": 1255
    },
    {
      "epoch": 3.4112598695502916,
      "grad_norm": 0.04345703125,
      "learning_rate": 1.3939281787341102e-06,
      "loss": 0.5009,
      "step": 1256
    },
    {
      "epoch": 3.4140061791967042,
      "grad_norm": 0.04638671875,
      "learning_rate": 1.3802410349585153e-06,
      "loss": 0.5052,
      "step": 1257
    },
    {
      "epoch": 3.416752488843117,
      "grad_norm": 0.0498046875,
      "learning_rate": 1.3666181795523004e-06,
      "loss": 0.6236,
      "step": 1258
    },
    {
      "epoch": 3.4194987984895295,
      "grad_norm": 0.053466796875,
      "learning_rate": 1.3530596768184478e-06,
      "loss": 0.5415,
      "step": 1259
    },
    {
      "epoch": 3.422245108135942,
      "grad_norm": 0.04443359375,
      "learning_rate": 1.3395655907561871e-06,
      "loss": 0.4757,
      "step": 1260
    },
    {
      "epoch": 3.424991417782355,
      "grad_norm": 0.045166015625,
      "learning_rate": 1.3261359850606792e-06,
      "loss": 0.4161,
      "step": 1261
    },
    {
      "epoch": 3.4277377274287675,
      "grad_norm": 0.048095703125,
      "learning_rate": 1.3127709231227281e-06,
      "loss": 0.5019,
      "step": 1262
    },
    {
      "epoch": 3.43048403707518,
      "grad_norm": 0.04248046875,
      "learning_rate": 1.2994704680284786e-06,
      "loss": 0.4118,
      "step": 1263
    },
    {
      "epoch": 3.4332303467215928,
      "grad_norm": 0.06298828125,
      "learning_rate": 1.2862346825591075e-06,
      "loss": 0.495,
      "step": 1264
    },
    {
      "epoch": 3.4359766563680054,
      "grad_norm": 0.049072265625,
      "learning_rate": 1.2730636291905462e-06,
      "loss": 0.4988,
      "step": 1265
    },
    {
      "epoch": 3.438722966014418,
      "grad_norm": 0.04345703125,
      "learning_rate": 1.2599573700931666e-06,
      "loss": 0.3351,
      "step": 1266
    },
    {
      "epoch": 3.4414692756608307,
      "grad_norm": 0.051025390625,
      "learning_rate": 1.2469159671315072e-06,
      "loss": 0.5976,
      "step": 1267
    },
    {
      "epoch": 3.4442155853072434,
      "grad_norm": 0.04541015625,
      "learning_rate": 1.2339394818639583e-06,
      "loss": 0.5267,
      "step": 1268
    },
    {
      "epoch": 3.446961894953656,
      "grad_norm": 0.05126953125,
      "learning_rate": 1.2210279755424981e-06,
      "loss": 0.4385,
      "step": 1269
    },
    {
      "epoch": 3.4497082046000687,
      "grad_norm": 0.04150390625,
      "learning_rate": 1.2081815091123755e-06,
      "loss": 0.4201,
      "step": 1270
    },
    {
      "epoch": 3.4524545142464813,
      "grad_norm": 0.049072265625,
      "learning_rate": 1.1954001432118482e-06,
      "loss": 0.5371,
      "step": 1271
    },
    {
      "epoch": 3.455200823892894,
      "grad_norm": 0.04296875,
      "learning_rate": 1.1826839381718752e-06,
      "loss": 0.4782,
      "step": 1272
    },
    {
      "epoch": 3.4579471335393066,
      "grad_norm": 0.046630859375,
      "learning_rate": 1.1700329540158473e-06,
      "loss": 0.4626,
      "step": 1273
    },
    {
      "epoch": 3.4606934431857193,
      "grad_norm": 0.048095703125,
      "learning_rate": 1.157447250459292e-06,
      "loss": 0.5723,
      "step": 1274
    },
    {
      "epoch": 3.4606934431857193,
      "eval_loss": 0.5037119388580322,
      "eval_runtime": 636.191,
      "eval_samples_per_second": 14.408,
      "eval_steps_per_second": 14.408,
      "step": 1274
    }
  ],
  "logging_steps": 1,
  "max_steps": 1456,
  "num_input_tokens_seen": 0,
  "num_train_epochs": 4,
  "save_steps": 91,
  "stateful_callbacks": {
    "TrainerControl": {
      "args": {
        "should_epoch_stop": false,
        "should_evaluate": false,
        "should_log": false,
        "should_save": true,
        "should_training_stop": false
      },
      "attributes": {}
    }
  },
  "total_flos": 3.582524910880162e+18,
  "train_batch_size": 1,
  "trial_name": null,
  "trial_params": null
}