diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,29539 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.0778410631950095, + "eval_steps": 500, + "global_step": 3687, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.00027122321670735016, + "grad_norm": 33.32280731201172, + "learning_rate": 5.405405405405406e-09, + "loss": 1.4933, + "num_input_tokens_seen": 2097152, + "step": 1 + }, + { + "epoch": 0.0005424464334147003, + "grad_norm": 35.11781692504883, + "learning_rate": 1.0810810810810811e-08, + "loss": 1.6295, + "num_input_tokens_seen": 4194304, + "step": 2 + }, + { + "epoch": 0.0008136696501220504, + "grad_norm": 39.06733703613281, + "learning_rate": 1.6216216216216218e-08, + "loss": 1.7417, + "num_input_tokens_seen": 6291456, + "step": 3 + }, + { + "epoch": 0.0010848928668294006, + "grad_norm": 32.58023452758789, + "learning_rate": 2.1621621621621623e-08, + "loss": 1.4688, + "num_input_tokens_seen": 8388608, + "step": 4 + }, + { + "epoch": 0.0013561160835367507, + "grad_norm": 40.15693664550781, + "learning_rate": 2.7027027027027028e-08, + "loss": 2.078, + "num_input_tokens_seen": 10485760, + "step": 5 + }, + { + "epoch": 0.0016273393002441008, + "grad_norm": 33.44337463378906, + "learning_rate": 3.2432432432432436e-08, + "loss": 1.6623, + "num_input_tokens_seen": 12582912, + "step": 6 + }, + { + "epoch": 0.001898562516951451, + "grad_norm": 31.053829193115234, + "learning_rate": 3.783783783783784e-08, + "loss": 1.5643, + "num_input_tokens_seen": 14680064, + "step": 7 + }, + { + "epoch": 0.0021697857336588013, + "grad_norm": 34.127193450927734, + "learning_rate": 4.3243243243243246e-08, + "loss": 1.7671, + "num_input_tokens_seen": 16777216, + "step": 8 + }, + { + "epoch": 0.0024410089503661514, + "grad_norm": 52.89729309082031, + "learning_rate": 4.864864864864865e-08, + "loss": 2.3356, + "num_input_tokens_seen": 18874368, + "step": 9 + }, + { + "epoch": 0.0027122321670735015, + "grad_norm": 44.45952606201172, + "learning_rate": 5.4054054054054056e-08, + "loss": 2.0531, + "num_input_tokens_seen": 20971520, + "step": 10 + }, + { + "epoch": 0.0029834553837808516, + "grad_norm": 43.71432113647461, + "learning_rate": 5.945945945945946e-08, + "loss": 2.2434, + "num_input_tokens_seen": 23068672, + "step": 11 + }, + { + "epoch": 0.0032546786004882017, + "grad_norm": 35.20856857299805, + "learning_rate": 6.486486486486487e-08, + "loss": 1.5967, + "num_input_tokens_seen": 25165824, + "step": 12 + }, + { + "epoch": 0.003525901817195552, + "grad_norm": 32.52849578857422, + "learning_rate": 7.027027027027027e-08, + "loss": 1.622, + "num_input_tokens_seen": 27262976, + "step": 13 + }, + { + "epoch": 0.003797125033902902, + "grad_norm": 42.55111312866211, + "learning_rate": 7.567567567567568e-08, + "loss": 2.1261, + "num_input_tokens_seen": 29360128, + "step": 14 + }, + { + "epoch": 0.0040683482506102524, + "grad_norm": 25.502803802490234, + "learning_rate": 8.108108108108108e-08, + "loss": 1.102, + "num_input_tokens_seen": 31457280, + "step": 15 + }, + { + "epoch": 0.0043395714673176026, + "grad_norm": 41.81550216674805, + "learning_rate": 8.648648648648649e-08, + "loss": 1.6793, + "num_input_tokens_seen": 33554432, + "step": 16 + }, + { + "epoch": 0.004610794684024953, + "grad_norm": 51.921199798583984, + "learning_rate": 9.189189189189189e-08, + "loss": 2.5367, + "num_input_tokens_seen": 35651584, + "step": 17 + }, + { + "epoch": 0.004882017900732303, + "grad_norm": 39.028221130371094, + "learning_rate": 9.72972972972973e-08, + "loss": 2.173, + "num_input_tokens_seen": 37748736, + "step": 18 + }, + { + "epoch": 0.005153241117439653, + "grad_norm": 27.318889617919922, + "learning_rate": 1.0270270270270271e-07, + "loss": 1.2697, + "num_input_tokens_seen": 39845888, + "step": 19 + }, + { + "epoch": 0.005424464334147003, + "grad_norm": 24.04189682006836, + "learning_rate": 1.0810810810810811e-07, + "loss": 1.1464, + "num_input_tokens_seen": 41943040, + "step": 20 + }, + { + "epoch": 0.005695687550854353, + "grad_norm": 29.57497787475586, + "learning_rate": 1.1351351351351351e-07, + "loss": 1.2522, + "num_input_tokens_seen": 44040192, + "step": 21 + }, + { + "epoch": 0.005966910767561703, + "grad_norm": 38.306819915771484, + "learning_rate": 1.1891891891891891e-07, + "loss": 2.1454, + "num_input_tokens_seen": 46137344, + "step": 22 + }, + { + "epoch": 0.006238133984269053, + "grad_norm": 39.06193542480469, + "learning_rate": 1.2432432432432432e-07, + "loss": 2.2203, + "num_input_tokens_seen": 48234496, + "step": 23 + }, + { + "epoch": 0.006509357200976403, + "grad_norm": 40.80998229980469, + "learning_rate": 1.2972972972972974e-07, + "loss": 2.4594, + "num_input_tokens_seen": 50331648, + "step": 24 + }, + { + "epoch": 0.0067805804176837535, + "grad_norm": 33.7674560546875, + "learning_rate": 1.3513513513513515e-07, + "loss": 1.7983, + "num_input_tokens_seen": 52428800, + "step": 25 + }, + { + "epoch": 0.007051803634391104, + "grad_norm": 27.341468811035156, + "learning_rate": 1.4054054054054055e-07, + "loss": 1.2909, + "num_input_tokens_seen": 54525952, + "step": 26 + }, + { + "epoch": 0.007323026851098454, + "grad_norm": 26.22152328491211, + "learning_rate": 1.4594594594594595e-07, + "loss": 1.0993, + "num_input_tokens_seen": 56623104, + "step": 27 + }, + { + "epoch": 0.007594250067805804, + "grad_norm": 26.978374481201172, + "learning_rate": 1.5135135135135135e-07, + "loss": 1.2144, + "num_input_tokens_seen": 58720256, + "step": 28 + }, + { + "epoch": 0.007865473284513154, + "grad_norm": 27.742353439331055, + "learning_rate": 1.5675675675675675e-07, + "loss": 1.3153, + "num_input_tokens_seen": 60817408, + "step": 29 + }, + { + "epoch": 0.008136696501220505, + "grad_norm": 36.712318420410156, + "learning_rate": 1.6216216216216215e-07, + "loss": 1.9966, + "num_input_tokens_seen": 62914560, + "step": 30 + }, + { + "epoch": 0.008407919717927854, + "grad_norm": 25.96659278869629, + "learning_rate": 1.6756756756756758e-07, + "loss": 1.226, + "num_input_tokens_seen": 65011712, + "step": 31 + }, + { + "epoch": 0.008679142934635205, + "grad_norm": 27.041015625, + "learning_rate": 1.7297297297297298e-07, + "loss": 1.3386, + "num_input_tokens_seen": 67108864, + "step": 32 + }, + { + "epoch": 0.008950366151342554, + "grad_norm": 21.146181106567383, + "learning_rate": 1.7837837837837838e-07, + "loss": 0.9403, + "num_input_tokens_seen": 69206016, + "step": 33 + }, + { + "epoch": 0.009221589368049905, + "grad_norm": 25.72652244567871, + "learning_rate": 1.8378378378378379e-07, + "loss": 1.4175, + "num_input_tokens_seen": 71303168, + "step": 34 + }, + { + "epoch": 0.009492812584757255, + "grad_norm": 42.69353485107422, + "learning_rate": 1.891891891891892e-07, + "loss": 2.6364, + "num_input_tokens_seen": 73400320, + "step": 35 + }, + { + "epoch": 0.009764035801464606, + "grad_norm": 24.534645080566406, + "learning_rate": 1.945945945945946e-07, + "loss": 1.4978, + "num_input_tokens_seen": 75497472, + "step": 36 + }, + { + "epoch": 0.010035259018171955, + "grad_norm": 22.76692771911621, + "learning_rate": 2e-07, + "loss": 1.2377, + "num_input_tokens_seen": 77594624, + "step": 37 + }, + { + "epoch": 0.010306482234879306, + "grad_norm": 22.586381912231445, + "learning_rate": 2.0540540540540542e-07, + "loss": 1.2021, + "num_input_tokens_seen": 79691776, + "step": 38 + }, + { + "epoch": 0.010577705451586655, + "grad_norm": 24.025348663330078, + "learning_rate": 2.1081081081081082e-07, + "loss": 1.0536, + "num_input_tokens_seen": 81788928, + "step": 39 + }, + { + "epoch": 0.010848928668294006, + "grad_norm": 33.128868103027344, + "learning_rate": 2.1621621621621622e-07, + "loss": 1.8307, + "num_input_tokens_seen": 83886080, + "step": 40 + }, + { + "epoch": 0.011120151885001357, + "grad_norm": 34.85871887207031, + "learning_rate": 2.2162162162162162e-07, + "loss": 2.1082, + "num_input_tokens_seen": 85983232, + "step": 41 + }, + { + "epoch": 0.011391375101708706, + "grad_norm": 22.261695861816406, + "learning_rate": 2.2702702702702703e-07, + "loss": 0.9698, + "num_input_tokens_seen": 88080384, + "step": 42 + }, + { + "epoch": 0.011662598318416057, + "grad_norm": 28.77890396118164, + "learning_rate": 2.3243243243243243e-07, + "loss": 1.2482, + "num_input_tokens_seen": 90177536, + "step": 43 + }, + { + "epoch": 0.011933821535123406, + "grad_norm": 22.587844848632812, + "learning_rate": 2.3783783783783783e-07, + "loss": 1.1466, + "num_input_tokens_seen": 92274688, + "step": 44 + }, + { + "epoch": 0.012205044751830757, + "grad_norm": 18.199480056762695, + "learning_rate": 2.4324324324324326e-07, + "loss": 0.9606, + "num_input_tokens_seen": 94371840, + "step": 45 + }, + { + "epoch": 0.012476267968538107, + "grad_norm": 27.911603927612305, + "learning_rate": 2.4864864864864863e-07, + "loss": 1.6163, + "num_input_tokens_seen": 96468992, + "step": 46 + }, + { + "epoch": 0.012747491185245458, + "grad_norm": 34.57044982910156, + "learning_rate": 2.5405405405405406e-07, + "loss": 1.4498, + "num_input_tokens_seen": 98566144, + "step": 47 + }, + { + "epoch": 0.013018714401952807, + "grad_norm": 21.867979049682617, + "learning_rate": 2.594594594594595e-07, + "loss": 1.3337, + "num_input_tokens_seen": 100663296, + "step": 48 + }, + { + "epoch": 0.013289937618660158, + "grad_norm": 19.342687606811523, + "learning_rate": 2.6486486486486486e-07, + "loss": 1.0142, + "num_input_tokens_seen": 102760448, + "step": 49 + }, + { + "epoch": 0.013561160835367507, + "grad_norm": 24.752653121948242, + "learning_rate": 2.702702702702703e-07, + "loss": 1.0282, + "num_input_tokens_seen": 104857600, + "step": 50 + }, + { + "epoch": 0.013832384052074858, + "grad_norm": 20.70758628845215, + "learning_rate": 2.7567567567567567e-07, + "loss": 0.8941, + "num_input_tokens_seen": 106954752, + "step": 51 + }, + { + "epoch": 0.014103607268782207, + "grad_norm": 22.01601219177246, + "learning_rate": 2.810810810810811e-07, + "loss": 0.8012, + "num_input_tokens_seen": 109051904, + "step": 52 + }, + { + "epoch": 0.014374830485489558, + "grad_norm": 34.209354400634766, + "learning_rate": 2.8648648648648647e-07, + "loss": 2.5913, + "num_input_tokens_seen": 111149056, + "step": 53 + }, + { + "epoch": 0.014646053702196907, + "grad_norm": 19.16649627685547, + "learning_rate": 2.918918918918919e-07, + "loss": 0.9785, + "num_input_tokens_seen": 113246208, + "step": 54 + }, + { + "epoch": 0.014917276918904258, + "grad_norm": 25.85431480407715, + "learning_rate": 2.972972972972973e-07, + "loss": 1.7306, + "num_input_tokens_seen": 115343360, + "step": 55 + }, + { + "epoch": 0.015188500135611608, + "grad_norm": 19.06846809387207, + "learning_rate": 3.027027027027027e-07, + "loss": 0.9716, + "num_input_tokens_seen": 117440512, + "step": 56 + }, + { + "epoch": 0.015459723352318959, + "grad_norm": 23.541532516479492, + "learning_rate": 3.0810810810810813e-07, + "loss": 1.4204, + "num_input_tokens_seen": 119537664, + "step": 57 + }, + { + "epoch": 0.015730946569026308, + "grad_norm": 16.523822784423828, + "learning_rate": 3.135135135135135e-07, + "loss": 0.7799, + "num_input_tokens_seen": 121634816, + "step": 58 + }, + { + "epoch": 0.01600216978573366, + "grad_norm": 21.27098274230957, + "learning_rate": 3.1891891891891893e-07, + "loss": 0.9598, + "num_input_tokens_seen": 123731968, + "step": 59 + }, + { + "epoch": 0.01627339300244101, + "grad_norm": 22.250896453857422, + "learning_rate": 3.243243243243243e-07, + "loss": 0.847, + "num_input_tokens_seen": 125829120, + "step": 60 + }, + { + "epoch": 0.01654461621914836, + "grad_norm": 37.0035514831543, + "learning_rate": 3.2972972972972973e-07, + "loss": 1.2704, + "num_input_tokens_seen": 127926272, + "step": 61 + }, + { + "epoch": 0.016815839435855708, + "grad_norm": 20.06436538696289, + "learning_rate": 3.3513513513513516e-07, + "loss": 0.9768, + "num_input_tokens_seen": 130023424, + "step": 62 + }, + { + "epoch": 0.01708706265256306, + "grad_norm": 20.020830154418945, + "learning_rate": 3.4054054054054054e-07, + "loss": 1.1647, + "num_input_tokens_seen": 132120576, + "step": 63 + }, + { + "epoch": 0.01735828586927041, + "grad_norm": 19.906391143798828, + "learning_rate": 3.4594594594594597e-07, + "loss": 0.8014, + "num_input_tokens_seen": 134217728, + "step": 64 + }, + { + "epoch": 0.01762950908597776, + "grad_norm": 18.05296516418457, + "learning_rate": 3.5135135135135134e-07, + "loss": 0.9887, + "num_input_tokens_seen": 136314880, + "step": 65 + }, + { + "epoch": 0.01790073230268511, + "grad_norm": 18.5097599029541, + "learning_rate": 3.5675675675675677e-07, + "loss": 0.9417, + "num_input_tokens_seen": 138412032, + "step": 66 + }, + { + "epoch": 0.01817195551939246, + "grad_norm": 21.897937774658203, + "learning_rate": 3.6216216216216214e-07, + "loss": 1.2979, + "num_input_tokens_seen": 140509184, + "step": 67 + }, + { + "epoch": 0.01844317873609981, + "grad_norm": 23.45905113220215, + "learning_rate": 3.6756756756756757e-07, + "loss": 1.2426, + "num_input_tokens_seen": 142606336, + "step": 68 + }, + { + "epoch": 0.01871440195280716, + "grad_norm": 21.97277069091797, + "learning_rate": 3.72972972972973e-07, + "loss": 1.4745, + "num_input_tokens_seen": 144703488, + "step": 69 + }, + { + "epoch": 0.01898562516951451, + "grad_norm": 24.767820358276367, + "learning_rate": 3.783783783783784e-07, + "loss": 0.6843, + "num_input_tokens_seen": 146800640, + "step": 70 + }, + { + "epoch": 0.01925684838622186, + "grad_norm": 20.10371971130371, + "learning_rate": 3.837837837837838e-07, + "loss": 1.1348, + "num_input_tokens_seen": 148897792, + "step": 71 + }, + { + "epoch": 0.01952807160292921, + "grad_norm": 24.476518630981445, + "learning_rate": 3.891891891891892e-07, + "loss": 1.7115, + "num_input_tokens_seen": 150994944, + "step": 72 + }, + { + "epoch": 0.019799294819636562, + "grad_norm": 19.302143096923828, + "learning_rate": 3.945945945945946e-07, + "loss": 0.7886, + "num_input_tokens_seen": 153092096, + "step": 73 + }, + { + "epoch": 0.02007051803634391, + "grad_norm": 14.634087562561035, + "learning_rate": 4e-07, + "loss": 0.5778, + "num_input_tokens_seen": 155189248, + "step": 74 + }, + { + "epoch": 0.02034174125305126, + "grad_norm": 14.836756706237793, + "learning_rate": 4.054054054054054e-07, + "loss": 0.6952, + "num_input_tokens_seen": 157286400, + "step": 75 + }, + { + "epoch": 0.02061296446975861, + "grad_norm": 15.128747940063477, + "learning_rate": 4.1081081081081084e-07, + "loss": 0.7274, + "num_input_tokens_seen": 159383552, + "step": 76 + }, + { + "epoch": 0.020884187686465962, + "grad_norm": 14.638301849365234, + "learning_rate": 4.162162162162162e-07, + "loss": 0.634, + "num_input_tokens_seen": 161480704, + "step": 77 + }, + { + "epoch": 0.02115541090317331, + "grad_norm": 23.832578659057617, + "learning_rate": 4.2162162162162164e-07, + "loss": 1.0147, + "num_input_tokens_seen": 163577856, + "step": 78 + }, + { + "epoch": 0.02142663411988066, + "grad_norm": 16.2668514251709, + "learning_rate": 4.27027027027027e-07, + "loss": 0.7479, + "num_input_tokens_seen": 165675008, + "step": 79 + }, + { + "epoch": 0.021697857336588012, + "grad_norm": 14.96450424194336, + "learning_rate": 4.3243243243243244e-07, + "loss": 0.6993, + "num_input_tokens_seen": 167772160, + "step": 80 + }, + { + "epoch": 0.021969080553295363, + "grad_norm": 19.316436767578125, + "learning_rate": 4.378378378378378e-07, + "loss": 0.6489, + "num_input_tokens_seen": 169869312, + "step": 81 + }, + { + "epoch": 0.022240303770002714, + "grad_norm": 14.418977737426758, + "learning_rate": 4.4324324324324325e-07, + "loss": 0.592, + "num_input_tokens_seen": 171966464, + "step": 82 + }, + { + "epoch": 0.02251152698671006, + "grad_norm": 19.933923721313477, + "learning_rate": 4.486486486486487e-07, + "loss": 0.7344, + "num_input_tokens_seen": 174063616, + "step": 83 + }, + { + "epoch": 0.022782750203417412, + "grad_norm": 20.08726692199707, + "learning_rate": 4.5405405405405405e-07, + "loss": 0.916, + "num_input_tokens_seen": 176160768, + "step": 84 + }, + { + "epoch": 0.023053973420124763, + "grad_norm": 18.082517623901367, + "learning_rate": 4.594594594594595e-07, + "loss": 0.9287, + "num_input_tokens_seen": 178257920, + "step": 85 + }, + { + "epoch": 0.023325196636832114, + "grad_norm": 15.847949981689453, + "learning_rate": 4.6486486486486485e-07, + "loss": 0.8382, + "num_input_tokens_seen": 180355072, + "step": 86 + }, + { + "epoch": 0.023596419853539462, + "grad_norm": 22.586856842041016, + "learning_rate": 4.702702702702703e-07, + "loss": 1.0967, + "num_input_tokens_seen": 182452224, + "step": 87 + }, + { + "epoch": 0.023867643070246813, + "grad_norm": 16.663034439086914, + "learning_rate": 4.7567567567567566e-07, + "loss": 0.675, + "num_input_tokens_seen": 184549376, + "step": 88 + }, + { + "epoch": 0.024138866286954164, + "grad_norm": 19.878616333007812, + "learning_rate": 4.810810810810811e-07, + "loss": 1.1732, + "num_input_tokens_seen": 186646528, + "step": 89 + }, + { + "epoch": 0.024410089503661515, + "grad_norm": 16.984296798706055, + "learning_rate": 4.864864864864865e-07, + "loss": 0.8503, + "num_input_tokens_seen": 188743680, + "step": 90 + }, + { + "epoch": 0.024681312720368862, + "grad_norm": 14.563621520996094, + "learning_rate": 4.918918918918919e-07, + "loss": 0.4946, + "num_input_tokens_seen": 190840832, + "step": 91 + }, + { + "epoch": 0.024952535937076213, + "grad_norm": 15.319222450256348, + "learning_rate": 4.972972972972973e-07, + "loss": 0.7149, + "num_input_tokens_seen": 192937984, + "step": 92 + }, + { + "epoch": 0.025223759153783564, + "grad_norm": 14.346334457397461, + "learning_rate": 5.027027027027027e-07, + "loss": 0.3925, + "num_input_tokens_seen": 195035136, + "step": 93 + }, + { + "epoch": 0.025494982370490915, + "grad_norm": 19.126708984375, + "learning_rate": 5.081081081081081e-07, + "loss": 0.9146, + "num_input_tokens_seen": 197132288, + "step": 94 + }, + { + "epoch": 0.025766205587198263, + "grad_norm": 18.894250869750977, + "learning_rate": 5.135135135135134e-07, + "loss": 0.8788, + "num_input_tokens_seen": 199229440, + "step": 95 + }, + { + "epoch": 0.026037428803905614, + "grad_norm": 19.886512756347656, + "learning_rate": 5.18918918918919e-07, + "loss": 0.9447, + "num_input_tokens_seen": 201326592, + "step": 96 + }, + { + "epoch": 0.026308652020612965, + "grad_norm": 16.53607940673828, + "learning_rate": 5.243243243243243e-07, + "loss": 0.7017, + "num_input_tokens_seen": 203423744, + "step": 97 + }, + { + "epoch": 0.026579875237320316, + "grad_norm": 21.681087493896484, + "learning_rate": 5.297297297297297e-07, + "loss": 1.0761, + "num_input_tokens_seen": 205520896, + "step": 98 + }, + { + "epoch": 0.026851098454027666, + "grad_norm": 28.16925048828125, + "learning_rate": 5.35135135135135e-07, + "loss": 1.9542, + "num_input_tokens_seen": 207618048, + "step": 99 + }, + { + "epoch": 0.027122321670735014, + "grad_norm": 16.13709831237793, + "learning_rate": 5.405405405405406e-07, + "loss": 0.5452, + "num_input_tokens_seen": 209715200, + "step": 100 + }, + { + "epoch": 0.027393544887442365, + "grad_norm": 12.437922477722168, + "learning_rate": 5.459459459459459e-07, + "loss": 0.4683, + "num_input_tokens_seen": 211812352, + "step": 101 + }, + { + "epoch": 0.027664768104149716, + "grad_norm": 15.048489570617676, + "learning_rate": 5.513513513513513e-07, + "loss": 0.6694, + "num_input_tokens_seen": 213909504, + "step": 102 + }, + { + "epoch": 0.027935991320857067, + "grad_norm": 20.845455169677734, + "learning_rate": 5.567567567567567e-07, + "loss": 0.9275, + "num_input_tokens_seen": 216006656, + "step": 103 + }, + { + "epoch": 0.028207214537564414, + "grad_norm": 15.5338716506958, + "learning_rate": 5.621621621621622e-07, + "loss": 0.6088, + "num_input_tokens_seen": 218103808, + "step": 104 + }, + { + "epoch": 0.028478437754271765, + "grad_norm": 12.675715446472168, + "learning_rate": 5.675675675675675e-07, + "loss": 0.6012, + "num_input_tokens_seen": 220200960, + "step": 105 + }, + { + "epoch": 0.028749660970979116, + "grad_norm": 16.053550720214844, + "learning_rate": 5.729729729729729e-07, + "loss": 0.5751, + "num_input_tokens_seen": 222298112, + "step": 106 + }, + { + "epoch": 0.029020884187686467, + "grad_norm": 18.570253372192383, + "learning_rate": 5.783783783783784e-07, + "loss": 0.8626, + "num_input_tokens_seen": 224395264, + "step": 107 + }, + { + "epoch": 0.029292107404393815, + "grad_norm": 15.664776802062988, + "learning_rate": 5.837837837837838e-07, + "loss": 0.6016, + "num_input_tokens_seen": 226492416, + "step": 108 + }, + { + "epoch": 0.029563330621101166, + "grad_norm": 17.67211151123047, + "learning_rate": 5.891891891891891e-07, + "loss": 0.8534, + "num_input_tokens_seen": 228589568, + "step": 109 + }, + { + "epoch": 0.029834553837808517, + "grad_norm": 23.127309799194336, + "learning_rate": 5.945945945945947e-07, + "loss": 1.1914, + "num_input_tokens_seen": 230686720, + "step": 110 + }, + { + "epoch": 0.030105777054515868, + "grad_norm": 18.636947631835938, + "learning_rate": 6e-07, + "loss": 0.9078, + "num_input_tokens_seen": 232783872, + "step": 111 + }, + { + "epoch": 0.030377000271223215, + "grad_norm": 20.506677627563477, + "learning_rate": 6.054054054054054e-07, + "loss": 1.1244, + "num_input_tokens_seen": 234881024, + "step": 112 + }, + { + "epoch": 0.030648223487930566, + "grad_norm": 15.10786247253418, + "learning_rate": 6.108108108108107e-07, + "loss": 0.5326, + "num_input_tokens_seen": 236978176, + "step": 113 + }, + { + "epoch": 0.030919446704637917, + "grad_norm": 14.648728370666504, + "learning_rate": 6.162162162162163e-07, + "loss": 0.468, + "num_input_tokens_seen": 239075328, + "step": 114 + }, + { + "epoch": 0.031190669921345268, + "grad_norm": 16.831241607666016, + "learning_rate": 6.216216216216216e-07, + "loss": 0.6454, + "num_input_tokens_seen": 241172480, + "step": 115 + }, + { + "epoch": 0.031461893138052616, + "grad_norm": 13.11506175994873, + "learning_rate": 6.27027027027027e-07, + "loss": 0.4922, + "num_input_tokens_seen": 243269632, + "step": 116 + }, + { + "epoch": 0.03173311635475997, + "grad_norm": 18.388965606689453, + "learning_rate": 6.324324324324324e-07, + "loss": 0.6721, + "num_input_tokens_seen": 245366784, + "step": 117 + }, + { + "epoch": 0.03200433957146732, + "grad_norm": 19.308448791503906, + "learning_rate": 6.378378378378379e-07, + "loss": 0.8231, + "num_input_tokens_seen": 247463936, + "step": 118 + }, + { + "epoch": 0.032275562788174665, + "grad_norm": 19.583574295043945, + "learning_rate": 6.432432432432432e-07, + "loss": 1.0588, + "num_input_tokens_seen": 249561088, + "step": 119 + }, + { + "epoch": 0.03254678600488202, + "grad_norm": 15.67451000213623, + "learning_rate": 6.486486486486486e-07, + "loss": 0.5752, + "num_input_tokens_seen": 251658240, + "step": 120 + }, + { + "epoch": 0.03281800922158937, + "grad_norm": 18.46150779724121, + "learning_rate": 6.54054054054054e-07, + "loss": 0.8636, + "num_input_tokens_seen": 253755392, + "step": 121 + }, + { + "epoch": 0.03308923243829672, + "grad_norm": 18.50663948059082, + "learning_rate": 6.594594594594595e-07, + "loss": 0.9201, + "num_input_tokens_seen": 255852544, + "step": 122 + }, + { + "epoch": 0.03336045565500407, + "grad_norm": 18.124380111694336, + "learning_rate": 6.648648648648648e-07, + "loss": 0.7413, + "num_input_tokens_seen": 257949696, + "step": 123 + }, + { + "epoch": 0.033631678871711417, + "grad_norm": 20.973876953125, + "learning_rate": 6.702702702702703e-07, + "loss": 0.9747, + "num_input_tokens_seen": 260046848, + "step": 124 + }, + { + "epoch": 0.03390290208841877, + "grad_norm": 15.687307357788086, + "learning_rate": 6.756756756756756e-07, + "loss": 0.487, + "num_input_tokens_seen": 262144000, + "step": 125 + }, + { + "epoch": 0.03417412530512612, + "grad_norm": 17.15902328491211, + "learning_rate": 6.810810810810811e-07, + "loss": 0.8454, + "num_input_tokens_seen": 264241152, + "step": 126 + }, + { + "epoch": 0.034445348521833466, + "grad_norm": 14.142715454101562, + "learning_rate": 6.864864864864864e-07, + "loss": 0.5951, + "num_input_tokens_seen": 266338304, + "step": 127 + }, + { + "epoch": 0.03471657173854082, + "grad_norm": 20.538616180419922, + "learning_rate": 6.918918918918919e-07, + "loss": 0.9529, + "num_input_tokens_seen": 268435456, + "step": 128 + }, + { + "epoch": 0.03498779495524817, + "grad_norm": 22.79062843322754, + "learning_rate": 6.972972972972973e-07, + "loss": 1.0202, + "num_input_tokens_seen": 270532608, + "step": 129 + }, + { + "epoch": 0.03525901817195552, + "grad_norm": 16.5878963470459, + "learning_rate": 7.027027027027027e-07, + "loss": 0.7667, + "num_input_tokens_seen": 272629760, + "step": 130 + }, + { + "epoch": 0.03553024138866287, + "grad_norm": 16.780086517333984, + "learning_rate": 7.081081081081081e-07, + "loss": 0.6279, + "num_input_tokens_seen": 274726912, + "step": 131 + }, + { + "epoch": 0.03580146460537022, + "grad_norm": 23.6799259185791, + "learning_rate": 7.135135135135135e-07, + "loss": 1.2914, + "num_input_tokens_seen": 276824064, + "step": 132 + }, + { + "epoch": 0.03607268782207757, + "grad_norm": 17.561779022216797, + "learning_rate": 7.189189189189189e-07, + "loss": 0.5839, + "num_input_tokens_seen": 278921216, + "step": 133 + }, + { + "epoch": 0.03634391103878492, + "grad_norm": 21.706342697143555, + "learning_rate": 7.243243243243243e-07, + "loss": 0.7747, + "num_input_tokens_seen": 281018368, + "step": 134 + }, + { + "epoch": 0.03661513425549227, + "grad_norm": 25.93421173095703, + "learning_rate": 7.297297297297297e-07, + "loss": 1.2564, + "num_input_tokens_seen": 283115520, + "step": 135 + }, + { + "epoch": 0.03688635747219962, + "grad_norm": 26.61385154724121, + "learning_rate": 7.351351351351351e-07, + "loss": 0.4117, + "num_input_tokens_seen": 285212672, + "step": 136 + }, + { + "epoch": 0.03715758068890697, + "grad_norm": 19.071443557739258, + "learning_rate": 7.405405405405405e-07, + "loss": 0.6143, + "num_input_tokens_seen": 287309824, + "step": 137 + }, + { + "epoch": 0.03742880390561432, + "grad_norm": 15.275123596191406, + "learning_rate": 7.45945945945946e-07, + "loss": 0.6617, + "num_input_tokens_seen": 289406976, + "step": 138 + }, + { + "epoch": 0.03770002712232167, + "grad_norm": 13.6419095993042, + "learning_rate": 7.513513513513513e-07, + "loss": 0.4291, + "num_input_tokens_seen": 291504128, + "step": 139 + }, + { + "epoch": 0.03797125033902902, + "grad_norm": 14.96066951751709, + "learning_rate": 7.567567567567568e-07, + "loss": 0.5393, + "num_input_tokens_seen": 293601280, + "step": 140 + }, + { + "epoch": 0.03824247355573637, + "grad_norm": 12.42337703704834, + "learning_rate": 7.621621621621621e-07, + "loss": 0.3799, + "num_input_tokens_seen": 295698432, + "step": 141 + }, + { + "epoch": 0.03851369677244372, + "grad_norm": 13.157722473144531, + "learning_rate": 7.675675675675676e-07, + "loss": 0.3632, + "num_input_tokens_seen": 297795584, + "step": 142 + }, + { + "epoch": 0.038784919989151075, + "grad_norm": 16.27894401550293, + "learning_rate": 7.729729729729729e-07, + "loss": 0.7346, + "num_input_tokens_seen": 299892736, + "step": 143 + }, + { + "epoch": 0.03905614320585842, + "grad_norm": 20.3560848236084, + "learning_rate": 7.783783783783784e-07, + "loss": 0.7771, + "num_input_tokens_seen": 301989888, + "step": 144 + }, + { + "epoch": 0.03932736642256577, + "grad_norm": 21.640209197998047, + "learning_rate": 7.837837837837838e-07, + "loss": 1.1799, + "num_input_tokens_seen": 304087040, + "step": 145 + }, + { + "epoch": 0.039598589639273124, + "grad_norm": 14.982572555541992, + "learning_rate": 7.891891891891892e-07, + "loss": 0.5518, + "num_input_tokens_seen": 306184192, + "step": 146 + }, + { + "epoch": 0.03986981285598047, + "grad_norm": 18.2089900970459, + "learning_rate": 7.945945945945945e-07, + "loss": 0.624, + "num_input_tokens_seen": 308281344, + "step": 147 + }, + { + "epoch": 0.04014103607268782, + "grad_norm": 19.922109603881836, + "learning_rate": 8e-07, + "loss": 0.8285, + "num_input_tokens_seen": 310378496, + "step": 148 + }, + { + "epoch": 0.040412259289395173, + "grad_norm": 15.228759765625, + "learning_rate": 8.054054054054054e-07, + "loss": 0.5316, + "num_input_tokens_seen": 312475648, + "step": 149 + }, + { + "epoch": 0.04068348250610252, + "grad_norm": 17.04909324645996, + "learning_rate": 8.108108108108108e-07, + "loss": 0.6752, + "num_input_tokens_seen": 314572800, + "step": 150 + }, + { + "epoch": 0.040954705722809875, + "grad_norm": 18.965408325195312, + "learning_rate": 8.162162162162161e-07, + "loss": 0.8268, + "num_input_tokens_seen": 316669952, + "step": 151 + }, + { + "epoch": 0.04122592893951722, + "grad_norm": 18.962045669555664, + "learning_rate": 8.216216216216217e-07, + "loss": 0.747, + "num_input_tokens_seen": 318767104, + "step": 152 + }, + { + "epoch": 0.04149715215622457, + "grad_norm": 19.22246551513672, + "learning_rate": 8.27027027027027e-07, + "loss": 0.6027, + "num_input_tokens_seen": 320864256, + "step": 153 + }, + { + "epoch": 0.041768375372931925, + "grad_norm": 15.074151039123535, + "learning_rate": 8.324324324324324e-07, + "loss": 0.5226, + "num_input_tokens_seen": 322961408, + "step": 154 + }, + { + "epoch": 0.04203959858963927, + "grad_norm": 21.02098274230957, + "learning_rate": 8.378378378378377e-07, + "loss": 0.9997, + "num_input_tokens_seen": 325058560, + "step": 155 + }, + { + "epoch": 0.04231082180634662, + "grad_norm": 18.864036560058594, + "learning_rate": 8.432432432432433e-07, + "loss": 0.7729, + "num_input_tokens_seen": 327155712, + "step": 156 + }, + { + "epoch": 0.042582045023053974, + "grad_norm": 18.284622192382812, + "learning_rate": 8.486486486486486e-07, + "loss": 0.7608, + "num_input_tokens_seen": 329252864, + "step": 157 + }, + { + "epoch": 0.04285326823976132, + "grad_norm": 16.468151092529297, + "learning_rate": 8.54054054054054e-07, + "loss": 0.7438, + "num_input_tokens_seen": 331350016, + "step": 158 + }, + { + "epoch": 0.043124491456468676, + "grad_norm": 15.429784774780273, + "learning_rate": 8.594594594594595e-07, + "loss": 0.6437, + "num_input_tokens_seen": 333447168, + "step": 159 + }, + { + "epoch": 0.043395714673176024, + "grad_norm": 24.898374557495117, + "learning_rate": 8.648648648648649e-07, + "loss": 1.0021, + "num_input_tokens_seen": 335544320, + "step": 160 + }, + { + "epoch": 0.04366693788988337, + "grad_norm": 15.75402545928955, + "learning_rate": 8.702702702702702e-07, + "loss": 0.5328, + "num_input_tokens_seen": 337641472, + "step": 161 + }, + { + "epoch": 0.043938161106590726, + "grad_norm": 21.882734298706055, + "learning_rate": 8.756756756756756e-07, + "loss": 0.9156, + "num_input_tokens_seen": 339738624, + "step": 162 + }, + { + "epoch": 0.04420938432329807, + "grad_norm": 21.429712295532227, + "learning_rate": 8.810810810810811e-07, + "loss": 1.2029, + "num_input_tokens_seen": 341835776, + "step": 163 + }, + { + "epoch": 0.04448060754000543, + "grad_norm": 12.5687837600708, + "learning_rate": 8.864864864864865e-07, + "loss": 0.3537, + "num_input_tokens_seen": 343932928, + "step": 164 + }, + { + "epoch": 0.044751830756712775, + "grad_norm": 19.376272201538086, + "learning_rate": 8.918918918918918e-07, + "loss": 0.646, + "num_input_tokens_seen": 346030080, + "step": 165 + }, + { + "epoch": 0.04502305397342012, + "grad_norm": 17.1505184173584, + "learning_rate": 8.972972972972974e-07, + "loss": 0.6591, + "num_input_tokens_seen": 348127232, + "step": 166 + }, + { + "epoch": 0.04529427719012748, + "grad_norm": 15.413045883178711, + "learning_rate": 9.027027027027027e-07, + "loss": 0.4214, + "num_input_tokens_seen": 350224384, + "step": 167 + }, + { + "epoch": 0.045565500406834825, + "grad_norm": 17.270416259765625, + "learning_rate": 9.081081081081081e-07, + "loss": 0.6298, + "num_input_tokens_seen": 352321536, + "step": 168 + }, + { + "epoch": 0.04583672362354217, + "grad_norm": 14.51281452178955, + "learning_rate": 9.135135135135134e-07, + "loss": 0.5393, + "num_input_tokens_seen": 354418688, + "step": 169 + }, + { + "epoch": 0.04610794684024953, + "grad_norm": 18.907087326049805, + "learning_rate": 9.18918918918919e-07, + "loss": 0.6491, + "num_input_tokens_seen": 356515840, + "step": 170 + }, + { + "epoch": 0.046379170056956874, + "grad_norm": 17.48542594909668, + "learning_rate": 9.243243243243243e-07, + "loss": 0.5932, + "num_input_tokens_seen": 358612992, + "step": 171 + }, + { + "epoch": 0.04665039327366423, + "grad_norm": 15.82155704498291, + "learning_rate": 9.297297297297297e-07, + "loss": 0.5997, + "num_input_tokens_seen": 360710144, + "step": 172 + }, + { + "epoch": 0.046921616490371576, + "grad_norm": 54.783302307128906, + "learning_rate": 9.351351351351351e-07, + "loss": 0.5147, + "num_input_tokens_seen": 362807296, + "step": 173 + }, + { + "epoch": 0.047192839707078924, + "grad_norm": 15.447076797485352, + "learning_rate": 9.405405405405406e-07, + "loss": 0.5626, + "num_input_tokens_seen": 364904448, + "step": 174 + }, + { + "epoch": 0.04746406292378628, + "grad_norm": 21.660140991210938, + "learning_rate": 9.459459459459459e-07, + "loss": 0.5531, + "num_input_tokens_seen": 367001600, + "step": 175 + }, + { + "epoch": 0.047735286140493625, + "grad_norm": 19.583106994628906, + "learning_rate": 9.513513513513513e-07, + "loss": 0.7348, + "num_input_tokens_seen": 369098752, + "step": 176 + }, + { + "epoch": 0.04800650935720097, + "grad_norm": 19.693716049194336, + "learning_rate": 9.567567567567567e-07, + "loss": 0.7629, + "num_input_tokens_seen": 371195904, + "step": 177 + }, + { + "epoch": 0.04827773257390833, + "grad_norm": 21.72786521911621, + "learning_rate": 9.621621621621622e-07, + "loss": 0.7271, + "num_input_tokens_seen": 373293056, + "step": 178 + }, + { + "epoch": 0.048548955790615675, + "grad_norm": 20.64727210998535, + "learning_rate": 9.675675675675676e-07, + "loss": 0.944, + "num_input_tokens_seen": 375390208, + "step": 179 + }, + { + "epoch": 0.04882017900732303, + "grad_norm": 15.752840042114258, + "learning_rate": 9.72972972972973e-07, + "loss": 0.4082, + "num_input_tokens_seen": 377487360, + "step": 180 + }, + { + "epoch": 0.04909140222403038, + "grad_norm": 16.297456741333008, + "learning_rate": 9.783783783783782e-07, + "loss": 0.5716, + "num_input_tokens_seen": 379584512, + "step": 181 + }, + { + "epoch": 0.049362625440737724, + "grad_norm": 16.411787033081055, + "learning_rate": 9.837837837837839e-07, + "loss": 0.3719, + "num_input_tokens_seen": 381681664, + "step": 182 + }, + { + "epoch": 0.04963384865744508, + "grad_norm": 17.700828552246094, + "learning_rate": 9.89189189189189e-07, + "loss": 0.6982, + "num_input_tokens_seen": 383778816, + "step": 183 + }, + { + "epoch": 0.049905071874152426, + "grad_norm": 13.723708152770996, + "learning_rate": 9.945945945945945e-07, + "loss": 0.4343, + "num_input_tokens_seen": 385875968, + "step": 184 + }, + { + "epoch": 0.05017629509085978, + "grad_norm": 18.31014633178711, + "learning_rate": 1e-06, + "loss": 0.511, + "num_input_tokens_seen": 387973120, + "step": 185 + }, + { + "epoch": 0.05044751830756713, + "grad_norm": 16.061124801635742, + "learning_rate": 9.99999818928562e-07, + "loss": 0.5106, + "num_input_tokens_seen": 390070272, + "step": 186 + }, + { + "epoch": 0.050718741524274476, + "grad_norm": 17.339305877685547, + "learning_rate": 9.999992757143933e-07, + "loss": 0.5723, + "num_input_tokens_seen": 392167424, + "step": 187 + }, + { + "epoch": 0.05098996474098183, + "grad_norm": 15.1113920211792, + "learning_rate": 9.999983703579313e-07, + "loss": 0.5508, + "num_input_tokens_seen": 394264576, + "step": 188 + }, + { + "epoch": 0.05126118795768918, + "grad_norm": 18.564332962036133, + "learning_rate": 9.999971028599045e-07, + "loss": 0.7198, + "num_input_tokens_seen": 396361728, + "step": 189 + }, + { + "epoch": 0.051532411174396525, + "grad_norm": 14.539531707763672, + "learning_rate": 9.99995473221333e-07, + "loss": 0.4645, + "num_input_tokens_seen": 398458880, + "step": 190 + }, + { + "epoch": 0.05180363439110388, + "grad_norm": 16.04762077331543, + "learning_rate": 9.999934814435284e-07, + "loss": 0.6121, + "num_input_tokens_seen": 400556032, + "step": 191 + }, + { + "epoch": 0.05207485760781123, + "grad_norm": 17.753110885620117, + "learning_rate": 9.999911275280933e-07, + "loss": 0.6252, + "num_input_tokens_seen": 402653184, + "step": 192 + }, + { + "epoch": 0.05234608082451858, + "grad_norm": 23.137664794921875, + "learning_rate": 9.999884114769223e-07, + "loss": 1.1105, + "num_input_tokens_seen": 404750336, + "step": 193 + }, + { + "epoch": 0.05261730404122593, + "grad_norm": 17.16234588623047, + "learning_rate": 9.99985333292201e-07, + "loss": 0.693, + "num_input_tokens_seen": 406847488, + "step": 194 + }, + { + "epoch": 0.05288852725793328, + "grad_norm": 25.552888870239258, + "learning_rate": 9.999818929764068e-07, + "loss": 1.1388, + "num_input_tokens_seen": 408944640, + "step": 195 + }, + { + "epoch": 0.05315975047464063, + "grad_norm": 17.101301193237305, + "learning_rate": 9.99978090532308e-07, + "loss": 0.5531, + "num_input_tokens_seen": 411041792, + "step": 196 + }, + { + "epoch": 0.05343097369134798, + "grad_norm": 16.60688591003418, + "learning_rate": 9.99973925962965e-07, + "loss": 0.7798, + "num_input_tokens_seen": 413138944, + "step": 197 + }, + { + "epoch": 0.05370219690805533, + "grad_norm": 12.576007843017578, + "learning_rate": 9.999693992717292e-07, + "loss": 0.4118, + "num_input_tokens_seen": 415236096, + "step": 198 + }, + { + "epoch": 0.05397342012476268, + "grad_norm": 12.9966402053833, + "learning_rate": 9.999645104622434e-07, + "loss": 0.4088, + "num_input_tokens_seen": 417333248, + "step": 199 + }, + { + "epoch": 0.05424464334147003, + "grad_norm": 16.39759635925293, + "learning_rate": 9.99959259538442e-07, + "loss": 0.5526, + "num_input_tokens_seen": 419430400, + "step": 200 + }, + { + "epoch": 0.05451586655817738, + "grad_norm": 14.523542404174805, + "learning_rate": 9.99953646504551e-07, + "loss": 0.4645, + "num_input_tokens_seen": 421527552, + "step": 201 + }, + { + "epoch": 0.05478708977488473, + "grad_norm": 17.604217529296875, + "learning_rate": 9.99947671365087e-07, + "loss": 0.794, + "num_input_tokens_seen": 423624704, + "step": 202 + }, + { + "epoch": 0.05505831299159208, + "grad_norm": 16.45903778076172, + "learning_rate": 9.99941334124859e-07, + "loss": 0.6843, + "num_input_tokens_seen": 425721856, + "step": 203 + }, + { + "epoch": 0.05532953620829943, + "grad_norm": 17.023611068725586, + "learning_rate": 9.999346347889667e-07, + "loss": 0.6436, + "num_input_tokens_seen": 427819008, + "step": 204 + }, + { + "epoch": 0.05560075942500678, + "grad_norm": 17.66231346130371, + "learning_rate": 9.999275733628017e-07, + "loss": 0.707, + "num_input_tokens_seen": 429916160, + "step": 205 + }, + { + "epoch": 0.055871982641714134, + "grad_norm": 14.247184753417969, + "learning_rate": 9.999201498520466e-07, + "loss": 0.5772, + "num_input_tokens_seen": 432013312, + "step": 206 + }, + { + "epoch": 0.05614320585842148, + "grad_norm": 16.290454864501953, + "learning_rate": 9.999123642626758e-07, + "loss": 0.5432, + "num_input_tokens_seen": 434110464, + "step": 207 + }, + { + "epoch": 0.05641442907512883, + "grad_norm": 16.772056579589844, + "learning_rate": 9.999042166009544e-07, + "loss": 0.5976, + "num_input_tokens_seen": 436207616, + "step": 208 + }, + { + "epoch": 0.05668565229183618, + "grad_norm": 10.778910636901855, + "learning_rate": 9.998957068734399e-07, + "loss": 0.3159, + "num_input_tokens_seen": 438304768, + "step": 209 + }, + { + "epoch": 0.05695687550854353, + "grad_norm": 21.58819580078125, + "learning_rate": 9.9988683508698e-07, + "loss": 0.9222, + "num_input_tokens_seen": 440401920, + "step": 210 + }, + { + "epoch": 0.05722809872525088, + "grad_norm": 21.694862365722656, + "learning_rate": 9.99877601248715e-07, + "loss": 1.0365, + "num_input_tokens_seen": 442499072, + "step": 211 + }, + { + "epoch": 0.05749932194195823, + "grad_norm": 20.887956619262695, + "learning_rate": 9.998680053660756e-07, + "loss": 0.5279, + "num_input_tokens_seen": 444596224, + "step": 212 + }, + { + "epoch": 0.05777054515866558, + "grad_norm": 20.843788146972656, + "learning_rate": 9.998580474467842e-07, + "loss": 0.7565, + "num_input_tokens_seen": 446693376, + "step": 213 + }, + { + "epoch": 0.058041768375372935, + "grad_norm": 13.617545127868652, + "learning_rate": 9.998477274988545e-07, + "loss": 0.4137, + "num_input_tokens_seen": 448790528, + "step": 214 + }, + { + "epoch": 0.05831299159208028, + "grad_norm": 11.583077430725098, + "learning_rate": 9.998370455305918e-07, + "loss": 0.4458, + "num_input_tokens_seen": 450887680, + "step": 215 + }, + { + "epoch": 0.05858421480878763, + "grad_norm": 16.68201446533203, + "learning_rate": 9.998260015505923e-07, + "loss": 0.6287, + "num_input_tokens_seen": 452984832, + "step": 216 + }, + { + "epoch": 0.058855438025494984, + "grad_norm": 13.695748329162598, + "learning_rate": 9.998145955677438e-07, + "loss": 0.4641, + "num_input_tokens_seen": 455081984, + "step": 217 + }, + { + "epoch": 0.05912666124220233, + "grad_norm": 12.062211990356445, + "learning_rate": 9.998028275912257e-07, + "loss": 0.3928, + "num_input_tokens_seen": 457179136, + "step": 218 + }, + { + "epoch": 0.059397884458909686, + "grad_norm": 19.77566146850586, + "learning_rate": 9.997906976305082e-07, + "loss": 0.5814, + "num_input_tokens_seen": 459276288, + "step": 219 + }, + { + "epoch": 0.059669107675617034, + "grad_norm": 19.376760482788086, + "learning_rate": 9.99778205695353e-07, + "loss": 0.8212, + "num_input_tokens_seen": 461373440, + "step": 220 + }, + { + "epoch": 0.05994033089232438, + "grad_norm": 13.375431060791016, + "learning_rate": 9.997653517958132e-07, + "loss": 0.4862, + "num_input_tokens_seen": 463470592, + "step": 221 + }, + { + "epoch": 0.060211554109031735, + "grad_norm": 16.914608001708984, + "learning_rate": 9.997521359422332e-07, + "loss": 0.6401, + "num_input_tokens_seen": 465567744, + "step": 222 + }, + { + "epoch": 0.06048277732573908, + "grad_norm": 13.095565795898438, + "learning_rate": 9.997385581452484e-07, + "loss": 0.4319, + "num_input_tokens_seen": 467664896, + "step": 223 + }, + { + "epoch": 0.06075400054244643, + "grad_norm": 15.450024604797363, + "learning_rate": 9.99724618415786e-07, + "loss": 0.3974, + "num_input_tokens_seen": 469762048, + "step": 224 + }, + { + "epoch": 0.061025223759153785, + "grad_norm": 15.416088104248047, + "learning_rate": 9.997103167650637e-07, + "loss": 0.7345, + "num_input_tokens_seen": 471859200, + "step": 225 + }, + { + "epoch": 0.06129644697586113, + "grad_norm": 12.959712982177734, + "learning_rate": 9.996956532045914e-07, + "loss": 0.4602, + "num_input_tokens_seen": 473956352, + "step": 226 + }, + { + "epoch": 0.06156767019256849, + "grad_norm": 24.262161254882812, + "learning_rate": 9.996806277461696e-07, + "loss": 0.9696, + "num_input_tokens_seen": 476053504, + "step": 227 + }, + { + "epoch": 0.061838893409275834, + "grad_norm": 21.882034301757812, + "learning_rate": 9.9966524040189e-07, + "loss": 0.7863, + "num_input_tokens_seen": 478150656, + "step": 228 + }, + { + "epoch": 0.06211011662598318, + "grad_norm": 17.270530700683594, + "learning_rate": 9.996494911841363e-07, + "loss": 0.6471, + "num_input_tokens_seen": 480247808, + "step": 229 + }, + { + "epoch": 0.062381339842690536, + "grad_norm": 13.247891426086426, + "learning_rate": 9.996333801055823e-07, + "loss": 0.3576, + "num_input_tokens_seen": 482344960, + "step": 230 + }, + { + "epoch": 0.06265256305939788, + "grad_norm": 13.374452590942383, + "learning_rate": 9.99616907179194e-07, + "loss": 0.4596, + "num_input_tokens_seen": 484442112, + "step": 231 + }, + { + "epoch": 0.06292378627610523, + "grad_norm": 18.631214141845703, + "learning_rate": 9.996000724182278e-07, + "loss": 0.7839, + "num_input_tokens_seen": 486539264, + "step": 232 + }, + { + "epoch": 0.06319500949281258, + "grad_norm": 18.132368087768555, + "learning_rate": 9.99582875836232e-07, + "loss": 0.636, + "num_input_tokens_seen": 488636416, + "step": 233 + }, + { + "epoch": 0.06346623270951994, + "grad_norm": 15.695167541503906, + "learning_rate": 9.995653174470456e-07, + "loss": 0.6983, + "num_input_tokens_seen": 490733568, + "step": 234 + }, + { + "epoch": 0.06373745592622729, + "grad_norm": 13.286776542663574, + "learning_rate": 9.99547397264799e-07, + "loss": 0.4567, + "num_input_tokens_seen": 492830720, + "step": 235 + }, + { + "epoch": 0.06400867914293464, + "grad_norm": 11.679216384887695, + "learning_rate": 9.995291153039135e-07, + "loss": 0.359, + "num_input_tokens_seen": 494927872, + "step": 236 + }, + { + "epoch": 0.06427990235964198, + "grad_norm": 16.616655349731445, + "learning_rate": 9.99510471579102e-07, + "loss": 0.6919, + "num_input_tokens_seen": 497025024, + "step": 237 + }, + { + "epoch": 0.06455112557634933, + "grad_norm": 15.427544593811035, + "learning_rate": 9.99491466105368e-07, + "loss": 0.5059, + "num_input_tokens_seen": 499122176, + "step": 238 + }, + { + "epoch": 0.06482234879305669, + "grad_norm": 38.47232437133789, + "learning_rate": 9.994720988980065e-07, + "loss": 0.6509, + "num_input_tokens_seen": 501219328, + "step": 239 + }, + { + "epoch": 0.06509357200976404, + "grad_norm": 12.464300155639648, + "learning_rate": 9.994523699726035e-07, + "loss": 0.3664, + "num_input_tokens_seen": 503316480, + "step": 240 + }, + { + "epoch": 0.06536479522647139, + "grad_norm": 25.51685905456543, + "learning_rate": 9.994322793450361e-07, + "loss": 0.9145, + "num_input_tokens_seen": 505413632, + "step": 241 + }, + { + "epoch": 0.06563601844317873, + "grad_norm": 18.906505584716797, + "learning_rate": 9.994118270314725e-07, + "loss": 0.7751, + "num_input_tokens_seen": 507510784, + "step": 242 + }, + { + "epoch": 0.06590724165988608, + "grad_norm": 22.564781188964844, + "learning_rate": 9.993910130483717e-07, + "loss": 0.8191, + "num_input_tokens_seen": 509607936, + "step": 243 + }, + { + "epoch": 0.06617846487659344, + "grad_norm": 12.493294715881348, + "learning_rate": 9.993698374124844e-07, + "loss": 0.4122, + "num_input_tokens_seen": 511705088, + "step": 244 + }, + { + "epoch": 0.06644968809330079, + "grad_norm": 16.386438369750977, + "learning_rate": 9.993483001408516e-07, + "loss": 0.6446, + "num_input_tokens_seen": 513802240, + "step": 245 + }, + { + "epoch": 0.06672091131000814, + "grad_norm": 20.80731773376465, + "learning_rate": 9.99326401250806e-07, + "loss": 0.7367, + "num_input_tokens_seen": 515899392, + "step": 246 + }, + { + "epoch": 0.06699213452671549, + "grad_norm": 21.987316131591797, + "learning_rate": 9.993041407599708e-07, + "loss": 0.5971, + "num_input_tokens_seen": 517996544, + "step": 247 + }, + { + "epoch": 0.06726335774342283, + "grad_norm": 14.191584587097168, + "learning_rate": 9.992815186862602e-07, + "loss": 0.468, + "num_input_tokens_seen": 520093696, + "step": 248 + }, + { + "epoch": 0.06753458096013018, + "grad_norm": 19.399492263793945, + "learning_rate": 9.9925853504788e-07, + "loss": 0.7622, + "num_input_tokens_seen": 522190848, + "step": 249 + }, + { + "epoch": 0.06780580417683754, + "grad_norm": 15.56145191192627, + "learning_rate": 9.992351898633262e-07, + "loss": 0.6956, + "num_input_tokens_seen": 524288000, + "step": 250 + }, + { + "epoch": 0.06807702739354489, + "grad_norm": 10.482128143310547, + "learning_rate": 9.992114831513863e-07, + "loss": 0.2901, + "num_input_tokens_seen": 526385152, + "step": 251 + }, + { + "epoch": 0.06834825061025224, + "grad_norm": 14.247305870056152, + "learning_rate": 9.991874149311386e-07, + "loss": 0.5265, + "num_input_tokens_seen": 528482304, + "step": 252 + }, + { + "epoch": 0.06861947382695958, + "grad_norm": 23.847684860229492, + "learning_rate": 9.991629852219523e-07, + "loss": 1.0187, + "num_input_tokens_seen": 530579456, + "step": 253 + }, + { + "epoch": 0.06889069704366693, + "grad_norm": 19.322067260742188, + "learning_rate": 9.991381940434873e-07, + "loss": 0.7402, + "num_input_tokens_seen": 532676608, + "step": 254 + }, + { + "epoch": 0.0691619202603743, + "grad_norm": 19.398937225341797, + "learning_rate": 9.991130414156946e-07, + "loss": 0.8914, + "num_input_tokens_seen": 534773760, + "step": 255 + }, + { + "epoch": 0.06943314347708164, + "grad_norm": 14.539494514465332, + "learning_rate": 9.990875273588161e-07, + "loss": 0.6505, + "num_input_tokens_seen": 536870912, + "step": 256 + }, + { + "epoch": 0.06970436669378899, + "grad_norm": 15.979605674743652, + "learning_rate": 9.99061651893385e-07, + "loss": 0.6311, + "num_input_tokens_seen": 538968064, + "step": 257 + }, + { + "epoch": 0.06997558991049634, + "grad_norm": 13.959792137145996, + "learning_rate": 9.990354150402242e-07, + "loss": 0.4779, + "num_input_tokens_seen": 541065216, + "step": 258 + }, + { + "epoch": 0.07024681312720368, + "grad_norm": 14.552385330200195, + "learning_rate": 9.990088168204487e-07, + "loss": 0.4992, + "num_input_tokens_seen": 543162368, + "step": 259 + }, + { + "epoch": 0.07051803634391104, + "grad_norm": 18.06981086730957, + "learning_rate": 9.989818572554633e-07, + "loss": 0.8364, + "num_input_tokens_seen": 545259520, + "step": 260 + }, + { + "epoch": 0.07078925956061839, + "grad_norm": 16.378019332885742, + "learning_rate": 9.989545363669644e-07, + "loss": 0.7462, + "num_input_tokens_seen": 547356672, + "step": 261 + }, + { + "epoch": 0.07106048277732574, + "grad_norm": 12.575121879577637, + "learning_rate": 9.989268541769383e-07, + "loss": 0.4064, + "num_input_tokens_seen": 549453824, + "step": 262 + }, + { + "epoch": 0.07133170599403309, + "grad_norm": 14.232921600341797, + "learning_rate": 9.988988107076632e-07, + "loss": 0.6201, + "num_input_tokens_seen": 551550976, + "step": 263 + }, + { + "epoch": 0.07160292921074043, + "grad_norm": 14.675073623657227, + "learning_rate": 9.98870405981707e-07, + "loss": 0.4924, + "num_input_tokens_seen": 553648128, + "step": 264 + }, + { + "epoch": 0.0718741524274478, + "grad_norm": 15.136611938476562, + "learning_rate": 9.988416400219288e-07, + "loss": 0.6525, + "num_input_tokens_seen": 555745280, + "step": 265 + }, + { + "epoch": 0.07214537564415514, + "grad_norm": 18.230609893798828, + "learning_rate": 9.988125128514785e-07, + "loss": 0.8809, + "num_input_tokens_seen": 557842432, + "step": 266 + }, + { + "epoch": 0.07241659886086249, + "grad_norm": 15.418828964233398, + "learning_rate": 9.987830244937964e-07, + "loss": 0.8042, + "num_input_tokens_seen": 559939584, + "step": 267 + }, + { + "epoch": 0.07268782207756984, + "grad_norm": 20.395606994628906, + "learning_rate": 9.987531749726137e-07, + "loss": 0.5648, + "num_input_tokens_seen": 562036736, + "step": 268 + }, + { + "epoch": 0.07295904529427719, + "grad_norm": 16.41619873046875, + "learning_rate": 9.98722964311952e-07, + "loss": 0.6025, + "num_input_tokens_seen": 564133888, + "step": 269 + }, + { + "epoch": 0.07323026851098453, + "grad_norm": 15.053420066833496, + "learning_rate": 9.986923925361238e-07, + "loss": 0.6662, + "num_input_tokens_seen": 566231040, + "step": 270 + }, + { + "epoch": 0.0735014917276919, + "grad_norm": 19.180273056030273, + "learning_rate": 9.98661459669732e-07, + "loss": 0.7407, + "num_input_tokens_seen": 568328192, + "step": 271 + }, + { + "epoch": 0.07377271494439924, + "grad_norm": 19.252796173095703, + "learning_rate": 9.986301657376705e-07, + "loss": 0.7506, + "num_input_tokens_seen": 570425344, + "step": 272 + }, + { + "epoch": 0.07404393816110659, + "grad_norm": 12.588926315307617, + "learning_rate": 9.985985107651231e-07, + "loss": 0.418, + "num_input_tokens_seen": 572522496, + "step": 273 + }, + { + "epoch": 0.07431516137781394, + "grad_norm": 19.830280303955078, + "learning_rate": 9.985664947775649e-07, + "loss": 0.9066, + "num_input_tokens_seen": 574619648, + "step": 274 + }, + { + "epoch": 0.07458638459452128, + "grad_norm": 12.818648338317871, + "learning_rate": 9.985341178007608e-07, + "loss": 0.5215, + "num_input_tokens_seen": 576716800, + "step": 275 + }, + { + "epoch": 0.07485760781122865, + "grad_norm": 18.200252532958984, + "learning_rate": 9.985013798607666e-07, + "loss": 0.8381, + "num_input_tokens_seen": 578813952, + "step": 276 + }, + { + "epoch": 0.075128831027936, + "grad_norm": 17.148433685302734, + "learning_rate": 9.98468280983929e-07, + "loss": 0.6161, + "num_input_tokens_seen": 580911104, + "step": 277 + }, + { + "epoch": 0.07540005424464334, + "grad_norm": 22.672407150268555, + "learning_rate": 9.984348211968837e-07, + "loss": 0.8064, + "num_input_tokens_seen": 583008256, + "step": 278 + }, + { + "epoch": 0.07567127746135069, + "grad_norm": 17.892045974731445, + "learning_rate": 9.984010005265592e-07, + "loss": 0.6633, + "num_input_tokens_seen": 585105408, + "step": 279 + }, + { + "epoch": 0.07594250067805804, + "grad_norm": 16.25893211364746, + "learning_rate": 9.98366819000172e-07, + "loss": 0.6054, + "num_input_tokens_seen": 587202560, + "step": 280 + }, + { + "epoch": 0.0762137238947654, + "grad_norm": 13.858470916748047, + "learning_rate": 9.983322766452305e-07, + "loss": 0.4507, + "num_input_tokens_seen": 589299712, + "step": 281 + }, + { + "epoch": 0.07648494711147275, + "grad_norm": 12.110980033874512, + "learning_rate": 9.98297373489533e-07, + "loss": 0.3055, + "num_input_tokens_seen": 591396864, + "step": 282 + }, + { + "epoch": 0.07675617032818009, + "grad_norm": 18.724361419677734, + "learning_rate": 9.982621095611686e-07, + "loss": 0.6864, + "num_input_tokens_seen": 593494016, + "step": 283 + }, + { + "epoch": 0.07702739354488744, + "grad_norm": 13.076481819152832, + "learning_rate": 9.98226484888516e-07, + "loss": 0.4906, + "num_input_tokens_seen": 595591168, + "step": 284 + }, + { + "epoch": 0.07729861676159479, + "grad_norm": 22.247615814208984, + "learning_rate": 9.981904995002443e-07, + "loss": 0.447, + "num_input_tokens_seen": 597688320, + "step": 285 + }, + { + "epoch": 0.07756983997830215, + "grad_norm": 15.656414031982422, + "learning_rate": 9.98154153425314e-07, + "loss": 0.5142, + "num_input_tokens_seen": 599785472, + "step": 286 + }, + { + "epoch": 0.0778410631950095, + "grad_norm": 16.987443923950195, + "learning_rate": 9.981174466929742e-07, + "loss": 0.6084, + "num_input_tokens_seen": 601882624, + "step": 287 + }, + { + "epoch": 0.07811228641171684, + "grad_norm": 15.571359634399414, + "learning_rate": 9.980803793327655e-07, + "loss": 0.5127, + "num_input_tokens_seen": 603979776, + "step": 288 + }, + { + "epoch": 0.07838350962842419, + "grad_norm": 17.951066970825195, + "learning_rate": 9.980429513745182e-07, + "loss": 0.8033, + "num_input_tokens_seen": 606076928, + "step": 289 + }, + { + "epoch": 0.07865473284513154, + "grad_norm": 24.38446617126465, + "learning_rate": 9.980051628483532e-07, + "loss": 1.0212, + "num_input_tokens_seen": 608174080, + "step": 290 + }, + { + "epoch": 0.07892595606183889, + "grad_norm": 14.629419326782227, + "learning_rate": 9.979670137846806e-07, + "loss": 0.6336, + "num_input_tokens_seen": 610271232, + "step": 291 + }, + { + "epoch": 0.07919717927854625, + "grad_norm": 23.52570343017578, + "learning_rate": 9.97928504214202e-07, + "loss": 1.1455, + "num_input_tokens_seen": 612368384, + "step": 292 + }, + { + "epoch": 0.0794684024952536, + "grad_norm": 20.963306427001953, + "learning_rate": 9.97889634167908e-07, + "loss": 0.9892, + "num_input_tokens_seen": 614465536, + "step": 293 + }, + { + "epoch": 0.07973962571196094, + "grad_norm": 16.687395095825195, + "learning_rate": 9.978504036770802e-07, + "loss": 0.3016, + "num_input_tokens_seen": 616562688, + "step": 294 + }, + { + "epoch": 0.08001084892866829, + "grad_norm": 11.436803817749023, + "learning_rate": 9.978108127732892e-07, + "loss": 0.378, + "num_input_tokens_seen": 618659840, + "step": 295 + }, + { + "epoch": 0.08028207214537564, + "grad_norm": 16.03774642944336, + "learning_rate": 9.977708614883965e-07, + "loss": 0.5094, + "num_input_tokens_seen": 620756992, + "step": 296 + }, + { + "epoch": 0.080553295362083, + "grad_norm": 25.998807907104492, + "learning_rate": 9.977305498545537e-07, + "loss": 0.7122, + "num_input_tokens_seen": 622854144, + "step": 297 + }, + { + "epoch": 0.08082451857879035, + "grad_norm": 12.087435722351074, + "learning_rate": 9.976898779042018e-07, + "loss": 0.455, + "num_input_tokens_seen": 624951296, + "step": 298 + }, + { + "epoch": 0.0810957417954977, + "grad_norm": 19.255043029785156, + "learning_rate": 9.976488456700717e-07, + "loss": 0.7966, + "num_input_tokens_seen": 627048448, + "step": 299 + }, + { + "epoch": 0.08136696501220504, + "grad_norm": 15.657422065734863, + "learning_rate": 9.97607453185185e-07, + "loss": 0.61, + "num_input_tokens_seen": 629145600, + "step": 300 + }, + { + "epoch": 0.08163818822891239, + "grad_norm": 17.275453567504883, + "learning_rate": 9.97565700482853e-07, + "loss": 0.7554, + "num_input_tokens_seen": 631242752, + "step": 301 + }, + { + "epoch": 0.08190941144561975, + "grad_norm": 16.451929092407227, + "learning_rate": 9.97523587596676e-07, + "loss": 0.5496, + "num_input_tokens_seen": 633339904, + "step": 302 + }, + { + "epoch": 0.0821806346623271, + "grad_norm": 11.221899032592773, + "learning_rate": 9.974811145605453e-07, + "loss": 0.2887, + "num_input_tokens_seen": 635437056, + "step": 303 + }, + { + "epoch": 0.08245185787903445, + "grad_norm": 15.957961082458496, + "learning_rate": 9.974382814086418e-07, + "loss": 0.3712, + "num_input_tokens_seen": 637534208, + "step": 304 + }, + { + "epoch": 0.0827230810957418, + "grad_norm": 14.90071964263916, + "learning_rate": 9.973950881754353e-07, + "loss": 0.4835, + "num_input_tokens_seen": 639631360, + "step": 305 + }, + { + "epoch": 0.08299430431244914, + "grad_norm": 17.702465057373047, + "learning_rate": 9.973515348956869e-07, + "loss": 0.6427, + "num_input_tokens_seen": 641728512, + "step": 306 + }, + { + "epoch": 0.0832655275291565, + "grad_norm": 13.002301216125488, + "learning_rate": 9.97307621604446e-07, + "loss": 0.5525, + "num_input_tokens_seen": 643825664, + "step": 307 + }, + { + "epoch": 0.08353675074586385, + "grad_norm": 12.444464683532715, + "learning_rate": 9.972633483370526e-07, + "loss": 0.4219, + "num_input_tokens_seen": 645922816, + "step": 308 + }, + { + "epoch": 0.0838079739625712, + "grad_norm": 34.6700439453125, + "learning_rate": 9.97218715129136e-07, + "loss": 0.8305, + "num_input_tokens_seen": 648019968, + "step": 309 + }, + { + "epoch": 0.08407919717927854, + "grad_norm": 13.552191734313965, + "learning_rate": 9.971737220166155e-07, + "loss": 0.4636, + "num_input_tokens_seen": 650117120, + "step": 310 + }, + { + "epoch": 0.08435042039598589, + "grad_norm": 15.92861270904541, + "learning_rate": 9.971283690356997e-07, + "loss": 0.68, + "num_input_tokens_seen": 652214272, + "step": 311 + }, + { + "epoch": 0.08462164361269324, + "grad_norm": 10.955565452575684, + "learning_rate": 9.97082656222887e-07, + "loss": 0.292, + "num_input_tokens_seen": 654311424, + "step": 312 + }, + { + "epoch": 0.0848928668294006, + "grad_norm": 16.476566314697266, + "learning_rate": 9.970365836149654e-07, + "loss": 0.6708, + "num_input_tokens_seen": 656408576, + "step": 313 + }, + { + "epoch": 0.08516409004610795, + "grad_norm": 21.01806640625, + "learning_rate": 9.969901512490121e-07, + "loss": 1.0652, + "num_input_tokens_seen": 658505728, + "step": 314 + }, + { + "epoch": 0.0854353132628153, + "grad_norm": 16.38351058959961, + "learning_rate": 9.969433591623946e-07, + "loss": 0.6264, + "num_input_tokens_seen": 660602880, + "step": 315 + }, + { + "epoch": 0.08570653647952264, + "grad_norm": 23.092439651489258, + "learning_rate": 9.96896207392769e-07, + "loss": 1.0512, + "num_input_tokens_seen": 662700032, + "step": 316 + }, + { + "epoch": 0.08597775969622999, + "grad_norm": 14.13205337524414, + "learning_rate": 9.968486959780813e-07, + "loss": 0.5465, + "num_input_tokens_seen": 664797184, + "step": 317 + }, + { + "epoch": 0.08624898291293735, + "grad_norm": 16.44294548034668, + "learning_rate": 9.96800824956567e-07, + "loss": 0.6146, + "num_input_tokens_seen": 666894336, + "step": 318 + }, + { + "epoch": 0.0865202061296447, + "grad_norm": 16.18340301513672, + "learning_rate": 9.967525943667506e-07, + "loss": 0.6067, + "num_input_tokens_seen": 668991488, + "step": 319 + }, + { + "epoch": 0.08679142934635205, + "grad_norm": 14.490288734436035, + "learning_rate": 9.967040042474467e-07, + "loss": 0.5758, + "num_input_tokens_seen": 671088640, + "step": 320 + }, + { + "epoch": 0.0870626525630594, + "grad_norm": 20.6240234375, + "learning_rate": 9.966550546377586e-07, + "loss": 0.8391, + "num_input_tokens_seen": 673185792, + "step": 321 + }, + { + "epoch": 0.08733387577976674, + "grad_norm": 12.437811851501465, + "learning_rate": 9.966057455770788e-07, + "loss": 0.3815, + "num_input_tokens_seen": 675282944, + "step": 322 + }, + { + "epoch": 0.0876050989964741, + "grad_norm": 17.535430908203125, + "learning_rate": 9.965560771050896e-07, + "loss": 0.7666, + "num_input_tokens_seen": 677380096, + "step": 323 + }, + { + "epoch": 0.08787632221318145, + "grad_norm": 17.357746124267578, + "learning_rate": 9.965060492617623e-07, + "loss": 0.6267, + "num_input_tokens_seen": 679477248, + "step": 324 + }, + { + "epoch": 0.0881475454298888, + "grad_norm": 19.566974639892578, + "learning_rate": 9.964556620873573e-07, + "loss": 0.6236, + "num_input_tokens_seen": 681574400, + "step": 325 + }, + { + "epoch": 0.08841876864659615, + "grad_norm": 14.313620567321777, + "learning_rate": 9.964049156224244e-07, + "loss": 0.5585, + "num_input_tokens_seen": 683671552, + "step": 326 + }, + { + "epoch": 0.0886899918633035, + "grad_norm": 13.445440292358398, + "learning_rate": 9.963538099078024e-07, + "loss": 0.4307, + "num_input_tokens_seen": 685768704, + "step": 327 + }, + { + "epoch": 0.08896121508001086, + "grad_norm": 17.229440689086914, + "learning_rate": 9.963023449846194e-07, + "loss": 0.7462, + "num_input_tokens_seen": 687865856, + "step": 328 + }, + { + "epoch": 0.0892324382967182, + "grad_norm": 20.201749801635742, + "learning_rate": 9.962505208942919e-07, + "loss": 0.9177, + "num_input_tokens_seen": 689963008, + "step": 329 + }, + { + "epoch": 0.08950366151342555, + "grad_norm": 13.015423774719238, + "learning_rate": 9.961983376785264e-07, + "loss": 0.4933, + "num_input_tokens_seen": 692060160, + "step": 330 + }, + { + "epoch": 0.0897748847301329, + "grad_norm": 11.634929656982422, + "learning_rate": 9.96145795379318e-07, + "loss": 0.3818, + "num_input_tokens_seen": 694157312, + "step": 331 + }, + { + "epoch": 0.09004610794684025, + "grad_norm": 18.479869842529297, + "learning_rate": 9.960928940389503e-07, + "loss": 0.8161, + "num_input_tokens_seen": 696254464, + "step": 332 + }, + { + "epoch": 0.09031733116354759, + "grad_norm": 12.271891593933105, + "learning_rate": 9.960396336999967e-07, + "loss": 0.451, + "num_input_tokens_seen": 698351616, + "step": 333 + }, + { + "epoch": 0.09058855438025495, + "grad_norm": 10.953863143920898, + "learning_rate": 9.95986014405319e-07, + "loss": 0.3789, + "num_input_tokens_seen": 700448768, + "step": 334 + }, + { + "epoch": 0.0908597775969623, + "grad_norm": 9.575960159301758, + "learning_rate": 9.959320361980679e-07, + "loss": 0.3936, + "num_input_tokens_seen": 702545920, + "step": 335 + }, + { + "epoch": 0.09113100081366965, + "grad_norm": 13.222553253173828, + "learning_rate": 9.95877699121683e-07, + "loss": 0.256, + "num_input_tokens_seen": 704643072, + "step": 336 + }, + { + "epoch": 0.091402224030377, + "grad_norm": 20.804859161376953, + "learning_rate": 9.95823003219893e-07, + "loss": 1.0641, + "num_input_tokens_seen": 706740224, + "step": 337 + }, + { + "epoch": 0.09167344724708434, + "grad_norm": 11.381640434265137, + "learning_rate": 9.957679485367144e-07, + "loss": 0.3291, + "num_input_tokens_seen": 708837376, + "step": 338 + }, + { + "epoch": 0.0919446704637917, + "grad_norm": 13.033456802368164, + "learning_rate": 9.95712535116454e-07, + "loss": 0.4397, + "num_input_tokens_seen": 710934528, + "step": 339 + }, + { + "epoch": 0.09221589368049905, + "grad_norm": 21.973115921020508, + "learning_rate": 9.956567630037058e-07, + "loss": 0.6069, + "num_input_tokens_seen": 713031680, + "step": 340 + }, + { + "epoch": 0.0924871168972064, + "grad_norm": 20.099735260009766, + "learning_rate": 9.95600632243353e-07, + "loss": 0.8007, + "num_input_tokens_seen": 715128832, + "step": 341 + }, + { + "epoch": 0.09275834011391375, + "grad_norm": 19.394807815551758, + "learning_rate": 9.95544142880568e-07, + "loss": 0.7435, + "num_input_tokens_seen": 717225984, + "step": 342 + }, + { + "epoch": 0.0930295633306211, + "grad_norm": 17.3151798248291, + "learning_rate": 9.954872949608108e-07, + "loss": 0.5055, + "num_input_tokens_seen": 719323136, + "step": 343 + }, + { + "epoch": 0.09330078654732846, + "grad_norm": 19.800769805908203, + "learning_rate": 9.954300885298309e-07, + "loss": 0.6907, + "num_input_tokens_seen": 721420288, + "step": 344 + }, + { + "epoch": 0.0935720097640358, + "grad_norm": 11.735955238342285, + "learning_rate": 9.953725236336653e-07, + "loss": 0.3009, + "num_input_tokens_seen": 723517440, + "step": 345 + }, + { + "epoch": 0.09384323298074315, + "grad_norm": 15.41151237487793, + "learning_rate": 9.953146003186407e-07, + "loss": 0.577, + "num_input_tokens_seen": 725614592, + "step": 346 + }, + { + "epoch": 0.0941144561974505, + "grad_norm": 21.002737045288086, + "learning_rate": 9.952563186313711e-07, + "loss": 0.9646, + "num_input_tokens_seen": 727711744, + "step": 347 + }, + { + "epoch": 0.09438567941415785, + "grad_norm": 8.514850616455078, + "learning_rate": 9.951976786187598e-07, + "loss": 0.1761, + "num_input_tokens_seen": 729808896, + "step": 348 + }, + { + "epoch": 0.09465690263086521, + "grad_norm": 17.795473098754883, + "learning_rate": 9.951386803279973e-07, + "loss": 0.5064, + "num_input_tokens_seen": 731906048, + "step": 349 + }, + { + "epoch": 0.09492812584757256, + "grad_norm": 15.562695503234863, + "learning_rate": 9.95079323806564e-07, + "loss": 0.5842, + "num_input_tokens_seen": 734003200, + "step": 350 + }, + { + "epoch": 0.0951993490642799, + "grad_norm": 14.850616455078125, + "learning_rate": 9.950196091022274e-07, + "loss": 0.4142, + "num_input_tokens_seen": 736100352, + "step": 351 + }, + { + "epoch": 0.09547057228098725, + "grad_norm": 13.248444557189941, + "learning_rate": 9.949595362630435e-07, + "loss": 0.3965, + "num_input_tokens_seen": 738197504, + "step": 352 + }, + { + "epoch": 0.0957417954976946, + "grad_norm": 19.31031608581543, + "learning_rate": 9.948991053373567e-07, + "loss": 0.6689, + "num_input_tokens_seen": 740294656, + "step": 353 + }, + { + "epoch": 0.09601301871440195, + "grad_norm": 18.403900146484375, + "learning_rate": 9.948383163738e-07, + "loss": 0.6139, + "num_input_tokens_seen": 742391808, + "step": 354 + }, + { + "epoch": 0.09628424193110931, + "grad_norm": 16.723413467407227, + "learning_rate": 9.947771694212933e-07, + "loss": 0.5709, + "num_input_tokens_seen": 744488960, + "step": 355 + }, + { + "epoch": 0.09655546514781665, + "grad_norm": 14.325742721557617, + "learning_rate": 9.947156645290456e-07, + "loss": 0.6557, + "num_input_tokens_seen": 746586112, + "step": 356 + }, + { + "epoch": 0.096826688364524, + "grad_norm": 19.636194229125977, + "learning_rate": 9.94653801746554e-07, + "loss": 0.8035, + "num_input_tokens_seen": 748683264, + "step": 357 + }, + { + "epoch": 0.09709791158123135, + "grad_norm": 13.98243236541748, + "learning_rate": 9.945915811236029e-07, + "loss": 0.508, + "num_input_tokens_seen": 750780416, + "step": 358 + }, + { + "epoch": 0.0973691347979387, + "grad_norm": 14.497775077819824, + "learning_rate": 9.945290027102654e-07, + "loss": 0.4487, + "num_input_tokens_seen": 752877568, + "step": 359 + }, + { + "epoch": 0.09764035801464606, + "grad_norm": 15.805448532104492, + "learning_rate": 9.944660665569023e-07, + "loss": 0.6489, + "num_input_tokens_seen": 754974720, + "step": 360 + }, + { + "epoch": 0.0979115812313534, + "grad_norm": 14.843584060668945, + "learning_rate": 9.944027727141617e-07, + "loss": 0.5395, + "num_input_tokens_seen": 757071872, + "step": 361 + }, + { + "epoch": 0.09818280444806075, + "grad_norm": 15.000771522521973, + "learning_rate": 9.943391212329805e-07, + "loss": 0.4855, + "num_input_tokens_seen": 759169024, + "step": 362 + }, + { + "epoch": 0.0984540276647681, + "grad_norm": 26.341798782348633, + "learning_rate": 9.942751121645828e-07, + "loss": 1.1353, + "num_input_tokens_seen": 761266176, + "step": 363 + }, + { + "epoch": 0.09872525088147545, + "grad_norm": 14.345602989196777, + "learning_rate": 9.94210745560481e-07, + "loss": 0.5821, + "num_input_tokens_seen": 763363328, + "step": 364 + }, + { + "epoch": 0.09899647409818281, + "grad_norm": 13.088194847106934, + "learning_rate": 9.941460214724747e-07, + "loss": 0.3652, + "num_input_tokens_seen": 765460480, + "step": 365 + }, + { + "epoch": 0.09926769731489016, + "grad_norm": 25.987003326416016, + "learning_rate": 9.94080939952651e-07, + "loss": 0.8981, + "num_input_tokens_seen": 767557632, + "step": 366 + }, + { + "epoch": 0.0995389205315975, + "grad_norm": 19.704673767089844, + "learning_rate": 9.940155010533855e-07, + "loss": 0.721, + "num_input_tokens_seen": 769654784, + "step": 367 + }, + { + "epoch": 0.09981014374830485, + "grad_norm": 15.577407836914062, + "learning_rate": 9.939497048273407e-07, + "loss": 0.741, + "num_input_tokens_seen": 771751936, + "step": 368 + }, + { + "epoch": 0.1000813669650122, + "grad_norm": 23.83604621887207, + "learning_rate": 9.938835513274672e-07, + "loss": 0.7845, + "num_input_tokens_seen": 773849088, + "step": 369 + }, + { + "epoch": 0.10035259018171956, + "grad_norm": 13.335168838500977, + "learning_rate": 9.938170406070025e-07, + "loss": 0.4799, + "num_input_tokens_seen": 775946240, + "step": 370 + }, + { + "epoch": 0.10062381339842691, + "grad_norm": 12.84589672088623, + "learning_rate": 9.937501727194721e-07, + "loss": 0.4814, + "num_input_tokens_seen": 778043392, + "step": 371 + }, + { + "epoch": 0.10089503661513426, + "grad_norm": 17.722841262817383, + "learning_rate": 9.936829477186884e-07, + "loss": 0.6229, + "num_input_tokens_seen": 780140544, + "step": 372 + }, + { + "epoch": 0.1011662598318416, + "grad_norm": 11.619221687316895, + "learning_rate": 9.93615365658752e-07, + "loss": 0.3687, + "num_input_tokens_seen": 782237696, + "step": 373 + }, + { + "epoch": 0.10143748304854895, + "grad_norm": 17.260656356811523, + "learning_rate": 9.9354742659405e-07, + "loss": 0.7215, + "num_input_tokens_seen": 784334848, + "step": 374 + }, + { + "epoch": 0.10170870626525631, + "grad_norm": 19.690866470336914, + "learning_rate": 9.934791305792575e-07, + "loss": 1.015, + "num_input_tokens_seen": 786432000, + "step": 375 + }, + { + "epoch": 0.10197992948196366, + "grad_norm": 22.19734001159668, + "learning_rate": 9.934104776693363e-07, + "loss": 1.135, + "num_input_tokens_seen": 788529152, + "step": 376 + }, + { + "epoch": 0.10225115269867101, + "grad_norm": 12.495186805725098, + "learning_rate": 9.933414679195354e-07, + "loss": 0.4608, + "num_input_tokens_seen": 790626304, + "step": 377 + }, + { + "epoch": 0.10252237591537836, + "grad_norm": 13.882384300231934, + "learning_rate": 9.932721013853917e-07, + "loss": 0.4727, + "num_input_tokens_seen": 792723456, + "step": 378 + }, + { + "epoch": 0.1027935991320857, + "grad_norm": 17.025941848754883, + "learning_rate": 9.932023781227287e-07, + "loss": 0.3467, + "num_input_tokens_seen": 794820608, + "step": 379 + }, + { + "epoch": 0.10306482234879305, + "grad_norm": 9.603739738464355, + "learning_rate": 9.931322981876567e-07, + "loss": 0.3002, + "num_input_tokens_seen": 796917760, + "step": 380 + }, + { + "epoch": 0.10333604556550041, + "grad_norm": 9.856173515319824, + "learning_rate": 9.930618616365737e-07, + "loss": 0.2998, + "num_input_tokens_seen": 799014912, + "step": 381 + }, + { + "epoch": 0.10360726878220776, + "grad_norm": 16.00568199157715, + "learning_rate": 9.92991068526164e-07, + "loss": 0.6395, + "num_input_tokens_seen": 801112064, + "step": 382 + }, + { + "epoch": 0.1038784919989151, + "grad_norm": 13.31304931640625, + "learning_rate": 9.929199189133996e-07, + "loss": 0.4683, + "num_input_tokens_seen": 803209216, + "step": 383 + }, + { + "epoch": 0.10414971521562245, + "grad_norm": 22.845258712768555, + "learning_rate": 9.928484128555388e-07, + "loss": 0.9509, + "num_input_tokens_seen": 805306368, + "step": 384 + }, + { + "epoch": 0.1044209384323298, + "grad_norm": 18.43478012084961, + "learning_rate": 9.92776550410127e-07, + "loss": 0.6683, + "num_input_tokens_seen": 807403520, + "step": 385 + }, + { + "epoch": 0.10469216164903716, + "grad_norm": 19.82917022705078, + "learning_rate": 9.927043316349962e-07, + "loss": 0.375, + "num_input_tokens_seen": 809500672, + "step": 386 + }, + { + "epoch": 0.10496338486574451, + "grad_norm": 17.08930206298828, + "learning_rate": 9.926317565882657e-07, + "loss": 0.4983, + "num_input_tokens_seen": 811597824, + "step": 387 + }, + { + "epoch": 0.10523460808245186, + "grad_norm": 13.114651679992676, + "learning_rate": 9.925588253283407e-07, + "loss": 0.433, + "num_input_tokens_seen": 813694976, + "step": 388 + }, + { + "epoch": 0.1055058312991592, + "grad_norm": 12.93115234375, + "learning_rate": 9.924855379139136e-07, + "loss": 0.4286, + "num_input_tokens_seen": 815792128, + "step": 389 + }, + { + "epoch": 0.10577705451586655, + "grad_norm": 16.018659591674805, + "learning_rate": 9.924118944039635e-07, + "loss": 0.5672, + "num_input_tokens_seen": 817889280, + "step": 390 + }, + { + "epoch": 0.10604827773257391, + "grad_norm": 17.725582122802734, + "learning_rate": 9.923378948577558e-07, + "loss": 0.7503, + "num_input_tokens_seen": 819986432, + "step": 391 + }, + { + "epoch": 0.10631950094928126, + "grad_norm": 12.33151912689209, + "learning_rate": 9.922635393348425e-07, + "loss": 0.4547, + "num_input_tokens_seen": 822083584, + "step": 392 + }, + { + "epoch": 0.10659072416598861, + "grad_norm": 15.34462833404541, + "learning_rate": 9.92188827895062e-07, + "loss": 0.4688, + "num_input_tokens_seen": 824180736, + "step": 393 + }, + { + "epoch": 0.10686194738269596, + "grad_norm": 20.61101722717285, + "learning_rate": 9.921137605985397e-07, + "loss": 0.8941, + "num_input_tokens_seen": 826277888, + "step": 394 + }, + { + "epoch": 0.1071331705994033, + "grad_norm": 14.963687896728516, + "learning_rate": 9.920383375056863e-07, + "loss": 0.584, + "num_input_tokens_seen": 828375040, + "step": 395 + }, + { + "epoch": 0.10740439381611067, + "grad_norm": 17.79062843322754, + "learning_rate": 9.919625586771998e-07, + "loss": 0.7246, + "num_input_tokens_seen": 830472192, + "step": 396 + }, + { + "epoch": 0.10767561703281801, + "grad_norm": 16.45888328552246, + "learning_rate": 9.918864241740639e-07, + "loss": 0.6275, + "num_input_tokens_seen": 832569344, + "step": 397 + }, + { + "epoch": 0.10794684024952536, + "grad_norm": 17.214998245239258, + "learning_rate": 9.918099340575487e-07, + "loss": 0.53, + "num_input_tokens_seen": 834666496, + "step": 398 + }, + { + "epoch": 0.10821806346623271, + "grad_norm": 18.772640228271484, + "learning_rate": 9.91733088389211e-07, + "loss": 0.5958, + "num_input_tokens_seen": 836763648, + "step": 399 + }, + { + "epoch": 0.10848928668294006, + "grad_norm": 13.285908699035645, + "learning_rate": 9.916558872308929e-07, + "loss": 0.3651, + "num_input_tokens_seen": 838860800, + "step": 400 + }, + { + "epoch": 0.1087605098996474, + "grad_norm": 14.000696182250977, + "learning_rate": 9.915783306447229e-07, + "loss": 0.5324, + "num_input_tokens_seen": 840957952, + "step": 401 + }, + { + "epoch": 0.10903173311635476, + "grad_norm": 14.921085357666016, + "learning_rate": 9.915004186931156e-07, + "loss": 0.5363, + "num_input_tokens_seen": 843055104, + "step": 402 + }, + { + "epoch": 0.10930295633306211, + "grad_norm": 13.49205493927002, + "learning_rate": 9.91422151438772e-07, + "loss": 0.3414, + "num_input_tokens_seen": 845152256, + "step": 403 + }, + { + "epoch": 0.10957417954976946, + "grad_norm": 20.21406364440918, + "learning_rate": 9.91343528944678e-07, + "loss": 0.8933, + "num_input_tokens_seen": 847249408, + "step": 404 + }, + { + "epoch": 0.10984540276647681, + "grad_norm": 11.765609741210938, + "learning_rate": 9.912645512741064e-07, + "loss": 0.3803, + "num_input_tokens_seen": 849346560, + "step": 405 + }, + { + "epoch": 0.11011662598318415, + "grad_norm": 22.0726318359375, + "learning_rate": 9.911852184906151e-07, + "loss": 0.5149, + "num_input_tokens_seen": 851443712, + "step": 406 + }, + { + "epoch": 0.11038784919989152, + "grad_norm": 14.043606758117676, + "learning_rate": 9.911055306580485e-07, + "loss": 0.5114, + "num_input_tokens_seen": 853540864, + "step": 407 + }, + { + "epoch": 0.11065907241659886, + "grad_norm": 12.334258079528809, + "learning_rate": 9.910254878405361e-07, + "loss": 0.3898, + "num_input_tokens_seen": 855638016, + "step": 408 + }, + { + "epoch": 0.11093029563330621, + "grad_norm": 11.179019927978516, + "learning_rate": 9.909450901024935e-07, + "loss": 0.3856, + "num_input_tokens_seen": 857735168, + "step": 409 + }, + { + "epoch": 0.11120151885001356, + "grad_norm": 17.625316619873047, + "learning_rate": 9.908643375086213e-07, + "loss": 0.7504, + "num_input_tokens_seen": 859832320, + "step": 410 + }, + { + "epoch": 0.1114727420667209, + "grad_norm": 11.948003768920898, + "learning_rate": 9.907832301239066e-07, + "loss": 0.4077, + "num_input_tokens_seen": 861929472, + "step": 411 + }, + { + "epoch": 0.11174396528342827, + "grad_norm": 22.031713485717773, + "learning_rate": 9.907017680136213e-07, + "loss": 0.8141, + "num_input_tokens_seen": 864026624, + "step": 412 + }, + { + "epoch": 0.11201518850013562, + "grad_norm": 14.142679214477539, + "learning_rate": 9.90619951243323e-07, + "loss": 0.5477, + "num_input_tokens_seen": 866123776, + "step": 413 + }, + { + "epoch": 0.11228641171684296, + "grad_norm": 16.43646240234375, + "learning_rate": 9.905377798788547e-07, + "loss": 0.6771, + "num_input_tokens_seen": 868220928, + "step": 414 + }, + { + "epoch": 0.11255763493355031, + "grad_norm": 14.17488956451416, + "learning_rate": 9.904552539863452e-07, + "loss": 0.4444, + "num_input_tokens_seen": 870318080, + "step": 415 + }, + { + "epoch": 0.11282885815025766, + "grad_norm": 17.454513549804688, + "learning_rate": 9.903723736322075e-07, + "loss": 0.7704, + "num_input_tokens_seen": 872415232, + "step": 416 + }, + { + "epoch": 0.11310008136696502, + "grad_norm": 18.859848022460938, + "learning_rate": 9.90289138883141e-07, + "loss": 0.813, + "num_input_tokens_seen": 874512384, + "step": 417 + }, + { + "epoch": 0.11337130458367237, + "grad_norm": 21.147764205932617, + "learning_rate": 9.9020554980613e-07, + "loss": 0.6002, + "num_input_tokens_seen": 876609536, + "step": 418 + }, + { + "epoch": 0.11364252780037971, + "grad_norm": 20.489925384521484, + "learning_rate": 9.901216064684434e-07, + "loss": 0.9423, + "num_input_tokens_seen": 878706688, + "step": 419 + }, + { + "epoch": 0.11391375101708706, + "grad_norm": 12.808001518249512, + "learning_rate": 9.900373089376357e-07, + "loss": 0.4228, + "num_input_tokens_seen": 880803840, + "step": 420 + }, + { + "epoch": 0.11418497423379441, + "grad_norm": 12.175069808959961, + "learning_rate": 9.899526572815465e-07, + "loss": 0.3785, + "num_input_tokens_seen": 882900992, + "step": 421 + }, + { + "epoch": 0.11445619745050176, + "grad_norm": 12.485739707946777, + "learning_rate": 9.898676515683001e-07, + "loss": 0.4116, + "num_input_tokens_seen": 884998144, + "step": 422 + }, + { + "epoch": 0.11472742066720912, + "grad_norm": 16.668554306030273, + "learning_rate": 9.897822918663062e-07, + "loss": 0.5853, + "num_input_tokens_seen": 887095296, + "step": 423 + }, + { + "epoch": 0.11499864388391647, + "grad_norm": 20.581716537475586, + "learning_rate": 9.896965782442584e-07, + "loss": 0.6277, + "num_input_tokens_seen": 889192448, + "step": 424 + }, + { + "epoch": 0.11526986710062381, + "grad_norm": 17.985069274902344, + "learning_rate": 9.896105107711365e-07, + "loss": 0.5649, + "num_input_tokens_seen": 891289600, + "step": 425 + }, + { + "epoch": 0.11554109031733116, + "grad_norm": 16.7568302154541, + "learning_rate": 9.895240895162037e-07, + "loss": 0.5382, + "num_input_tokens_seen": 893386752, + "step": 426 + }, + { + "epoch": 0.11581231353403851, + "grad_norm": 14.392632484436035, + "learning_rate": 9.89437314549009e-07, + "loss": 0.493, + "num_input_tokens_seen": 895483904, + "step": 427 + }, + { + "epoch": 0.11608353675074587, + "grad_norm": 11.781503677368164, + "learning_rate": 9.89350185939385e-07, + "loss": 0.2819, + "num_input_tokens_seen": 897581056, + "step": 428 + }, + { + "epoch": 0.11635475996745322, + "grad_norm": 18.502288818359375, + "learning_rate": 9.8926270375745e-07, + "loss": 0.8497, + "num_input_tokens_seen": 899678208, + "step": 429 + }, + { + "epoch": 0.11662598318416056, + "grad_norm": 13.56667709350586, + "learning_rate": 9.891748680736064e-07, + "loss": 0.4923, + "num_input_tokens_seen": 901775360, + "step": 430 + }, + { + "epoch": 0.11689720640086791, + "grad_norm": 19.49355125427246, + "learning_rate": 9.890866789585407e-07, + "loss": 0.6438, + "num_input_tokens_seen": 903872512, + "step": 431 + }, + { + "epoch": 0.11716842961757526, + "grad_norm": 11.730159759521484, + "learning_rate": 9.889981364832245e-07, + "loss": 0.3834, + "num_input_tokens_seen": 905969664, + "step": 432 + }, + { + "epoch": 0.11743965283428262, + "grad_norm": 10.967796325683594, + "learning_rate": 9.889092407189129e-07, + "loss": 0.3682, + "num_input_tokens_seen": 908066816, + "step": 433 + }, + { + "epoch": 0.11771087605098997, + "grad_norm": 16.377155303955078, + "learning_rate": 9.88819991737146e-07, + "loss": 0.454, + "num_input_tokens_seen": 910163968, + "step": 434 + }, + { + "epoch": 0.11798209926769732, + "grad_norm": 15.872443199157715, + "learning_rate": 9.887303896097483e-07, + "loss": 0.6401, + "num_input_tokens_seen": 912261120, + "step": 435 + }, + { + "epoch": 0.11825332248440466, + "grad_norm": 15.673075675964355, + "learning_rate": 9.88640434408828e-07, + "loss": 0.7127, + "num_input_tokens_seen": 914358272, + "step": 436 + }, + { + "epoch": 0.11852454570111201, + "grad_norm": 18.29340934753418, + "learning_rate": 9.885501262067776e-07, + "loss": 0.4442, + "num_input_tokens_seen": 916455424, + "step": 437 + }, + { + "epoch": 0.11879576891781937, + "grad_norm": 21.140331268310547, + "learning_rate": 9.884594650762734e-07, + "loss": 1.0107, + "num_input_tokens_seen": 918552576, + "step": 438 + }, + { + "epoch": 0.11906699213452672, + "grad_norm": 21.628915786743164, + "learning_rate": 9.883684510902767e-07, + "loss": 0.4715, + "num_input_tokens_seen": 920649728, + "step": 439 + }, + { + "epoch": 0.11933821535123407, + "grad_norm": 17.046772003173828, + "learning_rate": 9.882770843220316e-07, + "loss": 0.7166, + "num_input_tokens_seen": 922746880, + "step": 440 + }, + { + "epoch": 0.11960943856794141, + "grad_norm": 13.360857963562012, + "learning_rate": 9.881853648450667e-07, + "loss": 0.4021, + "num_input_tokens_seen": 924844032, + "step": 441 + }, + { + "epoch": 0.11988066178464876, + "grad_norm": 12.711376190185547, + "learning_rate": 9.880932927331942e-07, + "loss": 0.3842, + "num_input_tokens_seen": 926941184, + "step": 442 + }, + { + "epoch": 0.12015188500135611, + "grad_norm": 13.425957679748535, + "learning_rate": 9.880008680605104e-07, + "loss": 0.5039, + "num_input_tokens_seen": 929038336, + "step": 443 + }, + { + "epoch": 0.12042310821806347, + "grad_norm": 14.570152282714844, + "learning_rate": 9.879080909013955e-07, + "loss": 0.3667, + "num_input_tokens_seen": 931135488, + "step": 444 + }, + { + "epoch": 0.12069433143477082, + "grad_norm": 15.635329246520996, + "learning_rate": 9.878149613305125e-07, + "loss": 0.6197, + "num_input_tokens_seen": 933232640, + "step": 445 + }, + { + "epoch": 0.12096555465147817, + "grad_norm": 10.609899520874023, + "learning_rate": 9.877214794228087e-07, + "loss": 0.3233, + "num_input_tokens_seen": 935329792, + "step": 446 + }, + { + "epoch": 0.12123677786818551, + "grad_norm": 18.10693359375, + "learning_rate": 9.876276452535147e-07, + "loss": 0.664, + "num_input_tokens_seen": 937426944, + "step": 447 + }, + { + "epoch": 0.12150800108489286, + "grad_norm": 12.761764526367188, + "learning_rate": 9.87533458898145e-07, + "loss": 0.3728, + "num_input_tokens_seen": 939524096, + "step": 448 + }, + { + "epoch": 0.12177922430160022, + "grad_norm": 11.233665466308594, + "learning_rate": 9.874389204324967e-07, + "loss": 0.4342, + "num_input_tokens_seen": 941621248, + "step": 449 + }, + { + "epoch": 0.12205044751830757, + "grad_norm": 14.0928955078125, + "learning_rate": 9.873440299326513e-07, + "loss": 0.5682, + "num_input_tokens_seen": 943718400, + "step": 450 + }, + { + "epoch": 0.12232167073501492, + "grad_norm": 14.01526927947998, + "learning_rate": 9.872487874749726e-07, + "loss": 0.5666, + "num_input_tokens_seen": 945815552, + "step": 451 + }, + { + "epoch": 0.12259289395172226, + "grad_norm": 15.093655586242676, + "learning_rate": 9.871531931361084e-07, + "loss": 0.4609, + "num_input_tokens_seen": 947912704, + "step": 452 + }, + { + "epoch": 0.12286411716842961, + "grad_norm": 16.862146377563477, + "learning_rate": 9.870572469929892e-07, + "loss": 0.4889, + "num_input_tokens_seen": 950009856, + "step": 453 + }, + { + "epoch": 0.12313534038513697, + "grad_norm": 14.488612174987793, + "learning_rate": 9.869609491228288e-07, + "loss": 0.4386, + "num_input_tokens_seen": 952107008, + "step": 454 + }, + { + "epoch": 0.12340656360184432, + "grad_norm": 10.171273231506348, + "learning_rate": 9.868642996031243e-07, + "loss": 0.3064, + "num_input_tokens_seen": 954204160, + "step": 455 + }, + { + "epoch": 0.12367778681855167, + "grad_norm": 13.183239936828613, + "learning_rate": 9.867672985116553e-07, + "loss": 0.4111, + "num_input_tokens_seen": 956301312, + "step": 456 + }, + { + "epoch": 0.12394901003525902, + "grad_norm": 13.565240859985352, + "learning_rate": 9.866699459264846e-07, + "loss": 0.5021, + "num_input_tokens_seen": 958398464, + "step": 457 + }, + { + "epoch": 0.12422023325196636, + "grad_norm": 14.678339958190918, + "learning_rate": 9.865722419259582e-07, + "loss": 0.4652, + "num_input_tokens_seen": 960495616, + "step": 458 + }, + { + "epoch": 0.12449145646867373, + "grad_norm": 20.48042869567871, + "learning_rate": 9.864741865887042e-07, + "loss": 0.91, + "num_input_tokens_seen": 962592768, + "step": 459 + }, + { + "epoch": 0.12476267968538107, + "grad_norm": 20.1795711517334, + "learning_rate": 9.86375779993634e-07, + "loss": 0.6582, + "num_input_tokens_seen": 964689920, + "step": 460 + }, + { + "epoch": 0.1250339029020884, + "grad_norm": 28.812326431274414, + "learning_rate": 9.86277022219941e-07, + "loss": 0.6452, + "num_input_tokens_seen": 966787072, + "step": 461 + }, + { + "epoch": 0.12530512611879577, + "grad_norm": 21.05647850036621, + "learning_rate": 9.861779133471025e-07, + "loss": 0.9777, + "num_input_tokens_seen": 968884224, + "step": 462 + }, + { + "epoch": 0.12557634933550313, + "grad_norm": 18.095561981201172, + "learning_rate": 9.86078453454877e-07, + "loss": 0.3998, + "num_input_tokens_seen": 970981376, + "step": 463 + }, + { + "epoch": 0.12584757255221046, + "grad_norm": 12.681943893432617, + "learning_rate": 9.859786426233061e-07, + "loss": 0.4588, + "num_input_tokens_seen": 973078528, + "step": 464 + }, + { + "epoch": 0.12611879576891782, + "grad_norm": 11.929615020751953, + "learning_rate": 9.85878480932714e-07, + "loss": 0.3087, + "num_input_tokens_seen": 975175680, + "step": 465 + }, + { + "epoch": 0.12639001898562516, + "grad_norm": 20.00129508972168, + "learning_rate": 9.857779684637068e-07, + "loss": 0.8953, + "num_input_tokens_seen": 977272832, + "step": 466 + }, + { + "epoch": 0.12666124220233252, + "grad_norm": 12.261861801147461, + "learning_rate": 9.856771052971733e-07, + "loss": 0.3192, + "num_input_tokens_seen": 979369984, + "step": 467 + }, + { + "epoch": 0.12693246541903988, + "grad_norm": 15.92088794708252, + "learning_rate": 9.85575891514284e-07, + "loss": 0.5931, + "num_input_tokens_seen": 981467136, + "step": 468 + }, + { + "epoch": 0.12720368863574721, + "grad_norm": 12.561171531677246, + "learning_rate": 9.85474327196492e-07, + "loss": 0.3394, + "num_input_tokens_seen": 983564288, + "step": 469 + }, + { + "epoch": 0.12747491185245458, + "grad_norm": 16.519527435302734, + "learning_rate": 9.853724124255328e-07, + "loss": 0.4335, + "num_input_tokens_seen": 985661440, + "step": 470 + }, + { + "epoch": 0.1277461350691619, + "grad_norm": 23.729719161987305, + "learning_rate": 9.85270147283423e-07, + "loss": 0.6265, + "num_input_tokens_seen": 987758592, + "step": 471 + }, + { + "epoch": 0.12801735828586927, + "grad_norm": 20.1112003326416, + "learning_rate": 9.85167531852462e-07, + "loss": 0.8864, + "num_input_tokens_seen": 989855744, + "step": 472 + }, + { + "epoch": 0.12828858150257663, + "grad_norm": 11.423717498779297, + "learning_rate": 9.850645662152308e-07, + "loss": 0.3044, + "num_input_tokens_seen": 991952896, + "step": 473 + }, + { + "epoch": 0.12855980471928397, + "grad_norm": 15.77081298828125, + "learning_rate": 9.84961250454592e-07, + "loss": 0.4784, + "num_input_tokens_seen": 994050048, + "step": 474 + }, + { + "epoch": 0.12883102793599133, + "grad_norm": 14.118724822998047, + "learning_rate": 9.848575846536902e-07, + "loss": 0.4869, + "num_input_tokens_seen": 996147200, + "step": 475 + }, + { + "epoch": 0.12910225115269866, + "grad_norm": 14.419587135314941, + "learning_rate": 9.847535688959523e-07, + "loss": 0.5372, + "num_input_tokens_seen": 998244352, + "step": 476 + }, + { + "epoch": 0.12937347436940602, + "grad_norm": 17.4903507232666, + "learning_rate": 9.846492032650855e-07, + "loss": 0.5932, + "num_input_tokens_seen": 1000341504, + "step": 477 + }, + { + "epoch": 0.12964469758611338, + "grad_norm": 13.698707580566406, + "learning_rate": 9.845444878450794e-07, + "loss": 0.4332, + "num_input_tokens_seen": 1002438656, + "step": 478 + }, + { + "epoch": 0.12991592080282072, + "grad_norm": 20.30847930908203, + "learning_rate": 9.844394227202053e-07, + "loss": 0.5934, + "num_input_tokens_seen": 1004535808, + "step": 479 + }, + { + "epoch": 0.13018714401952808, + "grad_norm": 14.15876293182373, + "learning_rate": 9.843340079750154e-07, + "loss": 0.4459, + "num_input_tokens_seen": 1006632960, + "step": 480 + }, + { + "epoch": 0.1304583672362354, + "grad_norm": 12.466696739196777, + "learning_rate": 9.842282436943435e-07, + "loss": 0.3993, + "num_input_tokens_seen": 1008730112, + "step": 481 + }, + { + "epoch": 0.13072959045294277, + "grad_norm": 17.355510711669922, + "learning_rate": 9.841221299633049e-07, + "loss": 0.517, + "num_input_tokens_seen": 1010827264, + "step": 482 + }, + { + "epoch": 0.13100081366965013, + "grad_norm": 23.515899658203125, + "learning_rate": 9.840156668672953e-07, + "loss": 0.7, + "num_input_tokens_seen": 1012924416, + "step": 483 + }, + { + "epoch": 0.13127203688635747, + "grad_norm": 8.689852714538574, + "learning_rate": 9.839088544919927e-07, + "loss": 0.3053, + "num_input_tokens_seen": 1015021568, + "step": 484 + }, + { + "epoch": 0.13154326010306483, + "grad_norm": 16.73533058166504, + "learning_rate": 9.838016929233555e-07, + "loss": 0.4763, + "num_input_tokens_seen": 1017118720, + "step": 485 + }, + { + "epoch": 0.13181448331977216, + "grad_norm": 16.722810745239258, + "learning_rate": 9.836941822476232e-07, + "loss": 0.3931, + "num_input_tokens_seen": 1019215872, + "step": 486 + }, + { + "epoch": 0.13208570653647952, + "grad_norm": 16.654083251953125, + "learning_rate": 9.835863225513163e-07, + "loss": 0.63, + "num_input_tokens_seen": 1021313024, + "step": 487 + }, + { + "epoch": 0.13235692975318689, + "grad_norm": 17.620363235473633, + "learning_rate": 9.83478113921236e-07, + "loss": 0.8202, + "num_input_tokens_seen": 1023410176, + "step": 488 + }, + { + "epoch": 0.13262815296989422, + "grad_norm": 14.268228530883789, + "learning_rate": 9.833695564444652e-07, + "loss": 0.4924, + "num_input_tokens_seen": 1025507328, + "step": 489 + }, + { + "epoch": 0.13289937618660158, + "grad_norm": 12.854048728942871, + "learning_rate": 9.832606502083658e-07, + "loss": 0.4732, + "num_input_tokens_seen": 1027604480, + "step": 490 + }, + { + "epoch": 0.13317059940330891, + "grad_norm": 17.436908721923828, + "learning_rate": 9.83151395300582e-07, + "loss": 0.5987, + "num_input_tokens_seen": 1029701632, + "step": 491 + }, + { + "epoch": 0.13344182262001628, + "grad_norm": 13.574828147888184, + "learning_rate": 9.83041791809038e-07, + "loss": 0.4847, + "num_input_tokens_seen": 1031798784, + "step": 492 + }, + { + "epoch": 0.13371304583672364, + "grad_norm": 18.534931182861328, + "learning_rate": 9.829318398219385e-07, + "loss": 0.5305, + "num_input_tokens_seen": 1033895936, + "step": 493 + }, + { + "epoch": 0.13398426905343097, + "grad_norm": 13.970446586608887, + "learning_rate": 9.828215394277686e-07, + "loss": 0.3694, + "num_input_tokens_seen": 1035993088, + "step": 494 + }, + { + "epoch": 0.13425549227013833, + "grad_norm": 15.922163009643555, + "learning_rate": 9.827108907152937e-07, + "loss": 0.5241, + "num_input_tokens_seen": 1038090240, + "step": 495 + }, + { + "epoch": 0.13452671548684567, + "grad_norm": 15.5534029006958, + "learning_rate": 9.825998937735599e-07, + "loss": 0.6606, + "num_input_tokens_seen": 1040187392, + "step": 496 + }, + { + "epoch": 0.13479793870355303, + "grad_norm": 18.83726692199707, + "learning_rate": 9.824885486918932e-07, + "loss": 0.7119, + "num_input_tokens_seen": 1042284544, + "step": 497 + }, + { + "epoch": 0.13506916192026036, + "grad_norm": 19.1335506439209, + "learning_rate": 9.823768555599e-07, + "loss": 0.8065, + "num_input_tokens_seen": 1044381696, + "step": 498 + }, + { + "epoch": 0.13534038513696772, + "grad_norm": 13.459641456604004, + "learning_rate": 9.822648144674664e-07, + "loss": 0.324, + "num_input_tokens_seen": 1046478848, + "step": 499 + }, + { + "epoch": 0.13561160835367508, + "grad_norm": 11.397218704223633, + "learning_rate": 9.821524255047592e-07, + "loss": 0.3856, + "num_input_tokens_seen": 1048576000, + "step": 500 + }, + { + "epoch": 0.13588283157038242, + "grad_norm": 13.145438194274902, + "learning_rate": 9.820396887622245e-07, + "loss": 0.4602, + "num_input_tokens_seen": 1050673152, + "step": 501 + }, + { + "epoch": 0.13615405478708978, + "grad_norm": 12.989845275878906, + "learning_rate": 9.819266043305887e-07, + "loss": 0.3918, + "num_input_tokens_seen": 1052770304, + "step": 502 + }, + { + "epoch": 0.1364252780037971, + "grad_norm": 18.566017150878906, + "learning_rate": 9.818131723008576e-07, + "loss": 0.5483, + "num_input_tokens_seen": 1054867456, + "step": 503 + }, + { + "epoch": 0.13669650122050447, + "grad_norm": 20.60564613342285, + "learning_rate": 9.816993927643174e-07, + "loss": 0.8955, + "num_input_tokens_seen": 1056964608, + "step": 504 + }, + { + "epoch": 0.13696772443721184, + "grad_norm": 13.428114891052246, + "learning_rate": 9.815852658125332e-07, + "loss": 0.4461, + "num_input_tokens_seen": 1059061760, + "step": 505 + }, + { + "epoch": 0.13723894765391917, + "grad_norm": 15.920856475830078, + "learning_rate": 9.8147079153735e-07, + "loss": 0.6438, + "num_input_tokens_seen": 1061158912, + "step": 506 + }, + { + "epoch": 0.13751017087062653, + "grad_norm": 11.576249122619629, + "learning_rate": 9.813559700308925e-07, + "loss": 0.3101, + "num_input_tokens_seen": 1063256064, + "step": 507 + }, + { + "epoch": 0.13778139408733386, + "grad_norm": 20.885211944580078, + "learning_rate": 9.812408013855646e-07, + "loss": 0.7386, + "num_input_tokens_seen": 1065353216, + "step": 508 + }, + { + "epoch": 0.13805261730404123, + "grad_norm": 16.579875946044922, + "learning_rate": 9.811252856940496e-07, + "loss": 0.6981, + "num_input_tokens_seen": 1067450368, + "step": 509 + }, + { + "epoch": 0.1383238405207486, + "grad_norm": 10.057287216186523, + "learning_rate": 9.810094230493104e-07, + "loss": 0.2932, + "num_input_tokens_seen": 1069547520, + "step": 510 + }, + { + "epoch": 0.13859506373745592, + "grad_norm": 17.37000846862793, + "learning_rate": 9.808932135445885e-07, + "loss": 0.5647, + "num_input_tokens_seen": 1071644672, + "step": 511 + }, + { + "epoch": 0.13886628695416328, + "grad_norm": 15.917064666748047, + "learning_rate": 9.807766572734052e-07, + "loss": 0.7026, + "num_input_tokens_seen": 1073741824, + "step": 512 + }, + { + "epoch": 0.13913751017087062, + "grad_norm": 11.753175735473633, + "learning_rate": 9.806597543295603e-07, + "loss": 0.416, + "num_input_tokens_seen": 1075838976, + "step": 513 + }, + { + "epoch": 0.13940873338757798, + "grad_norm": 18.979516983032227, + "learning_rate": 9.80542504807133e-07, + "loss": 0.5972, + "num_input_tokens_seen": 1077936128, + "step": 514 + }, + { + "epoch": 0.13967995660428534, + "grad_norm": 28.44046401977539, + "learning_rate": 9.804249088004812e-07, + "loss": 0.4001, + "num_input_tokens_seen": 1080033280, + "step": 515 + }, + { + "epoch": 0.13995117982099267, + "grad_norm": 12.617019653320312, + "learning_rate": 9.803069664042416e-07, + "loss": 0.3327, + "num_input_tokens_seen": 1082130432, + "step": 516 + }, + { + "epoch": 0.14022240303770003, + "grad_norm": 14.718525886535645, + "learning_rate": 9.801886777133297e-07, + "loss": 0.5243, + "num_input_tokens_seen": 1084227584, + "step": 517 + }, + { + "epoch": 0.14049362625440737, + "grad_norm": 17.5297794342041, + "learning_rate": 9.8007004282294e-07, + "loss": 0.5279, + "num_input_tokens_seen": 1086324736, + "step": 518 + }, + { + "epoch": 0.14076484947111473, + "grad_norm": 13.261266708374023, + "learning_rate": 9.799510618285454e-07, + "loss": 0.4663, + "num_input_tokens_seen": 1088421888, + "step": 519 + }, + { + "epoch": 0.1410360726878221, + "grad_norm": 22.933935165405273, + "learning_rate": 9.79831734825897e-07, + "loss": 0.7519, + "num_input_tokens_seen": 1090519040, + "step": 520 + }, + { + "epoch": 0.14130729590452942, + "grad_norm": 20.257604598999023, + "learning_rate": 9.797120619110245e-07, + "loss": 0.7251, + "num_input_tokens_seen": 1092616192, + "step": 521 + }, + { + "epoch": 0.14157851912123678, + "grad_norm": 19.605735778808594, + "learning_rate": 9.795920431802365e-07, + "loss": 0.7534, + "num_input_tokens_seen": 1094713344, + "step": 522 + }, + { + "epoch": 0.14184974233794412, + "grad_norm": 13.443857192993164, + "learning_rate": 9.794716787301194e-07, + "loss": 0.5247, + "num_input_tokens_seen": 1096810496, + "step": 523 + }, + { + "epoch": 0.14212096555465148, + "grad_norm": 21.489839553833008, + "learning_rate": 9.793509686575378e-07, + "loss": 0.7223, + "num_input_tokens_seen": 1098907648, + "step": 524 + }, + { + "epoch": 0.14239218877135884, + "grad_norm": 15.88458251953125, + "learning_rate": 9.792299130596346e-07, + "loss": 0.6552, + "num_input_tokens_seen": 1101004800, + "step": 525 + }, + { + "epoch": 0.14266341198806617, + "grad_norm": 20.70093536376953, + "learning_rate": 9.79108512033831e-07, + "loss": 0.8404, + "num_input_tokens_seen": 1103101952, + "step": 526 + }, + { + "epoch": 0.14293463520477354, + "grad_norm": 17.892988204956055, + "learning_rate": 9.789867656778254e-07, + "loss": 0.6306, + "num_input_tokens_seen": 1105199104, + "step": 527 + }, + { + "epoch": 0.14320585842148087, + "grad_norm": 16.286060333251953, + "learning_rate": 9.788646740895952e-07, + "loss": 0.6665, + "num_input_tokens_seen": 1107296256, + "step": 528 + }, + { + "epoch": 0.14347708163818823, + "grad_norm": 14.654132843017578, + "learning_rate": 9.787422373673945e-07, + "loss": 0.5283, + "num_input_tokens_seen": 1109393408, + "step": 529 + }, + { + "epoch": 0.1437483048548956, + "grad_norm": 16.374067306518555, + "learning_rate": 9.786194556097564e-07, + "loss": 0.6534, + "num_input_tokens_seen": 1111490560, + "step": 530 + }, + { + "epoch": 0.14401952807160293, + "grad_norm": 19.840892791748047, + "learning_rate": 9.784963289154902e-07, + "loss": 0.928, + "num_input_tokens_seen": 1113587712, + "step": 531 + }, + { + "epoch": 0.1442907512883103, + "grad_norm": 12.525459289550781, + "learning_rate": 9.783728573836843e-07, + "loss": 0.4473, + "num_input_tokens_seen": 1115684864, + "step": 532 + }, + { + "epoch": 0.14456197450501762, + "grad_norm": 15.423880577087402, + "learning_rate": 9.782490411137035e-07, + "loss": 0.387, + "num_input_tokens_seen": 1117782016, + "step": 533 + }, + { + "epoch": 0.14483319772172498, + "grad_norm": 14.608306884765625, + "learning_rate": 9.781248802051904e-07, + "loss": 0.5885, + "num_input_tokens_seen": 1119879168, + "step": 534 + }, + { + "epoch": 0.14510442093843234, + "grad_norm": 17.79987144470215, + "learning_rate": 9.780003747580651e-07, + "loss": 0.674, + "num_input_tokens_seen": 1121976320, + "step": 535 + }, + { + "epoch": 0.14537564415513968, + "grad_norm": 11.311891555786133, + "learning_rate": 9.778755248725248e-07, + "loss": 0.4104, + "num_input_tokens_seen": 1124073472, + "step": 536 + }, + { + "epoch": 0.14564686737184704, + "grad_norm": 13.756656646728516, + "learning_rate": 9.77750330649044e-07, + "loss": 0.5469, + "num_input_tokens_seen": 1126170624, + "step": 537 + }, + { + "epoch": 0.14591809058855437, + "grad_norm": 19.89360809326172, + "learning_rate": 9.776247921883743e-07, + "loss": 0.9223, + "num_input_tokens_seen": 1128267776, + "step": 538 + }, + { + "epoch": 0.14618931380526173, + "grad_norm": 14.971158027648926, + "learning_rate": 9.774989095915442e-07, + "loss": 0.491, + "num_input_tokens_seen": 1130364928, + "step": 539 + }, + { + "epoch": 0.14646053702196907, + "grad_norm": 15.824216842651367, + "learning_rate": 9.77372682959859e-07, + "loss": 0.6727, + "num_input_tokens_seen": 1132462080, + "step": 540 + }, + { + "epoch": 0.14673176023867643, + "grad_norm": 16.35426902770996, + "learning_rate": 9.772461123949015e-07, + "loss": 0.7117, + "num_input_tokens_seen": 1134559232, + "step": 541 + }, + { + "epoch": 0.1470029834553838, + "grad_norm": 17.143768310546875, + "learning_rate": 9.771191979985303e-07, + "loss": 0.772, + "num_input_tokens_seen": 1136656384, + "step": 542 + }, + { + "epoch": 0.14727420667209112, + "grad_norm": 11.403589248657227, + "learning_rate": 9.76991939872882e-07, + "loss": 0.3995, + "num_input_tokens_seen": 1138753536, + "step": 543 + }, + { + "epoch": 0.14754542988879848, + "grad_norm": 13.465624809265137, + "learning_rate": 9.768643381203686e-07, + "loss": 0.5924, + "num_input_tokens_seen": 1140850688, + "step": 544 + }, + { + "epoch": 0.14781665310550582, + "grad_norm": 15.813559532165527, + "learning_rate": 9.767363928436793e-07, + "loss": 0.5763, + "num_input_tokens_seen": 1142947840, + "step": 545 + }, + { + "epoch": 0.14808787632221318, + "grad_norm": 17.59358787536621, + "learning_rate": 9.766081041457795e-07, + "loss": 0.6967, + "num_input_tokens_seen": 1145044992, + "step": 546 + }, + { + "epoch": 0.14835909953892054, + "grad_norm": 15.673956871032715, + "learning_rate": 9.764794721299113e-07, + "loss": 0.7236, + "num_input_tokens_seen": 1147142144, + "step": 547 + }, + { + "epoch": 0.14863032275562787, + "grad_norm": 13.424758911132812, + "learning_rate": 9.763504968995927e-07, + "loss": 0.4046, + "num_input_tokens_seen": 1149239296, + "step": 548 + }, + { + "epoch": 0.14890154597233524, + "grad_norm": 18.995798110961914, + "learning_rate": 9.762211785586178e-07, + "loss": 0.781, + "num_input_tokens_seen": 1151336448, + "step": 549 + }, + { + "epoch": 0.14917276918904257, + "grad_norm": 12.103052139282227, + "learning_rate": 9.760915172110574e-07, + "loss": 0.3885, + "num_input_tokens_seen": 1153433600, + "step": 550 + }, + { + "epoch": 0.14944399240574993, + "grad_norm": 12.997087478637695, + "learning_rate": 9.759615129612579e-07, + "loss": 0.3381, + "num_input_tokens_seen": 1155530752, + "step": 551 + }, + { + "epoch": 0.1497152156224573, + "grad_norm": 15.175359725952148, + "learning_rate": 9.75831165913842e-07, + "loss": 0.5939, + "num_input_tokens_seen": 1157627904, + "step": 552 + }, + { + "epoch": 0.14998643883916463, + "grad_norm": 15.981290817260742, + "learning_rate": 9.757004761737077e-07, + "loss": 0.5293, + "num_input_tokens_seen": 1159725056, + "step": 553 + }, + { + "epoch": 0.150257662055872, + "grad_norm": 11.93864631652832, + "learning_rate": 9.755694438460293e-07, + "loss": 0.4189, + "num_input_tokens_seen": 1161822208, + "step": 554 + }, + { + "epoch": 0.15052888527257932, + "grad_norm": 15.042703628540039, + "learning_rate": 9.754380690362565e-07, + "loss": 0.4993, + "num_input_tokens_seen": 1163919360, + "step": 555 + }, + { + "epoch": 0.15080010848928668, + "grad_norm": 14.069710731506348, + "learning_rate": 9.75306351850115e-07, + "loss": 0.472, + "num_input_tokens_seen": 1166016512, + "step": 556 + }, + { + "epoch": 0.15107133170599404, + "grad_norm": 15.970101356506348, + "learning_rate": 9.751742923936055e-07, + "loss": 0.4091, + "num_input_tokens_seen": 1168113664, + "step": 557 + }, + { + "epoch": 0.15134255492270138, + "grad_norm": 16.250965118408203, + "learning_rate": 9.75041890773005e-07, + "loss": 0.5406, + "num_input_tokens_seen": 1170210816, + "step": 558 + }, + { + "epoch": 0.15161377813940874, + "grad_norm": 17.741085052490234, + "learning_rate": 9.749091470948643e-07, + "loss": 0.4236, + "num_input_tokens_seen": 1172307968, + "step": 559 + }, + { + "epoch": 0.15188500135611607, + "grad_norm": 13.735379219055176, + "learning_rate": 9.747760614660111e-07, + "loss": 0.4524, + "num_input_tokens_seen": 1174405120, + "step": 560 + }, + { + "epoch": 0.15215622457282343, + "grad_norm": 14.853401184082031, + "learning_rate": 9.746426339935477e-07, + "loss": 0.4694, + "num_input_tokens_seen": 1176502272, + "step": 561 + }, + { + "epoch": 0.1524274477895308, + "grad_norm": 11.256040573120117, + "learning_rate": 9.745088647848515e-07, + "loss": 0.32, + "num_input_tokens_seen": 1178599424, + "step": 562 + }, + { + "epoch": 0.15269867100623813, + "grad_norm": 21.761676788330078, + "learning_rate": 9.743747539475744e-07, + "loss": 0.7391, + "num_input_tokens_seen": 1180696576, + "step": 563 + }, + { + "epoch": 0.1529698942229455, + "grad_norm": 15.350510597229004, + "learning_rate": 9.74240301589644e-07, + "loss": 0.5598, + "num_input_tokens_seen": 1182793728, + "step": 564 + }, + { + "epoch": 0.15324111743965282, + "grad_norm": 12.007080078125, + "learning_rate": 9.741055078192626e-07, + "loss": 0.4444, + "num_input_tokens_seen": 1184890880, + "step": 565 + }, + { + "epoch": 0.15351234065636019, + "grad_norm": 12.715327262878418, + "learning_rate": 9.739703727449068e-07, + "loss": 0.4789, + "num_input_tokens_seen": 1186988032, + "step": 566 + }, + { + "epoch": 0.15378356387306755, + "grad_norm": 10.164214134216309, + "learning_rate": 9.738348964753283e-07, + "loss": 0.3289, + "num_input_tokens_seen": 1189085184, + "step": 567 + }, + { + "epoch": 0.15405478708977488, + "grad_norm": 15.66612720489502, + "learning_rate": 9.736990791195532e-07, + "loss": 0.4524, + "num_input_tokens_seen": 1191182336, + "step": 568 + }, + { + "epoch": 0.15432601030648224, + "grad_norm": 12.935588836669922, + "learning_rate": 9.735629207868824e-07, + "loss": 0.3797, + "num_input_tokens_seen": 1193279488, + "step": 569 + }, + { + "epoch": 0.15459723352318958, + "grad_norm": 14.650870323181152, + "learning_rate": 9.734264215868904e-07, + "loss": 0.5074, + "num_input_tokens_seen": 1195376640, + "step": 570 + }, + { + "epoch": 0.15486845673989694, + "grad_norm": 16.69452667236328, + "learning_rate": 9.73289581629427e-07, + "loss": 0.4658, + "num_input_tokens_seen": 1197473792, + "step": 571 + }, + { + "epoch": 0.1551396799566043, + "grad_norm": 19.8432674407959, + "learning_rate": 9.73152401024616e-07, + "loss": 0.7099, + "num_input_tokens_seen": 1199570944, + "step": 572 + }, + { + "epoch": 0.15541090317331163, + "grad_norm": 17.770639419555664, + "learning_rate": 9.730148798828543e-07, + "loss": 0.5642, + "num_input_tokens_seen": 1201668096, + "step": 573 + }, + { + "epoch": 0.155682126390019, + "grad_norm": 20.076339721679688, + "learning_rate": 9.728770183148143e-07, + "loss": 0.7079, + "num_input_tokens_seen": 1203765248, + "step": 574 + }, + { + "epoch": 0.15595334960672633, + "grad_norm": 18.26179313659668, + "learning_rate": 9.727388164314415e-07, + "loss": 0.721, + "num_input_tokens_seen": 1205862400, + "step": 575 + }, + { + "epoch": 0.1562245728234337, + "grad_norm": 21.209247589111328, + "learning_rate": 9.72600274343956e-07, + "loss": 0.5755, + "num_input_tokens_seen": 1207959552, + "step": 576 + }, + { + "epoch": 0.15649579604014105, + "grad_norm": 16.366296768188477, + "learning_rate": 9.724613921638506e-07, + "loss": 0.3086, + "num_input_tokens_seen": 1210056704, + "step": 577 + }, + { + "epoch": 0.15676701925684838, + "grad_norm": 16.17742919921875, + "learning_rate": 9.723221700028928e-07, + "loss": 0.5038, + "num_input_tokens_seen": 1212153856, + "step": 578 + }, + { + "epoch": 0.15703824247355574, + "grad_norm": 19.471908569335938, + "learning_rate": 9.72182607973123e-07, + "loss": 0.6945, + "num_input_tokens_seen": 1214251008, + "step": 579 + }, + { + "epoch": 0.15730946569026308, + "grad_norm": 23.296817779541016, + "learning_rate": 9.720427061868558e-07, + "loss": 0.6556, + "num_input_tokens_seen": 1216348160, + "step": 580 + }, + { + "epoch": 0.15758068890697044, + "grad_norm": 11.207878112792969, + "learning_rate": 9.71902464756678e-07, + "loss": 0.3241, + "num_input_tokens_seen": 1218445312, + "step": 581 + }, + { + "epoch": 0.15785191212367777, + "grad_norm": 19.07700538635254, + "learning_rate": 9.717618837954517e-07, + "loss": 0.5486, + "num_input_tokens_seen": 1220542464, + "step": 582 + }, + { + "epoch": 0.15812313534038513, + "grad_norm": 13.274455070495605, + "learning_rate": 9.716209634163102e-07, + "loss": 0.5149, + "num_input_tokens_seen": 1222639616, + "step": 583 + }, + { + "epoch": 0.1583943585570925, + "grad_norm": 13.982866287231445, + "learning_rate": 9.714797037326616e-07, + "loss": 0.4192, + "num_input_tokens_seen": 1224736768, + "step": 584 + }, + { + "epoch": 0.15866558177379983, + "grad_norm": 17.50905990600586, + "learning_rate": 9.713381048581855e-07, + "loss": 0.7549, + "num_input_tokens_seen": 1226833920, + "step": 585 + }, + { + "epoch": 0.1589368049905072, + "grad_norm": 12.565037727355957, + "learning_rate": 9.71196166906836e-07, + "loss": 0.4246, + "num_input_tokens_seen": 1228931072, + "step": 586 + }, + { + "epoch": 0.15920802820721452, + "grad_norm": 14.987252235412598, + "learning_rate": 9.71053889992839e-07, + "loss": 0.6191, + "num_input_tokens_seen": 1231028224, + "step": 587 + }, + { + "epoch": 0.1594792514239219, + "grad_norm": 11.8529691696167, + "learning_rate": 9.709112742306936e-07, + "loss": 0.4308, + "num_input_tokens_seen": 1233125376, + "step": 588 + }, + { + "epoch": 0.15975047464062925, + "grad_norm": 11.237396240234375, + "learning_rate": 9.707683197351715e-07, + "loss": 0.375, + "num_input_tokens_seen": 1235222528, + "step": 589 + }, + { + "epoch": 0.16002169785733658, + "grad_norm": 11.192814826965332, + "learning_rate": 9.706250266213173e-07, + "loss": 0.2555, + "num_input_tokens_seen": 1237319680, + "step": 590 + }, + { + "epoch": 0.16029292107404394, + "grad_norm": 11.35294246673584, + "learning_rate": 9.704813950044476e-07, + "loss": 0.3542, + "num_input_tokens_seen": 1239416832, + "step": 591 + }, + { + "epoch": 0.16056414429075128, + "grad_norm": 17.085115432739258, + "learning_rate": 9.703374250001516e-07, + "loss": 0.7244, + "num_input_tokens_seen": 1241513984, + "step": 592 + }, + { + "epoch": 0.16083536750745864, + "grad_norm": 15.55121898651123, + "learning_rate": 9.70193116724291e-07, + "loss": 0.6381, + "num_input_tokens_seen": 1243611136, + "step": 593 + }, + { + "epoch": 0.161106590724166, + "grad_norm": 17.19528579711914, + "learning_rate": 9.700484702929996e-07, + "loss": 0.5695, + "num_input_tokens_seen": 1245708288, + "step": 594 + }, + { + "epoch": 0.16137781394087333, + "grad_norm": 14.350486755371094, + "learning_rate": 9.699034858226834e-07, + "loss": 0.5896, + "num_input_tokens_seen": 1247805440, + "step": 595 + }, + { + "epoch": 0.1616490371575807, + "grad_norm": 11.893250465393066, + "learning_rate": 9.697581634300202e-07, + "loss": 0.3935, + "num_input_tokens_seen": 1249902592, + "step": 596 + }, + { + "epoch": 0.16192026037428803, + "grad_norm": 17.247047424316406, + "learning_rate": 9.6961250323196e-07, + "loss": 0.8572, + "num_input_tokens_seen": 1251999744, + "step": 597 + }, + { + "epoch": 0.1621914835909954, + "grad_norm": 22.222002029418945, + "learning_rate": 9.69466505345725e-07, + "loss": 0.8848, + "num_input_tokens_seen": 1254096896, + "step": 598 + }, + { + "epoch": 0.16246270680770275, + "grad_norm": 17.794937133789062, + "learning_rate": 9.69320169888808e-07, + "loss": 0.7104, + "num_input_tokens_seen": 1256194048, + "step": 599 + }, + { + "epoch": 0.16273393002441008, + "grad_norm": 13.776861190795898, + "learning_rate": 9.691734969789746e-07, + "loss": 0.5435, + "num_input_tokens_seen": 1258291200, + "step": 600 + }, + { + "epoch": 0.16300515324111745, + "grad_norm": 15.114326477050781, + "learning_rate": 9.690264867342618e-07, + "loss": 0.4268, + "num_input_tokens_seen": 1260388352, + "step": 601 + }, + { + "epoch": 0.16327637645782478, + "grad_norm": 14.888233184814453, + "learning_rate": 9.688791392729775e-07, + "loss": 0.5672, + "num_input_tokens_seen": 1262485504, + "step": 602 + }, + { + "epoch": 0.16354759967453214, + "grad_norm": 12.982342720031738, + "learning_rate": 9.687314547137016e-07, + "loss": 0.4089, + "num_input_tokens_seen": 1264582656, + "step": 603 + }, + { + "epoch": 0.1638188228912395, + "grad_norm": 11.086604118347168, + "learning_rate": 9.685834331752846e-07, + "loss": 0.2905, + "num_input_tokens_seen": 1266679808, + "step": 604 + }, + { + "epoch": 0.16409004610794684, + "grad_norm": 20.390928268432617, + "learning_rate": 9.684350747768492e-07, + "loss": 0.9503, + "num_input_tokens_seen": 1268776960, + "step": 605 + }, + { + "epoch": 0.1643612693246542, + "grad_norm": 35.44593048095703, + "learning_rate": 9.68286379637788e-07, + "loss": 0.5335, + "num_input_tokens_seen": 1270874112, + "step": 606 + }, + { + "epoch": 0.16463249254136153, + "grad_norm": 11.213769912719727, + "learning_rate": 9.681373478777654e-07, + "loss": 0.3713, + "num_input_tokens_seen": 1272971264, + "step": 607 + }, + { + "epoch": 0.1649037157580689, + "grad_norm": 11.05787181854248, + "learning_rate": 9.679879796167166e-07, + "loss": 0.4049, + "num_input_tokens_seen": 1275068416, + "step": 608 + }, + { + "epoch": 0.16517493897477625, + "grad_norm": 15.489423751831055, + "learning_rate": 9.678382749748477e-07, + "loss": 0.5591, + "num_input_tokens_seen": 1277165568, + "step": 609 + }, + { + "epoch": 0.1654461621914836, + "grad_norm": 16.684755325317383, + "learning_rate": 9.676882340726345e-07, + "loss": 0.592, + "num_input_tokens_seen": 1279262720, + "step": 610 + }, + { + "epoch": 0.16571738540819095, + "grad_norm": 17.940027236938477, + "learning_rate": 9.675378570308253e-07, + "loss": 0.6475, + "num_input_tokens_seen": 1281359872, + "step": 611 + }, + { + "epoch": 0.16598860862489828, + "grad_norm": 11.783378601074219, + "learning_rate": 9.673871439704369e-07, + "loss": 0.3884, + "num_input_tokens_seen": 1283457024, + "step": 612 + }, + { + "epoch": 0.16625983184160564, + "grad_norm": 16.390104293823242, + "learning_rate": 9.672360950127578e-07, + "loss": 0.5195, + "num_input_tokens_seen": 1285554176, + "step": 613 + }, + { + "epoch": 0.166531055058313, + "grad_norm": 15.546297073364258, + "learning_rate": 9.670847102793464e-07, + "loss": 0.3534, + "num_input_tokens_seen": 1287651328, + "step": 614 + }, + { + "epoch": 0.16680227827502034, + "grad_norm": 15.699833869934082, + "learning_rate": 9.669329898920317e-07, + "loss": 0.5498, + "num_input_tokens_seen": 1289748480, + "step": 615 + }, + { + "epoch": 0.1670735014917277, + "grad_norm": 19.23993682861328, + "learning_rate": 9.66780933972912e-07, + "loss": 0.6282, + "num_input_tokens_seen": 1291845632, + "step": 616 + }, + { + "epoch": 0.16734472470843503, + "grad_norm": 9.65346908569336, + "learning_rate": 9.666285426443564e-07, + "loss": 0.2954, + "num_input_tokens_seen": 1293942784, + "step": 617 + }, + { + "epoch": 0.1676159479251424, + "grad_norm": 11.71711254119873, + "learning_rate": 9.664758160290036e-07, + "loss": 0.444, + "num_input_tokens_seen": 1296039936, + "step": 618 + }, + { + "epoch": 0.16788717114184976, + "grad_norm": 20.66088104248047, + "learning_rate": 9.66322754249762e-07, + "loss": 0.6955, + "num_input_tokens_seen": 1298137088, + "step": 619 + }, + { + "epoch": 0.1681583943585571, + "grad_norm": 20.292829513549805, + "learning_rate": 9.661693574298102e-07, + "loss": 0.8379, + "num_input_tokens_seen": 1300234240, + "step": 620 + }, + { + "epoch": 0.16842961757526445, + "grad_norm": 12.26496410369873, + "learning_rate": 9.66015625692596e-07, + "loss": 0.3796, + "num_input_tokens_seen": 1302331392, + "step": 621 + }, + { + "epoch": 0.16870084079197178, + "grad_norm": 13.123128890991211, + "learning_rate": 9.658615591618366e-07, + "loss": 0.2908, + "num_input_tokens_seen": 1304428544, + "step": 622 + }, + { + "epoch": 0.16897206400867915, + "grad_norm": 13.454907417297363, + "learning_rate": 9.657071579615191e-07, + "loss": 0.4747, + "num_input_tokens_seen": 1306525696, + "step": 623 + }, + { + "epoch": 0.16924328722538648, + "grad_norm": 17.779613494873047, + "learning_rate": 9.655524222159e-07, + "loss": 0.745, + "num_input_tokens_seen": 1308622848, + "step": 624 + }, + { + "epoch": 0.16951451044209384, + "grad_norm": 18.326885223388672, + "learning_rate": 9.653973520495042e-07, + "loss": 0.6889, + "num_input_tokens_seen": 1310720000, + "step": 625 + }, + { + "epoch": 0.1697857336588012, + "grad_norm": 15.276704788208008, + "learning_rate": 9.652419475871267e-07, + "loss": 0.4917, + "num_input_tokens_seen": 1312817152, + "step": 626 + }, + { + "epoch": 0.17005695687550854, + "grad_norm": 15.219459533691406, + "learning_rate": 9.650862089538307e-07, + "loss": 0.4999, + "num_input_tokens_seen": 1314914304, + "step": 627 + }, + { + "epoch": 0.1703281800922159, + "grad_norm": 16.05649757385254, + "learning_rate": 9.64930136274949e-07, + "loss": 0.5888, + "num_input_tokens_seen": 1317011456, + "step": 628 + }, + { + "epoch": 0.17059940330892323, + "grad_norm": 16.75012969970703, + "learning_rate": 9.647737296760828e-07, + "loss": 0.6463, + "num_input_tokens_seen": 1319108608, + "step": 629 + }, + { + "epoch": 0.1708706265256306, + "grad_norm": 14.331846237182617, + "learning_rate": 9.646169892831025e-07, + "loss": 0.4548, + "num_input_tokens_seen": 1321205760, + "step": 630 + }, + { + "epoch": 0.17114184974233795, + "grad_norm": 14.530692100524902, + "learning_rate": 9.644599152221465e-07, + "loss": 0.5452, + "num_input_tokens_seen": 1323302912, + "step": 631 + }, + { + "epoch": 0.1714130729590453, + "grad_norm": 13.921281814575195, + "learning_rate": 9.643025076196219e-07, + "loss": 0.599, + "num_input_tokens_seen": 1325400064, + "step": 632 + }, + { + "epoch": 0.17168429617575265, + "grad_norm": 14.689400672912598, + "learning_rate": 9.641447666022048e-07, + "loss": 0.5125, + "num_input_tokens_seen": 1327497216, + "step": 633 + }, + { + "epoch": 0.17195551939245998, + "grad_norm": 21.08767318725586, + "learning_rate": 9.639866922968387e-07, + "loss": 0.7141, + "num_input_tokens_seen": 1329594368, + "step": 634 + }, + { + "epoch": 0.17222674260916734, + "grad_norm": 21.691171646118164, + "learning_rate": 9.638282848307361e-07, + "loss": 0.7668, + "num_input_tokens_seen": 1331691520, + "step": 635 + }, + { + "epoch": 0.1724979658258747, + "grad_norm": 18.770170211791992, + "learning_rate": 9.636695443313773e-07, + "loss": 0.6828, + "num_input_tokens_seen": 1333788672, + "step": 636 + }, + { + "epoch": 0.17276918904258204, + "grad_norm": 18.908126831054688, + "learning_rate": 9.635104709265103e-07, + "loss": 0.8134, + "num_input_tokens_seen": 1335885824, + "step": 637 + }, + { + "epoch": 0.1730404122592894, + "grad_norm": 11.843361854553223, + "learning_rate": 9.633510647441518e-07, + "loss": 0.3847, + "num_input_tokens_seen": 1337982976, + "step": 638 + }, + { + "epoch": 0.17331163547599673, + "grad_norm": 19.700693130493164, + "learning_rate": 9.631913259125854e-07, + "loss": 0.9532, + "num_input_tokens_seen": 1340080128, + "step": 639 + }, + { + "epoch": 0.1735828586927041, + "grad_norm": 10.508553504943848, + "learning_rate": 9.630312545603631e-07, + "loss": 0.3257, + "num_input_tokens_seen": 1342177280, + "step": 640 + }, + { + "epoch": 0.17385408190941146, + "grad_norm": 19.769977569580078, + "learning_rate": 9.628708508163041e-07, + "loss": 0.7283, + "num_input_tokens_seen": 1344274432, + "step": 641 + }, + { + "epoch": 0.1741253051261188, + "grad_norm": 10.588939666748047, + "learning_rate": 9.627101148094952e-07, + "loss": 0.3603, + "num_input_tokens_seen": 1346371584, + "step": 642 + }, + { + "epoch": 0.17439652834282615, + "grad_norm": 16.81378936767578, + "learning_rate": 9.625490466692906e-07, + "loss": 0.7399, + "num_input_tokens_seen": 1348468736, + "step": 643 + }, + { + "epoch": 0.17466775155953349, + "grad_norm": 13.483939170837402, + "learning_rate": 9.623876465253122e-07, + "loss": 0.4786, + "num_input_tokens_seen": 1350565888, + "step": 644 + }, + { + "epoch": 0.17493897477624085, + "grad_norm": 16.27707290649414, + "learning_rate": 9.622259145074482e-07, + "loss": 0.4953, + "num_input_tokens_seen": 1352663040, + "step": 645 + }, + { + "epoch": 0.1752101979929482, + "grad_norm": 15.7435941696167, + "learning_rate": 9.620638507458547e-07, + "loss": 0.6285, + "num_input_tokens_seen": 1354760192, + "step": 646 + }, + { + "epoch": 0.17548142120965554, + "grad_norm": 20.336198806762695, + "learning_rate": 9.619014553709542e-07, + "loss": 0.8913, + "num_input_tokens_seen": 1356857344, + "step": 647 + }, + { + "epoch": 0.1757526444263629, + "grad_norm": 25.84923553466797, + "learning_rate": 9.617387285134364e-07, + "loss": 0.9168, + "num_input_tokens_seen": 1358954496, + "step": 648 + }, + { + "epoch": 0.17602386764307024, + "grad_norm": 12.170639991760254, + "learning_rate": 9.615756703042575e-07, + "loss": 0.4383, + "num_input_tokens_seen": 1361051648, + "step": 649 + }, + { + "epoch": 0.1762950908597776, + "grad_norm": 16.723989486694336, + "learning_rate": 9.61412280874641e-07, + "loss": 0.4632, + "num_input_tokens_seen": 1363148800, + "step": 650 + }, + { + "epoch": 0.17656631407648496, + "grad_norm": 16.861356735229492, + "learning_rate": 9.612485603560763e-07, + "loss": 0.8342, + "num_input_tokens_seen": 1365245952, + "step": 651 + }, + { + "epoch": 0.1768375372931923, + "grad_norm": 19.289173126220703, + "learning_rate": 9.610845088803194e-07, + "loss": 0.5916, + "num_input_tokens_seen": 1367343104, + "step": 652 + }, + { + "epoch": 0.17710876050989965, + "grad_norm": 17.8167781829834, + "learning_rate": 9.609201265793927e-07, + "loss": 0.5876, + "num_input_tokens_seen": 1369440256, + "step": 653 + }, + { + "epoch": 0.177379983726607, + "grad_norm": 13.8448486328125, + "learning_rate": 9.607554135855847e-07, + "loss": 0.524, + "num_input_tokens_seen": 1371537408, + "step": 654 + }, + { + "epoch": 0.17765120694331435, + "grad_norm": 14.017175674438477, + "learning_rate": 9.605903700314503e-07, + "loss": 0.4775, + "num_input_tokens_seen": 1373634560, + "step": 655 + }, + { + "epoch": 0.1779224301600217, + "grad_norm": 11.310769081115723, + "learning_rate": 9.604249960498102e-07, + "loss": 0.359, + "num_input_tokens_seen": 1375731712, + "step": 656 + }, + { + "epoch": 0.17819365337672904, + "grad_norm": 13.998455047607422, + "learning_rate": 9.602592917737512e-07, + "loss": 0.4158, + "num_input_tokens_seen": 1377828864, + "step": 657 + }, + { + "epoch": 0.1784648765934364, + "grad_norm": 21.207416534423828, + "learning_rate": 9.600932573366254e-07, + "loss": 0.8973, + "num_input_tokens_seen": 1379926016, + "step": 658 + }, + { + "epoch": 0.17873609981014374, + "grad_norm": 12.346809387207031, + "learning_rate": 9.599268928720518e-07, + "loss": 0.3974, + "num_input_tokens_seen": 1382023168, + "step": 659 + }, + { + "epoch": 0.1790073230268511, + "grad_norm": 16.258766174316406, + "learning_rate": 9.597601985139132e-07, + "loss": 0.3364, + "num_input_tokens_seen": 1384120320, + "step": 660 + }, + { + "epoch": 0.17927854624355846, + "grad_norm": 16.505359649658203, + "learning_rate": 9.595931743963596e-07, + "loss": 0.5018, + "num_input_tokens_seen": 1386217472, + "step": 661 + }, + { + "epoch": 0.1795497694602658, + "grad_norm": 15.472549438476562, + "learning_rate": 9.594258206538054e-07, + "loss": 0.4393, + "num_input_tokens_seen": 1388314624, + "step": 662 + }, + { + "epoch": 0.17982099267697316, + "grad_norm": 12.654108047485352, + "learning_rate": 9.592581374209306e-07, + "loss": 0.3907, + "num_input_tokens_seen": 1390411776, + "step": 663 + }, + { + "epoch": 0.1800922158936805, + "grad_norm": 15.311179161071777, + "learning_rate": 9.590901248326802e-07, + "loss": 0.5991, + "num_input_tokens_seen": 1392508928, + "step": 664 + }, + { + "epoch": 0.18036343911038785, + "grad_norm": 13.300477981567383, + "learning_rate": 9.589217830242645e-07, + "loss": 0.3853, + "num_input_tokens_seen": 1394606080, + "step": 665 + }, + { + "epoch": 0.18063466232709519, + "grad_norm": 14.14307975769043, + "learning_rate": 9.587531121311582e-07, + "loss": 0.3291, + "num_input_tokens_seen": 1396703232, + "step": 666 + }, + { + "epoch": 0.18090588554380255, + "grad_norm": 19.638458251953125, + "learning_rate": 9.585841122891016e-07, + "loss": 0.8024, + "num_input_tokens_seen": 1398800384, + "step": 667 + }, + { + "epoch": 0.1811771087605099, + "grad_norm": 18.304893493652344, + "learning_rate": 9.584147836340992e-07, + "loss": 0.6279, + "num_input_tokens_seen": 1400897536, + "step": 668 + }, + { + "epoch": 0.18144833197721724, + "grad_norm": 13.643768310546875, + "learning_rate": 9.582451263024202e-07, + "loss": 0.5066, + "num_input_tokens_seen": 1402994688, + "step": 669 + }, + { + "epoch": 0.1817195551939246, + "grad_norm": 17.128938674926758, + "learning_rate": 9.580751404305985e-07, + "loss": 0.7491, + "num_input_tokens_seen": 1405091840, + "step": 670 + }, + { + "epoch": 0.18199077841063194, + "grad_norm": 18.64154052734375, + "learning_rate": 9.579048261554321e-07, + "loss": 0.8889, + "num_input_tokens_seen": 1407188992, + "step": 671 + }, + { + "epoch": 0.1822620016273393, + "grad_norm": 14.319369316101074, + "learning_rate": 9.577341836139837e-07, + "loss": 0.4579, + "num_input_tokens_seen": 1409286144, + "step": 672 + }, + { + "epoch": 0.18253322484404666, + "grad_norm": 15.52761459350586, + "learning_rate": 9.575632129435796e-07, + "loss": 0.6158, + "num_input_tokens_seen": 1411383296, + "step": 673 + }, + { + "epoch": 0.182804448060754, + "grad_norm": 11.78895092010498, + "learning_rate": 9.573919142818109e-07, + "loss": 0.3947, + "num_input_tokens_seen": 1413480448, + "step": 674 + }, + { + "epoch": 0.18307567127746135, + "grad_norm": 17.20797348022461, + "learning_rate": 9.572202877665317e-07, + "loss": 0.6729, + "num_input_tokens_seen": 1415577600, + "step": 675 + }, + { + "epoch": 0.1833468944941687, + "grad_norm": 10.157732009887695, + "learning_rate": 9.57048333535861e-07, + "loss": 0.2969, + "num_input_tokens_seen": 1417674752, + "step": 676 + }, + { + "epoch": 0.18361811771087605, + "grad_norm": 20.86275863647461, + "learning_rate": 9.568760517281808e-07, + "loss": 0.9544, + "num_input_tokens_seen": 1419771904, + "step": 677 + }, + { + "epoch": 0.1838893409275834, + "grad_norm": 14.223817825317383, + "learning_rate": 9.56703442482137e-07, + "loss": 0.3624, + "num_input_tokens_seen": 1421869056, + "step": 678 + }, + { + "epoch": 0.18416056414429074, + "grad_norm": 16.383438110351562, + "learning_rate": 9.565305059366385e-07, + "loss": 0.8115, + "num_input_tokens_seen": 1423966208, + "step": 679 + }, + { + "epoch": 0.1844317873609981, + "grad_norm": 12.147848129272461, + "learning_rate": 9.563572422308588e-07, + "loss": 0.3638, + "num_input_tokens_seen": 1426063360, + "step": 680 + }, + { + "epoch": 0.18470301057770544, + "grad_norm": 15.152525901794434, + "learning_rate": 9.561836515042336e-07, + "loss": 0.5524, + "num_input_tokens_seen": 1428160512, + "step": 681 + }, + { + "epoch": 0.1849742337944128, + "grad_norm": 13.729914665222168, + "learning_rate": 9.56009733896462e-07, + "loss": 0.4642, + "num_input_tokens_seen": 1430257664, + "step": 682 + }, + { + "epoch": 0.18524545701112016, + "grad_norm": 16.161523818969727, + "learning_rate": 9.558354895475065e-07, + "loss": 0.6551, + "num_input_tokens_seen": 1432354816, + "step": 683 + }, + { + "epoch": 0.1855166802278275, + "grad_norm": 22.390674591064453, + "learning_rate": 9.55660918597592e-07, + "loss": 0.9867, + "num_input_tokens_seen": 1434451968, + "step": 684 + }, + { + "epoch": 0.18578790344453486, + "grad_norm": 17.674570083618164, + "learning_rate": 9.55486021187207e-07, + "loss": 0.6112, + "num_input_tokens_seen": 1436549120, + "step": 685 + }, + { + "epoch": 0.1860591266612422, + "grad_norm": 11.00310230255127, + "learning_rate": 9.553107974571018e-07, + "loss": 0.3261, + "num_input_tokens_seen": 1438646272, + "step": 686 + }, + { + "epoch": 0.18633034987794955, + "grad_norm": 18.365558624267578, + "learning_rate": 9.551352475482902e-07, + "loss": 0.7754, + "num_input_tokens_seen": 1440743424, + "step": 687 + }, + { + "epoch": 0.18660157309465691, + "grad_norm": 13.15377140045166, + "learning_rate": 9.549593716020478e-07, + "loss": 0.4411, + "num_input_tokens_seen": 1442840576, + "step": 688 + }, + { + "epoch": 0.18687279631136425, + "grad_norm": 11.530652046203613, + "learning_rate": 9.54783169759913e-07, + "loss": 0.4043, + "num_input_tokens_seen": 1444937728, + "step": 689 + }, + { + "epoch": 0.1871440195280716, + "grad_norm": 14.537562370300293, + "learning_rate": 9.546066421636867e-07, + "loss": 0.389, + "num_input_tokens_seen": 1447034880, + "step": 690 + }, + { + "epoch": 0.18741524274477894, + "grad_norm": 9.434741020202637, + "learning_rate": 9.54429788955431e-07, + "loss": 0.2483, + "num_input_tokens_seen": 1449132032, + "step": 691 + }, + { + "epoch": 0.1876864659614863, + "grad_norm": 15.039995193481445, + "learning_rate": 9.542526102774704e-07, + "loss": 0.5747, + "num_input_tokens_seen": 1451229184, + "step": 692 + }, + { + "epoch": 0.18795768917819367, + "grad_norm": 16.6597900390625, + "learning_rate": 9.540751062723923e-07, + "loss": 0.7014, + "num_input_tokens_seen": 1453326336, + "step": 693 + }, + { + "epoch": 0.188228912394901, + "grad_norm": 17.40535545349121, + "learning_rate": 9.53897277083045e-07, + "loss": 0.7324, + "num_input_tokens_seen": 1455423488, + "step": 694 + }, + { + "epoch": 0.18850013561160836, + "grad_norm": 14.920904159545898, + "learning_rate": 9.537191228525382e-07, + "loss": 0.4232, + "num_input_tokens_seen": 1457520640, + "step": 695 + }, + { + "epoch": 0.1887713588283157, + "grad_norm": 16.280458450317383, + "learning_rate": 9.535406437242444e-07, + "loss": 0.5262, + "num_input_tokens_seen": 1459617792, + "step": 696 + }, + { + "epoch": 0.18904258204502306, + "grad_norm": 13.120879173278809, + "learning_rate": 9.533618398417962e-07, + "loss": 0.534, + "num_input_tokens_seen": 1461714944, + "step": 697 + }, + { + "epoch": 0.18931380526173042, + "grad_norm": 16.71864891052246, + "learning_rate": 9.531827113490883e-07, + "loss": 0.6104, + "num_input_tokens_seen": 1463812096, + "step": 698 + }, + { + "epoch": 0.18958502847843775, + "grad_norm": 14.800333023071289, + "learning_rate": 9.530032583902767e-07, + "loss": 0.5911, + "num_input_tokens_seen": 1465909248, + "step": 699 + }, + { + "epoch": 0.1898562516951451, + "grad_norm": 15.830343246459961, + "learning_rate": 9.528234811097781e-07, + "loss": 0.6818, + "num_input_tokens_seen": 1468006400, + "step": 700 + }, + { + "epoch": 0.19012747491185245, + "grad_norm": 20.223512649536133, + "learning_rate": 9.526433796522702e-07, + "loss": 0.6553, + "num_input_tokens_seen": 1470103552, + "step": 701 + }, + { + "epoch": 0.1903986981285598, + "grad_norm": 12.725762367248535, + "learning_rate": 9.524629541626925e-07, + "loss": 0.4537, + "num_input_tokens_seen": 1472200704, + "step": 702 + }, + { + "epoch": 0.19066992134526717, + "grad_norm": 18.15487289428711, + "learning_rate": 9.522822047862438e-07, + "loss": 0.6253, + "num_input_tokens_seen": 1474297856, + "step": 703 + }, + { + "epoch": 0.1909411445619745, + "grad_norm": 17.46477699279785, + "learning_rate": 9.521011316683849e-07, + "loss": 0.6778, + "num_input_tokens_seen": 1476395008, + "step": 704 + }, + { + "epoch": 0.19121236777868186, + "grad_norm": 16.60508155822754, + "learning_rate": 9.519197349548364e-07, + "loss": 0.6406, + "num_input_tokens_seen": 1478492160, + "step": 705 + }, + { + "epoch": 0.1914835909953892, + "grad_norm": 14.466193199157715, + "learning_rate": 9.517380147915791e-07, + "loss": 0.4449, + "num_input_tokens_seen": 1480589312, + "step": 706 + }, + { + "epoch": 0.19175481421209656, + "grad_norm": 16.949872970581055, + "learning_rate": 9.515559713248549e-07, + "loss": 0.6814, + "num_input_tokens_seen": 1482686464, + "step": 707 + }, + { + "epoch": 0.1920260374288039, + "grad_norm": 9.922392845153809, + "learning_rate": 9.513736047011653e-07, + "loss": 0.3183, + "num_input_tokens_seen": 1484783616, + "step": 708 + }, + { + "epoch": 0.19229726064551125, + "grad_norm": 15.374136924743652, + "learning_rate": 9.511909150672721e-07, + "loss": 0.6533, + "num_input_tokens_seen": 1486880768, + "step": 709 + }, + { + "epoch": 0.19256848386221861, + "grad_norm": 16.63404083251953, + "learning_rate": 9.510079025701967e-07, + "loss": 0.5225, + "num_input_tokens_seen": 1488977920, + "step": 710 + }, + { + "epoch": 0.19283970707892595, + "grad_norm": 13.690201759338379, + "learning_rate": 9.508245673572209e-07, + "loss": 0.3678, + "num_input_tokens_seen": 1491075072, + "step": 711 + }, + { + "epoch": 0.1931109302956333, + "grad_norm": 12.991211891174316, + "learning_rate": 9.50640909575886e-07, + "loss": 0.352, + "num_input_tokens_seen": 1493172224, + "step": 712 + }, + { + "epoch": 0.19338215351234064, + "grad_norm": 18.690780639648438, + "learning_rate": 9.504569293739923e-07, + "loss": 0.7666, + "num_input_tokens_seen": 1495269376, + "step": 713 + }, + { + "epoch": 0.193653376729048, + "grad_norm": 15.247132301330566, + "learning_rate": 9.502726268996005e-07, + "loss": 0.5814, + "num_input_tokens_seen": 1497366528, + "step": 714 + }, + { + "epoch": 0.19392459994575537, + "grad_norm": 10.226189613342285, + "learning_rate": 9.5008800230103e-07, + "loss": 0.2609, + "num_input_tokens_seen": 1499463680, + "step": 715 + }, + { + "epoch": 0.1941958231624627, + "grad_norm": 19.02218246459961, + "learning_rate": 9.499030557268599e-07, + "loss": 0.6701, + "num_input_tokens_seen": 1501560832, + "step": 716 + }, + { + "epoch": 0.19446704637917006, + "grad_norm": 14.307744979858398, + "learning_rate": 9.497177873259279e-07, + "loss": 0.4341, + "num_input_tokens_seen": 1503657984, + "step": 717 + }, + { + "epoch": 0.1947382695958774, + "grad_norm": 15.487143516540527, + "learning_rate": 9.495321972473311e-07, + "loss": 0.3923, + "num_input_tokens_seen": 1505755136, + "step": 718 + }, + { + "epoch": 0.19500949281258476, + "grad_norm": 11.34209156036377, + "learning_rate": 9.493462856404251e-07, + "loss": 0.4535, + "num_input_tokens_seen": 1507852288, + "step": 719 + }, + { + "epoch": 0.19528071602929212, + "grad_norm": 13.599776268005371, + "learning_rate": 9.491600526548247e-07, + "loss": 0.4414, + "num_input_tokens_seen": 1509949440, + "step": 720 + }, + { + "epoch": 0.19555193924599945, + "grad_norm": 12.897449493408203, + "learning_rate": 9.489734984404033e-07, + "loss": 0.4073, + "num_input_tokens_seen": 1512046592, + "step": 721 + }, + { + "epoch": 0.1958231624627068, + "grad_norm": 16.54181480407715, + "learning_rate": 9.487866231472922e-07, + "loss": 0.5141, + "num_input_tokens_seen": 1514143744, + "step": 722 + }, + { + "epoch": 0.19609438567941415, + "grad_norm": 9.380181312561035, + "learning_rate": 9.48599426925882e-07, + "loss": 0.2837, + "num_input_tokens_seen": 1516240896, + "step": 723 + }, + { + "epoch": 0.1963656088961215, + "grad_norm": 17.064441680908203, + "learning_rate": 9.484119099268206e-07, + "loss": 0.6393, + "num_input_tokens_seen": 1518338048, + "step": 724 + }, + { + "epoch": 0.19663683211282887, + "grad_norm": 12.202215194702148, + "learning_rate": 9.482240723010148e-07, + "loss": 0.3783, + "num_input_tokens_seen": 1520435200, + "step": 725 + }, + { + "epoch": 0.1969080553295362, + "grad_norm": 20.838542938232422, + "learning_rate": 9.480359141996295e-07, + "loss": 0.7739, + "num_input_tokens_seen": 1522532352, + "step": 726 + }, + { + "epoch": 0.19717927854624356, + "grad_norm": 13.571410179138184, + "learning_rate": 9.478474357740864e-07, + "loss": 0.4288, + "num_input_tokens_seen": 1524629504, + "step": 727 + }, + { + "epoch": 0.1974505017629509, + "grad_norm": 13.46505355834961, + "learning_rate": 9.476586371760665e-07, + "loss": 0.4699, + "num_input_tokens_seen": 1526726656, + "step": 728 + }, + { + "epoch": 0.19772172497965826, + "grad_norm": 16.68222427368164, + "learning_rate": 9.474695185575072e-07, + "loss": 0.782, + "num_input_tokens_seen": 1528823808, + "step": 729 + }, + { + "epoch": 0.19799294819636562, + "grad_norm": 11.398141860961914, + "learning_rate": 9.472800800706044e-07, + "loss": 0.3531, + "num_input_tokens_seen": 1530920960, + "step": 730 + }, + { + "epoch": 0.19826417141307295, + "grad_norm": 17.724573135375977, + "learning_rate": 9.470903218678108e-07, + "loss": 0.8683, + "num_input_tokens_seen": 1533018112, + "step": 731 + }, + { + "epoch": 0.19853539462978032, + "grad_norm": 20.757295608520508, + "learning_rate": 9.469002441018366e-07, + "loss": 0.8129, + "num_input_tokens_seen": 1535115264, + "step": 732 + }, + { + "epoch": 0.19880661784648765, + "grad_norm": 16.623191833496094, + "learning_rate": 9.467098469256488e-07, + "loss": 0.5387, + "num_input_tokens_seen": 1537212416, + "step": 733 + }, + { + "epoch": 0.199077841063195, + "grad_norm": 11.821105003356934, + "learning_rate": 9.465191304924725e-07, + "loss": 0.3834, + "num_input_tokens_seen": 1539309568, + "step": 734 + }, + { + "epoch": 0.19934906427990237, + "grad_norm": 16.763580322265625, + "learning_rate": 9.463280949557885e-07, + "loss": 0.5618, + "num_input_tokens_seen": 1541406720, + "step": 735 + }, + { + "epoch": 0.1996202874966097, + "grad_norm": 14.250548362731934, + "learning_rate": 9.46136740469335e-07, + "loss": 0.5418, + "num_input_tokens_seen": 1543503872, + "step": 736 + }, + { + "epoch": 0.19989151071331707, + "grad_norm": 22.913976669311523, + "learning_rate": 9.45945067187107e-07, + "loss": 0.4692, + "num_input_tokens_seen": 1545601024, + "step": 737 + }, + { + "epoch": 0.2001627339300244, + "grad_norm": 15.227788925170898, + "learning_rate": 9.457530752633557e-07, + "loss": 0.522, + "num_input_tokens_seen": 1547698176, + "step": 738 + }, + { + "epoch": 0.20043395714673176, + "grad_norm": 12.005767822265625, + "learning_rate": 9.455607648525889e-07, + "loss": 0.2608, + "num_input_tokens_seen": 1549795328, + "step": 739 + }, + { + "epoch": 0.20070518036343912, + "grad_norm": 11.50250244140625, + "learning_rate": 9.45368136109571e-07, + "loss": 0.3001, + "num_input_tokens_seen": 1551892480, + "step": 740 + }, + { + "epoch": 0.20097640358014646, + "grad_norm": 19.086029052734375, + "learning_rate": 9.451751891893217e-07, + "loss": 0.6888, + "num_input_tokens_seen": 1553989632, + "step": 741 + }, + { + "epoch": 0.20124762679685382, + "grad_norm": 14.965021133422852, + "learning_rate": 9.449819242471179e-07, + "loss": 0.6411, + "num_input_tokens_seen": 1556086784, + "step": 742 + }, + { + "epoch": 0.20151885001356115, + "grad_norm": 12.905089378356934, + "learning_rate": 9.447883414384916e-07, + "loss": 0.4511, + "num_input_tokens_seen": 1558183936, + "step": 743 + }, + { + "epoch": 0.2017900732302685, + "grad_norm": 15.355229377746582, + "learning_rate": 9.445944409192308e-07, + "loss": 0.3277, + "num_input_tokens_seen": 1560281088, + "step": 744 + }, + { + "epoch": 0.20206129644697587, + "grad_norm": 12.57699203491211, + "learning_rate": 9.444002228453796e-07, + "loss": 0.3726, + "num_input_tokens_seen": 1562378240, + "step": 745 + }, + { + "epoch": 0.2023325196636832, + "grad_norm": 17.249740600585938, + "learning_rate": 9.442056873732369e-07, + "loss": 0.7487, + "num_input_tokens_seen": 1564475392, + "step": 746 + }, + { + "epoch": 0.20260374288039057, + "grad_norm": 19.854642868041992, + "learning_rate": 9.440108346593579e-07, + "loss": 0.6115, + "num_input_tokens_seen": 1566572544, + "step": 747 + }, + { + "epoch": 0.2028749660970979, + "grad_norm": 21.570199966430664, + "learning_rate": 9.438156648605521e-07, + "loss": 0.8311, + "num_input_tokens_seen": 1568669696, + "step": 748 + }, + { + "epoch": 0.20314618931380526, + "grad_norm": 14.75474739074707, + "learning_rate": 9.436201781338852e-07, + "loss": 0.4656, + "num_input_tokens_seen": 1570766848, + "step": 749 + }, + { + "epoch": 0.20341741253051263, + "grad_norm": 11.912012100219727, + "learning_rate": 9.434243746366771e-07, + "loss": 0.4637, + "num_input_tokens_seen": 1572864000, + "step": 750 + }, + { + "epoch": 0.20368863574721996, + "grad_norm": 18.71771240234375, + "learning_rate": 9.432282545265034e-07, + "loss": 0.5026, + "num_input_tokens_seen": 1574961152, + "step": 751 + }, + { + "epoch": 0.20395985896392732, + "grad_norm": 11.189618110656738, + "learning_rate": 9.430318179611938e-07, + "loss": 0.3925, + "num_input_tokens_seen": 1577058304, + "step": 752 + }, + { + "epoch": 0.20423108218063465, + "grad_norm": 15.765860557556152, + "learning_rate": 9.42835065098833e-07, + "loss": 0.6212, + "num_input_tokens_seen": 1579155456, + "step": 753 + }, + { + "epoch": 0.20450230539734202, + "grad_norm": 19.61184310913086, + "learning_rate": 9.426379960977605e-07, + "loss": 0.582, + "num_input_tokens_seen": 1581252608, + "step": 754 + }, + { + "epoch": 0.20477352861404935, + "grad_norm": 9.36821174621582, + "learning_rate": 9.424406111165697e-07, + "loss": 0.2286, + "num_input_tokens_seen": 1583349760, + "step": 755 + }, + { + "epoch": 0.2050447518307567, + "grad_norm": 25.251680374145508, + "learning_rate": 9.422429103141084e-07, + "loss": 1.0384, + "num_input_tokens_seen": 1585446912, + "step": 756 + }, + { + "epoch": 0.20531597504746407, + "grad_norm": 14.59071159362793, + "learning_rate": 9.42044893849479e-07, + "loss": 0.3857, + "num_input_tokens_seen": 1587544064, + "step": 757 + }, + { + "epoch": 0.2055871982641714, + "grad_norm": 16.961990356445312, + "learning_rate": 9.418465618820374e-07, + "loss": 0.6815, + "num_input_tokens_seen": 1589641216, + "step": 758 + }, + { + "epoch": 0.20585842148087877, + "grad_norm": 10.979452133178711, + "learning_rate": 9.416479145713936e-07, + "loss": 0.3136, + "num_input_tokens_seen": 1591738368, + "step": 759 + }, + { + "epoch": 0.2061296446975861, + "grad_norm": 14.783207893371582, + "learning_rate": 9.414489520774114e-07, + "loss": 0.5799, + "num_input_tokens_seen": 1593835520, + "step": 760 + }, + { + "epoch": 0.20640086791429346, + "grad_norm": 19.04448890686035, + "learning_rate": 9.412496745602084e-07, + "loss": 0.6763, + "num_input_tokens_seen": 1595932672, + "step": 761 + }, + { + "epoch": 0.20667209113100082, + "grad_norm": 24.7120418548584, + "learning_rate": 9.410500821801556e-07, + "loss": 0.6357, + "num_input_tokens_seen": 1598029824, + "step": 762 + }, + { + "epoch": 0.20694331434770816, + "grad_norm": 18.664295196533203, + "learning_rate": 9.408501750978769e-07, + "loss": 0.7047, + "num_input_tokens_seen": 1600126976, + "step": 763 + }, + { + "epoch": 0.20721453756441552, + "grad_norm": 12.567000389099121, + "learning_rate": 9.406499534742503e-07, + "loss": 0.3672, + "num_input_tokens_seen": 1602224128, + "step": 764 + }, + { + "epoch": 0.20748576078112285, + "grad_norm": 16.22792625427246, + "learning_rate": 9.404494174704068e-07, + "loss": 0.4399, + "num_input_tokens_seen": 1604321280, + "step": 765 + }, + { + "epoch": 0.2077569839978302, + "grad_norm": 11.791084289550781, + "learning_rate": 9.402485672477296e-07, + "loss": 0.3927, + "num_input_tokens_seen": 1606418432, + "step": 766 + }, + { + "epoch": 0.20802820721453757, + "grad_norm": 12.438189506530762, + "learning_rate": 9.400474029678555e-07, + "loss": 0.3697, + "num_input_tokens_seen": 1608515584, + "step": 767 + }, + { + "epoch": 0.2082994304312449, + "grad_norm": 14.400805473327637, + "learning_rate": 9.39845924792674e-07, + "loss": 0.5527, + "num_input_tokens_seen": 1610612736, + "step": 768 + }, + { + "epoch": 0.20857065364795227, + "grad_norm": 21.340023040771484, + "learning_rate": 9.396441328843268e-07, + "loss": 0.7031, + "num_input_tokens_seen": 1612709888, + "step": 769 + }, + { + "epoch": 0.2088418768646596, + "grad_norm": 18.95412254333496, + "learning_rate": 9.394420274052088e-07, + "loss": 0.6709, + "num_input_tokens_seen": 1614807040, + "step": 770 + }, + { + "epoch": 0.20911310008136696, + "grad_norm": 15.32834243774414, + "learning_rate": 9.392396085179662e-07, + "loss": 0.4891, + "num_input_tokens_seen": 1616904192, + "step": 771 + }, + { + "epoch": 0.20938432329807433, + "grad_norm": 14.922652244567871, + "learning_rate": 9.390368763854985e-07, + "loss": 0.5442, + "num_input_tokens_seen": 1619001344, + "step": 772 + }, + { + "epoch": 0.20965554651478166, + "grad_norm": 15.146223068237305, + "learning_rate": 9.388338311709566e-07, + "loss": 0.4359, + "num_input_tokens_seen": 1621098496, + "step": 773 + }, + { + "epoch": 0.20992676973148902, + "grad_norm": 18.45183753967285, + "learning_rate": 9.386304730377437e-07, + "loss": 0.7346, + "num_input_tokens_seen": 1623195648, + "step": 774 + }, + { + "epoch": 0.21019799294819635, + "grad_norm": 15.511632919311523, + "learning_rate": 9.384268021495145e-07, + "loss": 0.4655, + "num_input_tokens_seen": 1625292800, + "step": 775 + }, + { + "epoch": 0.21046921616490372, + "grad_norm": 12.674488067626953, + "learning_rate": 9.382228186701756e-07, + "loss": 0.4528, + "num_input_tokens_seen": 1627389952, + "step": 776 + }, + { + "epoch": 0.21074043938161108, + "grad_norm": 17.0650634765625, + "learning_rate": 9.380185227638854e-07, + "loss": 0.4335, + "num_input_tokens_seen": 1629487104, + "step": 777 + }, + { + "epoch": 0.2110116625983184, + "grad_norm": 13.060535430908203, + "learning_rate": 9.378139145950532e-07, + "loss": 0.4389, + "num_input_tokens_seen": 1631584256, + "step": 778 + }, + { + "epoch": 0.21128288581502577, + "grad_norm": 11.724872589111328, + "learning_rate": 9.376089943283398e-07, + "loss": 0.326, + "num_input_tokens_seen": 1633681408, + "step": 779 + }, + { + "epoch": 0.2115541090317331, + "grad_norm": 20.703407287597656, + "learning_rate": 9.374037621286574e-07, + "loss": 0.579, + "num_input_tokens_seen": 1635778560, + "step": 780 + }, + { + "epoch": 0.21182533224844047, + "grad_norm": 13.414408683776855, + "learning_rate": 9.371982181611692e-07, + "loss": 0.4883, + "num_input_tokens_seen": 1637875712, + "step": 781 + }, + { + "epoch": 0.21209655546514783, + "grad_norm": 15.019268989562988, + "learning_rate": 9.369923625912888e-07, + "loss": 0.495, + "num_input_tokens_seen": 1639972864, + "step": 782 + }, + { + "epoch": 0.21236777868185516, + "grad_norm": 11.590291976928711, + "learning_rate": 9.367861955846813e-07, + "loss": 0.3606, + "num_input_tokens_seen": 1642070016, + "step": 783 + }, + { + "epoch": 0.21263900189856252, + "grad_norm": 13.894659996032715, + "learning_rate": 9.365797173072619e-07, + "loss": 0.4129, + "num_input_tokens_seen": 1644167168, + "step": 784 + }, + { + "epoch": 0.21291022511526986, + "grad_norm": 17.303585052490234, + "learning_rate": 9.363729279251965e-07, + "loss": 0.634, + "num_input_tokens_seen": 1646264320, + "step": 785 + }, + { + "epoch": 0.21318144833197722, + "grad_norm": 15.712759017944336, + "learning_rate": 9.361658276049012e-07, + "loss": 0.6141, + "num_input_tokens_seen": 1648361472, + "step": 786 + }, + { + "epoch": 0.21345267154868458, + "grad_norm": 12.570011138916016, + "learning_rate": 9.359584165130426e-07, + "loss": 0.3697, + "num_input_tokens_seen": 1650458624, + "step": 787 + }, + { + "epoch": 0.21372389476539191, + "grad_norm": 17.42540740966797, + "learning_rate": 9.357506948165372e-07, + "loss": 0.6994, + "num_input_tokens_seen": 1652555776, + "step": 788 + }, + { + "epoch": 0.21399511798209928, + "grad_norm": 13.891875267028809, + "learning_rate": 9.355426626825516e-07, + "loss": 0.3642, + "num_input_tokens_seen": 1654652928, + "step": 789 + }, + { + "epoch": 0.2142663411988066, + "grad_norm": 14.662867546081543, + "learning_rate": 9.353343202785019e-07, + "loss": 0.4218, + "num_input_tokens_seen": 1656750080, + "step": 790 + }, + { + "epoch": 0.21453756441551397, + "grad_norm": 16.362796783447266, + "learning_rate": 9.351256677720542e-07, + "loss": 0.5427, + "num_input_tokens_seen": 1658847232, + "step": 791 + }, + { + "epoch": 0.21480878763222133, + "grad_norm": 16.475540161132812, + "learning_rate": 9.349167053311245e-07, + "loss": 0.6179, + "num_input_tokens_seen": 1660944384, + "step": 792 + }, + { + "epoch": 0.21508001084892867, + "grad_norm": 11.967279434204102, + "learning_rate": 9.347074331238774e-07, + "loss": 0.4276, + "num_input_tokens_seen": 1663041536, + "step": 793 + }, + { + "epoch": 0.21535123406563603, + "grad_norm": 10.945496559143066, + "learning_rate": 9.344978513187271e-07, + "loss": 0.2828, + "num_input_tokens_seen": 1665138688, + "step": 794 + }, + { + "epoch": 0.21562245728234336, + "grad_norm": 21.34172821044922, + "learning_rate": 9.342879600843376e-07, + "loss": 0.9658, + "num_input_tokens_seen": 1667235840, + "step": 795 + }, + { + "epoch": 0.21589368049905072, + "grad_norm": 13.481067657470703, + "learning_rate": 9.34077759589621e-07, + "loss": 0.4159, + "num_input_tokens_seen": 1669332992, + "step": 796 + }, + { + "epoch": 0.21616490371575806, + "grad_norm": 21.213205337524414, + "learning_rate": 9.338672500037387e-07, + "loss": 0.8269, + "num_input_tokens_seen": 1671430144, + "step": 797 + }, + { + "epoch": 0.21643612693246542, + "grad_norm": 12.533496856689453, + "learning_rate": 9.336564314961008e-07, + "loss": 0.4823, + "num_input_tokens_seen": 1673527296, + "step": 798 + }, + { + "epoch": 0.21670735014917278, + "grad_norm": 13.096843719482422, + "learning_rate": 9.334453042363661e-07, + "loss": 0.3558, + "num_input_tokens_seen": 1675624448, + "step": 799 + }, + { + "epoch": 0.2169785733658801, + "grad_norm": 21.356908798217773, + "learning_rate": 9.332338683944415e-07, + "loss": 0.8749, + "num_input_tokens_seen": 1677721600, + "step": 800 + }, + { + "epoch": 0.21724979658258747, + "grad_norm": 11.876173973083496, + "learning_rate": 9.33022124140483e-07, + "loss": 0.3332, + "num_input_tokens_seen": 1679818752, + "step": 801 + }, + { + "epoch": 0.2175210197992948, + "grad_norm": 10.453437805175781, + "learning_rate": 9.32810071644894e-07, + "loss": 0.3932, + "num_input_tokens_seen": 1681915904, + "step": 802 + }, + { + "epoch": 0.21779224301600217, + "grad_norm": 10.991945266723633, + "learning_rate": 9.325977110783263e-07, + "loss": 0.323, + "num_input_tokens_seen": 1684013056, + "step": 803 + }, + { + "epoch": 0.21806346623270953, + "grad_norm": 15.58512020111084, + "learning_rate": 9.323850426116797e-07, + "loss": 0.5063, + "num_input_tokens_seen": 1686110208, + "step": 804 + }, + { + "epoch": 0.21833468944941686, + "grad_norm": 18.722299575805664, + "learning_rate": 9.321720664161017e-07, + "loss": 0.5303, + "num_input_tokens_seen": 1688207360, + "step": 805 + }, + { + "epoch": 0.21860591266612422, + "grad_norm": 17.243942260742188, + "learning_rate": 9.319587826629872e-07, + "loss": 0.4814, + "num_input_tokens_seen": 1690304512, + "step": 806 + }, + { + "epoch": 0.21887713588283156, + "grad_norm": 14.880188941955566, + "learning_rate": 9.317451915239792e-07, + "loss": 0.6408, + "num_input_tokens_seen": 1692401664, + "step": 807 + }, + { + "epoch": 0.21914835909953892, + "grad_norm": 14.305965423583984, + "learning_rate": 9.315312931709674e-07, + "loss": 0.5187, + "num_input_tokens_seen": 1694498816, + "step": 808 + }, + { + "epoch": 0.21941958231624628, + "grad_norm": 18.705963134765625, + "learning_rate": 9.313170877760892e-07, + "loss": 0.6711, + "num_input_tokens_seen": 1696595968, + "step": 809 + }, + { + "epoch": 0.21969080553295361, + "grad_norm": 14.972651481628418, + "learning_rate": 9.311025755117291e-07, + "loss": 0.4127, + "num_input_tokens_seen": 1698693120, + "step": 810 + }, + { + "epoch": 0.21996202874966098, + "grad_norm": 35.52400207519531, + "learning_rate": 9.308877565505181e-07, + "loss": 0.7297, + "num_input_tokens_seen": 1700790272, + "step": 811 + }, + { + "epoch": 0.2202332519663683, + "grad_norm": 11.86293888092041, + "learning_rate": 9.306726310653346e-07, + "loss": 0.3399, + "num_input_tokens_seen": 1702887424, + "step": 812 + }, + { + "epoch": 0.22050447518307567, + "grad_norm": 15.077362060546875, + "learning_rate": 9.304571992293032e-07, + "loss": 0.574, + "num_input_tokens_seen": 1704984576, + "step": 813 + }, + { + "epoch": 0.22077569839978303, + "grad_norm": 24.58783531188965, + "learning_rate": 9.302414612157954e-07, + "loss": 0.935, + "num_input_tokens_seen": 1707081728, + "step": 814 + }, + { + "epoch": 0.22104692161649037, + "grad_norm": 18.85881996154785, + "learning_rate": 9.300254171984289e-07, + "loss": 0.5214, + "num_input_tokens_seen": 1709178880, + "step": 815 + }, + { + "epoch": 0.22131814483319773, + "grad_norm": 23.3087158203125, + "learning_rate": 9.298090673510677e-07, + "loss": 0.6932, + "num_input_tokens_seen": 1711276032, + "step": 816 + }, + { + "epoch": 0.22158936804990506, + "grad_norm": 15.919090270996094, + "learning_rate": 9.295924118478218e-07, + "loss": 0.5998, + "num_input_tokens_seen": 1713373184, + "step": 817 + }, + { + "epoch": 0.22186059126661242, + "grad_norm": 19.33287239074707, + "learning_rate": 9.293754508630473e-07, + "loss": 0.7676, + "num_input_tokens_seen": 1715470336, + "step": 818 + }, + { + "epoch": 0.22213181448331978, + "grad_norm": 13.031168937683105, + "learning_rate": 9.291581845713466e-07, + "loss": 0.4777, + "num_input_tokens_seen": 1717567488, + "step": 819 + }, + { + "epoch": 0.22240303770002712, + "grad_norm": 13.561074256896973, + "learning_rate": 9.289406131475665e-07, + "loss": 0.4002, + "num_input_tokens_seen": 1719664640, + "step": 820 + }, + { + "epoch": 0.22267426091673448, + "grad_norm": 16.580875396728516, + "learning_rate": 9.287227367668012e-07, + "loss": 0.6063, + "num_input_tokens_seen": 1721761792, + "step": 821 + }, + { + "epoch": 0.2229454841334418, + "grad_norm": 11.243143081665039, + "learning_rate": 9.285045556043885e-07, + "loss": 0.3497, + "num_input_tokens_seen": 1723858944, + "step": 822 + }, + { + "epoch": 0.22321670735014917, + "grad_norm": 16.73326873779297, + "learning_rate": 9.282860698359128e-07, + "loss": 0.7562, + "num_input_tokens_seen": 1725956096, + "step": 823 + }, + { + "epoch": 0.22348793056685654, + "grad_norm": 19.721914291381836, + "learning_rate": 9.280672796372029e-07, + "loss": 0.9825, + "num_input_tokens_seen": 1728053248, + "step": 824 + }, + { + "epoch": 0.22375915378356387, + "grad_norm": 18.111730575561523, + "learning_rate": 9.278481851843327e-07, + "loss": 0.7322, + "num_input_tokens_seen": 1730150400, + "step": 825 + }, + { + "epoch": 0.22403037700027123, + "grad_norm": 20.524099349975586, + "learning_rate": 9.276287866536215e-07, + "loss": 0.7254, + "num_input_tokens_seen": 1732247552, + "step": 826 + }, + { + "epoch": 0.22430160021697856, + "grad_norm": 13.87217903137207, + "learning_rate": 9.274090842216326e-07, + "loss": 0.3707, + "num_input_tokens_seen": 1734344704, + "step": 827 + }, + { + "epoch": 0.22457282343368593, + "grad_norm": 19.11530113220215, + "learning_rate": 9.271890780651741e-07, + "loss": 0.7017, + "num_input_tokens_seen": 1736441856, + "step": 828 + }, + { + "epoch": 0.2248440466503933, + "grad_norm": 9.669900894165039, + "learning_rate": 9.269687683612987e-07, + "loss": 0.2848, + "num_input_tokens_seen": 1738539008, + "step": 829 + }, + { + "epoch": 0.22511526986710062, + "grad_norm": 14.931427955627441, + "learning_rate": 9.267481552873033e-07, + "loss": 0.4288, + "num_input_tokens_seen": 1740636160, + "step": 830 + }, + { + "epoch": 0.22538649308380798, + "grad_norm": 16.54037094116211, + "learning_rate": 9.265272390207289e-07, + "loss": 0.5282, + "num_input_tokens_seen": 1742733312, + "step": 831 + }, + { + "epoch": 0.22565771630051532, + "grad_norm": 25.891746520996094, + "learning_rate": 9.263060197393603e-07, + "loss": 0.5665, + "num_input_tokens_seen": 1744830464, + "step": 832 + }, + { + "epoch": 0.22592893951722268, + "grad_norm": 18.841882705688477, + "learning_rate": 9.260844976212268e-07, + "loss": 0.8732, + "num_input_tokens_seen": 1746927616, + "step": 833 + }, + { + "epoch": 0.22620016273393004, + "grad_norm": 12.339317321777344, + "learning_rate": 9.258626728446004e-07, + "loss": 0.4302, + "num_input_tokens_seen": 1749024768, + "step": 834 + }, + { + "epoch": 0.22647138595063737, + "grad_norm": 18.42291831970215, + "learning_rate": 9.256405455879977e-07, + "loss": 0.7744, + "num_input_tokens_seen": 1751121920, + "step": 835 + }, + { + "epoch": 0.22674260916734473, + "grad_norm": 22.923080444335938, + "learning_rate": 9.25418116030178e-07, + "loss": 0.726, + "num_input_tokens_seen": 1753219072, + "step": 836 + }, + { + "epoch": 0.22701383238405207, + "grad_norm": 12.434412956237793, + "learning_rate": 9.251953843501443e-07, + "loss": 0.3722, + "num_input_tokens_seen": 1755316224, + "step": 837 + }, + { + "epoch": 0.22728505560075943, + "grad_norm": 18.6683292388916, + "learning_rate": 9.249723507271425e-07, + "loss": 0.5697, + "num_input_tokens_seen": 1757413376, + "step": 838 + }, + { + "epoch": 0.22755627881746676, + "grad_norm": 14.937889099121094, + "learning_rate": 9.247490153406617e-07, + "loss": 0.5072, + "num_input_tokens_seen": 1759510528, + "step": 839 + }, + { + "epoch": 0.22782750203417412, + "grad_norm": 12.447941780090332, + "learning_rate": 9.245253783704334e-07, + "loss": 0.3661, + "num_input_tokens_seen": 1761607680, + "step": 840 + }, + { + "epoch": 0.22809872525088148, + "grad_norm": 11.856642723083496, + "learning_rate": 9.243014399964324e-07, + "loss": 0.3809, + "num_input_tokens_seen": 1763704832, + "step": 841 + }, + { + "epoch": 0.22836994846758882, + "grad_norm": 10.702392578125, + "learning_rate": 9.240772003988758e-07, + "loss": 0.3321, + "num_input_tokens_seen": 1765801984, + "step": 842 + }, + { + "epoch": 0.22864117168429618, + "grad_norm": 13.526901245117188, + "learning_rate": 9.238526597582229e-07, + "loss": 0.5027, + "num_input_tokens_seen": 1767899136, + "step": 843 + }, + { + "epoch": 0.2289123949010035, + "grad_norm": 22.885881423950195, + "learning_rate": 9.236278182551758e-07, + "loss": 0.7197, + "num_input_tokens_seen": 1769996288, + "step": 844 + }, + { + "epoch": 0.22918361811771087, + "grad_norm": 20.40297508239746, + "learning_rate": 9.23402676070678e-07, + "loss": 0.5759, + "num_input_tokens_seen": 1772093440, + "step": 845 + }, + { + "epoch": 0.22945484133441824, + "grad_norm": 19.0419921875, + "learning_rate": 9.231772333859154e-07, + "loss": 0.6163, + "num_input_tokens_seen": 1774190592, + "step": 846 + }, + { + "epoch": 0.22972606455112557, + "grad_norm": 12.630133628845215, + "learning_rate": 9.22951490382316e-07, + "loss": 0.3334, + "num_input_tokens_seen": 1776287744, + "step": 847 + }, + { + "epoch": 0.22999728776783293, + "grad_norm": 11.405631065368652, + "learning_rate": 9.22725447241549e-07, + "loss": 0.3219, + "num_input_tokens_seen": 1778384896, + "step": 848 + }, + { + "epoch": 0.23026851098454026, + "grad_norm": 14.75662899017334, + "learning_rate": 9.224991041455252e-07, + "loss": 0.4205, + "num_input_tokens_seen": 1780482048, + "step": 849 + }, + { + "epoch": 0.23053973420124763, + "grad_norm": 16.446325302124023, + "learning_rate": 9.222724612763971e-07, + "loss": 0.6833, + "num_input_tokens_seen": 1782579200, + "step": 850 + }, + { + "epoch": 0.230810957417955, + "grad_norm": 14.62348747253418, + "learning_rate": 9.220455188165582e-07, + "loss": 0.5999, + "num_input_tokens_seen": 1784676352, + "step": 851 + }, + { + "epoch": 0.23108218063466232, + "grad_norm": 20.784975051879883, + "learning_rate": 9.218182769486433e-07, + "loss": 0.7337, + "num_input_tokens_seen": 1786773504, + "step": 852 + }, + { + "epoch": 0.23135340385136968, + "grad_norm": 16.6190128326416, + "learning_rate": 9.215907358555276e-07, + "loss": 0.6725, + "num_input_tokens_seen": 1788870656, + "step": 853 + }, + { + "epoch": 0.23162462706807702, + "grad_norm": 19.758882522583008, + "learning_rate": 9.213628957203277e-07, + "loss": 0.8521, + "num_input_tokens_seen": 1790967808, + "step": 854 + }, + { + "epoch": 0.23189585028478438, + "grad_norm": 21.532390594482422, + "learning_rate": 9.21134756726401e-07, + "loss": 0.7921, + "num_input_tokens_seen": 1793064960, + "step": 855 + }, + { + "epoch": 0.23216707350149174, + "grad_norm": 19.4443302154541, + "learning_rate": 9.209063190573445e-07, + "loss": 0.673, + "num_input_tokens_seen": 1795162112, + "step": 856 + }, + { + "epoch": 0.23243829671819907, + "grad_norm": 14.834383964538574, + "learning_rate": 9.206775828969967e-07, + "loss": 0.3932, + "num_input_tokens_seen": 1797259264, + "step": 857 + }, + { + "epoch": 0.23270951993490643, + "grad_norm": 12.486455917358398, + "learning_rate": 9.204485484294355e-07, + "loss": 0.37, + "num_input_tokens_seen": 1799356416, + "step": 858 + }, + { + "epoch": 0.23298074315161377, + "grad_norm": 15.540726661682129, + "learning_rate": 9.202192158389791e-07, + "loss": 0.6087, + "num_input_tokens_seen": 1801453568, + "step": 859 + }, + { + "epoch": 0.23325196636832113, + "grad_norm": 16.4538631439209, + "learning_rate": 9.199895853101856e-07, + "loss": 0.7186, + "num_input_tokens_seen": 1803550720, + "step": 860 + }, + { + "epoch": 0.2335231895850285, + "grad_norm": 18.737592697143555, + "learning_rate": 9.197596570278529e-07, + "loss": 0.7831, + "num_input_tokens_seen": 1805647872, + "step": 861 + }, + { + "epoch": 0.23379441280173582, + "grad_norm": 13.706647872924805, + "learning_rate": 9.19529431177019e-07, + "loss": 0.3793, + "num_input_tokens_seen": 1807745024, + "step": 862 + }, + { + "epoch": 0.23406563601844319, + "grad_norm": 11.725102424621582, + "learning_rate": 9.192989079429603e-07, + "loss": 0.1998, + "num_input_tokens_seen": 1809842176, + "step": 863 + }, + { + "epoch": 0.23433685923515052, + "grad_norm": 15.088400840759277, + "learning_rate": 9.190680875111934e-07, + "loss": 0.4601, + "num_input_tokens_seen": 1811939328, + "step": 864 + }, + { + "epoch": 0.23460808245185788, + "grad_norm": 14.075953483581543, + "learning_rate": 9.188369700674735e-07, + "loss": 0.4754, + "num_input_tokens_seen": 1814036480, + "step": 865 + }, + { + "epoch": 0.23487930566856524, + "grad_norm": 19.2708740234375, + "learning_rate": 9.186055557977957e-07, + "loss": 0.5167, + "num_input_tokens_seen": 1816133632, + "step": 866 + }, + { + "epoch": 0.23515052888527258, + "grad_norm": 18.750097274780273, + "learning_rate": 9.18373844888393e-07, + "loss": 0.7142, + "num_input_tokens_seen": 1818230784, + "step": 867 + }, + { + "epoch": 0.23542175210197994, + "grad_norm": 13.886237144470215, + "learning_rate": 9.181418375257374e-07, + "loss": 0.4226, + "num_input_tokens_seen": 1820327936, + "step": 868 + }, + { + "epoch": 0.23569297531868727, + "grad_norm": 18.76662254333496, + "learning_rate": 9.179095338965401e-07, + "loss": 0.6679, + "num_input_tokens_seen": 1822425088, + "step": 869 + }, + { + "epoch": 0.23596419853539463, + "grad_norm": 13.589776039123535, + "learning_rate": 9.176769341877497e-07, + "loss": 0.4632, + "num_input_tokens_seen": 1824522240, + "step": 870 + }, + { + "epoch": 0.236235421752102, + "grad_norm": 16.83132553100586, + "learning_rate": 9.17444038586554e-07, + "loss": 0.5446, + "num_input_tokens_seen": 1826619392, + "step": 871 + }, + { + "epoch": 0.23650664496880933, + "grad_norm": 26.304426193237305, + "learning_rate": 9.172108472803782e-07, + "loss": 0.6821, + "num_input_tokens_seen": 1828716544, + "step": 872 + }, + { + "epoch": 0.2367778681855167, + "grad_norm": 18.7716121673584, + "learning_rate": 9.16977360456886e-07, + "loss": 0.4784, + "num_input_tokens_seen": 1830813696, + "step": 873 + }, + { + "epoch": 0.23704909140222402, + "grad_norm": 11.739683151245117, + "learning_rate": 9.167435783039786e-07, + "loss": 0.3455, + "num_input_tokens_seen": 1832910848, + "step": 874 + }, + { + "epoch": 0.23732031461893138, + "grad_norm": 14.229680061340332, + "learning_rate": 9.165095010097949e-07, + "loss": 0.6126, + "num_input_tokens_seen": 1835008000, + "step": 875 + }, + { + "epoch": 0.23759153783563874, + "grad_norm": 15.295489311218262, + "learning_rate": 9.162751287627116e-07, + "loss": 0.6215, + "num_input_tokens_seen": 1837105152, + "step": 876 + }, + { + "epoch": 0.23786276105234608, + "grad_norm": 13.08266830444336, + "learning_rate": 9.160404617513424e-07, + "loss": 0.4853, + "num_input_tokens_seen": 1839202304, + "step": 877 + }, + { + "epoch": 0.23813398426905344, + "grad_norm": 18.74358558654785, + "learning_rate": 9.158055001645385e-07, + "loss": 0.7992, + "num_input_tokens_seen": 1841299456, + "step": 878 + }, + { + "epoch": 0.23840520748576077, + "grad_norm": 14.171740531921387, + "learning_rate": 9.155702441913881e-07, + "loss": 0.4748, + "num_input_tokens_seen": 1843396608, + "step": 879 + }, + { + "epoch": 0.23867643070246813, + "grad_norm": 15.582724571228027, + "learning_rate": 9.15334694021216e-07, + "loss": 0.5375, + "num_input_tokens_seen": 1845493760, + "step": 880 + }, + { + "epoch": 0.23894765391917547, + "grad_norm": 16.704227447509766, + "learning_rate": 9.150988498435843e-07, + "loss": 0.5576, + "num_input_tokens_seen": 1847590912, + "step": 881 + }, + { + "epoch": 0.23921887713588283, + "grad_norm": 12.607769966125488, + "learning_rate": 9.148627118482912e-07, + "loss": 0.287, + "num_input_tokens_seen": 1849688064, + "step": 882 + }, + { + "epoch": 0.2394901003525902, + "grad_norm": 16.634292602539062, + "learning_rate": 9.146262802253717e-07, + "loss": 0.752, + "num_input_tokens_seen": 1851785216, + "step": 883 + }, + { + "epoch": 0.23976132356929752, + "grad_norm": 13.200092315673828, + "learning_rate": 9.14389555165097e-07, + "loss": 0.3663, + "num_input_tokens_seen": 1853882368, + "step": 884 + }, + { + "epoch": 0.24003254678600489, + "grad_norm": 15.857352256774902, + "learning_rate": 9.141525368579742e-07, + "loss": 0.507, + "num_input_tokens_seen": 1855979520, + "step": 885 + }, + { + "epoch": 0.24030377000271222, + "grad_norm": 15.330340385437012, + "learning_rate": 9.139152254947469e-07, + "loss": 0.4418, + "num_input_tokens_seen": 1858076672, + "step": 886 + }, + { + "epoch": 0.24057499321941958, + "grad_norm": 12.231500625610352, + "learning_rate": 9.136776212663942e-07, + "loss": 0.4606, + "num_input_tokens_seen": 1860173824, + "step": 887 + }, + { + "epoch": 0.24084621643612694, + "grad_norm": 19.30156898498535, + "learning_rate": 9.134397243641307e-07, + "loss": 0.6261, + "num_input_tokens_seen": 1862270976, + "step": 888 + }, + { + "epoch": 0.24111743965283428, + "grad_norm": 23.425487518310547, + "learning_rate": 9.132015349794069e-07, + "loss": 0.4225, + "num_input_tokens_seen": 1864368128, + "step": 889 + }, + { + "epoch": 0.24138866286954164, + "grad_norm": 9.727029800415039, + "learning_rate": 9.129630533039086e-07, + "loss": 0.2508, + "num_input_tokens_seen": 1866465280, + "step": 890 + }, + { + "epoch": 0.24165988608624897, + "grad_norm": 20.487241744995117, + "learning_rate": 9.127242795295569e-07, + "loss": 0.7279, + "num_input_tokens_seen": 1868562432, + "step": 891 + }, + { + "epoch": 0.24193110930295633, + "grad_norm": 16.180301666259766, + "learning_rate": 9.124852138485076e-07, + "loss": 0.5699, + "num_input_tokens_seen": 1870659584, + "step": 892 + }, + { + "epoch": 0.2422023325196637, + "grad_norm": 20.350139617919922, + "learning_rate": 9.12245856453152e-07, + "loss": 0.856, + "num_input_tokens_seen": 1872756736, + "step": 893 + }, + { + "epoch": 0.24247355573637103, + "grad_norm": 18.013940811157227, + "learning_rate": 9.120062075361155e-07, + "loss": 0.6445, + "num_input_tokens_seen": 1874853888, + "step": 894 + }, + { + "epoch": 0.2427447789530784, + "grad_norm": 16.004846572875977, + "learning_rate": 9.117662672902584e-07, + "loss": 0.6282, + "num_input_tokens_seen": 1876951040, + "step": 895 + }, + { + "epoch": 0.24301600216978572, + "grad_norm": 14.721831321716309, + "learning_rate": 9.115260359086757e-07, + "loss": 0.3772, + "num_input_tokens_seen": 1879048192, + "step": 896 + }, + { + "epoch": 0.24328722538649308, + "grad_norm": 19.249744415283203, + "learning_rate": 9.112855135846964e-07, + "loss": 0.5968, + "num_input_tokens_seen": 1881145344, + "step": 897 + }, + { + "epoch": 0.24355844860320044, + "grad_norm": 12.756824493408203, + "learning_rate": 9.110447005118836e-07, + "loss": 0.4154, + "num_input_tokens_seen": 1883242496, + "step": 898 + }, + { + "epoch": 0.24382967181990778, + "grad_norm": 11.557883262634277, + "learning_rate": 9.108035968840348e-07, + "loss": 0.3552, + "num_input_tokens_seen": 1885339648, + "step": 899 + }, + { + "epoch": 0.24410089503661514, + "grad_norm": 19.659439086914062, + "learning_rate": 9.105622028951806e-07, + "loss": 0.6714, + "num_input_tokens_seen": 1887436800, + "step": 900 + }, + { + "epoch": 0.24437211825332247, + "grad_norm": 15.646737098693848, + "learning_rate": 9.103205187395861e-07, + "loss": 0.367, + "num_input_tokens_seen": 1889533952, + "step": 901 + }, + { + "epoch": 0.24464334147002983, + "grad_norm": 13.878365516662598, + "learning_rate": 9.100785446117493e-07, + "loss": 0.4032, + "num_input_tokens_seen": 1891631104, + "step": 902 + }, + { + "epoch": 0.2449145646867372, + "grad_norm": 16.992267608642578, + "learning_rate": 9.098362807064017e-07, + "loss": 0.5326, + "num_input_tokens_seen": 1893728256, + "step": 903 + }, + { + "epoch": 0.24518578790344453, + "grad_norm": 12.592440605163574, + "learning_rate": 9.095937272185083e-07, + "loss": 0.3813, + "num_input_tokens_seen": 1895825408, + "step": 904 + }, + { + "epoch": 0.2454570111201519, + "grad_norm": 14.522198677062988, + "learning_rate": 9.093508843432667e-07, + "loss": 0.449, + "num_input_tokens_seen": 1897922560, + "step": 905 + }, + { + "epoch": 0.24572823433685922, + "grad_norm": 19.913555145263672, + "learning_rate": 9.091077522761078e-07, + "loss": 0.6901, + "num_input_tokens_seen": 1900019712, + "step": 906 + }, + { + "epoch": 0.2459994575535666, + "grad_norm": 18.202465057373047, + "learning_rate": 9.088643312126948e-07, + "loss": 0.7029, + "num_input_tokens_seen": 1902116864, + "step": 907 + }, + { + "epoch": 0.24627068077027395, + "grad_norm": 13.385407447814941, + "learning_rate": 9.086206213489239e-07, + "loss": 0.4506, + "num_input_tokens_seen": 1904214016, + "step": 908 + }, + { + "epoch": 0.24654190398698128, + "grad_norm": 17.774667739868164, + "learning_rate": 9.083766228809234e-07, + "loss": 0.6725, + "num_input_tokens_seen": 1906311168, + "step": 909 + }, + { + "epoch": 0.24681312720368864, + "grad_norm": 14.268959999084473, + "learning_rate": 9.081323360050543e-07, + "loss": 0.4936, + "num_input_tokens_seen": 1908408320, + "step": 910 + }, + { + "epoch": 0.24708435042039598, + "grad_norm": 15.435829162597656, + "learning_rate": 9.078877609179088e-07, + "loss": 0.4626, + "num_input_tokens_seen": 1910505472, + "step": 911 + }, + { + "epoch": 0.24735557363710334, + "grad_norm": 11.141834259033203, + "learning_rate": 9.076428978163121e-07, + "loss": 0.337, + "num_input_tokens_seen": 1912602624, + "step": 912 + }, + { + "epoch": 0.2476267968538107, + "grad_norm": 21.919893264770508, + "learning_rate": 9.073977468973206e-07, + "loss": 0.993, + "num_input_tokens_seen": 1914699776, + "step": 913 + }, + { + "epoch": 0.24789802007051803, + "grad_norm": 15.084315299987793, + "learning_rate": 9.071523083582223e-07, + "loss": 0.3735, + "num_input_tokens_seen": 1916796928, + "step": 914 + }, + { + "epoch": 0.2481692432872254, + "grad_norm": 19.048635482788086, + "learning_rate": 9.06906582396537e-07, + "loss": 0.7275, + "num_input_tokens_seen": 1918894080, + "step": 915 + }, + { + "epoch": 0.24844046650393273, + "grad_norm": 11.731395721435547, + "learning_rate": 9.066605692100155e-07, + "loss": 0.388, + "num_input_tokens_seen": 1920991232, + "step": 916 + }, + { + "epoch": 0.2487116897206401, + "grad_norm": 11.196650505065918, + "learning_rate": 9.064142689966397e-07, + "loss": 0.3481, + "num_input_tokens_seen": 1923088384, + "step": 917 + }, + { + "epoch": 0.24898291293734745, + "grad_norm": 21.21782684326172, + "learning_rate": 9.061676819546229e-07, + "loss": 0.863, + "num_input_tokens_seen": 1925185536, + "step": 918 + }, + { + "epoch": 0.24925413615405478, + "grad_norm": 14.912806510925293, + "learning_rate": 9.059208082824087e-07, + "loss": 0.5098, + "num_input_tokens_seen": 1927282688, + "step": 919 + }, + { + "epoch": 0.24952535937076215, + "grad_norm": 18.304712295532227, + "learning_rate": 9.05673648178672e-07, + "loss": 0.6095, + "num_input_tokens_seen": 1929379840, + "step": 920 + }, + { + "epoch": 0.24979658258746948, + "grad_norm": 11.65080738067627, + "learning_rate": 9.054262018423175e-07, + "loss": 0.3207, + "num_input_tokens_seen": 1931476992, + "step": 921 + }, + { + "epoch": 0.2500678058041768, + "grad_norm": 21.233457565307617, + "learning_rate": 9.051784694724808e-07, + "loss": 0.8062, + "num_input_tokens_seen": 1933574144, + "step": 922 + }, + { + "epoch": 0.2503390290208842, + "grad_norm": 9.506631851196289, + "learning_rate": 9.049304512685274e-07, + "loss": 0.2804, + "num_input_tokens_seen": 1935671296, + "step": 923 + }, + { + "epoch": 0.25061025223759154, + "grad_norm": 10.630436897277832, + "learning_rate": 9.046821474300527e-07, + "loss": 0.3033, + "num_input_tokens_seen": 1937768448, + "step": 924 + }, + { + "epoch": 0.2508814754542989, + "grad_norm": 18.508195877075195, + "learning_rate": 9.044335581568827e-07, + "loss": 0.6053, + "num_input_tokens_seen": 1939865600, + "step": 925 + }, + { + "epoch": 0.25115269867100626, + "grad_norm": 11.025355339050293, + "learning_rate": 9.041846836490723e-07, + "loss": 0.3725, + "num_input_tokens_seen": 1941962752, + "step": 926 + }, + { + "epoch": 0.25142392188771356, + "grad_norm": 13.680339813232422, + "learning_rate": 9.03935524106906e-07, + "loss": 0.4114, + "num_input_tokens_seen": 1944059904, + "step": 927 + }, + { + "epoch": 0.2516951451044209, + "grad_norm": 10.082228660583496, + "learning_rate": 9.036860797308984e-07, + "loss": 0.2863, + "num_input_tokens_seen": 1946157056, + "step": 928 + }, + { + "epoch": 0.2519663683211283, + "grad_norm": 15.916239738464355, + "learning_rate": 9.034363507217925e-07, + "loss": 0.452, + "num_input_tokens_seen": 1948254208, + "step": 929 + }, + { + "epoch": 0.25223759153783565, + "grad_norm": 20.79644012451172, + "learning_rate": 9.031863372805606e-07, + "loss": 0.6497, + "num_input_tokens_seen": 1950351360, + "step": 930 + }, + { + "epoch": 0.252508814754543, + "grad_norm": 15.72608757019043, + "learning_rate": 9.029360396084043e-07, + "loss": 0.3006, + "num_input_tokens_seen": 1952448512, + "step": 931 + }, + { + "epoch": 0.2527800379712503, + "grad_norm": 13.770834922790527, + "learning_rate": 9.026854579067537e-07, + "loss": 0.5086, + "num_input_tokens_seen": 1954545664, + "step": 932 + }, + { + "epoch": 0.2530512611879577, + "grad_norm": 18.707565307617188, + "learning_rate": 9.024345923772671e-07, + "loss": 0.513, + "num_input_tokens_seen": 1956642816, + "step": 933 + }, + { + "epoch": 0.25332248440466504, + "grad_norm": 11.008857727050781, + "learning_rate": 9.021834432218317e-07, + "loss": 0.336, + "num_input_tokens_seen": 1958739968, + "step": 934 + }, + { + "epoch": 0.2535937076213724, + "grad_norm": 13.265803337097168, + "learning_rate": 9.019320106425629e-07, + "loss": 0.3631, + "num_input_tokens_seen": 1960837120, + "step": 935 + }, + { + "epoch": 0.25386493083807976, + "grad_norm": 11.919256210327148, + "learning_rate": 9.016802948418038e-07, + "loss": 0.2662, + "num_input_tokens_seen": 1962934272, + "step": 936 + }, + { + "epoch": 0.25413615405478707, + "grad_norm": 13.845582962036133, + "learning_rate": 9.014282960221257e-07, + "loss": 0.5253, + "num_input_tokens_seen": 1965031424, + "step": 937 + }, + { + "epoch": 0.25440737727149443, + "grad_norm": 14.255606651306152, + "learning_rate": 9.01176014386328e-07, + "loss": 0.444, + "num_input_tokens_seen": 1967128576, + "step": 938 + }, + { + "epoch": 0.2546786004882018, + "grad_norm": 19.263900756835938, + "learning_rate": 9.009234501374371e-07, + "loss": 0.6345, + "num_input_tokens_seen": 1969225728, + "step": 939 + }, + { + "epoch": 0.25494982370490915, + "grad_norm": 21.17646026611328, + "learning_rate": 9.006706034787071e-07, + "loss": 0.5709, + "num_input_tokens_seen": 1971322880, + "step": 940 + }, + { + "epoch": 0.2552210469216165, + "grad_norm": 16.962255477905273, + "learning_rate": 9.004174746136196e-07, + "loss": 0.6367, + "num_input_tokens_seen": 1973420032, + "step": 941 + }, + { + "epoch": 0.2554922701383238, + "grad_norm": 12.953526496887207, + "learning_rate": 9.001640637458829e-07, + "loss": 0.4449, + "num_input_tokens_seen": 1975517184, + "step": 942 + }, + { + "epoch": 0.2557634933550312, + "grad_norm": 19.991397857666016, + "learning_rate": 8.999103710794323e-07, + "loss": 0.9314, + "num_input_tokens_seen": 1977614336, + "step": 943 + }, + { + "epoch": 0.25603471657173854, + "grad_norm": 12.952409744262695, + "learning_rate": 8.996563968184302e-07, + "loss": 0.3151, + "num_input_tokens_seen": 1979711488, + "step": 944 + }, + { + "epoch": 0.2563059397884459, + "grad_norm": 18.413883209228516, + "learning_rate": 8.994021411672653e-07, + "loss": 0.7246, + "num_input_tokens_seen": 1981808640, + "step": 945 + }, + { + "epoch": 0.25657716300515326, + "grad_norm": 15.939887046813965, + "learning_rate": 8.99147604330553e-07, + "loss": 0.5387, + "num_input_tokens_seen": 1983905792, + "step": 946 + }, + { + "epoch": 0.25684838622186057, + "grad_norm": 18.361181259155273, + "learning_rate": 8.988927865131347e-07, + "loss": 0.5397, + "num_input_tokens_seen": 1986002944, + "step": 947 + }, + { + "epoch": 0.25711960943856793, + "grad_norm": 16.950855255126953, + "learning_rate": 8.986376879200783e-07, + "loss": 0.5404, + "num_input_tokens_seen": 1988100096, + "step": 948 + }, + { + "epoch": 0.2573908326552753, + "grad_norm": 13.930158615112305, + "learning_rate": 8.983823087566772e-07, + "loss": 0.3202, + "num_input_tokens_seen": 1990197248, + "step": 949 + }, + { + "epoch": 0.25766205587198265, + "grad_norm": 18.551612854003906, + "learning_rate": 8.981266492284511e-07, + "loss": 0.5814, + "num_input_tokens_seen": 1992294400, + "step": 950 + }, + { + "epoch": 0.25793327908869, + "grad_norm": 17.812480926513672, + "learning_rate": 8.978707095411446e-07, + "loss": 0.6843, + "num_input_tokens_seen": 1994391552, + "step": 951 + }, + { + "epoch": 0.2582045023053973, + "grad_norm": 22.378063201904297, + "learning_rate": 8.976144899007288e-07, + "loss": 1.0731, + "num_input_tokens_seen": 1996488704, + "step": 952 + }, + { + "epoch": 0.2584757255221047, + "grad_norm": 12.730658531188965, + "learning_rate": 8.973579905133991e-07, + "loss": 0.4466, + "num_input_tokens_seen": 1998585856, + "step": 953 + }, + { + "epoch": 0.25874694873881204, + "grad_norm": 9.290279388427734, + "learning_rate": 8.971012115855766e-07, + "loss": 0.2807, + "num_input_tokens_seen": 2000683008, + "step": 954 + }, + { + "epoch": 0.2590181719555194, + "grad_norm": 14.961581230163574, + "learning_rate": 8.968441533239073e-07, + "loss": 0.5023, + "num_input_tokens_seen": 2002780160, + "step": 955 + }, + { + "epoch": 0.25928939517222677, + "grad_norm": 11.520310401916504, + "learning_rate": 8.965868159352616e-07, + "loss": 0.3769, + "num_input_tokens_seen": 2004877312, + "step": 956 + }, + { + "epoch": 0.2595606183889341, + "grad_norm": 13.574871063232422, + "learning_rate": 8.963291996267354e-07, + "loss": 0.4165, + "num_input_tokens_seen": 2006974464, + "step": 957 + }, + { + "epoch": 0.25983184160564143, + "grad_norm": 10.172213554382324, + "learning_rate": 8.960713046056478e-07, + "loss": 0.3482, + "num_input_tokens_seen": 2009071616, + "step": 958 + }, + { + "epoch": 0.2601030648223488, + "grad_norm": 15.749048233032227, + "learning_rate": 8.958131310795434e-07, + "loss": 0.4703, + "num_input_tokens_seen": 2011168768, + "step": 959 + }, + { + "epoch": 0.26037428803905616, + "grad_norm": 13.04376220703125, + "learning_rate": 8.955546792561902e-07, + "loss": 0.3705, + "num_input_tokens_seen": 2013265920, + "step": 960 + }, + { + "epoch": 0.2606455112557635, + "grad_norm": 13.472237586975098, + "learning_rate": 8.952959493435806e-07, + "loss": 0.4021, + "num_input_tokens_seen": 2015363072, + "step": 961 + }, + { + "epoch": 0.2609167344724708, + "grad_norm": 15.37488079071045, + "learning_rate": 8.950369415499304e-07, + "loss": 0.5569, + "num_input_tokens_seen": 2017460224, + "step": 962 + }, + { + "epoch": 0.2611879576891782, + "grad_norm": 18.64729881286621, + "learning_rate": 8.947776560836793e-07, + "loss": 0.7467, + "num_input_tokens_seen": 2019557376, + "step": 963 + }, + { + "epoch": 0.26145918090588555, + "grad_norm": 13.231241226196289, + "learning_rate": 8.945180931534902e-07, + "loss": 0.4492, + "num_input_tokens_seen": 2021654528, + "step": 964 + }, + { + "epoch": 0.2617304041225929, + "grad_norm": 13.122672080993652, + "learning_rate": 8.942582529682496e-07, + "loss": 0.387, + "num_input_tokens_seen": 2023751680, + "step": 965 + }, + { + "epoch": 0.26200162733930027, + "grad_norm": 12.181504249572754, + "learning_rate": 8.939981357370672e-07, + "loss": 0.3468, + "num_input_tokens_seen": 2025848832, + "step": 966 + }, + { + "epoch": 0.2622728505560076, + "grad_norm": 13.75951862335205, + "learning_rate": 8.937377416692752e-07, + "loss": 0.4132, + "num_input_tokens_seen": 2027945984, + "step": 967 + }, + { + "epoch": 0.26254407377271494, + "grad_norm": 15.17529582977295, + "learning_rate": 8.934770709744289e-07, + "loss": 0.5208, + "num_input_tokens_seen": 2030043136, + "step": 968 + }, + { + "epoch": 0.2628152969894223, + "grad_norm": 12.855225563049316, + "learning_rate": 8.93216123862306e-07, + "loss": 0.3606, + "num_input_tokens_seen": 2032140288, + "step": 969 + }, + { + "epoch": 0.26308652020612966, + "grad_norm": 23.8533992767334, + "learning_rate": 8.929549005429071e-07, + "loss": 1.1691, + "num_input_tokens_seen": 2034237440, + "step": 970 + }, + { + "epoch": 0.263357743422837, + "grad_norm": 12.444233894348145, + "learning_rate": 8.926934012264546e-07, + "loss": 0.4991, + "num_input_tokens_seen": 2036334592, + "step": 971 + }, + { + "epoch": 0.2636289666395443, + "grad_norm": 19.38515853881836, + "learning_rate": 8.924316261233933e-07, + "loss": 0.603, + "num_input_tokens_seen": 2038431744, + "step": 972 + }, + { + "epoch": 0.2639001898562517, + "grad_norm": 25.431591033935547, + "learning_rate": 8.921695754443898e-07, + "loss": 0.4341, + "num_input_tokens_seen": 2040528896, + "step": 973 + }, + { + "epoch": 0.26417141307295905, + "grad_norm": 14.830904960632324, + "learning_rate": 8.919072494003325e-07, + "loss": 0.4873, + "num_input_tokens_seen": 2042626048, + "step": 974 + }, + { + "epoch": 0.2644426362896664, + "grad_norm": 10.834352493286133, + "learning_rate": 8.916446482023313e-07, + "loss": 0.386, + "num_input_tokens_seen": 2044723200, + "step": 975 + }, + { + "epoch": 0.26471385950637377, + "grad_norm": 18.31397819519043, + "learning_rate": 8.913817720617178e-07, + "loss": 0.5712, + "num_input_tokens_seen": 2046820352, + "step": 976 + }, + { + "epoch": 0.2649850827230811, + "grad_norm": 14.418971061706543, + "learning_rate": 8.911186211900448e-07, + "loss": 0.3854, + "num_input_tokens_seen": 2048917504, + "step": 977 + }, + { + "epoch": 0.26525630593978844, + "grad_norm": 15.985840797424316, + "learning_rate": 8.908551957990858e-07, + "loss": 0.6277, + "num_input_tokens_seen": 2051014656, + "step": 978 + }, + { + "epoch": 0.2655275291564958, + "grad_norm": 20.523096084594727, + "learning_rate": 8.90591496100836e-07, + "loss": 0.6733, + "num_input_tokens_seen": 2053111808, + "step": 979 + }, + { + "epoch": 0.26579875237320316, + "grad_norm": 21.447572708129883, + "learning_rate": 8.903275223075104e-07, + "loss": 0.5786, + "num_input_tokens_seen": 2055208960, + "step": 980 + }, + { + "epoch": 0.2660699755899105, + "grad_norm": 15.837870597839355, + "learning_rate": 8.900632746315454e-07, + "loss": 0.6692, + "num_input_tokens_seen": 2057306112, + "step": 981 + }, + { + "epoch": 0.26634119880661783, + "grad_norm": 25.582645416259766, + "learning_rate": 8.897987532855972e-07, + "loss": 1.0872, + "num_input_tokens_seen": 2059403264, + "step": 982 + }, + { + "epoch": 0.2666124220233252, + "grad_norm": 17.301437377929688, + "learning_rate": 8.895339584825429e-07, + "loss": 0.5819, + "num_input_tokens_seen": 2061500416, + "step": 983 + }, + { + "epoch": 0.26688364524003255, + "grad_norm": 14.604111671447754, + "learning_rate": 8.892688904354787e-07, + "loss": 0.4918, + "num_input_tokens_seen": 2063597568, + "step": 984 + }, + { + "epoch": 0.2671548684567399, + "grad_norm": 12.52351188659668, + "learning_rate": 8.890035493577219e-07, + "loss": 0.388, + "num_input_tokens_seen": 2065694720, + "step": 985 + }, + { + "epoch": 0.2674260916734473, + "grad_norm": 21.226011276245117, + "learning_rate": 8.887379354628085e-07, + "loss": 0.9703, + "num_input_tokens_seen": 2067791872, + "step": 986 + }, + { + "epoch": 0.2676973148901546, + "grad_norm": 12.156089782714844, + "learning_rate": 8.884720489644945e-07, + "loss": 0.36, + "num_input_tokens_seen": 2069889024, + "step": 987 + }, + { + "epoch": 0.26796853810686194, + "grad_norm": 16.55057144165039, + "learning_rate": 8.882058900767555e-07, + "loss": 0.6824, + "num_input_tokens_seen": 2071986176, + "step": 988 + }, + { + "epoch": 0.2682397613235693, + "grad_norm": 19.046653747558594, + "learning_rate": 8.879394590137857e-07, + "loss": 0.6588, + "num_input_tokens_seen": 2074083328, + "step": 989 + }, + { + "epoch": 0.26851098454027666, + "grad_norm": 13.83050537109375, + "learning_rate": 8.876727559899989e-07, + "loss": 0.3678, + "num_input_tokens_seen": 2076180480, + "step": 990 + }, + { + "epoch": 0.26878220775698397, + "grad_norm": 23.640356063842773, + "learning_rate": 8.874057812200274e-07, + "loss": 0.9572, + "num_input_tokens_seen": 2078277632, + "step": 991 + }, + { + "epoch": 0.26905343097369133, + "grad_norm": 15.057960510253906, + "learning_rate": 8.871385349187225e-07, + "loss": 0.722, + "num_input_tokens_seen": 2080374784, + "step": 992 + }, + { + "epoch": 0.2693246541903987, + "grad_norm": 18.069377899169922, + "learning_rate": 8.868710173011538e-07, + "loss": 0.7448, + "num_input_tokens_seen": 2082471936, + "step": 993 + }, + { + "epoch": 0.26959587740710605, + "grad_norm": 13.327810287475586, + "learning_rate": 8.866032285826091e-07, + "loss": 0.5023, + "num_input_tokens_seen": 2084569088, + "step": 994 + }, + { + "epoch": 0.2698671006238134, + "grad_norm": 11.229975700378418, + "learning_rate": 8.86335168978595e-07, + "loss": 0.357, + "num_input_tokens_seen": 2086666240, + "step": 995 + }, + { + "epoch": 0.2701383238405207, + "grad_norm": 10.154101371765137, + "learning_rate": 8.860668387048353e-07, + "loss": 0.2903, + "num_input_tokens_seen": 2088763392, + "step": 996 + }, + { + "epoch": 0.2704095470572281, + "grad_norm": 12.260966300964355, + "learning_rate": 8.85798237977272e-07, + "loss": 0.3106, + "num_input_tokens_seen": 2090860544, + "step": 997 + }, + { + "epoch": 0.27068077027393544, + "grad_norm": 18.005229949951172, + "learning_rate": 8.85529367012065e-07, + "loss": 0.6471, + "num_input_tokens_seen": 2092957696, + "step": 998 + }, + { + "epoch": 0.2709519934906428, + "grad_norm": 13.38134765625, + "learning_rate": 8.852602260255911e-07, + "loss": 0.4971, + "num_input_tokens_seen": 2095054848, + "step": 999 + }, + { + "epoch": 0.27122321670735017, + "grad_norm": 17.00292205810547, + "learning_rate": 8.849908152344451e-07, + "loss": 0.7469, + "num_input_tokens_seen": 2097152000, + "step": 1000 + }, + { + "epoch": 0.00027122321670735016, + "grad_norm": 21.211387634277344, + "learning_rate": 8.847211348554382e-07, + "loss": 0.5452, + "num_input_tokens_seen": 2099249152, + "step": 1001 + }, + { + "epoch": 0.0005424464334147003, + "grad_norm": 15.134449005126953, + "learning_rate": 8.844511851055991e-07, + "loss": 0.494, + "num_input_tokens_seen": 2101346304, + "step": 1002 + }, + { + "epoch": 0.0008136696501220504, + "grad_norm": 15.855742454528809, + "learning_rate": 8.841809662021731e-07, + "loss": 0.5585, + "num_input_tokens_seen": 2103443456, + "step": 1003 + }, + { + "epoch": 0.0010848928668294006, + "grad_norm": 25.325538635253906, + "learning_rate": 8.839104783626219e-07, + "loss": 0.9896, + "num_input_tokens_seen": 2105540608, + "step": 1004 + }, + { + "epoch": 0.0013561160835367507, + "grad_norm": 12.955252647399902, + "learning_rate": 8.836397218046239e-07, + "loss": 0.423, + "num_input_tokens_seen": 2107637760, + "step": 1005 + }, + { + "epoch": 0.0016273393002441008, + "grad_norm": 18.2934513092041, + "learning_rate": 8.83368696746074e-07, + "loss": 0.5536, + "num_input_tokens_seen": 2109734912, + "step": 1006 + }, + { + "epoch": 0.001898562516951451, + "grad_norm": 19.734365463256836, + "learning_rate": 8.830974034050824e-07, + "loss": 0.8425, + "num_input_tokens_seen": 2111832064, + "step": 1007 + }, + { + "epoch": 0.0021697857336588013, + "grad_norm": 9.506952285766602, + "learning_rate": 8.828258419999759e-07, + "loss": 0.2779, + "num_input_tokens_seen": 2113929216, + "step": 1008 + }, + { + "epoch": 0.0024410089503661514, + "grad_norm": 10.620465278625488, + "learning_rate": 8.825540127492965e-07, + "loss": 0.3105, + "num_input_tokens_seen": 2116026368, + "step": 1009 + }, + { + "epoch": 0.0027122321670735015, + "grad_norm": 20.96041488647461, + "learning_rate": 8.822819158718026e-07, + "loss": 0.775, + "num_input_tokens_seen": 2118123520, + "step": 1010 + }, + { + "epoch": 0.0029834553837808516, + "grad_norm": 15.075592994689941, + "learning_rate": 8.820095515864669e-07, + "loss": 0.505, + "num_input_tokens_seen": 2120220672, + "step": 1011 + }, + { + "epoch": 0.0032546786004882017, + "grad_norm": 19.47187042236328, + "learning_rate": 8.81736920112478e-07, + "loss": 0.6136, + "num_input_tokens_seen": 2122317824, + "step": 1012 + }, + { + "epoch": 0.003525901817195552, + "grad_norm": 17.397031784057617, + "learning_rate": 8.814640216692391e-07, + "loss": 0.4918, + "num_input_tokens_seen": 2124414976, + "step": 1013 + }, + { + "epoch": 0.003797125033902902, + "grad_norm": 23.4683895111084, + "learning_rate": 8.81190856476369e-07, + "loss": 0.8976, + "num_input_tokens_seen": 2126512128, + "step": 1014 + }, + { + "epoch": 0.0040683482506102524, + "grad_norm": 21.26357078552246, + "learning_rate": 8.809174247537003e-07, + "loss": 0.6695, + "num_input_tokens_seen": 2128609280, + "step": 1015 + }, + { + "epoch": 0.0043395714673176026, + "grad_norm": 9.867420196533203, + "learning_rate": 8.806437267212805e-07, + "loss": 0.2329, + "num_input_tokens_seen": 2130706432, + "step": 1016 + }, + { + "epoch": 0.004610794684024953, + "grad_norm": 26.58327293395996, + "learning_rate": 8.803697625993713e-07, + "loss": 0.8958, + "num_input_tokens_seen": 2132803584, + "step": 1017 + }, + { + "epoch": 0.004882017900732303, + "grad_norm": 15.50483226776123, + "learning_rate": 8.800955326084487e-07, + "loss": 0.62, + "num_input_tokens_seen": 2134900736, + "step": 1018 + }, + { + "epoch": 0.005153241117439653, + "grad_norm": 9.97493839263916, + "learning_rate": 8.798210369692025e-07, + "loss": 0.255, + "num_input_tokens_seen": 2136997888, + "step": 1019 + }, + { + "epoch": 0.005424464334147003, + "grad_norm": 13.728446960449219, + "learning_rate": 8.795462759025364e-07, + "loss": 0.5253, + "num_input_tokens_seen": 2139095040, + "step": 1020 + }, + { + "epoch": 0.005695687550854353, + "grad_norm": 13.042479515075684, + "learning_rate": 8.792712496295677e-07, + "loss": 0.3554, + "num_input_tokens_seen": 2141192192, + "step": 1021 + }, + { + "epoch": 0.005966910767561703, + "grad_norm": 14.499848365783691, + "learning_rate": 8.789959583716268e-07, + "loss": 0.4398, + "num_input_tokens_seen": 2143289344, + "step": 1022 + }, + { + "epoch": 0.006238133984269053, + "grad_norm": 12.071355819702148, + "learning_rate": 8.787204023502579e-07, + "loss": 0.3154, + "num_input_tokens_seen": 2145386496, + "step": 1023 + }, + { + "epoch": 0.006509357200976403, + "grad_norm": 16.29543113708496, + "learning_rate": 8.78444581787218e-07, + "loss": 0.4772, + "num_input_tokens_seen": 2147483648, + "step": 1024 + }, + { + "epoch": 0.0067805804176837535, + "grad_norm": 14.60418701171875, + "learning_rate": 8.781684969044769e-07, + "loss": 0.5242, + "num_input_tokens_seen": 2149580800, + "step": 1025 + }, + { + "epoch": 0.007051803634391104, + "grad_norm": 16.060916900634766, + "learning_rate": 8.778921479242173e-07, + "loss": 0.6014, + "num_input_tokens_seen": 2151677952, + "step": 1026 + }, + { + "epoch": 0.007323026851098454, + "grad_norm": 11.258378028869629, + "learning_rate": 8.776155350688342e-07, + "loss": 0.2425, + "num_input_tokens_seen": 2153775104, + "step": 1027 + }, + { + "epoch": 0.007594250067805804, + "grad_norm": 13.625775337219238, + "learning_rate": 8.773386585609352e-07, + "loss": 0.3129, + "num_input_tokens_seen": 2155872256, + "step": 1028 + }, + { + "epoch": 0.007865473284513154, + "grad_norm": 14.161125183105469, + "learning_rate": 8.770615186233398e-07, + "loss": 0.4769, + "num_input_tokens_seen": 2157969408, + "step": 1029 + }, + { + "epoch": 0.008136696501220505, + "grad_norm": 14.752073287963867, + "learning_rate": 8.7678411547908e-07, + "loss": 0.4059, + "num_input_tokens_seen": 2160066560, + "step": 1030 + }, + { + "epoch": 0.008407919717927854, + "grad_norm": 19.5262451171875, + "learning_rate": 8.76506449351399e-07, + "loss": 0.9232, + "num_input_tokens_seen": 2162163712, + "step": 1031 + }, + { + "epoch": 0.008679142934635205, + "grad_norm": 23.627025604248047, + "learning_rate": 8.762285204637522e-07, + "loss": 0.3903, + "num_input_tokens_seen": 2164260864, + "step": 1032 + }, + { + "epoch": 0.008950366151342554, + "grad_norm": 22.0424861907959, + "learning_rate": 8.75950329039806e-07, + "loss": 0.9677, + "num_input_tokens_seen": 2166358016, + "step": 1033 + }, + { + "epoch": 0.009221589368049905, + "grad_norm": 16.031373977661133, + "learning_rate": 8.756718753034381e-07, + "loss": 0.4666, + "num_input_tokens_seen": 2168455168, + "step": 1034 + }, + { + "epoch": 0.009492812584757255, + "grad_norm": 12.060325622558594, + "learning_rate": 8.75393159478738e-07, + "loss": 0.4239, + "num_input_tokens_seen": 2170552320, + "step": 1035 + }, + { + "epoch": 0.009764035801464606, + "grad_norm": 17.0775089263916, + "learning_rate": 8.751141817900052e-07, + "loss": 0.5839, + "num_input_tokens_seen": 2172649472, + "step": 1036 + }, + { + "epoch": 0.010035259018171955, + "grad_norm": 12.101827621459961, + "learning_rate": 8.748349424617504e-07, + "loss": 0.3156, + "num_input_tokens_seen": 2174746624, + "step": 1037 + }, + { + "epoch": 0.010306482234879306, + "grad_norm": 19.348678588867188, + "learning_rate": 8.745554417186946e-07, + "loss": 0.7375, + "num_input_tokens_seen": 2176843776, + "step": 1038 + }, + { + "epoch": 0.010577705451586655, + "grad_norm": 20.518571853637695, + "learning_rate": 8.742756797857698e-07, + "loss": 0.7549, + "num_input_tokens_seen": 2178940928, + "step": 1039 + }, + { + "epoch": 0.010848928668294006, + "grad_norm": 29.709476470947266, + "learning_rate": 8.739956568881174e-07, + "loss": 0.5175, + "num_input_tokens_seen": 2181038080, + "step": 1040 + }, + { + "epoch": 0.011120151885001357, + "grad_norm": 16.73153305053711, + "learning_rate": 8.737153732510894e-07, + "loss": 0.5272, + "num_input_tokens_seen": 2183135232, + "step": 1041 + }, + { + "epoch": 0.011391375101708706, + "grad_norm": 11.584559440612793, + "learning_rate": 8.734348291002472e-07, + "loss": 0.4049, + "num_input_tokens_seen": 2185232384, + "step": 1042 + }, + { + "epoch": 0.011662598318416057, + "grad_norm": 16.299226760864258, + "learning_rate": 8.731540246613621e-07, + "loss": 0.6874, + "num_input_tokens_seen": 2187329536, + "step": 1043 + }, + { + "epoch": 0.011933821535123406, + "grad_norm": 12.125165939331055, + "learning_rate": 8.728729601604149e-07, + "loss": 0.2929, + "num_input_tokens_seen": 2189426688, + "step": 1044 + }, + { + "epoch": 0.012205044751830757, + "grad_norm": 15.17818546295166, + "learning_rate": 8.725916358235956e-07, + "loss": 0.5507, + "num_input_tokens_seen": 2191523840, + "step": 1045 + }, + { + "epoch": 0.012476267968538107, + "grad_norm": 15.40369987487793, + "learning_rate": 8.723100518773034e-07, + "loss": 0.7183, + "num_input_tokens_seen": 2193620992, + "step": 1046 + }, + { + "epoch": 0.012747491185245458, + "grad_norm": 13.198980331420898, + "learning_rate": 8.720282085481463e-07, + "loss": 0.3474, + "num_input_tokens_seen": 2195718144, + "step": 1047 + }, + { + "epoch": 0.013018714401952807, + "grad_norm": 13.224248886108398, + "learning_rate": 8.717461060629408e-07, + "loss": 0.4623, + "num_input_tokens_seen": 2197815296, + "step": 1048 + }, + { + "epoch": 0.013289937618660158, + "grad_norm": 21.415082931518555, + "learning_rate": 8.714637446487127e-07, + "loss": 0.6044, + "num_input_tokens_seen": 2199912448, + "step": 1049 + }, + { + "epoch": 0.013561160835367507, + "grad_norm": 17.107685089111328, + "learning_rate": 8.711811245326955e-07, + "loss": 0.4799, + "num_input_tokens_seen": 2202009600, + "step": 1050 + }, + { + "epoch": 0.013832384052074858, + "grad_norm": 16.153474807739258, + "learning_rate": 8.70898245942331e-07, + "loss": 0.5492, + "num_input_tokens_seen": 2204106752, + "step": 1051 + }, + { + "epoch": 0.014103607268782207, + "grad_norm": 14.7495756149292, + "learning_rate": 8.706151091052693e-07, + "loss": 0.6168, + "num_input_tokens_seen": 2206203904, + "step": 1052 + }, + { + "epoch": 0.014374830485489558, + "grad_norm": 15.120250701904297, + "learning_rate": 8.703317142493681e-07, + "loss": 0.5481, + "num_input_tokens_seen": 2208301056, + "step": 1053 + }, + { + "epoch": 0.014646053702196907, + "grad_norm": 18.195310592651367, + "learning_rate": 8.700480616026928e-07, + "loss": 0.6981, + "num_input_tokens_seen": 2210398208, + "step": 1054 + }, + { + "epoch": 0.014917276918904258, + "grad_norm": 15.834187507629395, + "learning_rate": 8.697641513935164e-07, + "loss": 0.5543, + "num_input_tokens_seen": 2212495360, + "step": 1055 + }, + { + "epoch": 0.015188500135611608, + "grad_norm": 13.692312240600586, + "learning_rate": 8.694799838503186e-07, + "loss": 0.5255, + "num_input_tokens_seen": 2214592512, + "step": 1056 + }, + { + "epoch": 0.015459723352318959, + "grad_norm": 11.233868598937988, + "learning_rate": 8.691955592017872e-07, + "loss": 0.3443, + "num_input_tokens_seen": 2216689664, + "step": 1057 + }, + { + "epoch": 0.015730946569026308, + "grad_norm": 20.017642974853516, + "learning_rate": 8.689108776768159e-07, + "loss": 0.7494, + "num_input_tokens_seen": 2218786816, + "step": 1058 + }, + { + "epoch": 0.01600216978573366, + "grad_norm": 14.512331008911133, + "learning_rate": 8.686259395045056e-07, + "loss": 0.6063, + "num_input_tokens_seen": 2220883968, + "step": 1059 + }, + { + "epoch": 0.01627339300244101, + "grad_norm": 15.613311767578125, + "learning_rate": 8.68340744914164e-07, + "loss": 0.4511, + "num_input_tokens_seen": 2222981120, + "step": 1060 + }, + { + "epoch": 0.01654461621914836, + "grad_norm": 13.627348899841309, + "learning_rate": 8.680552941353045e-07, + "loss": 0.3019, + "num_input_tokens_seen": 2225078272, + "step": 1061 + }, + { + "epoch": 0.016815839435855708, + "grad_norm": 16.624229431152344, + "learning_rate": 8.677695873976473e-07, + "loss": 0.701, + "num_input_tokens_seen": 2227175424, + "step": 1062 + }, + { + "epoch": 0.01708706265256306, + "grad_norm": 11.855605125427246, + "learning_rate": 8.674836249311182e-07, + "loss": 0.4085, + "num_input_tokens_seen": 2229272576, + "step": 1063 + }, + { + "epoch": 0.01735828586927041, + "grad_norm": 16.31574249267578, + "learning_rate": 8.671974069658488e-07, + "loss": 0.6829, + "num_input_tokens_seen": 2231369728, + "step": 1064 + }, + { + "epoch": 0.01762950908597776, + "grad_norm": 16.633039474487305, + "learning_rate": 8.669109337321767e-07, + "loss": 0.603, + "num_input_tokens_seen": 2233466880, + "step": 1065 + }, + { + "epoch": 0.01790073230268511, + "grad_norm": 12.263994216918945, + "learning_rate": 8.666242054606444e-07, + "loss": 0.4269, + "num_input_tokens_seen": 2235564032, + "step": 1066 + }, + { + "epoch": 0.01817195551939246, + "grad_norm": 13.086897850036621, + "learning_rate": 8.66337222382e-07, + "loss": 0.5071, + "num_input_tokens_seen": 2237661184, + "step": 1067 + }, + { + "epoch": 0.01844317873609981, + "grad_norm": 18.24944305419922, + "learning_rate": 8.660499847271965e-07, + "loss": 0.7171, + "num_input_tokens_seen": 2239758336, + "step": 1068 + }, + { + "epoch": 0.01871440195280716, + "grad_norm": 12.880998611450195, + "learning_rate": 8.657624927273919e-07, + "loss": 0.3733, + "num_input_tokens_seen": 2241855488, + "step": 1069 + }, + { + "epoch": 0.01898562516951451, + "grad_norm": 11.316634178161621, + "learning_rate": 8.654747466139488e-07, + "loss": 0.417, + "num_input_tokens_seen": 2243952640, + "step": 1070 + }, + { + "epoch": 0.01925684838622186, + "grad_norm": 16.127735137939453, + "learning_rate": 8.651867466184344e-07, + "loss": 0.515, + "num_input_tokens_seen": 2246049792, + "step": 1071 + }, + { + "epoch": 0.01952807160292921, + "grad_norm": 13.648701667785645, + "learning_rate": 8.6489849297262e-07, + "loss": 0.4598, + "num_input_tokens_seen": 2248146944, + "step": 1072 + }, + { + "epoch": 0.019799294819636562, + "grad_norm": 18.77680778503418, + "learning_rate": 8.646099859084812e-07, + "loss": 0.4691, + "num_input_tokens_seen": 2250244096, + "step": 1073 + }, + { + "epoch": 0.02007051803634391, + "grad_norm": 14.230118751525879, + "learning_rate": 8.643212256581978e-07, + "loss": 0.4609, + "num_input_tokens_seen": 2252341248, + "step": 1074 + }, + { + "epoch": 0.02034174125305126, + "grad_norm": 20.32296371459961, + "learning_rate": 8.640322124541525e-07, + "loss": 0.4302, + "num_input_tokens_seen": 2254438400, + "step": 1075 + }, + { + "epoch": 0.02061296446975861, + "grad_norm": 19.631397247314453, + "learning_rate": 8.637429465289324e-07, + "loss": 0.7439, + "num_input_tokens_seen": 2256535552, + "step": 1076 + }, + { + "epoch": 0.020884187686465962, + "grad_norm": 20.95248794555664, + "learning_rate": 8.63453428115328e-07, + "loss": 0.9491, + "num_input_tokens_seen": 2258632704, + "step": 1077 + }, + { + "epoch": 0.02115541090317331, + "grad_norm": 23.594327926635742, + "learning_rate": 8.631636574463321e-07, + "loss": 1.0095, + "num_input_tokens_seen": 2260729856, + "step": 1078 + }, + { + "epoch": 0.02142663411988066, + "grad_norm": 13.66223430633545, + "learning_rate": 8.628736347551417e-07, + "loss": 0.4976, + "num_input_tokens_seen": 2262827008, + "step": 1079 + }, + { + "epoch": 0.021697857336588012, + "grad_norm": 14.790290832519531, + "learning_rate": 8.625833602751559e-07, + "loss": 0.4426, + "num_input_tokens_seen": 2264924160, + "step": 1080 + }, + { + "epoch": 0.021969080553295363, + "grad_norm": 22.70588493347168, + "learning_rate": 8.622928342399762e-07, + "loss": 0.9969, + "num_input_tokens_seen": 2267021312, + "step": 1081 + }, + { + "epoch": 0.022240303770002714, + "grad_norm": 15.693807601928711, + "learning_rate": 8.620020568834072e-07, + "loss": 0.6167, + "num_input_tokens_seen": 2269118464, + "step": 1082 + }, + { + "epoch": 0.02251152698671006, + "grad_norm": 14.573871612548828, + "learning_rate": 8.617110284394553e-07, + "loss": 0.5203, + "num_input_tokens_seen": 2271215616, + "step": 1083 + }, + { + "epoch": 0.022782750203417412, + "grad_norm": 12.033546447753906, + "learning_rate": 8.614197491423293e-07, + "loss": 0.3329, + "num_input_tokens_seen": 2273312768, + "step": 1084 + }, + { + "epoch": 0.023053973420124763, + "grad_norm": 21.395675659179688, + "learning_rate": 8.611282192264396e-07, + "loss": 0.8502, + "num_input_tokens_seen": 2275409920, + "step": 1085 + }, + { + "epoch": 0.023325196636832114, + "grad_norm": 10.336832046508789, + "learning_rate": 8.608364389263984e-07, + "loss": 0.3252, + "num_input_tokens_seen": 2277507072, + "step": 1086 + }, + { + "epoch": 0.023596419853539462, + "grad_norm": 14.438992500305176, + "learning_rate": 8.605444084770192e-07, + "loss": 0.4895, + "num_input_tokens_seen": 2279604224, + "step": 1087 + }, + { + "epoch": 0.023867643070246813, + "grad_norm": 12.839303970336914, + "learning_rate": 8.602521281133173e-07, + "loss": 0.4126, + "num_input_tokens_seen": 2281701376, + "step": 1088 + }, + { + "epoch": 0.024138866286954164, + "grad_norm": 15.832900047302246, + "learning_rate": 8.599595980705085e-07, + "loss": 0.298, + "num_input_tokens_seen": 2283798528, + "step": 1089 + }, + { + "epoch": 0.024410089503661515, + "grad_norm": 16.126102447509766, + "learning_rate": 8.596668185840102e-07, + "loss": 0.5055, + "num_input_tokens_seen": 2285895680, + "step": 1090 + }, + { + "epoch": 0.024681312720368862, + "grad_norm": 13.678728103637695, + "learning_rate": 8.593737898894398e-07, + "loss": 0.5144, + "num_input_tokens_seen": 2287992832, + "step": 1091 + }, + { + "epoch": 0.024952535937076213, + "grad_norm": 13.563033103942871, + "learning_rate": 8.59080512222616e-07, + "loss": 0.5175, + "num_input_tokens_seen": 2290089984, + "step": 1092 + }, + { + "epoch": 0.025223759153783564, + "grad_norm": 22.787578582763672, + "learning_rate": 8.587869858195574e-07, + "loss": 0.662, + "num_input_tokens_seen": 2292187136, + "step": 1093 + }, + { + "epoch": 0.025494982370490915, + "grad_norm": 12.561062812805176, + "learning_rate": 8.584932109164826e-07, + "loss": 0.4915, + "num_input_tokens_seen": 2294284288, + "step": 1094 + }, + { + "epoch": 0.025766205587198263, + "grad_norm": 10.324007034301758, + "learning_rate": 8.581991877498109e-07, + "loss": 0.3419, + "num_input_tokens_seen": 2296381440, + "step": 1095 + }, + { + "epoch": 0.026037428803905614, + "grad_norm": 16.793432235717773, + "learning_rate": 8.579049165561607e-07, + "loss": 0.5506, + "num_input_tokens_seen": 2298478592, + "step": 1096 + }, + { + "epoch": 0.026308652020612965, + "grad_norm": 18.08753776550293, + "learning_rate": 8.576103975723502e-07, + "loss": 0.6403, + "num_input_tokens_seen": 2300575744, + "step": 1097 + }, + { + "epoch": 0.026579875237320316, + "grad_norm": 14.949502944946289, + "learning_rate": 8.573156310353974e-07, + "loss": 0.5685, + "num_input_tokens_seen": 2302672896, + "step": 1098 + }, + { + "epoch": 0.026851098454027666, + "grad_norm": 14.36096477508545, + "learning_rate": 8.570206171825188e-07, + "loss": 0.5644, + "num_input_tokens_seen": 2304770048, + "step": 1099 + }, + { + "epoch": 0.027122321670735014, + "grad_norm": 13.551562309265137, + "learning_rate": 8.567253562511306e-07, + "loss": 0.439, + "num_input_tokens_seen": 2306867200, + "step": 1100 + }, + { + "epoch": 0.027393544887442365, + "grad_norm": 18.999597549438477, + "learning_rate": 8.564298484788472e-07, + "loss": 0.4937, + "num_input_tokens_seen": 2308964352, + "step": 1101 + }, + { + "epoch": 0.027664768104149716, + "grad_norm": 10.107467651367188, + "learning_rate": 8.561340941034825e-07, + "loss": 0.2738, + "num_input_tokens_seen": 2311061504, + "step": 1102 + }, + { + "epoch": 0.027935991320857067, + "grad_norm": 19.000905990600586, + "learning_rate": 8.55838093363048e-07, + "loss": 0.6454, + "num_input_tokens_seen": 2313158656, + "step": 1103 + }, + { + "epoch": 0.028207214537564414, + "grad_norm": 31.94487190246582, + "learning_rate": 8.555418464957542e-07, + "loss": 0.6641, + "num_input_tokens_seen": 2315255808, + "step": 1104 + }, + { + "epoch": 0.028478437754271765, + "grad_norm": 16.993053436279297, + "learning_rate": 8.552453537400089e-07, + "loss": 0.4975, + "num_input_tokens_seen": 2317352960, + "step": 1105 + }, + { + "epoch": 0.028749660970979116, + "grad_norm": 21.718708038330078, + "learning_rate": 8.549486153344183e-07, + "loss": 0.7325, + "num_input_tokens_seen": 2319450112, + "step": 1106 + }, + { + "epoch": 0.029020884187686467, + "grad_norm": 16.96457862854004, + "learning_rate": 8.546516315177863e-07, + "loss": 0.5915, + "num_input_tokens_seen": 2321547264, + "step": 1107 + }, + { + "epoch": 0.029292107404393815, + "grad_norm": 15.942675590515137, + "learning_rate": 8.543544025291143e-07, + "loss": 0.5334, + "num_input_tokens_seen": 2323644416, + "step": 1108 + }, + { + "epoch": 0.029563330621101166, + "grad_norm": 14.432857513427734, + "learning_rate": 8.540569286076004e-07, + "loss": 0.5069, + "num_input_tokens_seen": 2325741568, + "step": 1109 + }, + { + "epoch": 0.029834553837808517, + "grad_norm": 16.693790435791016, + "learning_rate": 8.537592099926407e-07, + "loss": 0.6163, + "num_input_tokens_seen": 2327838720, + "step": 1110 + }, + { + "epoch": 0.030105777054515868, + "grad_norm": 13.320590019226074, + "learning_rate": 8.534612469238278e-07, + "loss": 0.3806, + "num_input_tokens_seen": 2329935872, + "step": 1111 + }, + { + "epoch": 0.030377000271223215, + "grad_norm": 20.88117790222168, + "learning_rate": 8.531630396409507e-07, + "loss": 0.6614, + "num_input_tokens_seen": 2332033024, + "step": 1112 + }, + { + "epoch": 0.030648223487930566, + "grad_norm": 12.577048301696777, + "learning_rate": 8.528645883839956e-07, + "loss": 0.489, + "num_input_tokens_seen": 2334130176, + "step": 1113 + }, + { + "epoch": 0.030919446704637917, + "grad_norm": 16.97111701965332, + "learning_rate": 8.525658933931448e-07, + "loss": 0.5687, + "num_input_tokens_seen": 2336227328, + "step": 1114 + }, + { + "epoch": 0.031190669921345268, + "grad_norm": 15.8145112991333, + "learning_rate": 8.522669549087762e-07, + "loss": 0.556, + "num_input_tokens_seen": 2338324480, + "step": 1115 + }, + { + "epoch": 0.031461893138052616, + "grad_norm": 16.058319091796875, + "learning_rate": 8.519677731714645e-07, + "loss": 0.7402, + "num_input_tokens_seen": 2340421632, + "step": 1116 + }, + { + "epoch": 0.03173311635475997, + "grad_norm": 11.520198822021484, + "learning_rate": 8.516683484219797e-07, + "loss": 0.4213, + "num_input_tokens_seen": 2342518784, + "step": 1117 + }, + { + "epoch": 0.03200433957146732, + "grad_norm": 18.839998245239258, + "learning_rate": 8.513686809012875e-07, + "loss": 0.5229, + "num_input_tokens_seen": 2344615936, + "step": 1118 + }, + { + "epoch": 0.032275562788174665, + "grad_norm": 16.22575569152832, + "learning_rate": 8.510687708505489e-07, + "loss": 0.5792, + "num_input_tokens_seen": 2346713088, + "step": 1119 + }, + { + "epoch": 0.03254678600488202, + "grad_norm": 15.674661636352539, + "learning_rate": 8.507686185111199e-07, + "loss": 0.5386, + "num_input_tokens_seen": 2348810240, + "step": 1120 + }, + { + "epoch": 0.03281800922158937, + "grad_norm": 18.9416446685791, + "learning_rate": 8.504682241245516e-07, + "loss": 0.3716, + "num_input_tokens_seen": 2350907392, + "step": 1121 + }, + { + "epoch": 0.03308923243829672, + "grad_norm": 10.184718132019043, + "learning_rate": 8.501675879325906e-07, + "loss": 0.2198, + "num_input_tokens_seen": 2353004544, + "step": 1122 + }, + { + "epoch": 0.03336045565500407, + "grad_norm": 12.775616645812988, + "learning_rate": 8.498667101771769e-07, + "loss": 0.4381, + "num_input_tokens_seen": 2355101696, + "step": 1123 + }, + { + "epoch": 0.033631678871711417, + "grad_norm": 19.200225830078125, + "learning_rate": 8.495655911004456e-07, + "loss": 0.6905, + "num_input_tokens_seen": 2357198848, + "step": 1124 + }, + { + "epoch": 0.03390290208841877, + "grad_norm": 16.312833786010742, + "learning_rate": 8.492642309447257e-07, + "loss": 0.551, + "num_input_tokens_seen": 2359296000, + "step": 1125 + }, + { + "epoch": 0.03417412530512612, + "grad_norm": 12.200955390930176, + "learning_rate": 8.489626299525409e-07, + "loss": 0.3892, + "num_input_tokens_seen": 2361393152, + "step": 1126 + }, + { + "epoch": 0.034445348521833466, + "grad_norm": 12.41970157623291, + "learning_rate": 8.486607883666077e-07, + "loss": 0.4058, + "num_input_tokens_seen": 2363490304, + "step": 1127 + }, + { + "epoch": 0.03471657173854082, + "grad_norm": 17.284793853759766, + "learning_rate": 8.483587064298372e-07, + "loss": 0.5063, + "num_input_tokens_seen": 2365587456, + "step": 1128 + }, + { + "epoch": 0.03498779495524817, + "grad_norm": 12.992026329040527, + "learning_rate": 8.480563843853328e-07, + "loss": 0.3901, + "num_input_tokens_seen": 2367684608, + "step": 1129 + }, + { + "epoch": 0.03525901817195552, + "grad_norm": 25.536312103271484, + "learning_rate": 8.477538224763923e-07, + "loss": 0.6989, + "num_input_tokens_seen": 2369781760, + "step": 1130 + }, + { + "epoch": 0.03553024138866287, + "grad_norm": 14.612037658691406, + "learning_rate": 8.474510209465058e-07, + "loss": 0.4448, + "num_input_tokens_seen": 2371878912, + "step": 1131 + }, + { + "epoch": 0.03580146460537022, + "grad_norm": 17.197641372680664, + "learning_rate": 8.471479800393565e-07, + "loss": 0.5952, + "num_input_tokens_seen": 2373976064, + "step": 1132 + }, + { + "epoch": 0.03607268782207757, + "grad_norm": 15.149774551391602, + "learning_rate": 8.468446999988202e-07, + "loss": 0.5679, + "num_input_tokens_seen": 2376073216, + "step": 1133 + }, + { + "epoch": 0.03634391103878492, + "grad_norm": 20.161657333374023, + "learning_rate": 8.465411810689653e-07, + "loss": 0.9363, + "num_input_tokens_seen": 2378170368, + "step": 1134 + }, + { + "epoch": 0.03661513425549227, + "grad_norm": 13.918910026550293, + "learning_rate": 8.462374234940517e-07, + "loss": 0.4595, + "num_input_tokens_seen": 2380267520, + "step": 1135 + }, + { + "epoch": 0.03688635747219962, + "grad_norm": 11.941801071166992, + "learning_rate": 8.459334275185325e-07, + "loss": 0.3769, + "num_input_tokens_seen": 2382364672, + "step": 1136 + }, + { + "epoch": 0.03715758068890697, + "grad_norm": 23.785158157348633, + "learning_rate": 8.456291933870521e-07, + "loss": 1.0131, + "num_input_tokens_seen": 2384461824, + "step": 1137 + }, + { + "epoch": 0.03742880390561432, + "grad_norm": 12.919363021850586, + "learning_rate": 8.453247213444463e-07, + "loss": 0.2438, + "num_input_tokens_seen": 2386558976, + "step": 1138 + }, + { + "epoch": 0.03770002712232167, + "grad_norm": 14.640851020812988, + "learning_rate": 8.450200116357428e-07, + "loss": 0.4749, + "num_input_tokens_seen": 2388656128, + "step": 1139 + }, + { + "epoch": 0.03797125033902902, + "grad_norm": 10.543968200683594, + "learning_rate": 8.4471506450616e-07, + "loss": 0.3616, + "num_input_tokens_seen": 2390753280, + "step": 1140 + }, + { + "epoch": 0.03824247355573637, + "grad_norm": 15.393962860107422, + "learning_rate": 8.444098802011083e-07, + "loss": 0.4972, + "num_input_tokens_seen": 2392850432, + "step": 1141 + }, + { + "epoch": 0.03851369677244372, + "grad_norm": 79.32931518554688, + "learning_rate": 8.441044589661881e-07, + "loss": 0.5945, + "num_input_tokens_seen": 2394947584, + "step": 1142 + }, + { + "epoch": 0.038784919989151075, + "grad_norm": 13.988775253295898, + "learning_rate": 8.437988010471907e-07, + "loss": 0.5593, + "num_input_tokens_seen": 2397044736, + "step": 1143 + }, + { + "epoch": 0.03905614320585842, + "grad_norm": 14.631237983703613, + "learning_rate": 8.434929066900982e-07, + "loss": 0.5196, + "num_input_tokens_seen": 2399141888, + "step": 1144 + }, + { + "epoch": 0.03932736642256577, + "grad_norm": 23.943798065185547, + "learning_rate": 8.431867761410826e-07, + "loss": 0.8291, + "num_input_tokens_seen": 2401239040, + "step": 1145 + }, + { + "epoch": 0.039598589639273124, + "grad_norm": 13.034646034240723, + "learning_rate": 8.42880409646506e-07, + "loss": 0.4124, + "num_input_tokens_seen": 2403336192, + "step": 1146 + }, + { + "epoch": 0.03986981285598047, + "grad_norm": 16.411720275878906, + "learning_rate": 8.42573807452921e-07, + "loss": 0.4623, + "num_input_tokens_seen": 2405433344, + "step": 1147 + }, + { + "epoch": 0.04014103607268782, + "grad_norm": 18.51169204711914, + "learning_rate": 8.422669698070687e-07, + "loss": 0.5727, + "num_input_tokens_seen": 2407530496, + "step": 1148 + }, + { + "epoch": 0.040412259289395173, + "grad_norm": 10.572891235351562, + "learning_rate": 8.419598969558808e-07, + "loss": 0.2515, + "num_input_tokens_seen": 2409627648, + "step": 1149 + }, + { + "epoch": 0.04068348250610252, + "grad_norm": 10.220544815063477, + "learning_rate": 8.416525891464776e-07, + "loss": 0.3125, + "num_input_tokens_seen": 2411724800, + "step": 1150 + }, + { + "epoch": 0.040954705722809875, + "grad_norm": 14.300618171691895, + "learning_rate": 8.413450466261691e-07, + "loss": 0.429, + "num_input_tokens_seen": 2413821952, + "step": 1151 + }, + { + "epoch": 0.04122592893951722, + "grad_norm": 14.614937782287598, + "learning_rate": 8.410372696424535e-07, + "loss": 0.3731, + "num_input_tokens_seen": 2415919104, + "step": 1152 + }, + { + "epoch": 0.04149715215622457, + "grad_norm": 12.267914772033691, + "learning_rate": 8.40729258443018e-07, + "loss": 0.4024, + "num_input_tokens_seen": 2418016256, + "step": 1153 + }, + { + "epoch": 0.041768375372931925, + "grad_norm": 12.720690727233887, + "learning_rate": 8.404210132757385e-07, + "loss": 0.3267, + "num_input_tokens_seen": 2420113408, + "step": 1154 + }, + { + "epoch": 0.04203959858963927, + "grad_norm": 15.737774848937988, + "learning_rate": 8.401125343886787e-07, + "loss": 0.5473, + "num_input_tokens_seen": 2422210560, + "step": 1155 + }, + { + "epoch": 0.04231082180634662, + "grad_norm": 9.714899063110352, + "learning_rate": 8.398038220300908e-07, + "loss": 0.3099, + "num_input_tokens_seen": 2424307712, + "step": 1156 + }, + { + "epoch": 0.042582045023053974, + "grad_norm": 18.72726058959961, + "learning_rate": 8.39494876448415e-07, + "loss": 0.7204, + "num_input_tokens_seen": 2426404864, + "step": 1157 + }, + { + "epoch": 0.04285326823976132, + "grad_norm": 19.449329376220703, + "learning_rate": 8.391856978922785e-07, + "loss": 0.8302, + "num_input_tokens_seen": 2428502016, + "step": 1158 + }, + { + "epoch": 0.043124491456468676, + "grad_norm": 12.25051212310791, + "learning_rate": 8.38876286610497e-07, + "loss": 0.2353, + "num_input_tokens_seen": 2430599168, + "step": 1159 + }, + { + "epoch": 0.043395714673176024, + "grad_norm": 17.40892791748047, + "learning_rate": 8.385666428520723e-07, + "loss": 0.6826, + "num_input_tokens_seen": 2432696320, + "step": 1160 + }, + { + "epoch": 0.04366693788988337, + "grad_norm": 20.962051391601562, + "learning_rate": 8.382567668661943e-07, + "loss": 0.694, + "num_input_tokens_seen": 2434793472, + "step": 1161 + }, + { + "epoch": 0.043938161106590726, + "grad_norm": 17.564403533935547, + "learning_rate": 8.379466589022393e-07, + "loss": 0.6455, + "num_input_tokens_seen": 2436890624, + "step": 1162 + }, + { + "epoch": 0.04420938432329807, + "grad_norm": 16.586353302001953, + "learning_rate": 8.376363192097703e-07, + "loss": 0.5219, + "num_input_tokens_seen": 2438987776, + "step": 1163 + }, + { + "epoch": 0.04448060754000543, + "grad_norm": 20.66847801208496, + "learning_rate": 8.37325748038537e-07, + "loss": 0.6066, + "num_input_tokens_seen": 2441084928, + "step": 1164 + }, + { + "epoch": 0.044751830756712775, + "grad_norm": 17.191225051879883, + "learning_rate": 8.370149456384754e-07, + "loss": 0.4234, + "num_input_tokens_seen": 2443182080, + "step": 1165 + }, + { + "epoch": 0.04502305397342012, + "grad_norm": 13.013045310974121, + "learning_rate": 8.36703912259707e-07, + "loss": 0.3087, + "num_input_tokens_seen": 2445279232, + "step": 1166 + }, + { + "epoch": 0.04529427719012748, + "grad_norm": 22.083459854125977, + "learning_rate": 8.363926481525402e-07, + "loss": 0.7541, + "num_input_tokens_seen": 2447376384, + "step": 1167 + }, + { + "epoch": 0.045565500406834825, + "grad_norm": 14.312165260314941, + "learning_rate": 8.360811535674682e-07, + "loss": 0.4852, + "num_input_tokens_seen": 2449473536, + "step": 1168 + }, + { + "epoch": 0.04583672362354217, + "grad_norm": 20.598003387451172, + "learning_rate": 8.357694287551698e-07, + "loss": 0.7828, + "num_input_tokens_seen": 2451570688, + "step": 1169 + }, + { + "epoch": 0.04610794684024953, + "grad_norm": 19.325916290283203, + "learning_rate": 8.354574739665096e-07, + "loss": 0.4868, + "num_input_tokens_seen": 2453667840, + "step": 1170 + }, + { + "epoch": 0.046379170056956874, + "grad_norm": 12.34057903289795, + "learning_rate": 8.351452894525368e-07, + "loss": 0.3321, + "num_input_tokens_seen": 2455764992, + "step": 1171 + }, + { + "epoch": 0.04665039327366423, + "grad_norm": 12.274550437927246, + "learning_rate": 8.348328754644855e-07, + "loss": 0.4123, + "num_input_tokens_seen": 2457862144, + "step": 1172 + }, + { + "epoch": 0.046921616490371576, + "grad_norm": 14.192599296569824, + "learning_rate": 8.34520232253775e-07, + "loss": 0.5802, + "num_input_tokens_seen": 2459959296, + "step": 1173 + }, + { + "epoch": 0.047192839707078924, + "grad_norm": 20.330703735351562, + "learning_rate": 8.342073600720082e-07, + "loss": 0.6071, + "num_input_tokens_seen": 2462056448, + "step": 1174 + }, + { + "epoch": 0.04746406292378628, + "grad_norm": 16.746747970581055, + "learning_rate": 8.33894259170973e-07, + "loss": 0.4686, + "num_input_tokens_seen": 2464153600, + "step": 1175 + }, + { + "epoch": 0.047735286140493625, + "grad_norm": 13.190831184387207, + "learning_rate": 8.335809298026409e-07, + "loss": 0.329, + "num_input_tokens_seen": 2466250752, + "step": 1176 + }, + { + "epoch": 0.04800650935720097, + "grad_norm": 14.298337936401367, + "learning_rate": 8.332673722191677e-07, + "loss": 0.396, + "num_input_tokens_seen": 2468347904, + "step": 1177 + }, + { + "epoch": 0.04827773257390833, + "grad_norm": 14.075898170471191, + "learning_rate": 8.329535866728922e-07, + "loss": 0.4024, + "num_input_tokens_seen": 2470445056, + "step": 1178 + }, + { + "epoch": 0.048548955790615675, + "grad_norm": 15.5481595993042, + "learning_rate": 8.326395734163375e-07, + "loss": 0.6025, + "num_input_tokens_seen": 2472542208, + "step": 1179 + }, + { + "epoch": 0.04882017900732303, + "grad_norm": 12.739563941955566, + "learning_rate": 8.323253327022094e-07, + "loss": 0.3926, + "num_input_tokens_seen": 2474639360, + "step": 1180 + }, + { + "epoch": 0.04909140222403038, + "grad_norm": 10.208687782287598, + "learning_rate": 8.320108647833967e-07, + "loss": 0.3176, + "num_input_tokens_seen": 2476736512, + "step": 1181 + }, + { + "epoch": 0.049362625440737724, + "grad_norm": 9.83712387084961, + "learning_rate": 8.316961699129714e-07, + "loss": 0.247, + "num_input_tokens_seen": 2478833664, + "step": 1182 + }, + { + "epoch": 0.04963384865744508, + "grad_norm": 19.233394622802734, + "learning_rate": 8.313812483441879e-07, + "loss": 0.7097, + "num_input_tokens_seen": 2480930816, + "step": 1183 + }, + { + "epoch": 0.049905071874152426, + "grad_norm": 12.595343589782715, + "learning_rate": 8.310661003304829e-07, + "loss": 0.33, + "num_input_tokens_seen": 2483027968, + "step": 1184 + }, + { + "epoch": 0.05017629509085978, + "grad_norm": 14.636686325073242, + "learning_rate": 8.30750726125476e-07, + "loss": 0.4683, + "num_input_tokens_seen": 2485125120, + "step": 1185 + }, + { + "epoch": 0.05044751830756713, + "grad_norm": 13.301069259643555, + "learning_rate": 8.304351259829678e-07, + "loss": 0.4655, + "num_input_tokens_seen": 2487222272, + "step": 1186 + }, + { + "epoch": 0.050718741524274476, + "grad_norm": 13.354597091674805, + "learning_rate": 8.301193001569418e-07, + "loss": 0.3759, + "num_input_tokens_seen": 2489319424, + "step": 1187 + }, + { + "epoch": 0.05098996474098183, + "grad_norm": 12.58713150024414, + "learning_rate": 8.298032489015623e-07, + "loss": 0.3448, + "num_input_tokens_seen": 2491416576, + "step": 1188 + }, + { + "epoch": 0.05126118795768918, + "grad_norm": 19.054805755615234, + "learning_rate": 8.294869724711752e-07, + "loss": 0.586, + "num_input_tokens_seen": 2493513728, + "step": 1189 + }, + { + "epoch": 0.051532411174396525, + "grad_norm": 11.936301231384277, + "learning_rate": 8.291704711203082e-07, + "loss": 0.3833, + "num_input_tokens_seen": 2495610880, + "step": 1190 + }, + { + "epoch": 0.05180363439110388, + "grad_norm": 18.23284912109375, + "learning_rate": 8.288537451036691e-07, + "loss": 0.4227, + "num_input_tokens_seen": 2497708032, + "step": 1191 + }, + { + "epoch": 0.05207485760781123, + "grad_norm": 14.435386657714844, + "learning_rate": 8.28536794676147e-07, + "loss": 0.3232, + "num_input_tokens_seen": 2499805184, + "step": 1192 + }, + { + "epoch": 0.05234608082451858, + "grad_norm": 13.897747039794922, + "learning_rate": 8.282196200928119e-07, + "loss": 0.4033, + "num_input_tokens_seen": 2501902336, + "step": 1193 + }, + { + "epoch": 0.05261730404122593, + "grad_norm": 11.174321174621582, + "learning_rate": 8.279022216089135e-07, + "loss": 0.3174, + "num_input_tokens_seen": 2503999488, + "step": 1194 + }, + { + "epoch": 0.05288852725793328, + "grad_norm": 15.450333595275879, + "learning_rate": 8.275845994798821e-07, + "loss": 0.4879, + "num_input_tokens_seen": 2506096640, + "step": 1195 + }, + { + "epoch": 0.05315975047464063, + "grad_norm": 12.975201606750488, + "learning_rate": 8.272667539613281e-07, + "loss": 0.3927, + "num_input_tokens_seen": 2508193792, + "step": 1196 + }, + { + "epoch": 0.05343097369134798, + "grad_norm": 20.552017211914062, + "learning_rate": 8.26948685309041e-07, + "loss": 0.6859, + "num_input_tokens_seen": 2510290944, + "step": 1197 + }, + { + "epoch": 0.05370219690805533, + "grad_norm": 14.239463806152344, + "learning_rate": 8.266303937789908e-07, + "loss": 0.4479, + "num_input_tokens_seen": 2512388096, + "step": 1198 + }, + { + "epoch": 0.05397342012476268, + "grad_norm": 13.604122161865234, + "learning_rate": 8.263118796273263e-07, + "loss": 0.5931, + "num_input_tokens_seen": 2514485248, + "step": 1199 + }, + { + "epoch": 0.05424464334147003, + "grad_norm": 22.870668411254883, + "learning_rate": 8.259931431103754e-07, + "loss": 0.4736, + "num_input_tokens_seen": 2516582400, + "step": 1200 + }, + { + "epoch": 0.00027122321670735016, + "grad_norm": 16.450984954833984, + "learning_rate": 8.256741844846452e-07, + "loss": 0.5646, + "num_input_tokens_seen": 2518679552, + "step": 1201 + }, + { + "epoch": 0.0005424464334147003, + "grad_norm": 15.368687629699707, + "learning_rate": 8.253550040068216e-07, + "loss": 0.4454, + "num_input_tokens_seen": 2520776704, + "step": 1202 + }, + { + "epoch": 0.0008136696501220504, + "grad_norm": 15.554755210876465, + "learning_rate": 8.250356019337688e-07, + "loss": 0.4901, + "num_input_tokens_seen": 2522873856, + "step": 1203 + }, + { + "epoch": 0.0010848928668294006, + "grad_norm": 19.45119285583496, + "learning_rate": 8.247159785225295e-07, + "loss": 0.8747, + "num_input_tokens_seen": 2524971008, + "step": 1204 + }, + { + "epoch": 0.0013561160835367507, + "grad_norm": 19.58152198791504, + "learning_rate": 8.243961340303245e-07, + "loss": 0.6606, + "num_input_tokens_seen": 2527068160, + "step": 1205 + }, + { + "epoch": 0.0016273393002441008, + "grad_norm": 7.020671367645264, + "learning_rate": 8.240760687145521e-07, + "loss": 0.1893, + "num_input_tokens_seen": 2529165312, + "step": 1206 + }, + { + "epoch": 0.001898562516951451, + "grad_norm": 14.452476501464844, + "learning_rate": 8.237557828327891e-07, + "loss": 0.5641, + "num_input_tokens_seen": 2531262464, + "step": 1207 + }, + { + "epoch": 0.0021697857336588013, + "grad_norm": 17.215316772460938, + "learning_rate": 8.234352766427894e-07, + "loss": 0.6351, + "num_input_tokens_seen": 2533359616, + "step": 1208 + }, + { + "epoch": 0.0024410089503661514, + "grad_norm": 16.62022590637207, + "learning_rate": 8.231145504024838e-07, + "loss": 0.6488, + "num_input_tokens_seen": 2535456768, + "step": 1209 + }, + { + "epoch": 0.0027122321670735015, + "grad_norm": 9.606282234191895, + "learning_rate": 8.22793604369981e-07, + "loss": 0.353, + "num_input_tokens_seen": 2537553920, + "step": 1210 + }, + { + "epoch": 0.0029834553837808516, + "grad_norm": 12.205081939697266, + "learning_rate": 8.224724388035659e-07, + "loss": 0.3574, + "num_input_tokens_seen": 2539651072, + "step": 1211 + }, + { + "epoch": 0.0032546786004882017, + "grad_norm": 16.890460968017578, + "learning_rate": 8.221510539617003e-07, + "loss": 0.4982, + "num_input_tokens_seen": 2541748224, + "step": 1212 + }, + { + "epoch": 0.003525901817195552, + "grad_norm": 11.31159496307373, + "learning_rate": 8.218294501030226e-07, + "loss": 0.3511, + "num_input_tokens_seen": 2543845376, + "step": 1213 + }, + { + "epoch": 0.003797125033902902, + "grad_norm": 14.827753067016602, + "learning_rate": 8.215076274863476e-07, + "loss": 0.5236, + "num_input_tokens_seen": 2545942528, + "step": 1214 + }, + { + "epoch": 0.0040683482506102524, + "grad_norm": 13.192427635192871, + "learning_rate": 8.211855863706654e-07, + "loss": 0.4119, + "num_input_tokens_seen": 2548039680, + "step": 1215 + }, + { + "epoch": 0.0043395714673176026, + "grad_norm": 14.75649356842041, + "learning_rate": 8.208633270151426e-07, + "loss": 0.5348, + "num_input_tokens_seen": 2550136832, + "step": 1216 + }, + { + "epoch": 0.004610794684024953, + "grad_norm": 20.833091735839844, + "learning_rate": 8.205408496791216e-07, + "loss": 0.5757, + "num_input_tokens_seen": 2552233984, + "step": 1217 + }, + { + "epoch": 0.004882017900732303, + "grad_norm": 11.868510246276855, + "learning_rate": 8.202181546221193e-07, + "loss": 0.3579, + "num_input_tokens_seen": 2554331136, + "step": 1218 + }, + { + "epoch": 0.005153241117439653, + "grad_norm": 17.338224411010742, + "learning_rate": 8.19895242103829e-07, + "loss": 0.6951, + "num_input_tokens_seen": 2556428288, + "step": 1219 + }, + { + "epoch": 0.005424464334147003, + "grad_norm": 23.777769088745117, + "learning_rate": 8.19572112384118e-07, + "loss": 0.6578, + "num_input_tokens_seen": 2558525440, + "step": 1220 + }, + { + "epoch": 0.005695687550854353, + "grad_norm": 13.37986946105957, + "learning_rate": 8.192487657230288e-07, + "loss": 0.3095, + "num_input_tokens_seen": 2560622592, + "step": 1221 + }, + { + "epoch": 0.005966910767561703, + "grad_norm": 12.245566368103027, + "learning_rate": 8.18925202380779e-07, + "loss": 0.3178, + "num_input_tokens_seen": 2562719744, + "step": 1222 + }, + { + "epoch": 0.006238133984269053, + "grad_norm": 13.670166969299316, + "learning_rate": 8.186014226177594e-07, + "loss": 0.491, + "num_input_tokens_seen": 2564816896, + "step": 1223 + }, + { + "epoch": 0.006509357200976403, + "grad_norm": 13.172407150268555, + "learning_rate": 8.18277426694536e-07, + "loss": 0.4929, + "num_input_tokens_seen": 2566914048, + "step": 1224 + }, + { + "epoch": 0.0067805804176837535, + "grad_norm": 14.171710014343262, + "learning_rate": 8.179532148718483e-07, + "loss": 0.5016, + "num_input_tokens_seen": 2569011200, + "step": 1225 + }, + { + "epoch": 0.007051803634391104, + "grad_norm": 25.743188858032227, + "learning_rate": 8.176287874106097e-07, + "loss": 0.6181, + "num_input_tokens_seen": 2571108352, + "step": 1226 + }, + { + "epoch": 0.007323026851098454, + "grad_norm": 16.81450653076172, + "learning_rate": 8.173041445719069e-07, + "loss": 0.6927, + "num_input_tokens_seen": 2573205504, + "step": 1227 + }, + { + "epoch": 0.007594250067805804, + "grad_norm": 21.463520050048828, + "learning_rate": 8.169792866170003e-07, + "loss": 0.817, + "num_input_tokens_seen": 2575302656, + "step": 1228 + }, + { + "epoch": 0.007865473284513154, + "grad_norm": 12.594096183776855, + "learning_rate": 8.166542138073232e-07, + "loss": 0.5051, + "num_input_tokens_seen": 2577399808, + "step": 1229 + }, + { + "epoch": 0.008136696501220505, + "grad_norm": 14.551346778869629, + "learning_rate": 8.163289264044817e-07, + "loss": 0.5802, + "num_input_tokens_seen": 2579496960, + "step": 1230 + }, + { + "epoch": 0.008407919717927854, + "grad_norm": 14.674225807189941, + "learning_rate": 8.160034246702548e-07, + "loss": 0.5542, + "num_input_tokens_seen": 2581594112, + "step": 1231 + }, + { + "epoch": 0.008679142934635205, + "grad_norm": 14.190738677978516, + "learning_rate": 8.156777088665939e-07, + "loss": 0.44, + "num_input_tokens_seen": 2583691264, + "step": 1232 + }, + { + "epoch": 0.008950366151342554, + "grad_norm": 11.013334274291992, + "learning_rate": 8.153517792556226e-07, + "loss": 0.2932, + "num_input_tokens_seen": 2585788416, + "step": 1233 + }, + { + "epoch": 0.009221589368049905, + "grad_norm": 16.620725631713867, + "learning_rate": 8.15025636099637e-07, + "loss": 0.6058, + "num_input_tokens_seen": 2587885568, + "step": 1234 + }, + { + "epoch": 0.009492812584757255, + "grad_norm": 13.847301483154297, + "learning_rate": 8.146992796611042e-07, + "loss": 0.3527, + "num_input_tokens_seen": 2589982720, + "step": 1235 + }, + { + "epoch": 0.009764035801464606, + "grad_norm": 18.135631561279297, + "learning_rate": 8.143727102026638e-07, + "loss": 0.852, + "num_input_tokens_seen": 2592079872, + "step": 1236 + }, + { + "epoch": 0.010035259018171955, + "grad_norm": 11.433444023132324, + "learning_rate": 8.140459279871264e-07, + "loss": 0.2981, + "num_input_tokens_seen": 2594177024, + "step": 1237 + }, + { + "epoch": 0.010306482234879306, + "grad_norm": 20.82098960876465, + "learning_rate": 8.137189332774738e-07, + "loss": 0.7318, + "num_input_tokens_seen": 2596274176, + "step": 1238 + }, + { + "epoch": 0.010577705451586655, + "grad_norm": 12.118500709533691, + "learning_rate": 8.133917263368589e-07, + "loss": 0.3293, + "num_input_tokens_seen": 2598371328, + "step": 1239 + }, + { + "epoch": 0.010848928668294006, + "grad_norm": 19.532363891601562, + "learning_rate": 8.130643074286056e-07, + "loss": 0.458, + "num_input_tokens_seen": 2600468480, + "step": 1240 + }, + { + "epoch": 0.011120151885001357, + "grad_norm": 16.201066970825195, + "learning_rate": 8.127366768162077e-07, + "loss": 0.4731, + "num_input_tokens_seen": 2602565632, + "step": 1241 + }, + { + "epoch": 0.011391375101708706, + "grad_norm": 13.628470420837402, + "learning_rate": 8.124088347633304e-07, + "loss": 0.4137, + "num_input_tokens_seen": 2604662784, + "step": 1242 + }, + { + "epoch": 0.011662598318416057, + "grad_norm": 12.295647621154785, + "learning_rate": 8.120807815338083e-07, + "loss": 0.3935, + "num_input_tokens_seen": 2606759936, + "step": 1243 + }, + { + "epoch": 0.011933821535123406, + "grad_norm": 12.40659236907959, + "learning_rate": 8.11752517391646e-07, + "loss": 0.3719, + "num_input_tokens_seen": 2608857088, + "step": 1244 + }, + { + "epoch": 0.012205044751830757, + "grad_norm": 12.558025360107422, + "learning_rate": 8.114240426010183e-07, + "loss": 0.3506, + "num_input_tokens_seen": 2610954240, + "step": 1245 + }, + { + "epoch": 0.012476267968538107, + "grad_norm": 18.552087783813477, + "learning_rate": 8.11095357426269e-07, + "loss": 0.6331, + "num_input_tokens_seen": 2613051392, + "step": 1246 + }, + { + "epoch": 0.012747491185245458, + "grad_norm": 14.407464981079102, + "learning_rate": 8.107664621319113e-07, + "loss": 0.5219, + "num_input_tokens_seen": 2615148544, + "step": 1247 + }, + { + "epoch": 0.013018714401952807, + "grad_norm": 14.350251197814941, + "learning_rate": 8.10437356982628e-07, + "loss": 0.4026, + "num_input_tokens_seen": 2617245696, + "step": 1248 + }, + { + "epoch": 0.013289937618660158, + "grad_norm": 13.879180908203125, + "learning_rate": 8.1010804224327e-07, + "loss": 0.4596, + "num_input_tokens_seen": 2619342848, + "step": 1249 + }, + { + "epoch": 0.013561160835367507, + "grad_norm": 12.347127914428711, + "learning_rate": 8.097785181788574e-07, + "loss": 0.3675, + "num_input_tokens_seen": 2621440000, + "step": 1250 + }, + { + "epoch": 0.013832384052074858, + "grad_norm": 13.129162788391113, + "learning_rate": 8.09448785054579e-07, + "loss": 0.3485, + "num_input_tokens_seen": 2623537152, + "step": 1251 + }, + { + "epoch": 0.014103607268782207, + "grad_norm": 19.645524978637695, + "learning_rate": 8.091188431357908e-07, + "loss": 0.9547, + "num_input_tokens_seen": 2625634304, + "step": 1252 + }, + { + "epoch": 0.014374830485489558, + "grad_norm": 19.11149787902832, + "learning_rate": 8.087886926880181e-07, + "loss": 0.8976, + "num_input_tokens_seen": 2627731456, + "step": 1253 + }, + { + "epoch": 0.014646053702196907, + "grad_norm": 12.85779094696045, + "learning_rate": 8.084583339769531e-07, + "loss": 0.4815, + "num_input_tokens_seen": 2629828608, + "step": 1254 + }, + { + "epoch": 0.014917276918904258, + "grad_norm": 13.494819641113281, + "learning_rate": 8.081277672684557e-07, + "loss": 0.4236, + "num_input_tokens_seen": 2631925760, + "step": 1255 + }, + { + "epoch": 0.015188500135611608, + "grad_norm": 13.848788261413574, + "learning_rate": 8.077969928285541e-07, + "loss": 0.5823, + "num_input_tokens_seen": 2634022912, + "step": 1256 + }, + { + "epoch": 0.015459723352318959, + "grad_norm": 15.541146278381348, + "learning_rate": 8.074660109234424e-07, + "loss": 0.456, + "num_input_tokens_seen": 2636120064, + "step": 1257 + }, + { + "epoch": 0.015730946569026308, + "grad_norm": 18.899259567260742, + "learning_rate": 8.071348218194823e-07, + "loss": 0.5888, + "num_input_tokens_seen": 2638217216, + "step": 1258 + }, + { + "epoch": 0.01600216978573366, + "grad_norm": 17.18483543395996, + "learning_rate": 8.068034257832026e-07, + "loss": 0.7188, + "num_input_tokens_seen": 2640314368, + "step": 1259 + }, + { + "epoch": 0.01627339300244101, + "grad_norm": 15.19528579711914, + "learning_rate": 8.064718230812976e-07, + "loss": 0.4918, + "num_input_tokens_seen": 2642411520, + "step": 1260 + }, + { + "epoch": 0.01654461621914836, + "grad_norm": 16.64455795288086, + "learning_rate": 8.06140013980629e-07, + "loss": 0.7336, + "num_input_tokens_seen": 2644508672, + "step": 1261 + }, + { + "epoch": 0.016815839435855708, + "grad_norm": 10.2529296875, + "learning_rate": 8.05807998748224e-07, + "loss": 0.3127, + "num_input_tokens_seen": 2646605824, + "step": 1262 + }, + { + "epoch": 0.01708706265256306, + "grad_norm": 10.849603652954102, + "learning_rate": 8.05475777651276e-07, + "loss": 0.3275, + "num_input_tokens_seen": 2648702976, + "step": 1263 + }, + { + "epoch": 0.01735828586927041, + "grad_norm": 18.223087310791016, + "learning_rate": 8.051433509571435e-07, + "loss": 0.6497, + "num_input_tokens_seen": 2650800128, + "step": 1264 + }, + { + "epoch": 0.01762950908597776, + "grad_norm": 16.752042770385742, + "learning_rate": 8.04810718933351e-07, + "loss": 0.5918, + "num_input_tokens_seen": 2652897280, + "step": 1265 + }, + { + "epoch": 0.01790073230268511, + "grad_norm": 12.526599884033203, + "learning_rate": 8.044778818475884e-07, + "loss": 0.3929, + "num_input_tokens_seen": 2654994432, + "step": 1266 + }, + { + "epoch": 0.01817195551939246, + "grad_norm": 15.401561737060547, + "learning_rate": 8.0414483996771e-07, + "loss": 0.6053, + "num_input_tokens_seen": 2657091584, + "step": 1267 + }, + { + "epoch": 0.01844317873609981, + "grad_norm": 18.10499382019043, + "learning_rate": 8.038115935617355e-07, + "loss": 0.8571, + "num_input_tokens_seen": 2659188736, + "step": 1268 + }, + { + "epoch": 0.01871440195280716, + "grad_norm": 11.17979621887207, + "learning_rate": 8.034781428978484e-07, + "loss": 0.3619, + "num_input_tokens_seen": 2661285888, + "step": 1269 + }, + { + "epoch": 0.01898562516951451, + "grad_norm": 17.378992080688477, + "learning_rate": 8.031444882443976e-07, + "loss": 0.231, + "num_input_tokens_seen": 2663383040, + "step": 1270 + }, + { + "epoch": 0.01925684838622186, + "grad_norm": 17.07963752746582, + "learning_rate": 8.028106298698957e-07, + "loss": 0.6219, + "num_input_tokens_seen": 2665480192, + "step": 1271 + }, + { + "epoch": 0.01952807160292921, + "grad_norm": 15.965147972106934, + "learning_rate": 8.024765680430188e-07, + "loss": 0.3175, + "num_input_tokens_seen": 2667577344, + "step": 1272 + }, + { + "epoch": 0.019799294819636562, + "grad_norm": 13.383700370788574, + "learning_rate": 8.021423030326075e-07, + "loss": 0.5492, + "num_input_tokens_seen": 2669674496, + "step": 1273 + }, + { + "epoch": 0.02007051803634391, + "grad_norm": 8.79860782623291, + "learning_rate": 8.018078351076653e-07, + "loss": 0.2271, + "num_input_tokens_seen": 2671771648, + "step": 1274 + }, + { + "epoch": 0.02034174125305126, + "grad_norm": 10.821599006652832, + "learning_rate": 8.014731645373595e-07, + "loss": 0.3977, + "num_input_tokens_seen": 2673868800, + "step": 1275 + }, + { + "epoch": 0.02061296446975861, + "grad_norm": 15.333173751831055, + "learning_rate": 8.011382915910203e-07, + "loss": 0.4958, + "num_input_tokens_seen": 2675965952, + "step": 1276 + }, + { + "epoch": 0.020884187686465962, + "grad_norm": 10.879945755004883, + "learning_rate": 8.008032165381403e-07, + "loss": 0.3322, + "num_input_tokens_seen": 2678063104, + "step": 1277 + }, + { + "epoch": 0.02115541090317331, + "grad_norm": 16.250761032104492, + "learning_rate": 8.004679396483756e-07, + "loss": 0.266, + "num_input_tokens_seen": 2680160256, + "step": 1278 + }, + { + "epoch": 0.02142663411988066, + "grad_norm": 20.67441749572754, + "learning_rate": 8.001324611915441e-07, + "loss": 0.82, + "num_input_tokens_seen": 2682257408, + "step": 1279 + }, + { + "epoch": 0.021697857336588012, + "grad_norm": 14.254646301269531, + "learning_rate": 7.99796781437626e-07, + "loss": 0.4862, + "num_input_tokens_seen": 2684354560, + "step": 1280 + }, + { + "epoch": 0.021969080553295363, + "grad_norm": 11.306888580322266, + "learning_rate": 7.994609006567635e-07, + "loss": 0.287, + "num_input_tokens_seen": 2686451712, + "step": 1281 + }, + { + "epoch": 0.022240303770002714, + "grad_norm": 17.26317596435547, + "learning_rate": 7.99124819119261e-07, + "loss": 0.4729, + "num_input_tokens_seen": 2688548864, + "step": 1282 + }, + { + "epoch": 0.02251152698671006, + "grad_norm": 19.987943649291992, + "learning_rate": 7.987885370955841e-07, + "loss": 0.6808, + "num_input_tokens_seen": 2690646016, + "step": 1283 + }, + { + "epoch": 0.022782750203417412, + "grad_norm": 16.031089782714844, + "learning_rate": 7.984520548563594e-07, + "loss": 0.6981, + "num_input_tokens_seen": 2692743168, + "step": 1284 + }, + { + "epoch": 0.023053973420124763, + "grad_norm": 14.085775375366211, + "learning_rate": 7.981153726723755e-07, + "loss": 0.4403, + "num_input_tokens_seen": 2694840320, + "step": 1285 + }, + { + "epoch": 0.023325196636832114, + "grad_norm": 9.78898811340332, + "learning_rate": 7.977784908145809e-07, + "loss": 0.3044, + "num_input_tokens_seen": 2696937472, + "step": 1286 + }, + { + "epoch": 0.023596419853539462, + "grad_norm": 9.197432518005371, + "learning_rate": 7.974414095540858e-07, + "loss": 0.1947, + "num_input_tokens_seen": 2699034624, + "step": 1287 + }, + { + "epoch": 0.023867643070246813, + "grad_norm": 9.063929557800293, + "learning_rate": 7.971041291621601e-07, + "loss": 0.3043, + "num_input_tokens_seen": 2701131776, + "step": 1288 + }, + { + "epoch": 0.024138866286954164, + "grad_norm": 15.241850852966309, + "learning_rate": 7.967666499102341e-07, + "loss": 0.4811, + "num_input_tokens_seen": 2703228928, + "step": 1289 + }, + { + "epoch": 0.024410089503661515, + "grad_norm": 12.951624870300293, + "learning_rate": 7.964289720698986e-07, + "loss": 0.375, + "num_input_tokens_seen": 2705326080, + "step": 1290 + }, + { + "epoch": 0.024681312720368862, + "grad_norm": 15.193502426147461, + "learning_rate": 7.960910959129037e-07, + "loss": 0.6131, + "num_input_tokens_seen": 2707423232, + "step": 1291 + }, + { + "epoch": 0.024952535937076213, + "grad_norm": 16.782381057739258, + "learning_rate": 7.957530217111591e-07, + "loss": 0.5336, + "num_input_tokens_seen": 2709520384, + "step": 1292 + }, + { + "epoch": 0.025223759153783564, + "grad_norm": 19.65116310119629, + "learning_rate": 7.954147497367343e-07, + "loss": 0.8438, + "num_input_tokens_seen": 2711617536, + "step": 1293 + }, + { + "epoch": 0.025494982370490915, + "grad_norm": 13.590276718139648, + "learning_rate": 7.950762802618576e-07, + "loss": 0.4738, + "num_input_tokens_seen": 2713714688, + "step": 1294 + }, + { + "epoch": 0.025766205587198263, + "grad_norm": 15.773996353149414, + "learning_rate": 7.947376135589164e-07, + "loss": 0.595, + "num_input_tokens_seen": 2715811840, + "step": 1295 + }, + { + "epoch": 0.026037428803905614, + "grad_norm": 23.006914138793945, + "learning_rate": 7.943987499004569e-07, + "loss": 0.7875, + "num_input_tokens_seen": 2717908992, + "step": 1296 + }, + { + "epoch": 0.026308652020612965, + "grad_norm": 18.493175506591797, + "learning_rate": 7.940596895591835e-07, + "loss": 0.6582, + "num_input_tokens_seen": 2720006144, + "step": 1297 + }, + { + "epoch": 0.026579875237320316, + "grad_norm": 17.077844619750977, + "learning_rate": 7.93720432807959e-07, + "loss": 0.3572, + "num_input_tokens_seen": 2722103296, + "step": 1298 + }, + { + "epoch": 0.026851098454027666, + "grad_norm": 17.862571716308594, + "learning_rate": 7.933809799198045e-07, + "loss": 0.6255, + "num_input_tokens_seen": 2724200448, + "step": 1299 + }, + { + "epoch": 0.027122321670735014, + "grad_norm": 12.72701644897461, + "learning_rate": 7.930413311678987e-07, + "loss": 0.4388, + "num_input_tokens_seen": 2726297600, + "step": 1300 + }, + { + "epoch": 0.027393544887442365, + "grad_norm": 27.155799865722656, + "learning_rate": 7.92701486825578e-07, + "loss": 0.3302, + "num_input_tokens_seen": 2728394752, + "step": 1301 + }, + { + "epoch": 0.027664768104149716, + "grad_norm": 17.29124641418457, + "learning_rate": 7.923614471663361e-07, + "loss": 0.5452, + "num_input_tokens_seen": 2730491904, + "step": 1302 + }, + { + "epoch": 0.027935991320857067, + "grad_norm": 10.107339859008789, + "learning_rate": 7.920212124638241e-07, + "loss": 0.2633, + "num_input_tokens_seen": 2732589056, + "step": 1303 + }, + { + "epoch": 0.028207214537564414, + "grad_norm": 12.023981094360352, + "learning_rate": 7.916807829918499e-07, + "loss": 0.3122, + "num_input_tokens_seen": 2734686208, + "step": 1304 + }, + { + "epoch": 0.028478437754271765, + "grad_norm": 22.489961624145508, + "learning_rate": 7.913401590243781e-07, + "loss": 0.3229, + "num_input_tokens_seen": 2736783360, + "step": 1305 + }, + { + "epoch": 0.028749660970979116, + "grad_norm": 28.63505744934082, + "learning_rate": 7.909993408355302e-07, + "loss": 0.5431, + "num_input_tokens_seen": 2738880512, + "step": 1306 + }, + { + "epoch": 0.029020884187686467, + "grad_norm": 17.687850952148438, + "learning_rate": 7.906583286995834e-07, + "loss": 0.6771, + "num_input_tokens_seen": 2740977664, + "step": 1307 + }, + { + "epoch": 0.029292107404393815, + "grad_norm": 10.240657806396484, + "learning_rate": 7.903171228909714e-07, + "loss": 0.276, + "num_input_tokens_seen": 2743074816, + "step": 1308 + }, + { + "epoch": 0.029563330621101166, + "grad_norm": 12.936287879943848, + "learning_rate": 7.899757236842836e-07, + "loss": 0.3849, + "num_input_tokens_seen": 2745171968, + "step": 1309 + }, + { + "epoch": 0.029834553837808517, + "grad_norm": 19.912282943725586, + "learning_rate": 7.89634131354265e-07, + "loss": 0.7212, + "num_input_tokens_seen": 2747269120, + "step": 1310 + }, + { + "epoch": 0.030105777054515868, + "grad_norm": 16.74636459350586, + "learning_rate": 7.892923461758165e-07, + "loss": 0.5772, + "num_input_tokens_seen": 2749366272, + "step": 1311 + }, + { + "epoch": 0.030377000271223215, + "grad_norm": 13.208150863647461, + "learning_rate": 7.889503684239933e-07, + "loss": 0.4566, + "num_input_tokens_seen": 2751463424, + "step": 1312 + }, + { + "epoch": 0.030648223487930566, + "grad_norm": 9.817733764648438, + "learning_rate": 7.886081983740066e-07, + "loss": 0.2727, + "num_input_tokens_seen": 2753560576, + "step": 1313 + }, + { + "epoch": 0.030919446704637917, + "grad_norm": 15.745170593261719, + "learning_rate": 7.882658363012214e-07, + "loss": 0.5364, + "num_input_tokens_seen": 2755657728, + "step": 1314 + }, + { + "epoch": 0.031190669921345268, + "grad_norm": 19.3269100189209, + "learning_rate": 7.879232824811579e-07, + "loss": 0.4603, + "num_input_tokens_seen": 2757754880, + "step": 1315 + }, + { + "epoch": 0.031461893138052616, + "grad_norm": 21.108278274536133, + "learning_rate": 7.875805371894904e-07, + "loss": 0.8861, + "num_input_tokens_seen": 2759852032, + "step": 1316 + }, + { + "epoch": 0.03173311635475997, + "grad_norm": 21.515796661376953, + "learning_rate": 7.872376007020469e-07, + "loss": 0.5236, + "num_input_tokens_seen": 2761949184, + "step": 1317 + }, + { + "epoch": 0.03200433957146732, + "grad_norm": 11.72025203704834, + "learning_rate": 7.8689447329481e-07, + "loss": 0.3716, + "num_input_tokens_seen": 2764046336, + "step": 1318 + }, + { + "epoch": 0.032275562788174665, + "grad_norm": 19.262054443359375, + "learning_rate": 7.865511552439156e-07, + "loss": 0.5643, + "num_input_tokens_seen": 2766143488, + "step": 1319 + }, + { + "epoch": 0.03254678600488202, + "grad_norm": 13.74807357788086, + "learning_rate": 7.862076468256529e-07, + "loss": 0.5372, + "num_input_tokens_seen": 2768240640, + "step": 1320 + }, + { + "epoch": 0.03281800922158937, + "grad_norm": 15.6185941696167, + "learning_rate": 7.858639483164643e-07, + "loss": 0.4638, + "num_input_tokens_seen": 2770337792, + "step": 1321 + }, + { + "epoch": 0.03308923243829672, + "grad_norm": 14.039484024047852, + "learning_rate": 7.855200599929456e-07, + "loss": 0.5236, + "num_input_tokens_seen": 2772434944, + "step": 1322 + }, + { + "epoch": 0.03336045565500407, + "grad_norm": 14.383334159851074, + "learning_rate": 7.851759821318447e-07, + "loss": 0.4256, + "num_input_tokens_seen": 2774532096, + "step": 1323 + }, + { + "epoch": 0.033631678871711417, + "grad_norm": 18.412899017333984, + "learning_rate": 7.848317150100624e-07, + "loss": 0.5748, + "num_input_tokens_seen": 2776629248, + "step": 1324 + }, + { + "epoch": 0.03390290208841877, + "grad_norm": 12.141471862792969, + "learning_rate": 7.84487258904652e-07, + "loss": 0.4111, + "num_input_tokens_seen": 2778726400, + "step": 1325 + }, + { + "epoch": 0.03417412530512612, + "grad_norm": 22.77025032043457, + "learning_rate": 7.841426140928184e-07, + "loss": 0.8071, + "num_input_tokens_seen": 2780823552, + "step": 1326 + }, + { + "epoch": 0.034445348521833466, + "grad_norm": 27.863340377807617, + "learning_rate": 7.837977808519189e-07, + "loss": 0.7943, + "num_input_tokens_seen": 2782920704, + "step": 1327 + }, + { + "epoch": 0.03471657173854082, + "grad_norm": 19.592632293701172, + "learning_rate": 7.834527594594618e-07, + "loss": 0.6998, + "num_input_tokens_seen": 2785017856, + "step": 1328 + }, + { + "epoch": 0.03498779495524817, + "grad_norm": 19.130634307861328, + "learning_rate": 7.831075501931078e-07, + "loss": 0.5628, + "num_input_tokens_seen": 2787115008, + "step": 1329 + }, + { + "epoch": 0.03525901817195552, + "grad_norm": 16.67599105834961, + "learning_rate": 7.827621533306677e-07, + "loss": 0.6865, + "num_input_tokens_seen": 2789212160, + "step": 1330 + }, + { + "epoch": 0.03553024138866287, + "grad_norm": 13.267627716064453, + "learning_rate": 7.824165691501036e-07, + "loss": 0.4898, + "num_input_tokens_seen": 2791309312, + "step": 1331 + }, + { + "epoch": 0.03580146460537022, + "grad_norm": 17.262802124023438, + "learning_rate": 7.820707979295289e-07, + "loss": 0.3767, + "num_input_tokens_seen": 2793406464, + "step": 1332 + }, + { + "epoch": 0.03607268782207757, + "grad_norm": 21.728710174560547, + "learning_rate": 7.81724839947207e-07, + "loss": 0.7031, + "num_input_tokens_seen": 2795503616, + "step": 1333 + }, + { + "epoch": 0.03634391103878492, + "grad_norm": 20.991073608398438, + "learning_rate": 7.813786954815517e-07, + "loss": 0.4953, + "num_input_tokens_seen": 2797600768, + "step": 1334 + }, + { + "epoch": 0.03661513425549227, + "grad_norm": 14.685553550720215, + "learning_rate": 7.810323648111269e-07, + "loss": 0.3939, + "num_input_tokens_seen": 2799697920, + "step": 1335 + }, + { + "epoch": 0.03688635747219962, + "grad_norm": 14.160144805908203, + "learning_rate": 7.806858482146463e-07, + "loss": 0.4822, + "num_input_tokens_seen": 2801795072, + "step": 1336 + }, + { + "epoch": 0.03715758068890697, + "grad_norm": 16.465970993041992, + "learning_rate": 7.803391459709733e-07, + "loss": 0.5157, + "num_input_tokens_seen": 2803892224, + "step": 1337 + }, + { + "epoch": 0.03742880390561432, + "grad_norm": 12.798428535461426, + "learning_rate": 7.799922583591206e-07, + "loss": 0.4207, + "num_input_tokens_seen": 2805989376, + "step": 1338 + }, + { + "epoch": 0.03770002712232167, + "grad_norm": 12.751097679138184, + "learning_rate": 7.796451856582504e-07, + "loss": 0.4589, + "num_input_tokens_seen": 2808086528, + "step": 1339 + }, + { + "epoch": 0.03797125033902902, + "grad_norm": 15.334579467773438, + "learning_rate": 7.792979281476734e-07, + "loss": 0.3559, + "num_input_tokens_seen": 2810183680, + "step": 1340 + }, + { + "epoch": 0.03824247355573637, + "grad_norm": 13.238780975341797, + "learning_rate": 7.789504861068492e-07, + "loss": 0.4655, + "num_input_tokens_seen": 2812280832, + "step": 1341 + }, + { + "epoch": 0.03851369677244372, + "grad_norm": 15.215901374816895, + "learning_rate": 7.78602859815386e-07, + "loss": 0.5126, + "num_input_tokens_seen": 2814377984, + "step": 1342 + }, + { + "epoch": 0.038784919989151075, + "grad_norm": 13.011480331420898, + "learning_rate": 7.782550495530402e-07, + "loss": 0.327, + "num_input_tokens_seen": 2816475136, + "step": 1343 + }, + { + "epoch": 0.03905614320585842, + "grad_norm": 18.614702224731445, + "learning_rate": 7.779070555997162e-07, + "loss": 0.7977, + "num_input_tokens_seen": 2818572288, + "step": 1344 + }, + { + "epoch": 0.03932736642256577, + "grad_norm": 18.841096878051758, + "learning_rate": 7.775588782354666e-07, + "loss": 0.7646, + "num_input_tokens_seen": 2820669440, + "step": 1345 + }, + { + "epoch": 0.039598589639273124, + "grad_norm": 20.057369232177734, + "learning_rate": 7.77210517740491e-07, + "loss": 0.7922, + "num_input_tokens_seen": 2822766592, + "step": 1346 + }, + { + "epoch": 0.03986981285598047, + "grad_norm": 10.478140830993652, + "learning_rate": 7.768619743951366e-07, + "loss": 0.2467, + "num_input_tokens_seen": 2824863744, + "step": 1347 + }, + { + "epoch": 0.04014103607268782, + "grad_norm": 10.633049964904785, + "learning_rate": 7.765132484798978e-07, + "loss": 0.3628, + "num_input_tokens_seen": 2826960896, + "step": 1348 + }, + { + "epoch": 0.040412259289395173, + "grad_norm": 18.010910034179688, + "learning_rate": 7.761643402754163e-07, + "loss": 0.6482, + "num_input_tokens_seen": 2829058048, + "step": 1349 + }, + { + "epoch": 0.04068348250610252, + "grad_norm": 20.25141716003418, + "learning_rate": 7.7581525006248e-07, + "loss": 0.5458, + "num_input_tokens_seen": 2831155200, + "step": 1350 + }, + { + "epoch": 0.040954705722809875, + "grad_norm": 16.61979866027832, + "learning_rate": 7.754659781220232e-07, + "loss": 0.5326, + "num_input_tokens_seen": 2833252352, + "step": 1351 + }, + { + "epoch": 0.04122592893951722, + "grad_norm": 11.642721176147461, + "learning_rate": 7.751165247351269e-07, + "loss": 0.3478, + "num_input_tokens_seen": 2835349504, + "step": 1352 + }, + { + "epoch": 0.04149715215622457, + "grad_norm": 15.440996170043945, + "learning_rate": 7.747668901830178e-07, + "loss": 0.559, + "num_input_tokens_seen": 2837446656, + "step": 1353 + }, + { + "epoch": 0.041768375372931925, + "grad_norm": 15.072626113891602, + "learning_rate": 7.744170747470685e-07, + "loss": 0.5729, + "num_input_tokens_seen": 2839543808, + "step": 1354 + }, + { + "epoch": 0.04203959858963927, + "grad_norm": 19.414575576782227, + "learning_rate": 7.740670787087972e-07, + "loss": 0.9754, + "num_input_tokens_seen": 2841640960, + "step": 1355 + }, + { + "epoch": 0.04231082180634662, + "grad_norm": 14.357555389404297, + "learning_rate": 7.737169023498672e-07, + "loss": 0.3421, + "num_input_tokens_seen": 2843738112, + "step": 1356 + }, + { + "epoch": 0.042582045023053974, + "grad_norm": 13.045211791992188, + "learning_rate": 7.733665459520873e-07, + "loss": 0.3127, + "num_input_tokens_seen": 2845835264, + "step": 1357 + }, + { + "epoch": 0.04285326823976132, + "grad_norm": 11.419096946716309, + "learning_rate": 7.730160097974109e-07, + "loss": 0.3253, + "num_input_tokens_seen": 2847932416, + "step": 1358 + }, + { + "epoch": 0.043124491456468676, + "grad_norm": 14.921306610107422, + "learning_rate": 7.726652941679365e-07, + "loss": 0.46, + "num_input_tokens_seen": 2850029568, + "step": 1359 + }, + { + "epoch": 0.043395714673176024, + "grad_norm": 12.84131908416748, + "learning_rate": 7.723143993459061e-07, + "loss": 0.4562, + "num_input_tokens_seen": 2852126720, + "step": 1360 + }, + { + "epoch": 0.04366693788988337, + "grad_norm": 15.976582527160645, + "learning_rate": 7.719633256137067e-07, + "loss": 0.5137, + "num_input_tokens_seen": 2854223872, + "step": 1361 + }, + { + "epoch": 0.043938161106590726, + "grad_norm": 10.385905265808105, + "learning_rate": 7.716120732538696e-07, + "loss": 0.2903, + "num_input_tokens_seen": 2856321024, + "step": 1362 + }, + { + "epoch": 0.04420938432329807, + "grad_norm": 17.1414737701416, + "learning_rate": 7.712606425490687e-07, + "loss": 0.7363, + "num_input_tokens_seen": 2858418176, + "step": 1363 + }, + { + "epoch": 0.04448060754000543, + "grad_norm": 20.92815399169922, + "learning_rate": 7.70909033782122e-07, + "loss": 0.5892, + "num_input_tokens_seen": 2860515328, + "step": 1364 + }, + { + "epoch": 0.044751830756712775, + "grad_norm": 15.580632209777832, + "learning_rate": 7.705572472359913e-07, + "loss": 0.5212, + "num_input_tokens_seen": 2862612480, + "step": 1365 + }, + { + "epoch": 0.04502305397342012, + "grad_norm": 13.09688949584961, + "learning_rate": 7.702052831937811e-07, + "loss": 0.5206, + "num_input_tokens_seen": 2864709632, + "step": 1366 + }, + { + "epoch": 0.04529427719012748, + "grad_norm": 17.451766967773438, + "learning_rate": 7.698531419387382e-07, + "loss": 0.6266, + "num_input_tokens_seen": 2866806784, + "step": 1367 + }, + { + "epoch": 0.045565500406834825, + "grad_norm": 16.42080307006836, + "learning_rate": 7.695008237542526e-07, + "loss": 0.3435, + "num_input_tokens_seen": 2868903936, + "step": 1368 + }, + { + "epoch": 0.04583672362354217, + "grad_norm": 21.22711753845215, + "learning_rate": 7.691483289238569e-07, + "loss": 0.9744, + "num_input_tokens_seen": 2871001088, + "step": 1369 + }, + { + "epoch": 0.04610794684024953, + "grad_norm": 14.438827514648438, + "learning_rate": 7.687956577312251e-07, + "loss": 0.5828, + "num_input_tokens_seen": 2873098240, + "step": 1370 + }, + { + "epoch": 0.046379170056956874, + "grad_norm": 13.411489486694336, + "learning_rate": 7.684428104601739e-07, + "loss": 0.3964, + "num_input_tokens_seen": 2875195392, + "step": 1371 + }, + { + "epoch": 0.04665039327366423, + "grad_norm": 11.179100036621094, + "learning_rate": 7.680897873946611e-07, + "loss": 0.3965, + "num_input_tokens_seen": 2877292544, + "step": 1372 + }, + { + "epoch": 0.046921616490371576, + "grad_norm": 13.811481475830078, + "learning_rate": 7.677365888187864e-07, + "loss": 0.472, + "num_input_tokens_seen": 2879389696, + "step": 1373 + }, + { + "epoch": 0.047192839707078924, + "grad_norm": 13.83006763458252, + "learning_rate": 7.673832150167906e-07, + "loss": 0.4331, + "num_input_tokens_seen": 2881486848, + "step": 1374 + }, + { + "epoch": 0.04746406292378628, + "grad_norm": 11.234846115112305, + "learning_rate": 7.670296662730552e-07, + "loss": 0.3489, + "num_input_tokens_seen": 2883584000, + "step": 1375 + }, + { + "epoch": 0.047735286140493625, + "grad_norm": 15.245635986328125, + "learning_rate": 7.66675942872103e-07, + "loss": 0.5866, + "num_input_tokens_seen": 2885681152, + "step": 1376 + }, + { + "epoch": 0.04800650935720097, + "grad_norm": 16.57020378112793, + "learning_rate": 7.663220450985973e-07, + "loss": 0.4715, + "num_input_tokens_seen": 2887778304, + "step": 1377 + }, + { + "epoch": 0.04827773257390833, + "grad_norm": 18.864973068237305, + "learning_rate": 7.659679732373413e-07, + "loss": 0.6713, + "num_input_tokens_seen": 2889875456, + "step": 1378 + }, + { + "epoch": 0.048548955790615675, + "grad_norm": 10.371908187866211, + "learning_rate": 7.656137275732786e-07, + "loss": 0.3513, + "num_input_tokens_seen": 2891972608, + "step": 1379 + }, + { + "epoch": 0.04882017900732303, + "grad_norm": 18.576017379760742, + "learning_rate": 7.652593083914927e-07, + "loss": 0.628, + "num_input_tokens_seen": 2894069760, + "step": 1380 + }, + { + "epoch": 0.04909140222403038, + "grad_norm": 17.365699768066406, + "learning_rate": 7.649047159772064e-07, + "loss": 0.6301, + "num_input_tokens_seen": 2896166912, + "step": 1381 + }, + { + "epoch": 0.049362625440737724, + "grad_norm": 13.766050338745117, + "learning_rate": 7.645499506157827e-07, + "loss": 0.432, + "num_input_tokens_seen": 2898264064, + "step": 1382 + }, + { + "epoch": 0.04963384865744508, + "grad_norm": 15.873371124267578, + "learning_rate": 7.641950125927228e-07, + "loss": 0.613, + "num_input_tokens_seen": 2900361216, + "step": 1383 + }, + { + "epoch": 0.049905071874152426, + "grad_norm": 10.771862983703613, + "learning_rate": 7.638399021936675e-07, + "loss": 0.3301, + "num_input_tokens_seen": 2902458368, + "step": 1384 + }, + { + "epoch": 0.05017629509085978, + "grad_norm": 26.106536865234375, + "learning_rate": 7.634846197043963e-07, + "loss": 0.4127, + "num_input_tokens_seen": 2904555520, + "step": 1385 + }, + { + "epoch": 0.05044751830756713, + "grad_norm": 16.403791427612305, + "learning_rate": 7.631291654108266e-07, + "loss": 0.6469, + "num_input_tokens_seen": 2906652672, + "step": 1386 + }, + { + "epoch": 0.050718741524274476, + "grad_norm": 25.416994094848633, + "learning_rate": 7.627735395990149e-07, + "loss": 0.8975, + "num_input_tokens_seen": 2908749824, + "step": 1387 + }, + { + "epoch": 0.05098996474098183, + "grad_norm": 26.157794952392578, + "learning_rate": 7.624177425551552e-07, + "loss": 1.0145, + "num_input_tokens_seen": 2910846976, + "step": 1388 + }, + { + "epoch": 0.05126118795768918, + "grad_norm": 20.74803924560547, + "learning_rate": 7.620617745655793e-07, + "loss": 0.7406, + "num_input_tokens_seen": 2912944128, + "step": 1389 + }, + { + "epoch": 0.051532411174396525, + "grad_norm": 15.389482498168945, + "learning_rate": 7.617056359167568e-07, + "loss": 0.4884, + "num_input_tokens_seen": 2915041280, + "step": 1390 + }, + { + "epoch": 0.05180363439110388, + "grad_norm": 15.435873985290527, + "learning_rate": 7.613493268952947e-07, + "loss": 0.5215, + "num_input_tokens_seen": 2917138432, + "step": 1391 + }, + { + "epoch": 0.05207485760781123, + "grad_norm": 12.54256534576416, + "learning_rate": 7.609928477879365e-07, + "loss": 0.3087, + "num_input_tokens_seen": 2919235584, + "step": 1392 + }, + { + "epoch": 0.05234608082451858, + "grad_norm": 15.502516746520996, + "learning_rate": 7.606361988815633e-07, + "loss": 0.5513, + "num_input_tokens_seen": 2921332736, + "step": 1393 + }, + { + "epoch": 0.05261730404122593, + "grad_norm": 13.708436965942383, + "learning_rate": 7.602793804631927e-07, + "loss": 0.273, + "num_input_tokens_seen": 2923429888, + "step": 1394 + }, + { + "epoch": 0.05288852725793328, + "grad_norm": 15.411025047302246, + "learning_rate": 7.599223928199781e-07, + "loss": 0.4805, + "num_input_tokens_seen": 2925527040, + "step": 1395 + }, + { + "epoch": 0.05315975047464063, + "grad_norm": 16.42882537841797, + "learning_rate": 7.595652362392103e-07, + "loss": 0.6015, + "num_input_tokens_seen": 2927624192, + "step": 1396 + }, + { + "epoch": 0.05343097369134798, + "grad_norm": 17.059511184692383, + "learning_rate": 7.592079110083146e-07, + "loss": 0.711, + "num_input_tokens_seen": 2929721344, + "step": 1397 + }, + { + "epoch": 0.05370219690805533, + "grad_norm": 15.640413284301758, + "learning_rate": 7.588504174148532e-07, + "loss": 0.3908, + "num_input_tokens_seen": 2931818496, + "step": 1398 + }, + { + "epoch": 0.05397342012476268, + "grad_norm": 18.543855667114258, + "learning_rate": 7.584927557465233e-07, + "loss": 0.7099, + "num_input_tokens_seen": 2933915648, + "step": 1399 + }, + { + "epoch": 0.05424464334147003, + "grad_norm": 11.331782341003418, + "learning_rate": 7.581349262911573e-07, + "loss": 0.3008, + "num_input_tokens_seen": 2936012800, + "step": 1400 + }, + { + "epoch": 0.05451586655817738, + "grad_norm": 13.517926216125488, + "learning_rate": 7.577769293367226e-07, + "loss": 0.451, + "num_input_tokens_seen": 2938109952, + "step": 1401 + }, + { + "epoch": 0.05478708977488473, + "grad_norm": 15.007146835327148, + "learning_rate": 7.574187651713218e-07, + "loss": 0.4981, + "num_input_tokens_seen": 2940207104, + "step": 1402 + }, + { + "epoch": 0.05505831299159208, + "grad_norm": 13.51999282836914, + "learning_rate": 7.570604340831916e-07, + "loss": 0.4342, + "num_input_tokens_seen": 2942304256, + "step": 1403 + }, + { + "epoch": 0.05532953620829943, + "grad_norm": 9.16441822052002, + "learning_rate": 7.567019363607032e-07, + "loss": 0.2665, + "num_input_tokens_seen": 2944401408, + "step": 1404 + }, + { + "epoch": 0.05560075942500678, + "grad_norm": 13.39521598815918, + "learning_rate": 7.563432722923621e-07, + "loss": 0.4227, + "num_input_tokens_seen": 2946498560, + "step": 1405 + }, + { + "epoch": 0.055871982641714134, + "grad_norm": 12.353023529052734, + "learning_rate": 7.559844421668074e-07, + "loss": 0.3066, + "num_input_tokens_seen": 2948595712, + "step": 1406 + }, + { + "epoch": 0.05614320585842148, + "grad_norm": 14.080292701721191, + "learning_rate": 7.556254462728122e-07, + "loss": 0.5427, + "num_input_tokens_seen": 2950692864, + "step": 1407 + }, + { + "epoch": 0.05641442907512883, + "grad_norm": 14.238213539123535, + "learning_rate": 7.552662848992822e-07, + "loss": 0.4679, + "num_input_tokens_seen": 2952790016, + "step": 1408 + }, + { + "epoch": 0.05668565229183618, + "grad_norm": 13.333026885986328, + "learning_rate": 7.54906958335257e-07, + "loss": 0.4319, + "num_input_tokens_seen": 2954887168, + "step": 1409 + }, + { + "epoch": 0.05695687550854353, + "grad_norm": 10.981861114501953, + "learning_rate": 7.545474668699091e-07, + "loss": 0.3214, + "num_input_tokens_seen": 2956984320, + "step": 1410 + }, + { + "epoch": 0.05722809872525088, + "grad_norm": 10.62180233001709, + "learning_rate": 7.541878107925435e-07, + "loss": 0.3711, + "num_input_tokens_seen": 2959081472, + "step": 1411 + }, + { + "epoch": 0.05749932194195823, + "grad_norm": 13.643486976623535, + "learning_rate": 7.538279903925977e-07, + "loss": 0.2983, + "num_input_tokens_seen": 2961178624, + "step": 1412 + }, + { + "epoch": 0.05777054515866558, + "grad_norm": 15.046182632446289, + "learning_rate": 7.534680059596414e-07, + "loss": 0.5167, + "num_input_tokens_seen": 2963275776, + "step": 1413 + }, + { + "epoch": 0.058041768375372935, + "grad_norm": 22.363344192504883, + "learning_rate": 7.531078577833765e-07, + "loss": 0.6654, + "num_input_tokens_seen": 2965372928, + "step": 1414 + }, + { + "epoch": 0.05831299159208028, + "grad_norm": 12.185877799987793, + "learning_rate": 7.527475461536363e-07, + "loss": 0.3195, + "num_input_tokens_seen": 2967470080, + "step": 1415 + }, + { + "epoch": 0.05858421480878763, + "grad_norm": 18.2763614654541, + "learning_rate": 7.523870713603864e-07, + "loss": 0.3883, + "num_input_tokens_seen": 2969567232, + "step": 1416 + }, + { + "epoch": 0.058855438025494984, + "grad_norm": 12.003350257873535, + "learning_rate": 7.520264336937227e-07, + "loss": 0.4257, + "num_input_tokens_seen": 2971664384, + "step": 1417 + }, + { + "epoch": 0.05912666124220233, + "grad_norm": 10.348063468933105, + "learning_rate": 7.516656334438727e-07, + "loss": 0.3473, + "num_input_tokens_seen": 2973761536, + "step": 1418 + }, + { + "epoch": 0.059397884458909686, + "grad_norm": 18.062877655029297, + "learning_rate": 7.513046709011951e-07, + "loss": 0.5485, + "num_input_tokens_seen": 2975858688, + "step": 1419 + }, + { + "epoch": 0.059669107675617034, + "grad_norm": 11.781852722167969, + "learning_rate": 7.509435463561785e-07, + "loss": 0.3139, + "num_input_tokens_seen": 2977955840, + "step": 1420 + }, + { + "epoch": 0.05994033089232438, + "grad_norm": 17.999149322509766, + "learning_rate": 7.505822600994423e-07, + "loss": 0.8586, + "num_input_tokens_seen": 2980052992, + "step": 1421 + }, + { + "epoch": 0.060211554109031735, + "grad_norm": 19.359094619750977, + "learning_rate": 7.502208124217357e-07, + "loss": 0.5113, + "num_input_tokens_seen": 2982150144, + "step": 1422 + }, + { + "epoch": 0.06048277732573908, + "grad_norm": 18.1299991607666, + "learning_rate": 7.498592036139383e-07, + "loss": 0.6946, + "num_input_tokens_seen": 2984247296, + "step": 1423 + }, + { + "epoch": 0.06075400054244643, + "grad_norm": 12.68889045715332, + "learning_rate": 7.494974339670591e-07, + "loss": 0.3701, + "num_input_tokens_seen": 2986344448, + "step": 1424 + }, + { + "epoch": 0.061025223759153785, + "grad_norm": 12.457084655761719, + "learning_rate": 7.491355037722365e-07, + "loss": 0.3567, + "num_input_tokens_seen": 2988441600, + "step": 1425 + }, + { + "epoch": 0.06129644697586113, + "grad_norm": 21.13746452331543, + "learning_rate": 7.487734133207382e-07, + "loss": 0.6114, + "num_input_tokens_seen": 2990538752, + "step": 1426 + }, + { + "epoch": 0.06156767019256849, + "grad_norm": 10.113631248474121, + "learning_rate": 7.484111629039607e-07, + "loss": 0.2361, + "num_input_tokens_seen": 2992635904, + "step": 1427 + }, + { + "epoch": 0.061838893409275834, + "grad_norm": 18.88080406188965, + "learning_rate": 7.480487528134292e-07, + "loss": 0.7382, + "num_input_tokens_seen": 2994733056, + "step": 1428 + }, + { + "epoch": 0.06211011662598318, + "grad_norm": 13.798151969909668, + "learning_rate": 7.47686183340798e-07, + "loss": 0.3842, + "num_input_tokens_seen": 2996830208, + "step": 1429 + }, + { + "epoch": 0.062381339842690536, + "grad_norm": 17.79098892211914, + "learning_rate": 7.473234547778489e-07, + "loss": 0.6524, + "num_input_tokens_seen": 2998927360, + "step": 1430 + }, + { + "epoch": 0.06265256305939788, + "grad_norm": 14.120518684387207, + "learning_rate": 7.46960567416492e-07, + "loss": 0.476, + "num_input_tokens_seen": 3001024512, + "step": 1431 + }, + { + "epoch": 0.06292378627610523, + "grad_norm": 12.636813163757324, + "learning_rate": 7.465975215487655e-07, + "loss": 0.3001, + "num_input_tokens_seen": 3003121664, + "step": 1432 + }, + { + "epoch": 0.06319500949281258, + "grad_norm": 18.47852897644043, + "learning_rate": 7.462343174668346e-07, + "loss": 0.6856, + "num_input_tokens_seen": 3005218816, + "step": 1433 + }, + { + "epoch": 0.06346623270951994, + "grad_norm": 15.712651252746582, + "learning_rate": 7.458709554629924e-07, + "loss": 0.569, + "num_input_tokens_seen": 3007315968, + "step": 1434 + }, + { + "epoch": 0.06373745592622729, + "grad_norm": 20.598899841308594, + "learning_rate": 7.455074358296586e-07, + "loss": 0.6715, + "num_input_tokens_seen": 3009413120, + "step": 1435 + }, + { + "epoch": 0.06400867914293464, + "grad_norm": 11.868573188781738, + "learning_rate": 7.451437588593802e-07, + "loss": 0.4039, + "num_input_tokens_seen": 3011510272, + "step": 1436 + }, + { + "epoch": 0.06427990235964198, + "grad_norm": 16.7470703125, + "learning_rate": 7.447799248448303e-07, + "loss": 0.5598, + "num_input_tokens_seen": 3013607424, + "step": 1437 + }, + { + "epoch": 0.06455112557634933, + "grad_norm": 16.11475944519043, + "learning_rate": 7.444159340788088e-07, + "loss": 0.5864, + "num_input_tokens_seen": 3015704576, + "step": 1438 + }, + { + "epoch": 0.06482234879305669, + "grad_norm": 12.091106414794922, + "learning_rate": 7.440517868542417e-07, + "loss": 0.394, + "num_input_tokens_seen": 3017801728, + "step": 1439 + }, + { + "epoch": 0.06509357200976404, + "grad_norm": 13.48214340209961, + "learning_rate": 7.436874834641807e-07, + "loss": 0.365, + "num_input_tokens_seen": 3019898880, + "step": 1440 + }, + { + "epoch": 0.06536479522647139, + "grad_norm": 18.32723045349121, + "learning_rate": 7.433230242018035e-07, + "loss": 0.6249, + "num_input_tokens_seen": 3021996032, + "step": 1441 + }, + { + "epoch": 0.06563601844317873, + "grad_norm": 13.360453605651855, + "learning_rate": 7.429584093604128e-07, + "loss": 0.4254, + "num_input_tokens_seen": 3024093184, + "step": 1442 + }, + { + "epoch": 0.06590724165988608, + "grad_norm": 16.321001052856445, + "learning_rate": 7.425936392334368e-07, + "loss": 0.6535, + "num_input_tokens_seen": 3026190336, + "step": 1443 + }, + { + "epoch": 0.06617846487659344, + "grad_norm": 15.212583541870117, + "learning_rate": 7.422287141144287e-07, + "loss": 0.6856, + "num_input_tokens_seen": 3028287488, + "step": 1444 + }, + { + "epoch": 0.06644968809330079, + "grad_norm": 20.179534912109375, + "learning_rate": 7.418636342970665e-07, + "loss": 0.8147, + "num_input_tokens_seen": 3030384640, + "step": 1445 + }, + { + "epoch": 0.06672091131000814, + "grad_norm": 11.90196418762207, + "learning_rate": 7.414984000751521e-07, + "loss": 0.3257, + "num_input_tokens_seen": 3032481792, + "step": 1446 + }, + { + "epoch": 0.06699213452671549, + "grad_norm": 12.35843563079834, + "learning_rate": 7.411330117426125e-07, + "loss": 0.3769, + "num_input_tokens_seen": 3034578944, + "step": 1447 + }, + { + "epoch": 0.06726335774342283, + "grad_norm": 9.548952102661133, + "learning_rate": 7.407674695934983e-07, + "loss": 0.2638, + "num_input_tokens_seen": 3036676096, + "step": 1448 + }, + { + "epoch": 0.06753458096013018, + "grad_norm": 14.05826187133789, + "learning_rate": 7.404017739219836e-07, + "loss": 0.4847, + "num_input_tokens_seen": 3038773248, + "step": 1449 + }, + { + "epoch": 0.06780580417683754, + "grad_norm": 13.904762268066406, + "learning_rate": 7.400359250223667e-07, + "loss": 0.3108, + "num_input_tokens_seen": 3040870400, + "step": 1450 + }, + { + "epoch": 0.06807702739354489, + "grad_norm": 15.42446517944336, + "learning_rate": 7.396699231890689e-07, + "loss": 0.59, + "num_input_tokens_seen": 3042967552, + "step": 1451 + }, + { + "epoch": 0.06834825061025224, + "grad_norm": 16.7812442779541, + "learning_rate": 7.393037687166342e-07, + "loss": 0.591, + "num_input_tokens_seen": 3045064704, + "step": 1452 + }, + { + "epoch": 0.06861947382695958, + "grad_norm": 14.538595199584961, + "learning_rate": 7.3893746189973e-07, + "loss": 0.4907, + "num_input_tokens_seen": 3047161856, + "step": 1453 + }, + { + "epoch": 0.06889069704366693, + "grad_norm": 15.920239448547363, + "learning_rate": 7.385710030331461e-07, + "loss": 0.4527, + "num_input_tokens_seen": 3049259008, + "step": 1454 + }, + { + "epoch": 0.0691619202603743, + "grad_norm": 11.364521026611328, + "learning_rate": 7.382043924117945e-07, + "loss": 0.4068, + "num_input_tokens_seen": 3051356160, + "step": 1455 + }, + { + "epoch": 0.06943314347708164, + "grad_norm": 15.396533012390137, + "learning_rate": 7.378376303307099e-07, + "loss": 0.285, + "num_input_tokens_seen": 3053453312, + "step": 1456 + }, + { + "epoch": 0.06970436669378899, + "grad_norm": 14.151115417480469, + "learning_rate": 7.374707170850479e-07, + "loss": 0.4752, + "num_input_tokens_seen": 3055550464, + "step": 1457 + }, + { + "epoch": 0.06997558991049634, + "grad_norm": 10.992759704589844, + "learning_rate": 7.371036529700866e-07, + "loss": 0.4462, + "num_input_tokens_seen": 3057647616, + "step": 1458 + }, + { + "epoch": 0.07024681312720368, + "grad_norm": 12.24362564086914, + "learning_rate": 7.367364382812253e-07, + "loss": 0.3646, + "num_input_tokens_seen": 3059744768, + "step": 1459 + }, + { + "epoch": 0.07051803634391104, + "grad_norm": 16.037370681762695, + "learning_rate": 7.363690733139842e-07, + "loss": 0.6663, + "num_input_tokens_seen": 3061841920, + "step": 1460 + }, + { + "epoch": 0.07078925956061839, + "grad_norm": 18.999717712402344, + "learning_rate": 7.360015583640049e-07, + "loss": 0.7606, + "num_input_tokens_seen": 3063939072, + "step": 1461 + }, + { + "epoch": 0.07106048277732574, + "grad_norm": 10.200159072875977, + "learning_rate": 7.356338937270492e-07, + "loss": 0.3097, + "num_input_tokens_seen": 3066036224, + "step": 1462 + }, + { + "epoch": 0.07133170599403309, + "grad_norm": 20.75838279724121, + "learning_rate": 7.352660796989999e-07, + "loss": 0.3529, + "num_input_tokens_seen": 3068133376, + "step": 1463 + }, + { + "epoch": 0.07160292921074043, + "grad_norm": 15.318017959594727, + "learning_rate": 7.348981165758595e-07, + "loss": 0.4486, + "num_input_tokens_seen": 3070230528, + "step": 1464 + }, + { + "epoch": 0.0718741524274478, + "grad_norm": 11.934765815734863, + "learning_rate": 7.345300046537507e-07, + "loss": 0.411, + "num_input_tokens_seen": 3072327680, + "step": 1465 + }, + { + "epoch": 0.07214537564415514, + "grad_norm": 16.084491729736328, + "learning_rate": 7.341617442289159e-07, + "loss": 0.5918, + "num_input_tokens_seen": 3074424832, + "step": 1466 + }, + { + "epoch": 0.07241659886086249, + "grad_norm": 15.078495979309082, + "learning_rate": 7.337933355977175e-07, + "loss": 0.495, + "num_input_tokens_seen": 3076521984, + "step": 1467 + }, + { + "epoch": 0.07268782207756984, + "grad_norm": 12.67852783203125, + "learning_rate": 7.334247790566364e-07, + "loss": 0.3726, + "num_input_tokens_seen": 3078619136, + "step": 1468 + }, + { + "epoch": 0.07295904529427719, + "grad_norm": 12.588438034057617, + "learning_rate": 7.330560749022728e-07, + "loss": 0.364, + "num_input_tokens_seen": 3080716288, + "step": 1469 + }, + { + "epoch": 0.07323026851098453, + "grad_norm": 31.776931762695312, + "learning_rate": 7.326872234313459e-07, + "loss": 0.5082, + "num_input_tokens_seen": 3082813440, + "step": 1470 + }, + { + "epoch": 0.0735014917276919, + "grad_norm": 16.448945999145508, + "learning_rate": 7.323182249406936e-07, + "loss": 0.4378, + "num_input_tokens_seen": 3084910592, + "step": 1471 + }, + { + "epoch": 0.07377271494439924, + "grad_norm": 19.21479606628418, + "learning_rate": 7.319490797272714e-07, + "loss": 0.6422, + "num_input_tokens_seen": 3087007744, + "step": 1472 + }, + { + "epoch": 0.07404393816110659, + "grad_norm": 12.941019058227539, + "learning_rate": 7.315797880881535e-07, + "loss": 0.4457, + "num_input_tokens_seen": 3089104896, + "step": 1473 + }, + { + "epoch": 0.07431516137781394, + "grad_norm": 11.186482429504395, + "learning_rate": 7.312103503205318e-07, + "loss": 0.4441, + "num_input_tokens_seen": 3091202048, + "step": 1474 + }, + { + "epoch": 0.07458638459452128, + "grad_norm": 9.005460739135742, + "learning_rate": 7.308407667217158e-07, + "loss": 0.1681, + "num_input_tokens_seen": 3093299200, + "step": 1475 + }, + { + "epoch": 0.07485760781122865, + "grad_norm": 15.578516960144043, + "learning_rate": 7.304710375891323e-07, + "loss": 0.5316, + "num_input_tokens_seen": 3095396352, + "step": 1476 + }, + { + "epoch": 0.075128831027936, + "grad_norm": 22.51955795288086, + "learning_rate": 7.30101163220325e-07, + "loss": 0.798, + "num_input_tokens_seen": 3097493504, + "step": 1477 + }, + { + "epoch": 0.07540005424464334, + "grad_norm": 14.802974700927734, + "learning_rate": 7.29731143912955e-07, + "loss": 0.6361, + "num_input_tokens_seen": 3099590656, + "step": 1478 + }, + { + "epoch": 0.07567127746135069, + "grad_norm": 13.981009483337402, + "learning_rate": 7.293609799647996e-07, + "loss": 0.507, + "num_input_tokens_seen": 3101687808, + "step": 1479 + }, + { + "epoch": 0.07594250067805804, + "grad_norm": 17.046981811523438, + "learning_rate": 7.289906716737528e-07, + "loss": 0.5821, + "num_input_tokens_seen": 3103784960, + "step": 1480 + }, + { + "epoch": 0.0762137238947654, + "grad_norm": 11.836614608764648, + "learning_rate": 7.286202193378244e-07, + "loss": 0.3149, + "num_input_tokens_seen": 3105882112, + "step": 1481 + }, + { + "epoch": 0.07648494711147275, + "grad_norm": 14.480379104614258, + "learning_rate": 7.282496232551406e-07, + "loss": 0.5075, + "num_input_tokens_seen": 3107979264, + "step": 1482 + }, + { + "epoch": 0.07675617032818009, + "grad_norm": 17.524093627929688, + "learning_rate": 7.278788837239429e-07, + "loss": 0.5825, + "num_input_tokens_seen": 3110076416, + "step": 1483 + }, + { + "epoch": 0.07702739354488744, + "grad_norm": 10.272031784057617, + "learning_rate": 7.275080010425883e-07, + "loss": 0.3495, + "num_input_tokens_seen": 3112173568, + "step": 1484 + }, + { + "epoch": 0.07729861676159479, + "grad_norm": 14.482696533203125, + "learning_rate": 7.271369755095494e-07, + "loss": 0.5638, + "num_input_tokens_seen": 3114270720, + "step": 1485 + }, + { + "epoch": 0.07756983997830215, + "grad_norm": 17.09143829345703, + "learning_rate": 7.267658074234128e-07, + "loss": 0.7527, + "num_input_tokens_seen": 3116367872, + "step": 1486 + }, + { + "epoch": 0.0778410631950095, + "grad_norm": 15.939448356628418, + "learning_rate": 7.26394497082881e-07, + "loss": 0.5614, + "num_input_tokens_seen": 3118465024, + "step": 1487 + }, + { + "epoch": 0.07811228641171684, + "grad_norm": 13.810276985168457, + "learning_rate": 7.260230447867703e-07, + "loss": 0.4629, + "num_input_tokens_seen": 3120562176, + "step": 1488 + }, + { + "epoch": 0.07838350962842419, + "grad_norm": 14.860830307006836, + "learning_rate": 7.256514508340114e-07, + "loss": 0.5521, + "num_input_tokens_seen": 3122659328, + "step": 1489 + }, + { + "epoch": 0.07865473284513154, + "grad_norm": 17.182228088378906, + "learning_rate": 7.252797155236488e-07, + "loss": 0.624, + "num_input_tokens_seen": 3124756480, + "step": 1490 + }, + { + "epoch": 0.07892595606183889, + "grad_norm": 15.805337905883789, + "learning_rate": 7.249078391548409e-07, + "loss": 0.596, + "num_input_tokens_seen": 3126853632, + "step": 1491 + }, + { + "epoch": 0.07919717927854625, + "grad_norm": 9.75827407836914, + "learning_rate": 7.245358220268599e-07, + "loss": 0.2837, + "num_input_tokens_seen": 3128950784, + "step": 1492 + }, + { + "epoch": 0.0794684024952536, + "grad_norm": 22.41292953491211, + "learning_rate": 7.24163664439091e-07, + "loss": 0.6901, + "num_input_tokens_seen": 3131047936, + "step": 1493 + }, + { + "epoch": 0.07973962571196094, + "grad_norm": 9.785002708435059, + "learning_rate": 7.237913666910322e-07, + "loss": 0.3265, + "num_input_tokens_seen": 3133145088, + "step": 1494 + }, + { + "epoch": 0.08001084892866829, + "grad_norm": 14.571457862854004, + "learning_rate": 7.234189290822947e-07, + "loss": 0.6036, + "num_input_tokens_seen": 3135242240, + "step": 1495 + }, + { + "epoch": 0.08028207214537564, + "grad_norm": 13.217930793762207, + "learning_rate": 7.230463519126024e-07, + "loss": 0.4217, + "num_input_tokens_seen": 3137339392, + "step": 1496 + }, + { + "epoch": 0.080553295362083, + "grad_norm": 12.922200202941895, + "learning_rate": 7.226736354817908e-07, + "loss": 0.4309, + "num_input_tokens_seen": 3139436544, + "step": 1497 + }, + { + "epoch": 0.08082451857879035, + "grad_norm": 26.469648361206055, + "learning_rate": 7.223007800898082e-07, + "loss": 1.4187, + "num_input_tokens_seen": 3141533696, + "step": 1498 + }, + { + "epoch": 0.0810957417954977, + "grad_norm": 16.948135375976562, + "learning_rate": 7.219277860367143e-07, + "loss": 0.6953, + "num_input_tokens_seen": 3143630848, + "step": 1499 + }, + { + "epoch": 0.08136696501220504, + "grad_norm": 28.913721084594727, + "learning_rate": 7.215546536226805e-07, + "loss": 0.7974, + "num_input_tokens_seen": 3145728000, + "step": 1500 + }, + { + "epoch": 0.08163818822891239, + "grad_norm": 12.902350425720215, + "learning_rate": 7.211813831479896e-07, + "loss": 0.4603, + "num_input_tokens_seen": 3147825152, + "step": 1501 + }, + { + "epoch": 0.08190941144561975, + "grad_norm": 13.105647087097168, + "learning_rate": 7.208079749130356e-07, + "loss": 0.4464, + "num_input_tokens_seen": 3149922304, + "step": 1502 + }, + { + "epoch": 0.0821806346623271, + "grad_norm": 22.866016387939453, + "learning_rate": 7.204344292183228e-07, + "loss": 0.5586, + "num_input_tokens_seen": 3152019456, + "step": 1503 + }, + { + "epoch": 0.08245185787903445, + "grad_norm": 12.577240943908691, + "learning_rate": 7.200607463644673e-07, + "loss": 0.3706, + "num_input_tokens_seen": 3154116608, + "step": 1504 + }, + { + "epoch": 0.0827230810957418, + "grad_norm": 16.995012283325195, + "learning_rate": 7.196869266521941e-07, + "loss": 0.25, + "num_input_tokens_seen": 3156213760, + "step": 1505 + }, + { + "epoch": 0.08299430431244914, + "grad_norm": 13.941088676452637, + "learning_rate": 7.193129703823395e-07, + "loss": 0.5905, + "num_input_tokens_seen": 3158310912, + "step": 1506 + }, + { + "epoch": 0.0832655275291565, + "grad_norm": 17.264551162719727, + "learning_rate": 7.189388778558491e-07, + "loss": 0.5209, + "num_input_tokens_seen": 3160408064, + "step": 1507 + }, + { + "epoch": 0.08353675074586385, + "grad_norm": 9.845773696899414, + "learning_rate": 7.185646493737785e-07, + "loss": 0.306, + "num_input_tokens_seen": 3162505216, + "step": 1508 + }, + { + "epoch": 0.0838079739625712, + "grad_norm": 12.871875762939453, + "learning_rate": 7.181902852372924e-07, + "loss": 0.3667, + "num_input_tokens_seen": 3164602368, + "step": 1509 + }, + { + "epoch": 0.08407919717927854, + "grad_norm": 14.597813606262207, + "learning_rate": 7.17815785747665e-07, + "loss": 0.5319, + "num_input_tokens_seen": 3166699520, + "step": 1510 + }, + { + "epoch": 0.08435042039598589, + "grad_norm": 15.713737487792969, + "learning_rate": 7.174411512062789e-07, + "loss": 0.4813, + "num_input_tokens_seen": 3168796672, + "step": 1511 + }, + { + "epoch": 0.08462164361269324, + "grad_norm": 15.4694242477417, + "learning_rate": 7.170663819146259e-07, + "loss": 0.5327, + "num_input_tokens_seen": 3170893824, + "step": 1512 + }, + { + "epoch": 0.0848928668294006, + "grad_norm": 16.36274528503418, + "learning_rate": 7.166914781743062e-07, + "loss": 0.7225, + "num_input_tokens_seen": 3172990976, + "step": 1513 + }, + { + "epoch": 0.08516409004610795, + "grad_norm": 10.62022876739502, + "learning_rate": 7.16316440287028e-07, + "loss": 0.3582, + "num_input_tokens_seen": 3175088128, + "step": 1514 + }, + { + "epoch": 0.0854353132628153, + "grad_norm": 19.60194206237793, + "learning_rate": 7.159412685546073e-07, + "loss": 0.5609, + "num_input_tokens_seen": 3177185280, + "step": 1515 + }, + { + "epoch": 0.08570653647952264, + "grad_norm": 16.77193260192871, + "learning_rate": 7.155659632789683e-07, + "loss": 0.7256, + "num_input_tokens_seen": 3179282432, + "step": 1516 + }, + { + "epoch": 0.08597775969622999, + "grad_norm": 15.905506134033203, + "learning_rate": 7.151905247621422e-07, + "loss": 0.6207, + "num_input_tokens_seen": 3181379584, + "step": 1517 + }, + { + "epoch": 0.08624898291293735, + "grad_norm": 13.54134750366211, + "learning_rate": 7.148149533062678e-07, + "loss": 0.4583, + "num_input_tokens_seen": 3183476736, + "step": 1518 + }, + { + "epoch": 0.0865202061296447, + "grad_norm": 16.33243751525879, + "learning_rate": 7.144392492135908e-07, + "loss": 0.4972, + "num_input_tokens_seen": 3185573888, + "step": 1519 + }, + { + "epoch": 0.08679142934635205, + "grad_norm": 17.10910987854004, + "learning_rate": 7.140634127864632e-07, + "loss": 0.6943, + "num_input_tokens_seen": 3187671040, + "step": 1520 + }, + { + "epoch": 0.0870626525630594, + "grad_norm": 19.709762573242188, + "learning_rate": 7.136874443273442e-07, + "loss": 0.4614, + "num_input_tokens_seen": 3189768192, + "step": 1521 + }, + { + "epoch": 0.08733387577976674, + "grad_norm": 18.73311424255371, + "learning_rate": 7.133113441387988e-07, + "loss": 0.7051, + "num_input_tokens_seen": 3191865344, + "step": 1522 + }, + { + "epoch": 0.0876050989964741, + "grad_norm": 9.619464874267578, + "learning_rate": 7.129351125234979e-07, + "loss": 0.2741, + "num_input_tokens_seen": 3193962496, + "step": 1523 + }, + { + "epoch": 0.08787632221318145, + "grad_norm": 16.603302001953125, + "learning_rate": 7.125587497842189e-07, + "loss": 0.666, + "num_input_tokens_seen": 3196059648, + "step": 1524 + }, + { + "epoch": 0.0881475454298888, + "grad_norm": 10.391305923461914, + "learning_rate": 7.121822562238436e-07, + "loss": 0.282, + "num_input_tokens_seen": 3198156800, + "step": 1525 + }, + { + "epoch": 0.08841876864659615, + "grad_norm": 15.561429977416992, + "learning_rate": 7.118056321453601e-07, + "loss": 0.5204, + "num_input_tokens_seen": 3200253952, + "step": 1526 + }, + { + "epoch": 0.0886899918633035, + "grad_norm": 15.614664077758789, + "learning_rate": 7.11428877851861e-07, + "loss": 0.5045, + "num_input_tokens_seen": 3202351104, + "step": 1527 + }, + { + "epoch": 0.08896121508001086, + "grad_norm": 13.624761581420898, + "learning_rate": 7.110519936465438e-07, + "loss": 0.4001, + "num_input_tokens_seen": 3204448256, + "step": 1528 + }, + { + "epoch": 0.0892324382967182, + "grad_norm": 15.247365951538086, + "learning_rate": 7.106749798327106e-07, + "loss": 0.3207, + "num_input_tokens_seen": 3206545408, + "step": 1529 + }, + { + "epoch": 0.08950366151342555, + "grad_norm": 12.790914535522461, + "learning_rate": 7.102978367137679e-07, + "loss": 0.403, + "num_input_tokens_seen": 3208642560, + "step": 1530 + }, + { + "epoch": 0.0897748847301329, + "grad_norm": 16.698665618896484, + "learning_rate": 7.099205645932258e-07, + "loss": 0.4211, + "num_input_tokens_seen": 3210739712, + "step": 1531 + }, + { + "epoch": 0.09004610794684025, + "grad_norm": 16.94194221496582, + "learning_rate": 7.095431637746988e-07, + "loss": 0.6433, + "num_input_tokens_seen": 3212836864, + "step": 1532 + }, + { + "epoch": 0.09031733116354759, + "grad_norm": 12.014008522033691, + "learning_rate": 7.091656345619047e-07, + "loss": 0.2988, + "num_input_tokens_seen": 3214934016, + "step": 1533 + }, + { + "epoch": 0.09058855438025495, + "grad_norm": 14.939805030822754, + "learning_rate": 7.087879772586647e-07, + "loss": 0.523, + "num_input_tokens_seen": 3217031168, + "step": 1534 + }, + { + "epoch": 0.0908597775969623, + "grad_norm": 21.286935806274414, + "learning_rate": 7.084101921689029e-07, + "loss": 0.5719, + "num_input_tokens_seen": 3219128320, + "step": 1535 + }, + { + "epoch": 0.09113100081366965, + "grad_norm": 15.480658531188965, + "learning_rate": 7.080322795966462e-07, + "loss": 0.536, + "num_input_tokens_seen": 3221225472, + "step": 1536 + }, + { + "epoch": 0.091402224030377, + "grad_norm": 14.61371898651123, + "learning_rate": 7.076542398460247e-07, + "loss": 0.48, + "num_input_tokens_seen": 3223322624, + "step": 1537 + }, + { + "epoch": 0.09167344724708434, + "grad_norm": 15.25113296508789, + "learning_rate": 7.0727607322127e-07, + "loss": 0.4631, + "num_input_tokens_seen": 3225419776, + "step": 1538 + }, + { + "epoch": 0.0919446704637917, + "grad_norm": 16.593374252319336, + "learning_rate": 7.068977800267164e-07, + "loss": 0.4729, + "num_input_tokens_seen": 3227516928, + "step": 1539 + }, + { + "epoch": 0.09221589368049905, + "grad_norm": 11.405816078186035, + "learning_rate": 7.065193605667999e-07, + "loss": 0.3467, + "num_input_tokens_seen": 3229614080, + "step": 1540 + }, + { + "epoch": 0.0924871168972064, + "grad_norm": 15.258930206298828, + "learning_rate": 7.06140815146058e-07, + "loss": 0.3913, + "num_input_tokens_seen": 3231711232, + "step": 1541 + }, + { + "epoch": 0.09275834011391375, + "grad_norm": 19.230300903320312, + "learning_rate": 7.057621440691296e-07, + "loss": 0.6633, + "num_input_tokens_seen": 3233808384, + "step": 1542 + }, + { + "epoch": 0.0930295633306211, + "grad_norm": 17.410680770874023, + "learning_rate": 7.053833476407549e-07, + "loss": 0.6296, + "num_input_tokens_seen": 3235905536, + "step": 1543 + }, + { + "epoch": 0.09330078654732846, + "grad_norm": 12.020039558410645, + "learning_rate": 7.050044261657748e-07, + "loss": 0.319, + "num_input_tokens_seen": 3238002688, + "step": 1544 + }, + { + "epoch": 0.0935720097640358, + "grad_norm": 10.420241355895996, + "learning_rate": 7.04625379949131e-07, + "loss": 0.2475, + "num_input_tokens_seen": 3240099840, + "step": 1545 + }, + { + "epoch": 0.09384323298074315, + "grad_norm": 14.750914573669434, + "learning_rate": 7.042462092958651e-07, + "loss": 0.4473, + "num_input_tokens_seen": 3242196992, + "step": 1546 + }, + { + "epoch": 0.0941144561974505, + "grad_norm": 12.780601501464844, + "learning_rate": 7.038669145111195e-07, + "loss": 0.409, + "num_input_tokens_seen": 3244294144, + "step": 1547 + }, + { + "epoch": 0.09438567941415785, + "grad_norm": 16.467487335205078, + "learning_rate": 7.034874959001363e-07, + "loss": 0.4919, + "num_input_tokens_seen": 3246391296, + "step": 1548 + }, + { + "epoch": 0.09465690263086521, + "grad_norm": 13.055885314941406, + "learning_rate": 7.031079537682569e-07, + "loss": 0.3439, + "num_input_tokens_seen": 3248488448, + "step": 1549 + }, + { + "epoch": 0.09492812584757256, + "grad_norm": 17.769779205322266, + "learning_rate": 7.027282884209227e-07, + "loss": 0.6011, + "num_input_tokens_seen": 3250585600, + "step": 1550 + }, + { + "epoch": 0.0951993490642799, + "grad_norm": 15.251014709472656, + "learning_rate": 7.023485001636737e-07, + "loss": 0.3822, + "num_input_tokens_seen": 3252682752, + "step": 1551 + }, + { + "epoch": 0.09547057228098725, + "grad_norm": 14.227032661437988, + "learning_rate": 7.019685893021488e-07, + "loss": 0.5338, + "num_input_tokens_seen": 3254779904, + "step": 1552 + }, + { + "epoch": 0.0957417954976946, + "grad_norm": 10.746809005737305, + "learning_rate": 7.015885561420863e-07, + "loss": 0.3099, + "num_input_tokens_seen": 3256877056, + "step": 1553 + }, + { + "epoch": 0.09601301871440195, + "grad_norm": 10.69054889678955, + "learning_rate": 7.012084009893221e-07, + "loss": 0.273, + "num_input_tokens_seen": 3258974208, + "step": 1554 + }, + { + "epoch": 0.09628424193110931, + "grad_norm": 15.015727996826172, + "learning_rate": 7.008281241497908e-07, + "loss": 0.6222, + "num_input_tokens_seen": 3261071360, + "step": 1555 + }, + { + "epoch": 0.09655546514781665, + "grad_norm": 15.404815673828125, + "learning_rate": 7.004477259295244e-07, + "loss": 0.4674, + "num_input_tokens_seen": 3263168512, + "step": 1556 + }, + { + "epoch": 0.096826688364524, + "grad_norm": 7.422001838684082, + "learning_rate": 7.000672066346532e-07, + "loss": 0.1853, + "num_input_tokens_seen": 3265265664, + "step": 1557 + }, + { + "epoch": 0.09709791158123135, + "grad_norm": 11.59667682647705, + "learning_rate": 6.996865665714046e-07, + "loss": 0.3281, + "num_input_tokens_seen": 3267362816, + "step": 1558 + }, + { + "epoch": 0.0973691347979387, + "grad_norm": 17.11145782470703, + "learning_rate": 6.993058060461035e-07, + "loss": 0.6195, + "num_input_tokens_seen": 3269459968, + "step": 1559 + }, + { + "epoch": 0.09764035801464606, + "grad_norm": 18.34768295288086, + "learning_rate": 6.989249253651708e-07, + "loss": 0.5665, + "num_input_tokens_seen": 3271557120, + "step": 1560 + }, + { + "epoch": 0.0979115812313534, + "grad_norm": 11.051552772521973, + "learning_rate": 6.985439248351254e-07, + "loss": 0.2874, + "num_input_tokens_seen": 3273654272, + "step": 1561 + }, + { + "epoch": 0.09818280444806075, + "grad_norm": 23.903244018554688, + "learning_rate": 6.981628047625818e-07, + "loss": 0.925, + "num_input_tokens_seen": 3275751424, + "step": 1562 + }, + { + "epoch": 0.0984540276647681, + "grad_norm": 16.90290641784668, + "learning_rate": 6.977815654542508e-07, + "loss": 0.6217, + "num_input_tokens_seen": 3277848576, + "step": 1563 + }, + { + "epoch": 0.09872525088147545, + "grad_norm": 9.315507888793945, + "learning_rate": 6.974002072169395e-07, + "loss": 0.1993, + "num_input_tokens_seen": 3279945728, + "step": 1564 + }, + { + "epoch": 0.09899647409818281, + "grad_norm": 15.298419952392578, + "learning_rate": 6.970187303575505e-07, + "loss": 0.336, + "num_input_tokens_seen": 3282042880, + "step": 1565 + }, + { + "epoch": 0.09926769731489016, + "grad_norm": 20.24376106262207, + "learning_rate": 6.966371351830818e-07, + "loss": 0.8491, + "num_input_tokens_seen": 3284140032, + "step": 1566 + }, + { + "epoch": 0.0995389205315975, + "grad_norm": 11.215019226074219, + "learning_rate": 6.962554220006265e-07, + "loss": 0.3051, + "num_input_tokens_seen": 3286237184, + "step": 1567 + }, + { + "epoch": 0.09981014374830485, + "grad_norm": 16.658477783203125, + "learning_rate": 6.958735911173729e-07, + "loss": 0.5136, + "num_input_tokens_seen": 3288334336, + "step": 1568 + }, + { + "epoch": 0.1000813669650122, + "grad_norm": 19.485260009765625, + "learning_rate": 6.954916428406045e-07, + "loss": 0.6946, + "num_input_tokens_seen": 3290431488, + "step": 1569 + }, + { + "epoch": 0.10035259018171956, + "grad_norm": 15.814196586608887, + "learning_rate": 6.95109577477698e-07, + "loss": 0.5309, + "num_input_tokens_seen": 3292528640, + "step": 1570 + }, + { + "epoch": 0.10062381339842691, + "grad_norm": 14.682167053222656, + "learning_rate": 6.947273953361255e-07, + "loss": 0.5357, + "num_input_tokens_seen": 3294625792, + "step": 1571 + }, + { + "epoch": 0.10089503661513426, + "grad_norm": 16.301332473754883, + "learning_rate": 6.943450967234524e-07, + "loss": 0.388, + "num_input_tokens_seen": 3296722944, + "step": 1572 + }, + { + "epoch": 0.1011662598318416, + "grad_norm": 11.630681991577148, + "learning_rate": 6.939626819473384e-07, + "loss": 0.3335, + "num_input_tokens_seen": 3298820096, + "step": 1573 + }, + { + "epoch": 0.10143748304854895, + "grad_norm": 21.778146743774414, + "learning_rate": 6.93580151315536e-07, + "loss": 0.791, + "num_input_tokens_seen": 3300917248, + "step": 1574 + }, + { + "epoch": 0.10170870626525631, + "grad_norm": 16.302976608276367, + "learning_rate": 6.931975051358914e-07, + "loss": 0.3302, + "num_input_tokens_seen": 3303014400, + "step": 1575 + }, + { + "epoch": 0.10197992948196366, + "grad_norm": 14.719386100769043, + "learning_rate": 6.928147437163439e-07, + "loss": 0.5707, + "num_input_tokens_seen": 3305111552, + "step": 1576 + }, + { + "epoch": 0.10225115269867101, + "grad_norm": 10.73705768585205, + "learning_rate": 6.92431867364925e-07, + "loss": 0.3245, + "num_input_tokens_seen": 3307208704, + "step": 1577 + }, + { + "epoch": 0.10252237591537836, + "grad_norm": 21.941381454467773, + "learning_rate": 6.920488763897593e-07, + "loss": 0.8567, + "num_input_tokens_seen": 3309305856, + "step": 1578 + }, + { + "epoch": 0.1027935991320857, + "grad_norm": 20.246137619018555, + "learning_rate": 6.916657710990632e-07, + "loss": 0.5603, + "num_input_tokens_seen": 3311403008, + "step": 1579 + }, + { + "epoch": 0.10306482234879305, + "grad_norm": 15.607077598571777, + "learning_rate": 6.912825518011452e-07, + "loss": 0.6866, + "num_input_tokens_seen": 3313500160, + "step": 1580 + }, + { + "epoch": 0.10333604556550041, + "grad_norm": 13.166605949401855, + "learning_rate": 6.90899218804406e-07, + "loss": 0.4176, + "num_input_tokens_seen": 3315597312, + "step": 1581 + }, + { + "epoch": 0.10360726878220776, + "grad_norm": 14.276864051818848, + "learning_rate": 6.905157724173369e-07, + "loss": 0.3567, + "num_input_tokens_seen": 3317694464, + "step": 1582 + }, + { + "epoch": 0.1038784919989151, + "grad_norm": 15.237847328186035, + "learning_rate": 6.901322129485212e-07, + "loss": 0.5268, + "num_input_tokens_seen": 3319791616, + "step": 1583 + }, + { + "epoch": 0.10414971521562245, + "grad_norm": 10.99821949005127, + "learning_rate": 6.897485407066329e-07, + "loss": 0.3016, + "num_input_tokens_seen": 3321888768, + "step": 1584 + }, + { + "epoch": 0.1044209384323298, + "grad_norm": 13.16862964630127, + "learning_rate": 6.893647560004369e-07, + "loss": 0.5617, + "num_input_tokens_seen": 3323985920, + "step": 1585 + }, + { + "epoch": 0.10469216164903716, + "grad_norm": 15.644988059997559, + "learning_rate": 6.889808591387885e-07, + "loss": 0.4087, + "num_input_tokens_seen": 3326083072, + "step": 1586 + }, + { + "epoch": 0.10496338486574451, + "grad_norm": 14.807315826416016, + "learning_rate": 6.885968504306334e-07, + "loss": 0.4284, + "num_input_tokens_seen": 3328180224, + "step": 1587 + }, + { + "epoch": 0.10523460808245186, + "grad_norm": 18.928443908691406, + "learning_rate": 6.882127301850069e-07, + "loss": 0.7086, + "num_input_tokens_seen": 3330277376, + "step": 1588 + }, + { + "epoch": 0.1055058312991592, + "grad_norm": 17.395771026611328, + "learning_rate": 6.878284987110345e-07, + "loss": 0.5846, + "num_input_tokens_seen": 3332374528, + "step": 1589 + }, + { + "epoch": 0.10577705451586655, + "grad_norm": 10.824260711669922, + "learning_rate": 6.874441563179313e-07, + "loss": 0.3061, + "num_input_tokens_seen": 3334471680, + "step": 1590 + }, + { + "epoch": 0.10604827773257391, + "grad_norm": 18.152671813964844, + "learning_rate": 6.870597033150012e-07, + "loss": 0.664, + "num_input_tokens_seen": 3336568832, + "step": 1591 + }, + { + "epoch": 0.10631950094928126, + "grad_norm": 13.247352600097656, + "learning_rate": 6.866751400116374e-07, + "loss": 0.4636, + "num_input_tokens_seen": 3338665984, + "step": 1592 + }, + { + "epoch": 0.10659072416598861, + "grad_norm": 26.162593841552734, + "learning_rate": 6.862904667173216e-07, + "loss": 0.7328, + "num_input_tokens_seen": 3340763136, + "step": 1593 + }, + { + "epoch": 0.10686194738269596, + "grad_norm": 22.150165557861328, + "learning_rate": 6.859056837416245e-07, + "loss": 0.9154, + "num_input_tokens_seen": 3342860288, + "step": 1594 + }, + { + "epoch": 0.1071331705994033, + "grad_norm": 10.474020957946777, + "learning_rate": 6.855207913942048e-07, + "loss": 0.388, + "num_input_tokens_seen": 3344957440, + "step": 1595 + }, + { + "epoch": 0.10740439381611067, + "grad_norm": 13.631202697753906, + "learning_rate": 6.85135789984809e-07, + "loss": 0.4211, + "num_input_tokens_seen": 3347054592, + "step": 1596 + }, + { + "epoch": 0.10767561703281801, + "grad_norm": 14.622827529907227, + "learning_rate": 6.847506798232719e-07, + "loss": 0.5837, + "num_input_tokens_seen": 3349151744, + "step": 1597 + }, + { + "epoch": 0.10794684024952536, + "grad_norm": 16.3840389251709, + "learning_rate": 6.843654612195152e-07, + "loss": 0.4376, + "num_input_tokens_seen": 3351248896, + "step": 1598 + }, + { + "epoch": 0.10821806346623271, + "grad_norm": 16.240615844726562, + "learning_rate": 6.839801344835484e-07, + "loss": 0.6028, + "num_input_tokens_seen": 3353346048, + "step": 1599 + }, + { + "epoch": 0.10848928668294006, + "grad_norm": 13.917694091796875, + "learning_rate": 6.835946999254677e-07, + "loss": 0.6392, + "num_input_tokens_seen": 3355443200, + "step": 1600 + }, + { + "epoch": 0.1087605098996474, + "grad_norm": 11.366427421569824, + "learning_rate": 6.83209157855456e-07, + "loss": 0.3998, + "num_input_tokens_seen": 3357540352, + "step": 1601 + }, + { + "epoch": 0.10903173311635476, + "grad_norm": 14.33171272277832, + "learning_rate": 6.828235085837831e-07, + "loss": 0.4276, + "num_input_tokens_seen": 3359637504, + "step": 1602 + }, + { + "epoch": 0.10930295633306211, + "grad_norm": 12.968931198120117, + "learning_rate": 6.824377524208047e-07, + "loss": 0.489, + "num_input_tokens_seen": 3361734656, + "step": 1603 + }, + { + "epoch": 0.10957417954976946, + "grad_norm": 15.882153511047363, + "learning_rate": 6.820518896769629e-07, + "loss": 0.5021, + "num_input_tokens_seen": 3363831808, + "step": 1604 + }, + { + "epoch": 0.10984540276647681, + "grad_norm": 10.402983665466309, + "learning_rate": 6.816659206627853e-07, + "loss": 0.2281, + "num_input_tokens_seen": 3365928960, + "step": 1605 + }, + { + "epoch": 0.11011662598318415, + "grad_norm": 14.085593223571777, + "learning_rate": 6.812798456888849e-07, + "loss": 0.4662, + "num_input_tokens_seen": 3368026112, + "step": 1606 + }, + { + "epoch": 0.11038784919989152, + "grad_norm": 14.953266143798828, + "learning_rate": 6.808936650659605e-07, + "loss": 0.6349, + "num_input_tokens_seen": 3370123264, + "step": 1607 + }, + { + "epoch": 0.11065907241659886, + "grad_norm": 15.086980819702148, + "learning_rate": 6.805073791047951e-07, + "loss": 0.6972, + "num_input_tokens_seen": 3372220416, + "step": 1608 + }, + { + "epoch": 0.11093029563330621, + "grad_norm": 14.327410697937012, + "learning_rate": 6.80120988116257e-07, + "loss": 0.3387, + "num_input_tokens_seen": 3374317568, + "step": 1609 + }, + { + "epoch": 0.11120151885001356, + "grad_norm": 13.357967376708984, + "learning_rate": 6.797344924112995e-07, + "loss": 0.3773, + "num_input_tokens_seen": 3376414720, + "step": 1610 + }, + { + "epoch": 0.1114727420667209, + "grad_norm": 16.17928695678711, + "learning_rate": 6.793478923009592e-07, + "loss": 0.6131, + "num_input_tokens_seen": 3378511872, + "step": 1611 + }, + { + "epoch": 0.11174396528342827, + "grad_norm": 9.867973327636719, + "learning_rate": 6.789611880963569e-07, + "loss": 0.2377, + "num_input_tokens_seen": 3380609024, + "step": 1612 + }, + { + "epoch": 0.11201518850013562, + "grad_norm": 13.116447448730469, + "learning_rate": 6.78574380108698e-07, + "loss": 0.4105, + "num_input_tokens_seen": 3382706176, + "step": 1613 + }, + { + "epoch": 0.11228641171684296, + "grad_norm": 13.960227012634277, + "learning_rate": 6.781874686492706e-07, + "loss": 0.475, + "num_input_tokens_seen": 3384803328, + "step": 1614 + }, + { + "epoch": 0.11255763493355031, + "grad_norm": 11.566934585571289, + "learning_rate": 6.778004540294464e-07, + "loss": 0.3708, + "num_input_tokens_seen": 3386900480, + "step": 1615 + }, + { + "epoch": 0.11282885815025766, + "grad_norm": 21.169614791870117, + "learning_rate": 6.774133365606801e-07, + "loss": 0.8485, + "num_input_tokens_seen": 3388997632, + "step": 1616 + }, + { + "epoch": 0.11310008136696502, + "grad_norm": 20.027353286743164, + "learning_rate": 6.770261165545085e-07, + "loss": 0.8033, + "num_input_tokens_seen": 3391094784, + "step": 1617 + }, + { + "epoch": 0.11337130458367237, + "grad_norm": 19.385282516479492, + "learning_rate": 6.766387943225524e-07, + "loss": 0.7747, + "num_input_tokens_seen": 3393191936, + "step": 1618 + }, + { + "epoch": 0.11364252780037971, + "grad_norm": 11.849488258361816, + "learning_rate": 6.762513701765135e-07, + "loss": 0.4182, + "num_input_tokens_seen": 3395289088, + "step": 1619 + }, + { + "epoch": 0.11391375101708706, + "grad_norm": 11.74028205871582, + "learning_rate": 6.75863844428176e-07, + "loss": 0.3773, + "num_input_tokens_seen": 3397386240, + "step": 1620 + }, + { + "epoch": 0.11418497423379441, + "grad_norm": 16.50984001159668, + "learning_rate": 6.754762173894061e-07, + "loss": 0.5555, + "num_input_tokens_seen": 3399483392, + "step": 1621 + }, + { + "epoch": 0.11445619745050176, + "grad_norm": 10.819229125976562, + "learning_rate": 6.750884893721511e-07, + "loss": 0.2958, + "num_input_tokens_seen": 3401580544, + "step": 1622 + }, + { + "epoch": 0.11472742066720912, + "grad_norm": 14.737079620361328, + "learning_rate": 6.747006606884398e-07, + "loss": 0.5203, + "num_input_tokens_seen": 3403677696, + "step": 1623 + }, + { + "epoch": 0.11499864388391647, + "grad_norm": 19.5377197265625, + "learning_rate": 6.74312731650382e-07, + "loss": 0.5925, + "num_input_tokens_seen": 3405774848, + "step": 1624 + }, + { + "epoch": 0.11526986710062381, + "grad_norm": 17.311817169189453, + "learning_rate": 6.739247025701683e-07, + "loss": 0.5197, + "num_input_tokens_seen": 3407872000, + "step": 1625 + }, + { + "epoch": 0.11554109031733116, + "grad_norm": 33.43289566040039, + "learning_rate": 6.735365737600695e-07, + "loss": 0.8992, + "num_input_tokens_seen": 3409969152, + "step": 1626 + }, + { + "epoch": 0.11581231353403851, + "grad_norm": 12.524608612060547, + "learning_rate": 6.731483455324374e-07, + "loss": 0.4096, + "num_input_tokens_seen": 3412066304, + "step": 1627 + }, + { + "epoch": 0.11608353675074587, + "grad_norm": 9.463518142700195, + "learning_rate": 6.727600181997026e-07, + "loss": 0.2309, + "num_input_tokens_seen": 3414163456, + "step": 1628 + }, + { + "epoch": 0.11635475996745322, + "grad_norm": 16.343246459960938, + "learning_rate": 6.723715920743767e-07, + "loss": 0.5599, + "num_input_tokens_seen": 3416260608, + "step": 1629 + }, + { + "epoch": 0.11662598318416056, + "grad_norm": 12.191635131835938, + "learning_rate": 6.7198306746905e-07, + "loss": 0.3698, + "num_input_tokens_seen": 3418357760, + "step": 1630 + }, + { + "epoch": 0.11689720640086791, + "grad_norm": 16.117708206176758, + "learning_rate": 6.715944446963924e-07, + "loss": 0.576, + "num_input_tokens_seen": 3420454912, + "step": 1631 + }, + { + "epoch": 0.11716842961757526, + "grad_norm": 19.668415069580078, + "learning_rate": 6.712057240691527e-07, + "loss": 0.9028, + "num_input_tokens_seen": 3422552064, + "step": 1632 + }, + { + "epoch": 0.11743965283428262, + "grad_norm": 11.864884376525879, + "learning_rate": 6.708169059001586e-07, + "loss": 0.4027, + "num_input_tokens_seen": 3424649216, + "step": 1633 + }, + { + "epoch": 0.11771087605098997, + "grad_norm": 15.043469429016113, + "learning_rate": 6.704279905023159e-07, + "loss": 0.579, + "num_input_tokens_seen": 3426746368, + "step": 1634 + }, + { + "epoch": 0.11798209926769732, + "grad_norm": 21.475412368774414, + "learning_rate": 6.700389781886091e-07, + "loss": 0.6607, + "num_input_tokens_seen": 3428843520, + "step": 1635 + }, + { + "epoch": 0.11825332248440466, + "grad_norm": 13.802813529968262, + "learning_rate": 6.696498692721006e-07, + "loss": 0.3898, + "num_input_tokens_seen": 3430940672, + "step": 1636 + }, + { + "epoch": 0.11852454570111201, + "grad_norm": 15.97105598449707, + "learning_rate": 6.692606640659302e-07, + "loss": 0.5449, + "num_input_tokens_seen": 3433037824, + "step": 1637 + }, + { + "epoch": 0.11879576891781937, + "grad_norm": 18.830066680908203, + "learning_rate": 6.688713628833157e-07, + "loss": 0.7314, + "num_input_tokens_seen": 3435134976, + "step": 1638 + }, + { + "epoch": 0.11906699213452672, + "grad_norm": 28.016523361206055, + "learning_rate": 6.684819660375516e-07, + "loss": 0.4921, + "num_input_tokens_seen": 3437232128, + "step": 1639 + }, + { + "epoch": 0.11933821535123407, + "grad_norm": 20.086233139038086, + "learning_rate": 6.6809247384201e-07, + "loss": 0.6695, + "num_input_tokens_seen": 3439329280, + "step": 1640 + }, + { + "epoch": 0.11960943856794141, + "grad_norm": 12.381033897399902, + "learning_rate": 6.67702886610139e-07, + "loss": 0.433, + "num_input_tokens_seen": 3441426432, + "step": 1641 + }, + { + "epoch": 0.11988066178464876, + "grad_norm": 16.39443016052246, + "learning_rate": 6.673132046554639e-07, + "loss": 0.6587, + "num_input_tokens_seen": 3443523584, + "step": 1642 + }, + { + "epoch": 0.12015188500135611, + "grad_norm": 9.596854209899902, + "learning_rate": 6.669234282915857e-07, + "loss": 0.2475, + "num_input_tokens_seen": 3445620736, + "step": 1643 + }, + { + "epoch": 0.12042310821806347, + "grad_norm": 14.690909385681152, + "learning_rate": 6.665335578321819e-07, + "loss": 0.4947, + "num_input_tokens_seen": 3447717888, + "step": 1644 + }, + { + "epoch": 0.12069433143477082, + "grad_norm": 19.749242782592773, + "learning_rate": 6.661435935910048e-07, + "loss": 0.6977, + "num_input_tokens_seen": 3449815040, + "step": 1645 + }, + { + "epoch": 0.12096555465147817, + "grad_norm": 11.498983383178711, + "learning_rate": 6.657535358818833e-07, + "loss": 0.3786, + "num_input_tokens_seen": 3451912192, + "step": 1646 + }, + { + "epoch": 0.12123677786818551, + "grad_norm": 18.334880828857422, + "learning_rate": 6.653633850187211e-07, + "loss": 0.7039, + "num_input_tokens_seen": 3454009344, + "step": 1647 + }, + { + "epoch": 0.12150800108489286, + "grad_norm": 14.584370613098145, + "learning_rate": 6.649731413154964e-07, + "loss": 0.4845, + "num_input_tokens_seen": 3456106496, + "step": 1648 + }, + { + "epoch": 0.12177922430160022, + "grad_norm": 14.279857635498047, + "learning_rate": 6.645828050862626e-07, + "loss": 0.5301, + "num_input_tokens_seen": 3458203648, + "step": 1649 + }, + { + "epoch": 0.12205044751830757, + "grad_norm": 9.879186630249023, + "learning_rate": 6.641923766451475e-07, + "loss": 0.2123, + "num_input_tokens_seen": 3460300800, + "step": 1650 + }, + { + "epoch": 0.12232167073501492, + "grad_norm": 11.519014358520508, + "learning_rate": 6.63801856306353e-07, + "loss": 0.4437, + "num_input_tokens_seen": 3462397952, + "step": 1651 + }, + { + "epoch": 0.12259289395172226, + "grad_norm": 15.039155960083008, + "learning_rate": 6.634112443841551e-07, + "loss": 0.5951, + "num_input_tokens_seen": 3464495104, + "step": 1652 + }, + { + "epoch": 0.12286411716842961, + "grad_norm": 14.848783493041992, + "learning_rate": 6.630205411929032e-07, + "loss": 0.5456, + "num_input_tokens_seen": 3466592256, + "step": 1653 + }, + { + "epoch": 0.12313534038513697, + "grad_norm": 10.93878173828125, + "learning_rate": 6.626297470470205e-07, + "loss": 0.3535, + "num_input_tokens_seen": 3468689408, + "step": 1654 + }, + { + "epoch": 0.12340656360184432, + "grad_norm": 11.482619285583496, + "learning_rate": 6.622388622610034e-07, + "loss": 0.2725, + "num_input_tokens_seen": 3470786560, + "step": 1655 + }, + { + "epoch": 0.12367778681855167, + "grad_norm": 12.639951705932617, + "learning_rate": 6.618478871494209e-07, + "loss": 0.3996, + "num_input_tokens_seen": 3472883712, + "step": 1656 + }, + { + "epoch": 0.12394901003525902, + "grad_norm": 12.591337203979492, + "learning_rate": 6.61456822026915e-07, + "loss": 0.3903, + "num_input_tokens_seen": 3474980864, + "step": 1657 + }, + { + "epoch": 0.12422023325196636, + "grad_norm": 18.513185501098633, + "learning_rate": 6.610656672081999e-07, + "loss": 0.4721, + "num_input_tokens_seen": 3477078016, + "step": 1658 + }, + { + "epoch": 0.12449145646867373, + "grad_norm": 16.166847229003906, + "learning_rate": 6.606744230080622e-07, + "loss": 0.5814, + "num_input_tokens_seen": 3479175168, + "step": 1659 + }, + { + "epoch": 0.12476267968538107, + "grad_norm": 19.309938430786133, + "learning_rate": 6.602830897413603e-07, + "loss": 0.7281, + "num_input_tokens_seen": 3481272320, + "step": 1660 + }, + { + "epoch": 0.1250339029020884, + "grad_norm": 20.90107536315918, + "learning_rate": 6.598916677230243e-07, + "loss": 0.9278, + "num_input_tokens_seen": 3483369472, + "step": 1661 + }, + { + "epoch": 0.12530512611879577, + "grad_norm": 14.196072578430176, + "learning_rate": 6.59500157268056e-07, + "loss": 0.488, + "num_input_tokens_seen": 3485466624, + "step": 1662 + }, + { + "epoch": 0.12557634933550313, + "grad_norm": 14.922718048095703, + "learning_rate": 6.591085586915279e-07, + "loss": 0.4756, + "num_input_tokens_seen": 3487563776, + "step": 1663 + }, + { + "epoch": 0.12584757255221046, + "grad_norm": 19.60527992248535, + "learning_rate": 6.587168723085836e-07, + "loss": 0.7925, + "num_input_tokens_seen": 3489660928, + "step": 1664 + }, + { + "epoch": 0.12611879576891782, + "grad_norm": 13.591885566711426, + "learning_rate": 6.583250984344374e-07, + "loss": 0.4419, + "num_input_tokens_seen": 3491758080, + "step": 1665 + }, + { + "epoch": 0.12639001898562516, + "grad_norm": 13.59085464477539, + "learning_rate": 6.57933237384374e-07, + "loss": 0.4151, + "num_input_tokens_seen": 3493855232, + "step": 1666 + }, + { + "epoch": 0.12666124220233252, + "grad_norm": 18.338226318359375, + "learning_rate": 6.575412894737484e-07, + "loss": 0.6672, + "num_input_tokens_seen": 3495952384, + "step": 1667 + }, + { + "epoch": 0.12693246541903988, + "grad_norm": 19.015380859375, + "learning_rate": 6.571492550179853e-07, + "loss": 0.9037, + "num_input_tokens_seen": 3498049536, + "step": 1668 + }, + { + "epoch": 0.12720368863574721, + "grad_norm": 15.90075969696045, + "learning_rate": 6.567571343325791e-07, + "loss": 0.6357, + "num_input_tokens_seen": 3500146688, + "step": 1669 + }, + { + "epoch": 0.12747491185245458, + "grad_norm": 17.481428146362305, + "learning_rate": 6.563649277330935e-07, + "loss": 0.5654, + "num_input_tokens_seen": 3502243840, + "step": 1670 + }, + { + "epoch": 0.1277461350691619, + "grad_norm": 13.026409149169922, + "learning_rate": 6.559726355351617e-07, + "loss": 0.4402, + "num_input_tokens_seen": 3504340992, + "step": 1671 + }, + { + "epoch": 0.12801735828586927, + "grad_norm": 12.516473770141602, + "learning_rate": 6.555802580544853e-07, + "loss": 0.3762, + "num_input_tokens_seen": 3506438144, + "step": 1672 + }, + { + "epoch": 0.12828858150257663, + "grad_norm": 16.42319107055664, + "learning_rate": 6.551877956068349e-07, + "loss": 0.4072, + "num_input_tokens_seen": 3508535296, + "step": 1673 + }, + { + "epoch": 0.12855980471928397, + "grad_norm": 23.695953369140625, + "learning_rate": 6.547952485080491e-07, + "loss": 0.6364, + "num_input_tokens_seen": 3510632448, + "step": 1674 + }, + { + "epoch": 0.12883102793599133, + "grad_norm": 11.105734825134277, + "learning_rate": 6.544026170740352e-07, + "loss": 0.3398, + "num_input_tokens_seen": 3512729600, + "step": 1675 + }, + { + "epoch": 0.12910225115269866, + "grad_norm": 12.451509475708008, + "learning_rate": 6.540099016207679e-07, + "loss": 0.42, + "num_input_tokens_seen": 3514826752, + "step": 1676 + }, + { + "epoch": 0.12937347436940602, + "grad_norm": 12.851999282836914, + "learning_rate": 6.536171024642896e-07, + "loss": 0.4315, + "num_input_tokens_seen": 3516923904, + "step": 1677 + }, + { + "epoch": 0.12964469758611338, + "grad_norm": 14.279557228088379, + "learning_rate": 6.532242199207103e-07, + "loss": 0.4199, + "num_input_tokens_seen": 3519021056, + "step": 1678 + }, + { + "epoch": 0.12991592080282072, + "grad_norm": 12.546601295471191, + "learning_rate": 6.528312543062066e-07, + "loss": 0.2372, + "num_input_tokens_seen": 3521118208, + "step": 1679 + }, + { + "epoch": 0.13018714401952808, + "grad_norm": 17.787132263183594, + "learning_rate": 6.524382059370226e-07, + "loss": 0.6352, + "num_input_tokens_seen": 3523215360, + "step": 1680 + }, + { + "epoch": 0.1304583672362354, + "grad_norm": 14.964465141296387, + "learning_rate": 6.520450751294685e-07, + "loss": 0.5617, + "num_input_tokens_seen": 3525312512, + "step": 1681 + }, + { + "epoch": 0.13072959045294277, + "grad_norm": 24.104333877563477, + "learning_rate": 6.516518621999209e-07, + "loss": 0.7431, + "num_input_tokens_seen": 3527409664, + "step": 1682 + }, + { + "epoch": 0.13100081366965013, + "grad_norm": 15.481319427490234, + "learning_rate": 6.512585674648227e-07, + "loss": 0.5991, + "num_input_tokens_seen": 3529506816, + "step": 1683 + }, + { + "epoch": 0.13127203688635747, + "grad_norm": 9.043676376342773, + "learning_rate": 6.50865191240683e-07, + "loss": 0.2865, + "num_input_tokens_seen": 3531603968, + "step": 1684 + }, + { + "epoch": 0.13154326010306483, + "grad_norm": 15.140732765197754, + "learning_rate": 6.504717338440751e-07, + "loss": 0.5084, + "num_input_tokens_seen": 3533701120, + "step": 1685 + }, + { + "epoch": 0.13181448331977216, + "grad_norm": 10.798766136169434, + "learning_rate": 6.500781955916393e-07, + "loss": 0.3529, + "num_input_tokens_seen": 3535798272, + "step": 1686 + }, + { + "epoch": 0.13208570653647952, + "grad_norm": 15.705323219299316, + "learning_rate": 6.4968457680008e-07, + "loss": 0.5974, + "num_input_tokens_seen": 3537895424, + "step": 1687 + }, + { + "epoch": 0.13235692975318689, + "grad_norm": 12.985349655151367, + "learning_rate": 6.492908777861664e-07, + "loss": 0.3212, + "num_input_tokens_seen": 3539992576, + "step": 1688 + }, + { + "epoch": 0.13262815296989422, + "grad_norm": 17.304018020629883, + "learning_rate": 6.488970988667327e-07, + "loss": 0.7088, + "num_input_tokens_seen": 3542089728, + "step": 1689 + }, + { + "epoch": 0.13289937618660158, + "grad_norm": 24.675233840942383, + "learning_rate": 6.485032403586772e-07, + "loss": 0.9613, + "num_input_tokens_seen": 3544186880, + "step": 1690 + }, + { + "epoch": 0.13317059940330891, + "grad_norm": 14.140911102294922, + "learning_rate": 6.481093025789621e-07, + "loss": 0.3862, + "num_input_tokens_seen": 3546284032, + "step": 1691 + }, + { + "epoch": 0.13344182262001628, + "grad_norm": 17.960968017578125, + "learning_rate": 6.477152858446136e-07, + "loss": 0.5059, + "num_input_tokens_seen": 3548381184, + "step": 1692 + }, + { + "epoch": 0.13371304583672364, + "grad_norm": 14.560555458068848, + "learning_rate": 6.473211904727216e-07, + "loss": 0.5256, + "num_input_tokens_seen": 3550478336, + "step": 1693 + }, + { + "epoch": 0.13398426905343097, + "grad_norm": 15.947834014892578, + "learning_rate": 6.469270167804386e-07, + "loss": 0.5166, + "num_input_tokens_seen": 3552575488, + "step": 1694 + }, + { + "epoch": 0.13425549227013833, + "grad_norm": 13.651203155517578, + "learning_rate": 6.46532765084981e-07, + "loss": 0.4082, + "num_input_tokens_seen": 3554672640, + "step": 1695 + }, + { + "epoch": 0.13452671548684567, + "grad_norm": 11.05508041381836, + "learning_rate": 6.461384357036274e-07, + "loss": 0.3031, + "num_input_tokens_seen": 3556769792, + "step": 1696 + }, + { + "epoch": 0.13479793870355303, + "grad_norm": 18.280160903930664, + "learning_rate": 6.457440289537191e-07, + "loss": 0.5375, + "num_input_tokens_seen": 3558866944, + "step": 1697 + }, + { + "epoch": 0.13506916192026036, + "grad_norm": 10.277115821838379, + "learning_rate": 6.453495451526595e-07, + "loss": 0.2785, + "num_input_tokens_seen": 3560964096, + "step": 1698 + }, + { + "epoch": 0.13534038513696772, + "grad_norm": 13.42917537689209, + "learning_rate": 6.449549846179144e-07, + "loss": 0.3343, + "num_input_tokens_seen": 3563061248, + "step": 1699 + }, + { + "epoch": 0.13561160835367508, + "grad_norm": 20.96129035949707, + "learning_rate": 6.445603476670109e-07, + "loss": 0.5684, + "num_input_tokens_seen": 3565158400, + "step": 1700 + }, + { + "epoch": 0.13588283157038242, + "grad_norm": 18.72563934326172, + "learning_rate": 6.44165634617538e-07, + "loss": 0.8015, + "num_input_tokens_seen": 3567255552, + "step": 1701 + }, + { + "epoch": 0.13615405478708978, + "grad_norm": 12.687376022338867, + "learning_rate": 6.437708457871455e-07, + "loss": 0.4409, + "num_input_tokens_seen": 3569352704, + "step": 1702 + }, + { + "epoch": 0.1364252780037971, + "grad_norm": 17.298492431640625, + "learning_rate": 6.433759814935447e-07, + "loss": 0.56, + "num_input_tokens_seen": 3571449856, + "step": 1703 + }, + { + "epoch": 0.13669650122050447, + "grad_norm": 10.709501266479492, + "learning_rate": 6.429810420545072e-07, + "loss": 0.3482, + "num_input_tokens_seen": 3573547008, + "step": 1704 + }, + { + "epoch": 0.13696772443721184, + "grad_norm": 15.926192283630371, + "learning_rate": 6.425860277878651e-07, + "loss": 0.6368, + "num_input_tokens_seen": 3575644160, + "step": 1705 + }, + { + "epoch": 0.13723894765391917, + "grad_norm": 13.588170051574707, + "learning_rate": 6.42190939011511e-07, + "loss": 0.4412, + "num_input_tokens_seen": 3577741312, + "step": 1706 + }, + { + "epoch": 0.13751017087062653, + "grad_norm": 13.829466819763184, + "learning_rate": 6.417957760433974e-07, + "loss": 0.4696, + "num_input_tokens_seen": 3579838464, + "step": 1707 + }, + { + "epoch": 0.13778139408733386, + "grad_norm": 17.291702270507812, + "learning_rate": 6.414005392015364e-07, + "loss": 0.4773, + "num_input_tokens_seen": 3581935616, + "step": 1708 + }, + { + "epoch": 0.13805261730404123, + "grad_norm": 15.568994522094727, + "learning_rate": 6.410052288039994e-07, + "loss": 0.631, + "num_input_tokens_seen": 3584032768, + "step": 1709 + }, + { + "epoch": 0.1383238405207486, + "grad_norm": 13.949945449829102, + "learning_rate": 6.406098451689171e-07, + "loss": 0.3946, + "num_input_tokens_seen": 3586129920, + "step": 1710 + }, + { + "epoch": 0.13859506373745592, + "grad_norm": 18.95940589904785, + "learning_rate": 6.402143886144797e-07, + "loss": 0.453, + "num_input_tokens_seen": 3588227072, + "step": 1711 + }, + { + "epoch": 0.13886628695416328, + "grad_norm": 17.928314208984375, + "learning_rate": 6.39818859458935e-07, + "loss": 0.4933, + "num_input_tokens_seen": 3590324224, + "step": 1712 + }, + { + "epoch": 0.13913751017087062, + "grad_norm": 24.465429306030273, + "learning_rate": 6.394232580205903e-07, + "loss": 0.9862, + "num_input_tokens_seen": 3592421376, + "step": 1713 + }, + { + "epoch": 0.13940873338757798, + "grad_norm": 16.95507049560547, + "learning_rate": 6.390275846178102e-07, + "loss": 0.6797, + "num_input_tokens_seen": 3594518528, + "step": 1714 + }, + { + "epoch": 0.13967995660428534, + "grad_norm": 12.937496185302734, + "learning_rate": 6.386318395690178e-07, + "loss": 0.296, + "num_input_tokens_seen": 3596615680, + "step": 1715 + }, + { + "epoch": 0.13995117982099267, + "grad_norm": 16.558284759521484, + "learning_rate": 6.382360231926935e-07, + "loss": 0.4922, + "num_input_tokens_seen": 3598712832, + "step": 1716 + }, + { + "epoch": 0.14022240303770003, + "grad_norm": 15.192840576171875, + "learning_rate": 6.378401358073754e-07, + "loss": 0.4696, + "num_input_tokens_seen": 3600809984, + "step": 1717 + }, + { + "epoch": 0.14049362625440737, + "grad_norm": 13.925806045532227, + "learning_rate": 6.374441777316587e-07, + "loss": 0.539, + "num_input_tokens_seen": 3602907136, + "step": 1718 + }, + { + "epoch": 0.14076484947111473, + "grad_norm": 12.96522331237793, + "learning_rate": 6.370481492841952e-07, + "loss": 0.438, + "num_input_tokens_seen": 3605004288, + "step": 1719 + }, + { + "epoch": 0.1410360726878221, + "grad_norm": 13.147494316101074, + "learning_rate": 6.366520507836934e-07, + "loss": 0.4557, + "num_input_tokens_seen": 3607101440, + "step": 1720 + }, + { + "epoch": 0.14130729590452942, + "grad_norm": 14.254191398620605, + "learning_rate": 6.362558825489187e-07, + "loss": 0.337, + "num_input_tokens_seen": 3609198592, + "step": 1721 + }, + { + "epoch": 0.14157851912123678, + "grad_norm": 16.945711135864258, + "learning_rate": 6.35859644898692e-07, + "loss": 0.565, + "num_input_tokens_seen": 3611295744, + "step": 1722 + }, + { + "epoch": 0.14184974233794412, + "grad_norm": 16.973342895507812, + "learning_rate": 6.354633381518901e-07, + "loss": 0.6576, + "num_input_tokens_seen": 3613392896, + "step": 1723 + }, + { + "epoch": 0.14212096555465148, + "grad_norm": 15.234251022338867, + "learning_rate": 6.350669626274461e-07, + "loss": 0.4715, + "num_input_tokens_seen": 3615490048, + "step": 1724 + }, + { + "epoch": 0.14239218877135884, + "grad_norm": 14.848320007324219, + "learning_rate": 6.346705186443474e-07, + "loss": 0.4606, + "num_input_tokens_seen": 3617587200, + "step": 1725 + }, + { + "epoch": 0.14266341198806617, + "grad_norm": 14.895830154418945, + "learning_rate": 6.342740065216371e-07, + "loss": 0.5635, + "num_input_tokens_seen": 3619684352, + "step": 1726 + }, + { + "epoch": 0.14293463520477354, + "grad_norm": 12.153023719787598, + "learning_rate": 6.338774265784134e-07, + "loss": 0.3342, + "num_input_tokens_seen": 3621781504, + "step": 1727 + }, + { + "epoch": 0.14320585842148087, + "grad_norm": 14.42286205291748, + "learning_rate": 6.334807791338286e-07, + "loss": 0.4809, + "num_input_tokens_seen": 3623878656, + "step": 1728 + }, + { + "epoch": 0.14347708163818823, + "grad_norm": 18.29535675048828, + "learning_rate": 6.330840645070894e-07, + "loss": 0.6635, + "num_input_tokens_seen": 3625975808, + "step": 1729 + }, + { + "epoch": 0.1437483048548956, + "grad_norm": 14.65046215057373, + "learning_rate": 6.326872830174566e-07, + "loss": 0.5119, + "num_input_tokens_seen": 3628072960, + "step": 1730 + }, + { + "epoch": 0.14401952807160293, + "grad_norm": 15.046059608459473, + "learning_rate": 6.32290434984245e-07, + "loss": 0.4347, + "num_input_tokens_seen": 3630170112, + "step": 1731 + }, + { + "epoch": 0.1442907512883103, + "grad_norm": 15.064492225646973, + "learning_rate": 6.318935207268227e-07, + "loss": 0.5375, + "num_input_tokens_seen": 3632267264, + "step": 1732 + }, + { + "epoch": 0.14456197450501762, + "grad_norm": 12.767688751220703, + "learning_rate": 6.314965405646113e-07, + "loss": 0.5119, + "num_input_tokens_seen": 3634364416, + "step": 1733 + }, + { + "epoch": 0.14483319772172498, + "grad_norm": 22.676115036010742, + "learning_rate": 6.31099494817085e-07, + "loss": 0.7369, + "num_input_tokens_seen": 3636461568, + "step": 1734 + }, + { + "epoch": 0.14510442093843234, + "grad_norm": 12.052305221557617, + "learning_rate": 6.307023838037714e-07, + "loss": 0.4235, + "num_input_tokens_seen": 3638558720, + "step": 1735 + }, + { + "epoch": 0.14537564415513968, + "grad_norm": 21.096866607666016, + "learning_rate": 6.303052078442503e-07, + "loss": 0.6735, + "num_input_tokens_seen": 3640655872, + "step": 1736 + }, + { + "epoch": 0.14564686737184704, + "grad_norm": 13.492481231689453, + "learning_rate": 6.299079672581537e-07, + "loss": 0.4236, + "num_input_tokens_seen": 3642753024, + "step": 1737 + }, + { + "epoch": 0.14591809058855437, + "grad_norm": 15.035043716430664, + "learning_rate": 6.295106623651661e-07, + "loss": 0.5539, + "num_input_tokens_seen": 3644850176, + "step": 1738 + }, + { + "epoch": 0.14618931380526173, + "grad_norm": 11.176956176757812, + "learning_rate": 6.291132934850225e-07, + "loss": 0.3057, + "num_input_tokens_seen": 3646947328, + "step": 1739 + }, + { + "epoch": 0.14646053702196907, + "grad_norm": 19.135766983032227, + "learning_rate": 6.28715860937511e-07, + "loss": 0.8421, + "num_input_tokens_seen": 3649044480, + "step": 1740 + }, + { + "epoch": 0.14673176023867643, + "grad_norm": 16.260631561279297, + "learning_rate": 6.283183650424701e-07, + "loss": 0.528, + "num_input_tokens_seen": 3651141632, + "step": 1741 + }, + { + "epoch": 0.1470029834553838, + "grad_norm": 18.478513717651367, + "learning_rate": 6.279208061197892e-07, + "loss": 0.7138, + "num_input_tokens_seen": 3653238784, + "step": 1742 + }, + { + "epoch": 0.14727420667209112, + "grad_norm": 13.567277908325195, + "learning_rate": 6.275231844894086e-07, + "loss": 0.43, + "num_input_tokens_seen": 3655335936, + "step": 1743 + }, + { + "epoch": 0.14754542988879848, + "grad_norm": 16.15982437133789, + "learning_rate": 6.271255004713192e-07, + "loss": 0.494, + "num_input_tokens_seen": 3657433088, + "step": 1744 + }, + { + "epoch": 0.14781665310550582, + "grad_norm": 10.626032829284668, + "learning_rate": 6.267277543855618e-07, + "loss": 0.3169, + "num_input_tokens_seen": 3659530240, + "step": 1745 + }, + { + "epoch": 0.14808787632221318, + "grad_norm": 16.266632080078125, + "learning_rate": 6.263299465522274e-07, + "loss": 0.5772, + "num_input_tokens_seen": 3661627392, + "step": 1746 + }, + { + "epoch": 0.14835909953892054, + "grad_norm": 12.792734146118164, + "learning_rate": 6.259320772914566e-07, + "loss": 0.4696, + "num_input_tokens_seen": 3663724544, + "step": 1747 + }, + { + "epoch": 0.14863032275562787, + "grad_norm": 18.667470932006836, + "learning_rate": 6.255341469234393e-07, + "loss": 0.8274, + "num_input_tokens_seen": 3665821696, + "step": 1748 + }, + { + "epoch": 0.14890154597233524, + "grad_norm": 15.332829475402832, + "learning_rate": 6.25136155768415e-07, + "loss": 0.6095, + "num_input_tokens_seen": 3667918848, + "step": 1749 + }, + { + "epoch": 0.14917276918904257, + "grad_norm": 12.824024200439453, + "learning_rate": 6.247381041466716e-07, + "loss": 0.4388, + "num_input_tokens_seen": 3670016000, + "step": 1750 + }, + { + "epoch": 0.14944399240574993, + "grad_norm": 9.92809772491455, + "learning_rate": 6.243399923785459e-07, + "loss": 0.2921, + "num_input_tokens_seen": 3672113152, + "step": 1751 + }, + { + "epoch": 0.1497152156224573, + "grad_norm": 15.665802001953125, + "learning_rate": 6.239418207844232e-07, + "loss": 0.6068, + "num_input_tokens_seen": 3674210304, + "step": 1752 + }, + { + "epoch": 0.14998643883916463, + "grad_norm": 15.602997779846191, + "learning_rate": 6.23543589684737e-07, + "loss": 0.5003, + "num_input_tokens_seen": 3676307456, + "step": 1753 + }, + { + "epoch": 0.150257662055872, + "grad_norm": 15.580018043518066, + "learning_rate": 6.231452993999683e-07, + "loss": 0.5314, + "num_input_tokens_seen": 3678404608, + "step": 1754 + }, + { + "epoch": 0.15052888527257932, + "grad_norm": 18.052650451660156, + "learning_rate": 6.22746950250646e-07, + "loss": 0.5794, + "num_input_tokens_seen": 3680501760, + "step": 1755 + }, + { + "epoch": 0.15080010848928668, + "grad_norm": 14.942179679870605, + "learning_rate": 6.223485425573463e-07, + "loss": 0.4202, + "num_input_tokens_seen": 3682598912, + "step": 1756 + }, + { + "epoch": 0.15107133170599404, + "grad_norm": 12.076883316040039, + "learning_rate": 6.219500766406926e-07, + "loss": 0.385, + "num_input_tokens_seen": 3684696064, + "step": 1757 + }, + { + "epoch": 0.15134255492270138, + "grad_norm": 16.620975494384766, + "learning_rate": 6.215515528213553e-07, + "loss": 0.5812, + "num_input_tokens_seen": 3686793216, + "step": 1758 + }, + { + "epoch": 0.15161377813940874, + "grad_norm": 11.731767654418945, + "learning_rate": 6.211529714200509e-07, + "loss": 0.376, + "num_input_tokens_seen": 3688890368, + "step": 1759 + }, + { + "epoch": 0.15188500135611607, + "grad_norm": 24.166730880737305, + "learning_rate": 6.207543327575426e-07, + "loss": 0.5209, + "num_input_tokens_seen": 3690987520, + "step": 1760 + }, + { + "epoch": 0.15215622457282343, + "grad_norm": 14.723969459533691, + "learning_rate": 6.203556371546399e-07, + "loss": 0.4827, + "num_input_tokens_seen": 3693084672, + "step": 1761 + }, + { + "epoch": 0.1524274477895308, + "grad_norm": 14.81787395477295, + "learning_rate": 6.199568849321975e-07, + "loss": 0.4707, + "num_input_tokens_seen": 3695181824, + "step": 1762 + }, + { + "epoch": 0.15269867100623813, + "grad_norm": 15.406253814697266, + "learning_rate": 6.195580764111163e-07, + "loss": 0.4198, + "num_input_tokens_seen": 3697278976, + "step": 1763 + }, + { + "epoch": 0.1529698942229455, + "grad_norm": 12.230108261108398, + "learning_rate": 6.191592119123419e-07, + "loss": 0.2609, + "num_input_tokens_seen": 3699376128, + "step": 1764 + }, + { + "epoch": 0.15324111743965282, + "grad_norm": 7.24866247177124, + "learning_rate": 6.187602917568655e-07, + "loss": 0.1833, + "num_input_tokens_seen": 3701473280, + "step": 1765 + }, + { + "epoch": 0.15351234065636019, + "grad_norm": 10.991303443908691, + "learning_rate": 6.183613162657229e-07, + "loss": 0.2878, + "num_input_tokens_seen": 3703570432, + "step": 1766 + }, + { + "epoch": 0.15378356387306755, + "grad_norm": 13.555655479431152, + "learning_rate": 6.179622857599942e-07, + "loss": 0.4992, + "num_input_tokens_seen": 3705667584, + "step": 1767 + }, + { + "epoch": 0.15405478708977488, + "grad_norm": 21.7625675201416, + "learning_rate": 6.175632005608043e-07, + "loss": 0.6617, + "num_input_tokens_seen": 3707764736, + "step": 1768 + }, + { + "epoch": 0.15432601030648224, + "grad_norm": 16.64663314819336, + "learning_rate": 6.171640609893213e-07, + "loss": 0.4283, + "num_input_tokens_seen": 3709861888, + "step": 1769 + }, + { + "epoch": 0.15459723352318958, + "grad_norm": 14.173835754394531, + "learning_rate": 6.167648673667581e-07, + "loss": 0.4511, + "num_input_tokens_seen": 3711959040, + "step": 1770 + }, + { + "epoch": 0.15486845673989694, + "grad_norm": 19.849355697631836, + "learning_rate": 6.1636562001437e-07, + "loss": 0.619, + "num_input_tokens_seen": 3714056192, + "step": 1771 + }, + { + "epoch": 0.1551396799566043, + "grad_norm": 13.180706024169922, + "learning_rate": 6.159663192534561e-07, + "loss": 0.486, + "num_input_tokens_seen": 3716153344, + "step": 1772 + }, + { + "epoch": 0.15541090317331163, + "grad_norm": 10.8189697265625, + "learning_rate": 6.155669654053592e-07, + "loss": 0.3495, + "num_input_tokens_seen": 3718250496, + "step": 1773 + }, + { + "epoch": 0.155682126390019, + "grad_norm": 28.293182373046875, + "learning_rate": 6.151675587914631e-07, + "loss": 0.7353, + "num_input_tokens_seen": 3720347648, + "step": 1774 + }, + { + "epoch": 0.15595334960672633, + "grad_norm": 21.1544132232666, + "learning_rate": 6.147680997331958e-07, + "loss": 0.8318, + "num_input_tokens_seen": 3722444800, + "step": 1775 + }, + { + "epoch": 0.1562245728234337, + "grad_norm": 16.06313705444336, + "learning_rate": 6.143685885520263e-07, + "loss": 0.5683, + "num_input_tokens_seen": 3724541952, + "step": 1776 + }, + { + "epoch": 0.15649579604014105, + "grad_norm": 14.972505569458008, + "learning_rate": 6.13969025569466e-07, + "loss": 0.3172, + "num_input_tokens_seen": 3726639104, + "step": 1777 + }, + { + "epoch": 0.15676701925684838, + "grad_norm": 11.951261520385742, + "learning_rate": 6.135694111070684e-07, + "loss": 0.3382, + "num_input_tokens_seen": 3728736256, + "step": 1778 + }, + { + "epoch": 0.15703824247355574, + "grad_norm": 15.1442289352417, + "learning_rate": 6.131697454864277e-07, + "loss": 0.433, + "num_input_tokens_seen": 3730833408, + "step": 1779 + }, + { + "epoch": 0.15730946569026308, + "grad_norm": 18.318092346191406, + "learning_rate": 6.127700290291794e-07, + "loss": 0.6716, + "num_input_tokens_seen": 3732930560, + "step": 1780 + }, + { + "epoch": 0.15758068890697044, + "grad_norm": 17.144010543823242, + "learning_rate": 6.123702620570005e-07, + "loss": 0.5492, + "num_input_tokens_seen": 3735027712, + "step": 1781 + }, + { + "epoch": 0.15785191212367777, + "grad_norm": 18.71530532836914, + "learning_rate": 6.11970444891608e-07, + "loss": 0.5079, + "num_input_tokens_seen": 3737124864, + "step": 1782 + }, + { + "epoch": 0.15812313534038513, + "grad_norm": 11.793743133544922, + "learning_rate": 6.115705778547597e-07, + "loss": 0.258, + "num_input_tokens_seen": 3739222016, + "step": 1783 + }, + { + "epoch": 0.1583943585570925, + "grad_norm": 18.33780288696289, + "learning_rate": 6.111706612682532e-07, + "loss": 0.6809, + "num_input_tokens_seen": 3741319168, + "step": 1784 + }, + { + "epoch": 0.15866558177379983, + "grad_norm": 12.799018859863281, + "learning_rate": 6.107706954539261e-07, + "loss": 0.3735, + "num_input_tokens_seen": 3743416320, + "step": 1785 + }, + { + "epoch": 0.1589368049905072, + "grad_norm": 14.640045166015625, + "learning_rate": 6.103706807336559e-07, + "loss": 0.5625, + "num_input_tokens_seen": 3745513472, + "step": 1786 + }, + { + "epoch": 0.15920802820721452, + "grad_norm": 11.989145278930664, + "learning_rate": 6.099706174293592e-07, + "loss": 0.4499, + "num_input_tokens_seen": 3747610624, + "step": 1787 + }, + { + "epoch": 0.1594792514239219, + "grad_norm": 14.482452392578125, + "learning_rate": 6.095705058629914e-07, + "loss": 0.3148, + "num_input_tokens_seen": 3749707776, + "step": 1788 + }, + { + "epoch": 0.15975047464062925, + "grad_norm": 12.022076606750488, + "learning_rate": 6.091703463565475e-07, + "loss": 0.3346, + "num_input_tokens_seen": 3751804928, + "step": 1789 + }, + { + "epoch": 0.16002169785733658, + "grad_norm": 14.825411796569824, + "learning_rate": 6.087701392320606e-07, + "loss": 0.5093, + "num_input_tokens_seen": 3753902080, + "step": 1790 + }, + { + "epoch": 0.16029292107404394, + "grad_norm": 12.915584564208984, + "learning_rate": 6.083698848116018e-07, + "loss": 0.3152, + "num_input_tokens_seen": 3755999232, + "step": 1791 + }, + { + "epoch": 0.16056414429075128, + "grad_norm": 21.649930953979492, + "learning_rate": 6.079695834172808e-07, + "loss": 0.9817, + "num_input_tokens_seen": 3758096384, + "step": 1792 + }, + { + "epoch": 0.16083536750745864, + "grad_norm": 15.43675422668457, + "learning_rate": 6.075692353712451e-07, + "loss": 0.4618, + "num_input_tokens_seen": 3760193536, + "step": 1793 + }, + { + "epoch": 0.161106590724166, + "grad_norm": 16.996694564819336, + "learning_rate": 6.071688409956793e-07, + "loss": 0.5975, + "num_input_tokens_seen": 3762290688, + "step": 1794 + }, + { + "epoch": 0.16137781394087333, + "grad_norm": 16.773155212402344, + "learning_rate": 6.06768400612806e-07, + "loss": 0.5768, + "num_input_tokens_seen": 3764387840, + "step": 1795 + }, + { + "epoch": 0.1616490371575807, + "grad_norm": 22.683000564575195, + "learning_rate": 6.063679145448838e-07, + "loss": 0.7318, + "num_input_tokens_seen": 3766484992, + "step": 1796 + }, + { + "epoch": 0.16192026037428803, + "grad_norm": 13.12799072265625, + "learning_rate": 6.05967383114209e-07, + "loss": 0.3581, + "num_input_tokens_seen": 3768582144, + "step": 1797 + }, + { + "epoch": 0.1621914835909954, + "grad_norm": 18.787900924682617, + "learning_rate": 6.055668066431142e-07, + "loss": 0.6844, + "num_input_tokens_seen": 3770679296, + "step": 1798 + }, + { + "epoch": 0.16246270680770275, + "grad_norm": 10.89427661895752, + "learning_rate": 6.051661854539677e-07, + "loss": 0.3149, + "num_input_tokens_seen": 3772776448, + "step": 1799 + }, + { + "epoch": 0.16273393002441008, + "grad_norm": 18.79866600036621, + "learning_rate": 6.047655198691742e-07, + "loss": 0.8429, + "num_input_tokens_seen": 3774873600, + "step": 1800 + }, + { + "epoch": 0.16300515324111745, + "grad_norm": 13.372797966003418, + "learning_rate": 6.043648102111745e-07, + "loss": 0.2438, + "num_input_tokens_seen": 3776970752, + "step": 1801 + }, + { + "epoch": 0.16327637645782478, + "grad_norm": 18.644472122192383, + "learning_rate": 6.039640568024443e-07, + "loss": 0.6407, + "num_input_tokens_seen": 3779067904, + "step": 1802 + }, + { + "epoch": 0.16354759967453214, + "grad_norm": 15.050209045410156, + "learning_rate": 6.035632599654946e-07, + "loss": 0.5603, + "num_input_tokens_seen": 3781165056, + "step": 1803 + }, + { + "epoch": 0.1638188228912395, + "grad_norm": 12.185089111328125, + "learning_rate": 6.031624200228715e-07, + "loss": 0.3552, + "num_input_tokens_seen": 3783262208, + "step": 1804 + }, + { + "epoch": 0.16409004610794684, + "grad_norm": 13.226753234863281, + "learning_rate": 6.027615372971558e-07, + "loss": 0.4319, + "num_input_tokens_seen": 3785359360, + "step": 1805 + }, + { + "epoch": 0.1643612693246542, + "grad_norm": 13.992337226867676, + "learning_rate": 6.023606121109626e-07, + "loss": 0.4704, + "num_input_tokens_seen": 3787456512, + "step": 1806 + }, + { + "epoch": 0.16463249254136153, + "grad_norm": 17.444171905517578, + "learning_rate": 6.019596447869413e-07, + "loss": 0.4482, + "num_input_tokens_seen": 3789553664, + "step": 1807 + }, + { + "epoch": 0.1649037157580689, + "grad_norm": 14.175294876098633, + "learning_rate": 6.015586356477749e-07, + "loss": 0.4118, + "num_input_tokens_seen": 3791650816, + "step": 1808 + }, + { + "epoch": 0.16517493897477625, + "grad_norm": 18.26668930053711, + "learning_rate": 6.011575850161805e-07, + "loss": 0.5852, + "num_input_tokens_seen": 3793747968, + "step": 1809 + }, + { + "epoch": 0.1654461621914836, + "grad_norm": 15.502846717834473, + "learning_rate": 6.007564932149086e-07, + "loss": 0.5226, + "num_input_tokens_seen": 3795845120, + "step": 1810 + }, + { + "epoch": 0.16571738540819095, + "grad_norm": 13.446937561035156, + "learning_rate": 6.003553605667423e-07, + "loss": 0.4124, + "num_input_tokens_seen": 3797942272, + "step": 1811 + }, + { + "epoch": 0.16598860862489828, + "grad_norm": 13.046964645385742, + "learning_rate": 5.999541873944979e-07, + "loss": 0.4353, + "num_input_tokens_seen": 3800039424, + "step": 1812 + }, + { + "epoch": 0.16625983184160564, + "grad_norm": 11.632257461547852, + "learning_rate": 5.995529740210244e-07, + "loss": 0.3998, + "num_input_tokens_seen": 3802136576, + "step": 1813 + }, + { + "epoch": 0.166531055058313, + "grad_norm": 12.484574317932129, + "learning_rate": 5.99151720769203e-07, + "loss": 0.318, + "num_input_tokens_seen": 3804233728, + "step": 1814 + }, + { + "epoch": 0.16680227827502034, + "grad_norm": 11.363998413085938, + "learning_rate": 5.987504279619473e-07, + "loss": 0.4028, + "num_input_tokens_seen": 3806330880, + "step": 1815 + }, + { + "epoch": 0.1670735014917277, + "grad_norm": 18.80327033996582, + "learning_rate": 5.98349095922202e-07, + "loss": 0.3596, + "num_input_tokens_seen": 3808428032, + "step": 1816 + }, + { + "epoch": 0.16734472470843503, + "grad_norm": 27.840991973876953, + "learning_rate": 5.979477249729442e-07, + "loss": 1.0105, + "num_input_tokens_seen": 3810525184, + "step": 1817 + }, + { + "epoch": 0.1676159479251424, + "grad_norm": 13.593608856201172, + "learning_rate": 5.975463154371822e-07, + "loss": 0.5061, + "num_input_tokens_seen": 3812622336, + "step": 1818 + }, + { + "epoch": 0.16788717114184976, + "grad_norm": 14.841533660888672, + "learning_rate": 5.971448676379544e-07, + "loss": 0.4594, + "num_input_tokens_seen": 3814719488, + "step": 1819 + }, + { + "epoch": 0.1681583943585571, + "grad_norm": 10.198492050170898, + "learning_rate": 5.967433818983311e-07, + "loss": 0.3427, + "num_input_tokens_seen": 3816816640, + "step": 1820 + }, + { + "epoch": 0.16842961757526445, + "grad_norm": 12.751827239990234, + "learning_rate": 5.963418585414129e-07, + "loss": 0.3955, + "num_input_tokens_seen": 3818913792, + "step": 1821 + }, + { + "epoch": 0.16870084079197178, + "grad_norm": 11.35025405883789, + "learning_rate": 5.959402978903306e-07, + "loss": 0.2443, + "num_input_tokens_seen": 3821010944, + "step": 1822 + }, + { + "epoch": 0.16897206400867915, + "grad_norm": 16.358675003051758, + "learning_rate": 5.955387002682445e-07, + "loss": 0.7063, + "num_input_tokens_seen": 3823108096, + "step": 1823 + }, + { + "epoch": 0.16924328722538648, + "grad_norm": 10.499540328979492, + "learning_rate": 5.951370659983452e-07, + "loss": 0.3716, + "num_input_tokens_seen": 3825205248, + "step": 1824 + }, + { + "epoch": 0.16951451044209384, + "grad_norm": 9.75558853149414, + "learning_rate": 5.94735395403853e-07, + "loss": 0.2681, + "num_input_tokens_seen": 3827302400, + "step": 1825 + }, + { + "epoch": 0.1697857336588012, + "grad_norm": 21.415407180786133, + "learning_rate": 5.943336888080167e-07, + "loss": 0.7333, + "num_input_tokens_seen": 3829399552, + "step": 1826 + }, + { + "epoch": 0.17005695687550854, + "grad_norm": 16.07489585876465, + "learning_rate": 5.939319465341148e-07, + "loss": 0.5555, + "num_input_tokens_seen": 3831496704, + "step": 1827 + }, + { + "epoch": 0.1703281800922159, + "grad_norm": 16.00339126586914, + "learning_rate": 5.93530168905454e-07, + "loss": 0.4852, + "num_input_tokens_seen": 3833593856, + "step": 1828 + }, + { + "epoch": 0.17059940330892323, + "grad_norm": 15.944879531860352, + "learning_rate": 5.931283562453696e-07, + "loss": 0.511, + "num_input_tokens_seen": 3835691008, + "step": 1829 + }, + { + "epoch": 0.1708706265256306, + "grad_norm": 12.453446388244629, + "learning_rate": 5.927265088772255e-07, + "loss": 0.4054, + "num_input_tokens_seen": 3837788160, + "step": 1830 + }, + { + "epoch": 0.17114184974233795, + "grad_norm": 16.834558486938477, + "learning_rate": 5.923246271244127e-07, + "loss": 0.6094, + "num_input_tokens_seen": 3839885312, + "step": 1831 + }, + { + "epoch": 0.1714130729590453, + "grad_norm": 20.197141647338867, + "learning_rate": 5.919227113103508e-07, + "loss": 0.4895, + "num_input_tokens_seen": 3841982464, + "step": 1832 + }, + { + "epoch": 0.17168429617575265, + "grad_norm": 10.21462631225586, + "learning_rate": 5.915207617584858e-07, + "loss": 0.2921, + "num_input_tokens_seen": 3844079616, + "step": 1833 + }, + { + "epoch": 0.17195551939245998, + "grad_norm": 15.528749465942383, + "learning_rate": 5.911187787922918e-07, + "loss": 0.5211, + "num_input_tokens_seen": 3846176768, + "step": 1834 + }, + { + "epoch": 0.17222674260916734, + "grad_norm": 13.571805000305176, + "learning_rate": 5.907167627352697e-07, + "loss": 0.3398, + "num_input_tokens_seen": 3848273920, + "step": 1835 + }, + { + "epoch": 0.1724979658258747, + "grad_norm": 12.700458526611328, + "learning_rate": 5.90314713910946e-07, + "loss": 0.4671, + "num_input_tokens_seen": 3850371072, + "step": 1836 + }, + { + "epoch": 0.17276918904258204, + "grad_norm": 12.876029968261719, + "learning_rate": 5.899126326428746e-07, + "loss": 0.4635, + "num_input_tokens_seen": 3852468224, + "step": 1837 + }, + { + "epoch": 0.1730404122592894, + "grad_norm": 19.769418716430664, + "learning_rate": 5.895105192546353e-07, + "loss": 0.7521, + "num_input_tokens_seen": 3854565376, + "step": 1838 + }, + { + "epoch": 0.17331163547599673, + "grad_norm": 10.773146629333496, + "learning_rate": 5.891083740698337e-07, + "loss": 0.2714, + "num_input_tokens_seen": 3856662528, + "step": 1839 + }, + { + "epoch": 0.1735828586927041, + "grad_norm": 20.43178939819336, + "learning_rate": 5.887061974121007e-07, + "loss": 0.4969, + "num_input_tokens_seen": 3858759680, + "step": 1840 + }, + { + "epoch": 0.17385408190941146, + "grad_norm": 16.35590934753418, + "learning_rate": 5.883039896050928e-07, + "loss": 0.6751, + "num_input_tokens_seen": 3860856832, + "step": 1841 + }, + { + "epoch": 0.1741253051261188, + "grad_norm": 13.783025741577148, + "learning_rate": 5.879017509724919e-07, + "loss": 0.4219, + "num_input_tokens_seen": 3862953984, + "step": 1842 + }, + { + "epoch": 0.17439652834282615, + "grad_norm": 16.210859298706055, + "learning_rate": 5.874994818380039e-07, + "loss": 0.5322, + "num_input_tokens_seen": 3865051136, + "step": 1843 + }, + { + "epoch": 0.17466775155953349, + "grad_norm": 14.429267883300781, + "learning_rate": 5.8709718252536e-07, + "loss": 0.4487, + "num_input_tokens_seen": 3867148288, + "step": 1844 + }, + { + "epoch": 0.17493897477624085, + "grad_norm": 13.477580070495605, + "learning_rate": 5.866948533583152e-07, + "loss": 0.3541, + "num_input_tokens_seen": 3869245440, + "step": 1845 + }, + { + "epoch": 0.1752101979929482, + "grad_norm": 10.946706771850586, + "learning_rate": 5.862924946606487e-07, + "loss": 0.247, + "num_input_tokens_seen": 3871342592, + "step": 1846 + }, + { + "epoch": 0.17548142120965554, + "grad_norm": 17.450336456298828, + "learning_rate": 5.858901067561637e-07, + "loss": 0.5306, + "num_input_tokens_seen": 3873439744, + "step": 1847 + }, + { + "epoch": 0.1757526444263629, + "grad_norm": 18.952194213867188, + "learning_rate": 5.854876899686864e-07, + "loss": 0.8947, + "num_input_tokens_seen": 3875536896, + "step": 1848 + }, + { + "epoch": 0.17602386764307024, + "grad_norm": 16.47836685180664, + "learning_rate": 5.850852446220666e-07, + "loss": 0.6694, + "num_input_tokens_seen": 3877634048, + "step": 1849 + }, + { + "epoch": 0.1762950908597776, + "grad_norm": 14.285144805908203, + "learning_rate": 5.84682771040177e-07, + "loss": 0.3916, + "num_input_tokens_seen": 3879731200, + "step": 1850 + }, + { + "epoch": 0.17656631407648496, + "grad_norm": 11.325536727905273, + "learning_rate": 5.842802695469131e-07, + "loss": 0.3398, + "num_input_tokens_seen": 3881828352, + "step": 1851 + }, + { + "epoch": 0.1768375372931923, + "grad_norm": 14.129776954650879, + "learning_rate": 5.838777404661927e-07, + "loss": 0.3623, + "num_input_tokens_seen": 3883925504, + "step": 1852 + }, + { + "epoch": 0.17710876050989965, + "grad_norm": 16.82501792907715, + "learning_rate": 5.83475184121956e-07, + "loss": 0.707, + "num_input_tokens_seen": 3886022656, + "step": 1853 + }, + { + "epoch": 0.177379983726607, + "grad_norm": 21.06279182434082, + "learning_rate": 5.830726008381648e-07, + "loss": 0.704, + "num_input_tokens_seen": 3888119808, + "step": 1854 + }, + { + "epoch": 0.17765120694331435, + "grad_norm": 14.380115509033203, + "learning_rate": 5.826699909388031e-07, + "loss": 0.4179, + "num_input_tokens_seen": 3890216960, + "step": 1855 + }, + { + "epoch": 0.1779224301600217, + "grad_norm": 15.511592864990234, + "learning_rate": 5.822673547478757e-07, + "loss": 0.652, + "num_input_tokens_seen": 3892314112, + "step": 1856 + }, + { + "epoch": 0.17819365337672904, + "grad_norm": 16.734813690185547, + "learning_rate": 5.818646925894092e-07, + "loss": 0.6038, + "num_input_tokens_seen": 3894411264, + "step": 1857 + }, + { + "epoch": 0.1784648765934364, + "grad_norm": 9.239023208618164, + "learning_rate": 5.814620047874505e-07, + "loss": 0.2753, + "num_input_tokens_seen": 3896508416, + "step": 1858 + }, + { + "epoch": 0.17873609981014374, + "grad_norm": 14.336617469787598, + "learning_rate": 5.810592916660677e-07, + "loss": 0.4859, + "num_input_tokens_seen": 3898605568, + "step": 1859 + }, + { + "epoch": 0.1790073230268511, + "grad_norm": 13.082498550415039, + "learning_rate": 5.806565535493489e-07, + "loss": 0.3663, + "num_input_tokens_seen": 3900702720, + "step": 1860 + }, + { + "epoch": 0.17927854624355846, + "grad_norm": 12.747309684753418, + "learning_rate": 5.802537907614023e-07, + "loss": 0.3627, + "num_input_tokens_seen": 3902799872, + "step": 1861 + }, + { + "epoch": 0.1795497694602658, + "grad_norm": 19.392642974853516, + "learning_rate": 5.798510036263561e-07, + "loss": 0.8228, + "num_input_tokens_seen": 3904897024, + "step": 1862 + }, + { + "epoch": 0.17982099267697316, + "grad_norm": 14.767632484436035, + "learning_rate": 5.794481924683581e-07, + "loss": 0.5703, + "num_input_tokens_seen": 3906994176, + "step": 1863 + }, + { + "epoch": 0.1800922158936805, + "grad_norm": 17.085071563720703, + "learning_rate": 5.790453576115756e-07, + "loss": 0.4947, + "num_input_tokens_seen": 3909091328, + "step": 1864 + }, + { + "epoch": 0.18036343911038785, + "grad_norm": 10.42067813873291, + "learning_rate": 5.786424993801942e-07, + "loss": 0.2922, + "num_input_tokens_seen": 3911188480, + "step": 1865 + }, + { + "epoch": 0.18063466232709519, + "grad_norm": 12.602887153625488, + "learning_rate": 5.782396180984194e-07, + "loss": 0.353, + "num_input_tokens_seen": 3913285632, + "step": 1866 + }, + { + "epoch": 0.18090588554380255, + "grad_norm": 13.061318397521973, + "learning_rate": 5.778367140904746e-07, + "loss": 0.4512, + "num_input_tokens_seen": 3915382784, + "step": 1867 + }, + { + "epoch": 0.1811771087605099, + "grad_norm": 18.99317741394043, + "learning_rate": 5.774337876806016e-07, + "loss": 0.6254, + "num_input_tokens_seen": 3917479936, + "step": 1868 + }, + { + "epoch": 0.18144833197721724, + "grad_norm": 16.320486068725586, + "learning_rate": 5.770308391930601e-07, + "loss": 0.5455, + "num_input_tokens_seen": 3919577088, + "step": 1869 + }, + { + "epoch": 0.1817195551939246, + "grad_norm": 14.826481819152832, + "learning_rate": 5.766278689521278e-07, + "loss": 0.4508, + "num_input_tokens_seen": 3921674240, + "step": 1870 + }, + { + "epoch": 0.18199077841063194, + "grad_norm": 16.660938262939453, + "learning_rate": 5.762248772820999e-07, + "loss": 0.3395, + "num_input_tokens_seen": 3923771392, + "step": 1871 + }, + { + "epoch": 0.1822620016273393, + "grad_norm": 11.43323040008545, + "learning_rate": 5.758218645072887e-07, + "loss": 0.3886, + "num_input_tokens_seen": 3925868544, + "step": 1872 + }, + { + "epoch": 0.18253322484404666, + "grad_norm": 19.8907470703125, + "learning_rate": 5.754188309520235e-07, + "loss": 0.6173, + "num_input_tokens_seen": 3927965696, + "step": 1873 + }, + { + "epoch": 0.182804448060754, + "grad_norm": 26.984996795654297, + "learning_rate": 5.750157769406504e-07, + "loss": 0.6616, + "num_input_tokens_seen": 3930062848, + "step": 1874 + }, + { + "epoch": 0.18307567127746135, + "grad_norm": 15.848214149475098, + "learning_rate": 5.74612702797532e-07, + "loss": 0.7792, + "num_input_tokens_seen": 3932160000, + "step": 1875 + }, + { + "epoch": 0.1833468944941687, + "grad_norm": 17.739269256591797, + "learning_rate": 5.74209608847047e-07, + "loss": 0.6638, + "num_input_tokens_seen": 3934257152, + "step": 1876 + }, + { + "epoch": 0.18361811771087605, + "grad_norm": 21.376651763916016, + "learning_rate": 5.7380649541359e-07, + "loss": 0.7258, + "num_input_tokens_seen": 3936354304, + "step": 1877 + }, + { + "epoch": 0.1838893409275834, + "grad_norm": 11.857672691345215, + "learning_rate": 5.734033628215714e-07, + "loss": 0.4125, + "num_input_tokens_seen": 3938451456, + "step": 1878 + }, + { + "epoch": 0.18416056414429074, + "grad_norm": 17.930675506591797, + "learning_rate": 5.730002113954169e-07, + "loss": 0.6946, + "num_input_tokens_seen": 3940548608, + "step": 1879 + }, + { + "epoch": 0.1844317873609981, + "grad_norm": 18.330848693847656, + "learning_rate": 5.725970414595675e-07, + "loss": 0.6827, + "num_input_tokens_seen": 3942645760, + "step": 1880 + }, + { + "epoch": 0.18470301057770544, + "grad_norm": 16.09734535217285, + "learning_rate": 5.721938533384791e-07, + "loss": 0.6227, + "num_input_tokens_seen": 3944742912, + "step": 1881 + }, + { + "epoch": 0.1849742337944128, + "grad_norm": 14.484498023986816, + "learning_rate": 5.717906473566219e-07, + "loss": 0.4096, + "num_input_tokens_seen": 3946840064, + "step": 1882 + }, + { + "epoch": 0.18524545701112016, + "grad_norm": 21.29253578186035, + "learning_rate": 5.71387423838481e-07, + "loss": 0.658, + "num_input_tokens_seen": 3948937216, + "step": 1883 + }, + { + "epoch": 0.1855166802278275, + "grad_norm": 12.313188552856445, + "learning_rate": 5.709841831085553e-07, + "loss": 0.3509, + "num_input_tokens_seen": 3951034368, + "step": 1884 + }, + { + "epoch": 0.18578790344453486, + "grad_norm": 12.124674797058105, + "learning_rate": 5.705809254913576e-07, + "loss": 0.3977, + "num_input_tokens_seen": 3953131520, + "step": 1885 + }, + { + "epoch": 0.1860591266612422, + "grad_norm": 19.086606979370117, + "learning_rate": 5.70177651311414e-07, + "loss": 0.7218, + "num_input_tokens_seen": 3955228672, + "step": 1886 + }, + { + "epoch": 0.18633034987794955, + "grad_norm": 11.468111991882324, + "learning_rate": 5.697743608932646e-07, + "loss": 0.3488, + "num_input_tokens_seen": 3957325824, + "step": 1887 + }, + { + "epoch": 0.18660157309465691, + "grad_norm": 22.346908569335938, + "learning_rate": 5.693710545614621e-07, + "loss": 0.9688, + "num_input_tokens_seen": 3959422976, + "step": 1888 + }, + { + "epoch": 0.18687279631136425, + "grad_norm": 18.068498611450195, + "learning_rate": 5.689677326405719e-07, + "loss": 0.7224, + "num_input_tokens_seen": 3961520128, + "step": 1889 + }, + { + "epoch": 0.1871440195280716, + "grad_norm": 21.589921951293945, + "learning_rate": 5.685643954551722e-07, + "loss": 0.8734, + "num_input_tokens_seen": 3963617280, + "step": 1890 + }, + { + "epoch": 0.18741524274477894, + "grad_norm": 15.714263916015625, + "learning_rate": 5.681610433298535e-07, + "loss": 0.5104, + "num_input_tokens_seen": 3965714432, + "step": 1891 + }, + { + "epoch": 0.1876864659614863, + "grad_norm": 12.886143684387207, + "learning_rate": 5.677576765892182e-07, + "loss": 0.4624, + "num_input_tokens_seen": 3967811584, + "step": 1892 + }, + { + "epoch": 0.18795768917819367, + "grad_norm": 20.087732315063477, + "learning_rate": 5.673542955578806e-07, + "loss": 0.4943, + "num_input_tokens_seen": 3969908736, + "step": 1893 + }, + { + "epoch": 0.188228912394901, + "grad_norm": 12.997749328613281, + "learning_rate": 5.669509005604663e-07, + "loss": 0.3788, + "num_input_tokens_seen": 3972005888, + "step": 1894 + }, + { + "epoch": 0.18850013561160836, + "grad_norm": 18.928436279296875, + "learning_rate": 5.665474919216122e-07, + "loss": 0.7576, + "num_input_tokens_seen": 3974103040, + "step": 1895 + }, + { + "epoch": 0.1887713588283157, + "grad_norm": 12.904560089111328, + "learning_rate": 5.661440699659663e-07, + "loss": 0.4746, + "num_input_tokens_seen": 3976200192, + "step": 1896 + }, + { + "epoch": 0.18904258204502306, + "grad_norm": 13.534509658813477, + "learning_rate": 5.657406350181872e-07, + "loss": 0.4028, + "num_input_tokens_seen": 3978297344, + "step": 1897 + }, + { + "epoch": 0.18931380526173042, + "grad_norm": 18.946544647216797, + "learning_rate": 5.65337187402944e-07, + "loss": 0.7937, + "num_input_tokens_seen": 3980394496, + "step": 1898 + }, + { + "epoch": 0.18958502847843775, + "grad_norm": 21.76073455810547, + "learning_rate": 5.64933727444916e-07, + "loss": 0.5819, + "num_input_tokens_seen": 3982491648, + "step": 1899 + }, + { + "epoch": 0.1898562516951451, + "grad_norm": 13.195499420166016, + "learning_rate": 5.645302554687925e-07, + "loss": 0.4144, + "num_input_tokens_seen": 3984588800, + "step": 1900 + }, + { + "epoch": 0.19012747491185245, + "grad_norm": 17.136215209960938, + "learning_rate": 5.641267717992723e-07, + "loss": 0.4328, + "num_input_tokens_seen": 3986685952, + "step": 1901 + }, + { + "epoch": 0.1903986981285598, + "grad_norm": 13.509635925292969, + "learning_rate": 5.637232767610637e-07, + "loss": 0.4364, + "num_input_tokens_seen": 3988783104, + "step": 1902 + }, + { + "epoch": 0.19066992134526717, + "grad_norm": 16.542097091674805, + "learning_rate": 5.63319770678884e-07, + "loss": 0.4641, + "num_input_tokens_seen": 3990880256, + "step": 1903 + }, + { + "epoch": 0.1909411445619745, + "grad_norm": 13.608541488647461, + "learning_rate": 5.6291625387746e-07, + "loss": 0.403, + "num_input_tokens_seen": 3992977408, + "step": 1904 + }, + { + "epoch": 0.19121236777868186, + "grad_norm": 15.277917861938477, + "learning_rate": 5.625127266815263e-07, + "loss": 0.3942, + "num_input_tokens_seen": 3995074560, + "step": 1905 + }, + { + "epoch": 0.1914835909953892, + "grad_norm": 20.041263580322266, + "learning_rate": 5.621091894158261e-07, + "loss": 0.3556, + "num_input_tokens_seen": 3997171712, + "step": 1906 + }, + { + "epoch": 0.19175481421209656, + "grad_norm": 17.10890769958496, + "learning_rate": 5.617056424051113e-07, + "loss": 0.5611, + "num_input_tokens_seen": 3999268864, + "step": 1907 + }, + { + "epoch": 0.1920260374288039, + "grad_norm": 19.5421199798584, + "learning_rate": 5.613020859741408e-07, + "loss": 0.5567, + "num_input_tokens_seen": 4001366016, + "step": 1908 + }, + { + "epoch": 0.19229726064551125, + "grad_norm": 20.52210235595703, + "learning_rate": 5.608985204476817e-07, + "loss": 0.7141, + "num_input_tokens_seen": 4003463168, + "step": 1909 + }, + { + "epoch": 0.19256848386221861, + "grad_norm": 26.15365982055664, + "learning_rate": 5.60494946150508e-07, + "loss": 0.5505, + "num_input_tokens_seen": 4005560320, + "step": 1910 + }, + { + "epoch": 0.19283970707892595, + "grad_norm": 14.202539443969727, + "learning_rate": 5.600913634074009e-07, + "loss": 0.4245, + "num_input_tokens_seen": 4007657472, + "step": 1911 + }, + { + "epoch": 0.1931109302956333, + "grad_norm": 17.88532257080078, + "learning_rate": 5.596877725431487e-07, + "loss": 0.4943, + "num_input_tokens_seen": 4009754624, + "step": 1912 + }, + { + "epoch": 0.19338215351234064, + "grad_norm": 10.894256591796875, + "learning_rate": 5.592841738825457e-07, + "loss": 0.3248, + "num_input_tokens_seen": 4011851776, + "step": 1913 + }, + { + "epoch": 0.193653376729048, + "grad_norm": 14.625932693481445, + "learning_rate": 5.588805677503928e-07, + "loss": 0.4324, + "num_input_tokens_seen": 4013948928, + "step": 1914 + }, + { + "epoch": 0.19392459994575537, + "grad_norm": 12.210338592529297, + "learning_rate": 5.58476954471497e-07, + "loss": 0.4082, + "num_input_tokens_seen": 4016046080, + "step": 1915 + }, + { + "epoch": 0.1941958231624627, + "grad_norm": 14.13292407989502, + "learning_rate": 5.580733343706708e-07, + "loss": 0.6738, + "num_input_tokens_seen": 4018143232, + "step": 1916 + }, + { + "epoch": 0.19446704637917006, + "grad_norm": 17.628400802612305, + "learning_rate": 5.576697077727323e-07, + "loss": 0.529, + "num_input_tokens_seen": 4020240384, + "step": 1917 + }, + { + "epoch": 0.1947382695958774, + "grad_norm": 16.035490036010742, + "learning_rate": 5.572660750025049e-07, + "loss": 0.4977, + "num_input_tokens_seen": 4022337536, + "step": 1918 + }, + { + "epoch": 0.19500949281258476, + "grad_norm": 12.913204193115234, + "learning_rate": 5.568624363848166e-07, + "loss": 0.3954, + "num_input_tokens_seen": 4024434688, + "step": 1919 + }, + { + "epoch": 0.19528071602929212, + "grad_norm": 10.610041618347168, + "learning_rate": 5.564587922445008e-07, + "loss": 0.3049, + "num_input_tokens_seen": 4026531840, + "step": 1920 + }, + { + "epoch": 0.19555193924599945, + "grad_norm": 12.461469650268555, + "learning_rate": 5.560551429063949e-07, + "loss": 0.3567, + "num_input_tokens_seen": 4028628992, + "step": 1921 + }, + { + "epoch": 0.1958231624627068, + "grad_norm": 19.98131561279297, + "learning_rate": 5.556514886953403e-07, + "loss": 0.7164, + "num_input_tokens_seen": 4030726144, + "step": 1922 + }, + { + "epoch": 0.19609438567941415, + "grad_norm": 14.228310585021973, + "learning_rate": 5.552478299361826e-07, + "loss": 0.4572, + "num_input_tokens_seen": 4032823296, + "step": 1923 + }, + { + "epoch": 0.1963656088961215, + "grad_norm": 14.937055587768555, + "learning_rate": 5.548441669537712e-07, + "loss": 0.6819, + "num_input_tokens_seen": 4034920448, + "step": 1924 + }, + { + "epoch": 0.19663683211282887, + "grad_norm": 18.690763473510742, + "learning_rate": 5.544405000729584e-07, + "loss": 0.6, + "num_input_tokens_seen": 4037017600, + "step": 1925 + }, + { + "epoch": 0.1969080553295362, + "grad_norm": 18.642099380493164, + "learning_rate": 5.540368296186002e-07, + "loss": 0.4619, + "num_input_tokens_seen": 4039114752, + "step": 1926 + }, + { + "epoch": 0.19717927854624356, + "grad_norm": 11.240645408630371, + "learning_rate": 5.53633155915555e-07, + "loss": 0.3568, + "num_input_tokens_seen": 4041211904, + "step": 1927 + }, + { + "epoch": 0.1974505017629509, + "grad_norm": 19.68428611755371, + "learning_rate": 5.532294792886843e-07, + "loss": 0.8633, + "num_input_tokens_seen": 4043309056, + "step": 1928 + }, + { + "epoch": 0.19772172497965826, + "grad_norm": 15.801280975341797, + "learning_rate": 5.528258000628518e-07, + "loss": 0.4886, + "num_input_tokens_seen": 4045406208, + "step": 1929 + }, + { + "epoch": 0.19799294819636562, + "grad_norm": 14.434005737304688, + "learning_rate": 5.524221185629224e-07, + "loss": 0.5922, + "num_input_tokens_seen": 4047503360, + "step": 1930 + }, + { + "epoch": 0.19826417141307295, + "grad_norm": 12.741336822509766, + "learning_rate": 5.520184351137646e-07, + "loss": 0.4185, + "num_input_tokens_seen": 4049600512, + "step": 1931 + }, + { + "epoch": 0.19853539462978032, + "grad_norm": 11.868021965026855, + "learning_rate": 5.51614750040247e-07, + "loss": 0.3508, + "num_input_tokens_seen": 4051697664, + "step": 1932 + }, + { + "epoch": 0.19880661784648765, + "grad_norm": 18.497217178344727, + "learning_rate": 5.512110636672396e-07, + "loss": 0.5296, + "num_input_tokens_seen": 4053794816, + "step": 1933 + }, + { + "epoch": 0.199077841063195, + "grad_norm": 16.076515197753906, + "learning_rate": 5.508073763196146e-07, + "loss": 0.3934, + "num_input_tokens_seen": 4055891968, + "step": 1934 + }, + { + "epoch": 0.19934906427990237, + "grad_norm": 14.377509117126465, + "learning_rate": 5.504036883222438e-07, + "loss": 0.4811, + "num_input_tokens_seen": 4057989120, + "step": 1935 + }, + { + "epoch": 0.1996202874966097, + "grad_norm": 20.239578247070312, + "learning_rate": 5.5e-07, + "loss": 0.7046, + "num_input_tokens_seen": 4060086272, + "step": 1936 + }, + { + "epoch": 0.19989151071331707, + "grad_norm": 15.798903465270996, + "learning_rate": 5.495963116777562e-07, + "loss": 0.3267, + "num_input_tokens_seen": 4062183424, + "step": 1937 + }, + { + "epoch": 0.2001627339300244, + "grad_norm": 17.59996795654297, + "learning_rate": 5.491926236803854e-07, + "loss": 0.726, + "num_input_tokens_seen": 4064280576, + "step": 1938 + }, + { + "epoch": 0.20043395714673176, + "grad_norm": 14.18375301361084, + "learning_rate": 5.487889363327603e-07, + "loss": 0.5171, + "num_input_tokens_seen": 4066377728, + "step": 1939 + }, + { + "epoch": 0.20070518036343912, + "grad_norm": 12.889824867248535, + "learning_rate": 5.483852499597532e-07, + "loss": 0.3304, + "num_input_tokens_seen": 4068474880, + "step": 1940 + }, + { + "epoch": 0.20097640358014646, + "grad_norm": 18.258941650390625, + "learning_rate": 5.479815648862355e-07, + "loss": 0.6463, + "num_input_tokens_seen": 4070572032, + "step": 1941 + }, + { + "epoch": 0.20124762679685382, + "grad_norm": 9.00694465637207, + "learning_rate": 5.475778814370776e-07, + "loss": 0.2284, + "num_input_tokens_seen": 4072669184, + "step": 1942 + }, + { + "epoch": 0.20151885001356115, + "grad_norm": 11.947250366210938, + "learning_rate": 5.471741999371483e-07, + "loss": 0.3377, + "num_input_tokens_seen": 4074766336, + "step": 1943 + }, + { + "epoch": 0.2017900732302685, + "grad_norm": 14.319008827209473, + "learning_rate": 5.467705207113156e-07, + "loss": 0.3758, + "num_input_tokens_seen": 4076863488, + "step": 1944 + }, + { + "epoch": 0.20206129644697587, + "grad_norm": 15.36997127532959, + "learning_rate": 5.46366844084445e-07, + "loss": 0.5119, + "num_input_tokens_seen": 4078960640, + "step": 1945 + }, + { + "epoch": 0.2023325196636832, + "grad_norm": 16.01044273376465, + "learning_rate": 5.459631703813998e-07, + "loss": 0.4673, + "num_input_tokens_seen": 4081057792, + "step": 1946 + }, + { + "epoch": 0.20260374288039057, + "grad_norm": 12.556026458740234, + "learning_rate": 5.455594999270416e-07, + "loss": 0.3059, + "num_input_tokens_seen": 4083154944, + "step": 1947 + }, + { + "epoch": 0.2028749660970979, + "grad_norm": 17.68223762512207, + "learning_rate": 5.451558330462289e-07, + "loss": 0.6006, + "num_input_tokens_seen": 4085252096, + "step": 1948 + }, + { + "epoch": 0.20314618931380526, + "grad_norm": 15.037524223327637, + "learning_rate": 5.447521700638174e-07, + "loss": 0.4331, + "num_input_tokens_seen": 4087349248, + "step": 1949 + }, + { + "epoch": 0.20341741253051263, + "grad_norm": 13.928112030029297, + "learning_rate": 5.443485113046597e-07, + "loss": 0.4327, + "num_input_tokens_seen": 4089446400, + "step": 1950 + }, + { + "epoch": 0.20368863574721996, + "grad_norm": 11.044869422912598, + "learning_rate": 5.439448570936052e-07, + "loss": 0.2896, + "num_input_tokens_seen": 4091543552, + "step": 1951 + }, + { + "epoch": 0.20395985896392732, + "grad_norm": 14.348036766052246, + "learning_rate": 5.435412077554991e-07, + "loss": 0.3981, + "num_input_tokens_seen": 4093640704, + "step": 1952 + }, + { + "epoch": 0.20423108218063465, + "grad_norm": 14.003972053527832, + "learning_rate": 5.431375636151833e-07, + "loss": 0.3766, + "num_input_tokens_seen": 4095737856, + "step": 1953 + }, + { + "epoch": 0.20450230539734202, + "grad_norm": 18.595943450927734, + "learning_rate": 5.427339249974952e-07, + "loss": 0.7953, + "num_input_tokens_seen": 4097835008, + "step": 1954 + }, + { + "epoch": 0.20477352861404935, + "grad_norm": 12.756117820739746, + "learning_rate": 5.423302922272677e-07, + "loss": 0.4171, + "num_input_tokens_seen": 4099932160, + "step": 1955 + }, + { + "epoch": 0.2050447518307567, + "grad_norm": 17.62868881225586, + "learning_rate": 5.419266656293293e-07, + "loss": 0.5959, + "num_input_tokens_seen": 4102029312, + "step": 1956 + }, + { + "epoch": 0.20531597504746407, + "grad_norm": 14.314393997192383, + "learning_rate": 5.415230455285031e-07, + "loss": 0.4849, + "num_input_tokens_seen": 4104126464, + "step": 1957 + }, + { + "epoch": 0.2055871982641714, + "grad_norm": 11.385714530944824, + "learning_rate": 5.411194322496073e-07, + "loss": 0.341, + "num_input_tokens_seen": 4106223616, + "step": 1958 + }, + { + "epoch": 0.20585842148087877, + "grad_norm": 23.866220474243164, + "learning_rate": 5.407158261174543e-07, + "loss": 0.3017, + "num_input_tokens_seen": 4108320768, + "step": 1959 + }, + { + "epoch": 0.2061296446975861, + "grad_norm": 12.497730255126953, + "learning_rate": 5.403122274568514e-07, + "loss": 0.2759, + "num_input_tokens_seen": 4110417920, + "step": 1960 + }, + { + "epoch": 0.20640086791429346, + "grad_norm": 35.96780776977539, + "learning_rate": 5.399086365925991e-07, + "loss": 0.9278, + "num_input_tokens_seen": 4112515072, + "step": 1961 + }, + { + "epoch": 0.20667209113100082, + "grad_norm": 13.677523612976074, + "learning_rate": 5.395050538494922e-07, + "loss": 0.2937, + "num_input_tokens_seen": 4114612224, + "step": 1962 + }, + { + "epoch": 0.20694331434770816, + "grad_norm": 9.891523361206055, + "learning_rate": 5.391014795523184e-07, + "loss": 0.2558, + "num_input_tokens_seen": 4116709376, + "step": 1963 + }, + { + "epoch": 0.20721453756441552, + "grad_norm": 15.449033737182617, + "learning_rate": 5.386979140258592e-07, + "loss": 0.5304, + "num_input_tokens_seen": 4118806528, + "step": 1964 + }, + { + "epoch": 0.20748576078112285, + "grad_norm": 12.77426815032959, + "learning_rate": 5.382943575948888e-07, + "loss": 0.3884, + "num_input_tokens_seen": 4120903680, + "step": 1965 + }, + { + "epoch": 0.2077569839978302, + "grad_norm": 27.787797927856445, + "learning_rate": 5.378908105841738e-07, + "loss": 0.5596, + "num_input_tokens_seen": 4123000832, + "step": 1966 + }, + { + "epoch": 0.20802820721453757, + "grad_norm": 17.83061408996582, + "learning_rate": 5.374872733184737e-07, + "loss": 0.5323, + "num_input_tokens_seen": 4125097984, + "step": 1967 + }, + { + "epoch": 0.2082994304312449, + "grad_norm": 16.810670852661133, + "learning_rate": 5.370837461225402e-07, + "loss": 0.6982, + "num_input_tokens_seen": 4127195136, + "step": 1968 + }, + { + "epoch": 0.20857065364795227, + "grad_norm": 16.809240341186523, + "learning_rate": 5.36680229321116e-07, + "loss": 0.4564, + "num_input_tokens_seen": 4129292288, + "step": 1969 + }, + { + "epoch": 0.2088418768646596, + "grad_norm": 16.903656005859375, + "learning_rate": 5.362767232389365e-07, + "loss": 0.5688, + "num_input_tokens_seen": 4131389440, + "step": 1970 + }, + { + "epoch": 0.20911310008136696, + "grad_norm": 11.472921371459961, + "learning_rate": 5.358732282007278e-07, + "loss": 0.3912, + "num_input_tokens_seen": 4133486592, + "step": 1971 + }, + { + "epoch": 0.20938432329807433, + "grad_norm": 15.253931999206543, + "learning_rate": 5.354697445312074e-07, + "loss": 0.5295, + "num_input_tokens_seen": 4135583744, + "step": 1972 + }, + { + "epoch": 0.20965554651478166, + "grad_norm": 14.575647354125977, + "learning_rate": 5.35066272555084e-07, + "loss": 0.6172, + "num_input_tokens_seen": 4137680896, + "step": 1973 + }, + { + "epoch": 0.20992676973148902, + "grad_norm": 15.925987243652344, + "learning_rate": 5.346628125970562e-07, + "loss": 0.4215, + "num_input_tokens_seen": 4139778048, + "step": 1974 + }, + { + "epoch": 0.21019799294819635, + "grad_norm": 21.621906280517578, + "learning_rate": 5.342593649818129e-07, + "loss": 0.6736, + "num_input_tokens_seen": 4141875200, + "step": 1975 + }, + { + "epoch": 0.21046921616490372, + "grad_norm": 13.804909706115723, + "learning_rate": 5.338559300340338e-07, + "loss": 0.4555, + "num_input_tokens_seen": 4143972352, + "step": 1976 + }, + { + "epoch": 0.21074043938161108, + "grad_norm": 15.145031929016113, + "learning_rate": 5.334525080783879e-07, + "loss": 0.485, + "num_input_tokens_seen": 4146069504, + "step": 1977 + }, + { + "epoch": 0.2110116625983184, + "grad_norm": 12.25986099243164, + "learning_rate": 5.330490994395338e-07, + "loss": 0.3228, + "num_input_tokens_seen": 4148166656, + "step": 1978 + }, + { + "epoch": 0.21128288581502577, + "grad_norm": 15.601759910583496, + "learning_rate": 5.326457044421195e-07, + "loss": 0.5354, + "num_input_tokens_seen": 4150263808, + "step": 1979 + }, + { + "epoch": 0.2115541090317331, + "grad_norm": 12.419604301452637, + "learning_rate": 5.322423234107818e-07, + "loss": 0.4416, + "num_input_tokens_seen": 4152360960, + "step": 1980 + }, + { + "epoch": 0.21182533224844047, + "grad_norm": 17.01136589050293, + "learning_rate": 5.318389566701465e-07, + "loss": 0.5834, + "num_input_tokens_seen": 4154458112, + "step": 1981 + }, + { + "epoch": 0.21209655546514783, + "grad_norm": 13.978080749511719, + "learning_rate": 5.314356045448278e-07, + "loss": 0.388, + "num_input_tokens_seen": 4156555264, + "step": 1982 + }, + { + "epoch": 0.21236777868185516, + "grad_norm": 18.458215713500977, + "learning_rate": 5.310322673594282e-07, + "loss": 0.4498, + "num_input_tokens_seen": 4158652416, + "step": 1983 + }, + { + "epoch": 0.21263900189856252, + "grad_norm": 13.268664360046387, + "learning_rate": 5.306289454385379e-07, + "loss": 0.2379, + "num_input_tokens_seen": 4160749568, + "step": 1984 + }, + { + "epoch": 0.21291022511526986, + "grad_norm": 14.612197875976562, + "learning_rate": 5.302256391067354e-07, + "loss": 0.5217, + "num_input_tokens_seen": 4162846720, + "step": 1985 + }, + { + "epoch": 0.21318144833197722, + "grad_norm": 14.71393871307373, + "learning_rate": 5.29822348688586e-07, + "loss": 0.4559, + "num_input_tokens_seen": 4164943872, + "step": 1986 + }, + { + "epoch": 0.21345267154868458, + "grad_norm": 18.54011344909668, + "learning_rate": 5.294190745086426e-07, + "loss": 0.5078, + "num_input_tokens_seen": 4167041024, + "step": 1987 + }, + { + "epoch": 0.21372389476539191, + "grad_norm": 22.229833602905273, + "learning_rate": 5.290158168914447e-07, + "loss": 0.6627, + "num_input_tokens_seen": 4169138176, + "step": 1988 + }, + { + "epoch": 0.21399511798209928, + "grad_norm": 15.733107566833496, + "learning_rate": 5.28612576161519e-07, + "loss": 0.6356, + "num_input_tokens_seen": 4171235328, + "step": 1989 + }, + { + "epoch": 0.2142663411988066, + "grad_norm": 16.396329879760742, + "learning_rate": 5.282093526433781e-07, + "loss": 0.5705, + "num_input_tokens_seen": 4173332480, + "step": 1990 + }, + { + "epoch": 0.21453756441551397, + "grad_norm": 10.583121299743652, + "learning_rate": 5.27806146661521e-07, + "loss": 0.2994, + "num_input_tokens_seen": 4175429632, + "step": 1991 + }, + { + "epoch": 0.21480878763222133, + "grad_norm": 14.593506813049316, + "learning_rate": 5.274029585404326e-07, + "loss": 0.3533, + "num_input_tokens_seen": 4177526784, + "step": 1992 + }, + { + "epoch": 0.21508001084892867, + "grad_norm": 12.540976524353027, + "learning_rate": 5.269997886045833e-07, + "loss": 0.4252, + "num_input_tokens_seen": 4179623936, + "step": 1993 + }, + { + "epoch": 0.21535123406563603, + "grad_norm": 15.491146087646484, + "learning_rate": 5.265966371784287e-07, + "loss": 0.5759, + "num_input_tokens_seen": 4181721088, + "step": 1994 + }, + { + "epoch": 0.21562245728234336, + "grad_norm": 13.454513549804688, + "learning_rate": 5.261935045864101e-07, + "loss": 0.3348, + "num_input_tokens_seen": 4183818240, + "step": 1995 + }, + { + "epoch": 0.21589368049905072, + "grad_norm": 13.391287803649902, + "learning_rate": 5.257903911529532e-07, + "loss": 0.416, + "num_input_tokens_seen": 4185915392, + "step": 1996 + }, + { + "epoch": 0.21616490371575806, + "grad_norm": 11.039448738098145, + "learning_rate": 5.253872972024681e-07, + "loss": 0.2622, + "num_input_tokens_seen": 4188012544, + "step": 1997 + }, + { + "epoch": 0.21643612693246542, + "grad_norm": 12.495340347290039, + "learning_rate": 5.249842230593497e-07, + "loss": 0.3805, + "num_input_tokens_seen": 4190109696, + "step": 1998 + }, + { + "epoch": 0.21670735014917278, + "grad_norm": 18.3435001373291, + "learning_rate": 5.245811690479766e-07, + "loss": 0.4074, + "num_input_tokens_seen": 4192206848, + "step": 1999 + }, + { + "epoch": 0.2169785733658801, + "grad_norm": 13.659356117248535, + "learning_rate": 5.241781354927113e-07, + "loss": 0.4192, + "num_input_tokens_seen": 4194304000, + "step": 2000 + }, + { + "epoch": 0.00027122321670735016, + "grad_norm": 23.389102935791016, + "learning_rate": 5.237751227179001e-07, + "loss": 0.5657, + "num_input_tokens_seen": 4196401152, + "step": 2001 + }, + { + "epoch": 0.0005424464334147003, + "grad_norm": 17.303232192993164, + "learning_rate": 5.233721310478722e-07, + "loss": 0.5492, + "num_input_tokens_seen": 4198498304, + "step": 2002 + }, + { + "epoch": 0.0008136696501220504, + "grad_norm": 9.405117988586426, + "learning_rate": 5.2296916080694e-07, + "loss": 0.2159, + "num_input_tokens_seen": 4200595456, + "step": 2003 + }, + { + "epoch": 0.0010848928668294006, + "grad_norm": 17.866546630859375, + "learning_rate": 5.225662123193985e-07, + "loss": 0.4617, + "num_input_tokens_seen": 4202692608, + "step": 2004 + }, + { + "epoch": 0.0013561160835367507, + "grad_norm": 17.711912155151367, + "learning_rate": 5.221632859095254e-07, + "loss": 0.375, + "num_input_tokens_seen": 4204789760, + "step": 2005 + }, + { + "epoch": 0.0016273393002441008, + "grad_norm": 17.354616165161133, + "learning_rate": 5.217603819015805e-07, + "loss": 0.5621, + "num_input_tokens_seen": 4206886912, + "step": 2006 + }, + { + "epoch": 0.001898562516951451, + "grad_norm": 18.786190032958984, + "learning_rate": 5.213575006198057e-07, + "loss": 0.7147, + "num_input_tokens_seen": 4208984064, + "step": 2007 + }, + { + "epoch": 0.0021697857336588013, + "grad_norm": 16.878746032714844, + "learning_rate": 5.209546423884246e-07, + "loss": 0.5038, + "num_input_tokens_seen": 4211081216, + "step": 2008 + }, + { + "epoch": 0.0024410089503661514, + "grad_norm": 13.854147911071777, + "learning_rate": 5.20551807531642e-07, + "loss": 0.3361, + "num_input_tokens_seen": 4213178368, + "step": 2009 + }, + { + "epoch": 0.0027122321670735015, + "grad_norm": 16.524831771850586, + "learning_rate": 5.20148996373644e-07, + "loss": 0.5854, + "num_input_tokens_seen": 4215275520, + "step": 2010 + }, + { + "epoch": 0.0029834553837808516, + "grad_norm": 11.64826488494873, + "learning_rate": 5.197462092385978e-07, + "loss": 0.3409, + "num_input_tokens_seen": 4217372672, + "step": 2011 + }, + { + "epoch": 0.0032546786004882017, + "grad_norm": 9.127813339233398, + "learning_rate": 5.193434464506511e-07, + "loss": 0.2305, + "num_input_tokens_seen": 4219469824, + "step": 2012 + }, + { + "epoch": 0.003525901817195552, + "grad_norm": 15.029864311218262, + "learning_rate": 5.189407083339324e-07, + "loss": 0.4268, + "num_input_tokens_seen": 4221566976, + "step": 2013 + }, + { + "epoch": 0.003797125033902902, + "grad_norm": 9.265511512756348, + "learning_rate": 5.185379952125494e-07, + "loss": 0.2192, + "num_input_tokens_seen": 4223664128, + "step": 2014 + }, + { + "epoch": 0.0040683482506102524, + "grad_norm": 15.287160873413086, + "learning_rate": 5.18135307410591e-07, + "loss": 0.4158, + "num_input_tokens_seen": 4225761280, + "step": 2015 + }, + { + "epoch": 0.0043395714673176026, + "grad_norm": 20.016700744628906, + "learning_rate": 5.177326452521242e-07, + "loss": 0.3771, + "num_input_tokens_seen": 4227858432, + "step": 2016 + }, + { + "epoch": 0.004610794684024953, + "grad_norm": 19.859949111938477, + "learning_rate": 5.17330009061197e-07, + "loss": 0.467, + "num_input_tokens_seen": 4229955584, + "step": 2017 + }, + { + "epoch": 0.004882017900732303, + "grad_norm": 16.03641700744629, + "learning_rate": 5.169273991618351e-07, + "loss": 0.43, + "num_input_tokens_seen": 4232052736, + "step": 2018 + }, + { + "epoch": 0.005153241117439653, + "grad_norm": 19.55535316467285, + "learning_rate": 5.165248158780441e-07, + "loss": 0.4449, + "num_input_tokens_seen": 4234149888, + "step": 2019 + }, + { + "epoch": 0.005424464334147003, + "grad_norm": 19.658199310302734, + "learning_rate": 5.161222595338073e-07, + "loss": 0.842, + "num_input_tokens_seen": 4236247040, + "step": 2020 + }, + { + "epoch": 0.005695687550854353, + "grad_norm": 14.803008079528809, + "learning_rate": 5.157197304530869e-07, + "loss": 0.4334, + "num_input_tokens_seen": 4238344192, + "step": 2021 + }, + { + "epoch": 0.005966910767561703, + "grad_norm": 16.178064346313477, + "learning_rate": 5.15317228959823e-07, + "loss": 0.4884, + "num_input_tokens_seen": 4240441344, + "step": 2022 + }, + { + "epoch": 0.006238133984269053, + "grad_norm": 20.011199951171875, + "learning_rate": 5.149147553779335e-07, + "loss": 0.7486, + "num_input_tokens_seen": 4242538496, + "step": 2023 + }, + { + "epoch": 0.006509357200976403, + "grad_norm": 12.02757740020752, + "learning_rate": 5.145123100313137e-07, + "loss": 0.3126, + "num_input_tokens_seen": 4244635648, + "step": 2024 + }, + { + "epoch": 0.0067805804176837535, + "grad_norm": 18.81563377380371, + "learning_rate": 5.141098932438365e-07, + "loss": 0.4878, + "num_input_tokens_seen": 4246732800, + "step": 2025 + }, + { + "epoch": 0.007051803634391104, + "grad_norm": 19.94472312927246, + "learning_rate": 5.137075053393512e-07, + "loss": 0.5076, + "num_input_tokens_seen": 4248829952, + "step": 2026 + }, + { + "epoch": 0.007323026851098454, + "grad_norm": 23.192970275878906, + "learning_rate": 5.133051466416849e-07, + "loss": 0.7827, + "num_input_tokens_seen": 4250927104, + "step": 2027 + }, + { + "epoch": 0.007594250067805804, + "grad_norm": 14.543439865112305, + "learning_rate": 5.129028174746399e-07, + "loss": 0.3696, + "num_input_tokens_seen": 4253024256, + "step": 2028 + }, + { + "epoch": 0.007865473284513154, + "grad_norm": 21.16613006591797, + "learning_rate": 5.12500518161996e-07, + "loss": 1.0285, + "num_input_tokens_seen": 4255121408, + "step": 2029 + }, + { + "epoch": 0.008136696501220505, + "grad_norm": 15.31861686706543, + "learning_rate": 5.120982490275081e-07, + "loss": 0.468, + "num_input_tokens_seen": 4257218560, + "step": 2030 + }, + { + "epoch": 0.008407919717927854, + "grad_norm": 17.255081176757812, + "learning_rate": 5.116960103949071e-07, + "loss": 0.7309, + "num_input_tokens_seen": 4259315712, + "step": 2031 + }, + { + "epoch": 0.008679142934635205, + "grad_norm": 15.113636016845703, + "learning_rate": 5.112938025878994e-07, + "loss": 0.3611, + "num_input_tokens_seen": 4261412864, + "step": 2032 + }, + { + "epoch": 0.008950366151342554, + "grad_norm": 9.508875846862793, + "learning_rate": 5.108916259301663e-07, + "loss": 0.2398, + "num_input_tokens_seen": 4263510016, + "step": 2033 + }, + { + "epoch": 0.009221589368049905, + "grad_norm": 10.597322463989258, + "learning_rate": 5.104894807453647e-07, + "loss": 0.3246, + "num_input_tokens_seen": 4265607168, + "step": 2034 + }, + { + "epoch": 0.009492812584757255, + "grad_norm": 17.956418991088867, + "learning_rate": 5.100873673571253e-07, + "loss": 0.5213, + "num_input_tokens_seen": 4267704320, + "step": 2035 + }, + { + "epoch": 0.009764035801464606, + "grad_norm": 16.9884090423584, + "learning_rate": 5.096852860890541e-07, + "loss": 0.3541, + "num_input_tokens_seen": 4269801472, + "step": 2036 + }, + { + "epoch": 0.010035259018171955, + "grad_norm": 22.198923110961914, + "learning_rate": 5.092832372647304e-07, + "loss": 0.5872, + "num_input_tokens_seen": 4271898624, + "step": 2037 + }, + { + "epoch": 0.010306482234879306, + "grad_norm": 16.18609046936035, + "learning_rate": 5.08881221207708e-07, + "loss": 0.5318, + "num_input_tokens_seen": 4273995776, + "step": 2038 + }, + { + "epoch": 0.010577705451586655, + "grad_norm": 17.77985191345215, + "learning_rate": 5.084792382415141e-07, + "loss": 0.2986, + "num_input_tokens_seen": 4276092928, + "step": 2039 + }, + { + "epoch": 0.010848928668294006, + "grad_norm": 17.145601272583008, + "learning_rate": 5.080772886896493e-07, + "loss": 0.4883, + "num_input_tokens_seen": 4278190080, + "step": 2040 + }, + { + "epoch": 0.011120151885001357, + "grad_norm": 13.083086013793945, + "learning_rate": 5.076753728755871e-07, + "loss": 0.3112, + "num_input_tokens_seen": 4280287232, + "step": 2041 + }, + { + "epoch": 0.011391375101708706, + "grad_norm": 17.41132354736328, + "learning_rate": 5.072734911227746e-07, + "loss": 0.4163, + "num_input_tokens_seen": 4282384384, + "step": 2042 + }, + { + "epoch": 0.011662598318416057, + "grad_norm": 8.932369232177734, + "learning_rate": 5.068716437546305e-07, + "loss": 0.2173, + "num_input_tokens_seen": 4284481536, + "step": 2043 + }, + { + "epoch": 0.011933821535123406, + "grad_norm": 16.91132164001465, + "learning_rate": 5.06469831094546e-07, + "loss": 0.6219, + "num_input_tokens_seen": 4286578688, + "step": 2044 + }, + { + "epoch": 0.012205044751830757, + "grad_norm": 15.44660758972168, + "learning_rate": 5.060680534658852e-07, + "loss": 0.4811, + "num_input_tokens_seen": 4288675840, + "step": 2045 + }, + { + "epoch": 0.012476267968538107, + "grad_norm": 11.51047134399414, + "learning_rate": 5.056663111919833e-07, + "loss": 0.4082, + "num_input_tokens_seen": 4290772992, + "step": 2046 + }, + { + "epoch": 0.012747491185245458, + "grad_norm": 9.88651180267334, + "learning_rate": 5.052646045961471e-07, + "loss": 0.2301, + "num_input_tokens_seen": 4292870144, + "step": 2047 + }, + { + "epoch": 0.013018714401952807, + "grad_norm": 14.237520217895508, + "learning_rate": 5.048629340016548e-07, + "loss": 0.4273, + "num_input_tokens_seen": 4294967296, + "step": 2048 + }, + { + "epoch": 0.013289937618660158, + "grad_norm": 14.176645278930664, + "learning_rate": 5.044612997317557e-07, + "loss": 0.3895, + "num_input_tokens_seen": 4297064448, + "step": 2049 + }, + { + "epoch": 0.013561160835367507, + "grad_norm": 18.345462799072266, + "learning_rate": 5.040597021096695e-07, + "loss": 0.7524, + "num_input_tokens_seen": 4299161600, + "step": 2050 + }, + { + "epoch": 0.013832384052074858, + "grad_norm": 15.396586418151855, + "learning_rate": 5.03658141458587e-07, + "loss": 0.4844, + "num_input_tokens_seen": 4301258752, + "step": 2051 + }, + { + "epoch": 0.014103607268782207, + "grad_norm": 17.37771224975586, + "learning_rate": 5.032566181016688e-07, + "loss": 0.381, + "num_input_tokens_seen": 4303355904, + "step": 2052 + }, + { + "epoch": 0.014374830485489558, + "grad_norm": 15.787535667419434, + "learning_rate": 5.028551323620458e-07, + "loss": 0.6099, + "num_input_tokens_seen": 4305453056, + "step": 2053 + }, + { + "epoch": 0.014646053702196907, + "grad_norm": 14.97440242767334, + "learning_rate": 5.024536845628181e-07, + "loss": 0.4472, + "num_input_tokens_seen": 4307550208, + "step": 2054 + }, + { + "epoch": 0.014917276918904258, + "grad_norm": 26.290592193603516, + "learning_rate": 5.020522750270558e-07, + "loss": 0.7158, + "num_input_tokens_seen": 4309647360, + "step": 2055 + }, + { + "epoch": 0.015188500135611608, + "grad_norm": 14.337723731994629, + "learning_rate": 5.01650904077798e-07, + "loss": 0.3129, + "num_input_tokens_seen": 4311744512, + "step": 2056 + }, + { + "epoch": 0.015459723352318959, + "grad_norm": 20.845989227294922, + "learning_rate": 5.012495720380527e-07, + "loss": 0.7573, + "num_input_tokens_seen": 4313841664, + "step": 2057 + }, + { + "epoch": 0.015730946569026308, + "grad_norm": 13.03821086883545, + "learning_rate": 5.008482792307968e-07, + "loss": 0.3595, + "num_input_tokens_seen": 4315938816, + "step": 2058 + }, + { + "epoch": 0.01600216978573366, + "grad_norm": 11.9432954788208, + "learning_rate": 5.004470259789757e-07, + "loss": 0.3451, + "num_input_tokens_seen": 4318035968, + "step": 2059 + }, + { + "epoch": 0.01627339300244101, + "grad_norm": 11.00033950805664, + "learning_rate": 5.000458126055021e-07, + "loss": 0.3564, + "num_input_tokens_seen": 4320133120, + "step": 2060 + }, + { + "epoch": 0.01654461621914836, + "grad_norm": 18.752838134765625, + "learning_rate": 4.996446394332578e-07, + "loss": 0.5661, + "num_input_tokens_seen": 4322230272, + "step": 2061 + }, + { + "epoch": 0.016815839435855708, + "grad_norm": 12.34859848022461, + "learning_rate": 4.992435067850914e-07, + "loss": 0.3662, + "num_input_tokens_seen": 4324327424, + "step": 2062 + }, + { + "epoch": 0.01708706265256306, + "grad_norm": 17.15289306640625, + "learning_rate": 4.988424149838192e-07, + "loss": 0.508, + "num_input_tokens_seen": 4326424576, + "step": 2063 + }, + { + "epoch": 0.01735828586927041, + "grad_norm": 31.281986236572266, + "learning_rate": 4.984413643522251e-07, + "loss": 0.8106, + "num_input_tokens_seen": 4328521728, + "step": 2064 + }, + { + "epoch": 0.01762950908597776, + "grad_norm": 12.578858375549316, + "learning_rate": 4.980403552130589e-07, + "loss": 0.3856, + "num_input_tokens_seen": 4330618880, + "step": 2065 + }, + { + "epoch": 0.01790073230268511, + "grad_norm": 14.403166770935059, + "learning_rate": 4.976393878890375e-07, + "loss": 0.4615, + "num_input_tokens_seen": 4332716032, + "step": 2066 + }, + { + "epoch": 0.01817195551939246, + "grad_norm": 14.294715881347656, + "learning_rate": 4.972384627028442e-07, + "loss": 0.4769, + "num_input_tokens_seen": 4334813184, + "step": 2067 + }, + { + "epoch": 0.01844317873609981, + "grad_norm": 16.856159210205078, + "learning_rate": 4.968375799771285e-07, + "loss": 0.3869, + "num_input_tokens_seen": 4336910336, + "step": 2068 + }, + { + "epoch": 0.01871440195280716, + "grad_norm": 11.613068580627441, + "learning_rate": 4.964367400345053e-07, + "loss": 0.3629, + "num_input_tokens_seen": 4339007488, + "step": 2069 + }, + { + "epoch": 0.01898562516951451, + "grad_norm": 17.622570037841797, + "learning_rate": 4.960359431975555e-07, + "loss": 0.738, + "num_input_tokens_seen": 4341104640, + "step": 2070 + }, + { + "epoch": 0.01925684838622186, + "grad_norm": 21.49909782409668, + "learning_rate": 4.956351897888256e-07, + "loss": 0.7802, + "num_input_tokens_seen": 4343201792, + "step": 2071 + }, + { + "epoch": 0.01952807160292921, + "grad_norm": 17.356836318969727, + "learning_rate": 4.952344801308258e-07, + "loss": 0.6596, + "num_input_tokens_seen": 4345298944, + "step": 2072 + }, + { + "epoch": 0.019799294819636562, + "grad_norm": 12.582341194152832, + "learning_rate": 4.948338145460324e-07, + "loss": 0.373, + "num_input_tokens_seen": 4347396096, + "step": 2073 + }, + { + "epoch": 0.02007051803634391, + "grad_norm": 18.529993057250977, + "learning_rate": 4.944331933568858e-07, + "loss": 0.7155, + "num_input_tokens_seen": 4349493248, + "step": 2074 + }, + { + "epoch": 0.02034174125305126, + "grad_norm": 16.39361572265625, + "learning_rate": 4.94032616885791e-07, + "loss": 0.6209, + "num_input_tokens_seen": 4351590400, + "step": 2075 + }, + { + "epoch": 0.02061296446975861, + "grad_norm": 12.521406173706055, + "learning_rate": 4.936320854551163e-07, + "loss": 0.4043, + "num_input_tokens_seen": 4353687552, + "step": 2076 + }, + { + "epoch": 0.020884187686465962, + "grad_norm": 9.356000900268555, + "learning_rate": 4.932315993871942e-07, + "loss": 0.2696, + "num_input_tokens_seen": 4355784704, + "step": 2077 + }, + { + "epoch": 0.02115541090317331, + "grad_norm": 13.908049583435059, + "learning_rate": 4.928311590043206e-07, + "loss": 0.3287, + "num_input_tokens_seen": 4357881856, + "step": 2078 + }, + { + "epoch": 0.02142663411988066, + "grad_norm": 17.52366065979004, + "learning_rate": 4.924307646287549e-07, + "loss": 0.7094, + "num_input_tokens_seen": 4359979008, + "step": 2079 + }, + { + "epoch": 0.021697857336588012, + "grad_norm": 21.215076446533203, + "learning_rate": 4.920304165827192e-07, + "loss": 0.7868, + "num_input_tokens_seen": 4362076160, + "step": 2080 + }, + { + "epoch": 0.021969080553295363, + "grad_norm": 24.048057556152344, + "learning_rate": 4.916301151883983e-07, + "loss": 0.4977, + "num_input_tokens_seen": 4364173312, + "step": 2081 + }, + { + "epoch": 0.022240303770002714, + "grad_norm": 18.432392120361328, + "learning_rate": 4.912298607679396e-07, + "loss": 0.6461, + "num_input_tokens_seen": 4366270464, + "step": 2082 + }, + { + "epoch": 0.02251152698671006, + "grad_norm": 16.7590389251709, + "learning_rate": 4.908296536434525e-07, + "loss": 0.4585, + "num_input_tokens_seen": 4368367616, + "step": 2083 + }, + { + "epoch": 0.022782750203417412, + "grad_norm": 21.218637466430664, + "learning_rate": 4.904294941370086e-07, + "loss": 0.7689, + "num_input_tokens_seen": 4370464768, + "step": 2084 + }, + { + "epoch": 0.023053973420124763, + "grad_norm": 13.70450210571289, + "learning_rate": 4.900293825706408e-07, + "loss": 0.3011, + "num_input_tokens_seen": 4372561920, + "step": 2085 + }, + { + "epoch": 0.023325196636832114, + "grad_norm": 12.731541633605957, + "learning_rate": 4.89629319266344e-07, + "loss": 0.444, + "num_input_tokens_seen": 4374659072, + "step": 2086 + }, + { + "epoch": 0.023596419853539462, + "grad_norm": 13.77962589263916, + "learning_rate": 4.892293045460738e-07, + "loss": 0.4021, + "num_input_tokens_seen": 4376756224, + "step": 2087 + }, + { + "epoch": 0.023867643070246813, + "grad_norm": 14.275106430053711, + "learning_rate": 4.88829338731747e-07, + "loss": 0.3847, + "num_input_tokens_seen": 4378853376, + "step": 2088 + }, + { + "epoch": 0.024138866286954164, + "grad_norm": 11.474467277526855, + "learning_rate": 4.884294221452405e-07, + "loss": 0.3539, + "num_input_tokens_seen": 4380950528, + "step": 2089 + }, + { + "epoch": 0.024410089503661515, + "grad_norm": 17.160314559936523, + "learning_rate": 4.88029555108392e-07, + "loss": 0.6452, + "num_input_tokens_seen": 4383047680, + "step": 2090 + }, + { + "epoch": 0.024681312720368862, + "grad_norm": 10.082201957702637, + "learning_rate": 4.876297379429995e-07, + "loss": 0.2754, + "num_input_tokens_seen": 4385144832, + "step": 2091 + }, + { + "epoch": 0.024952535937076213, + "grad_norm": 16.620037078857422, + "learning_rate": 4.872299709708206e-07, + "loss": 0.4956, + "num_input_tokens_seen": 4387241984, + "step": 2092 + }, + { + "epoch": 0.025223759153783564, + "grad_norm": 13.162679672241211, + "learning_rate": 4.868302545135725e-07, + "loss": 0.3437, + "num_input_tokens_seen": 4389339136, + "step": 2093 + }, + { + "epoch": 0.025494982370490915, + "grad_norm": 19.05023765563965, + "learning_rate": 4.864305888929318e-07, + "loss": 0.9092, + "num_input_tokens_seen": 4391436288, + "step": 2094 + }, + { + "epoch": 0.025766205587198263, + "grad_norm": 19.761568069458008, + "learning_rate": 4.860309744305339e-07, + "loss": 0.8281, + "num_input_tokens_seen": 4393533440, + "step": 2095 + }, + { + "epoch": 0.026037428803905614, + "grad_norm": 18.267724990844727, + "learning_rate": 4.856314114479738e-07, + "loss": 0.7678, + "num_input_tokens_seen": 4395630592, + "step": 2096 + }, + { + "epoch": 0.026308652020612965, + "grad_norm": 12.300085067749023, + "learning_rate": 4.852319002668044e-07, + "loss": 0.3752, + "num_input_tokens_seen": 4397727744, + "step": 2097 + }, + { + "epoch": 0.026579875237320316, + "grad_norm": 15.5980806350708, + "learning_rate": 4.848324412085367e-07, + "loss": 0.5343, + "num_input_tokens_seen": 4399824896, + "step": 2098 + }, + { + "epoch": 0.026851098454027666, + "grad_norm": 17.482070922851562, + "learning_rate": 4.84433034594641e-07, + "loss": 0.5354, + "num_input_tokens_seen": 4401922048, + "step": 2099 + }, + { + "epoch": 0.027122321670735014, + "grad_norm": 14.687858581542969, + "learning_rate": 4.840336807465439e-07, + "loss": 0.4802, + "num_input_tokens_seen": 4404019200, + "step": 2100 + }, + { + "epoch": 0.027393544887442365, + "grad_norm": 13.101119995117188, + "learning_rate": 4.836343799856302e-07, + "loss": 0.4752, + "num_input_tokens_seen": 4406116352, + "step": 2101 + }, + { + "epoch": 0.027664768104149716, + "grad_norm": 8.815567970275879, + "learning_rate": 4.83235132633242e-07, + "loss": 0.2088, + "num_input_tokens_seen": 4408213504, + "step": 2102 + }, + { + "epoch": 0.027935991320857067, + "grad_norm": 9.5147066116333, + "learning_rate": 4.828359390106786e-07, + "loss": 0.2341, + "num_input_tokens_seen": 4410310656, + "step": 2103 + }, + { + "epoch": 0.028207214537564414, + "grad_norm": 12.234147071838379, + "learning_rate": 4.824367994391958e-07, + "loss": 0.4088, + "num_input_tokens_seen": 4412407808, + "step": 2104 + }, + { + "epoch": 0.028478437754271765, + "grad_norm": 11.71864128112793, + "learning_rate": 4.820377142400058e-07, + "loss": 0.2418, + "num_input_tokens_seen": 4414504960, + "step": 2105 + }, + { + "epoch": 0.028749660970979116, + "grad_norm": 22.507688522338867, + "learning_rate": 4.816386837342771e-07, + "loss": 0.5335, + "num_input_tokens_seen": 4416602112, + "step": 2106 + }, + { + "epoch": 0.029020884187686467, + "grad_norm": 15.12206745147705, + "learning_rate": 4.812397082431345e-07, + "loss": 0.5299, + "num_input_tokens_seen": 4418699264, + "step": 2107 + }, + { + "epoch": 0.029292107404393815, + "grad_norm": 23.719499588012695, + "learning_rate": 4.808407880876581e-07, + "loss": 0.5769, + "num_input_tokens_seen": 4420796416, + "step": 2108 + }, + { + "epoch": 0.029563330621101166, + "grad_norm": 15.224874496459961, + "learning_rate": 4.804419235888838e-07, + "loss": 0.5696, + "num_input_tokens_seen": 4422893568, + "step": 2109 + }, + { + "epoch": 0.029834553837808517, + "grad_norm": 15.45009994506836, + "learning_rate": 4.800431150678026e-07, + "loss": 0.4187, + "num_input_tokens_seen": 4424990720, + "step": 2110 + }, + { + "epoch": 0.030105777054515868, + "grad_norm": 18.44724464416504, + "learning_rate": 4.796443628453603e-07, + "loss": 0.6679, + "num_input_tokens_seen": 4427087872, + "step": 2111 + }, + { + "epoch": 0.030377000271223215, + "grad_norm": 15.730889320373535, + "learning_rate": 4.792456672424574e-07, + "loss": 0.4933, + "num_input_tokens_seen": 4429185024, + "step": 2112 + }, + { + "epoch": 0.030648223487930566, + "grad_norm": 10.117815971374512, + "learning_rate": 4.788470285799492e-07, + "loss": 0.2543, + "num_input_tokens_seen": 4431282176, + "step": 2113 + }, + { + "epoch": 0.030919446704637917, + "grad_norm": 14.274503707885742, + "learning_rate": 4.784484471786447e-07, + "loss": 0.4122, + "num_input_tokens_seen": 4433379328, + "step": 2114 + }, + { + "epoch": 0.031190669921345268, + "grad_norm": 17.943662643432617, + "learning_rate": 4.780499233593073e-07, + "loss": 0.5615, + "num_input_tokens_seen": 4435476480, + "step": 2115 + }, + { + "epoch": 0.031461893138052616, + "grad_norm": 12.883108139038086, + "learning_rate": 4.776514574426538e-07, + "loss": 0.3805, + "num_input_tokens_seen": 4437573632, + "step": 2116 + }, + { + "epoch": 0.03173311635475997, + "grad_norm": 11.446172714233398, + "learning_rate": 4.772530497493541e-07, + "loss": 0.3671, + "num_input_tokens_seen": 4439670784, + "step": 2117 + }, + { + "epoch": 0.03200433957146732, + "grad_norm": 15.992230415344238, + "learning_rate": 4.768547006000317e-07, + "loss": 0.5018, + "num_input_tokens_seen": 4441767936, + "step": 2118 + }, + { + "epoch": 0.032275562788174665, + "grad_norm": 15.411177635192871, + "learning_rate": 4.76456410315263e-07, + "loss": 0.5656, + "num_input_tokens_seen": 4443865088, + "step": 2119 + }, + { + "epoch": 0.03254678600488202, + "grad_norm": 13.412214279174805, + "learning_rate": 4.7605817921557666e-07, + "loss": 0.3278, + "num_input_tokens_seen": 4445962240, + "step": 2120 + }, + { + "epoch": 0.03281800922158937, + "grad_norm": 10.45663833618164, + "learning_rate": 4.756600076214541e-07, + "loss": 0.3316, + "num_input_tokens_seen": 4448059392, + "step": 2121 + }, + { + "epoch": 0.03308923243829672, + "grad_norm": 17.893962860107422, + "learning_rate": 4.7526189585332855e-07, + "loss": 0.5041, + "num_input_tokens_seen": 4450156544, + "step": 2122 + }, + { + "epoch": 0.03336045565500407, + "grad_norm": 18.3114013671875, + "learning_rate": 4.748638442315851e-07, + "loss": 0.7515, + "num_input_tokens_seen": 4452253696, + "step": 2123 + }, + { + "epoch": 0.033631678871711417, + "grad_norm": 19.698787689208984, + "learning_rate": 4.7446585307656074e-07, + "loss": 0.425, + "num_input_tokens_seen": 4454350848, + "step": 2124 + }, + { + "epoch": 0.03390290208841877, + "grad_norm": 15.37483024597168, + "learning_rate": 4.740679227085436e-07, + "loss": 0.6293, + "num_input_tokens_seen": 4456448000, + "step": 2125 + }, + { + "epoch": 0.03417412530512612, + "grad_norm": 36.20682907104492, + "learning_rate": 4.7367005344777255e-07, + "loss": 0.5484, + "num_input_tokens_seen": 4458545152, + "step": 2126 + }, + { + "epoch": 0.034445348521833466, + "grad_norm": 14.473278045654297, + "learning_rate": 4.732722456144381e-07, + "loss": 0.4536, + "num_input_tokens_seen": 4460642304, + "step": 2127 + }, + { + "epoch": 0.03471657173854082, + "grad_norm": 13.409770965576172, + "learning_rate": 4.7287449952868084e-07, + "loss": 0.445, + "num_input_tokens_seen": 4462739456, + "step": 2128 + }, + { + "epoch": 0.03498779495524817, + "grad_norm": 11.895949363708496, + "learning_rate": 4.724768155105914e-07, + "loss": 0.2948, + "num_input_tokens_seen": 4464836608, + "step": 2129 + }, + { + "epoch": 0.03525901817195552, + "grad_norm": 18.184831619262695, + "learning_rate": 4.7207919388021076e-07, + "loss": 0.6784, + "num_input_tokens_seen": 4466933760, + "step": 2130 + }, + { + "epoch": 0.03553024138866287, + "grad_norm": 15.693188667297363, + "learning_rate": 4.7168163495752977e-07, + "loss": 0.4952, + "num_input_tokens_seen": 4469030912, + "step": 2131 + }, + { + "epoch": 0.03580146460537022, + "grad_norm": 14.902560234069824, + "learning_rate": 4.7128413906248885e-07, + "loss": 0.5021, + "num_input_tokens_seen": 4471128064, + "step": 2132 + }, + { + "epoch": 0.03607268782207757, + "grad_norm": 13.470413208007812, + "learning_rate": 4.708867065149775e-07, + "loss": 0.4083, + "num_input_tokens_seen": 4473225216, + "step": 2133 + }, + { + "epoch": 0.03634391103878492, + "grad_norm": 18.67334747314453, + "learning_rate": 4.704893376348342e-07, + "loss": 0.6536, + "num_input_tokens_seen": 4475322368, + "step": 2134 + }, + { + "epoch": 0.03661513425549227, + "grad_norm": 16.05646324157715, + "learning_rate": 4.700920327418463e-07, + "loss": 0.6247, + "num_input_tokens_seen": 4477419520, + "step": 2135 + }, + { + "epoch": 0.03688635747219962, + "grad_norm": 12.6869535446167, + "learning_rate": 4.696947921557498e-07, + "loss": 0.3876, + "num_input_tokens_seen": 4479516672, + "step": 2136 + }, + { + "epoch": 0.03715758068890697, + "grad_norm": 13.598705291748047, + "learning_rate": 4.6929761619622866e-07, + "loss": 0.4517, + "num_input_tokens_seen": 4481613824, + "step": 2137 + }, + { + "epoch": 0.03742880390561432, + "grad_norm": 12.73837661743164, + "learning_rate": 4.68900505182915e-07, + "loss": 0.3866, + "num_input_tokens_seen": 4483710976, + "step": 2138 + }, + { + "epoch": 0.03770002712232167, + "grad_norm": 10.884099006652832, + "learning_rate": 4.6850345943538896e-07, + "loss": 0.2201, + "num_input_tokens_seen": 4485808128, + "step": 2139 + }, + { + "epoch": 0.03797125033902902, + "grad_norm": 18.487733840942383, + "learning_rate": 4.6810647927317735e-07, + "loss": 0.5549, + "num_input_tokens_seen": 4487905280, + "step": 2140 + }, + { + "epoch": 0.03824247355573637, + "grad_norm": 22.46426010131836, + "learning_rate": 4.677095650157551e-07, + "loss": 0.4995, + "num_input_tokens_seen": 4490002432, + "step": 2141 + }, + { + "epoch": 0.03851369677244372, + "grad_norm": 16.194440841674805, + "learning_rate": 4.6731271698254326e-07, + "loss": 0.5513, + "num_input_tokens_seen": 4492099584, + "step": 2142 + }, + { + "epoch": 0.038784919989151075, + "grad_norm": 20.5705623626709, + "learning_rate": 4.669159354929105e-07, + "loss": 0.9542, + "num_input_tokens_seen": 4494196736, + "step": 2143 + }, + { + "epoch": 0.03905614320585842, + "grad_norm": 9.620034217834473, + "learning_rate": 4.6651922086617134e-07, + "loss": 0.2924, + "num_input_tokens_seen": 4496293888, + "step": 2144 + }, + { + "epoch": 0.03932736642256577, + "grad_norm": 15.673210144042969, + "learning_rate": 4.661225734215867e-07, + "loss": 0.4557, + "num_input_tokens_seen": 4498391040, + "step": 2145 + }, + { + "epoch": 0.039598589639273124, + "grad_norm": 10.837203025817871, + "learning_rate": 4.657259934783628e-07, + "loss": 0.3044, + "num_input_tokens_seen": 4500488192, + "step": 2146 + }, + { + "epoch": 0.03986981285598047, + "grad_norm": 15.021329879760742, + "learning_rate": 4.6532948135565264e-07, + "loss": 0.5526, + "num_input_tokens_seen": 4502585344, + "step": 2147 + }, + { + "epoch": 0.04014103607268782, + "grad_norm": 15.685761451721191, + "learning_rate": 4.6493303737255397e-07, + "loss": 0.5141, + "num_input_tokens_seen": 4504682496, + "step": 2148 + }, + { + "epoch": 0.040412259289395173, + "grad_norm": 17.80036163330078, + "learning_rate": 4.645366618481098e-07, + "loss": 0.6464, + "num_input_tokens_seen": 4506779648, + "step": 2149 + }, + { + "epoch": 0.04068348250610252, + "grad_norm": 17.086820602416992, + "learning_rate": 4.641403551013081e-07, + "loss": 0.8128, + "num_input_tokens_seen": 4508876800, + "step": 2150 + }, + { + "epoch": 0.040954705722809875, + "grad_norm": 13.269697189331055, + "learning_rate": 4.637441174510813e-07, + "loss": 0.4149, + "num_input_tokens_seen": 4510973952, + "step": 2151 + }, + { + "epoch": 0.04122592893951722, + "grad_norm": 14.31224250793457, + "learning_rate": 4.633479492163066e-07, + "loss": 0.4536, + "num_input_tokens_seen": 4513071104, + "step": 2152 + }, + { + "epoch": 0.04149715215622457, + "grad_norm": 13.99878215789795, + "learning_rate": 4.6295185071580487e-07, + "loss": 0.4298, + "num_input_tokens_seen": 4515168256, + "step": 2153 + }, + { + "epoch": 0.041768375372931925, + "grad_norm": 13.95372200012207, + "learning_rate": 4.6255582226834133e-07, + "loss": 0.4801, + "num_input_tokens_seen": 4517265408, + "step": 2154 + }, + { + "epoch": 0.04203959858963927, + "grad_norm": 19.383800506591797, + "learning_rate": 4.6215986419262444e-07, + "loss": 0.6272, + "num_input_tokens_seen": 4519362560, + "step": 2155 + }, + { + "epoch": 0.04231082180634662, + "grad_norm": 22.379093170166016, + "learning_rate": 4.617639768073066e-07, + "loss": 0.9261, + "num_input_tokens_seen": 4521459712, + "step": 2156 + }, + { + "epoch": 0.042582045023053974, + "grad_norm": 9.793371200561523, + "learning_rate": 4.613681604309824e-07, + "loss": 0.3081, + "num_input_tokens_seen": 4523556864, + "step": 2157 + }, + { + "epoch": 0.04285326823976132, + "grad_norm": 16.70575714111328, + "learning_rate": 4.609724153821898e-07, + "loss": 0.5864, + "num_input_tokens_seen": 4525654016, + "step": 2158 + }, + { + "epoch": 0.043124491456468676, + "grad_norm": 15.424320220947266, + "learning_rate": 4.6057674197940974e-07, + "loss": 0.5138, + "num_input_tokens_seen": 4527751168, + "step": 2159 + }, + { + "epoch": 0.043395714673176024, + "grad_norm": 12.337757110595703, + "learning_rate": 4.6018114054106494e-07, + "loss": 0.3479, + "num_input_tokens_seen": 4529848320, + "step": 2160 + }, + { + "epoch": 0.04366693788988337, + "grad_norm": 12.336735725402832, + "learning_rate": 4.597856113855203e-07, + "loss": 0.3803, + "num_input_tokens_seen": 4531945472, + "step": 2161 + }, + { + "epoch": 0.043938161106590726, + "grad_norm": 18.751192092895508, + "learning_rate": 4.593901548310828e-07, + "loss": 0.6854, + "num_input_tokens_seen": 4534042624, + "step": 2162 + }, + { + "epoch": 0.04420938432329807, + "grad_norm": 13.754616737365723, + "learning_rate": 4.5899477119600073e-07, + "loss": 0.483, + "num_input_tokens_seen": 4536139776, + "step": 2163 + }, + { + "epoch": 0.04448060754000543, + "grad_norm": 22.812694549560547, + "learning_rate": 4.585994607984637e-07, + "loss": 0.5172, + "num_input_tokens_seen": 4538236928, + "step": 2164 + }, + { + "epoch": 0.044751830756712775, + "grad_norm": 13.95494270324707, + "learning_rate": 4.582042239566026e-07, + "loss": 0.5708, + "num_input_tokens_seen": 4540334080, + "step": 2165 + }, + { + "epoch": 0.04502305397342012, + "grad_norm": 14.064531326293945, + "learning_rate": 4.578090609884889e-07, + "loss": 0.3845, + "num_input_tokens_seen": 4542431232, + "step": 2166 + }, + { + "epoch": 0.04529427719012748, + "grad_norm": 13.075774192810059, + "learning_rate": 4.57413972212135e-07, + "loss": 0.3608, + "num_input_tokens_seen": 4544528384, + "step": 2167 + }, + { + "epoch": 0.045565500406834825, + "grad_norm": 12.075623512268066, + "learning_rate": 4.5701895794549293e-07, + "loss": 0.3836, + "num_input_tokens_seen": 4546625536, + "step": 2168 + }, + { + "epoch": 0.04583672362354217, + "grad_norm": 14.203977584838867, + "learning_rate": 4.566240185064554e-07, + "loss": 0.3646, + "num_input_tokens_seen": 4548722688, + "step": 2169 + }, + { + "epoch": 0.04610794684024953, + "grad_norm": 16.41777229309082, + "learning_rate": 4.5622915421285446e-07, + "loss": 0.5869, + "num_input_tokens_seen": 4550819840, + "step": 2170 + }, + { + "epoch": 0.046379170056956874, + "grad_norm": 5.759540557861328, + "learning_rate": 4.558343653824619e-07, + "loss": 0.1035, + "num_input_tokens_seen": 4552916992, + "step": 2171 + }, + { + "epoch": 0.04665039327366423, + "grad_norm": 11.012629508972168, + "learning_rate": 4.55439652332989e-07, + "loss": 0.3469, + "num_input_tokens_seen": 4555014144, + "step": 2172 + }, + { + "epoch": 0.046921616490371576, + "grad_norm": 16.583019256591797, + "learning_rate": 4.550450153820857e-07, + "loss": 0.4748, + "num_input_tokens_seen": 4557111296, + "step": 2173 + }, + { + "epoch": 0.047192839707078924, + "grad_norm": 12.1753511428833, + "learning_rate": 4.5465045484734044e-07, + "loss": 0.4094, + "num_input_tokens_seen": 4559208448, + "step": 2174 + }, + { + "epoch": 0.04746406292378628, + "grad_norm": 12.30431842803955, + "learning_rate": 4.5425597104628086e-07, + "loss": 0.3467, + "num_input_tokens_seen": 4561305600, + "step": 2175 + }, + { + "epoch": 0.047735286140493625, + "grad_norm": 14.555771827697754, + "learning_rate": 4.5386156429637256e-07, + "loss": 0.4869, + "num_input_tokens_seen": 4563402752, + "step": 2176 + }, + { + "epoch": 0.04800650935720097, + "grad_norm": 14.621318817138672, + "learning_rate": 4.534672349150189e-07, + "loss": 0.3947, + "num_input_tokens_seen": 4565499904, + "step": 2177 + }, + { + "epoch": 0.04827773257390833, + "grad_norm": 14.621907234191895, + "learning_rate": 4.530729832195612e-07, + "loss": 0.3238, + "num_input_tokens_seen": 4567597056, + "step": 2178 + }, + { + "epoch": 0.048548955790615675, + "grad_norm": 17.090391159057617, + "learning_rate": 4.5267880952727845e-07, + "loss": 0.4632, + "num_input_tokens_seen": 4569694208, + "step": 2179 + }, + { + "epoch": 0.04882017900732303, + "grad_norm": 22.62020492553711, + "learning_rate": 4.5228471415538637e-07, + "loss": 0.9961, + "num_input_tokens_seen": 4571791360, + "step": 2180 + }, + { + "epoch": 0.04909140222403038, + "grad_norm": 15.037042617797852, + "learning_rate": 4.518906974210379e-07, + "loss": 0.4256, + "num_input_tokens_seen": 4573888512, + "step": 2181 + }, + { + "epoch": 0.049362625440737724, + "grad_norm": 13.915372848510742, + "learning_rate": 4.514967596413228e-07, + "loss": 0.4098, + "num_input_tokens_seen": 4575985664, + "step": 2182 + }, + { + "epoch": 0.04963384865744508, + "grad_norm": 14.323426246643066, + "learning_rate": 4.511029011332672e-07, + "loss": 0.3687, + "num_input_tokens_seen": 4578082816, + "step": 2183 + }, + { + "epoch": 0.049905071874152426, + "grad_norm": 16.720897674560547, + "learning_rate": 4.507091222138335e-07, + "loss": 0.6297, + "num_input_tokens_seen": 4580179968, + "step": 2184 + }, + { + "epoch": 0.05017629509085978, + "grad_norm": 12.606710433959961, + "learning_rate": 4.503154231999202e-07, + "loss": 0.2422, + "num_input_tokens_seen": 4582277120, + "step": 2185 + }, + { + "epoch": 0.05044751830756713, + "grad_norm": 19.402252197265625, + "learning_rate": 4.499218044083608e-07, + "loss": 0.7964, + "num_input_tokens_seen": 4584374272, + "step": 2186 + }, + { + "epoch": 0.050718741524274476, + "grad_norm": 13.136670112609863, + "learning_rate": 4.495282661559248e-07, + "loss": 0.3877, + "num_input_tokens_seen": 4586471424, + "step": 2187 + }, + { + "epoch": 0.05098996474098183, + "grad_norm": 16.292800903320312, + "learning_rate": 4.491348087593171e-07, + "loss": 0.6262, + "num_input_tokens_seen": 4588568576, + "step": 2188 + }, + { + "epoch": 0.05126118795768918, + "grad_norm": 9.792939186096191, + "learning_rate": 4.4874143253517706e-07, + "loss": 0.2886, + "num_input_tokens_seen": 4590665728, + "step": 2189 + }, + { + "epoch": 0.051532411174396525, + "grad_norm": 20.708683013916016, + "learning_rate": 4.483481378000791e-07, + "loss": 0.8719, + "num_input_tokens_seen": 4592762880, + "step": 2190 + }, + { + "epoch": 0.05180363439110388, + "grad_norm": 21.062313079833984, + "learning_rate": 4.4795492487053155e-07, + "loss": 0.5302, + "num_input_tokens_seen": 4594860032, + "step": 2191 + }, + { + "epoch": 0.05207485760781123, + "grad_norm": 11.393677711486816, + "learning_rate": 4.4756179406297744e-07, + "loss": 0.3503, + "num_input_tokens_seen": 4596957184, + "step": 2192 + }, + { + "epoch": 0.05234608082451858, + "grad_norm": 13.3013334274292, + "learning_rate": 4.4716874569379336e-07, + "loss": 0.4068, + "num_input_tokens_seen": 4599054336, + "step": 2193 + }, + { + "epoch": 0.05261730404122593, + "grad_norm": 14.940890312194824, + "learning_rate": 4.467757800792897e-07, + "loss": 0.3933, + "num_input_tokens_seen": 4601151488, + "step": 2194 + }, + { + "epoch": 0.05288852725793328, + "grad_norm": 18.109926223754883, + "learning_rate": 4.4638289753571025e-07, + "loss": 0.5994, + "num_input_tokens_seen": 4603248640, + "step": 2195 + }, + { + "epoch": 0.05315975047464063, + "grad_norm": 11.895615577697754, + "learning_rate": 4.459900983792321e-07, + "loss": 0.355, + "num_input_tokens_seen": 4605345792, + "step": 2196 + }, + { + "epoch": 0.05343097369134798, + "grad_norm": 12.7399320602417, + "learning_rate": 4.455973829259648e-07, + "loss": 0.3251, + "num_input_tokens_seen": 4607442944, + "step": 2197 + }, + { + "epoch": 0.05370219690805533, + "grad_norm": 14.225614547729492, + "learning_rate": 4.4520475149195093e-07, + "loss": 0.4504, + "num_input_tokens_seen": 4609540096, + "step": 2198 + }, + { + "epoch": 0.05397342012476268, + "grad_norm": 12.06645679473877, + "learning_rate": 4.4481220439316514e-07, + "loss": 0.3057, + "num_input_tokens_seen": 4611637248, + "step": 2199 + }, + { + "epoch": 0.05424464334147003, + "grad_norm": 17.66695785522461, + "learning_rate": 4.444197419455147e-07, + "loss": 0.7025, + "num_input_tokens_seen": 4613734400, + "step": 2200 + }, + { + "epoch": 0.05451586655817738, + "grad_norm": 14.438081741333008, + "learning_rate": 4.4402736446483824e-07, + "loss": 0.5549, + "num_input_tokens_seen": 4615831552, + "step": 2201 + }, + { + "epoch": 0.05478708977488473, + "grad_norm": 21.305463790893555, + "learning_rate": 4.436350722669065e-07, + "loss": 0.4533, + "num_input_tokens_seen": 4617928704, + "step": 2202 + }, + { + "epoch": 0.05505831299159208, + "grad_norm": 17.13565444946289, + "learning_rate": 4.43242865667421e-07, + "loss": 0.5579, + "num_input_tokens_seen": 4620025856, + "step": 2203 + }, + { + "epoch": 0.05532953620829943, + "grad_norm": 16.301301956176758, + "learning_rate": 4.428507449820147e-07, + "loss": 0.5478, + "num_input_tokens_seen": 4622123008, + "step": 2204 + }, + { + "epoch": 0.05560075942500678, + "grad_norm": 17.26921272277832, + "learning_rate": 4.424587105262516e-07, + "loss": 0.6332, + "num_input_tokens_seen": 4624220160, + "step": 2205 + }, + { + "epoch": 0.055871982641714134, + "grad_norm": 16.908565521240234, + "learning_rate": 4.4206676261562603e-07, + "loss": 0.6461, + "num_input_tokens_seen": 4626317312, + "step": 2206 + }, + { + "epoch": 0.05614320585842148, + "grad_norm": 15.162359237670898, + "learning_rate": 4.4167490156556274e-07, + "loss": 0.4447, + "num_input_tokens_seen": 4628414464, + "step": 2207 + }, + { + "epoch": 0.05641442907512883, + "grad_norm": 12.857962608337402, + "learning_rate": 4.4128312769141664e-07, + "loss": 0.4044, + "num_input_tokens_seen": 4630511616, + "step": 2208 + }, + { + "epoch": 0.05668565229183618, + "grad_norm": 12.184791564941406, + "learning_rate": 4.408914413084722e-07, + "loss": 0.3242, + "num_input_tokens_seen": 4632608768, + "step": 2209 + }, + { + "epoch": 0.05695687550854353, + "grad_norm": 19.207063674926758, + "learning_rate": 4.4049984273194406e-07, + "loss": 0.7085, + "num_input_tokens_seen": 4634705920, + "step": 2210 + }, + { + "epoch": 0.05722809872525088, + "grad_norm": 19.66663932800293, + "learning_rate": 4.401083322769756e-07, + "loss": 0.7906, + "num_input_tokens_seen": 4636803072, + "step": 2211 + }, + { + "epoch": 0.05749932194195823, + "grad_norm": 13.727479934692383, + "learning_rate": 4.3971691025863964e-07, + "loss": 0.4326, + "num_input_tokens_seen": 4638900224, + "step": 2212 + }, + { + "epoch": 0.05777054515866558, + "grad_norm": 15.684118270874023, + "learning_rate": 4.3932557699193794e-07, + "loss": 0.5288, + "num_input_tokens_seen": 4640997376, + "step": 2213 + }, + { + "epoch": 0.058041768375372935, + "grad_norm": 12.542720794677734, + "learning_rate": 4.389343327918004e-07, + "loss": 0.401, + "num_input_tokens_seen": 4643094528, + "step": 2214 + }, + { + "epoch": 0.05831299159208028, + "grad_norm": 13.282032012939453, + "learning_rate": 4.3854317797308515e-07, + "loss": 0.4052, + "num_input_tokens_seen": 4645191680, + "step": 2215 + }, + { + "epoch": 0.05858421480878763, + "grad_norm": 12.997306823730469, + "learning_rate": 4.381521128505791e-07, + "loss": 0.3949, + "num_input_tokens_seen": 4647288832, + "step": 2216 + }, + { + "epoch": 0.058855438025494984, + "grad_norm": 10.497172355651855, + "learning_rate": 4.377611377389966e-07, + "loss": 0.2387, + "num_input_tokens_seen": 4649385984, + "step": 2217 + }, + { + "epoch": 0.05912666124220233, + "grad_norm": 11.194908142089844, + "learning_rate": 4.3737025295297937e-07, + "loss": 0.3324, + "num_input_tokens_seen": 4651483136, + "step": 2218 + }, + { + "epoch": 0.059397884458909686, + "grad_norm": 16.6368408203125, + "learning_rate": 4.3697945880709686e-07, + "loss": 0.5526, + "num_input_tokens_seen": 4653580288, + "step": 2219 + }, + { + "epoch": 0.059669107675617034, + "grad_norm": 17.438602447509766, + "learning_rate": 4.3658875561584494e-07, + "loss": 0.5616, + "num_input_tokens_seen": 4655677440, + "step": 2220 + }, + { + "epoch": 0.05994033089232438, + "grad_norm": 10.89105224609375, + "learning_rate": 4.36198143693647e-07, + "loss": 0.2007, + "num_input_tokens_seen": 4657774592, + "step": 2221 + }, + { + "epoch": 0.060211554109031735, + "grad_norm": 10.294246673583984, + "learning_rate": 4.3580762335485253e-07, + "loss": 0.2698, + "num_input_tokens_seen": 4659871744, + "step": 2222 + }, + { + "epoch": 0.06048277732573908, + "grad_norm": 22.007549285888672, + "learning_rate": 4.3541719491373743e-07, + "loss": 0.7437, + "num_input_tokens_seen": 4661968896, + "step": 2223 + }, + { + "epoch": 0.06075400054244643, + "grad_norm": 18.49791717529297, + "learning_rate": 4.350268586845035e-07, + "loss": 0.6865, + "num_input_tokens_seen": 4664066048, + "step": 2224 + }, + { + "epoch": 0.061025223759153785, + "grad_norm": 13.56779956817627, + "learning_rate": 4.34636614981279e-07, + "loss": 0.4945, + "num_input_tokens_seen": 4666163200, + "step": 2225 + }, + { + "epoch": 0.06129644697586113, + "grad_norm": 14.960966110229492, + "learning_rate": 4.342464641181166e-07, + "loss": 0.4721, + "num_input_tokens_seen": 4668260352, + "step": 2226 + }, + { + "epoch": 0.06156767019256849, + "grad_norm": 11.218042373657227, + "learning_rate": 4.3385640640899524e-07, + "loss": 0.3769, + "num_input_tokens_seen": 4670357504, + "step": 2227 + }, + { + "epoch": 0.061838893409275834, + "grad_norm": 13.593111991882324, + "learning_rate": 4.3346644216781823e-07, + "loss": 0.4218, + "num_input_tokens_seen": 4672454656, + "step": 2228 + }, + { + "epoch": 0.06211011662598318, + "grad_norm": 11.54797077178955, + "learning_rate": 4.3307657170841417e-07, + "loss": 0.2934, + "num_input_tokens_seen": 4674551808, + "step": 2229 + }, + { + "epoch": 0.062381339842690536, + "grad_norm": 17.747589111328125, + "learning_rate": 4.326867953445362e-07, + "loss": 0.6918, + "num_input_tokens_seen": 4676648960, + "step": 2230 + }, + { + "epoch": 0.06265256305939788, + "grad_norm": 14.534211158752441, + "learning_rate": 4.322971133898611e-07, + "loss": 0.4678, + "num_input_tokens_seen": 4678746112, + "step": 2231 + }, + { + "epoch": 0.06292378627610523, + "grad_norm": 15.299826622009277, + "learning_rate": 4.319075261579901e-07, + "loss": 0.3943, + "num_input_tokens_seen": 4680843264, + "step": 2232 + }, + { + "epoch": 0.06319500949281258, + "grad_norm": 14.455315589904785, + "learning_rate": 4.3151803396244833e-07, + "loss": 0.5782, + "num_input_tokens_seen": 4682940416, + "step": 2233 + }, + { + "epoch": 0.06346623270951994, + "grad_norm": 18.758607864379883, + "learning_rate": 4.3112863711668435e-07, + "loss": 0.3796, + "num_input_tokens_seen": 4685037568, + "step": 2234 + }, + { + "epoch": 0.06373745592622729, + "grad_norm": 19.549890518188477, + "learning_rate": 4.3073933593406974e-07, + "loss": 0.9138, + "num_input_tokens_seen": 4687134720, + "step": 2235 + }, + { + "epoch": 0.06400867914293464, + "grad_norm": 11.473661422729492, + "learning_rate": 4.303501307278994e-07, + "loss": 0.3049, + "num_input_tokens_seen": 4689231872, + "step": 2236 + }, + { + "epoch": 0.06427990235964198, + "grad_norm": 12.786169052124023, + "learning_rate": 4.299610218113908e-07, + "loss": 0.4731, + "num_input_tokens_seen": 4691329024, + "step": 2237 + }, + { + "epoch": 0.06455112557634933, + "grad_norm": 15.072799682617188, + "learning_rate": 4.2957200949768414e-07, + "loss": 0.3859, + "num_input_tokens_seen": 4693426176, + "step": 2238 + }, + { + "epoch": 0.06482234879305669, + "grad_norm": 20.06634521484375, + "learning_rate": 4.2918309409984145e-07, + "loss": 0.72, + "num_input_tokens_seen": 4695523328, + "step": 2239 + }, + { + "epoch": 0.06509357200976404, + "grad_norm": 8.56005573272705, + "learning_rate": 4.2879427593084714e-07, + "loss": 0.1702, + "num_input_tokens_seen": 4697620480, + "step": 2240 + }, + { + "epoch": 0.06536479522647139, + "grad_norm": 14.993252754211426, + "learning_rate": 4.2840555530360756e-07, + "loss": 0.403, + "num_input_tokens_seen": 4699717632, + "step": 2241 + }, + { + "epoch": 0.06563601844317873, + "grad_norm": 13.477447509765625, + "learning_rate": 4.280169325309502e-07, + "loss": 0.3717, + "num_input_tokens_seen": 4701814784, + "step": 2242 + }, + { + "epoch": 0.06590724165988608, + "grad_norm": 16.231098175048828, + "learning_rate": 4.276284079256235e-07, + "loss": 0.3061, + "num_input_tokens_seen": 4703911936, + "step": 2243 + }, + { + "epoch": 0.06617846487659344, + "grad_norm": 16.42155647277832, + "learning_rate": 4.272399818002974e-07, + "loss": 0.561, + "num_input_tokens_seen": 4706009088, + "step": 2244 + }, + { + "epoch": 0.06644968809330079, + "grad_norm": 14.12633228302002, + "learning_rate": 4.268516544675628e-07, + "loss": 0.293, + "num_input_tokens_seen": 4708106240, + "step": 2245 + }, + { + "epoch": 0.06672091131000814, + "grad_norm": 14.42664623260498, + "learning_rate": 4.2646342623993035e-07, + "loss": 0.3539, + "num_input_tokens_seen": 4710203392, + "step": 2246 + }, + { + "epoch": 0.06699213452671549, + "grad_norm": 11.090810775756836, + "learning_rate": 4.2607529742983174e-07, + "loss": 0.3227, + "num_input_tokens_seen": 4712300544, + "step": 2247 + }, + { + "epoch": 0.06726335774342283, + "grad_norm": 21.97173500061035, + "learning_rate": 4.2568726834961797e-07, + "loss": 0.7627, + "num_input_tokens_seen": 4714397696, + "step": 2248 + }, + { + "epoch": 0.06753458096013018, + "grad_norm": 10.479046821594238, + "learning_rate": 4.252993393115601e-07, + "loss": 0.2715, + "num_input_tokens_seen": 4716494848, + "step": 2249 + }, + { + "epoch": 0.06780580417683754, + "grad_norm": 20.145353317260742, + "learning_rate": 4.249115106278489e-07, + "loss": 0.4898, + "num_input_tokens_seen": 4718592000, + "step": 2250 + }, + { + "epoch": 0.06807702739354489, + "grad_norm": 18.290271759033203, + "learning_rate": 4.2452378261059397e-07, + "loss": 0.5795, + "num_input_tokens_seen": 4720689152, + "step": 2251 + }, + { + "epoch": 0.06834825061025224, + "grad_norm": 16.21881103515625, + "learning_rate": 4.24136155571824e-07, + "loss": 0.48, + "num_input_tokens_seen": 4722786304, + "step": 2252 + }, + { + "epoch": 0.06861947382695958, + "grad_norm": 14.017351150512695, + "learning_rate": 4.2374862982348657e-07, + "loss": 0.431, + "num_input_tokens_seen": 4724883456, + "step": 2253 + }, + { + "epoch": 0.06889069704366693, + "grad_norm": 20.59170150756836, + "learning_rate": 4.233612056774477e-07, + "loss": 0.776, + "num_input_tokens_seen": 4726980608, + "step": 2254 + }, + { + "epoch": 0.0691619202603743, + "grad_norm": 13.373947143554688, + "learning_rate": 4.2297388344549146e-07, + "loss": 0.3047, + "num_input_tokens_seen": 4729077760, + "step": 2255 + }, + { + "epoch": 0.06943314347708164, + "grad_norm": 15.071645736694336, + "learning_rate": 4.2258666343932004e-07, + "loss": 0.4074, + "num_input_tokens_seen": 4731174912, + "step": 2256 + }, + { + "epoch": 0.06970436669378899, + "grad_norm": 14.832840919494629, + "learning_rate": 4.2219954597055354e-07, + "loss": 0.3752, + "num_input_tokens_seen": 4733272064, + "step": 2257 + }, + { + "epoch": 0.06997558991049634, + "grad_norm": 11.459256172180176, + "learning_rate": 4.2181253135072925e-07, + "loss": 0.3155, + "num_input_tokens_seen": 4735369216, + "step": 2258 + }, + { + "epoch": 0.07024681312720368, + "grad_norm": 22.15343475341797, + "learning_rate": 4.21425619891302e-07, + "loss": 0.6153, + "num_input_tokens_seen": 4737466368, + "step": 2259 + }, + { + "epoch": 0.07051803634391104, + "grad_norm": 21.2708740234375, + "learning_rate": 4.21038811903643e-07, + "loss": 0.6954, + "num_input_tokens_seen": 4739563520, + "step": 2260 + }, + { + "epoch": 0.07078925956061839, + "grad_norm": 17.88982582092285, + "learning_rate": 4.206521076990409e-07, + "loss": 0.6006, + "num_input_tokens_seen": 4741660672, + "step": 2261 + }, + { + "epoch": 0.07106048277732574, + "grad_norm": 14.392377853393555, + "learning_rate": 4.202655075887005e-07, + "loss": 0.451, + "num_input_tokens_seen": 4743757824, + "step": 2262 + }, + { + "epoch": 0.07133170599403309, + "grad_norm": 12.109110832214355, + "learning_rate": 4.1987901188374286e-07, + "loss": 0.2934, + "num_input_tokens_seen": 4745854976, + "step": 2263 + }, + { + "epoch": 0.07160292921074043, + "grad_norm": 14.256464958190918, + "learning_rate": 4.194926208952051e-07, + "loss": 0.4602, + "num_input_tokens_seen": 4747952128, + "step": 2264 + }, + { + "epoch": 0.0718741524274478, + "grad_norm": 11.486831665039062, + "learning_rate": 4.191063349340397e-07, + "loss": 0.3412, + "num_input_tokens_seen": 4750049280, + "step": 2265 + }, + { + "epoch": 0.07214537564415514, + "grad_norm": 12.454242706298828, + "learning_rate": 4.1872015431111505e-07, + "loss": 0.3429, + "num_input_tokens_seen": 4752146432, + "step": 2266 + }, + { + "epoch": 0.07241659886086249, + "grad_norm": 17.0006103515625, + "learning_rate": 4.1833407933721476e-07, + "loss": 0.6483, + "num_input_tokens_seen": 4754243584, + "step": 2267 + }, + { + "epoch": 0.07268782207756984, + "grad_norm": 15.69262981414795, + "learning_rate": 4.17948110323037e-07, + "loss": 0.3751, + "num_input_tokens_seen": 4756340736, + "step": 2268 + }, + { + "epoch": 0.07295904529427719, + "grad_norm": 12.206964492797852, + "learning_rate": 4.1756224757919513e-07, + "loss": 0.3653, + "num_input_tokens_seen": 4758437888, + "step": 2269 + }, + { + "epoch": 0.07323026851098453, + "grad_norm": 16.216537475585938, + "learning_rate": 4.17176491416217e-07, + "loss": 0.5839, + "num_input_tokens_seen": 4760535040, + "step": 2270 + }, + { + "epoch": 0.0735014917276919, + "grad_norm": 11.075773239135742, + "learning_rate": 4.1679084214454405e-07, + "loss": 0.3098, + "num_input_tokens_seen": 4762632192, + "step": 2271 + }, + { + "epoch": 0.07377271494439924, + "grad_norm": 16.252365112304688, + "learning_rate": 4.1640530007453245e-07, + "loss": 0.4551, + "num_input_tokens_seen": 4764729344, + "step": 2272 + }, + { + "epoch": 0.07404393816110659, + "grad_norm": 11.240346908569336, + "learning_rate": 4.1601986551645163e-07, + "loss": 0.2972, + "num_input_tokens_seen": 4766826496, + "step": 2273 + }, + { + "epoch": 0.07431516137781394, + "grad_norm": 13.240141868591309, + "learning_rate": 4.156345387804847e-07, + "loss": 0.3682, + "num_input_tokens_seen": 4768923648, + "step": 2274 + }, + { + "epoch": 0.07458638459452128, + "grad_norm": 10.888014793395996, + "learning_rate": 4.152493201767281e-07, + "loss": 0.336, + "num_input_tokens_seen": 4771020800, + "step": 2275 + }, + { + "epoch": 0.07485760781122865, + "grad_norm": 18.147985458374023, + "learning_rate": 4.1486421001519087e-07, + "loss": 0.6925, + "num_input_tokens_seen": 4773117952, + "step": 2276 + }, + { + "epoch": 0.075128831027936, + "grad_norm": 13.516504287719727, + "learning_rate": 4.1447920860579524e-07, + "loss": 0.3624, + "num_input_tokens_seen": 4775215104, + "step": 2277 + }, + { + "epoch": 0.07540005424464334, + "grad_norm": 16.765531539916992, + "learning_rate": 4.1409431625837545e-07, + "loss": 0.5945, + "num_input_tokens_seen": 4777312256, + "step": 2278 + }, + { + "epoch": 0.07567127746135069, + "grad_norm": 15.376947402954102, + "learning_rate": 4.137095332826784e-07, + "loss": 0.4172, + "num_input_tokens_seen": 4779409408, + "step": 2279 + }, + { + "epoch": 0.07594250067805804, + "grad_norm": 15.102494239807129, + "learning_rate": 4.1332485998836277e-07, + "loss": 0.4686, + "num_input_tokens_seen": 4781506560, + "step": 2280 + }, + { + "epoch": 0.0762137238947654, + "grad_norm": 18.478492736816406, + "learning_rate": 4.129402966849987e-07, + "loss": 0.2808, + "num_input_tokens_seen": 4783603712, + "step": 2281 + }, + { + "epoch": 0.07648494711147275, + "grad_norm": 16.129724502563477, + "learning_rate": 4.1255584368206877e-07, + "loss": 0.5027, + "num_input_tokens_seen": 4785700864, + "step": 2282 + }, + { + "epoch": 0.07675617032818009, + "grad_norm": 16.364654541015625, + "learning_rate": 4.121715012889655e-07, + "loss": 0.4003, + "num_input_tokens_seen": 4787798016, + "step": 2283 + }, + { + "epoch": 0.07702739354488744, + "grad_norm": 12.427045822143555, + "learning_rate": 4.1178726981499313e-07, + "loss": 0.393, + "num_input_tokens_seen": 4789895168, + "step": 2284 + }, + { + "epoch": 0.07729861676159479, + "grad_norm": 27.002498626708984, + "learning_rate": 4.1140314956936663e-07, + "loss": 0.6929, + "num_input_tokens_seen": 4791992320, + "step": 2285 + }, + { + "epoch": 0.07756983997830215, + "grad_norm": 11.09440803527832, + "learning_rate": 4.1101914086121137e-07, + "loss": 0.2603, + "num_input_tokens_seen": 4794089472, + "step": 2286 + }, + { + "epoch": 0.0778410631950095, + "grad_norm": 15.960718154907227, + "learning_rate": 4.106352439995632e-07, + "loss": 0.482, + "num_input_tokens_seen": 4796186624, + "step": 2287 + }, + { + "epoch": 0.07811228641171684, + "grad_norm": 14.623363494873047, + "learning_rate": 4.102514592933671e-07, + "loss": 0.3698, + "num_input_tokens_seen": 4798283776, + "step": 2288 + }, + { + "epoch": 0.07838350962842419, + "grad_norm": 21.368053436279297, + "learning_rate": 4.098677870514788e-07, + "loss": 0.4633, + "num_input_tokens_seen": 4800380928, + "step": 2289 + }, + { + "epoch": 0.07865473284513154, + "grad_norm": 21.473129272460938, + "learning_rate": 4.094842275826631e-07, + "loss": 0.4671, + "num_input_tokens_seen": 4802478080, + "step": 2290 + }, + { + "epoch": 0.07892595606183889, + "grad_norm": 18.82076644897461, + "learning_rate": 4.091007811955941e-07, + "loss": 0.4475, + "num_input_tokens_seen": 4804575232, + "step": 2291 + }, + { + "epoch": 0.07919717927854625, + "grad_norm": 15.747272491455078, + "learning_rate": 4.0871744819885467e-07, + "loss": 0.5627, + "num_input_tokens_seen": 4806672384, + "step": 2292 + }, + { + "epoch": 0.0794684024952536, + "grad_norm": 18.276941299438477, + "learning_rate": 4.0833422890093684e-07, + "loss": 0.6141, + "num_input_tokens_seen": 4808769536, + "step": 2293 + }, + { + "epoch": 0.07973962571196094, + "grad_norm": 18.258466720581055, + "learning_rate": 4.0795112361024075e-07, + "loss": 0.5932, + "num_input_tokens_seen": 4810866688, + "step": 2294 + }, + { + "epoch": 0.08001084892866829, + "grad_norm": 12.13038158416748, + "learning_rate": 4.0756813263507496e-07, + "loss": 0.3242, + "num_input_tokens_seen": 4812963840, + "step": 2295 + }, + { + "epoch": 0.08028207214537564, + "grad_norm": 13.339944839477539, + "learning_rate": 4.0718525628365617e-07, + "loss": 0.3595, + "num_input_tokens_seen": 4815060992, + "step": 2296 + }, + { + "epoch": 0.080553295362083, + "grad_norm": 24.59098243713379, + "learning_rate": 4.0680249486410845e-07, + "loss": 0.4553, + "num_input_tokens_seen": 4817158144, + "step": 2297 + }, + { + "epoch": 0.08082451857879035, + "grad_norm": 18.36080551147461, + "learning_rate": 4.0641984868446386e-07, + "loss": 0.5813, + "num_input_tokens_seen": 4819255296, + "step": 2298 + }, + { + "epoch": 0.0810957417954977, + "grad_norm": 12.506386756896973, + "learning_rate": 4.0603731805266175e-07, + "loss": 0.369, + "num_input_tokens_seen": 4821352448, + "step": 2299 + }, + { + "epoch": 0.08136696501220504, + "grad_norm": 14.215474128723145, + "learning_rate": 4.056549032765476e-07, + "loss": 0.4783, + "num_input_tokens_seen": 4823449600, + "step": 2300 + }, + { + "epoch": 0.08163818822891239, + "grad_norm": 25.43838119506836, + "learning_rate": 4.0527260466387446e-07, + "loss": 0.8514, + "num_input_tokens_seen": 4825546752, + "step": 2301 + }, + { + "epoch": 0.08190941144561975, + "grad_norm": 17.24186134338379, + "learning_rate": 4.0489042252230197e-07, + "loss": 0.5077, + "num_input_tokens_seen": 4827643904, + "step": 2302 + }, + { + "epoch": 0.0821806346623271, + "grad_norm": 13.061026573181152, + "learning_rate": 4.045083571593955e-07, + "loss": 0.3865, + "num_input_tokens_seen": 4829741056, + "step": 2303 + }, + { + "epoch": 0.08245185787903445, + "grad_norm": 19.54876136779785, + "learning_rate": 4.0412640888262697e-07, + "loss": 0.7823, + "num_input_tokens_seen": 4831838208, + "step": 2304 + }, + { + "epoch": 0.0827230810957418, + "grad_norm": 13.487531661987305, + "learning_rate": 4.0374457799937354e-07, + "loss": 0.4054, + "num_input_tokens_seen": 4833935360, + "step": 2305 + }, + { + "epoch": 0.08299430431244914, + "grad_norm": 15.34504222869873, + "learning_rate": 4.033628648169184e-07, + "loss": 0.3901, + "num_input_tokens_seen": 4836032512, + "step": 2306 + }, + { + "epoch": 0.0832655275291565, + "grad_norm": 23.76385498046875, + "learning_rate": 4.0298126964244947e-07, + "loss": 0.5777, + "num_input_tokens_seen": 4838129664, + "step": 2307 + }, + { + "epoch": 0.08353675074586385, + "grad_norm": 11.67712116241455, + "learning_rate": 4.025997927830604e-07, + "loss": 0.3415, + "num_input_tokens_seen": 4840226816, + "step": 2308 + }, + { + "epoch": 0.0838079739625712, + "grad_norm": 18.445024490356445, + "learning_rate": 4.022184345457492e-07, + "loss": 0.3676, + "num_input_tokens_seen": 4842323968, + "step": 2309 + }, + { + "epoch": 0.08407919717927854, + "grad_norm": 15.196893692016602, + "learning_rate": 4.0183719523741846e-07, + "loss": 0.4861, + "num_input_tokens_seen": 4844421120, + "step": 2310 + }, + { + "epoch": 0.08435042039598589, + "grad_norm": 16.10106658935547, + "learning_rate": 4.014560751648747e-07, + "loss": 0.4878, + "num_input_tokens_seen": 4846518272, + "step": 2311 + }, + { + "epoch": 0.08462164361269324, + "grad_norm": 17.08901023864746, + "learning_rate": 4.0107507463482924e-07, + "loss": 0.4657, + "num_input_tokens_seen": 4848615424, + "step": 2312 + }, + { + "epoch": 0.0848928668294006, + "grad_norm": 16.41490364074707, + "learning_rate": 4.0069419395389657e-07, + "loss": 0.5807, + "num_input_tokens_seen": 4850712576, + "step": 2313 + }, + { + "epoch": 0.08516409004610795, + "grad_norm": 15.139748573303223, + "learning_rate": 4.0031343342859526e-07, + "loss": 0.5088, + "num_input_tokens_seen": 4852809728, + "step": 2314 + }, + { + "epoch": 0.0854353132628153, + "grad_norm": 16.5777645111084, + "learning_rate": 3.9993279336534667e-07, + "loss": 0.6122, + "num_input_tokens_seen": 4854906880, + "step": 2315 + }, + { + "epoch": 0.08570653647952264, + "grad_norm": 14.561357498168945, + "learning_rate": 3.995522740704758e-07, + "loss": 0.5766, + "num_input_tokens_seen": 4857004032, + "step": 2316 + }, + { + "epoch": 0.08597775969622999, + "grad_norm": 17.497167587280273, + "learning_rate": 3.991718758502094e-07, + "loss": 0.4089, + "num_input_tokens_seen": 4859101184, + "step": 2317 + }, + { + "epoch": 0.08624898291293735, + "grad_norm": 19.5342960357666, + "learning_rate": 3.987915990106779e-07, + "loss": 0.5453, + "num_input_tokens_seen": 4861198336, + "step": 2318 + }, + { + "epoch": 0.0865202061296447, + "grad_norm": 14.03099250793457, + "learning_rate": 3.984114438579137e-07, + "loss": 0.4075, + "num_input_tokens_seen": 4863295488, + "step": 2319 + }, + { + "epoch": 0.08679142934635205, + "grad_norm": 12.443893432617188, + "learning_rate": 3.980314106978512e-07, + "loss": 0.4093, + "num_input_tokens_seen": 4865392640, + "step": 2320 + }, + { + "epoch": 0.0870626525630594, + "grad_norm": 11.860504150390625, + "learning_rate": 3.976514998363265e-07, + "loss": 0.3936, + "num_input_tokens_seen": 4867489792, + "step": 2321 + }, + { + "epoch": 0.08733387577976674, + "grad_norm": 14.789926528930664, + "learning_rate": 3.972717115790773e-07, + "loss": 0.372, + "num_input_tokens_seen": 4869586944, + "step": 2322 + }, + { + "epoch": 0.0876050989964741, + "grad_norm": 18.579681396484375, + "learning_rate": 3.968920462317431e-07, + "loss": 0.5335, + "num_input_tokens_seen": 4871684096, + "step": 2323 + }, + { + "epoch": 0.08787632221318145, + "grad_norm": 12.13874340057373, + "learning_rate": 3.965125040998637e-07, + "loss": 0.3787, + "num_input_tokens_seen": 4873781248, + "step": 2324 + }, + { + "epoch": 0.0881475454298888, + "grad_norm": 14.869010925292969, + "learning_rate": 3.961330854888805e-07, + "loss": 0.4379, + "num_input_tokens_seen": 4875878400, + "step": 2325 + }, + { + "epoch": 0.08841876864659615, + "grad_norm": 10.25628662109375, + "learning_rate": 3.9575379070413485e-07, + "loss": 0.2783, + "num_input_tokens_seen": 4877975552, + "step": 2326 + }, + { + "epoch": 0.0886899918633035, + "grad_norm": 26.221960067749023, + "learning_rate": 3.953746200508693e-07, + "loss": 0.7503, + "num_input_tokens_seen": 4880072704, + "step": 2327 + }, + { + "epoch": 0.08896121508001086, + "grad_norm": 17.73170280456543, + "learning_rate": 3.9499557383422534e-07, + "loss": 0.6836, + "num_input_tokens_seen": 4882169856, + "step": 2328 + }, + { + "epoch": 0.0892324382967182, + "grad_norm": 14.907843589782715, + "learning_rate": 3.94616652359245e-07, + "loss": 0.3363, + "num_input_tokens_seen": 4884267008, + "step": 2329 + }, + { + "epoch": 0.08950366151342555, + "grad_norm": 8.722765922546387, + "learning_rate": 3.942378559308703e-07, + "loss": 0.2435, + "num_input_tokens_seen": 4886364160, + "step": 2330 + }, + { + "epoch": 0.0897748847301329, + "grad_norm": 17.718971252441406, + "learning_rate": 3.93859184853942e-07, + "loss": 0.4271, + "num_input_tokens_seen": 4888461312, + "step": 2331 + }, + { + "epoch": 0.09004610794684025, + "grad_norm": 12.278681755065918, + "learning_rate": 3.934806394332001e-07, + "loss": 0.3989, + "num_input_tokens_seen": 4890558464, + "step": 2332 + }, + { + "epoch": 0.09031733116354759, + "grad_norm": 19.766408920288086, + "learning_rate": 3.9310221997328363e-07, + "loss": 0.5485, + "num_input_tokens_seen": 4892655616, + "step": 2333 + }, + { + "epoch": 0.09058855438025495, + "grad_norm": 18.075057983398438, + "learning_rate": 3.9272392677873e-07, + "loss": 0.6252, + "num_input_tokens_seen": 4894752768, + "step": 2334 + }, + { + "epoch": 0.0908597775969623, + "grad_norm": 21.87603187561035, + "learning_rate": 3.9234576015397536e-07, + "loss": 0.8567, + "num_input_tokens_seen": 4896849920, + "step": 2335 + }, + { + "epoch": 0.09113100081366965, + "grad_norm": 11.228300094604492, + "learning_rate": 3.9196772040335367e-07, + "loss": 0.3036, + "num_input_tokens_seen": 4898947072, + "step": 2336 + }, + { + "epoch": 0.091402224030377, + "grad_norm": 25.12662696838379, + "learning_rate": 3.915898078310972e-07, + "loss": 0.674, + "num_input_tokens_seen": 4901044224, + "step": 2337 + }, + { + "epoch": 0.09167344724708434, + "grad_norm": 19.210277557373047, + "learning_rate": 3.9121202274133525e-07, + "loss": 0.6185, + "num_input_tokens_seen": 4903141376, + "step": 2338 + }, + { + "epoch": 0.0919446704637917, + "grad_norm": 15.579242706298828, + "learning_rate": 3.9083436543809536e-07, + "loss": 0.625, + "num_input_tokens_seen": 4905238528, + "step": 2339 + }, + { + "epoch": 0.09221589368049905, + "grad_norm": 15.06516170501709, + "learning_rate": 3.904568362253011e-07, + "loss": 0.3804, + "num_input_tokens_seen": 4907335680, + "step": 2340 + }, + { + "epoch": 0.0924871168972064, + "grad_norm": 15.726207733154297, + "learning_rate": 3.9007943540677426e-07, + "loss": 0.48, + "num_input_tokens_seen": 4909432832, + "step": 2341 + }, + { + "epoch": 0.09275834011391375, + "grad_norm": 16.687849044799805, + "learning_rate": 3.897021632862321e-07, + "loss": 0.4229, + "num_input_tokens_seen": 4911529984, + "step": 2342 + }, + { + "epoch": 0.0930295633306211, + "grad_norm": 13.342137336730957, + "learning_rate": 3.893250201672893e-07, + "loss": 0.4569, + "num_input_tokens_seen": 4913627136, + "step": 2343 + }, + { + "epoch": 0.09330078654732846, + "grad_norm": 14.77047061920166, + "learning_rate": 3.889480063534563e-07, + "loss": 0.4165, + "num_input_tokens_seen": 4915724288, + "step": 2344 + }, + { + "epoch": 0.0935720097640358, + "grad_norm": 10.642462730407715, + "learning_rate": 3.88571122148139e-07, + "loss": 0.2877, + "num_input_tokens_seen": 4917821440, + "step": 2345 + }, + { + "epoch": 0.09384323298074315, + "grad_norm": 13.166287422180176, + "learning_rate": 3.881943678546399e-07, + "loss": 0.4986, + "num_input_tokens_seen": 4919918592, + "step": 2346 + }, + { + "epoch": 0.0941144561974505, + "grad_norm": 25.655261993408203, + "learning_rate": 3.878177437761564e-07, + "loss": 0.9066, + "num_input_tokens_seen": 4922015744, + "step": 2347 + }, + { + "epoch": 0.09438567941415785, + "grad_norm": 17.33159828186035, + "learning_rate": 3.8744125021578123e-07, + "loss": 0.4746, + "num_input_tokens_seen": 4924112896, + "step": 2348 + }, + { + "epoch": 0.09465690263086521, + "grad_norm": 18.616378784179688, + "learning_rate": 3.87064887476502e-07, + "loss": 0.6, + "num_input_tokens_seen": 4926210048, + "step": 2349 + }, + { + "epoch": 0.09492812584757256, + "grad_norm": 16.331174850463867, + "learning_rate": 3.8668865586120124e-07, + "loss": 0.4109, + "num_input_tokens_seen": 4928307200, + "step": 2350 + }, + { + "epoch": 0.0951993490642799, + "grad_norm": 23.782737731933594, + "learning_rate": 3.8631255567265573e-07, + "loss": 0.5772, + "num_input_tokens_seen": 4930404352, + "step": 2351 + }, + { + "epoch": 0.09547057228098725, + "grad_norm": 14.399580001831055, + "learning_rate": 3.859365872135367e-07, + "loss": 0.408, + "num_input_tokens_seen": 4932501504, + "step": 2352 + }, + { + "epoch": 0.0957417954976946, + "grad_norm": 13.540172576904297, + "learning_rate": 3.8556075078640925e-07, + "loss": 0.4367, + "num_input_tokens_seen": 4934598656, + "step": 2353 + }, + { + "epoch": 0.09601301871440195, + "grad_norm": 16.12992286682129, + "learning_rate": 3.85185046693732e-07, + "loss": 0.5215, + "num_input_tokens_seen": 4936695808, + "step": 2354 + }, + { + "epoch": 0.09628424193110931, + "grad_norm": 13.96000862121582, + "learning_rate": 3.8480947523785767e-07, + "loss": 0.4359, + "num_input_tokens_seen": 4938792960, + "step": 2355 + }, + { + "epoch": 0.09655546514781665, + "grad_norm": 16.476837158203125, + "learning_rate": 3.844340367210318e-07, + "loss": 0.518, + "num_input_tokens_seen": 4940890112, + "step": 2356 + }, + { + "epoch": 0.096826688364524, + "grad_norm": 11.905656814575195, + "learning_rate": 3.840587314453928e-07, + "loss": 0.3644, + "num_input_tokens_seen": 4942987264, + "step": 2357 + }, + { + "epoch": 0.09709791158123135, + "grad_norm": 16.627532958984375, + "learning_rate": 3.8368355971297204e-07, + "loss": 0.3056, + "num_input_tokens_seen": 4945084416, + "step": 2358 + }, + { + "epoch": 0.0973691347979387, + "grad_norm": 14.9150972366333, + "learning_rate": 3.8330852182569374e-07, + "loss": 0.3175, + "num_input_tokens_seen": 4947181568, + "step": 2359 + }, + { + "epoch": 0.09764035801464606, + "grad_norm": 15.67542839050293, + "learning_rate": 3.8293361808537404e-07, + "loss": 0.5347, + "num_input_tokens_seen": 4949278720, + "step": 2360 + }, + { + "epoch": 0.0979115812313534, + "grad_norm": 18.2048397064209, + "learning_rate": 3.825588487937211e-07, + "loss": 0.5031, + "num_input_tokens_seen": 4951375872, + "step": 2361 + }, + { + "epoch": 0.09818280444806075, + "grad_norm": 11.449779510498047, + "learning_rate": 3.821842142523352e-07, + "loss": 0.266, + "num_input_tokens_seen": 4953473024, + "step": 2362 + }, + { + "epoch": 0.0984540276647681, + "grad_norm": 16.407081604003906, + "learning_rate": 3.818097147627076e-07, + "loss": 0.509, + "num_input_tokens_seen": 4955570176, + "step": 2363 + }, + { + "epoch": 0.09872525088147545, + "grad_norm": 19.54366683959961, + "learning_rate": 3.8143535062622154e-07, + "loss": 0.6122, + "num_input_tokens_seen": 4957667328, + "step": 2364 + }, + { + "epoch": 0.09899647409818281, + "grad_norm": 19.319889068603516, + "learning_rate": 3.8106112214415087e-07, + "loss": 0.6662, + "num_input_tokens_seen": 4959764480, + "step": 2365 + }, + { + "epoch": 0.09926769731489016, + "grad_norm": 14.861392974853516, + "learning_rate": 3.8068702961766053e-07, + "loss": 0.479, + "num_input_tokens_seen": 4961861632, + "step": 2366 + }, + { + "epoch": 0.0995389205315975, + "grad_norm": 8.038253784179688, + "learning_rate": 3.8031307334780594e-07, + "loss": 0.1862, + "num_input_tokens_seen": 4963958784, + "step": 2367 + }, + { + "epoch": 0.09981014374830485, + "grad_norm": 13.285365104675293, + "learning_rate": 3.7993925363553294e-07, + "loss": 0.3774, + "num_input_tokens_seen": 4966055936, + "step": 2368 + }, + { + "epoch": 0.1000813669650122, + "grad_norm": 15.662388801574707, + "learning_rate": 3.795655707816772e-07, + "loss": 0.5155, + "num_input_tokens_seen": 4968153088, + "step": 2369 + }, + { + "epoch": 0.10035259018171956, + "grad_norm": 15.881659507751465, + "learning_rate": 3.791920250869645e-07, + "loss": 0.5164, + "num_input_tokens_seen": 4970250240, + "step": 2370 + }, + { + "epoch": 0.10062381339842691, + "grad_norm": 23.03855323791504, + "learning_rate": 3.7881861685201033e-07, + "loss": 1.0028, + "num_input_tokens_seen": 4972347392, + "step": 2371 + }, + { + "epoch": 0.10089503661513426, + "grad_norm": 11.0081148147583, + "learning_rate": 3.784453463773194e-07, + "loss": 0.2919, + "num_input_tokens_seen": 4974444544, + "step": 2372 + }, + { + "epoch": 0.1011662598318416, + "grad_norm": 20.63175392150879, + "learning_rate": 3.780722139632858e-07, + "loss": 0.5551, + "num_input_tokens_seen": 4976541696, + "step": 2373 + }, + { + "epoch": 0.10143748304854895, + "grad_norm": 20.467164993286133, + "learning_rate": 3.776992199101918e-07, + "loss": 0.7329, + "num_input_tokens_seen": 4978638848, + "step": 2374 + }, + { + "epoch": 0.10170870626525631, + "grad_norm": 17.156503677368164, + "learning_rate": 3.773263645182091e-07, + "loss": 0.5547, + "num_input_tokens_seen": 4980736000, + "step": 2375 + }, + { + "epoch": 0.10197992948196366, + "grad_norm": 18.2172794342041, + "learning_rate": 3.769536480873976e-07, + "loss": 0.3412, + "num_input_tokens_seen": 4982833152, + "step": 2376 + }, + { + "epoch": 0.10225115269867101, + "grad_norm": 19.269289016723633, + "learning_rate": 3.765810709177052e-07, + "loss": 0.6615, + "num_input_tokens_seen": 4984930304, + "step": 2377 + }, + { + "epoch": 0.10252237591537836, + "grad_norm": 18.64261817932129, + "learning_rate": 3.762086333089678e-07, + "loss": 0.7011, + "num_input_tokens_seen": 4987027456, + "step": 2378 + }, + { + "epoch": 0.1027935991320857, + "grad_norm": 18.141159057617188, + "learning_rate": 3.758363355609092e-07, + "loss": 0.4578, + "num_input_tokens_seen": 4989124608, + "step": 2379 + }, + { + "epoch": 0.10306482234879305, + "grad_norm": 12.744918823242188, + "learning_rate": 3.7546417797314023e-07, + "loss": 0.3624, + "num_input_tokens_seen": 4991221760, + "step": 2380 + }, + { + "epoch": 0.10333604556550041, + "grad_norm": 17.12135887145996, + "learning_rate": 3.7509216084515916e-07, + "loss": 0.427, + "num_input_tokens_seen": 4993318912, + "step": 2381 + }, + { + "epoch": 0.10360726878220776, + "grad_norm": 16.627965927124023, + "learning_rate": 3.747202844763514e-07, + "loss": 0.5558, + "num_input_tokens_seen": 4995416064, + "step": 2382 + }, + { + "epoch": 0.1038784919989151, + "grad_norm": 19.756608963012695, + "learning_rate": 3.743485491659887e-07, + "loss": 0.6773, + "num_input_tokens_seen": 4997513216, + "step": 2383 + }, + { + "epoch": 0.10414971521562245, + "grad_norm": 14.47923755645752, + "learning_rate": 3.739769552132298e-07, + "loss": 0.4862, + "num_input_tokens_seen": 4999610368, + "step": 2384 + }, + { + "epoch": 0.1044209384323298, + "grad_norm": 12.582110404968262, + "learning_rate": 3.73605502917119e-07, + "loss": 0.2835, + "num_input_tokens_seen": 5001707520, + "step": 2385 + }, + { + "epoch": 0.10469216164903716, + "grad_norm": 15.337152481079102, + "learning_rate": 3.7323419257658716e-07, + "loss": 0.5735, + "num_input_tokens_seen": 5003804672, + "step": 2386 + }, + { + "epoch": 0.10496338486574451, + "grad_norm": 12.261470794677734, + "learning_rate": 3.728630244904507e-07, + "loss": 0.3295, + "num_input_tokens_seen": 5005901824, + "step": 2387 + }, + { + "epoch": 0.10523460808245186, + "grad_norm": 14.014972686767578, + "learning_rate": 3.724919989574116e-07, + "loss": 0.4343, + "num_input_tokens_seen": 5007998976, + "step": 2388 + }, + { + "epoch": 0.1055058312991592, + "grad_norm": 19.110898971557617, + "learning_rate": 3.7212111627605704e-07, + "loss": 0.783, + "num_input_tokens_seen": 5010096128, + "step": 2389 + }, + { + "epoch": 0.10577705451586655, + "grad_norm": 15.27743148803711, + "learning_rate": 3.717503767448593e-07, + "loss": 0.5481, + "num_input_tokens_seen": 5012193280, + "step": 2390 + }, + { + "epoch": 0.10604827773257391, + "grad_norm": 20.981355667114258, + "learning_rate": 3.7137978066217555e-07, + "loss": 0.5404, + "num_input_tokens_seen": 5014290432, + "step": 2391 + }, + { + "epoch": 0.10631950094928126, + "grad_norm": 19.111934661865234, + "learning_rate": 3.710093283262472e-07, + "loss": 0.7735, + "num_input_tokens_seen": 5016387584, + "step": 2392 + }, + { + "epoch": 0.10659072416598861, + "grad_norm": 13.102584838867188, + "learning_rate": 3.706390200352003e-07, + "loss": 0.3887, + "num_input_tokens_seen": 5018484736, + "step": 2393 + }, + { + "epoch": 0.10686194738269596, + "grad_norm": 23.91544532775879, + "learning_rate": 3.7026885608704494e-07, + "loss": 0.5647, + "num_input_tokens_seen": 5020581888, + "step": 2394 + }, + { + "epoch": 0.1071331705994033, + "grad_norm": 14.230181694030762, + "learning_rate": 3.6989883677967483e-07, + "loss": 0.3571, + "num_input_tokens_seen": 5022679040, + "step": 2395 + }, + { + "epoch": 0.10740439381611067, + "grad_norm": 14.798707008361816, + "learning_rate": 3.6952896241086783e-07, + "loss": 0.4492, + "num_input_tokens_seen": 5024776192, + "step": 2396 + }, + { + "epoch": 0.10767561703281801, + "grad_norm": 13.679465293884277, + "learning_rate": 3.6915923327828423e-07, + "loss": 0.3954, + "num_input_tokens_seen": 5026873344, + "step": 2397 + }, + { + "epoch": 0.10794684024952536, + "grad_norm": 12.650426864624023, + "learning_rate": 3.6878964967946813e-07, + "loss": 0.3741, + "num_input_tokens_seen": 5028970496, + "step": 2398 + }, + { + "epoch": 0.10821806346623271, + "grad_norm": 15.329103469848633, + "learning_rate": 3.6842021191184636e-07, + "loss": 0.4692, + "num_input_tokens_seen": 5031067648, + "step": 2399 + }, + { + "epoch": 0.10848928668294006, + "grad_norm": 20.871158599853516, + "learning_rate": 3.6805092027272853e-07, + "loss": 0.8217, + "num_input_tokens_seen": 5033164800, + "step": 2400 + }, + { + "epoch": 0.00027122321670735016, + "grad_norm": 14.983903884887695, + "learning_rate": 3.6768177505930655e-07, + "loss": 0.4378, + "num_input_tokens_seen": 5035261952, + "step": 2401 + }, + { + "epoch": 0.0005424464334147003, + "grad_norm": 17.2396240234375, + "learning_rate": 3.67312776568654e-07, + "loss": 0.5583, + "num_input_tokens_seen": 5037359104, + "step": 2402 + }, + { + "epoch": 0.0008136696501220504, + "grad_norm": 18.06374740600586, + "learning_rate": 3.669439250977272e-07, + "loss": 0.6734, + "num_input_tokens_seen": 5039456256, + "step": 2403 + }, + { + "epoch": 0.0010848928668294006, + "grad_norm": 16.376707077026367, + "learning_rate": 3.665752209433637e-07, + "loss": 0.5364, + "num_input_tokens_seen": 5041553408, + "step": 2404 + }, + { + "epoch": 0.0013561160835367507, + "grad_norm": 12.664093017578125, + "learning_rate": 3.6620666440228254e-07, + "loss": 0.3744, + "num_input_tokens_seen": 5043650560, + "step": 2405 + }, + { + "epoch": 0.0016273393002441008, + "grad_norm": 11.83484935760498, + "learning_rate": 3.6583825577108397e-07, + "loss": 0.2879, + "num_input_tokens_seen": 5045747712, + "step": 2406 + }, + { + "epoch": 0.001898562516951451, + "grad_norm": 30.771093368530273, + "learning_rate": 3.654699953462494e-07, + "loss": 0.5672, + "num_input_tokens_seen": 5047844864, + "step": 2407 + }, + { + "epoch": 0.0021697857336588013, + "grad_norm": 12.490880012512207, + "learning_rate": 3.651018834241406e-07, + "loss": 0.352, + "num_input_tokens_seen": 5049942016, + "step": 2408 + }, + { + "epoch": 0.0024410089503661514, + "grad_norm": 12.38934326171875, + "learning_rate": 3.6473392030100014e-07, + "loss": 0.377, + "num_input_tokens_seen": 5052039168, + "step": 2409 + }, + { + "epoch": 0.0027122321670735015, + "grad_norm": 15.052350044250488, + "learning_rate": 3.6436610627295074e-07, + "loss": 0.4124, + "num_input_tokens_seen": 5054136320, + "step": 2410 + }, + { + "epoch": 0.0029834553837808516, + "grad_norm": 16.288164138793945, + "learning_rate": 3.639984416359949e-07, + "loss": 0.6888, + "num_input_tokens_seen": 5056233472, + "step": 2411 + }, + { + "epoch": 0.0032546786004882017, + "grad_norm": 10.213361740112305, + "learning_rate": 3.636309266860156e-07, + "loss": 0.3183, + "num_input_tokens_seen": 5058330624, + "step": 2412 + }, + { + "epoch": 0.003525901817195552, + "grad_norm": 15.612173080444336, + "learning_rate": 3.6326356171877483e-07, + "loss": 0.5927, + "num_input_tokens_seen": 5060427776, + "step": 2413 + }, + { + "epoch": 0.003797125033902902, + "grad_norm": 15.306659698486328, + "learning_rate": 3.6289634702991343e-07, + "loss": 0.3537, + "num_input_tokens_seen": 5062524928, + "step": 2414 + }, + { + "epoch": 0.0040683482506102524, + "grad_norm": 13.14424991607666, + "learning_rate": 3.625292829149521e-07, + "loss": 0.2652, + "num_input_tokens_seen": 5064622080, + "step": 2415 + }, + { + "epoch": 0.0043395714673176026, + "grad_norm": 22.333417892456055, + "learning_rate": 3.6216236966929015e-07, + "loss": 0.8254, + "num_input_tokens_seen": 5066719232, + "step": 2416 + }, + { + "epoch": 0.004610794684024953, + "grad_norm": 15.718981742858887, + "learning_rate": 3.6179560758820527e-07, + "loss": 0.3925, + "num_input_tokens_seen": 5068816384, + "step": 2417 + }, + { + "epoch": 0.004882017900732303, + "grad_norm": 16.70728302001953, + "learning_rate": 3.61428996966854e-07, + "loss": 0.589, + "num_input_tokens_seen": 5070913536, + "step": 2418 + }, + { + "epoch": 0.005153241117439653, + "grad_norm": 12.905607223510742, + "learning_rate": 3.610625381002701e-07, + "loss": 0.3304, + "num_input_tokens_seen": 5073010688, + "step": 2419 + }, + { + "epoch": 0.005424464334147003, + "grad_norm": 15.081153869628906, + "learning_rate": 3.606962312833659e-07, + "loss": 0.4949, + "num_input_tokens_seen": 5075107840, + "step": 2420 + }, + { + "epoch": 0.005695687550854353, + "grad_norm": 13.203760147094727, + "learning_rate": 3.603300768109311e-07, + "loss": 0.3444, + "num_input_tokens_seen": 5077204992, + "step": 2421 + }, + { + "epoch": 0.005966910767561703, + "grad_norm": 21.240129470825195, + "learning_rate": 3.5996407497763305e-07, + "loss": 0.6882, + "num_input_tokens_seen": 5079302144, + "step": 2422 + }, + { + "epoch": 0.006238133984269053, + "grad_norm": 13.102483749389648, + "learning_rate": 3.5959822607801617e-07, + "loss": 0.3807, + "num_input_tokens_seen": 5081399296, + "step": 2423 + }, + { + "epoch": 0.006509357200976403, + "grad_norm": 17.356565475463867, + "learning_rate": 3.592325304065018e-07, + "loss": 0.5024, + "num_input_tokens_seen": 5083496448, + "step": 2424 + }, + { + "epoch": 0.0067805804176837535, + "grad_norm": 38.09556198120117, + "learning_rate": 3.588669882573875e-07, + "loss": 0.5346, + "num_input_tokens_seen": 5085593600, + "step": 2425 + }, + { + "epoch": 0.007051803634391104, + "grad_norm": 24.49935531616211, + "learning_rate": 3.5850159992484787e-07, + "loss": 0.9886, + "num_input_tokens_seen": 5087690752, + "step": 2426 + }, + { + "epoch": 0.007323026851098454, + "grad_norm": 15.095088005065918, + "learning_rate": 3.581363657029336e-07, + "loss": 0.3287, + "num_input_tokens_seen": 5089787904, + "step": 2427 + }, + { + "epoch": 0.007594250067805804, + "grad_norm": 8.799110412597656, + "learning_rate": 3.5777128588557126e-07, + "loss": 0.2119, + "num_input_tokens_seen": 5091885056, + "step": 2428 + }, + { + "epoch": 0.007865473284513154, + "grad_norm": 15.774507522583008, + "learning_rate": 3.574063607665633e-07, + "loss": 0.6325, + "num_input_tokens_seen": 5093982208, + "step": 2429 + }, + { + "epoch": 0.008136696501220505, + "grad_norm": 20.946643829345703, + "learning_rate": 3.570415906395873e-07, + "loss": 0.5039, + "num_input_tokens_seen": 5096079360, + "step": 2430 + }, + { + "epoch": 0.008407919717927854, + "grad_norm": 14.345913887023926, + "learning_rate": 3.5667697579819655e-07, + "loss": 0.4575, + "num_input_tokens_seen": 5098176512, + "step": 2431 + }, + { + "epoch": 0.008679142934635205, + "grad_norm": 14.178055763244629, + "learning_rate": 3.563125165358193e-07, + "loss": 0.5042, + "num_input_tokens_seen": 5100273664, + "step": 2432 + }, + { + "epoch": 0.008950366151342554, + "grad_norm": 17.350387573242188, + "learning_rate": 3.559482131457583e-07, + "loss": 0.5089, + "num_input_tokens_seen": 5102370816, + "step": 2433 + }, + { + "epoch": 0.009221589368049905, + "grad_norm": 18.10607147216797, + "learning_rate": 3.5558406592119115e-07, + "loss": 0.4904, + "num_input_tokens_seen": 5104467968, + "step": 2434 + }, + { + "epoch": 0.009492812584757255, + "grad_norm": 15.926411628723145, + "learning_rate": 3.552200751551697e-07, + "loss": 0.5084, + "num_input_tokens_seen": 5106565120, + "step": 2435 + }, + { + "epoch": 0.009764035801464606, + "grad_norm": 17.589426040649414, + "learning_rate": 3.548562411406201e-07, + "loss": 0.4039, + "num_input_tokens_seen": 5108662272, + "step": 2436 + }, + { + "epoch": 0.010035259018171955, + "grad_norm": 15.575359344482422, + "learning_rate": 3.544925641703413e-07, + "loss": 0.4175, + "num_input_tokens_seen": 5110759424, + "step": 2437 + }, + { + "epoch": 0.010306482234879306, + "grad_norm": 21.171030044555664, + "learning_rate": 3.5412904453700754e-07, + "loss": 0.6827, + "num_input_tokens_seen": 5112856576, + "step": 2438 + }, + { + "epoch": 0.010577705451586655, + "grad_norm": 18.639638900756836, + "learning_rate": 3.537656825331653e-07, + "loss": 0.6128, + "num_input_tokens_seen": 5114953728, + "step": 2439 + }, + { + "epoch": 0.010848928668294006, + "grad_norm": 19.739368438720703, + "learning_rate": 3.534024784512345e-07, + "loss": 0.5211, + "num_input_tokens_seen": 5117050880, + "step": 2440 + }, + { + "epoch": 0.011120151885001357, + "grad_norm": 14.70162296295166, + "learning_rate": 3.5303943258350813e-07, + "loss": 0.3884, + "num_input_tokens_seen": 5119148032, + "step": 2441 + }, + { + "epoch": 0.011391375101708706, + "grad_norm": 15.77631950378418, + "learning_rate": 3.526765452221512e-07, + "loss": 0.3623, + "num_input_tokens_seen": 5121245184, + "step": 2442 + }, + { + "epoch": 0.011662598318416057, + "grad_norm": 16.529884338378906, + "learning_rate": 3.523138166592021e-07, + "loss": 0.4567, + "num_input_tokens_seen": 5123342336, + "step": 2443 + }, + { + "epoch": 0.011933821535123406, + "grad_norm": 19.871736526489258, + "learning_rate": 3.5195124718657075e-07, + "loss": 0.7492, + "num_input_tokens_seen": 5125439488, + "step": 2444 + }, + { + "epoch": 0.012205044751830757, + "grad_norm": 9.847620964050293, + "learning_rate": 3.5158883709603946e-07, + "loss": 0.2599, + "num_input_tokens_seen": 5127536640, + "step": 2445 + }, + { + "epoch": 0.012476267968538107, + "grad_norm": 14.648107528686523, + "learning_rate": 3.5122658667926177e-07, + "loss": 0.5326, + "num_input_tokens_seen": 5129633792, + "step": 2446 + }, + { + "epoch": 0.012747491185245458, + "grad_norm": 14.89171314239502, + "learning_rate": 3.5086449622776346e-07, + "loss": 0.583, + "num_input_tokens_seen": 5131730944, + "step": 2447 + }, + { + "epoch": 0.013018714401952807, + "grad_norm": 14.507434844970703, + "learning_rate": 3.505025660329408e-07, + "loss": 0.4423, + "num_input_tokens_seen": 5133828096, + "step": 2448 + }, + { + "epoch": 0.013289937618660158, + "grad_norm": 13.03704833984375, + "learning_rate": 3.5014079638606164e-07, + "loss": 0.3409, + "num_input_tokens_seen": 5135925248, + "step": 2449 + }, + { + "epoch": 0.013561160835367507, + "grad_norm": 12.08354377746582, + "learning_rate": 3.497791875782643e-07, + "loss": 0.3119, + "num_input_tokens_seen": 5138022400, + "step": 2450 + }, + { + "epoch": 0.013832384052074858, + "grad_norm": 14.133995056152344, + "learning_rate": 3.4941773990055777e-07, + "loss": 0.404, + "num_input_tokens_seen": 5140119552, + "step": 2451 + }, + { + "epoch": 0.014103607268782207, + "grad_norm": 19.439821243286133, + "learning_rate": 3.490564536438215e-07, + "loss": 0.6819, + "num_input_tokens_seen": 5142216704, + "step": 2452 + }, + { + "epoch": 0.014374830485489558, + "grad_norm": 14.30314826965332, + "learning_rate": 3.4869532909880485e-07, + "loss": 0.4549, + "num_input_tokens_seen": 5144313856, + "step": 2453 + }, + { + "epoch": 0.014646053702196907, + "grad_norm": 15.834941864013672, + "learning_rate": 3.483343665561271e-07, + "loss": 0.3616, + "num_input_tokens_seen": 5146411008, + "step": 2454 + }, + { + "epoch": 0.014917276918904258, + "grad_norm": 17.570751190185547, + "learning_rate": 3.479735663062773e-07, + "loss": 0.4955, + "num_input_tokens_seen": 5148508160, + "step": 2455 + }, + { + "epoch": 0.015188500135611608, + "grad_norm": 11.63994026184082, + "learning_rate": 3.4761292863961354e-07, + "loss": 0.357, + "num_input_tokens_seen": 5150605312, + "step": 2456 + }, + { + "epoch": 0.015459723352318959, + "grad_norm": 13.828446388244629, + "learning_rate": 3.4725245384636347e-07, + "loss": 0.3693, + "num_input_tokens_seen": 5152702464, + "step": 2457 + }, + { + "epoch": 0.015730946569026308, + "grad_norm": 13.007326126098633, + "learning_rate": 3.4689214221662364e-07, + "loss": 0.2658, + "num_input_tokens_seen": 5154799616, + "step": 2458 + }, + { + "epoch": 0.01600216978573366, + "grad_norm": 15.117969512939453, + "learning_rate": 3.465319940403587e-07, + "loss": 0.5309, + "num_input_tokens_seen": 5156896768, + "step": 2459 + }, + { + "epoch": 0.01627339300244101, + "grad_norm": 12.923035621643066, + "learning_rate": 3.4617200960740247e-07, + "loss": 0.3563, + "num_input_tokens_seen": 5158993920, + "step": 2460 + }, + { + "epoch": 0.01654461621914836, + "grad_norm": 10.716797828674316, + "learning_rate": 3.4581218920745663e-07, + "loss": 0.3419, + "num_input_tokens_seen": 5161091072, + "step": 2461 + }, + { + "epoch": 0.016815839435855708, + "grad_norm": 14.895060539245605, + "learning_rate": 3.454525331300908e-07, + "loss": 0.3487, + "num_input_tokens_seen": 5163188224, + "step": 2462 + }, + { + "epoch": 0.01708706265256306, + "grad_norm": 17.449199676513672, + "learning_rate": 3.450930416647429e-07, + "loss": 0.4316, + "num_input_tokens_seen": 5165285376, + "step": 2463 + }, + { + "epoch": 0.01735828586927041, + "grad_norm": 13.589836120605469, + "learning_rate": 3.4473371510071795e-07, + "loss": 0.458, + "num_input_tokens_seen": 5167382528, + "step": 2464 + }, + { + "epoch": 0.01762950908597776, + "grad_norm": 17.81792640686035, + "learning_rate": 3.4437455372718795e-07, + "loss": 0.6118, + "num_input_tokens_seen": 5169479680, + "step": 2465 + }, + { + "epoch": 0.01790073230268511, + "grad_norm": 15.582667350769043, + "learning_rate": 3.440155578331925e-07, + "loss": 0.4143, + "num_input_tokens_seen": 5171576832, + "step": 2466 + }, + { + "epoch": 0.01817195551939246, + "grad_norm": 13.768962860107422, + "learning_rate": 3.4365672770763783e-07, + "loss": 0.3377, + "num_input_tokens_seen": 5173673984, + "step": 2467 + }, + { + "epoch": 0.01844317873609981, + "grad_norm": 14.987922668457031, + "learning_rate": 3.432980636392967e-07, + "loss": 0.4133, + "num_input_tokens_seen": 5175771136, + "step": 2468 + }, + { + "epoch": 0.01871440195280716, + "grad_norm": 14.550777435302734, + "learning_rate": 3.429395659168084e-07, + "loss": 0.3833, + "num_input_tokens_seen": 5177868288, + "step": 2469 + }, + { + "epoch": 0.01898562516951451, + "grad_norm": 17.40359115600586, + "learning_rate": 3.425812348286782e-07, + "loss": 0.5748, + "num_input_tokens_seen": 5179965440, + "step": 2470 + }, + { + "epoch": 0.01925684838622186, + "grad_norm": 21.7893123626709, + "learning_rate": 3.422230706632774e-07, + "loss": 0.8634, + "num_input_tokens_seen": 5182062592, + "step": 2471 + }, + { + "epoch": 0.01952807160292921, + "grad_norm": 17.623117446899414, + "learning_rate": 3.418650737088427e-07, + "loss": 0.4959, + "num_input_tokens_seen": 5184159744, + "step": 2472 + }, + { + "epoch": 0.019799294819636562, + "grad_norm": 15.7328519821167, + "learning_rate": 3.415072442534767e-07, + "loss": 0.3269, + "num_input_tokens_seen": 5186256896, + "step": 2473 + }, + { + "epoch": 0.02007051803634391, + "grad_norm": 14.757015228271484, + "learning_rate": 3.411495825851467e-07, + "loss": 0.3758, + "num_input_tokens_seen": 5188354048, + "step": 2474 + }, + { + "epoch": 0.02034174125305126, + "grad_norm": 12.561208724975586, + "learning_rate": 3.4079208899168545e-07, + "loss": 0.3777, + "num_input_tokens_seen": 5190451200, + "step": 2475 + }, + { + "epoch": 0.02061296446975861, + "grad_norm": 20.02168083190918, + "learning_rate": 3.404347637607899e-07, + "loss": 0.4442, + "num_input_tokens_seen": 5192548352, + "step": 2476 + }, + { + "epoch": 0.020884187686465962, + "grad_norm": 16.987884521484375, + "learning_rate": 3.400776071800219e-07, + "loss": 0.5238, + "num_input_tokens_seen": 5194645504, + "step": 2477 + }, + { + "epoch": 0.02115541090317331, + "grad_norm": 16.009906768798828, + "learning_rate": 3.3972061953680734e-07, + "loss": 0.2866, + "num_input_tokens_seen": 5196742656, + "step": 2478 + }, + { + "epoch": 0.02142663411988066, + "grad_norm": 15.513143539428711, + "learning_rate": 3.3936380111843666e-07, + "loss": 0.4394, + "num_input_tokens_seen": 5198839808, + "step": 2479 + }, + { + "epoch": 0.021697857336588012, + "grad_norm": 12.317339897155762, + "learning_rate": 3.390071522120635e-07, + "loss": 0.3811, + "num_input_tokens_seen": 5200936960, + "step": 2480 + }, + { + "epoch": 0.021969080553295363, + "grad_norm": 13.320518493652344, + "learning_rate": 3.3865067310470554e-07, + "loss": 0.3729, + "num_input_tokens_seen": 5203034112, + "step": 2481 + }, + { + "epoch": 0.022240303770002714, + "grad_norm": 11.673955917358398, + "learning_rate": 3.3829436408324316e-07, + "loss": 0.3577, + "num_input_tokens_seen": 5205131264, + "step": 2482 + }, + { + "epoch": 0.02251152698671006, + "grad_norm": 15.189332962036133, + "learning_rate": 3.3793822543442074e-07, + "loss": 0.4539, + "num_input_tokens_seen": 5207228416, + "step": 2483 + }, + { + "epoch": 0.022782750203417412, + "grad_norm": 13.745153427124023, + "learning_rate": 3.3758225744484483e-07, + "loss": 0.3885, + "num_input_tokens_seen": 5209325568, + "step": 2484 + }, + { + "epoch": 0.023053973420124763, + "grad_norm": 13.338912963867188, + "learning_rate": 3.372264604009851e-07, + "loss": 0.4417, + "num_input_tokens_seen": 5211422720, + "step": 2485 + }, + { + "epoch": 0.023325196636832114, + "grad_norm": 15.356630325317383, + "learning_rate": 3.3687083458917344e-07, + "loss": 0.4507, + "num_input_tokens_seen": 5213519872, + "step": 2486 + }, + { + "epoch": 0.023596419853539462, + "grad_norm": 16.816343307495117, + "learning_rate": 3.3651538029560377e-07, + "loss": 0.4607, + "num_input_tokens_seen": 5215617024, + "step": 2487 + }, + { + "epoch": 0.023867643070246813, + "grad_norm": 13.717020034790039, + "learning_rate": 3.361600978063325e-07, + "loss": 0.4352, + "num_input_tokens_seen": 5217714176, + "step": 2488 + }, + { + "epoch": 0.024138866286954164, + "grad_norm": 17.89053726196289, + "learning_rate": 3.358049874072771e-07, + "loss": 0.5094, + "num_input_tokens_seen": 5219811328, + "step": 2489 + }, + { + "epoch": 0.024410089503661515, + "grad_norm": 13.972254753112793, + "learning_rate": 3.3545004938421734e-07, + "loss": 0.4116, + "num_input_tokens_seen": 5221908480, + "step": 2490 + }, + { + "epoch": 0.024681312720368862, + "grad_norm": 20.813188552856445, + "learning_rate": 3.3509528402279357e-07, + "loss": 0.7909, + "num_input_tokens_seen": 5224005632, + "step": 2491 + }, + { + "epoch": 0.024952535937076213, + "grad_norm": 23.339984893798828, + "learning_rate": 3.347406916085074e-07, + "loss": 0.9871, + "num_input_tokens_seen": 5226102784, + "step": 2492 + }, + { + "epoch": 0.025223759153783564, + "grad_norm": 19.53771209716797, + "learning_rate": 3.3438627242672164e-07, + "loss": 0.7506, + "num_input_tokens_seen": 5228199936, + "step": 2493 + }, + { + "epoch": 0.025494982370490915, + "grad_norm": 13.463472366333008, + "learning_rate": 3.3403202676265875e-07, + "loss": 0.4105, + "num_input_tokens_seen": 5230297088, + "step": 2494 + }, + { + "epoch": 0.025766205587198263, + "grad_norm": 11.80983829498291, + "learning_rate": 3.336779549014026e-07, + "loss": 0.2789, + "num_input_tokens_seen": 5232394240, + "step": 2495 + }, + { + "epoch": 0.026037428803905614, + "grad_norm": 11.199295997619629, + "learning_rate": 3.333240571278968e-07, + "loss": 0.3236, + "num_input_tokens_seen": 5234491392, + "step": 2496 + }, + { + "epoch": 0.026308652020612965, + "grad_norm": 11.20341968536377, + "learning_rate": 3.3297033372694473e-07, + "loss": 0.2513, + "num_input_tokens_seen": 5236588544, + "step": 2497 + }, + { + "epoch": 0.026579875237320316, + "grad_norm": 19.574689865112305, + "learning_rate": 3.3261678498320954e-07, + "loss": 0.8343, + "num_input_tokens_seen": 5238685696, + "step": 2498 + }, + { + "epoch": 0.026851098454027666, + "grad_norm": 17.769323348999023, + "learning_rate": 3.3226341118121367e-07, + "loss": 0.6071, + "num_input_tokens_seen": 5240782848, + "step": 2499 + }, + { + "epoch": 0.027122321670735014, + "grad_norm": 10.659870147705078, + "learning_rate": 3.319102126053389e-07, + "loss": 0.3469, + "num_input_tokens_seen": 5242880000, + "step": 2500 + }, + { + "epoch": 0.027393544887442365, + "grad_norm": 14.309873580932617, + "learning_rate": 3.315571895398261e-07, + "loss": 0.2899, + "num_input_tokens_seen": 5244977152, + "step": 2501 + }, + { + "epoch": 0.027664768104149716, + "grad_norm": 12.785883903503418, + "learning_rate": 3.312043422687749e-07, + "loss": 0.3751, + "num_input_tokens_seen": 5247074304, + "step": 2502 + }, + { + "epoch": 0.027935991320857067, + "grad_norm": 9.752639770507812, + "learning_rate": 3.3085167107614297e-07, + "loss": 0.2094, + "num_input_tokens_seen": 5249171456, + "step": 2503 + }, + { + "epoch": 0.028207214537564414, + "grad_norm": 20.031442642211914, + "learning_rate": 3.3049917624574737e-07, + "loss": 0.6936, + "num_input_tokens_seen": 5251268608, + "step": 2504 + }, + { + "epoch": 0.028478437754271765, + "grad_norm": 12.938445091247559, + "learning_rate": 3.301468580612619e-07, + "loss": 0.3843, + "num_input_tokens_seen": 5253365760, + "step": 2505 + }, + { + "epoch": 0.028749660970979116, + "grad_norm": 15.744908332824707, + "learning_rate": 3.2979471680621903e-07, + "loss": 0.3618, + "num_input_tokens_seen": 5255462912, + "step": 2506 + }, + { + "epoch": 0.029020884187686467, + "grad_norm": 18.101648330688477, + "learning_rate": 3.2944275276400857e-07, + "loss": 0.6241, + "num_input_tokens_seen": 5257560064, + "step": 2507 + }, + { + "epoch": 0.029292107404393815, + "grad_norm": 15.829540252685547, + "learning_rate": 3.290909662178779e-07, + "loss": 0.4732, + "num_input_tokens_seen": 5259657216, + "step": 2508 + }, + { + "epoch": 0.029563330621101166, + "grad_norm": 10.271330833435059, + "learning_rate": 3.2873935745093145e-07, + "loss": 0.1894, + "num_input_tokens_seen": 5261754368, + "step": 2509 + }, + { + "epoch": 0.029834553837808517, + "grad_norm": 14.364543914794922, + "learning_rate": 3.283879267461305e-07, + "loss": 0.4071, + "num_input_tokens_seen": 5263851520, + "step": 2510 + }, + { + "epoch": 0.030105777054515868, + "grad_norm": 14.449692726135254, + "learning_rate": 3.280366743862931e-07, + "loss": 0.2794, + "num_input_tokens_seen": 5265948672, + "step": 2511 + }, + { + "epoch": 0.030377000271223215, + "grad_norm": 14.511991500854492, + "learning_rate": 3.276856006540939e-07, + "loss": 0.4112, + "num_input_tokens_seen": 5268045824, + "step": 2512 + }, + { + "epoch": 0.030648223487930566, + "grad_norm": 19.19399642944336, + "learning_rate": 3.2733470583206357e-07, + "loss": 0.7822, + "num_input_tokens_seen": 5270142976, + "step": 2513 + }, + { + "epoch": 0.030919446704637917, + "grad_norm": 23.72905158996582, + "learning_rate": 3.2698399020258895e-07, + "loss": 0.7289, + "num_input_tokens_seen": 5272240128, + "step": 2514 + }, + { + "epoch": 0.031190669921345268, + "grad_norm": 12.964669227600098, + "learning_rate": 3.266334540479128e-07, + "loss": 0.3558, + "num_input_tokens_seen": 5274337280, + "step": 2515 + }, + { + "epoch": 0.031461893138052616, + "grad_norm": 16.30192756652832, + "learning_rate": 3.262830976501329e-07, + "loss": 0.4661, + "num_input_tokens_seen": 5276434432, + "step": 2516 + }, + { + "epoch": 0.03173311635475997, + "grad_norm": 17.11813735961914, + "learning_rate": 3.2593292129120295e-07, + "loss": 0.4023, + "num_input_tokens_seen": 5278531584, + "step": 2517 + }, + { + "epoch": 0.03200433957146732, + "grad_norm": 13.747211456298828, + "learning_rate": 3.2558292525293156e-07, + "loss": 0.396, + "num_input_tokens_seen": 5280628736, + "step": 2518 + }, + { + "epoch": 0.032275562788174665, + "grad_norm": 15.1743803024292, + "learning_rate": 3.2523310981698213e-07, + "loss": 0.4891, + "num_input_tokens_seen": 5282725888, + "step": 2519 + }, + { + "epoch": 0.03254678600488202, + "grad_norm": 21.062551498413086, + "learning_rate": 3.24883475264873e-07, + "loss": 0.7359, + "num_input_tokens_seen": 5284823040, + "step": 2520 + }, + { + "epoch": 0.03281800922158937, + "grad_norm": 12.402223587036133, + "learning_rate": 3.2453402187797684e-07, + "loss": 0.5071, + "num_input_tokens_seen": 5286920192, + "step": 2521 + }, + { + "epoch": 0.03308923243829672, + "grad_norm": 16.109939575195312, + "learning_rate": 3.241847499375201e-07, + "loss": 0.4946, + "num_input_tokens_seen": 5289017344, + "step": 2522 + }, + { + "epoch": 0.03336045565500407, + "grad_norm": 10.402777671813965, + "learning_rate": 3.238356597245837e-07, + "loss": 0.2449, + "num_input_tokens_seen": 5291114496, + "step": 2523 + }, + { + "epoch": 0.033631678871711417, + "grad_norm": 11.929893493652344, + "learning_rate": 3.2348675152010217e-07, + "loss": 0.4098, + "num_input_tokens_seen": 5293211648, + "step": 2524 + }, + { + "epoch": 0.03390290208841877, + "grad_norm": 16.611042022705078, + "learning_rate": 3.2313802560486353e-07, + "loss": 0.5525, + "num_input_tokens_seen": 5295308800, + "step": 2525 + }, + { + "epoch": 0.03417412530512612, + "grad_norm": 18.180625915527344, + "learning_rate": 3.2278948225950916e-07, + "loss": 0.654, + "num_input_tokens_seen": 5297405952, + "step": 2526 + }, + { + "epoch": 0.034445348521833466, + "grad_norm": 13.363505363464355, + "learning_rate": 3.2244112176453343e-07, + "loss": 0.4729, + "num_input_tokens_seen": 5299503104, + "step": 2527 + }, + { + "epoch": 0.03471657173854082, + "grad_norm": 20.741127014160156, + "learning_rate": 3.2209294440028366e-07, + "loss": 0.7143, + "num_input_tokens_seen": 5301600256, + "step": 2528 + }, + { + "epoch": 0.03498779495524817, + "grad_norm": 13.14611530303955, + "learning_rate": 3.2174495044695973e-07, + "loss": 0.3495, + "num_input_tokens_seen": 5303697408, + "step": 2529 + }, + { + "epoch": 0.03525901817195552, + "grad_norm": 16.342422485351562, + "learning_rate": 3.2139714018461396e-07, + "loss": 0.5414, + "num_input_tokens_seen": 5305794560, + "step": 2530 + }, + { + "epoch": 0.03553024138866287, + "grad_norm": 16.03512191772461, + "learning_rate": 3.2104951389315073e-07, + "loss": 0.5638, + "num_input_tokens_seen": 5307891712, + "step": 2531 + }, + { + "epoch": 0.03580146460537022, + "grad_norm": 17.44066047668457, + "learning_rate": 3.207020718523266e-07, + "loss": 0.6175, + "num_input_tokens_seen": 5309988864, + "step": 2532 + }, + { + "epoch": 0.03607268782207757, + "grad_norm": 13.835807800292969, + "learning_rate": 3.2035481434174966e-07, + "loss": 0.4456, + "num_input_tokens_seen": 5312086016, + "step": 2533 + }, + { + "epoch": 0.03634391103878492, + "grad_norm": 16.608179092407227, + "learning_rate": 3.200077416408794e-07, + "loss": 0.4694, + "num_input_tokens_seen": 5314183168, + "step": 2534 + }, + { + "epoch": 0.03661513425549227, + "grad_norm": 23.50665855407715, + "learning_rate": 3.196608540290266e-07, + "loss": 0.597, + "num_input_tokens_seen": 5316280320, + "step": 2535 + }, + { + "epoch": 0.03688635747219962, + "grad_norm": 19.46283721923828, + "learning_rate": 3.193141517853536e-07, + "loss": 0.5605, + "num_input_tokens_seen": 5318377472, + "step": 2536 + }, + { + "epoch": 0.03715758068890697, + "grad_norm": 14.471436500549316, + "learning_rate": 3.1896763518887305e-07, + "loss": 0.4885, + "num_input_tokens_seen": 5320474624, + "step": 2537 + }, + { + "epoch": 0.03742880390561432, + "grad_norm": 22.120628356933594, + "learning_rate": 3.186213045184484e-07, + "loss": 0.8826, + "num_input_tokens_seen": 5322571776, + "step": 2538 + }, + { + "epoch": 0.03770002712232167, + "grad_norm": 11.51758861541748, + "learning_rate": 3.1827516005279306e-07, + "loss": 0.3493, + "num_input_tokens_seen": 5324668928, + "step": 2539 + }, + { + "epoch": 0.03797125033902902, + "grad_norm": 14.106184959411621, + "learning_rate": 3.1792920207047114e-07, + "loss": 0.4086, + "num_input_tokens_seen": 5326766080, + "step": 2540 + }, + { + "epoch": 0.03824247355573637, + "grad_norm": 15.168310165405273, + "learning_rate": 3.175834308498964e-07, + "loss": 0.4476, + "num_input_tokens_seen": 5328863232, + "step": 2541 + }, + { + "epoch": 0.03851369677244372, + "grad_norm": 16.02938461303711, + "learning_rate": 3.172378466693325e-07, + "loss": 0.3896, + "num_input_tokens_seen": 5330960384, + "step": 2542 + }, + { + "epoch": 0.038784919989151075, + "grad_norm": 21.418581008911133, + "learning_rate": 3.168924498068923e-07, + "loss": 0.9796, + "num_input_tokens_seen": 5333057536, + "step": 2543 + }, + { + "epoch": 0.03905614320585842, + "grad_norm": 16.576860427856445, + "learning_rate": 3.1654724054053805e-07, + "loss": 0.4524, + "num_input_tokens_seen": 5335154688, + "step": 2544 + }, + { + "epoch": 0.03932736642256577, + "grad_norm": 14.02090072631836, + "learning_rate": 3.1620221914808115e-07, + "loss": 0.4474, + "num_input_tokens_seen": 5337251840, + "step": 2545 + }, + { + "epoch": 0.039598589639273124, + "grad_norm": 14.15399169921875, + "learning_rate": 3.1585738590718157e-07, + "loss": 0.3879, + "num_input_tokens_seen": 5339348992, + "step": 2546 + }, + { + "epoch": 0.03986981285598047, + "grad_norm": 21.25953483581543, + "learning_rate": 3.1551274109534805e-07, + "loss": 0.6324, + "num_input_tokens_seen": 5341446144, + "step": 2547 + }, + { + "epoch": 0.04014103607268782, + "grad_norm": 17.72227668762207, + "learning_rate": 3.151682849899376e-07, + "loss": 0.7032, + "num_input_tokens_seen": 5343543296, + "step": 2548 + }, + { + "epoch": 0.040412259289395173, + "grad_norm": 14.34379768371582, + "learning_rate": 3.148240178681553e-07, + "loss": 0.44, + "num_input_tokens_seen": 5345640448, + "step": 2549 + }, + { + "epoch": 0.04068348250610252, + "grad_norm": 18.565998077392578, + "learning_rate": 3.1447994000705456e-07, + "loss": 0.6165, + "num_input_tokens_seen": 5347737600, + "step": 2550 + }, + { + "epoch": 0.040954705722809875, + "grad_norm": 11.777302742004395, + "learning_rate": 3.141360516835356e-07, + "loss": 0.3904, + "num_input_tokens_seen": 5349834752, + "step": 2551 + }, + { + "epoch": 0.04122592893951722, + "grad_norm": 7.871033191680908, + "learning_rate": 3.1379235317434703e-07, + "loss": 0.1931, + "num_input_tokens_seen": 5351931904, + "step": 2552 + }, + { + "epoch": 0.04149715215622457, + "grad_norm": 13.070982933044434, + "learning_rate": 3.134488447560843e-07, + "loss": 0.3965, + "num_input_tokens_seen": 5354029056, + "step": 2553 + }, + { + "epoch": 0.041768375372931925, + "grad_norm": 19.49721908569336, + "learning_rate": 3.1310552670518987e-07, + "loss": 0.8075, + "num_input_tokens_seen": 5356126208, + "step": 2554 + }, + { + "epoch": 0.04203959858963927, + "grad_norm": 15.62477970123291, + "learning_rate": 3.127623992979532e-07, + "loss": 0.4351, + "num_input_tokens_seen": 5358223360, + "step": 2555 + }, + { + "epoch": 0.04231082180634662, + "grad_norm": 13.880343437194824, + "learning_rate": 3.124194628105098e-07, + "loss": 0.3508, + "num_input_tokens_seen": 5360320512, + "step": 2556 + }, + { + "epoch": 0.042582045023053974, + "grad_norm": 17.306838989257812, + "learning_rate": 3.120767175188422e-07, + "loss": 0.5777, + "num_input_tokens_seen": 5362417664, + "step": 2557 + }, + { + "epoch": 0.04285326823976132, + "grad_norm": 18.195343017578125, + "learning_rate": 3.1173416369877864e-07, + "loss": 0.7024, + "num_input_tokens_seen": 5364514816, + "step": 2558 + }, + { + "epoch": 0.043124491456468676, + "grad_norm": 16.49274444580078, + "learning_rate": 3.1139180162599346e-07, + "loss": 0.519, + "num_input_tokens_seen": 5366611968, + "step": 2559 + }, + { + "epoch": 0.043395714673176024, + "grad_norm": 12.536412239074707, + "learning_rate": 3.110496315760065e-07, + "loss": 0.3714, + "num_input_tokens_seen": 5368709120, + "step": 2560 + }, + { + "epoch": 0.04366693788988337, + "grad_norm": 14.73096752166748, + "learning_rate": 3.107076538241835e-07, + "loss": 0.5109, + "num_input_tokens_seen": 5370806272, + "step": 2561 + }, + { + "epoch": 0.043938161106590726, + "grad_norm": 12.924612998962402, + "learning_rate": 3.103658686457349e-07, + "loss": 0.4013, + "num_input_tokens_seen": 5372903424, + "step": 2562 + }, + { + "epoch": 0.04420938432329807, + "grad_norm": 13.50550365447998, + "learning_rate": 3.1002427631571646e-07, + "loss": 0.3039, + "num_input_tokens_seen": 5375000576, + "step": 2563 + }, + { + "epoch": 0.04448060754000543, + "grad_norm": 12.640801429748535, + "learning_rate": 3.0968287710902866e-07, + "loss": 0.3334, + "num_input_tokens_seen": 5377097728, + "step": 2564 + }, + { + "epoch": 0.044751830756712775, + "grad_norm": 15.517276763916016, + "learning_rate": 3.0934167130041666e-07, + "loss": 0.6337, + "num_input_tokens_seen": 5379194880, + "step": 2565 + }, + { + "epoch": 0.04502305397342012, + "grad_norm": 16.423845291137695, + "learning_rate": 3.090006591644698e-07, + "loss": 0.637, + "num_input_tokens_seen": 5381292032, + "step": 2566 + }, + { + "epoch": 0.04529427719012748, + "grad_norm": 13.07593822479248, + "learning_rate": 3.0865984097562183e-07, + "loss": 0.2506, + "num_input_tokens_seen": 5383389184, + "step": 2567 + }, + { + "epoch": 0.045565500406834825, + "grad_norm": 21.440753936767578, + "learning_rate": 3.083192170081501e-07, + "loss": 0.5519, + "num_input_tokens_seen": 5385486336, + "step": 2568 + }, + { + "epoch": 0.04583672362354217, + "grad_norm": 14.802675247192383, + "learning_rate": 3.079787875361759e-07, + "loss": 0.476, + "num_input_tokens_seen": 5387583488, + "step": 2569 + }, + { + "epoch": 0.04610794684024953, + "grad_norm": 17.272769927978516, + "learning_rate": 3.0763855283366386e-07, + "loss": 0.4442, + "num_input_tokens_seen": 5389680640, + "step": 2570 + }, + { + "epoch": 0.046379170056956874, + "grad_norm": 10.331268310546875, + "learning_rate": 3.07298513174422e-07, + "loss": 0.3056, + "num_input_tokens_seen": 5391777792, + "step": 2571 + }, + { + "epoch": 0.04665039327366423, + "grad_norm": 15.530325889587402, + "learning_rate": 3.0695866883210143e-07, + "loss": 0.504, + "num_input_tokens_seen": 5393874944, + "step": 2572 + }, + { + "epoch": 0.046921616490371576, + "grad_norm": 20.427730560302734, + "learning_rate": 3.0661902008019556e-07, + "loss": 0.7343, + "num_input_tokens_seen": 5395972096, + "step": 2573 + }, + { + "epoch": 0.047192839707078924, + "grad_norm": 22.61469078063965, + "learning_rate": 3.062795671920411e-07, + "loss": 0.8057, + "num_input_tokens_seen": 5398069248, + "step": 2574 + }, + { + "epoch": 0.04746406292378628, + "grad_norm": 11.421932220458984, + "learning_rate": 3.059403104408166e-07, + "loss": 0.3017, + "num_input_tokens_seen": 5400166400, + "step": 2575 + }, + { + "epoch": 0.047735286140493625, + "grad_norm": 11.474665641784668, + "learning_rate": 3.0560125009954296e-07, + "loss": 0.2996, + "num_input_tokens_seen": 5402263552, + "step": 2576 + }, + { + "epoch": 0.04800650935720097, + "grad_norm": 26.75455665588379, + "learning_rate": 3.0526238644108335e-07, + "loss": 0.5857, + "num_input_tokens_seen": 5404360704, + "step": 2577 + }, + { + "epoch": 0.04827773257390833, + "grad_norm": 15.098592758178711, + "learning_rate": 3.049237197381424e-07, + "loss": 0.5197, + "num_input_tokens_seen": 5406457856, + "step": 2578 + }, + { + "epoch": 0.048548955790615675, + "grad_norm": 15.30495548248291, + "learning_rate": 3.045852502632657e-07, + "loss": 0.5844, + "num_input_tokens_seen": 5408555008, + "step": 2579 + }, + { + "epoch": 0.04882017900732303, + "grad_norm": 8.980935096740723, + "learning_rate": 3.042469782888409e-07, + "loss": 0.216, + "num_input_tokens_seen": 5410652160, + "step": 2580 + }, + { + "epoch": 0.04909140222403038, + "grad_norm": 20.313331604003906, + "learning_rate": 3.0390890408709645e-07, + "loss": 0.4768, + "num_input_tokens_seen": 5412749312, + "step": 2581 + }, + { + "epoch": 0.049362625440737724, + "grad_norm": 11.710707664489746, + "learning_rate": 3.0357102793010145e-07, + "loss": 0.3278, + "num_input_tokens_seen": 5414846464, + "step": 2582 + }, + { + "epoch": 0.04963384865744508, + "grad_norm": 11.84463119506836, + "learning_rate": 3.032333500897659e-07, + "loss": 0.209, + "num_input_tokens_seen": 5416943616, + "step": 2583 + }, + { + "epoch": 0.049905071874152426, + "grad_norm": 15.556648254394531, + "learning_rate": 3.0289587083784e-07, + "loss": 0.4434, + "num_input_tokens_seen": 5419040768, + "step": 2584 + }, + { + "epoch": 0.05017629509085978, + "grad_norm": 10.980393409729004, + "learning_rate": 3.0255859044591425e-07, + "loss": 0.2918, + "num_input_tokens_seen": 5421137920, + "step": 2585 + }, + { + "epoch": 0.05044751830756713, + "grad_norm": 9.970746994018555, + "learning_rate": 3.02221509185419e-07, + "loss": 0.2593, + "num_input_tokens_seen": 5423235072, + "step": 2586 + }, + { + "epoch": 0.050718741524274476, + "grad_norm": 12.755905151367188, + "learning_rate": 3.0188462732762457e-07, + "loss": 0.4138, + "num_input_tokens_seen": 5425332224, + "step": 2587 + }, + { + "epoch": 0.05098996474098183, + "grad_norm": 15.958523750305176, + "learning_rate": 3.015479451436406e-07, + "loss": 0.4157, + "num_input_tokens_seen": 5427429376, + "step": 2588 + }, + { + "epoch": 0.05126118795768918, + "grad_norm": 16.09610366821289, + "learning_rate": 3.0121146290441593e-07, + "loss": 0.453, + "num_input_tokens_seen": 5429526528, + "step": 2589 + }, + { + "epoch": 0.051532411174396525, + "grad_norm": 18.44511604309082, + "learning_rate": 3.0087518088073905e-07, + "loss": 0.675, + "num_input_tokens_seen": 5431623680, + "step": 2590 + }, + { + "epoch": 0.05180363439110388, + "grad_norm": 15.760098457336426, + "learning_rate": 3.005390993432366e-07, + "loss": 0.4527, + "num_input_tokens_seen": 5433720832, + "step": 2591 + }, + { + "epoch": 0.05207485760781123, + "grad_norm": 25.40867805480957, + "learning_rate": 3.00203218562374e-07, + "loss": 0.398, + "num_input_tokens_seen": 5435817984, + "step": 2592 + }, + { + "epoch": 0.05234608082451858, + "grad_norm": 11.827055931091309, + "learning_rate": 2.9986753880845596e-07, + "loss": 0.2695, + "num_input_tokens_seen": 5437915136, + "step": 2593 + }, + { + "epoch": 0.05261730404122593, + "grad_norm": 11.77489185333252, + "learning_rate": 2.9953206035162433e-07, + "loss": 0.2848, + "num_input_tokens_seen": 5440012288, + "step": 2594 + }, + { + "epoch": 0.05288852725793328, + "grad_norm": 13.155770301818848, + "learning_rate": 2.991967834618597e-07, + "loss": 0.3632, + "num_input_tokens_seen": 5442109440, + "step": 2595 + }, + { + "epoch": 0.05315975047464063, + "grad_norm": 10.677423477172852, + "learning_rate": 2.9886170840897977e-07, + "loss": 0.3251, + "num_input_tokens_seen": 5444206592, + "step": 2596 + }, + { + "epoch": 0.05343097369134798, + "grad_norm": 15.714378356933594, + "learning_rate": 2.9852683546264047e-07, + "loss": 0.4266, + "num_input_tokens_seen": 5446303744, + "step": 2597 + }, + { + "epoch": 0.05370219690805533, + "grad_norm": 21.693063735961914, + "learning_rate": 2.9819216489233467e-07, + "loss": 0.7394, + "num_input_tokens_seen": 5448400896, + "step": 2598 + }, + { + "epoch": 0.05397342012476268, + "grad_norm": 11.83186149597168, + "learning_rate": 2.978576969673926e-07, + "loss": 0.3142, + "num_input_tokens_seen": 5450498048, + "step": 2599 + }, + { + "epoch": 0.05424464334147003, + "grad_norm": 14.470967292785645, + "learning_rate": 2.9752343195698125e-07, + "loss": 0.4464, + "num_input_tokens_seen": 5452595200, + "step": 2600 + }, + { + "epoch": 0.05451586655817738, + "grad_norm": 14.670758247375488, + "learning_rate": 2.9718937013010444e-07, + "loss": 0.354, + "num_input_tokens_seen": 5454692352, + "step": 2601 + }, + { + "epoch": 0.05478708977488473, + "grad_norm": 7.69711971282959, + "learning_rate": 2.9685551175560235e-07, + "loss": 0.2038, + "num_input_tokens_seen": 5456789504, + "step": 2602 + }, + { + "epoch": 0.05505831299159208, + "grad_norm": 17.91798973083496, + "learning_rate": 2.965218571021516e-07, + "loss": 0.7645, + "num_input_tokens_seen": 5458886656, + "step": 2603 + }, + { + "epoch": 0.05532953620829943, + "grad_norm": 14.40876579284668, + "learning_rate": 2.9618840643826464e-07, + "loss": 0.4033, + "num_input_tokens_seen": 5460983808, + "step": 2604 + }, + { + "epoch": 0.05560075942500678, + "grad_norm": 17.744657516479492, + "learning_rate": 2.958551600322899e-07, + "loss": 0.5836, + "num_input_tokens_seen": 5463080960, + "step": 2605 + }, + { + "epoch": 0.055871982641714134, + "grad_norm": 14.996540069580078, + "learning_rate": 2.9552211815241156e-07, + "loss": 0.4728, + "num_input_tokens_seen": 5465178112, + "step": 2606 + }, + { + "epoch": 0.05614320585842148, + "grad_norm": 13.542115211486816, + "learning_rate": 2.9518928106664897e-07, + "loss": 0.3398, + "num_input_tokens_seen": 5467275264, + "step": 2607 + }, + { + "epoch": 0.05641442907512883, + "grad_norm": 17.0628719329834, + "learning_rate": 2.9485664904285643e-07, + "loss": 0.5308, + "num_input_tokens_seen": 5469372416, + "step": 2608 + }, + { + "epoch": 0.05668565229183618, + "grad_norm": 16.092180252075195, + "learning_rate": 2.94524222348724e-07, + "loss": 0.4883, + "num_input_tokens_seen": 5471469568, + "step": 2609 + }, + { + "epoch": 0.05695687550854353, + "grad_norm": 16.60764503479004, + "learning_rate": 2.9419200125177585e-07, + "loss": 0.4317, + "num_input_tokens_seen": 5473566720, + "step": 2610 + }, + { + "epoch": 0.05722809872525088, + "grad_norm": 9.960092544555664, + "learning_rate": 2.938599860193709e-07, + "loss": 0.2807, + "num_input_tokens_seen": 5475663872, + "step": 2611 + }, + { + "epoch": 0.05749932194195823, + "grad_norm": 8.7996187210083, + "learning_rate": 2.935281769187025e-07, + "loss": 0.198, + "num_input_tokens_seen": 5477761024, + "step": 2612 + }, + { + "epoch": 0.05777054515866558, + "grad_norm": 16.473575592041016, + "learning_rate": 2.9319657421679757e-07, + "loss": 0.4382, + "num_input_tokens_seen": 5479858176, + "step": 2613 + }, + { + "epoch": 0.058041768375372935, + "grad_norm": 13.767109870910645, + "learning_rate": 2.928651781805177e-07, + "loss": 0.3479, + "num_input_tokens_seen": 5481955328, + "step": 2614 + }, + { + "epoch": 0.05831299159208028, + "grad_norm": 15.12543773651123, + "learning_rate": 2.9253398907655775e-07, + "loss": 0.2618, + "num_input_tokens_seen": 5484052480, + "step": 2615 + }, + { + "epoch": 0.05858421480878763, + "grad_norm": 10.026293754577637, + "learning_rate": 2.9220300717144597e-07, + "loss": 0.2574, + "num_input_tokens_seen": 5486149632, + "step": 2616 + }, + { + "epoch": 0.058855438025494984, + "grad_norm": 10.266706466674805, + "learning_rate": 2.9187223273154406e-07, + "loss": 0.2729, + "num_input_tokens_seen": 5488246784, + "step": 2617 + }, + { + "epoch": 0.05912666124220233, + "grad_norm": 16.52318000793457, + "learning_rate": 2.91541666023047e-07, + "loss": 0.5708, + "num_input_tokens_seen": 5490343936, + "step": 2618 + }, + { + "epoch": 0.059397884458909686, + "grad_norm": 23.888343811035156, + "learning_rate": 2.9121130731198204e-07, + "loss": 0.5555, + "num_input_tokens_seen": 5492441088, + "step": 2619 + }, + { + "epoch": 0.059669107675617034, + "grad_norm": 12.785634994506836, + "learning_rate": 2.9088115686420917e-07, + "loss": 0.4039, + "num_input_tokens_seen": 5494538240, + "step": 2620 + }, + { + "epoch": 0.05994033089232438, + "grad_norm": 15.309110641479492, + "learning_rate": 2.9055121494542115e-07, + "loss": 0.3993, + "num_input_tokens_seen": 5496635392, + "step": 2621 + }, + { + "epoch": 0.060211554109031735, + "grad_norm": 12.312765121459961, + "learning_rate": 2.9022148182114247e-07, + "loss": 0.3103, + "num_input_tokens_seen": 5498732544, + "step": 2622 + }, + { + "epoch": 0.06048277732573908, + "grad_norm": 20.542030334472656, + "learning_rate": 2.8989195775673e-07, + "loss": 0.5541, + "num_input_tokens_seen": 5500829696, + "step": 2623 + }, + { + "epoch": 0.06075400054244643, + "grad_norm": 10.995097160339355, + "learning_rate": 2.89562643017372e-07, + "loss": 0.3218, + "num_input_tokens_seen": 5502926848, + "step": 2624 + }, + { + "epoch": 0.061025223759153785, + "grad_norm": 20.725732803344727, + "learning_rate": 2.8923353786808857e-07, + "loss": 0.4933, + "num_input_tokens_seen": 5505024000, + "step": 2625 + }, + { + "epoch": 0.06129644697586113, + "grad_norm": 20.948585510253906, + "learning_rate": 2.8890464257373105e-07, + "loss": 0.8481, + "num_input_tokens_seen": 5507121152, + "step": 2626 + }, + { + "epoch": 0.06156767019256849, + "grad_norm": 12.35352611541748, + "learning_rate": 2.8857595739898164e-07, + "loss": 0.3946, + "num_input_tokens_seen": 5509218304, + "step": 2627 + }, + { + "epoch": 0.061838893409275834, + "grad_norm": 15.794177055358887, + "learning_rate": 2.8824748260835386e-07, + "loss": 0.6322, + "num_input_tokens_seen": 5511315456, + "step": 2628 + }, + { + "epoch": 0.06211011662598318, + "grad_norm": 20.326683044433594, + "learning_rate": 2.879192184661918e-07, + "loss": 0.6344, + "num_input_tokens_seen": 5513412608, + "step": 2629 + }, + { + "epoch": 0.062381339842690536, + "grad_norm": 16.63013458251953, + "learning_rate": 2.8759116523666973e-07, + "loss": 0.3775, + "num_input_tokens_seen": 5515509760, + "step": 2630 + }, + { + "epoch": 0.06265256305939788, + "grad_norm": 17.79449462890625, + "learning_rate": 2.872633231837923e-07, + "loss": 0.6498, + "num_input_tokens_seen": 5517606912, + "step": 2631 + }, + { + "epoch": 0.06292378627610523, + "grad_norm": 14.09510612487793, + "learning_rate": 2.869356925713946e-07, + "loss": 0.3053, + "num_input_tokens_seen": 5519704064, + "step": 2632 + }, + { + "epoch": 0.06319500949281258, + "grad_norm": 22.424938201904297, + "learning_rate": 2.86608273663141e-07, + "loss": 0.7264, + "num_input_tokens_seen": 5521801216, + "step": 2633 + }, + { + "epoch": 0.06346623270951994, + "grad_norm": 19.868343353271484, + "learning_rate": 2.8628106672252614e-07, + "loss": 0.7189, + "num_input_tokens_seen": 5523898368, + "step": 2634 + }, + { + "epoch": 0.06373745592622729, + "grad_norm": 11.038687705993652, + "learning_rate": 2.859540720128737e-07, + "loss": 0.316, + "num_input_tokens_seen": 5525995520, + "step": 2635 + }, + { + "epoch": 0.06400867914293464, + "grad_norm": 20.234725952148438, + "learning_rate": 2.856272897973362e-07, + "loss": 0.4332, + "num_input_tokens_seen": 5528092672, + "step": 2636 + }, + { + "epoch": 0.06427990235964198, + "grad_norm": 17.736494064331055, + "learning_rate": 2.853007203388958e-07, + "loss": 0.5984, + "num_input_tokens_seen": 5530189824, + "step": 2637 + }, + { + "epoch": 0.06455112557634933, + "grad_norm": 13.119430541992188, + "learning_rate": 2.849743639003631e-07, + "loss": 0.318, + "num_input_tokens_seen": 5532286976, + "step": 2638 + }, + { + "epoch": 0.06482234879305669, + "grad_norm": 17.339921951293945, + "learning_rate": 2.846482207443773e-07, + "loss": 0.4915, + "num_input_tokens_seen": 5534384128, + "step": 2639 + }, + { + "epoch": 0.06509357200976404, + "grad_norm": 12.775398254394531, + "learning_rate": 2.843222911334061e-07, + "loss": 0.3488, + "num_input_tokens_seen": 5536481280, + "step": 2640 + }, + { + "epoch": 0.06536479522647139, + "grad_norm": 19.59604263305664, + "learning_rate": 2.839965753297452e-07, + "loss": 0.6288, + "num_input_tokens_seen": 5538578432, + "step": 2641 + }, + { + "epoch": 0.06563601844317873, + "grad_norm": 11.156400680541992, + "learning_rate": 2.8367107359551835e-07, + "loss": 0.3287, + "num_input_tokens_seen": 5540675584, + "step": 2642 + }, + { + "epoch": 0.06590724165988608, + "grad_norm": 17.5681095123291, + "learning_rate": 2.8334578619267683e-07, + "loss": 0.6211, + "num_input_tokens_seen": 5542772736, + "step": 2643 + }, + { + "epoch": 0.06617846487659344, + "grad_norm": 11.513365745544434, + "learning_rate": 2.830207133829997e-07, + "loss": 0.2376, + "num_input_tokens_seen": 5544869888, + "step": 2644 + }, + { + "epoch": 0.06644968809330079, + "grad_norm": 18.175596237182617, + "learning_rate": 2.8269585542809305e-07, + "loss": 0.5273, + "num_input_tokens_seen": 5546967040, + "step": 2645 + }, + { + "epoch": 0.06672091131000814, + "grad_norm": 15.030982971191406, + "learning_rate": 2.8237121258939036e-07, + "loss": 0.3895, + "num_input_tokens_seen": 5549064192, + "step": 2646 + }, + { + "epoch": 0.06699213452671549, + "grad_norm": 23.305875778198242, + "learning_rate": 2.8204678512815185e-07, + "loss": 0.5009, + "num_input_tokens_seen": 5551161344, + "step": 2647 + }, + { + "epoch": 0.06726335774342283, + "grad_norm": 19.74225425720215, + "learning_rate": 2.8172257330546414e-07, + "loss": 0.5924, + "num_input_tokens_seen": 5553258496, + "step": 2648 + }, + { + "epoch": 0.06753458096013018, + "grad_norm": 13.259350776672363, + "learning_rate": 2.8139857738224055e-07, + "loss": 0.3439, + "num_input_tokens_seen": 5555355648, + "step": 2649 + }, + { + "epoch": 0.06780580417683754, + "grad_norm": 13.237799644470215, + "learning_rate": 2.81074797619221e-07, + "loss": 0.378, + "num_input_tokens_seen": 5557452800, + "step": 2650 + }, + { + "epoch": 0.06807702739354489, + "grad_norm": 14.329866409301758, + "learning_rate": 2.8075123427697093e-07, + "loss": 0.3658, + "num_input_tokens_seen": 5559549952, + "step": 2651 + }, + { + "epoch": 0.06834825061025224, + "grad_norm": 9.912088394165039, + "learning_rate": 2.8042788761588204e-07, + "loss": 0.2397, + "num_input_tokens_seen": 5561647104, + "step": 2652 + }, + { + "epoch": 0.06861947382695958, + "grad_norm": 19.02084732055664, + "learning_rate": 2.8010475789617105e-07, + "loss": 0.6075, + "num_input_tokens_seen": 5563744256, + "step": 2653 + }, + { + "epoch": 0.06889069704366693, + "grad_norm": 15.462594032287598, + "learning_rate": 2.797818453778806e-07, + "loss": 0.5206, + "num_input_tokens_seen": 5565841408, + "step": 2654 + }, + { + "epoch": 0.0691619202603743, + "grad_norm": 12.81130313873291, + "learning_rate": 2.794591503208785e-07, + "loss": 0.4167, + "num_input_tokens_seen": 5567938560, + "step": 2655 + }, + { + "epoch": 0.06943314347708164, + "grad_norm": 21.61235809326172, + "learning_rate": 2.791366729848574e-07, + "loss": 0.665, + "num_input_tokens_seen": 5570035712, + "step": 2656 + }, + { + "epoch": 0.06970436669378899, + "grad_norm": 27.101455688476562, + "learning_rate": 2.7881441362933464e-07, + "loss": 0.6858, + "num_input_tokens_seen": 5572132864, + "step": 2657 + }, + { + "epoch": 0.06997558991049634, + "grad_norm": 17.959209442138672, + "learning_rate": 2.784923725136525e-07, + "loss": 0.5518, + "num_input_tokens_seen": 5574230016, + "step": 2658 + }, + { + "epoch": 0.07024681312720368, + "grad_norm": 10.713489532470703, + "learning_rate": 2.781705498969773e-07, + "loss": 0.2727, + "num_input_tokens_seen": 5576327168, + "step": 2659 + }, + { + "epoch": 0.07051803634391104, + "grad_norm": 13.588273048400879, + "learning_rate": 2.7784894603829966e-07, + "loss": 0.4162, + "num_input_tokens_seen": 5578424320, + "step": 2660 + }, + { + "epoch": 0.07078925956061839, + "grad_norm": 14.655363082885742, + "learning_rate": 2.7752756119643416e-07, + "loss": 0.4192, + "num_input_tokens_seen": 5580521472, + "step": 2661 + }, + { + "epoch": 0.07106048277732574, + "grad_norm": 11.51174545288086, + "learning_rate": 2.77206395630019e-07, + "loss": 0.2364, + "num_input_tokens_seen": 5582618624, + "step": 2662 + }, + { + "epoch": 0.07133170599403309, + "grad_norm": 14.545519828796387, + "learning_rate": 2.7688544959751615e-07, + "loss": 0.4139, + "num_input_tokens_seen": 5584715776, + "step": 2663 + }, + { + "epoch": 0.07160292921074043, + "grad_norm": 18.516597747802734, + "learning_rate": 2.765647233572108e-07, + "loss": 0.4624, + "num_input_tokens_seen": 5586812928, + "step": 2664 + }, + { + "epoch": 0.0718741524274478, + "grad_norm": 16.15486717224121, + "learning_rate": 2.7624421716721086e-07, + "loss": 0.6008, + "num_input_tokens_seen": 5588910080, + "step": 2665 + }, + { + "epoch": 0.07214537564415514, + "grad_norm": 14.004398345947266, + "learning_rate": 2.7592393128544784e-07, + "loss": 0.4258, + "num_input_tokens_seen": 5591007232, + "step": 2666 + }, + { + "epoch": 0.07241659886086249, + "grad_norm": 18.64406967163086, + "learning_rate": 2.7560386596967553e-07, + "loss": 0.3099, + "num_input_tokens_seen": 5593104384, + "step": 2667 + }, + { + "epoch": 0.07268782207756984, + "grad_norm": 23.5781307220459, + "learning_rate": 2.7528402147747045e-07, + "loss": 0.8708, + "num_input_tokens_seen": 5595201536, + "step": 2668 + }, + { + "epoch": 0.07295904529427719, + "grad_norm": 16.206514358520508, + "learning_rate": 2.7496439806623124e-07, + "loss": 0.5205, + "num_input_tokens_seen": 5597298688, + "step": 2669 + }, + { + "epoch": 0.07323026851098453, + "grad_norm": 13.137125968933105, + "learning_rate": 2.746449959931784e-07, + "loss": 0.3308, + "num_input_tokens_seen": 5599395840, + "step": 2670 + }, + { + "epoch": 0.0735014917276919, + "grad_norm": 16.53045082092285, + "learning_rate": 2.743258155153548e-07, + "loss": 0.3276, + "num_input_tokens_seen": 5601492992, + "step": 2671 + }, + { + "epoch": 0.07377271494439924, + "grad_norm": 17.230728149414062, + "learning_rate": 2.740068568896247e-07, + "loss": 0.662, + "num_input_tokens_seen": 5603590144, + "step": 2672 + }, + { + "epoch": 0.07404393816110659, + "grad_norm": 16.93472671508789, + "learning_rate": 2.7368812037267387e-07, + "loss": 0.6562, + "num_input_tokens_seen": 5605687296, + "step": 2673 + }, + { + "epoch": 0.07431516137781394, + "grad_norm": 24.3737735748291, + "learning_rate": 2.7336960622100907e-07, + "loss": 0.8715, + "num_input_tokens_seen": 5607784448, + "step": 2674 + }, + { + "epoch": 0.07458638459452128, + "grad_norm": 10.38357162475586, + "learning_rate": 2.7305131469095906e-07, + "loss": 0.2184, + "num_input_tokens_seen": 5609881600, + "step": 2675 + }, + { + "epoch": 0.07485760781122865, + "grad_norm": 17.325672149658203, + "learning_rate": 2.7273324603867203e-07, + "loss": 0.5331, + "num_input_tokens_seen": 5611978752, + "step": 2676 + }, + { + "epoch": 0.075128831027936, + "grad_norm": 17.097993850708008, + "learning_rate": 2.7241540052011787e-07, + "loss": 0.6643, + "num_input_tokens_seen": 5614075904, + "step": 2677 + }, + { + "epoch": 0.07540005424464334, + "grad_norm": 9.489489555358887, + "learning_rate": 2.720977783910865e-07, + "loss": 0.2313, + "num_input_tokens_seen": 5616173056, + "step": 2678 + }, + { + "epoch": 0.07567127746135069, + "grad_norm": 16.426868438720703, + "learning_rate": 2.717803799071881e-07, + "loss": 0.4908, + "num_input_tokens_seen": 5618270208, + "step": 2679 + }, + { + "epoch": 0.07594250067805804, + "grad_norm": 16.419797897338867, + "learning_rate": 2.714632053238529e-07, + "loss": 0.6232, + "num_input_tokens_seen": 5620367360, + "step": 2680 + }, + { + "epoch": 0.0762137238947654, + "grad_norm": 14.052704811096191, + "learning_rate": 2.711462548963309e-07, + "loss": 0.3795, + "num_input_tokens_seen": 5622464512, + "step": 2681 + }, + { + "epoch": 0.07648494711147275, + "grad_norm": 14.086715698242188, + "learning_rate": 2.708295288796918e-07, + "loss": 0.4398, + "num_input_tokens_seen": 5624561664, + "step": 2682 + }, + { + "epoch": 0.07675617032818009, + "grad_norm": 17.156911849975586, + "learning_rate": 2.7051302752882467e-07, + "loss": 0.5515, + "num_input_tokens_seen": 5626658816, + "step": 2683 + }, + { + "epoch": 0.07702739354488744, + "grad_norm": 14.56786823272705, + "learning_rate": 2.7019675109843777e-07, + "loss": 0.3963, + "num_input_tokens_seen": 5628755968, + "step": 2684 + }, + { + "epoch": 0.07729861676159479, + "grad_norm": 19.659238815307617, + "learning_rate": 2.6988069984305817e-07, + "loss": 0.4958, + "num_input_tokens_seen": 5630853120, + "step": 2685 + }, + { + "epoch": 0.07756983997830215, + "grad_norm": 17.3908634185791, + "learning_rate": 2.6956487401703207e-07, + "loss": 0.5915, + "num_input_tokens_seen": 5632950272, + "step": 2686 + }, + { + "epoch": 0.0778410631950095, + "grad_norm": 24.59274673461914, + "learning_rate": 2.692492738745241e-07, + "loss": 0.6446, + "num_input_tokens_seen": 5635047424, + "step": 2687 + }, + { + "epoch": 0.07811228641171684, + "grad_norm": 21.10965347290039, + "learning_rate": 2.689338996695171e-07, + "loss": 0.7802, + "num_input_tokens_seen": 5637144576, + "step": 2688 + }, + { + "epoch": 0.07838350962842419, + "grad_norm": 12.290288925170898, + "learning_rate": 2.686187516558122e-07, + "loss": 0.3851, + "num_input_tokens_seen": 5639241728, + "step": 2689 + }, + { + "epoch": 0.07865473284513154, + "grad_norm": 17.741069793701172, + "learning_rate": 2.683038300870285e-07, + "loss": 0.5609, + "num_input_tokens_seen": 5641338880, + "step": 2690 + }, + { + "epoch": 0.07892595606183889, + "grad_norm": 11.127710342407227, + "learning_rate": 2.679891352166032e-07, + "loss": 0.3566, + "num_input_tokens_seen": 5643436032, + "step": 2691 + }, + { + "epoch": 0.07919717927854625, + "grad_norm": 17.833637237548828, + "learning_rate": 2.6767466729779074e-07, + "loss": 0.5932, + "num_input_tokens_seen": 5645533184, + "step": 2692 + }, + { + "epoch": 0.0794684024952536, + "grad_norm": 18.04448127746582, + "learning_rate": 2.6736042658366255e-07, + "loss": 0.5263, + "num_input_tokens_seen": 5647630336, + "step": 2693 + }, + { + "epoch": 0.07973962571196094, + "grad_norm": 16.75208282470703, + "learning_rate": 2.6704641332710783e-07, + "loss": 0.5939, + "num_input_tokens_seen": 5649727488, + "step": 2694 + }, + { + "epoch": 0.08001084892866829, + "grad_norm": 18.136194229125977, + "learning_rate": 2.6673262778083246e-07, + "loss": 0.5117, + "num_input_tokens_seen": 5651824640, + "step": 2695 + }, + { + "epoch": 0.08028207214537564, + "grad_norm": 14.582008361816406, + "learning_rate": 2.6641907019735914e-07, + "loss": 0.4515, + "num_input_tokens_seen": 5653921792, + "step": 2696 + }, + { + "epoch": 0.080553295362083, + "grad_norm": 9.552046775817871, + "learning_rate": 2.6610574082902704e-07, + "loss": 0.2099, + "num_input_tokens_seen": 5656018944, + "step": 2697 + }, + { + "epoch": 0.08082451857879035, + "grad_norm": 15.047245979309082, + "learning_rate": 2.657926399279918e-07, + "loss": 0.5305, + "num_input_tokens_seen": 5658116096, + "step": 2698 + }, + { + "epoch": 0.0810957417954977, + "grad_norm": 11.793010711669922, + "learning_rate": 2.65479767746225e-07, + "loss": 0.2842, + "num_input_tokens_seen": 5660213248, + "step": 2699 + }, + { + "epoch": 0.08136696501220504, + "grad_norm": 11.611842155456543, + "learning_rate": 2.651671245355144e-07, + "loss": 0.2676, + "num_input_tokens_seen": 5662310400, + "step": 2700 + }, + { + "epoch": 0.08163818822891239, + "grad_norm": 14.877463340759277, + "learning_rate": 2.6485471054746315e-07, + "loss": 0.4741, + "num_input_tokens_seen": 5664407552, + "step": 2701 + }, + { + "epoch": 0.08190941144561975, + "grad_norm": 16.487375259399414, + "learning_rate": 2.645425260334904e-07, + "loss": 0.4078, + "num_input_tokens_seen": 5666504704, + "step": 2702 + }, + { + "epoch": 0.0821806346623271, + "grad_norm": 21.23032569885254, + "learning_rate": 2.6423057124483015e-07, + "loss": 0.7662, + "num_input_tokens_seen": 5668601856, + "step": 2703 + }, + { + "epoch": 0.08245185787903445, + "grad_norm": 9.899909973144531, + "learning_rate": 2.6391884643253197e-07, + "loss": 0.2434, + "num_input_tokens_seen": 5670699008, + "step": 2704 + }, + { + "epoch": 0.0827230810957418, + "grad_norm": 13.111746788024902, + "learning_rate": 2.6360735184745984e-07, + "loss": 0.3323, + "num_input_tokens_seen": 5672796160, + "step": 2705 + }, + { + "epoch": 0.08299430431244914, + "grad_norm": 13.60911750793457, + "learning_rate": 2.6329608774029285e-07, + "loss": 0.4452, + "num_input_tokens_seen": 5674893312, + "step": 2706 + }, + { + "epoch": 0.0832655275291565, + "grad_norm": 19.949071884155273, + "learning_rate": 2.6298505436152457e-07, + "loss": 0.6172, + "num_input_tokens_seen": 5676990464, + "step": 2707 + }, + { + "epoch": 0.08353675074586385, + "grad_norm": 24.41508674621582, + "learning_rate": 2.626742519614629e-07, + "loss": 0.3937, + "num_input_tokens_seen": 5679087616, + "step": 2708 + }, + { + "epoch": 0.0838079739625712, + "grad_norm": 14.442510604858398, + "learning_rate": 2.623636807902298e-07, + "loss": 0.4276, + "num_input_tokens_seen": 5681184768, + "step": 2709 + }, + { + "epoch": 0.08407919717927854, + "grad_norm": 20.97626495361328, + "learning_rate": 2.620533410977609e-07, + "loss": 0.3205, + "num_input_tokens_seen": 5683281920, + "step": 2710 + }, + { + "epoch": 0.08435042039598589, + "grad_norm": 17.24020004272461, + "learning_rate": 2.617432331338059e-07, + "loss": 0.4976, + "num_input_tokens_seen": 5685379072, + "step": 2711 + }, + { + "epoch": 0.08462164361269324, + "grad_norm": 25.12828826904297, + "learning_rate": 2.614333571479279e-07, + "loss": 0.7029, + "num_input_tokens_seen": 5687476224, + "step": 2712 + }, + { + "epoch": 0.0848928668294006, + "grad_norm": 10.766886711120605, + "learning_rate": 2.6112371338950325e-07, + "loss": 0.2783, + "num_input_tokens_seen": 5689573376, + "step": 2713 + }, + { + "epoch": 0.08516409004610795, + "grad_norm": 19.415145874023438, + "learning_rate": 2.608143021077215e-07, + "loss": 0.6848, + "num_input_tokens_seen": 5691670528, + "step": 2714 + }, + { + "epoch": 0.0854353132628153, + "grad_norm": 11.980889320373535, + "learning_rate": 2.6050512355158503e-07, + "loss": 0.35, + "num_input_tokens_seen": 5693767680, + "step": 2715 + }, + { + "epoch": 0.08570653647952264, + "grad_norm": 18.769763946533203, + "learning_rate": 2.601961779699091e-07, + "loss": 0.6817, + "num_input_tokens_seen": 5695864832, + "step": 2716 + }, + { + "epoch": 0.08597775969622999, + "grad_norm": 11.969738960266113, + "learning_rate": 2.5988746561132136e-07, + "loss": 0.3525, + "num_input_tokens_seen": 5697961984, + "step": 2717 + }, + { + "epoch": 0.08624898291293735, + "grad_norm": 13.127973556518555, + "learning_rate": 2.595789867242616e-07, + "loss": 0.3985, + "num_input_tokens_seen": 5700059136, + "step": 2718 + }, + { + "epoch": 0.0865202061296447, + "grad_norm": 14.819026947021484, + "learning_rate": 2.5927074155698203e-07, + "loss": 0.3679, + "num_input_tokens_seen": 5702156288, + "step": 2719 + }, + { + "epoch": 0.08679142934635205, + "grad_norm": 19.240089416503906, + "learning_rate": 2.589627303575465e-07, + "loss": 0.7484, + "num_input_tokens_seen": 5704253440, + "step": 2720 + }, + { + "epoch": 0.0870626525630594, + "grad_norm": 15.054210662841797, + "learning_rate": 2.5865495337383103e-07, + "loss": 0.4892, + "num_input_tokens_seen": 5706350592, + "step": 2721 + }, + { + "epoch": 0.08733387577976674, + "grad_norm": 20.892154693603516, + "learning_rate": 2.5834741085352223e-07, + "loss": 0.4666, + "num_input_tokens_seen": 5708447744, + "step": 2722 + }, + { + "epoch": 0.0876050989964741, + "grad_norm": 17.323524475097656, + "learning_rate": 2.5804010304411914e-07, + "loss": 0.7108, + "num_input_tokens_seen": 5710544896, + "step": 2723 + }, + { + "epoch": 0.08787632221318145, + "grad_norm": 9.577323913574219, + "learning_rate": 2.5773303019293123e-07, + "loss": 0.2308, + "num_input_tokens_seen": 5712642048, + "step": 2724 + }, + { + "epoch": 0.0881475454298888, + "grad_norm": 8.77543830871582, + "learning_rate": 2.5742619254707905e-07, + "loss": 0.1873, + "num_input_tokens_seen": 5714739200, + "step": 2725 + }, + { + "epoch": 0.08841876864659615, + "grad_norm": 16.472463607788086, + "learning_rate": 2.5711959035349396e-07, + "loss": 0.5128, + "num_input_tokens_seen": 5716836352, + "step": 2726 + }, + { + "epoch": 0.0886899918633035, + "grad_norm": 10.295055389404297, + "learning_rate": 2.568132238589175e-07, + "loss": 0.2232, + "num_input_tokens_seen": 5718933504, + "step": 2727 + }, + { + "epoch": 0.08896121508001086, + "grad_norm": 17.552204132080078, + "learning_rate": 2.565070933099019e-07, + "loss": 0.4109, + "num_input_tokens_seen": 5721030656, + "step": 2728 + }, + { + "epoch": 0.0892324382967182, + "grad_norm": 14.377561569213867, + "learning_rate": 2.5620119895280935e-07, + "loss": 0.3614, + "num_input_tokens_seen": 5723127808, + "step": 2729 + }, + { + "epoch": 0.08950366151342555, + "grad_norm": 14.117269515991211, + "learning_rate": 2.5589554103381195e-07, + "loss": 0.3619, + "num_input_tokens_seen": 5725224960, + "step": 2730 + }, + { + "epoch": 0.0897748847301329, + "grad_norm": 17.120271682739258, + "learning_rate": 2.5559011979889155e-07, + "loss": 0.4312, + "num_input_tokens_seen": 5727322112, + "step": 2731 + }, + { + "epoch": 0.09004610794684025, + "grad_norm": 16.064838409423828, + "learning_rate": 2.552849354938399e-07, + "loss": 0.3673, + "num_input_tokens_seen": 5729419264, + "step": 2732 + }, + { + "epoch": 0.09031733116354759, + "grad_norm": 12.923369407653809, + "learning_rate": 2.549799883642573e-07, + "loss": 0.3846, + "num_input_tokens_seen": 5731516416, + "step": 2733 + }, + { + "epoch": 0.09058855438025495, + "grad_norm": 10.39802074432373, + "learning_rate": 2.5467527865555366e-07, + "loss": 0.2633, + "num_input_tokens_seen": 5733613568, + "step": 2734 + }, + { + "epoch": 0.0908597775969623, + "grad_norm": 21.36480140686035, + "learning_rate": 2.5437080661294785e-07, + "loss": 0.627, + "num_input_tokens_seen": 5735710720, + "step": 2735 + }, + { + "epoch": 0.09113100081366965, + "grad_norm": 18.12044906616211, + "learning_rate": 2.5406657248146735e-07, + "loss": 0.5679, + "num_input_tokens_seen": 5737807872, + "step": 2736 + }, + { + "epoch": 0.091402224030377, + "grad_norm": 17.808435440063477, + "learning_rate": 2.5376257650594823e-07, + "loss": 0.5542, + "num_input_tokens_seen": 5739905024, + "step": 2737 + }, + { + "epoch": 0.09167344724708434, + "grad_norm": 11.723854064941406, + "learning_rate": 2.5345881893103484e-07, + "loss": 0.2498, + "num_input_tokens_seen": 5742002176, + "step": 2738 + }, + { + "epoch": 0.0919446704637917, + "grad_norm": 21.123031616210938, + "learning_rate": 2.5315530000117973e-07, + "loss": 0.7246, + "num_input_tokens_seen": 5744099328, + "step": 2739 + }, + { + "epoch": 0.09221589368049905, + "grad_norm": 21.011804580688477, + "learning_rate": 2.528520199606434e-07, + "loss": 0.7669, + "num_input_tokens_seen": 5746196480, + "step": 2740 + }, + { + "epoch": 0.0924871168972064, + "grad_norm": 16.76370620727539, + "learning_rate": 2.525489790534941e-07, + "loss": 0.6061, + "num_input_tokens_seen": 5748293632, + "step": 2741 + }, + { + "epoch": 0.09275834011391375, + "grad_norm": 20.120906829833984, + "learning_rate": 2.5224617752360766e-07, + "loss": 0.7128, + "num_input_tokens_seen": 5750390784, + "step": 2742 + }, + { + "epoch": 0.0930295633306211, + "grad_norm": 12.05733871459961, + "learning_rate": 2.519436156146671e-07, + "loss": 0.4059, + "num_input_tokens_seen": 5752487936, + "step": 2743 + }, + { + "epoch": 0.09330078654732846, + "grad_norm": 14.433719635009766, + "learning_rate": 2.51641293570163e-07, + "loss": 0.5306, + "num_input_tokens_seen": 5754585088, + "step": 2744 + }, + { + "epoch": 0.0935720097640358, + "grad_norm": 10.016216278076172, + "learning_rate": 2.513392116333922e-07, + "loss": 0.2541, + "num_input_tokens_seen": 5756682240, + "step": 2745 + }, + { + "epoch": 0.09384323298074315, + "grad_norm": 15.06049919128418, + "learning_rate": 2.510373700474592e-07, + "loss": 0.4363, + "num_input_tokens_seen": 5758779392, + "step": 2746 + }, + { + "epoch": 0.0941144561974505, + "grad_norm": 12.102774620056152, + "learning_rate": 2.5073576905527407e-07, + "loss": 0.3181, + "num_input_tokens_seen": 5760876544, + "step": 2747 + }, + { + "epoch": 0.09438567941415785, + "grad_norm": 16.28287124633789, + "learning_rate": 2.5043440889955434e-07, + "loss": 0.5829, + "num_input_tokens_seen": 5762973696, + "step": 2748 + }, + { + "epoch": 0.09465690263086521, + "grad_norm": 17.23906135559082, + "learning_rate": 2.501332898228232e-07, + "loss": 0.4747, + "num_input_tokens_seen": 5765070848, + "step": 2749 + }, + { + "epoch": 0.09492812584757256, + "grad_norm": 16.309682846069336, + "learning_rate": 2.4983241206740945e-07, + "loss": 0.3862, + "num_input_tokens_seen": 5767168000, + "step": 2750 + }, + { + "epoch": 0.0951993490642799, + "grad_norm": 16.106428146362305, + "learning_rate": 2.495317758754483e-07, + "loss": 0.5324, + "num_input_tokens_seen": 5769265152, + "step": 2751 + }, + { + "epoch": 0.09547057228098725, + "grad_norm": 17.912614822387695, + "learning_rate": 2.492313814888802e-07, + "loss": 0.5044, + "num_input_tokens_seen": 5771362304, + "step": 2752 + }, + { + "epoch": 0.0957417954976946, + "grad_norm": 14.133460998535156, + "learning_rate": 2.4893122914945124e-07, + "loss": 0.3362, + "num_input_tokens_seen": 5773459456, + "step": 2753 + }, + { + "epoch": 0.09601301871440195, + "grad_norm": 15.399311065673828, + "learning_rate": 2.4863131909871247e-07, + "loss": 0.3883, + "num_input_tokens_seen": 5775556608, + "step": 2754 + }, + { + "epoch": 0.09628424193110931, + "grad_norm": 15.49337100982666, + "learning_rate": 2.483316515780202e-07, + "loss": 0.5335, + "num_input_tokens_seen": 5777653760, + "step": 2755 + }, + { + "epoch": 0.09655546514781665, + "grad_norm": 9.771089553833008, + "learning_rate": 2.480322268285354e-07, + "loss": 0.2648, + "num_input_tokens_seen": 5779750912, + "step": 2756 + }, + { + "epoch": 0.096826688364524, + "grad_norm": 12.803630828857422, + "learning_rate": 2.4773304509122374e-07, + "loss": 0.3456, + "num_input_tokens_seen": 5781848064, + "step": 2757 + }, + { + "epoch": 0.09709791158123135, + "grad_norm": 14.358229637145996, + "learning_rate": 2.474341066068553e-07, + "loss": 0.4647, + "num_input_tokens_seen": 5783945216, + "step": 2758 + }, + { + "epoch": 0.0973691347979387, + "grad_norm": 19.96444320678711, + "learning_rate": 2.4713541161600434e-07, + "loss": 0.7291, + "num_input_tokens_seen": 5786042368, + "step": 2759 + }, + { + "epoch": 0.09764035801464606, + "grad_norm": 15.138256072998047, + "learning_rate": 2.4683696035904926e-07, + "loss": 0.4123, + "num_input_tokens_seen": 5788139520, + "step": 2760 + }, + { + "epoch": 0.0979115812313534, + "grad_norm": 14.141944885253906, + "learning_rate": 2.465387530761724e-07, + "loss": 0.3199, + "num_input_tokens_seen": 5790236672, + "step": 2761 + }, + { + "epoch": 0.09818280444806075, + "grad_norm": 14.20753288269043, + "learning_rate": 2.462407900073594e-07, + "loss": 0.3559, + "num_input_tokens_seen": 5792333824, + "step": 2762 + }, + { + "epoch": 0.0984540276647681, + "grad_norm": 17.01785659790039, + "learning_rate": 2.459430713923995e-07, + "loss": 0.6968, + "num_input_tokens_seen": 5794430976, + "step": 2763 + }, + { + "epoch": 0.09872525088147545, + "grad_norm": 9.875651359558105, + "learning_rate": 2.4564559747088573e-07, + "loss": 0.2179, + "num_input_tokens_seen": 5796528128, + "step": 2764 + }, + { + "epoch": 0.09899647409818281, + "grad_norm": 16.090343475341797, + "learning_rate": 2.4534836848221355e-07, + "loss": 0.4962, + "num_input_tokens_seen": 5798625280, + "step": 2765 + }, + { + "epoch": 0.09926769731489016, + "grad_norm": 13.723416328430176, + "learning_rate": 2.450513846655817e-07, + "loss": 0.4518, + "num_input_tokens_seen": 5800722432, + "step": 2766 + }, + { + "epoch": 0.0995389205315975, + "grad_norm": 13.636754989624023, + "learning_rate": 2.4475464625999113e-07, + "loss": 0.4715, + "num_input_tokens_seen": 5802819584, + "step": 2767 + }, + { + "epoch": 0.09981014374830485, + "grad_norm": 16.98908233642578, + "learning_rate": 2.444581535042459e-07, + "loss": 0.4008, + "num_input_tokens_seen": 5804916736, + "step": 2768 + }, + { + "epoch": 0.1000813669650122, + "grad_norm": 23.043598175048828, + "learning_rate": 2.441619066369519e-07, + "loss": 0.988, + "num_input_tokens_seen": 5807013888, + "step": 2769 + }, + { + "epoch": 0.10035259018171956, + "grad_norm": 16.22281837463379, + "learning_rate": 2.438659058965175e-07, + "loss": 0.585, + "num_input_tokens_seen": 5809111040, + "step": 2770 + }, + { + "epoch": 0.10062381339842691, + "grad_norm": 13.878887176513672, + "learning_rate": 2.435701515211527e-07, + "loss": 0.3956, + "num_input_tokens_seen": 5811208192, + "step": 2771 + }, + { + "epoch": 0.10089503661513426, + "grad_norm": 15.073182106018066, + "learning_rate": 2.4327464374886955e-07, + "loss": 0.3474, + "num_input_tokens_seen": 5813305344, + "step": 2772 + }, + { + "epoch": 0.1011662598318416, + "grad_norm": 37.12259292602539, + "learning_rate": 2.4297938281748124e-07, + "loss": 0.4542, + "num_input_tokens_seen": 5815402496, + "step": 2773 + }, + { + "epoch": 0.10143748304854895, + "grad_norm": 21.452543258666992, + "learning_rate": 2.4268436896460267e-07, + "loss": 0.5917, + "num_input_tokens_seen": 5817499648, + "step": 2774 + }, + { + "epoch": 0.10170870626525631, + "grad_norm": 18.589229583740234, + "learning_rate": 2.423896024276497e-07, + "loss": 0.6739, + "num_input_tokens_seen": 5819596800, + "step": 2775 + }, + { + "epoch": 0.10197992948196366, + "grad_norm": 15.261178970336914, + "learning_rate": 2.4209508344383926e-07, + "loss": 0.5353, + "num_input_tokens_seen": 5821693952, + "step": 2776 + }, + { + "epoch": 0.10225115269867101, + "grad_norm": 19.2561092376709, + "learning_rate": 2.4180081225018906e-07, + "loss": 0.6213, + "num_input_tokens_seen": 5823791104, + "step": 2777 + }, + { + "epoch": 0.10252237591537836, + "grad_norm": 17.913654327392578, + "learning_rate": 2.4150678908351744e-07, + "loss": 0.748, + "num_input_tokens_seen": 5825888256, + "step": 2778 + }, + { + "epoch": 0.1027935991320857, + "grad_norm": 11.78269100189209, + "learning_rate": 2.4121301418044264e-07, + "loss": 0.3224, + "num_input_tokens_seen": 5827985408, + "step": 2779 + }, + { + "epoch": 0.10306482234879305, + "grad_norm": 11.825636863708496, + "learning_rate": 2.409194877773839e-07, + "loss": 0.2892, + "num_input_tokens_seen": 5830082560, + "step": 2780 + }, + { + "epoch": 0.10333604556550041, + "grad_norm": 21.131614685058594, + "learning_rate": 2.4062621011056006e-07, + "loss": 0.4957, + "num_input_tokens_seen": 5832179712, + "step": 2781 + }, + { + "epoch": 0.10360726878220776, + "grad_norm": 16.387815475463867, + "learning_rate": 2.4033318141598977e-07, + "loss": 0.3793, + "num_input_tokens_seen": 5834276864, + "step": 2782 + }, + { + "epoch": 0.1038784919989151, + "grad_norm": 15.578926086425781, + "learning_rate": 2.4004040192949155e-07, + "loss": 0.5131, + "num_input_tokens_seen": 5836374016, + "step": 2783 + }, + { + "epoch": 0.10414971521562245, + "grad_norm": 18.233552932739258, + "learning_rate": 2.3974787188668284e-07, + "loss": 0.3963, + "num_input_tokens_seen": 5838471168, + "step": 2784 + }, + { + "epoch": 0.1044209384323298, + "grad_norm": 16.372631072998047, + "learning_rate": 2.3945559152298085e-07, + "loss": 0.5632, + "num_input_tokens_seen": 5840568320, + "step": 2785 + }, + { + "epoch": 0.10469216164903716, + "grad_norm": 13.143184661865234, + "learning_rate": 2.3916356107360174e-07, + "loss": 0.4031, + "num_input_tokens_seen": 5842665472, + "step": 2786 + }, + { + "epoch": 0.10496338486574451, + "grad_norm": 17.36909294128418, + "learning_rate": 2.388717807735605e-07, + "loss": 0.5266, + "num_input_tokens_seen": 5844762624, + "step": 2787 + }, + { + "epoch": 0.10523460808245186, + "grad_norm": 11.10837173461914, + "learning_rate": 2.385802508576706e-07, + "loss": 0.2815, + "num_input_tokens_seen": 5846859776, + "step": 2788 + }, + { + "epoch": 0.1055058312991592, + "grad_norm": 16.8798885345459, + "learning_rate": 2.382889715605447e-07, + "loss": 0.2858, + "num_input_tokens_seen": 5848956928, + "step": 2789 + }, + { + "epoch": 0.10577705451586655, + "grad_norm": 17.28868293762207, + "learning_rate": 2.3799794311659286e-07, + "loss": 0.2763, + "num_input_tokens_seen": 5851054080, + "step": 2790 + }, + { + "epoch": 0.10604827773257391, + "grad_norm": 20.567678451538086, + "learning_rate": 2.3770716576002383e-07, + "loss": 0.5661, + "num_input_tokens_seen": 5853151232, + "step": 2791 + }, + { + "epoch": 0.10631950094928126, + "grad_norm": 15.973060607910156, + "learning_rate": 2.3741663972484416e-07, + "loss": 0.5525, + "num_input_tokens_seen": 5855248384, + "step": 2792 + }, + { + "epoch": 0.10659072416598861, + "grad_norm": 12.698427200317383, + "learning_rate": 2.3712636524485817e-07, + "loss": 0.2953, + "num_input_tokens_seen": 5857345536, + "step": 2793 + }, + { + "epoch": 0.10686194738269596, + "grad_norm": 11.15018081665039, + "learning_rate": 2.3683634255366773e-07, + "loss": 0.246, + "num_input_tokens_seen": 5859442688, + "step": 2794 + }, + { + "epoch": 0.1071331705994033, + "grad_norm": 18.49885368347168, + "learning_rate": 2.3654657188467203e-07, + "loss": 0.6083, + "num_input_tokens_seen": 5861539840, + "step": 2795 + }, + { + "epoch": 0.10740439381611067, + "grad_norm": 30.425281524658203, + "learning_rate": 2.3625705347106748e-07, + "loss": 0.5344, + "num_input_tokens_seen": 5863636992, + "step": 2796 + }, + { + "epoch": 0.10767561703281801, + "grad_norm": 14.430277824401855, + "learning_rate": 2.3596778754584752e-07, + "loss": 0.4491, + "num_input_tokens_seen": 5865734144, + "step": 2797 + }, + { + "epoch": 0.10794684024952536, + "grad_norm": 16.26645278930664, + "learning_rate": 2.356787743418023e-07, + "loss": 0.402, + "num_input_tokens_seen": 5867831296, + "step": 2798 + }, + { + "epoch": 0.10821806346623271, + "grad_norm": 9.986083030700684, + "learning_rate": 2.3539001409151867e-07, + "loss": 0.2545, + "num_input_tokens_seen": 5869928448, + "step": 2799 + }, + { + "epoch": 0.10848928668294006, + "grad_norm": 13.592569351196289, + "learning_rate": 2.351015070273799e-07, + "loss": 0.4131, + "num_input_tokens_seen": 5872025600, + "step": 2800 + }, + { + "epoch": 0.1087605098996474, + "grad_norm": 18.985036849975586, + "learning_rate": 2.3481325338156568e-07, + "loss": 0.7892, + "num_input_tokens_seen": 5874122752, + "step": 2801 + }, + { + "epoch": 0.10903173311635476, + "grad_norm": 10.7304048538208, + "learning_rate": 2.3452525338605126e-07, + "loss": 0.2678, + "num_input_tokens_seen": 5876219904, + "step": 2802 + }, + { + "epoch": 0.10930295633306211, + "grad_norm": 16.625381469726562, + "learning_rate": 2.3423750727260813e-07, + "loss": 0.3209, + "num_input_tokens_seen": 5878317056, + "step": 2803 + }, + { + "epoch": 0.10957417954976946, + "grad_norm": 13.496749877929688, + "learning_rate": 2.3395001527280343e-07, + "loss": 0.2586, + "num_input_tokens_seen": 5880414208, + "step": 2804 + }, + { + "epoch": 0.10984540276647681, + "grad_norm": 17.611549377441406, + "learning_rate": 2.3366277761799993e-07, + "loss": 0.6402, + "num_input_tokens_seen": 5882511360, + "step": 2805 + }, + { + "epoch": 0.11011662598318415, + "grad_norm": 16.595611572265625, + "learning_rate": 2.333757945393557e-07, + "loss": 0.5554, + "num_input_tokens_seen": 5884608512, + "step": 2806 + }, + { + "epoch": 0.11038784919989152, + "grad_norm": 14.515238761901855, + "learning_rate": 2.3308906626782333e-07, + "loss": 0.4586, + "num_input_tokens_seen": 5886705664, + "step": 2807 + }, + { + "epoch": 0.11065907241659886, + "grad_norm": 23.05605697631836, + "learning_rate": 2.3280259303415115e-07, + "loss": 0.6299, + "num_input_tokens_seen": 5888802816, + "step": 2808 + }, + { + "epoch": 0.11093029563330621, + "grad_norm": 10.770841598510742, + "learning_rate": 2.3251637506888178e-07, + "loss": 0.191, + "num_input_tokens_seen": 5890899968, + "step": 2809 + }, + { + "epoch": 0.11120151885001356, + "grad_norm": 12.446057319641113, + "learning_rate": 2.3223041260235265e-07, + "loss": 0.3576, + "num_input_tokens_seen": 5892997120, + "step": 2810 + }, + { + "epoch": 0.1114727420667209, + "grad_norm": 8.817853927612305, + "learning_rate": 2.3194470586469543e-07, + "loss": 0.2229, + "num_input_tokens_seen": 5895094272, + "step": 2811 + }, + { + "epoch": 0.11174396528342827, + "grad_norm": 16.82341957092285, + "learning_rate": 2.3165925508583598e-07, + "loss": 0.372, + "num_input_tokens_seen": 5897191424, + "step": 2812 + }, + { + "epoch": 0.11201518850013562, + "grad_norm": 19.0521183013916, + "learning_rate": 2.313740604954943e-07, + "loss": 0.4898, + "num_input_tokens_seen": 5899288576, + "step": 2813 + }, + { + "epoch": 0.11228641171684296, + "grad_norm": 16.593416213989258, + "learning_rate": 2.3108912232318413e-07, + "loss": 0.4401, + "num_input_tokens_seen": 5901385728, + "step": 2814 + }, + { + "epoch": 0.11255763493355031, + "grad_norm": 12.823646545410156, + "learning_rate": 2.3080444079821284e-07, + "loss": 0.3815, + "num_input_tokens_seen": 5903482880, + "step": 2815 + }, + { + "epoch": 0.11282885815025766, + "grad_norm": 26.622953414916992, + "learning_rate": 2.305200161496813e-07, + "loss": 0.9118, + "num_input_tokens_seen": 5905580032, + "step": 2816 + }, + { + "epoch": 0.11310008136696502, + "grad_norm": 16.885168075561523, + "learning_rate": 2.3023584860648364e-07, + "loss": 0.7591, + "num_input_tokens_seen": 5907677184, + "step": 2817 + }, + { + "epoch": 0.11337130458367237, + "grad_norm": 15.160221099853516, + "learning_rate": 2.2995193839730727e-07, + "loss": 0.4651, + "num_input_tokens_seen": 5909774336, + "step": 2818 + }, + { + "epoch": 0.11364252780037971, + "grad_norm": 14.561100959777832, + "learning_rate": 2.2966828575063196e-07, + "loss": 0.4091, + "num_input_tokens_seen": 5911871488, + "step": 2819 + }, + { + "epoch": 0.11391375101708706, + "grad_norm": 16.13934898376465, + "learning_rate": 2.2938489089473065e-07, + "loss": 0.483, + "num_input_tokens_seen": 5913968640, + "step": 2820 + }, + { + "epoch": 0.11418497423379441, + "grad_norm": 20.38286590576172, + "learning_rate": 2.2910175405766896e-07, + "loss": 0.7366, + "num_input_tokens_seen": 5916065792, + "step": 2821 + }, + { + "epoch": 0.11445619745050176, + "grad_norm": 18.793577194213867, + "learning_rate": 2.288188754673045e-07, + "loss": 0.4723, + "num_input_tokens_seen": 5918162944, + "step": 2822 + }, + { + "epoch": 0.11472742066720912, + "grad_norm": 18.476377487182617, + "learning_rate": 2.2853625535128735e-07, + "loss": 0.7003, + "num_input_tokens_seen": 5920260096, + "step": 2823 + }, + { + "epoch": 0.11499864388391647, + "grad_norm": 11.239076614379883, + "learning_rate": 2.2825389393705922e-07, + "loss": 0.2991, + "num_input_tokens_seen": 5922357248, + "step": 2824 + }, + { + "epoch": 0.11526986710062381, + "grad_norm": 17.184232711791992, + "learning_rate": 2.2797179145185384e-07, + "loss": 0.6045, + "num_input_tokens_seen": 5924454400, + "step": 2825 + }, + { + "epoch": 0.11554109031733116, + "grad_norm": 12.037182807922363, + "learning_rate": 2.2768994812269666e-07, + "loss": 0.2699, + "num_input_tokens_seen": 5926551552, + "step": 2826 + }, + { + "epoch": 0.11581231353403851, + "grad_norm": 12.547082901000977, + "learning_rate": 2.2740836417640432e-07, + "loss": 0.3864, + "num_input_tokens_seen": 5928648704, + "step": 2827 + }, + { + "epoch": 0.11608353675074587, + "grad_norm": 16.02481460571289, + "learning_rate": 2.271270398395851e-07, + "loss": 0.6835, + "num_input_tokens_seen": 5930745856, + "step": 2828 + }, + { + "epoch": 0.11635475996745322, + "grad_norm": 11.960121154785156, + "learning_rate": 2.2684597533863793e-07, + "loss": 0.3943, + "num_input_tokens_seen": 5932843008, + "step": 2829 + }, + { + "epoch": 0.11662598318416056, + "grad_norm": 11.560256958007812, + "learning_rate": 2.2656517089975286e-07, + "loss": 0.3184, + "num_input_tokens_seen": 5934940160, + "step": 2830 + }, + { + "epoch": 0.11689720640086791, + "grad_norm": 15.655224800109863, + "learning_rate": 2.2628462674891062e-07, + "loss": 0.5279, + "num_input_tokens_seen": 5937037312, + "step": 2831 + }, + { + "epoch": 0.11716842961757526, + "grad_norm": 15.328285217285156, + "learning_rate": 2.2600434311188253e-07, + "loss": 0.4113, + "num_input_tokens_seen": 5939134464, + "step": 2832 + }, + { + "epoch": 0.11743965283428262, + "grad_norm": 20.01653289794922, + "learning_rate": 2.2572432021423018e-07, + "loss": 0.5133, + "num_input_tokens_seen": 5941231616, + "step": 2833 + }, + { + "epoch": 0.11771087605098997, + "grad_norm": 18.01945686340332, + "learning_rate": 2.2544455828130528e-07, + "loss": 0.7168, + "num_input_tokens_seen": 5943328768, + "step": 2834 + }, + { + "epoch": 0.11798209926769732, + "grad_norm": 14.951272964477539, + "learning_rate": 2.251650575382498e-07, + "loss": 0.5018, + "num_input_tokens_seen": 5945425920, + "step": 2835 + }, + { + "epoch": 0.11825332248440466, + "grad_norm": 20.382293701171875, + "learning_rate": 2.2488581820999481e-07, + "loss": 0.5145, + "num_input_tokens_seen": 5947523072, + "step": 2836 + }, + { + "epoch": 0.11852454570111201, + "grad_norm": 17.990644454956055, + "learning_rate": 2.2460684052126195e-07, + "loss": 0.6625, + "num_input_tokens_seen": 5949620224, + "step": 2837 + }, + { + "epoch": 0.11879576891781937, + "grad_norm": 14.664972305297852, + "learning_rate": 2.243281246965617e-07, + "loss": 0.5635, + "num_input_tokens_seen": 5951717376, + "step": 2838 + }, + { + "epoch": 0.11906699213452672, + "grad_norm": 12.643186569213867, + "learning_rate": 2.24049670960194e-07, + "loss": 0.2588, + "num_input_tokens_seen": 5953814528, + "step": 2839 + }, + { + "epoch": 0.11933821535123407, + "grad_norm": 12.24855899810791, + "learning_rate": 2.2377147953624776e-07, + "loss": 0.3325, + "num_input_tokens_seen": 5955911680, + "step": 2840 + }, + { + "epoch": 0.11960943856794141, + "grad_norm": 18.2128849029541, + "learning_rate": 2.2349355064860104e-07, + "loss": 0.5419, + "num_input_tokens_seen": 5958008832, + "step": 2841 + }, + { + "epoch": 0.11988066178464876, + "grad_norm": 18.80755043029785, + "learning_rate": 2.2321588452092006e-07, + "loss": 0.7344, + "num_input_tokens_seen": 5960105984, + "step": 2842 + }, + { + "epoch": 0.12015188500135611, + "grad_norm": 9.964282035827637, + "learning_rate": 2.229384813766602e-07, + "loss": 0.3019, + "num_input_tokens_seen": 5962203136, + "step": 2843 + }, + { + "epoch": 0.12042310821806347, + "grad_norm": 14.523286819458008, + "learning_rate": 2.2266134143906496e-07, + "loss": 0.3197, + "num_input_tokens_seen": 5964300288, + "step": 2844 + }, + { + "epoch": 0.12069433143477082, + "grad_norm": 14.210477828979492, + "learning_rate": 2.2238446493116572e-07, + "loss": 0.378, + "num_input_tokens_seen": 5966397440, + "step": 2845 + }, + { + "epoch": 0.12096555465147817, + "grad_norm": 16.729429244995117, + "learning_rate": 2.2210785207578275e-07, + "loss": 0.5794, + "num_input_tokens_seen": 5968494592, + "step": 2846 + }, + { + "epoch": 0.12123677786818551, + "grad_norm": 11.647602081298828, + "learning_rate": 2.2183150309552308e-07, + "loss": 0.2484, + "num_input_tokens_seen": 5970591744, + "step": 2847 + }, + { + "epoch": 0.12150800108489286, + "grad_norm": 18.91074562072754, + "learning_rate": 2.2155541821278196e-07, + "loss": 0.4457, + "num_input_tokens_seen": 5972688896, + "step": 2848 + }, + { + "epoch": 0.12177922430160022, + "grad_norm": 13.002408027648926, + "learning_rate": 2.2127959764974203e-07, + "loss": 0.3013, + "num_input_tokens_seen": 5974786048, + "step": 2849 + }, + { + "epoch": 0.12205044751830757, + "grad_norm": 17.768815994262695, + "learning_rate": 2.2100404162837317e-07, + "loss": 0.5265, + "num_input_tokens_seen": 5976883200, + "step": 2850 + }, + { + "epoch": 0.12232167073501492, + "grad_norm": 17.112438201904297, + "learning_rate": 2.2072875037043232e-07, + "loss": 0.5274, + "num_input_tokens_seen": 5978980352, + "step": 2851 + }, + { + "epoch": 0.12259289395172226, + "grad_norm": 11.445368766784668, + "learning_rate": 2.204537240974636e-07, + "loss": 0.3618, + "num_input_tokens_seen": 5981077504, + "step": 2852 + }, + { + "epoch": 0.12286411716842961, + "grad_norm": 17.361818313598633, + "learning_rate": 2.2017896303079743e-07, + "loss": 0.5425, + "num_input_tokens_seen": 5983174656, + "step": 2853 + }, + { + "epoch": 0.12313534038513697, + "grad_norm": 19.054218292236328, + "learning_rate": 2.1990446739155128e-07, + "loss": 0.3945, + "num_input_tokens_seen": 5985271808, + "step": 2854 + }, + { + "epoch": 0.12340656360184432, + "grad_norm": 11.030587196350098, + "learning_rate": 2.1963023740062864e-07, + "loss": 0.2286, + "num_input_tokens_seen": 5987368960, + "step": 2855 + }, + { + "epoch": 0.12367778681855167, + "grad_norm": 12.42334270477295, + "learning_rate": 2.1935627327871948e-07, + "loss": 0.2749, + "num_input_tokens_seen": 5989466112, + "step": 2856 + }, + { + "epoch": 0.12394901003525902, + "grad_norm": 10.632824897766113, + "learning_rate": 2.1908257524629963e-07, + "loss": 0.4374, + "num_input_tokens_seen": 5991563264, + "step": 2857 + }, + { + "epoch": 0.12422023325196636, + "grad_norm": 19.065475463867188, + "learning_rate": 2.1880914352363106e-07, + "loss": 0.581, + "num_input_tokens_seen": 5993660416, + "step": 2858 + }, + { + "epoch": 0.12449145646867373, + "grad_norm": 16.23573112487793, + "learning_rate": 2.1853597833076088e-07, + "loss": 0.5547, + "num_input_tokens_seen": 5995757568, + "step": 2859 + }, + { + "epoch": 0.12476267968538107, + "grad_norm": 12.071977615356445, + "learning_rate": 2.1826307988752212e-07, + "loss": 0.4157, + "num_input_tokens_seen": 5997854720, + "step": 2860 + }, + { + "epoch": 0.1250339029020884, + "grad_norm": 14.335684776306152, + "learning_rate": 2.179904484135331e-07, + "loss": 0.3816, + "num_input_tokens_seen": 5999951872, + "step": 2861 + }, + { + "epoch": 0.12530512611879577, + "grad_norm": 14.643807411193848, + "learning_rate": 2.177180841281974e-07, + "loss": 0.4579, + "num_input_tokens_seen": 6002049024, + "step": 2862 + }, + { + "epoch": 0.12557634933550313, + "grad_norm": 12.312231063842773, + "learning_rate": 2.1744598725070347e-07, + "loss": 0.4013, + "num_input_tokens_seen": 6004146176, + "step": 2863 + }, + { + "epoch": 0.12584757255221046, + "grad_norm": 17.71323013305664, + "learning_rate": 2.1717415800002425e-07, + "loss": 0.5469, + "num_input_tokens_seen": 6006243328, + "step": 2864 + }, + { + "epoch": 0.12611879576891782, + "grad_norm": 17.643123626708984, + "learning_rate": 2.1690259659491768e-07, + "loss": 0.5235, + "num_input_tokens_seen": 6008340480, + "step": 2865 + }, + { + "epoch": 0.12639001898562516, + "grad_norm": 13.04069995880127, + "learning_rate": 2.166313032539261e-07, + "loss": 0.3721, + "num_input_tokens_seen": 6010437632, + "step": 2866 + }, + { + "epoch": 0.12666124220233252, + "grad_norm": 15.557741165161133, + "learning_rate": 2.1636027819537605e-07, + "loss": 0.5084, + "num_input_tokens_seen": 6012534784, + "step": 2867 + }, + { + "epoch": 0.12693246541903988, + "grad_norm": 13.02405071258545, + "learning_rate": 2.160895216373781e-07, + "loss": 0.2844, + "num_input_tokens_seen": 6014631936, + "step": 2868 + }, + { + "epoch": 0.12720368863574721, + "grad_norm": 7.915323257446289, + "learning_rate": 2.15819033797827e-07, + "loss": 0.1902, + "num_input_tokens_seen": 6016729088, + "step": 2869 + }, + { + "epoch": 0.12747491185245458, + "grad_norm": 13.274603843688965, + "learning_rate": 2.1554881489440092e-07, + "loss": 0.4097, + "num_input_tokens_seen": 6018826240, + "step": 2870 + }, + { + "epoch": 0.1277461350691619, + "grad_norm": 14.94298267364502, + "learning_rate": 2.1527886514456178e-07, + "loss": 0.5581, + "num_input_tokens_seen": 6020923392, + "step": 2871 + }, + { + "epoch": 0.12801735828586927, + "grad_norm": 11.724287033081055, + "learning_rate": 2.150091847655549e-07, + "loss": 0.3464, + "num_input_tokens_seen": 6023020544, + "step": 2872 + }, + { + "epoch": 0.12828858150257663, + "grad_norm": 8.7376127243042, + "learning_rate": 2.1473977397440878e-07, + "loss": 0.1962, + "num_input_tokens_seen": 6025117696, + "step": 2873 + }, + { + "epoch": 0.12855980471928397, + "grad_norm": 13.356949806213379, + "learning_rate": 2.14470632987935e-07, + "loss": 0.5469, + "num_input_tokens_seen": 6027214848, + "step": 2874 + }, + { + "epoch": 0.12883102793599133, + "grad_norm": 11.348970413208008, + "learning_rate": 2.1420176202272805e-07, + "loss": 0.2122, + "num_input_tokens_seen": 6029312000, + "step": 2875 + }, + { + "epoch": 0.12910225115269866, + "grad_norm": 12.910688400268555, + "learning_rate": 2.1393316129516482e-07, + "loss": 0.4277, + "num_input_tokens_seen": 6031409152, + "step": 2876 + }, + { + "epoch": 0.12937347436940602, + "grad_norm": 18.22507667541504, + "learning_rate": 2.1366483102140497e-07, + "loss": 0.3707, + "num_input_tokens_seen": 6033506304, + "step": 2877 + }, + { + "epoch": 0.12964469758611338, + "grad_norm": 12.903253555297852, + "learning_rate": 2.1339677141739074e-07, + "loss": 0.3805, + "num_input_tokens_seen": 6035603456, + "step": 2878 + }, + { + "epoch": 0.12991592080282072, + "grad_norm": 15.44893741607666, + "learning_rate": 2.1312898269884616e-07, + "loss": 0.5398, + "num_input_tokens_seen": 6037700608, + "step": 2879 + }, + { + "epoch": 0.13018714401952808, + "grad_norm": 9.813923835754395, + "learning_rate": 2.128614650812775e-07, + "loss": 0.2239, + "num_input_tokens_seen": 6039797760, + "step": 2880 + }, + { + "epoch": 0.1304583672362354, + "grad_norm": 23.024755477905273, + "learning_rate": 2.125942187799726e-07, + "loss": 0.667, + "num_input_tokens_seen": 6041894912, + "step": 2881 + }, + { + "epoch": 0.13072959045294277, + "grad_norm": 21.844196319580078, + "learning_rate": 2.123272440100012e-07, + "loss": 0.7983, + "num_input_tokens_seen": 6043992064, + "step": 2882 + }, + { + "epoch": 0.13100081366965013, + "grad_norm": 16.548011779785156, + "learning_rate": 2.1206054098621434e-07, + "loss": 0.7041, + "num_input_tokens_seen": 6046089216, + "step": 2883 + }, + { + "epoch": 0.13127203688635747, + "grad_norm": 21.630081176757812, + "learning_rate": 2.117941099232446e-07, + "loss": 0.713, + "num_input_tokens_seen": 6048186368, + "step": 2884 + }, + { + "epoch": 0.13154326010306483, + "grad_norm": 12.728462219238281, + "learning_rate": 2.1152795103550547e-07, + "loss": 0.294, + "num_input_tokens_seen": 6050283520, + "step": 2885 + }, + { + "epoch": 0.13181448331977216, + "grad_norm": 33.04969787597656, + "learning_rate": 2.1126206453719157e-07, + "loss": 0.6548, + "num_input_tokens_seen": 6052380672, + "step": 2886 + }, + { + "epoch": 0.13208570653647952, + "grad_norm": 12.736434936523438, + "learning_rate": 2.1099645064227817e-07, + "loss": 0.3612, + "num_input_tokens_seen": 6054477824, + "step": 2887 + }, + { + "epoch": 0.13235692975318689, + "grad_norm": 13.483380317687988, + "learning_rate": 2.1073110956452126e-07, + "loss": 0.3295, + "num_input_tokens_seen": 6056574976, + "step": 2888 + }, + { + "epoch": 0.13262815296989422, + "grad_norm": 12.762460708618164, + "learning_rate": 2.1046604151745723e-07, + "loss": 0.3028, + "num_input_tokens_seen": 6058672128, + "step": 2889 + }, + { + "epoch": 0.13289937618660158, + "grad_norm": 12.392841339111328, + "learning_rate": 2.1020124671440274e-07, + "loss": 0.3086, + "num_input_tokens_seen": 6060769280, + "step": 2890 + }, + { + "epoch": 0.13317059940330891, + "grad_norm": 15.763758659362793, + "learning_rate": 2.0993672536845458e-07, + "loss": 0.3993, + "num_input_tokens_seen": 6062866432, + "step": 2891 + }, + { + "epoch": 0.13344182262001628, + "grad_norm": 14.49795150756836, + "learning_rate": 2.0967247769248968e-07, + "loss": 0.458, + "num_input_tokens_seen": 6064963584, + "step": 2892 + }, + { + "epoch": 0.13371304583672364, + "grad_norm": 17.557239532470703, + "learning_rate": 2.0940850389916398e-07, + "loss": 0.2814, + "num_input_tokens_seen": 6067060736, + "step": 2893 + }, + { + "epoch": 0.13398426905343097, + "grad_norm": 9.931197166442871, + "learning_rate": 2.09144804200914e-07, + "loss": 0.1937, + "num_input_tokens_seen": 6069157888, + "step": 2894 + }, + { + "epoch": 0.13425549227013833, + "grad_norm": 10.081891059875488, + "learning_rate": 2.0888137880995514e-07, + "loss": 0.2684, + "num_input_tokens_seen": 6071255040, + "step": 2895 + }, + { + "epoch": 0.13452671548684567, + "grad_norm": 18.273103713989258, + "learning_rate": 2.086182279382821e-07, + "loss": 0.6209, + "num_input_tokens_seen": 6073352192, + "step": 2896 + }, + { + "epoch": 0.13479793870355303, + "grad_norm": 16.057680130004883, + "learning_rate": 2.083553517976686e-07, + "loss": 0.4178, + "num_input_tokens_seen": 6075449344, + "step": 2897 + }, + { + "epoch": 0.13506916192026036, + "grad_norm": 14.120661735534668, + "learning_rate": 2.0809275059966764e-07, + "loss": 0.477, + "num_input_tokens_seen": 6077546496, + "step": 2898 + }, + { + "epoch": 0.13534038513696772, + "grad_norm": 10.962662696838379, + "learning_rate": 2.0783042455561023e-07, + "loss": 0.3011, + "num_input_tokens_seen": 6079643648, + "step": 2899 + }, + { + "epoch": 0.13561160835367508, + "grad_norm": 12.212385177612305, + "learning_rate": 2.0756837387660676e-07, + "loss": 0.361, + "num_input_tokens_seen": 6081740800, + "step": 2900 + }, + { + "epoch": 0.13588283157038242, + "grad_norm": 15.496916770935059, + "learning_rate": 2.073065987735454e-07, + "loss": 0.3815, + "num_input_tokens_seen": 6083837952, + "step": 2901 + }, + { + "epoch": 0.13615405478708978, + "grad_norm": 13.568929672241211, + "learning_rate": 2.070450994570928e-07, + "loss": 0.4771, + "num_input_tokens_seen": 6085935104, + "step": 2902 + }, + { + "epoch": 0.1364252780037971, + "grad_norm": 15.480619430541992, + "learning_rate": 2.0678387613769397e-07, + "loss": 0.4991, + "num_input_tokens_seen": 6088032256, + "step": 2903 + }, + { + "epoch": 0.13669650122050447, + "grad_norm": 17.832595825195312, + "learning_rate": 2.0652292902557117e-07, + "loss": 0.5667, + "num_input_tokens_seen": 6090129408, + "step": 2904 + }, + { + "epoch": 0.13696772443721184, + "grad_norm": 19.86845588684082, + "learning_rate": 2.0626225833072487e-07, + "loss": 0.8251, + "num_input_tokens_seen": 6092226560, + "step": 2905 + }, + { + "epoch": 0.13723894765391917, + "grad_norm": 14.052447319030762, + "learning_rate": 2.0600186426293282e-07, + "loss": 0.4079, + "num_input_tokens_seen": 6094323712, + "step": 2906 + }, + { + "epoch": 0.13751017087062653, + "grad_norm": 13.390406608581543, + "learning_rate": 2.0574174703175034e-07, + "loss": 0.2496, + "num_input_tokens_seen": 6096420864, + "step": 2907 + }, + { + "epoch": 0.13778139408733386, + "grad_norm": 16.510652542114258, + "learning_rate": 2.0548190684650981e-07, + "loss": 0.483, + "num_input_tokens_seen": 6098518016, + "step": 2908 + }, + { + "epoch": 0.13805261730404123, + "grad_norm": 14.933103561401367, + "learning_rate": 2.052223439163208e-07, + "loss": 0.5226, + "num_input_tokens_seen": 6100615168, + "step": 2909 + }, + { + "epoch": 0.1383238405207486, + "grad_norm": 10.2757568359375, + "learning_rate": 2.049630584500696e-07, + "loss": 0.3106, + "num_input_tokens_seen": 6102712320, + "step": 2910 + }, + { + "epoch": 0.13859506373745592, + "grad_norm": 21.213306427001953, + "learning_rate": 2.0470405065641938e-07, + "loss": 0.6228, + "num_input_tokens_seen": 6104809472, + "step": 2911 + }, + { + "epoch": 0.13886628695416328, + "grad_norm": 17.050752639770508, + "learning_rate": 2.0444532074380973e-07, + "loss": 0.5536, + "num_input_tokens_seen": 6106906624, + "step": 2912 + }, + { + "epoch": 0.13913751017087062, + "grad_norm": 11.27585506439209, + "learning_rate": 2.0418686892045654e-07, + "loss": 0.2669, + "num_input_tokens_seen": 6109003776, + "step": 2913 + }, + { + "epoch": 0.13940873338757798, + "grad_norm": 16.672374725341797, + "learning_rate": 2.039286953943521e-07, + "loss": 0.5225, + "num_input_tokens_seen": 6111100928, + "step": 2914 + }, + { + "epoch": 0.13967995660428534, + "grad_norm": 13.986382484436035, + "learning_rate": 2.0367080037326472e-07, + "loss": 0.3568, + "num_input_tokens_seen": 6113198080, + "step": 2915 + }, + { + "epoch": 0.13995117982099267, + "grad_norm": 12.307029724121094, + "learning_rate": 2.0341318406473833e-07, + "loss": 0.3359, + "num_input_tokens_seen": 6115295232, + "step": 2916 + }, + { + "epoch": 0.14022240303770003, + "grad_norm": 16.73017120361328, + "learning_rate": 2.031558466760927e-07, + "loss": 0.4768, + "num_input_tokens_seen": 6117392384, + "step": 2917 + }, + { + "epoch": 0.14049362625440737, + "grad_norm": 18.181156158447266, + "learning_rate": 2.0289878841442325e-07, + "loss": 0.5648, + "num_input_tokens_seen": 6119489536, + "step": 2918 + }, + { + "epoch": 0.14076484947111473, + "grad_norm": 14.489134788513184, + "learning_rate": 2.0264200948660076e-07, + "loss": 0.3855, + "num_input_tokens_seen": 6121586688, + "step": 2919 + }, + { + "epoch": 0.1410360726878221, + "grad_norm": 15.638324737548828, + "learning_rate": 2.0238551009927125e-07, + "loss": 0.4921, + "num_input_tokens_seen": 6123683840, + "step": 2920 + }, + { + "epoch": 0.14130729590452942, + "grad_norm": 13.121326446533203, + "learning_rate": 2.021292904588553e-07, + "loss": 0.369, + "num_input_tokens_seen": 6125780992, + "step": 2921 + }, + { + "epoch": 0.14157851912123678, + "grad_norm": 16.779006958007812, + "learning_rate": 2.0187335077154905e-07, + "loss": 0.5329, + "num_input_tokens_seen": 6127878144, + "step": 2922 + }, + { + "epoch": 0.14184974233794412, + "grad_norm": 13.377449035644531, + "learning_rate": 2.0161769124332278e-07, + "loss": 0.4986, + "num_input_tokens_seen": 6129975296, + "step": 2923 + }, + { + "epoch": 0.14212096555465148, + "grad_norm": 11.84175968170166, + "learning_rate": 2.013623120799217e-07, + "loss": 0.349, + "num_input_tokens_seen": 6132072448, + "step": 2924 + }, + { + "epoch": 0.14239218877135884, + "grad_norm": 15.083674430847168, + "learning_rate": 2.0110721348686523e-07, + "loss": 0.4303, + "num_input_tokens_seen": 6134169600, + "step": 2925 + }, + { + "epoch": 0.14266341198806617, + "grad_norm": 13.313584327697754, + "learning_rate": 2.0085239566944702e-07, + "loss": 0.1782, + "num_input_tokens_seen": 6136266752, + "step": 2926 + }, + { + "epoch": 0.14293463520477354, + "grad_norm": 15.516534805297852, + "learning_rate": 2.0059785883273463e-07, + "loss": 0.4328, + "num_input_tokens_seen": 6138363904, + "step": 2927 + }, + { + "epoch": 0.14320585842148087, + "grad_norm": 15.41818618774414, + "learning_rate": 2.003436031815698e-07, + "loss": 0.4351, + "num_input_tokens_seen": 6140461056, + "step": 2928 + }, + { + "epoch": 0.14347708163818823, + "grad_norm": 11.988521575927734, + "learning_rate": 2.0008962892056762e-07, + "loss": 0.3346, + "num_input_tokens_seen": 6142558208, + "step": 2929 + }, + { + "epoch": 0.1437483048548956, + "grad_norm": 15.152950286865234, + "learning_rate": 1.998359362541171e-07, + "loss": 0.5238, + "num_input_tokens_seen": 6144655360, + "step": 2930 + }, + { + "epoch": 0.14401952807160293, + "grad_norm": 14.549203872680664, + "learning_rate": 1.9958252538638032e-07, + "loss": 0.4193, + "num_input_tokens_seen": 6146752512, + "step": 2931 + }, + { + "epoch": 0.1442907512883103, + "grad_norm": 13.985295295715332, + "learning_rate": 1.9932939652129283e-07, + "loss": 0.2768, + "num_input_tokens_seen": 6148849664, + "step": 2932 + }, + { + "epoch": 0.14456197450501762, + "grad_norm": 16.646543502807617, + "learning_rate": 1.990765498625629e-07, + "loss": 0.6969, + "num_input_tokens_seen": 6150946816, + "step": 2933 + }, + { + "epoch": 0.14483319772172498, + "grad_norm": 12.219400405883789, + "learning_rate": 1.988239856136719e-07, + "loss": 0.3376, + "num_input_tokens_seen": 6153043968, + "step": 2934 + }, + { + "epoch": 0.14510442093843234, + "grad_norm": 14.528848648071289, + "learning_rate": 1.9857170397787415e-07, + "loss": 0.4568, + "num_input_tokens_seen": 6155141120, + "step": 2935 + }, + { + "epoch": 0.14537564415513968, + "grad_norm": 12.83976936340332, + "learning_rate": 1.9831970515819625e-07, + "loss": 0.3684, + "num_input_tokens_seen": 6157238272, + "step": 2936 + }, + { + "epoch": 0.14564686737184704, + "grad_norm": 11.202268600463867, + "learning_rate": 1.980679893574373e-07, + "loss": 0.2945, + "num_input_tokens_seen": 6159335424, + "step": 2937 + }, + { + "epoch": 0.14591809058855437, + "grad_norm": 22.032058715820312, + "learning_rate": 1.9781655677816838e-07, + "loss": 0.4701, + "num_input_tokens_seen": 6161432576, + "step": 2938 + }, + { + "epoch": 0.14618931380526173, + "grad_norm": 14.524810791015625, + "learning_rate": 1.97565407622733e-07, + "loss": 0.3425, + "num_input_tokens_seen": 6163529728, + "step": 2939 + }, + { + "epoch": 0.14646053702196907, + "grad_norm": 17.290315628051758, + "learning_rate": 1.9731454209324644e-07, + "loss": 0.6485, + "num_input_tokens_seen": 6165626880, + "step": 2940 + }, + { + "epoch": 0.14673176023867643, + "grad_norm": 25.62712860107422, + "learning_rate": 1.9706396039159568e-07, + "loss": 0.5547, + "num_input_tokens_seen": 6167724032, + "step": 2941 + }, + { + "epoch": 0.1470029834553838, + "grad_norm": 29.259971618652344, + "learning_rate": 1.9681366271943945e-07, + "loss": 0.3389, + "num_input_tokens_seen": 6169821184, + "step": 2942 + }, + { + "epoch": 0.14727420667209112, + "grad_norm": 24.600215911865234, + "learning_rate": 1.965636492782077e-07, + "loss": 0.2235, + "num_input_tokens_seen": 6171918336, + "step": 2943 + }, + { + "epoch": 0.14754542988879848, + "grad_norm": 20.362010955810547, + "learning_rate": 1.963139202691017e-07, + "loss": 0.7951, + "num_input_tokens_seen": 6174015488, + "step": 2944 + }, + { + "epoch": 0.14781665310550582, + "grad_norm": 22.120241165161133, + "learning_rate": 1.9606447589309397e-07, + "loss": 0.7222, + "num_input_tokens_seen": 6176112640, + "step": 2945 + }, + { + "epoch": 0.14808787632221318, + "grad_norm": 9.955594062805176, + "learning_rate": 1.9581531635092773e-07, + "loss": 0.2238, + "num_input_tokens_seen": 6178209792, + "step": 2946 + }, + { + "epoch": 0.14835909953892054, + "grad_norm": 13.511630058288574, + "learning_rate": 1.9556644184311728e-07, + "loss": 0.3301, + "num_input_tokens_seen": 6180306944, + "step": 2947 + }, + { + "epoch": 0.14863032275562787, + "grad_norm": 14.704965591430664, + "learning_rate": 1.9531785256994716e-07, + "loss": 0.4311, + "num_input_tokens_seen": 6182404096, + "step": 2948 + }, + { + "epoch": 0.14890154597233524, + "grad_norm": 16.59981918334961, + "learning_rate": 1.9506954873147276e-07, + "loss": 0.5568, + "num_input_tokens_seen": 6184501248, + "step": 2949 + }, + { + "epoch": 0.14917276918904257, + "grad_norm": 12.132022857666016, + "learning_rate": 1.9482153052751921e-07, + "loss": 0.4439, + "num_input_tokens_seen": 6186598400, + "step": 2950 + }, + { + "epoch": 0.14944399240574993, + "grad_norm": 22.458187103271484, + "learning_rate": 1.9457379815768245e-07, + "loss": 0.543, + "num_input_tokens_seen": 6188695552, + "step": 2951 + }, + { + "epoch": 0.1497152156224573, + "grad_norm": 18.78033447265625, + "learning_rate": 1.9432635182132795e-07, + "loss": 0.4977, + "num_input_tokens_seen": 6190792704, + "step": 2952 + }, + { + "epoch": 0.14998643883916463, + "grad_norm": 22.32543182373047, + "learning_rate": 1.9407919171759112e-07, + "loss": 0.345, + "num_input_tokens_seen": 6192889856, + "step": 2953 + }, + { + "epoch": 0.150257662055872, + "grad_norm": 18.59827423095703, + "learning_rate": 1.9383231804537704e-07, + "loss": 0.4047, + "num_input_tokens_seen": 6194987008, + "step": 2954 + }, + { + "epoch": 0.15052888527257932, + "grad_norm": 12.72734260559082, + "learning_rate": 1.9358573100336034e-07, + "loss": 0.347, + "num_input_tokens_seen": 6197084160, + "step": 2955 + }, + { + "epoch": 0.15080010848928668, + "grad_norm": 14.508316040039062, + "learning_rate": 1.933394307899846e-07, + "loss": 0.4512, + "num_input_tokens_seen": 6199181312, + "step": 2956 + }, + { + "epoch": 0.15107133170599404, + "grad_norm": 19.344667434692383, + "learning_rate": 1.9309341760346304e-07, + "loss": 0.7429, + "num_input_tokens_seen": 6201278464, + "step": 2957 + }, + { + "epoch": 0.15134255492270138, + "grad_norm": 11.575425148010254, + "learning_rate": 1.9284769164177767e-07, + "loss": 0.2508, + "num_input_tokens_seen": 6203375616, + "step": 2958 + }, + { + "epoch": 0.15161377813940874, + "grad_norm": 19.255504608154297, + "learning_rate": 1.9260225310267925e-07, + "loss": 0.5238, + "num_input_tokens_seen": 6205472768, + "step": 2959 + }, + { + "epoch": 0.15188500135611607, + "grad_norm": 18.295150756835938, + "learning_rate": 1.9235710218368784e-07, + "loss": 0.5423, + "num_input_tokens_seen": 6207569920, + "step": 2960 + }, + { + "epoch": 0.15215622457282343, + "grad_norm": 16.604528427124023, + "learning_rate": 1.9211223908209114e-07, + "loss": 0.4414, + "num_input_tokens_seen": 6209667072, + "step": 2961 + }, + { + "epoch": 0.1524274477895308, + "grad_norm": 28.41756248474121, + "learning_rate": 1.9186766399494581e-07, + "loss": 0.4075, + "num_input_tokens_seen": 6211764224, + "step": 2962 + }, + { + "epoch": 0.15269867100623813, + "grad_norm": 13.66691780090332, + "learning_rate": 1.9162337711907657e-07, + "loss": 0.4251, + "num_input_tokens_seen": 6213861376, + "step": 2963 + }, + { + "epoch": 0.1529698942229455, + "grad_norm": 14.04883098602295, + "learning_rate": 1.9137937865107606e-07, + "loss": 0.19, + "num_input_tokens_seen": 6215958528, + "step": 2964 + }, + { + "epoch": 0.15324111743965282, + "grad_norm": 12.723421096801758, + "learning_rate": 1.9113566878730515e-07, + "loss": 0.3229, + "num_input_tokens_seen": 6218055680, + "step": 2965 + }, + { + "epoch": 0.15351234065636019, + "grad_norm": 28.06492042541504, + "learning_rate": 1.9089224772389223e-07, + "loss": 0.9522, + "num_input_tokens_seen": 6220152832, + "step": 2966 + }, + { + "epoch": 0.15378356387306755, + "grad_norm": 13.24654483795166, + "learning_rate": 1.9064911565673328e-07, + "loss": 0.3005, + "num_input_tokens_seen": 6222249984, + "step": 2967 + }, + { + "epoch": 0.15405478708977488, + "grad_norm": 15.379100799560547, + "learning_rate": 1.9040627278149168e-07, + "loss": 0.4348, + "num_input_tokens_seen": 6224347136, + "step": 2968 + }, + { + "epoch": 0.15432601030648224, + "grad_norm": 23.722576141357422, + "learning_rate": 1.9016371929359824e-07, + "loss": 0.9692, + "num_input_tokens_seen": 6226444288, + "step": 2969 + }, + { + "epoch": 0.15459723352318958, + "grad_norm": 19.0823917388916, + "learning_rate": 1.8992145538825066e-07, + "loss": 0.6268, + "num_input_tokens_seen": 6228541440, + "step": 2970 + }, + { + "epoch": 0.15486845673989694, + "grad_norm": 15.874011993408203, + "learning_rate": 1.8967948126041383e-07, + "loss": 0.4841, + "num_input_tokens_seen": 6230638592, + "step": 2971 + }, + { + "epoch": 0.1551396799566043, + "grad_norm": 14.604007720947266, + "learning_rate": 1.8943779710481938e-07, + "loss": 0.5069, + "num_input_tokens_seen": 6232735744, + "step": 2972 + }, + { + "epoch": 0.15541090317331163, + "grad_norm": 20.07442283630371, + "learning_rate": 1.891964031159653e-07, + "loss": 0.7613, + "num_input_tokens_seen": 6234832896, + "step": 2973 + }, + { + "epoch": 0.155682126390019, + "grad_norm": 18.36520767211914, + "learning_rate": 1.8895529948811638e-07, + "loss": 0.6046, + "num_input_tokens_seen": 6236930048, + "step": 2974 + }, + { + "epoch": 0.15595334960672633, + "grad_norm": 15.209831237792969, + "learning_rate": 1.8871448641530353e-07, + "loss": 0.4244, + "num_input_tokens_seen": 6239027200, + "step": 2975 + }, + { + "epoch": 0.1562245728234337, + "grad_norm": 11.920095443725586, + "learning_rate": 1.8847396409132423e-07, + "loss": 0.3894, + "num_input_tokens_seen": 6241124352, + "step": 2976 + }, + { + "epoch": 0.15649579604014105, + "grad_norm": 13.935537338256836, + "learning_rate": 1.8823373270974164e-07, + "loss": 0.4652, + "num_input_tokens_seen": 6243221504, + "step": 2977 + }, + { + "epoch": 0.15676701925684838, + "grad_norm": 11.79638957977295, + "learning_rate": 1.8799379246388463e-07, + "loss": 0.3007, + "num_input_tokens_seen": 6245318656, + "step": 2978 + }, + { + "epoch": 0.15703824247355574, + "grad_norm": 20.775779724121094, + "learning_rate": 1.8775414354684804e-07, + "loss": 0.8416, + "num_input_tokens_seen": 6247415808, + "step": 2979 + }, + { + "epoch": 0.15730946569026308, + "grad_norm": 9.75412368774414, + "learning_rate": 1.875147861514923e-07, + "loss": 0.2322, + "num_input_tokens_seen": 6249512960, + "step": 2980 + }, + { + "epoch": 0.15758068890697044, + "grad_norm": 11.689945220947266, + "learning_rate": 1.8727572047044308e-07, + "loss": 0.3366, + "num_input_tokens_seen": 6251610112, + "step": 2981 + }, + { + "epoch": 0.15785191212367777, + "grad_norm": 12.28046703338623, + "learning_rate": 1.8703694669609133e-07, + "loss": 0.3674, + "num_input_tokens_seen": 6253707264, + "step": 2982 + }, + { + "epoch": 0.15812313534038513, + "grad_norm": 14.490015029907227, + "learning_rate": 1.8679846502059306e-07, + "loss": 0.4055, + "num_input_tokens_seen": 6255804416, + "step": 2983 + }, + { + "epoch": 0.1583943585570925, + "grad_norm": 14.019413948059082, + "learning_rate": 1.8656027563586934e-07, + "loss": 0.4061, + "num_input_tokens_seen": 6257901568, + "step": 2984 + }, + { + "epoch": 0.15866558177379983, + "grad_norm": 10.86307144165039, + "learning_rate": 1.863223787336059e-07, + "loss": 0.1615, + "num_input_tokens_seen": 6259998720, + "step": 2985 + }, + { + "epoch": 0.1589368049905072, + "grad_norm": 16.09626007080078, + "learning_rate": 1.8608477450525308e-07, + "loss": 0.5691, + "num_input_tokens_seen": 6262095872, + "step": 2986 + }, + { + "epoch": 0.15920802820721452, + "grad_norm": 11.85105037689209, + "learning_rate": 1.8584746314202574e-07, + "loss": 0.3223, + "num_input_tokens_seen": 6264193024, + "step": 2987 + }, + { + "epoch": 0.1594792514239219, + "grad_norm": 14.974807739257812, + "learning_rate": 1.8561044483490301e-07, + "loss": 0.5907, + "num_input_tokens_seen": 6266290176, + "step": 2988 + }, + { + "epoch": 0.15975047464062925, + "grad_norm": 13.03945255279541, + "learning_rate": 1.8537371977462835e-07, + "loss": 0.4826, + "num_input_tokens_seen": 6268387328, + "step": 2989 + }, + { + "epoch": 0.16002169785733658, + "grad_norm": 19.678932189941406, + "learning_rate": 1.8513728815170885e-07, + "loss": 0.6468, + "num_input_tokens_seen": 6270484480, + "step": 2990 + }, + { + "epoch": 0.16029292107404394, + "grad_norm": 22.712543487548828, + "learning_rate": 1.8490115015641566e-07, + "loss": 0.5548, + "num_input_tokens_seen": 6272581632, + "step": 2991 + }, + { + "epoch": 0.16056414429075128, + "grad_norm": 22.17213249206543, + "learning_rate": 1.846653059787839e-07, + "loss": 0.8835, + "num_input_tokens_seen": 6274678784, + "step": 2992 + }, + { + "epoch": 0.16083536750745864, + "grad_norm": 22.142024993896484, + "learning_rate": 1.8442975580861185e-07, + "loss": 0.791, + "num_input_tokens_seen": 6276775936, + "step": 2993 + }, + { + "epoch": 0.161106590724166, + "grad_norm": 16.552152633666992, + "learning_rate": 1.8419449983546136e-07, + "loss": 0.3504, + "num_input_tokens_seen": 6278873088, + "step": 2994 + }, + { + "epoch": 0.16137781394087333, + "grad_norm": 12.928352355957031, + "learning_rate": 1.8395953824865762e-07, + "loss": 0.3077, + "num_input_tokens_seen": 6280970240, + "step": 2995 + }, + { + "epoch": 0.1616490371575807, + "grad_norm": 18.581602096557617, + "learning_rate": 1.8372487123728844e-07, + "loss": 0.5233, + "num_input_tokens_seen": 6283067392, + "step": 2996 + }, + { + "epoch": 0.16192026037428803, + "grad_norm": 14.512155532836914, + "learning_rate": 1.8349049899020514e-07, + "loss": 0.4539, + "num_input_tokens_seen": 6285164544, + "step": 2997 + }, + { + "epoch": 0.1621914835909954, + "grad_norm": 11.53394603729248, + "learning_rate": 1.832564216960215e-07, + "loss": 0.3514, + "num_input_tokens_seen": 6287261696, + "step": 2998 + }, + { + "epoch": 0.16246270680770275, + "grad_norm": 13.360198020935059, + "learning_rate": 1.8302263954311408e-07, + "loss": 0.3348, + "num_input_tokens_seen": 6289358848, + "step": 2999 + }, + { + "epoch": 0.16273393002441008, + "grad_norm": 17.085603713989258, + "learning_rate": 1.827891527196218e-07, + "loss": 0.7366, + "num_input_tokens_seen": 6291456000, + "step": 3000 + }, + { + "epoch": 0.16300515324111745, + "grad_norm": 14.20055866241455, + "learning_rate": 1.8255596141344605e-07, + "loss": 0.2508, + "num_input_tokens_seen": 6293553152, + "step": 3001 + }, + { + "epoch": 0.16327637645782478, + "grad_norm": 14.382533073425293, + "learning_rate": 1.823230658122502e-07, + "loss": 0.4234, + "num_input_tokens_seen": 6295650304, + "step": 3002 + }, + { + "epoch": 0.16354759967453214, + "grad_norm": 19.175596237182617, + "learning_rate": 1.820904661034599e-07, + "loss": 0.4568, + "num_input_tokens_seen": 6297747456, + "step": 3003 + }, + { + "epoch": 0.1638188228912395, + "grad_norm": 13.114893913269043, + "learning_rate": 1.8185816247426245e-07, + "loss": 0.3285, + "num_input_tokens_seen": 6299844608, + "step": 3004 + }, + { + "epoch": 0.16409004610794684, + "grad_norm": 19.748310089111328, + "learning_rate": 1.8162615511160701e-07, + "loss": 0.6501, + "num_input_tokens_seen": 6301941760, + "step": 3005 + }, + { + "epoch": 0.1643612693246542, + "grad_norm": 16.454687118530273, + "learning_rate": 1.8139444420220438e-07, + "loss": 0.332, + "num_input_tokens_seen": 6304038912, + "step": 3006 + }, + { + "epoch": 0.16463249254136153, + "grad_norm": 19.270048141479492, + "learning_rate": 1.8116302993252636e-07, + "loss": 0.8047, + "num_input_tokens_seen": 6306136064, + "step": 3007 + }, + { + "epoch": 0.1649037157580689, + "grad_norm": 10.644708633422852, + "learning_rate": 1.809319124888067e-07, + "loss": 0.2935, + "num_input_tokens_seen": 6308233216, + "step": 3008 + }, + { + "epoch": 0.16517493897477625, + "grad_norm": 10.366632461547852, + "learning_rate": 1.8070109205703972e-07, + "loss": 0.2606, + "num_input_tokens_seen": 6310330368, + "step": 3009 + }, + { + "epoch": 0.1654461621914836, + "grad_norm": 16.61554527282715, + "learning_rate": 1.80470568822981e-07, + "loss": 0.5537, + "num_input_tokens_seen": 6312427520, + "step": 3010 + }, + { + "epoch": 0.16571738540819095, + "grad_norm": 20.610681533813477, + "learning_rate": 1.8024034297214686e-07, + "loss": 0.86, + "num_input_tokens_seen": 6314524672, + "step": 3011 + }, + { + "epoch": 0.16598860862489828, + "grad_norm": 15.841955184936523, + "learning_rate": 1.8001041468981442e-07, + "loss": 0.4544, + "num_input_tokens_seen": 6316621824, + "step": 3012 + }, + { + "epoch": 0.16625983184160564, + "grad_norm": 16.40627670288086, + "learning_rate": 1.79780784161021e-07, + "loss": 0.4465, + "num_input_tokens_seen": 6318718976, + "step": 3013 + }, + { + "epoch": 0.166531055058313, + "grad_norm": 18.409828186035156, + "learning_rate": 1.7955145157056456e-07, + "loss": 0.6821, + "num_input_tokens_seen": 6320816128, + "step": 3014 + }, + { + "epoch": 0.16680227827502034, + "grad_norm": 17.844806671142578, + "learning_rate": 1.793224171030033e-07, + "loss": 0.6167, + "num_input_tokens_seen": 6322913280, + "step": 3015 + }, + { + "epoch": 0.1670735014917277, + "grad_norm": 12.6576566696167, + "learning_rate": 1.7909368094265532e-07, + "loss": 0.3874, + "num_input_tokens_seen": 6325010432, + "step": 3016 + }, + { + "epoch": 0.16734472470843503, + "grad_norm": 14.177213668823242, + "learning_rate": 1.788652432735991e-07, + "loss": 0.3264, + "num_input_tokens_seen": 6327107584, + "step": 3017 + }, + { + "epoch": 0.1676159479251424, + "grad_norm": 16.917213439941406, + "learning_rate": 1.7863710427967222e-07, + "loss": 0.2328, + "num_input_tokens_seen": 6329204736, + "step": 3018 + }, + { + "epoch": 0.16788717114184976, + "grad_norm": 21.851543426513672, + "learning_rate": 1.784092641444725e-07, + "loss": 0.5402, + "num_input_tokens_seen": 6331301888, + "step": 3019 + }, + { + "epoch": 0.1681583943585571, + "grad_norm": 21.96310043334961, + "learning_rate": 1.7818172305135683e-07, + "loss": 0.5764, + "num_input_tokens_seen": 6333399040, + "step": 3020 + }, + { + "epoch": 0.16842961757526445, + "grad_norm": 16.358335494995117, + "learning_rate": 1.7795448118344174e-07, + "loss": 0.4508, + "num_input_tokens_seen": 6335496192, + "step": 3021 + }, + { + "epoch": 0.16870084079197178, + "grad_norm": 19.15180206298828, + "learning_rate": 1.7772753872360286e-07, + "loss": 0.5401, + "num_input_tokens_seen": 6337593344, + "step": 3022 + }, + { + "epoch": 0.16897206400867915, + "grad_norm": 15.562478065490723, + "learning_rate": 1.7750089585447473e-07, + "loss": 0.4368, + "num_input_tokens_seen": 6339690496, + "step": 3023 + }, + { + "epoch": 0.16924328722538648, + "grad_norm": 14.692290306091309, + "learning_rate": 1.77274552758451e-07, + "loss": 0.3316, + "num_input_tokens_seen": 6341787648, + "step": 3024 + }, + { + "epoch": 0.16951451044209384, + "grad_norm": 12.466604232788086, + "learning_rate": 1.770485096176839e-07, + "loss": 0.3341, + "num_input_tokens_seen": 6343884800, + "step": 3025 + }, + { + "epoch": 0.1697857336588012, + "grad_norm": 15.91280746459961, + "learning_rate": 1.7682276661408448e-07, + "loss": 0.6699, + "num_input_tokens_seen": 6345981952, + "step": 3026 + }, + { + "epoch": 0.17005695687550854, + "grad_norm": 17.506284713745117, + "learning_rate": 1.7659732392932202e-07, + "loss": 0.4563, + "num_input_tokens_seen": 6348079104, + "step": 3027 + }, + { + "epoch": 0.1703281800922159, + "grad_norm": 11.954529762268066, + "learning_rate": 1.7637218174482422e-07, + "loss": 0.3093, + "num_input_tokens_seen": 6350176256, + "step": 3028 + }, + { + "epoch": 0.17059940330892323, + "grad_norm": 12.980586051940918, + "learning_rate": 1.7614734024177707e-07, + "loss": 0.4789, + "num_input_tokens_seen": 6352273408, + "step": 3029 + }, + { + "epoch": 0.1708706265256306, + "grad_norm": 16.40062713623047, + "learning_rate": 1.7592279960112424e-07, + "loss": 0.6716, + "num_input_tokens_seen": 6354370560, + "step": 3030 + }, + { + "epoch": 0.17114184974233795, + "grad_norm": 17.116609573364258, + "learning_rate": 1.7569856000356758e-07, + "loss": 0.3462, + "num_input_tokens_seen": 6356467712, + "step": 3031 + }, + { + "epoch": 0.1714130729590453, + "grad_norm": 17.028945922851562, + "learning_rate": 1.754746216295665e-07, + "loss": 0.64, + "num_input_tokens_seen": 6358564864, + "step": 3032 + }, + { + "epoch": 0.17168429617575265, + "grad_norm": 21.552492141723633, + "learning_rate": 1.7525098465933834e-07, + "loss": 0.7373, + "num_input_tokens_seen": 6360662016, + "step": 3033 + }, + { + "epoch": 0.17195551939245998, + "grad_norm": 15.610780715942383, + "learning_rate": 1.750276492728575e-07, + "loss": 0.2999, + "num_input_tokens_seen": 6362759168, + "step": 3034 + }, + { + "epoch": 0.17222674260916734, + "grad_norm": 11.220949172973633, + "learning_rate": 1.748046156498557e-07, + "loss": 0.3368, + "num_input_tokens_seen": 6364856320, + "step": 3035 + }, + { + "epoch": 0.1724979658258747, + "grad_norm": 16.892784118652344, + "learning_rate": 1.74581883969822e-07, + "loss": 0.571, + "num_input_tokens_seen": 6366953472, + "step": 3036 + }, + { + "epoch": 0.17276918904258204, + "grad_norm": 14.306039810180664, + "learning_rate": 1.7435945441200232e-07, + "loss": 0.4861, + "num_input_tokens_seen": 6369050624, + "step": 3037 + }, + { + "epoch": 0.1730404122592894, + "grad_norm": 16.46154022216797, + "learning_rate": 1.7413732715539954e-07, + "loss": 0.3976, + "num_input_tokens_seen": 6371147776, + "step": 3038 + }, + { + "epoch": 0.17331163547599673, + "grad_norm": 18.606840133666992, + "learning_rate": 1.7391550237877326e-07, + "loss": 0.6797, + "num_input_tokens_seen": 6373244928, + "step": 3039 + }, + { + "epoch": 0.1735828586927041, + "grad_norm": 15.219120979309082, + "learning_rate": 1.7369398026063958e-07, + "loss": 0.4798, + "num_input_tokens_seen": 6375342080, + "step": 3040 + }, + { + "epoch": 0.17385408190941146, + "grad_norm": 15.45386791229248, + "learning_rate": 1.7347276097927105e-07, + "loss": 0.6304, + "num_input_tokens_seen": 6377439232, + "step": 3041 + }, + { + "epoch": 0.1741253051261188, + "grad_norm": 15.114662170410156, + "learning_rate": 1.732518447126966e-07, + "loss": 0.4832, + "num_input_tokens_seen": 6379536384, + "step": 3042 + }, + { + "epoch": 0.17439652834282615, + "grad_norm": 20.997955322265625, + "learning_rate": 1.730312316387012e-07, + "loss": 0.4316, + "num_input_tokens_seen": 6381633536, + "step": 3043 + }, + { + "epoch": 0.17466775155953349, + "grad_norm": 13.511618614196777, + "learning_rate": 1.7281092193482582e-07, + "loss": 0.4517, + "num_input_tokens_seen": 6383730688, + "step": 3044 + }, + { + "epoch": 0.17493897477624085, + "grad_norm": 17.852827072143555, + "learning_rate": 1.7259091577836737e-07, + "loss": 0.5438, + "num_input_tokens_seen": 6385827840, + "step": 3045 + }, + { + "epoch": 0.1752101979929482, + "grad_norm": 15.084151268005371, + "learning_rate": 1.723712133463785e-07, + "loss": 0.5421, + "num_input_tokens_seen": 6387924992, + "step": 3046 + }, + { + "epoch": 0.17548142120965554, + "grad_norm": 12.10033893585205, + "learning_rate": 1.7215181481566727e-07, + "loss": 0.304, + "num_input_tokens_seen": 6390022144, + "step": 3047 + }, + { + "epoch": 0.1757526444263629, + "grad_norm": 12.1007719039917, + "learning_rate": 1.719327203627971e-07, + "loss": 0.2765, + "num_input_tokens_seen": 6392119296, + "step": 3048 + }, + { + "epoch": 0.17602386764307024, + "grad_norm": 14.346907615661621, + "learning_rate": 1.7171393016408715e-07, + "loss": 0.3953, + "num_input_tokens_seen": 6394216448, + "step": 3049 + }, + { + "epoch": 0.1762950908597776, + "grad_norm": 13.626975059509277, + "learning_rate": 1.7149544439561135e-07, + "loss": 0.3358, + "num_input_tokens_seen": 6396313600, + "step": 3050 + }, + { + "epoch": 0.17656631407648496, + "grad_norm": 15.245820999145508, + "learning_rate": 1.7127726323319874e-07, + "loss": 0.3938, + "num_input_tokens_seen": 6398410752, + "step": 3051 + }, + { + "epoch": 0.1768375372931923, + "grad_norm": 9.872340202331543, + "learning_rate": 1.7105938685243339e-07, + "loss": 0.2623, + "num_input_tokens_seen": 6400507904, + "step": 3052 + }, + { + "epoch": 0.17710876050989965, + "grad_norm": 15.143594741821289, + "learning_rate": 1.7084181542865355e-07, + "loss": 0.5046, + "num_input_tokens_seen": 6402605056, + "step": 3053 + }, + { + "epoch": 0.177379983726607, + "grad_norm": 17.209291458129883, + "learning_rate": 1.7062454913695264e-07, + "loss": 0.5746, + "num_input_tokens_seen": 6404702208, + "step": 3054 + }, + { + "epoch": 0.17765120694331435, + "grad_norm": 14.591382026672363, + "learning_rate": 1.7040758815217827e-07, + "loss": 0.3467, + "num_input_tokens_seen": 6406799360, + "step": 3055 + }, + { + "epoch": 0.1779224301600217, + "grad_norm": 16.17945098876953, + "learning_rate": 1.7019093264893236e-07, + "loss": 0.6641, + "num_input_tokens_seen": 6408896512, + "step": 3056 + }, + { + "epoch": 0.17819365337672904, + "grad_norm": 11.522351264953613, + "learning_rate": 1.6997458280157113e-07, + "loss": 0.3645, + "num_input_tokens_seen": 6410993664, + "step": 3057 + }, + { + "epoch": 0.1784648765934364, + "grad_norm": 19.618732452392578, + "learning_rate": 1.6975853878420457e-07, + "loss": 0.6604, + "num_input_tokens_seen": 6413090816, + "step": 3058 + }, + { + "epoch": 0.17873609981014374, + "grad_norm": 15.74427604675293, + "learning_rate": 1.6954280077069676e-07, + "loss": 0.3712, + "num_input_tokens_seen": 6415187968, + "step": 3059 + }, + { + "epoch": 0.1790073230268511, + "grad_norm": 16.00514793395996, + "learning_rate": 1.6932736893466536e-07, + "loss": 0.4623, + "num_input_tokens_seen": 6417285120, + "step": 3060 + }, + { + "epoch": 0.17927854624355846, + "grad_norm": 12.272290229797363, + "learning_rate": 1.691122434494818e-07, + "loss": 0.3603, + "num_input_tokens_seen": 6419382272, + "step": 3061 + }, + { + "epoch": 0.1795497694602658, + "grad_norm": 15.91757583618164, + "learning_rate": 1.688974244882709e-07, + "loss": 0.5306, + "num_input_tokens_seen": 6421479424, + "step": 3062 + }, + { + "epoch": 0.17982099267697316, + "grad_norm": 15.56859302520752, + "learning_rate": 1.686829122239108e-07, + "loss": 0.4703, + "num_input_tokens_seen": 6423576576, + "step": 3063 + }, + { + "epoch": 0.1800922158936805, + "grad_norm": 18.83671760559082, + "learning_rate": 1.684687068290326e-07, + "loss": 0.6076, + "num_input_tokens_seen": 6425673728, + "step": 3064 + }, + { + "epoch": 0.18036343911038785, + "grad_norm": 14.046809196472168, + "learning_rate": 1.682548084760208e-07, + "loss": 0.3576, + "num_input_tokens_seen": 6427770880, + "step": 3065 + }, + { + "epoch": 0.18063466232709519, + "grad_norm": 12.50328540802002, + "learning_rate": 1.6804121733701277e-07, + "loss": 0.3457, + "num_input_tokens_seen": 6429868032, + "step": 3066 + }, + { + "epoch": 0.18090588554380255, + "grad_norm": 22.34259605407715, + "learning_rate": 1.678279335838983e-07, + "loss": 1.0324, + "num_input_tokens_seen": 6431965184, + "step": 3067 + }, + { + "epoch": 0.1811771087605099, + "grad_norm": 12.437803268432617, + "learning_rate": 1.676149573883202e-07, + "loss": 0.3294, + "num_input_tokens_seen": 6434062336, + "step": 3068 + }, + { + "epoch": 0.18144833197721724, + "grad_norm": 19.77998161315918, + "learning_rate": 1.674022889216737e-07, + "loss": 0.4922, + "num_input_tokens_seen": 6436159488, + "step": 3069 + }, + { + "epoch": 0.1817195551939246, + "grad_norm": 16.024463653564453, + "learning_rate": 1.67189928355106e-07, + "loss": 0.3612, + "num_input_tokens_seen": 6438256640, + "step": 3070 + }, + { + "epoch": 0.18199077841063194, + "grad_norm": 18.659027099609375, + "learning_rate": 1.66977875859517e-07, + "loss": 0.5333, + "num_input_tokens_seen": 6440353792, + "step": 3071 + }, + { + "epoch": 0.1822620016273393, + "grad_norm": 12.65479850769043, + "learning_rate": 1.6676613160555846e-07, + "loss": 0.3801, + "num_input_tokens_seen": 6442450944, + "step": 3072 + }, + { + "epoch": 0.18253322484404666, + "grad_norm": 16.552722930908203, + "learning_rate": 1.665546957636339e-07, + "loss": 0.6134, + "num_input_tokens_seen": 6444548096, + "step": 3073 + }, + { + "epoch": 0.182804448060754, + "grad_norm": 11.289351463317871, + "learning_rate": 1.6634356850389926e-07, + "loss": 0.3346, + "num_input_tokens_seen": 6446645248, + "step": 3074 + }, + { + "epoch": 0.18307567127746135, + "grad_norm": 13.140058517456055, + "learning_rate": 1.6613274999626134e-07, + "loss": 0.4239, + "num_input_tokens_seen": 6448742400, + "step": 3075 + }, + { + "epoch": 0.1833468944941687, + "grad_norm": 16.135107040405273, + "learning_rate": 1.65922240410379e-07, + "loss": 0.5723, + "num_input_tokens_seen": 6450839552, + "step": 3076 + }, + { + "epoch": 0.18361811771087605, + "grad_norm": 12.902073860168457, + "learning_rate": 1.6571203991566234e-07, + "loss": 0.3716, + "num_input_tokens_seen": 6452936704, + "step": 3077 + }, + { + "epoch": 0.1838893409275834, + "grad_norm": 13.520687103271484, + "learning_rate": 1.6550214868127276e-07, + "loss": 0.272, + "num_input_tokens_seen": 6455033856, + "step": 3078 + }, + { + "epoch": 0.18416056414429074, + "grad_norm": 15.4642333984375, + "learning_rate": 1.6529256687612264e-07, + "loss": 0.4952, + "num_input_tokens_seen": 6457131008, + "step": 3079 + }, + { + "epoch": 0.1844317873609981, + "grad_norm": 11.49740982055664, + "learning_rate": 1.6508329466887548e-07, + "loss": 0.321, + "num_input_tokens_seen": 6459228160, + "step": 3080 + }, + { + "epoch": 0.18470301057770544, + "grad_norm": 19.20050048828125, + "learning_rate": 1.6487433222794566e-07, + "loss": 0.5862, + "num_input_tokens_seen": 6461325312, + "step": 3081 + }, + { + "epoch": 0.1849742337944128, + "grad_norm": 9.488527297973633, + "learning_rate": 1.6466567972149806e-07, + "loss": 0.2648, + "num_input_tokens_seen": 6463422464, + "step": 3082 + }, + { + "epoch": 0.18524545701112016, + "grad_norm": 12.057193756103516, + "learning_rate": 1.6445733731744842e-07, + "loss": 0.3546, + "num_input_tokens_seen": 6465519616, + "step": 3083 + }, + { + "epoch": 0.1855166802278275, + "grad_norm": 10.167229652404785, + "learning_rate": 1.642493051834627e-07, + "loss": 0.2825, + "num_input_tokens_seen": 6467616768, + "step": 3084 + }, + { + "epoch": 0.18578790344453486, + "grad_norm": 15.535594940185547, + "learning_rate": 1.6404158348695729e-07, + "loss": 0.4639, + "num_input_tokens_seen": 6469713920, + "step": 3085 + }, + { + "epoch": 0.1860591266612422, + "grad_norm": 14.652868270874023, + "learning_rate": 1.6383417239509878e-07, + "loss": 0.5855, + "num_input_tokens_seen": 6471811072, + "step": 3086 + }, + { + "epoch": 0.18633034987794955, + "grad_norm": 16.557931900024414, + "learning_rate": 1.6362707207480347e-07, + "loss": 0.729, + "num_input_tokens_seen": 6473908224, + "step": 3087 + }, + { + "epoch": 0.18660157309465691, + "grad_norm": 20.1795654296875, + "learning_rate": 1.6342028269273802e-07, + "loss": 0.694, + "num_input_tokens_seen": 6476005376, + "step": 3088 + }, + { + "epoch": 0.18687279631136425, + "grad_norm": 16.401769638061523, + "learning_rate": 1.6321380441531852e-07, + "loss": 0.6733, + "num_input_tokens_seen": 6478102528, + "step": 3089 + }, + { + "epoch": 0.1871440195280716, + "grad_norm": 12.176185607910156, + "learning_rate": 1.63007637408711e-07, + "loss": 0.3886, + "num_input_tokens_seen": 6480199680, + "step": 3090 + }, + { + "epoch": 0.18741524274477894, + "grad_norm": 14.493315696716309, + "learning_rate": 1.6280178183883081e-07, + "loss": 0.4193, + "num_input_tokens_seen": 6482296832, + "step": 3091 + }, + { + "epoch": 0.1876864659614863, + "grad_norm": 13.448966979980469, + "learning_rate": 1.6259623787134258e-07, + "loss": 0.3855, + "num_input_tokens_seen": 6484393984, + "step": 3092 + }, + { + "epoch": 0.18795768917819367, + "grad_norm": 14.371041297912598, + "learning_rate": 1.6239100567166026e-07, + "loss": 0.3938, + "num_input_tokens_seen": 6486491136, + "step": 3093 + }, + { + "epoch": 0.188228912394901, + "grad_norm": 11.803318977355957, + "learning_rate": 1.6218608540494693e-07, + "loss": 0.3518, + "num_input_tokens_seen": 6488588288, + "step": 3094 + }, + { + "epoch": 0.18850013561160836, + "grad_norm": 16.85552978515625, + "learning_rate": 1.619814772361147e-07, + "loss": 0.3771, + "num_input_tokens_seen": 6490685440, + "step": 3095 + }, + { + "epoch": 0.1887713588283157, + "grad_norm": 15.620574951171875, + "learning_rate": 1.6177718132982441e-07, + "loss": 0.6079, + "num_input_tokens_seen": 6492782592, + "step": 3096 + }, + { + "epoch": 0.18904258204502306, + "grad_norm": 20.06139373779297, + "learning_rate": 1.6157319785048555e-07, + "loss": 0.5848, + "num_input_tokens_seen": 6494879744, + "step": 3097 + }, + { + "epoch": 0.18931380526173042, + "grad_norm": 14.808846473693848, + "learning_rate": 1.6136952696225634e-07, + "loss": 0.3535, + "num_input_tokens_seen": 6496976896, + "step": 3098 + }, + { + "epoch": 0.18958502847843775, + "grad_norm": 16.144725799560547, + "learning_rate": 1.6116616882904332e-07, + "loss": 0.5276, + "num_input_tokens_seen": 6499074048, + "step": 3099 + }, + { + "epoch": 0.1898562516951451, + "grad_norm": 14.656734466552734, + "learning_rate": 1.6096312361450142e-07, + "loss": 0.4783, + "num_input_tokens_seen": 6501171200, + "step": 3100 + }, + { + "epoch": 0.19012747491185245, + "grad_norm": 16.5491943359375, + "learning_rate": 1.6076039148203373e-07, + "loss": 0.453, + "num_input_tokens_seen": 6503268352, + "step": 3101 + }, + { + "epoch": 0.1903986981285598, + "grad_norm": 9.266581535339355, + "learning_rate": 1.6055797259479125e-07, + "loss": 0.2459, + "num_input_tokens_seen": 6505365504, + "step": 3102 + }, + { + "epoch": 0.19066992134526717, + "grad_norm": 22.371381759643555, + "learning_rate": 1.6035586711567318e-07, + "loss": 0.6326, + "num_input_tokens_seen": 6507462656, + "step": 3103 + }, + { + "epoch": 0.1909411445619745, + "grad_norm": 17.789241790771484, + "learning_rate": 1.601540752073261e-07, + "loss": 0.6873, + "num_input_tokens_seen": 6509559808, + "step": 3104 + }, + { + "epoch": 0.19121236777868186, + "grad_norm": 17.25879669189453, + "learning_rate": 1.5995259703214445e-07, + "loss": 0.284, + "num_input_tokens_seen": 6511656960, + "step": 3105 + }, + { + "epoch": 0.1914835909953892, + "grad_norm": 12.322155952453613, + "learning_rate": 1.5975143275227039e-07, + "loss": 0.3976, + "num_input_tokens_seen": 6513754112, + "step": 3106 + }, + { + "epoch": 0.19175481421209656, + "grad_norm": 18.380308151245117, + "learning_rate": 1.5955058252959318e-07, + "loss": 0.4834, + "num_input_tokens_seen": 6515851264, + "step": 3107 + }, + { + "epoch": 0.1920260374288039, + "grad_norm": 17.78908920288086, + "learning_rate": 1.5935004652574947e-07, + "loss": 0.5889, + "num_input_tokens_seen": 6517948416, + "step": 3108 + }, + { + "epoch": 0.19229726064551125, + "grad_norm": 11.990644454956055, + "learning_rate": 1.591498249021231e-07, + "loss": 0.2529, + "num_input_tokens_seen": 6520045568, + "step": 3109 + }, + { + "epoch": 0.19256848386221861, + "grad_norm": 20.534889221191406, + "learning_rate": 1.5894991781984456e-07, + "loss": 0.6632, + "num_input_tokens_seen": 6522142720, + "step": 3110 + }, + { + "epoch": 0.19283970707892595, + "grad_norm": 7.436341285705566, + "learning_rate": 1.587503254397916e-07, + "loss": 0.1396, + "num_input_tokens_seen": 6524239872, + "step": 3111 + }, + { + "epoch": 0.1931109302956333, + "grad_norm": 11.600310325622559, + "learning_rate": 1.585510479225886e-07, + "loss": 0.242, + "num_input_tokens_seen": 6526337024, + "step": 3112 + }, + { + "epoch": 0.19338215351234064, + "grad_norm": 13.862635612487793, + "learning_rate": 1.5835208542860648e-07, + "loss": 0.3614, + "num_input_tokens_seen": 6528434176, + "step": 3113 + }, + { + "epoch": 0.193653376729048, + "grad_norm": 10.956169128417969, + "learning_rate": 1.581534381179627e-07, + "loss": 0.3421, + "num_input_tokens_seen": 6530531328, + "step": 3114 + }, + { + "epoch": 0.19392459994575537, + "grad_norm": 17.733055114746094, + "learning_rate": 1.5795510615052104e-07, + "loss": 0.5629, + "num_input_tokens_seen": 6532628480, + "step": 3115 + }, + { + "epoch": 0.1941958231624627, + "grad_norm": 17.17679214477539, + "learning_rate": 1.5775708968589155e-07, + "loss": 0.4017, + "num_input_tokens_seen": 6534725632, + "step": 3116 + }, + { + "epoch": 0.19446704637917006, + "grad_norm": 16.183635711669922, + "learning_rate": 1.575593888834303e-07, + "loss": 0.4437, + "num_input_tokens_seen": 6536822784, + "step": 3117 + }, + { + "epoch": 0.1947382695958774, + "grad_norm": 25.411428451538086, + "learning_rate": 1.5736200390223942e-07, + "loss": 0.8358, + "num_input_tokens_seen": 6538919936, + "step": 3118 + }, + { + "epoch": 0.19500949281258476, + "grad_norm": 28.588600158691406, + "learning_rate": 1.5716493490116684e-07, + "loss": 0.3588, + "num_input_tokens_seen": 6541017088, + "step": 3119 + }, + { + "epoch": 0.19528071602929212, + "grad_norm": 15.308911323547363, + "learning_rate": 1.5696818203880624e-07, + "loss": 0.4436, + "num_input_tokens_seen": 6543114240, + "step": 3120 + }, + { + "epoch": 0.19555193924599945, + "grad_norm": 14.383072853088379, + "learning_rate": 1.5677174547349655e-07, + "loss": 0.5685, + "num_input_tokens_seen": 6545211392, + "step": 3121 + }, + { + "epoch": 0.1958231624627068, + "grad_norm": 21.179271697998047, + "learning_rate": 1.565756253633228e-07, + "loss": 0.5654, + "num_input_tokens_seen": 6547308544, + "step": 3122 + }, + { + "epoch": 0.19609438567941415, + "grad_norm": 17.321855545043945, + "learning_rate": 1.5637982186611481e-07, + "loss": 0.4894, + "num_input_tokens_seen": 6549405696, + "step": 3123 + }, + { + "epoch": 0.1963656088961215, + "grad_norm": 13.507481575012207, + "learning_rate": 1.5618433513944787e-07, + "loss": 0.3423, + "num_input_tokens_seen": 6551502848, + "step": 3124 + }, + { + "epoch": 0.19663683211282887, + "grad_norm": 18.613161087036133, + "learning_rate": 1.5598916534064216e-07, + "loss": 0.5254, + "num_input_tokens_seen": 6553600000, + "step": 3125 + }, + { + "epoch": 0.1969080553295362, + "grad_norm": 13.173487663269043, + "learning_rate": 1.557943126267631e-07, + "loss": 0.4205, + "num_input_tokens_seen": 6555697152, + "step": 3126 + }, + { + "epoch": 0.19717927854624356, + "grad_norm": 13.874591827392578, + "learning_rate": 1.5559977715462044e-07, + "loss": 0.3249, + "num_input_tokens_seen": 6557794304, + "step": 3127 + }, + { + "epoch": 0.1974505017629509, + "grad_norm": 16.8161678314209, + "learning_rate": 1.5540555908076914e-07, + "loss": 0.6233, + "num_input_tokens_seen": 6559891456, + "step": 3128 + }, + { + "epoch": 0.19772172497965826, + "grad_norm": 16.189739227294922, + "learning_rate": 1.5521165856150841e-07, + "loss": 0.4502, + "num_input_tokens_seen": 6561988608, + "step": 3129 + }, + { + "epoch": 0.19799294819636562, + "grad_norm": 15.414895057678223, + "learning_rate": 1.55018075752882e-07, + "loss": 0.5046, + "num_input_tokens_seen": 6564085760, + "step": 3130 + }, + { + "epoch": 0.19826417141307295, + "grad_norm": 13.006385803222656, + "learning_rate": 1.5482481081067825e-07, + "loss": 0.3974, + "num_input_tokens_seen": 6566182912, + "step": 3131 + }, + { + "epoch": 0.19853539462978032, + "grad_norm": 15.883471488952637, + "learning_rate": 1.546318638904291e-07, + "loss": 0.4731, + "num_input_tokens_seen": 6568280064, + "step": 3132 + }, + { + "epoch": 0.19880661784648765, + "grad_norm": 11.235620498657227, + "learning_rate": 1.5443923514741103e-07, + "loss": 0.2018, + "num_input_tokens_seen": 6570377216, + "step": 3133 + }, + { + "epoch": 0.199077841063195, + "grad_norm": 15.52456283569336, + "learning_rate": 1.5424692473664429e-07, + "loss": 0.3148, + "num_input_tokens_seen": 6572474368, + "step": 3134 + }, + { + "epoch": 0.19934906427990237, + "grad_norm": 19.32282066345215, + "learning_rate": 1.5405493281289302e-07, + "loss": 0.6533, + "num_input_tokens_seen": 6574571520, + "step": 3135 + }, + { + "epoch": 0.1996202874966097, + "grad_norm": 18.52271270751953, + "learning_rate": 1.5386325953066494e-07, + "loss": 0.6963, + "num_input_tokens_seen": 6576668672, + "step": 3136 + }, + { + "epoch": 0.19989151071331707, + "grad_norm": 12.21414852142334, + "learning_rate": 1.5367190504421157e-07, + "loss": 0.3308, + "num_input_tokens_seen": 6578765824, + "step": 3137 + }, + { + "epoch": 0.2001627339300244, + "grad_norm": 18.998111724853516, + "learning_rate": 1.5348086950752753e-07, + "loss": 0.3845, + "num_input_tokens_seen": 6580862976, + "step": 3138 + }, + { + "epoch": 0.20043395714673176, + "grad_norm": 16.331748962402344, + "learning_rate": 1.532901530743511e-07, + "loss": 0.452, + "num_input_tokens_seen": 6582960128, + "step": 3139 + }, + { + "epoch": 0.20070518036343912, + "grad_norm": 13.402908325195312, + "learning_rate": 1.530997558981635e-07, + "loss": 0.4053, + "num_input_tokens_seen": 6585057280, + "step": 3140 + }, + { + "epoch": 0.20097640358014646, + "grad_norm": 17.369251251220703, + "learning_rate": 1.5290967813218919e-07, + "loss": 0.7292, + "num_input_tokens_seen": 6587154432, + "step": 3141 + }, + { + "epoch": 0.20124762679685382, + "grad_norm": 14.586211204528809, + "learning_rate": 1.527199199293955e-07, + "loss": 0.3989, + "num_input_tokens_seen": 6589251584, + "step": 3142 + }, + { + "epoch": 0.20151885001356115, + "grad_norm": 26.158672332763672, + "learning_rate": 1.525304814424927e-07, + "loss": 0.8601, + "num_input_tokens_seen": 6591348736, + "step": 3143 + }, + { + "epoch": 0.2017900732302685, + "grad_norm": 12.1050386428833, + "learning_rate": 1.5234136282393356e-07, + "loss": 0.3601, + "num_input_tokens_seen": 6593445888, + "step": 3144 + }, + { + "epoch": 0.20206129644697587, + "grad_norm": 19.57971954345703, + "learning_rate": 1.5215256422591357e-07, + "loss": 0.5809, + "num_input_tokens_seen": 6595543040, + "step": 3145 + }, + { + "epoch": 0.2023325196636832, + "grad_norm": 23.03449821472168, + "learning_rate": 1.5196408580037058e-07, + "loss": 0.2546, + "num_input_tokens_seen": 6597640192, + "step": 3146 + }, + { + "epoch": 0.20260374288039057, + "grad_norm": 16.772308349609375, + "learning_rate": 1.51775927698985e-07, + "loss": 0.2833, + "num_input_tokens_seen": 6599737344, + "step": 3147 + }, + { + "epoch": 0.2028749660970979, + "grad_norm": 12.067276954650879, + "learning_rate": 1.5158809007317926e-07, + "loss": 0.4126, + "num_input_tokens_seen": 6601834496, + "step": 3148 + }, + { + "epoch": 0.20314618931380526, + "grad_norm": 23.149063110351562, + "learning_rate": 1.514005730741181e-07, + "loss": 0.7943, + "num_input_tokens_seen": 6603931648, + "step": 3149 + }, + { + "epoch": 0.20341741253051263, + "grad_norm": 15.860441207885742, + "learning_rate": 1.512133768527077e-07, + "loss": 0.4636, + "num_input_tokens_seen": 6606028800, + "step": 3150 + }, + { + "epoch": 0.20368863574721996, + "grad_norm": 16.582843780517578, + "learning_rate": 1.510265015595967e-07, + "loss": 0.3957, + "num_input_tokens_seen": 6608125952, + "step": 3151 + }, + { + "epoch": 0.20395985896392732, + "grad_norm": 20.899213790893555, + "learning_rate": 1.5083994734517518e-07, + "loss": 0.6574, + "num_input_tokens_seen": 6610223104, + "step": 3152 + }, + { + "epoch": 0.20423108218063465, + "grad_norm": 12.626710891723633, + "learning_rate": 1.5065371435957484e-07, + "loss": 0.4247, + "num_input_tokens_seen": 6612320256, + "step": 3153 + }, + { + "epoch": 0.20450230539734202, + "grad_norm": 28.32889175415039, + "learning_rate": 1.5046780275266897e-07, + "loss": 0.7678, + "num_input_tokens_seen": 6614417408, + "step": 3154 + }, + { + "epoch": 0.20477352861404935, + "grad_norm": 16.073993682861328, + "learning_rate": 1.5028221267407207e-07, + "loss": 0.4877, + "num_input_tokens_seen": 6616514560, + "step": 3155 + }, + { + "epoch": 0.2050447518307567, + "grad_norm": 9.331436157226562, + "learning_rate": 1.5009694427314007e-07, + "loss": 0.2676, + "num_input_tokens_seen": 6618611712, + "step": 3156 + }, + { + "epoch": 0.20531597504746407, + "grad_norm": 13.70032787322998, + "learning_rate": 1.4991199769896983e-07, + "loss": 0.3683, + "num_input_tokens_seen": 6620708864, + "step": 3157 + }, + { + "epoch": 0.2055871982641714, + "grad_norm": 12.181900024414062, + "learning_rate": 1.497273731003994e-07, + "loss": 0.3804, + "num_input_tokens_seen": 6622806016, + "step": 3158 + }, + { + "epoch": 0.20585842148087877, + "grad_norm": 11.273839950561523, + "learning_rate": 1.4954307062600758e-07, + "loss": 0.3363, + "num_input_tokens_seen": 6624903168, + "step": 3159 + }, + { + "epoch": 0.2061296446975861, + "grad_norm": 18.64375114440918, + "learning_rate": 1.4935909042411412e-07, + "loss": 0.7101, + "num_input_tokens_seen": 6627000320, + "step": 3160 + }, + { + "epoch": 0.20640086791429346, + "grad_norm": 17.58132553100586, + "learning_rate": 1.4917543264277901e-07, + "loss": 0.5021, + "num_input_tokens_seen": 6629097472, + "step": 3161 + }, + { + "epoch": 0.20667209113100082, + "grad_norm": 11.847007751464844, + "learning_rate": 1.4899209742980317e-07, + "loss": 0.4044, + "num_input_tokens_seen": 6631194624, + "step": 3162 + }, + { + "epoch": 0.20694331434770816, + "grad_norm": 17.20060157775879, + "learning_rate": 1.488090849327279e-07, + "loss": 0.5288, + "num_input_tokens_seen": 6633291776, + "step": 3163 + }, + { + "epoch": 0.20721453756441552, + "grad_norm": 19.338298797607422, + "learning_rate": 1.4862639529883463e-07, + "loss": 0.3903, + "num_input_tokens_seen": 6635388928, + "step": 3164 + }, + { + "epoch": 0.20748576078112285, + "grad_norm": 13.957944869995117, + "learning_rate": 1.4844402867514503e-07, + "loss": 0.4285, + "num_input_tokens_seen": 6637486080, + "step": 3165 + }, + { + "epoch": 0.2077569839978302, + "grad_norm": 10.898870468139648, + "learning_rate": 1.4826198520842093e-07, + "loss": 0.3047, + "num_input_tokens_seen": 6639583232, + "step": 3166 + }, + { + "epoch": 0.20802820721453757, + "grad_norm": 12.25651741027832, + "learning_rate": 1.4808026504516374e-07, + "loss": 0.435, + "num_input_tokens_seen": 6641680384, + "step": 3167 + }, + { + "epoch": 0.2082994304312449, + "grad_norm": 17.442764282226562, + "learning_rate": 1.4789886833161506e-07, + "loss": 0.582, + "num_input_tokens_seen": 6643777536, + "step": 3168 + }, + { + "epoch": 0.20857065364795227, + "grad_norm": 10.577066421508789, + "learning_rate": 1.477177952137561e-07, + "loss": 0.3007, + "num_input_tokens_seen": 6645874688, + "step": 3169 + }, + { + "epoch": 0.2088418768646596, + "grad_norm": 18.119701385498047, + "learning_rate": 1.4753704583730754e-07, + "loss": 0.5214, + "num_input_tokens_seen": 6647971840, + "step": 3170 + }, + { + "epoch": 0.20911310008136696, + "grad_norm": 13.43950366973877, + "learning_rate": 1.4735662034772962e-07, + "loss": 0.4102, + "num_input_tokens_seen": 6650068992, + "step": 3171 + }, + { + "epoch": 0.20938432329807433, + "grad_norm": 16.050125122070312, + "learning_rate": 1.47176518890222e-07, + "loss": 0.3653, + "num_input_tokens_seen": 6652166144, + "step": 3172 + }, + { + "epoch": 0.20965554651478166, + "grad_norm": 10.848344802856445, + "learning_rate": 1.4699674160972337e-07, + "loss": 0.3367, + "num_input_tokens_seen": 6654263296, + "step": 3173 + }, + { + "epoch": 0.20992676973148902, + "grad_norm": 11.283642768859863, + "learning_rate": 1.4681728865091165e-07, + "loss": 0.2537, + "num_input_tokens_seen": 6656360448, + "step": 3174 + }, + { + "epoch": 0.21019799294819635, + "grad_norm": 19.65658187866211, + "learning_rate": 1.466381601582038e-07, + "loss": 0.521, + "num_input_tokens_seen": 6658457600, + "step": 3175 + }, + { + "epoch": 0.21046921616490372, + "grad_norm": 12.568547248840332, + "learning_rate": 1.4645935627575562e-07, + "loss": 0.2631, + "num_input_tokens_seen": 6660554752, + "step": 3176 + }, + { + "epoch": 0.21074043938161108, + "grad_norm": 17.255279541015625, + "learning_rate": 1.462808771474617e-07, + "loss": 0.6249, + "num_input_tokens_seen": 6662651904, + "step": 3177 + }, + { + "epoch": 0.2110116625983184, + "grad_norm": 10.227863311767578, + "learning_rate": 1.4610272291695503e-07, + "loss": 0.2266, + "num_input_tokens_seen": 6664749056, + "step": 3178 + }, + { + "epoch": 0.21128288581502577, + "grad_norm": 17.062204360961914, + "learning_rate": 1.4592489372760764e-07, + "loss": 0.4047, + "num_input_tokens_seen": 6666846208, + "step": 3179 + }, + { + "epoch": 0.2115541090317331, + "grad_norm": 13.168872833251953, + "learning_rate": 1.4574738972252953e-07, + "loss": 0.3819, + "num_input_tokens_seen": 6668943360, + "step": 3180 + }, + { + "epoch": 0.21182533224844047, + "grad_norm": 14.085567474365234, + "learning_rate": 1.4557021104456923e-07, + "loss": 0.4032, + "num_input_tokens_seen": 6671040512, + "step": 3181 + }, + { + "epoch": 0.21209655546514783, + "grad_norm": 14.73873519897461, + "learning_rate": 1.4539335783631346e-07, + "loss": 0.415, + "num_input_tokens_seen": 6673137664, + "step": 3182 + }, + { + "epoch": 0.21236777868185516, + "grad_norm": 12.23835277557373, + "learning_rate": 1.4521683024008687e-07, + "loss": 0.4222, + "num_input_tokens_seen": 6675234816, + "step": 3183 + }, + { + "epoch": 0.21263900189856252, + "grad_norm": 15.95200252532959, + "learning_rate": 1.450406283979521e-07, + "loss": 0.4986, + "num_input_tokens_seen": 6677331968, + "step": 3184 + }, + { + "epoch": 0.21291022511526986, + "grad_norm": 10.582295417785645, + "learning_rate": 1.4486475245170977e-07, + "loss": 0.2614, + "num_input_tokens_seen": 6679429120, + "step": 3185 + }, + { + "epoch": 0.21318144833197722, + "grad_norm": 20.6806583404541, + "learning_rate": 1.446892025428981e-07, + "loss": 0.4077, + "num_input_tokens_seen": 6681526272, + "step": 3186 + }, + { + "epoch": 0.21345267154868458, + "grad_norm": 10.347526550292969, + "learning_rate": 1.4451397881279298e-07, + "loss": 0.1838, + "num_input_tokens_seen": 6683623424, + "step": 3187 + }, + { + "epoch": 0.21372389476539191, + "grad_norm": 15.546320915222168, + "learning_rate": 1.4433908140240792e-07, + "loss": 0.4314, + "num_input_tokens_seen": 6685720576, + "step": 3188 + }, + { + "epoch": 0.21399511798209928, + "grad_norm": 14.439404487609863, + "learning_rate": 1.4416451045249354e-07, + "loss": 0.3794, + "num_input_tokens_seen": 6687817728, + "step": 3189 + }, + { + "epoch": 0.2142663411988066, + "grad_norm": 14.810487747192383, + "learning_rate": 1.43990266103538e-07, + "loss": 0.3653, + "num_input_tokens_seen": 6689914880, + "step": 3190 + }, + { + "epoch": 0.21453756441551397, + "grad_norm": 13.329632759094238, + "learning_rate": 1.4381634849576644e-07, + "loss": 0.3458, + "num_input_tokens_seen": 6692012032, + "step": 3191 + }, + { + "epoch": 0.21480878763222133, + "grad_norm": 11.149251937866211, + "learning_rate": 1.4364275776914112e-07, + "loss": 0.3134, + "num_input_tokens_seen": 6694109184, + "step": 3192 + }, + { + "epoch": 0.21508001084892867, + "grad_norm": 11.901870727539062, + "learning_rate": 1.4346949406336136e-07, + "loss": 0.3335, + "num_input_tokens_seen": 6696206336, + "step": 3193 + }, + { + "epoch": 0.21535123406563603, + "grad_norm": 10.92190933227539, + "learning_rate": 1.4329655751786318e-07, + "loss": 0.3104, + "num_input_tokens_seen": 6698303488, + "step": 3194 + }, + { + "epoch": 0.21562245728234336, + "grad_norm": 13.854144096374512, + "learning_rate": 1.4312394827181925e-07, + "loss": 0.4487, + "num_input_tokens_seen": 6700400640, + "step": 3195 + }, + { + "epoch": 0.21589368049905072, + "grad_norm": 13.08950138092041, + "learning_rate": 1.4295166646413898e-07, + "loss": 0.3734, + "num_input_tokens_seen": 6702497792, + "step": 3196 + }, + { + "epoch": 0.21616490371575806, + "grad_norm": 13.053932189941406, + "learning_rate": 1.4277971223346825e-07, + "loss": 0.3552, + "num_input_tokens_seen": 6704594944, + "step": 3197 + }, + { + "epoch": 0.21643612693246542, + "grad_norm": 17.404874801635742, + "learning_rate": 1.4260808571818917e-07, + "loss": 0.3981, + "num_input_tokens_seen": 6706692096, + "step": 3198 + }, + { + "epoch": 0.21670735014917278, + "grad_norm": 16.526578903198242, + "learning_rate": 1.4243678705642027e-07, + "loss": 0.3884, + "num_input_tokens_seen": 6708789248, + "step": 3199 + }, + { + "epoch": 0.2169785733658801, + "grad_norm": 16.661752700805664, + "learning_rate": 1.4226581638601635e-07, + "loss": 0.4724, + "num_input_tokens_seen": 6710886400, + "step": 3200 + }, + { + "epoch": 0.00027122321670735016, + "grad_norm": 15.5323486328125, + "learning_rate": 1.4209517384456786e-07, + "loss": 0.505, + "num_input_tokens_seen": 6712983552, + "step": 3201 + }, + { + "epoch": 0.0005424464334147003, + "grad_norm": 9.104025840759277, + "learning_rate": 1.4192485956940153e-07, + "loss": 0.2839, + "num_input_tokens_seen": 6715080704, + "step": 3202 + }, + { + "epoch": 0.0008136696501220504, + "grad_norm": 9.672213554382324, + "learning_rate": 1.4175487369757973e-07, + "loss": 0.2322, + "num_input_tokens_seen": 6717177856, + "step": 3203 + }, + { + "epoch": 0.0010848928668294006, + "grad_norm": 15.418540954589844, + "learning_rate": 1.4158521636590079e-07, + "loss": 0.5406, + "num_input_tokens_seen": 6719275008, + "step": 3204 + }, + { + "epoch": 0.0013561160835367507, + "grad_norm": 19.670473098754883, + "learning_rate": 1.4141588771089838e-07, + "loss": 0.7279, + "num_input_tokens_seen": 6721372160, + "step": 3205 + }, + { + "epoch": 0.0016273393002441008, + "grad_norm": 9.864188194274902, + "learning_rate": 1.412468878688418e-07, + "loss": 0.2668, + "num_input_tokens_seen": 6723469312, + "step": 3206 + }, + { + "epoch": 0.001898562516951451, + "grad_norm": 9.84155559539795, + "learning_rate": 1.410782169757356e-07, + "loss": 0.2388, + "num_input_tokens_seen": 6725566464, + "step": 3207 + }, + { + "epoch": 0.0021697857336588013, + "grad_norm": 12.789325714111328, + "learning_rate": 1.4090987516731977e-07, + "loss": 0.4207, + "num_input_tokens_seen": 6727663616, + "step": 3208 + }, + { + "epoch": 0.0024410089503661514, + "grad_norm": 13.97022819519043, + "learning_rate": 1.4074186257906934e-07, + "loss": 0.3951, + "num_input_tokens_seen": 6729760768, + "step": 3209 + }, + { + "epoch": 0.0027122321670735015, + "grad_norm": 20.8309268951416, + "learning_rate": 1.4057417934619453e-07, + "loss": 0.5764, + "num_input_tokens_seen": 6731857920, + "step": 3210 + }, + { + "epoch": 0.0029834553837808516, + "grad_norm": 17.302791595458984, + "learning_rate": 1.404068256036403e-07, + "loss": 0.5127, + "num_input_tokens_seen": 6733955072, + "step": 3211 + }, + { + "epoch": 0.0032546786004882017, + "grad_norm": 18.93463134765625, + "learning_rate": 1.4023980148608667e-07, + "loss": 0.4078, + "num_input_tokens_seen": 6736052224, + "step": 3212 + }, + { + "epoch": 0.003525901817195552, + "grad_norm": 9.369474411010742, + "learning_rate": 1.4007310712794827e-07, + "loss": 0.1545, + "num_input_tokens_seen": 6738149376, + "step": 3213 + }, + { + "epoch": 0.003797125033902902, + "grad_norm": 17.984560012817383, + "learning_rate": 1.3990674266337442e-07, + "loss": 0.5359, + "num_input_tokens_seen": 6740246528, + "step": 3214 + }, + { + "epoch": 0.0040683482506102524, + "grad_norm": 11.723960876464844, + "learning_rate": 1.3974070822624883e-07, + "loss": 0.3113, + "num_input_tokens_seen": 6742343680, + "step": 3215 + }, + { + "epoch": 0.0043395714673176026, + "grad_norm": 15.782196998596191, + "learning_rate": 1.3957500395018977e-07, + "loss": 0.5187, + "num_input_tokens_seen": 6744440832, + "step": 3216 + }, + { + "epoch": 0.004610794684024953, + "grad_norm": 17.8033390045166, + "learning_rate": 1.394096299685497e-07, + "loss": 0.5859, + "num_input_tokens_seen": 6746537984, + "step": 3217 + }, + { + "epoch": 0.004882017900732303, + "grad_norm": 12.155229568481445, + "learning_rate": 1.3924458641441532e-07, + "loss": 0.3767, + "num_input_tokens_seen": 6748635136, + "step": 3218 + }, + { + "epoch": 0.005153241117439653, + "grad_norm": 14.897374153137207, + "learning_rate": 1.3907987342060725e-07, + "loss": 0.5853, + "num_input_tokens_seen": 6750732288, + "step": 3219 + }, + { + "epoch": 0.005424464334147003, + "grad_norm": 18.14958381652832, + "learning_rate": 1.389154911196805e-07, + "loss": 0.5894, + "num_input_tokens_seen": 6752829440, + "step": 3220 + }, + { + "epoch": 0.005695687550854353, + "grad_norm": 12.628942489624023, + "learning_rate": 1.3875143964392355e-07, + "loss": 0.2947, + "num_input_tokens_seen": 6754926592, + "step": 3221 + }, + { + "epoch": 0.005966910767561703, + "grad_norm": 22.452190399169922, + "learning_rate": 1.3858771912535877e-07, + "loss": 0.8263, + "num_input_tokens_seen": 6757023744, + "step": 3222 + }, + { + "epoch": 0.006238133984269053, + "grad_norm": 9.461743354797363, + "learning_rate": 1.3842432969574238e-07, + "loss": 0.1619, + "num_input_tokens_seen": 6759120896, + "step": 3223 + }, + { + "epoch": 0.006509357200976403, + "grad_norm": 12.965723037719727, + "learning_rate": 1.382612714865637e-07, + "loss": 0.3498, + "num_input_tokens_seen": 6761218048, + "step": 3224 + }, + { + "epoch": 0.0067805804176837535, + "grad_norm": 20.17690086364746, + "learning_rate": 1.3809854462904596e-07, + "loss": 0.5391, + "num_input_tokens_seen": 6763315200, + "step": 3225 + }, + { + "epoch": 0.007051803634391104, + "grad_norm": 10.085182189941406, + "learning_rate": 1.3793614925414542e-07, + "loss": 0.1799, + "num_input_tokens_seen": 6765412352, + "step": 3226 + }, + { + "epoch": 0.007323026851098454, + "grad_norm": 11.336246490478516, + "learning_rate": 1.3777408549255182e-07, + "loss": 0.3041, + "num_input_tokens_seen": 6767509504, + "step": 3227 + }, + { + "epoch": 0.007594250067805804, + "grad_norm": 7.451137542724609, + "learning_rate": 1.3761235347468784e-07, + "loss": 0.1651, + "num_input_tokens_seen": 6769606656, + "step": 3228 + }, + { + "epoch": 0.007865473284513154, + "grad_norm": 11.728433609008789, + "learning_rate": 1.3745095333070926e-07, + "loss": 0.2284, + "num_input_tokens_seen": 6771703808, + "step": 3229 + }, + { + "epoch": 0.008136696501220505, + "grad_norm": 21.89021873474121, + "learning_rate": 1.3728988519050476e-07, + "loss": 0.9454, + "num_input_tokens_seen": 6773800960, + "step": 3230 + }, + { + "epoch": 0.008407919717927854, + "grad_norm": 13.923346519470215, + "learning_rate": 1.3712914918369595e-07, + "loss": 0.4593, + "num_input_tokens_seen": 6775898112, + "step": 3231 + }, + { + "epoch": 0.008679142934635205, + "grad_norm": 17.16434097290039, + "learning_rate": 1.369687454396369e-07, + "loss": 0.5116, + "num_input_tokens_seen": 6777995264, + "step": 3232 + }, + { + "epoch": 0.008950366151342554, + "grad_norm": 13.330866813659668, + "learning_rate": 1.3680867408741457e-07, + "loss": 0.3037, + "num_input_tokens_seen": 6780092416, + "step": 3233 + }, + { + "epoch": 0.009221589368049905, + "grad_norm": 28.39195442199707, + "learning_rate": 1.366489352558483e-07, + "loss": 0.6069, + "num_input_tokens_seen": 6782189568, + "step": 3234 + }, + { + "epoch": 0.009492812584757255, + "grad_norm": 23.18076515197754, + "learning_rate": 1.3648952907348957e-07, + "loss": 0.4345, + "num_input_tokens_seen": 6784286720, + "step": 3235 + }, + { + "epoch": 0.009764035801464606, + "grad_norm": 13.762514114379883, + "learning_rate": 1.3633045566862268e-07, + "loss": 0.4396, + "num_input_tokens_seen": 6786383872, + "step": 3236 + }, + { + "epoch": 0.010035259018171955, + "grad_norm": 13.543777465820312, + "learning_rate": 1.3617171516926382e-07, + "loss": 0.37, + "num_input_tokens_seen": 6788481024, + "step": 3237 + }, + { + "epoch": 0.010306482234879306, + "grad_norm": 12.794722557067871, + "learning_rate": 1.3601330770316122e-07, + "loss": 0.2349, + "num_input_tokens_seen": 6790578176, + "step": 3238 + }, + { + "epoch": 0.010577705451586655, + "grad_norm": 17.2214298248291, + "learning_rate": 1.3585523339779523e-07, + "loss": 0.3966, + "num_input_tokens_seen": 6792675328, + "step": 3239 + }, + { + "epoch": 0.010848928668294006, + "grad_norm": 21.573457717895508, + "learning_rate": 1.356974923803781e-07, + "loss": 0.7812, + "num_input_tokens_seen": 6794772480, + "step": 3240 + }, + { + "epoch": 0.011120151885001357, + "grad_norm": 15.258363723754883, + "learning_rate": 1.3554008477785367e-07, + "loss": 0.4462, + "num_input_tokens_seen": 6796869632, + "step": 3241 + }, + { + "epoch": 0.011391375101708706, + "grad_norm": 15.593113899230957, + "learning_rate": 1.3538301071689756e-07, + "loss": 0.3842, + "num_input_tokens_seen": 6798966784, + "step": 3242 + }, + { + "epoch": 0.011662598318416057, + "grad_norm": 15.24464225769043, + "learning_rate": 1.3522627032391715e-07, + "loss": 0.4702, + "num_input_tokens_seen": 6801063936, + "step": 3243 + }, + { + "epoch": 0.011933821535123406, + "grad_norm": 11.781341552734375, + "learning_rate": 1.3506986372505098e-07, + "loss": 0.3587, + "num_input_tokens_seen": 6803161088, + "step": 3244 + }, + { + "epoch": 0.012205044751830757, + "grad_norm": 19.120126724243164, + "learning_rate": 1.3491379104616938e-07, + "loss": 0.5466, + "num_input_tokens_seen": 6805258240, + "step": 3245 + }, + { + "epoch": 0.012476267968538107, + "grad_norm": 10.356958389282227, + "learning_rate": 1.3475805241287345e-07, + "loss": 0.2264, + "num_input_tokens_seen": 6807355392, + "step": 3246 + }, + { + "epoch": 0.012747491185245458, + "grad_norm": 17.21544075012207, + "learning_rate": 1.3460264795049577e-07, + "loss": 0.6428, + "num_input_tokens_seen": 6809452544, + "step": 3247 + }, + { + "epoch": 0.013018714401952807, + "grad_norm": 16.766895294189453, + "learning_rate": 1.344475777841e-07, + "loss": 0.3485, + "num_input_tokens_seen": 6811549696, + "step": 3248 + }, + { + "epoch": 0.013289937618660158, + "grad_norm": 11.51309585571289, + "learning_rate": 1.3429284203848074e-07, + "loss": 0.2704, + "num_input_tokens_seen": 6813646848, + "step": 3249 + }, + { + "epoch": 0.013561160835367507, + "grad_norm": 26.110618591308594, + "learning_rate": 1.3413844083816334e-07, + "loss": 1.218, + "num_input_tokens_seen": 6815744000, + "step": 3250 + }, + { + "epoch": 0.013832384052074858, + "grad_norm": 14.552084922790527, + "learning_rate": 1.3398437430740403e-07, + "loss": 0.3287, + "num_input_tokens_seen": 6817841152, + "step": 3251 + }, + { + "epoch": 0.014103607268782207, + "grad_norm": 11.300999641418457, + "learning_rate": 1.3383064257018978e-07, + "loss": 0.2987, + "num_input_tokens_seen": 6819938304, + "step": 3252 + }, + { + "epoch": 0.014374830485489558, + "grad_norm": 12.652029991149902, + "learning_rate": 1.3367724575023786e-07, + "loss": 0.3022, + "num_input_tokens_seen": 6822035456, + "step": 3253 + }, + { + "epoch": 0.014646053702196907, + "grad_norm": 18.908370971679688, + "learning_rate": 1.3352418397099638e-07, + "loss": 0.6609, + "num_input_tokens_seen": 6824132608, + "step": 3254 + }, + { + "epoch": 0.014917276918904258, + "grad_norm": 16.611713409423828, + "learning_rate": 1.333714573556435e-07, + "loss": 0.4249, + "num_input_tokens_seen": 6826229760, + "step": 3255 + }, + { + "epoch": 0.015188500135611608, + "grad_norm": 15.014352798461914, + "learning_rate": 1.332190660270879e-07, + "loss": 0.4945, + "num_input_tokens_seen": 6828326912, + "step": 3256 + }, + { + "epoch": 0.015459723352318959, + "grad_norm": 14.692705154418945, + "learning_rate": 1.3306701010796832e-07, + "loss": 0.4748, + "num_input_tokens_seen": 6830424064, + "step": 3257 + }, + { + "epoch": 0.015730946569026308, + "grad_norm": 15.815617561340332, + "learning_rate": 1.3291528972065347e-07, + "loss": 0.559, + "num_input_tokens_seen": 6832521216, + "step": 3258 + }, + { + "epoch": 0.01600216978573366, + "grad_norm": 13.448586463928223, + "learning_rate": 1.327639049872422e-07, + "loss": 0.3937, + "num_input_tokens_seen": 6834618368, + "step": 3259 + }, + { + "epoch": 0.01627339300244101, + "grad_norm": 14.515972137451172, + "learning_rate": 1.326128560295631e-07, + "loss": 0.4, + "num_input_tokens_seen": 6836715520, + "step": 3260 + }, + { + "epoch": 0.01654461621914836, + "grad_norm": 14.561549186706543, + "learning_rate": 1.324621429691748e-07, + "loss": 0.4807, + "num_input_tokens_seen": 6838812672, + "step": 3261 + }, + { + "epoch": 0.016815839435855708, + "grad_norm": 22.775854110717773, + "learning_rate": 1.3231176592736528e-07, + "loss": 0.545, + "num_input_tokens_seen": 6840909824, + "step": 3262 + }, + { + "epoch": 0.01708706265256306, + "grad_norm": 17.545738220214844, + "learning_rate": 1.3216172502515248e-07, + "loss": 0.6031, + "num_input_tokens_seen": 6843006976, + "step": 3263 + }, + { + "epoch": 0.01735828586927041, + "grad_norm": 19.9493350982666, + "learning_rate": 1.320120203832833e-07, + "loss": 0.5962, + "num_input_tokens_seen": 6845104128, + "step": 3264 + }, + { + "epoch": 0.01762950908597776, + "grad_norm": 16.441688537597656, + "learning_rate": 1.3186265212223459e-07, + "loss": 0.5388, + "num_input_tokens_seen": 6847201280, + "step": 3265 + }, + { + "epoch": 0.01790073230268511, + "grad_norm": 17.45583724975586, + "learning_rate": 1.3171362036221202e-07, + "loss": 0.394, + "num_input_tokens_seen": 6849298432, + "step": 3266 + }, + { + "epoch": 0.01817195551939246, + "grad_norm": 21.072965621948242, + "learning_rate": 1.3156492522315088e-07, + "loss": 0.7281, + "num_input_tokens_seen": 6851395584, + "step": 3267 + }, + { + "epoch": 0.01844317873609981, + "grad_norm": 12.47169303894043, + "learning_rate": 1.314165668247153e-07, + "loss": 0.4182, + "num_input_tokens_seen": 6853492736, + "step": 3268 + }, + { + "epoch": 0.01871440195280716, + "grad_norm": 17.800216674804688, + "learning_rate": 1.3126854528629843e-07, + "loss": 0.5971, + "num_input_tokens_seen": 6855589888, + "step": 3269 + }, + { + "epoch": 0.01898562516951451, + "grad_norm": 13.492378234863281, + "learning_rate": 1.3112086072702238e-07, + "loss": 0.3652, + "num_input_tokens_seen": 6857687040, + "step": 3270 + }, + { + "epoch": 0.01925684838622186, + "grad_norm": 16.468965530395508, + "learning_rate": 1.3097351326573814e-07, + "loss": 0.4573, + "num_input_tokens_seen": 6859784192, + "step": 3271 + }, + { + "epoch": 0.01952807160292921, + "grad_norm": 19.35951042175293, + "learning_rate": 1.3082650302102524e-07, + "loss": 0.6076, + "num_input_tokens_seen": 6861881344, + "step": 3272 + }, + { + "epoch": 0.019799294819636562, + "grad_norm": 8.793999671936035, + "learning_rate": 1.30679830111192e-07, + "loss": 0.184, + "num_input_tokens_seen": 6863978496, + "step": 3273 + }, + { + "epoch": 0.02007051803634391, + "grad_norm": 14.717864990234375, + "learning_rate": 1.3053349465427518e-07, + "loss": 0.4201, + "num_input_tokens_seen": 6866075648, + "step": 3274 + }, + { + "epoch": 0.02034174125305126, + "grad_norm": 13.672134399414062, + "learning_rate": 1.303874967680399e-07, + "loss": 0.3217, + "num_input_tokens_seen": 6868172800, + "step": 3275 + }, + { + "epoch": 0.02061296446975861, + "grad_norm": 13.778663635253906, + "learning_rate": 1.302418365699798e-07, + "loss": 0.2786, + "num_input_tokens_seen": 6870269952, + "step": 3276 + }, + { + "epoch": 0.020884187686465962, + "grad_norm": 24.35689353942871, + "learning_rate": 1.300965141773167e-07, + "loss": 0.935, + "num_input_tokens_seen": 6872367104, + "step": 3277 + }, + { + "epoch": 0.02115541090317331, + "grad_norm": 12.533529281616211, + "learning_rate": 1.2995152970700044e-07, + "loss": 0.3299, + "num_input_tokens_seen": 6874464256, + "step": 3278 + }, + { + "epoch": 0.02142663411988066, + "grad_norm": 15.271020889282227, + "learning_rate": 1.2980688327570906e-07, + "loss": 0.5584, + "num_input_tokens_seen": 6876561408, + "step": 3279 + }, + { + "epoch": 0.021697857336588012, + "grad_norm": 12.875490188598633, + "learning_rate": 1.296625749998485e-07, + "loss": 0.3324, + "num_input_tokens_seen": 6878658560, + "step": 3280 + }, + { + "epoch": 0.021969080553295363, + "grad_norm": 13.227240562438965, + "learning_rate": 1.2951860499555253e-07, + "loss": 0.2358, + "num_input_tokens_seen": 6880755712, + "step": 3281 + }, + { + "epoch": 0.022240303770002714, + "grad_norm": 16.106595993041992, + "learning_rate": 1.2937497337868274e-07, + "loss": 0.5393, + "num_input_tokens_seen": 6882852864, + "step": 3282 + }, + { + "epoch": 0.02251152698671006, + "grad_norm": 18.33268928527832, + "learning_rate": 1.2923168026482843e-07, + "loss": 0.5663, + "num_input_tokens_seen": 6884950016, + "step": 3283 + }, + { + "epoch": 0.022782750203417412, + "grad_norm": 13.891176223754883, + "learning_rate": 1.290887257693064e-07, + "loss": 0.3277, + "num_input_tokens_seen": 6887047168, + "step": 3284 + }, + { + "epoch": 0.023053973420124763, + "grad_norm": 22.6708984375, + "learning_rate": 1.2894611000716103e-07, + "loss": 0.539, + "num_input_tokens_seen": 6889144320, + "step": 3285 + }, + { + "epoch": 0.023325196636832114, + "grad_norm": 17.467336654663086, + "learning_rate": 1.28803833093164e-07, + "loss": 0.4937, + "num_input_tokens_seen": 6891241472, + "step": 3286 + }, + { + "epoch": 0.023596419853539462, + "grad_norm": 14.677574157714844, + "learning_rate": 1.286618951418144e-07, + "loss": 0.3119, + "num_input_tokens_seen": 6893338624, + "step": 3287 + }, + { + "epoch": 0.023867643070246813, + "grad_norm": 14.258770942687988, + "learning_rate": 1.285202962673385e-07, + "loss": 0.4642, + "num_input_tokens_seen": 6895435776, + "step": 3288 + }, + { + "epoch": 0.024138866286954164, + "grad_norm": 19.215715408325195, + "learning_rate": 1.2837903658368965e-07, + "loss": 0.6045, + "num_input_tokens_seen": 6897532928, + "step": 3289 + }, + { + "epoch": 0.024410089503661515, + "grad_norm": 16.469133377075195, + "learning_rate": 1.2823811620454833e-07, + "loss": 0.3608, + "num_input_tokens_seen": 6899630080, + "step": 3290 + }, + { + "epoch": 0.024681312720368862, + "grad_norm": 19.01442527770996, + "learning_rate": 1.2809753524332188e-07, + "loss": 0.5565, + "num_input_tokens_seen": 6901727232, + "step": 3291 + }, + { + "epoch": 0.024952535937076213, + "grad_norm": 18.02033042907715, + "learning_rate": 1.2795729381314434e-07, + "loss": 0.5867, + "num_input_tokens_seen": 6903824384, + "step": 3292 + }, + { + "epoch": 0.025223759153783564, + "grad_norm": 14.180427551269531, + "learning_rate": 1.2781739202687694e-07, + "loss": 0.4978, + "num_input_tokens_seen": 6905921536, + "step": 3293 + }, + { + "epoch": 0.025494982370490915, + "grad_norm": 18.889726638793945, + "learning_rate": 1.2767782999710715e-07, + "loss": 0.8013, + "num_input_tokens_seen": 6908018688, + "step": 3294 + }, + { + "epoch": 0.025766205587198263, + "grad_norm": 17.951160430908203, + "learning_rate": 1.2753860783614925e-07, + "loss": 0.6829, + "num_input_tokens_seen": 6910115840, + "step": 3295 + }, + { + "epoch": 0.026037428803905614, + "grad_norm": 13.510412216186523, + "learning_rate": 1.2739972565604396e-07, + "loss": 0.3965, + "num_input_tokens_seen": 6912212992, + "step": 3296 + }, + { + "epoch": 0.026308652020612965, + "grad_norm": 15.73293399810791, + "learning_rate": 1.2726118356855834e-07, + "loss": 0.6381, + "num_input_tokens_seen": 6914310144, + "step": 3297 + }, + { + "epoch": 0.026579875237320316, + "grad_norm": 16.247882843017578, + "learning_rate": 1.271229816851857e-07, + "loss": 0.3781, + "num_input_tokens_seen": 6916407296, + "step": 3298 + }, + { + "epoch": 0.026851098454027666, + "grad_norm": 15.698039054870605, + "learning_rate": 1.2698512011714578e-07, + "loss": 0.4067, + "num_input_tokens_seen": 6918504448, + "step": 3299 + }, + { + "epoch": 0.027122321670735014, + "grad_norm": 18.372669219970703, + "learning_rate": 1.2684759897538422e-07, + "loss": 0.4066, + "num_input_tokens_seen": 6920601600, + "step": 3300 + }, + { + "epoch": 0.027393544887442365, + "grad_norm": 17.756454467773438, + "learning_rate": 1.2671041837057287e-07, + "loss": 0.5158, + "num_input_tokens_seen": 6922698752, + "step": 3301 + }, + { + "epoch": 0.027664768104149716, + "grad_norm": 18.298004150390625, + "learning_rate": 1.2657357841310941e-07, + "loss": 0.449, + "num_input_tokens_seen": 6924795904, + "step": 3302 + }, + { + "epoch": 0.027935991320857067, + "grad_norm": 14.342238426208496, + "learning_rate": 1.2643707921311765e-07, + "loss": 0.341, + "num_input_tokens_seen": 6926893056, + "step": 3303 + }, + { + "epoch": 0.028207214537564414, + "grad_norm": 10.686249732971191, + "learning_rate": 1.2630092088044664e-07, + "loss": 0.2287, + "num_input_tokens_seen": 6928990208, + "step": 3304 + }, + { + "epoch": 0.028478437754271765, + "grad_norm": 13.416088104248047, + "learning_rate": 1.2616510352467158e-07, + "loss": 0.3559, + "num_input_tokens_seen": 6931087360, + "step": 3305 + }, + { + "epoch": 0.028749660970979116, + "grad_norm": 21.23432731628418, + "learning_rate": 1.2602962725509307e-07, + "loss": 0.7772, + "num_input_tokens_seen": 6933184512, + "step": 3306 + }, + { + "epoch": 0.029020884187686467, + "grad_norm": 16.567541122436523, + "learning_rate": 1.2589449218073728e-07, + "loss": 0.4244, + "num_input_tokens_seen": 6935281664, + "step": 3307 + }, + { + "epoch": 0.029292107404393815, + "grad_norm": 10.55752944946289, + "learning_rate": 1.2575969841035578e-07, + "loss": 0.3227, + "num_input_tokens_seen": 6937378816, + "step": 3308 + }, + { + "epoch": 0.029563330621101166, + "grad_norm": 17.855588912963867, + "learning_rate": 1.256252460524255e-07, + "loss": 0.63, + "num_input_tokens_seen": 6939475968, + "step": 3309 + }, + { + "epoch": 0.029834553837808517, + "grad_norm": 20.68912124633789, + "learning_rate": 1.2549113521514844e-07, + "loss": 0.5901, + "num_input_tokens_seen": 6941573120, + "step": 3310 + }, + { + "epoch": 0.030105777054515868, + "grad_norm": 11.569897651672363, + "learning_rate": 1.2535736600645212e-07, + "loss": 0.3302, + "num_input_tokens_seen": 6943670272, + "step": 3311 + }, + { + "epoch": 0.030377000271223215, + "grad_norm": 16.52699851989746, + "learning_rate": 1.2522393853398868e-07, + "loss": 0.4751, + "num_input_tokens_seen": 6945767424, + "step": 3312 + }, + { + "epoch": 0.030648223487930566, + "grad_norm": 15.689116477966309, + "learning_rate": 1.2509085290513563e-07, + "loss": 0.2153, + "num_input_tokens_seen": 6947864576, + "step": 3313 + }, + { + "epoch": 0.030919446704637917, + "grad_norm": 14.995532035827637, + "learning_rate": 1.2495810922699517e-07, + "loss": 0.4613, + "num_input_tokens_seen": 6949961728, + "step": 3314 + }, + { + "epoch": 0.031190669921345268, + "grad_norm": 12.99809455871582, + "learning_rate": 1.2482570760639439e-07, + "loss": 0.3528, + "num_input_tokens_seen": 6952058880, + "step": 3315 + }, + { + "epoch": 0.031461893138052616, + "grad_norm": 11.95545482635498, + "learning_rate": 1.24693648149885e-07, + "loss": 0.2344, + "num_input_tokens_seen": 6954156032, + "step": 3316 + }, + { + "epoch": 0.03173311635475997, + "grad_norm": 9.108811378479004, + "learning_rate": 1.2456193096374338e-07, + "loss": 0.2519, + "num_input_tokens_seen": 6956253184, + "step": 3317 + }, + { + "epoch": 0.03200433957146732, + "grad_norm": 31.258995056152344, + "learning_rate": 1.244305561539707e-07, + "loss": 0.6474, + "num_input_tokens_seen": 6958350336, + "step": 3318 + }, + { + "epoch": 0.032275562788174665, + "grad_norm": 16.379648208618164, + "learning_rate": 1.242995238262923e-07, + "loss": 0.496, + "num_input_tokens_seen": 6960447488, + "step": 3319 + }, + { + "epoch": 0.03254678600488202, + "grad_norm": 15.703734397888184, + "learning_rate": 1.2416883408615809e-07, + "loss": 0.4296, + "num_input_tokens_seen": 6962544640, + "step": 3320 + }, + { + "epoch": 0.03281800922158937, + "grad_norm": 14.436956405639648, + "learning_rate": 1.2403848703874202e-07, + "loss": 0.3409, + "num_input_tokens_seen": 6964641792, + "step": 3321 + }, + { + "epoch": 0.03308923243829672, + "grad_norm": 13.694087028503418, + "learning_rate": 1.2390848278894264e-07, + "loss": 0.4741, + "num_input_tokens_seen": 6966738944, + "step": 3322 + }, + { + "epoch": 0.03336045565500407, + "grad_norm": 15.317477226257324, + "learning_rate": 1.2377882144138222e-07, + "loss": 0.4036, + "num_input_tokens_seen": 6968836096, + "step": 3323 + }, + { + "epoch": 0.033631678871711417, + "grad_norm": 14.88959789276123, + "learning_rate": 1.2364950310040744e-07, + "loss": 0.5187, + "num_input_tokens_seen": 6970933248, + "step": 3324 + }, + { + "epoch": 0.03390290208841877, + "grad_norm": 10.9374361038208, + "learning_rate": 1.235205278700887e-07, + "loss": 0.2573, + "num_input_tokens_seen": 6973030400, + "step": 3325 + }, + { + "epoch": 0.03417412530512612, + "grad_norm": 20.69586181640625, + "learning_rate": 1.233918958542204e-07, + "loss": 0.7775, + "num_input_tokens_seen": 6975127552, + "step": 3326 + }, + { + "epoch": 0.034445348521833466, + "grad_norm": 16.079696655273438, + "learning_rate": 1.2326360715632069e-07, + "loss": 0.4014, + "num_input_tokens_seen": 6977224704, + "step": 3327 + }, + { + "epoch": 0.03471657173854082, + "grad_norm": 21.355201721191406, + "learning_rate": 1.231356618796314e-07, + "loss": 0.6694, + "num_input_tokens_seen": 6979321856, + "step": 3328 + }, + { + "epoch": 0.03498779495524817, + "grad_norm": 18.189098358154297, + "learning_rate": 1.2300806012711798e-07, + "loss": 0.6481, + "num_input_tokens_seen": 6981419008, + "step": 3329 + }, + { + "epoch": 0.03525901817195552, + "grad_norm": 8.602594375610352, + "learning_rate": 1.228808020014696e-07, + "loss": 0.1969, + "num_input_tokens_seen": 6983516160, + "step": 3330 + }, + { + "epoch": 0.03553024138866287, + "grad_norm": 14.284547805786133, + "learning_rate": 1.227538876050987e-07, + "loss": 0.4433, + "num_input_tokens_seen": 6985613312, + "step": 3331 + }, + { + "epoch": 0.03580146460537022, + "grad_norm": 15.345396995544434, + "learning_rate": 1.226273170401411e-07, + "loss": 0.3931, + "num_input_tokens_seen": 6987710464, + "step": 3332 + }, + { + "epoch": 0.03607268782207757, + "grad_norm": 19.23891258239746, + "learning_rate": 1.2250109040845589e-07, + "loss": 0.4526, + "num_input_tokens_seen": 6989807616, + "step": 3333 + }, + { + "epoch": 0.03634391103878492, + "grad_norm": 13.172293663024902, + "learning_rate": 1.2237520781162567e-07, + "loss": 0.251, + "num_input_tokens_seen": 6991904768, + "step": 3334 + }, + { + "epoch": 0.03661513425549227, + "grad_norm": 14.895088195800781, + "learning_rate": 1.222496693509559e-07, + "loss": 0.4675, + "num_input_tokens_seen": 6994001920, + "step": 3335 + }, + { + "epoch": 0.03688635747219962, + "grad_norm": 17.735687255859375, + "learning_rate": 1.2212447512747506e-07, + "loss": 0.4232, + "num_input_tokens_seen": 6996099072, + "step": 3336 + }, + { + "epoch": 0.03715758068890697, + "grad_norm": 10.56796646118164, + "learning_rate": 1.2199962524193492e-07, + "loss": 0.2311, + "num_input_tokens_seen": 6998196224, + "step": 3337 + }, + { + "epoch": 0.03742880390561432, + "grad_norm": 17.399538040161133, + "learning_rate": 1.218751197948096e-07, + "loss": 0.4626, + "num_input_tokens_seen": 7000293376, + "step": 3338 + }, + { + "epoch": 0.03770002712232167, + "grad_norm": 21.85934066772461, + "learning_rate": 1.2175095888629657e-07, + "loss": 0.7143, + "num_input_tokens_seen": 7002390528, + "step": 3339 + }, + { + "epoch": 0.03797125033902902, + "grad_norm": 20.307708740234375, + "learning_rate": 1.2162714261631575e-07, + "loss": 0.6817, + "num_input_tokens_seen": 7004487680, + "step": 3340 + }, + { + "epoch": 0.03824247355573637, + "grad_norm": 12.21007251739502, + "learning_rate": 1.215036710845097e-07, + "loss": 0.3089, + "num_input_tokens_seen": 7006584832, + "step": 3341 + }, + { + "epoch": 0.03851369677244372, + "grad_norm": 17.77985954284668, + "learning_rate": 1.213805443902437e-07, + "loss": 0.3941, + "num_input_tokens_seen": 7008681984, + "step": 3342 + }, + { + "epoch": 0.038784919989151075, + "grad_norm": 13.20972728729248, + "learning_rate": 1.212577626326054e-07, + "loss": 0.3183, + "num_input_tokens_seen": 7010779136, + "step": 3343 + }, + { + "epoch": 0.03905614320585842, + "grad_norm": 13.693172454833984, + "learning_rate": 1.2113532591040488e-07, + "loss": 0.3986, + "num_input_tokens_seen": 7012876288, + "step": 3344 + }, + { + "epoch": 0.03932736642256577, + "grad_norm": 18.79658317565918, + "learning_rate": 1.2101323432217454e-07, + "loss": 0.438, + "num_input_tokens_seen": 7014973440, + "step": 3345 + }, + { + "epoch": 0.039598589639273124, + "grad_norm": 15.080113410949707, + "learning_rate": 1.208914879661691e-07, + "loss": 0.3442, + "num_input_tokens_seen": 7017070592, + "step": 3346 + }, + { + "epoch": 0.03986981285598047, + "grad_norm": 10.353347778320312, + "learning_rate": 1.2077008694036527e-07, + "loss": 0.2322, + "num_input_tokens_seen": 7019167744, + "step": 3347 + }, + { + "epoch": 0.04014103607268782, + "grad_norm": 14.034728050231934, + "learning_rate": 1.2064903134246221e-07, + "loss": 0.4263, + "num_input_tokens_seen": 7021264896, + "step": 3348 + }, + { + "epoch": 0.040412259289395173, + "grad_norm": 16.963212966918945, + "learning_rate": 1.2052832126988053e-07, + "loss": 0.7119, + "num_input_tokens_seen": 7023362048, + "step": 3349 + }, + { + "epoch": 0.04068348250610252, + "grad_norm": 16.336538314819336, + "learning_rate": 1.2040795681976338e-07, + "loss": 0.5812, + "num_input_tokens_seen": 7025459200, + "step": 3350 + }, + { + "epoch": 0.040954705722809875, + "grad_norm": 17.866945266723633, + "learning_rate": 1.2028793808897537e-07, + "loss": 0.6111, + "num_input_tokens_seen": 7027556352, + "step": 3351 + }, + { + "epoch": 0.04122592893951722, + "grad_norm": 19.273788452148438, + "learning_rate": 1.20168265174103e-07, + "loss": 0.3194, + "num_input_tokens_seen": 7029653504, + "step": 3352 + }, + { + "epoch": 0.04149715215622457, + "grad_norm": 19.322425842285156, + "learning_rate": 1.2004893817145456e-07, + "loss": 0.3223, + "num_input_tokens_seen": 7031750656, + "step": 3353 + }, + { + "epoch": 0.041768375372931925, + "grad_norm": 19.382596969604492, + "learning_rate": 1.1992995717705982e-07, + "loss": 0.7825, + "num_input_tokens_seen": 7033847808, + "step": 3354 + }, + { + "epoch": 0.04203959858963927, + "grad_norm": 12.71596908569336, + "learning_rate": 1.198113222866702e-07, + "loss": 0.326, + "num_input_tokens_seen": 7035944960, + "step": 3355 + }, + { + "epoch": 0.04231082180634662, + "grad_norm": 12.268967628479004, + "learning_rate": 1.1969303359575845e-07, + "loss": 0.3764, + "num_input_tokens_seen": 7038042112, + "step": 3356 + }, + { + "epoch": 0.042582045023053974, + "grad_norm": 12.389196395874023, + "learning_rate": 1.1957509119951885e-07, + "loss": 0.3302, + "num_input_tokens_seen": 7040139264, + "step": 3357 + }, + { + "epoch": 0.04285326823976132, + "grad_norm": 13.899662971496582, + "learning_rate": 1.1945749519286694e-07, + "loss": 0.3105, + "num_input_tokens_seen": 7042236416, + "step": 3358 + }, + { + "epoch": 0.043124491456468676, + "grad_norm": 16.53765106201172, + "learning_rate": 1.193402456704396e-07, + "loss": 0.4031, + "num_input_tokens_seen": 7044333568, + "step": 3359 + }, + { + "epoch": 0.043395714673176024, + "grad_norm": 17.577390670776367, + "learning_rate": 1.1922334272659474e-07, + "loss": 0.5253, + "num_input_tokens_seen": 7046430720, + "step": 3360 + }, + { + "epoch": 0.04366693788988337, + "grad_norm": 14.368865966796875, + "learning_rate": 1.1910678645541137e-07, + "loss": 0.4683, + "num_input_tokens_seen": 7048527872, + "step": 3361 + }, + { + "epoch": 0.043938161106590726, + "grad_norm": 10.524388313293457, + "learning_rate": 1.1899057695068954e-07, + "loss": 0.2427, + "num_input_tokens_seen": 7050625024, + "step": 3362 + }, + { + "epoch": 0.04420938432329807, + "grad_norm": 18.85384178161621, + "learning_rate": 1.1887471430595028e-07, + "loss": 0.5473, + "num_input_tokens_seen": 7052722176, + "step": 3363 + }, + { + "epoch": 0.04448060754000543, + "grad_norm": 18.354103088378906, + "learning_rate": 1.187591986144354e-07, + "loss": 0.5499, + "num_input_tokens_seen": 7054819328, + "step": 3364 + }, + { + "epoch": 0.044751830756712775, + "grad_norm": 17.2979736328125, + "learning_rate": 1.1864402996910748e-07, + "loss": 0.7023, + "num_input_tokens_seen": 7056916480, + "step": 3365 + }, + { + "epoch": 0.04502305397342012, + "grad_norm": 14.632923126220703, + "learning_rate": 1.1852920846265002e-07, + "loss": 0.4373, + "num_input_tokens_seen": 7059013632, + "step": 3366 + }, + { + "epoch": 0.04529427719012748, + "grad_norm": 17.784364700317383, + "learning_rate": 1.1841473418746685e-07, + "loss": 0.6175, + "num_input_tokens_seen": 7061110784, + "step": 3367 + }, + { + "epoch": 0.045565500406834825, + "grad_norm": 13.559286117553711, + "learning_rate": 1.1830060723568256e-07, + "loss": 0.3332, + "num_input_tokens_seen": 7063207936, + "step": 3368 + }, + { + "epoch": 0.04583672362354217, + "grad_norm": 14.777471542358398, + "learning_rate": 1.1818682769914226e-07, + "loss": 0.5035, + "num_input_tokens_seen": 7065305088, + "step": 3369 + }, + { + "epoch": 0.04610794684024953, + "grad_norm": 18.318246841430664, + "learning_rate": 1.1807339566941123e-07, + "loss": 0.6274, + "num_input_tokens_seen": 7067402240, + "step": 3370 + }, + { + "epoch": 0.046379170056956874, + "grad_norm": 16.257017135620117, + "learning_rate": 1.1796031123777546e-07, + "loss": 0.6656, + "num_input_tokens_seen": 7069499392, + "step": 3371 + }, + { + "epoch": 0.04665039327366423, + "grad_norm": 16.286823272705078, + "learning_rate": 1.178475744952408e-07, + "loss": 0.3738, + "num_input_tokens_seen": 7071596544, + "step": 3372 + }, + { + "epoch": 0.046921616490371576, + "grad_norm": 13.061100959777832, + "learning_rate": 1.1773518553253352e-07, + "loss": 0.3498, + "num_input_tokens_seen": 7073693696, + "step": 3373 + }, + { + "epoch": 0.047192839707078924, + "grad_norm": 19.254472732543945, + "learning_rate": 1.1762314444010002e-07, + "loss": 0.5665, + "num_input_tokens_seen": 7075790848, + "step": 3374 + }, + { + "epoch": 0.04746406292378628, + "grad_norm": 9.76313591003418, + "learning_rate": 1.1751145130810675e-07, + "loss": 0.1893, + "num_input_tokens_seen": 7077888000, + "step": 3375 + }, + { + "epoch": 0.047735286140493625, + "grad_norm": 11.822184562683105, + "learning_rate": 1.1740010622644008e-07, + "loss": 0.3819, + "num_input_tokens_seen": 7079985152, + "step": 3376 + }, + { + "epoch": 0.04800650935720097, + "grad_norm": 19.319860458374023, + "learning_rate": 1.1728910928470629e-07, + "loss": 0.4923, + "num_input_tokens_seen": 7082082304, + "step": 3377 + }, + { + "epoch": 0.04827773257390833, + "grad_norm": 10.68764877319336, + "learning_rate": 1.1717846057223143e-07, + "loss": 0.2708, + "num_input_tokens_seen": 7084179456, + "step": 3378 + }, + { + "epoch": 0.048548955790615675, + "grad_norm": 18.42941665649414, + "learning_rate": 1.1706816017806142e-07, + "loss": 0.5531, + "num_input_tokens_seen": 7086276608, + "step": 3379 + }, + { + "epoch": 0.04882017900732303, + "grad_norm": 19.539121627807617, + "learning_rate": 1.1695820819096186e-07, + "loss": 0.5045, + "num_input_tokens_seen": 7088373760, + "step": 3380 + }, + { + "epoch": 0.04909140222403038, + "grad_norm": 22.56864356994629, + "learning_rate": 1.1684860469941785e-07, + "loss": 0.8717, + "num_input_tokens_seen": 7090470912, + "step": 3381 + }, + { + "epoch": 0.049362625440737724, + "grad_norm": 18.064945220947266, + "learning_rate": 1.1673934979163417e-07, + "loss": 0.5734, + "num_input_tokens_seen": 7092568064, + "step": 3382 + }, + { + "epoch": 0.04963384865744508, + "grad_norm": 12.753329277038574, + "learning_rate": 1.1663044355553493e-07, + "loss": 0.2754, + "num_input_tokens_seen": 7094665216, + "step": 3383 + }, + { + "epoch": 0.049905071874152426, + "grad_norm": 16.793869018554688, + "learning_rate": 1.1652188607876379e-07, + "loss": 0.4608, + "num_input_tokens_seen": 7096762368, + "step": 3384 + }, + { + "epoch": 0.05017629509085978, + "grad_norm": 18.119853973388672, + "learning_rate": 1.1641367744868372e-07, + "loss": 0.7662, + "num_input_tokens_seen": 7098859520, + "step": 3385 + }, + { + "epoch": 0.05044751830756713, + "grad_norm": 18.567729949951172, + "learning_rate": 1.1630581775237681e-07, + "loss": 0.3328, + "num_input_tokens_seen": 7100956672, + "step": 3386 + }, + { + "epoch": 0.050718741524274476, + "grad_norm": 10.636219024658203, + "learning_rate": 1.1619830707664446e-07, + "loss": 0.2226, + "num_input_tokens_seen": 7103053824, + "step": 3387 + }, + { + "epoch": 0.05098996474098183, + "grad_norm": 15.37312126159668, + "learning_rate": 1.160911455080073e-07, + "loss": 0.3721, + "num_input_tokens_seen": 7105150976, + "step": 3388 + }, + { + "epoch": 0.05126118795768918, + "grad_norm": 20.098434448242188, + "learning_rate": 1.1598433313270472e-07, + "loss": 0.5492, + "num_input_tokens_seen": 7107248128, + "step": 3389 + }, + { + "epoch": 0.051532411174396525, + "grad_norm": 17.885168075561523, + "learning_rate": 1.1587787003669523e-07, + "loss": 0.4829, + "num_input_tokens_seen": 7109345280, + "step": 3390 + }, + { + "epoch": 0.05180363439110388, + "grad_norm": 15.35325813293457, + "learning_rate": 1.1577175630565646e-07, + "loss": 0.4408, + "num_input_tokens_seen": 7111442432, + "step": 3391 + }, + { + "epoch": 0.05207485760781123, + "grad_norm": 29.616416931152344, + "learning_rate": 1.1566599202498456e-07, + "loss": 0.7706, + "num_input_tokens_seen": 7113539584, + "step": 3392 + }, + { + "epoch": 0.05234608082451858, + "grad_norm": 10.501668930053711, + "learning_rate": 1.1556057727979468e-07, + "loss": 0.296, + "num_input_tokens_seen": 7115636736, + "step": 3393 + }, + { + "epoch": 0.05261730404122593, + "grad_norm": 17.35365867614746, + "learning_rate": 1.1545551215492055e-07, + "loss": 0.6134, + "num_input_tokens_seen": 7117733888, + "step": 3394 + }, + { + "epoch": 0.05288852725793328, + "grad_norm": 22.813182830810547, + "learning_rate": 1.1535079673491458e-07, + "loss": 0.4376, + "num_input_tokens_seen": 7119831040, + "step": 3395 + }, + { + "epoch": 0.05315975047464063, + "grad_norm": 11.890691757202148, + "learning_rate": 1.1524643110404778e-07, + "loss": 0.3133, + "num_input_tokens_seen": 7121928192, + "step": 3396 + }, + { + "epoch": 0.05343097369134798, + "grad_norm": 13.04570484161377, + "learning_rate": 1.1514241534630963e-07, + "loss": 0.3544, + "num_input_tokens_seen": 7124025344, + "step": 3397 + }, + { + "epoch": 0.05370219690805533, + "grad_norm": 15.982794761657715, + "learning_rate": 1.1503874954540804e-07, + "loss": 0.4587, + "num_input_tokens_seen": 7126122496, + "step": 3398 + }, + { + "epoch": 0.05397342012476268, + "grad_norm": 17.98113441467285, + "learning_rate": 1.1493543378476927e-07, + "loss": 0.3568, + "num_input_tokens_seen": 7128219648, + "step": 3399 + }, + { + "epoch": 0.05424464334147003, + "grad_norm": 14.370156288146973, + "learning_rate": 1.1483246814753798e-07, + "loss": 0.464, + "num_input_tokens_seen": 7130316800, + "step": 3400 + }, + { + "epoch": 0.00027122321670735016, + "grad_norm": 12.226541519165039, + "learning_rate": 1.1472985271657697e-07, + "loss": 0.35, + "num_input_tokens_seen": 7132413952, + "step": 3401 + }, + { + "epoch": 0.0005424464334147003, + "grad_norm": 15.276947975158691, + "learning_rate": 1.1462758757446728e-07, + "loss": 0.5296, + "num_input_tokens_seen": 7134511104, + "step": 3402 + }, + { + "epoch": 0.0008136696501220504, + "grad_norm": 15.25047492980957, + "learning_rate": 1.1452567280350789e-07, + "loss": 0.3315, + "num_input_tokens_seen": 7136608256, + "step": 3403 + }, + { + "epoch": 0.0010848928668294006, + "grad_norm": 18.95487403869629, + "learning_rate": 1.1442410848571602e-07, + "loss": 0.369, + "num_input_tokens_seen": 7138705408, + "step": 3404 + }, + { + "epoch": 0.0013561160835367507, + "grad_norm": 15.813984870910645, + "learning_rate": 1.1432289470282683e-07, + "loss": 0.4243, + "num_input_tokens_seen": 7140802560, + "step": 3405 + }, + { + "epoch": 0.0016273393002441008, + "grad_norm": 15.477920532226562, + "learning_rate": 1.1422203153629312e-07, + "loss": 0.3994, + "num_input_tokens_seen": 7142899712, + "step": 3406 + }, + { + "epoch": 0.001898562516951451, + "grad_norm": 15.906928062438965, + "learning_rate": 1.1412151906728589e-07, + "loss": 0.4129, + "num_input_tokens_seen": 7144996864, + "step": 3407 + }, + { + "epoch": 0.0021697857336588013, + "grad_norm": 19.251590728759766, + "learning_rate": 1.1402135737669372e-07, + "loss": 0.7718, + "num_input_tokens_seen": 7147094016, + "step": 3408 + }, + { + "epoch": 0.0024410089503661514, + "grad_norm": 14.729852676391602, + "learning_rate": 1.1392154654512289e-07, + "loss": 0.4258, + "num_input_tokens_seen": 7149191168, + "step": 3409 + }, + { + "epoch": 0.0027122321670735015, + "grad_norm": 16.52750587463379, + "learning_rate": 1.1382208665289742e-07, + "loss": 0.6094, + "num_input_tokens_seen": 7151288320, + "step": 3410 + }, + { + "epoch": 0.0029834553837808516, + "grad_norm": 30.927867889404297, + "learning_rate": 1.1372297778005883e-07, + "loss": 0.7976, + "num_input_tokens_seen": 7153385472, + "step": 3411 + }, + { + "epoch": 0.0032546786004882017, + "grad_norm": 14.352428436279297, + "learning_rate": 1.1362422000636609e-07, + "loss": 0.3816, + "num_input_tokens_seen": 7155482624, + "step": 3412 + }, + { + "epoch": 0.003525901817195552, + "grad_norm": 18.680057525634766, + "learning_rate": 1.135258134112958e-07, + "loss": 0.4478, + "num_input_tokens_seen": 7157579776, + "step": 3413 + }, + { + "epoch": 0.003797125033902902, + "grad_norm": 13.85073471069336, + "learning_rate": 1.1342775807404177e-07, + "loss": 0.4313, + "num_input_tokens_seen": 7159676928, + "step": 3414 + }, + { + "epoch": 0.0040683482506102524, + "grad_norm": 12.328075408935547, + "learning_rate": 1.1333005407351516e-07, + "loss": 0.301, + "num_input_tokens_seen": 7161774080, + "step": 3415 + }, + { + "epoch": 0.0043395714673176026, + "grad_norm": 15.454139709472656, + "learning_rate": 1.1323270148834461e-07, + "loss": 0.6079, + "num_input_tokens_seen": 7163871232, + "step": 3416 + }, + { + "epoch": 0.004610794684024953, + "grad_norm": 17.02078628540039, + "learning_rate": 1.1313570039687571e-07, + "loss": 0.5509, + "num_input_tokens_seen": 7165968384, + "step": 3417 + }, + { + "epoch": 0.004882017900732303, + "grad_norm": 16.010780334472656, + "learning_rate": 1.1303905087717111e-07, + "loss": 0.5081, + "num_input_tokens_seen": 7168065536, + "step": 3418 + }, + { + "epoch": 0.005153241117439653, + "grad_norm": 12.919666290283203, + "learning_rate": 1.1294275300701085e-07, + "loss": 0.3705, + "num_input_tokens_seen": 7170162688, + "step": 3419 + }, + { + "epoch": 0.005424464334147003, + "grad_norm": 14.903799057006836, + "learning_rate": 1.1284680686389163e-07, + "loss": 0.5029, + "num_input_tokens_seen": 7172259840, + "step": 3420 + }, + { + "epoch": 0.005695687550854353, + "grad_norm": 18.086734771728516, + "learning_rate": 1.1275121252502738e-07, + "loss": 0.408, + "num_input_tokens_seen": 7174356992, + "step": 3421 + }, + { + "epoch": 0.005966910767561703, + "grad_norm": 13.25420093536377, + "learning_rate": 1.1265597006734872e-07, + "loss": 0.3967, + "num_input_tokens_seen": 7176454144, + "step": 3422 + }, + { + "epoch": 0.006238133984269053, + "grad_norm": 16.535417556762695, + "learning_rate": 1.1256107956750319e-07, + "loss": 0.4737, + "num_input_tokens_seen": 7178551296, + "step": 3423 + }, + { + "epoch": 0.006509357200976403, + "grad_norm": 13.090187072753906, + "learning_rate": 1.1246654110185501e-07, + "loss": 0.2047, + "num_input_tokens_seen": 7180648448, + "step": 3424 + }, + { + "epoch": 0.0067805804176837535, + "grad_norm": 21.00216293334961, + "learning_rate": 1.1237235474648516e-07, + "loss": 0.8687, + "num_input_tokens_seen": 7182745600, + "step": 3425 + }, + { + "epoch": 0.007051803634391104, + "grad_norm": 18.46866226196289, + "learning_rate": 1.1227852057719125e-07, + "loss": 0.5996, + "num_input_tokens_seen": 7184842752, + "step": 3426 + }, + { + "epoch": 0.007323026851098454, + "grad_norm": 21.20566749572754, + "learning_rate": 1.121850386694875e-07, + "loss": 0.4136, + "num_input_tokens_seen": 7186939904, + "step": 3427 + }, + { + "epoch": 0.007594250067805804, + "grad_norm": 15.836469650268555, + "learning_rate": 1.1209190909860453e-07, + "loss": 0.5273, + "num_input_tokens_seen": 7189037056, + "step": 3428 + }, + { + "epoch": 0.007865473284513154, + "grad_norm": 17.805950164794922, + "learning_rate": 1.119991319394894e-07, + "loss": 0.4879, + "num_input_tokens_seen": 7191134208, + "step": 3429 + }, + { + "epoch": 0.008136696501220505, + "grad_norm": 21.2362117767334, + "learning_rate": 1.1190670726680579e-07, + "loss": 0.6706, + "num_input_tokens_seen": 7193231360, + "step": 3430 + }, + { + "epoch": 0.008407919717927854, + "grad_norm": 12.916275024414062, + "learning_rate": 1.1181463515493336e-07, + "loss": 0.2483, + "num_input_tokens_seen": 7195328512, + "step": 3431 + }, + { + "epoch": 0.008679142934635205, + "grad_norm": 14.223495483398438, + "learning_rate": 1.1172291567796846e-07, + "loss": 0.3723, + "num_input_tokens_seen": 7197425664, + "step": 3432 + }, + { + "epoch": 0.008950366151342554, + "grad_norm": 12.331622123718262, + "learning_rate": 1.1163154890972333e-07, + "loss": 0.3586, + "num_input_tokens_seen": 7199522816, + "step": 3433 + }, + { + "epoch": 0.009221589368049905, + "grad_norm": 9.547661781311035, + "learning_rate": 1.1154053492372654e-07, + "loss": 0.2378, + "num_input_tokens_seen": 7201619968, + "step": 3434 + }, + { + "epoch": 0.009492812584757255, + "grad_norm": 19.91844367980957, + "learning_rate": 1.1144987379322254e-07, + "loss": 0.5829, + "num_input_tokens_seen": 7203717120, + "step": 3435 + }, + { + "epoch": 0.009764035801464606, + "grad_norm": 23.920061111450195, + "learning_rate": 1.1135956559117207e-07, + "loss": 0.7757, + "num_input_tokens_seen": 7205814272, + "step": 3436 + }, + { + "epoch": 0.010035259018171955, + "grad_norm": 9.220373153686523, + "learning_rate": 1.1126961039025168e-07, + "loss": 0.2022, + "num_input_tokens_seen": 7207911424, + "step": 3437 + }, + { + "epoch": 0.010306482234879306, + "grad_norm": 14.202072143554688, + "learning_rate": 1.111800082628539e-07, + "loss": 0.3518, + "num_input_tokens_seen": 7210008576, + "step": 3438 + }, + { + "epoch": 0.010577705451586655, + "grad_norm": 16.335010528564453, + "learning_rate": 1.1109075928108715e-07, + "loss": 0.5367, + "num_input_tokens_seen": 7212105728, + "step": 3439 + }, + { + "epoch": 0.010848928668294006, + "grad_norm": 20.81877326965332, + "learning_rate": 1.1100186351677567e-07, + "loss": 0.8455, + "num_input_tokens_seen": 7214202880, + "step": 3440 + }, + { + "epoch": 0.011120151885001357, + "grad_norm": 12.109017372131348, + "learning_rate": 1.1091332104145921e-07, + "loss": 0.404, + "num_input_tokens_seen": 7216300032, + "step": 3441 + }, + { + "epoch": 0.011391375101708706, + "grad_norm": 51.72409439086914, + "learning_rate": 1.1082513192639353e-07, + "loss": 0.4581, + "num_input_tokens_seen": 7218397184, + "step": 3442 + }, + { + "epoch": 0.011662598318416057, + "grad_norm": 11.072718620300293, + "learning_rate": 1.1073729624254984e-07, + "loss": 0.2449, + "num_input_tokens_seen": 7220494336, + "step": 3443 + }, + { + "epoch": 0.011933821535123406, + "grad_norm": 20.055789947509766, + "learning_rate": 1.1064981406061494e-07, + "loss": 0.6404, + "num_input_tokens_seen": 7222591488, + "step": 3444 + }, + { + "epoch": 0.012205044751830757, + "grad_norm": 12.380927085876465, + "learning_rate": 1.1056268545099117e-07, + "loss": 0.4006, + "num_input_tokens_seen": 7224688640, + "step": 3445 + }, + { + "epoch": 0.012476267968538107, + "grad_norm": 13.249361038208008, + "learning_rate": 1.1047591048379635e-07, + "loss": 0.2531, + "num_input_tokens_seen": 7226785792, + "step": 3446 + }, + { + "epoch": 0.012747491185245458, + "grad_norm": 16.427322387695312, + "learning_rate": 1.1038948922886355e-07, + "loss": 0.5931, + "num_input_tokens_seen": 7228882944, + "step": 3447 + }, + { + "epoch": 0.013018714401952807, + "grad_norm": 16.59769630432129, + "learning_rate": 1.1030342175574144e-07, + "loss": 0.4725, + "num_input_tokens_seen": 7230980096, + "step": 3448 + }, + { + "epoch": 0.013289937618660158, + "grad_norm": 25.384323120117188, + "learning_rate": 1.1021770813369378e-07, + "loss": 0.6082, + "num_input_tokens_seen": 7233077248, + "step": 3449 + }, + { + "epoch": 0.013561160835367507, + "grad_norm": 14.333030700683594, + "learning_rate": 1.1013234843169967e-07, + "loss": 0.2975, + "num_input_tokens_seen": 7235174400, + "step": 3450 + }, + { + "epoch": 0.013832384052074858, + "grad_norm": 11.298495292663574, + "learning_rate": 1.100473427184534e-07, + "loss": 0.2593, + "num_input_tokens_seen": 7237271552, + "step": 3451 + }, + { + "epoch": 0.014103607268782207, + "grad_norm": 16.831668853759766, + "learning_rate": 1.0996269106236425e-07, + "loss": 0.5843, + "num_input_tokens_seen": 7239368704, + "step": 3452 + }, + { + "epoch": 0.014374830485489558, + "grad_norm": 12.40096664428711, + "learning_rate": 1.0987839353155661e-07, + "loss": 0.3489, + "num_input_tokens_seen": 7241465856, + "step": 3453 + }, + { + "epoch": 0.014646053702196907, + "grad_norm": 14.534159660339355, + "learning_rate": 1.0979445019387e-07, + "loss": 0.4209, + "num_input_tokens_seen": 7243563008, + "step": 3454 + }, + { + "epoch": 0.014917276918904258, + "grad_norm": 14.342657089233398, + "learning_rate": 1.0971086111685883e-07, + "loss": 0.3707, + "num_input_tokens_seen": 7245660160, + "step": 3455 + }, + { + "epoch": 0.015188500135611608, + "grad_norm": 19.60882568359375, + "learning_rate": 1.0962762636779235e-07, + "loss": 0.5742, + "num_input_tokens_seen": 7247757312, + "step": 3456 + }, + { + "epoch": 0.015459723352318959, + "grad_norm": 18.226106643676758, + "learning_rate": 1.0954474601365482e-07, + "loss": 0.6129, + "num_input_tokens_seen": 7249854464, + "step": 3457 + }, + { + "epoch": 0.015730946569026308, + "grad_norm": 15.364544868469238, + "learning_rate": 1.094622201211451e-07, + "loss": 0.438, + "num_input_tokens_seen": 7251951616, + "step": 3458 + }, + { + "epoch": 0.01600216978573366, + "grad_norm": 13.685408592224121, + "learning_rate": 1.0938004875667689e-07, + "loss": 0.4399, + "num_input_tokens_seen": 7254048768, + "step": 3459 + }, + { + "epoch": 0.01627339300244101, + "grad_norm": 13.642942428588867, + "learning_rate": 1.0929823198637866e-07, + "loss": 0.3581, + "num_input_tokens_seen": 7256145920, + "step": 3460 + }, + { + "epoch": 0.01654461621914836, + "grad_norm": 15.478311538696289, + "learning_rate": 1.0921676987609335e-07, + "loss": 0.5739, + "num_input_tokens_seen": 7258243072, + "step": 3461 + }, + { + "epoch": 0.016815839435855708, + "grad_norm": 15.546080589294434, + "learning_rate": 1.0913566249137865e-07, + "loss": 0.6112, + "num_input_tokens_seen": 7260340224, + "step": 3462 + }, + { + "epoch": 0.01708706265256306, + "grad_norm": 14.256020545959473, + "learning_rate": 1.0905490989750656e-07, + "loss": 0.532, + "num_input_tokens_seen": 7262437376, + "step": 3463 + }, + { + "epoch": 0.01735828586927041, + "grad_norm": 18.604835510253906, + "learning_rate": 1.0897451215946378e-07, + "loss": 0.673, + "num_input_tokens_seen": 7264534528, + "step": 3464 + }, + { + "epoch": 0.01762950908597776, + "grad_norm": 17.011693954467773, + "learning_rate": 1.0889446934195141e-07, + "loss": 0.5489, + "num_input_tokens_seen": 7266631680, + "step": 3465 + }, + { + "epoch": 0.01790073230268511, + "grad_norm": 13.888276100158691, + "learning_rate": 1.0881478150938475e-07, + "loss": 0.4176, + "num_input_tokens_seen": 7268728832, + "step": 3466 + }, + { + "epoch": 0.01817195551939246, + "grad_norm": 15.176491737365723, + "learning_rate": 1.0873544872589361e-07, + "loss": 0.396, + "num_input_tokens_seen": 7270825984, + "step": 3467 + }, + { + "epoch": 0.01844317873609981, + "grad_norm": 14.927526473999023, + "learning_rate": 1.08656471055322e-07, + "loss": 0.4393, + "num_input_tokens_seen": 7272923136, + "step": 3468 + }, + { + "epoch": 0.01871440195280716, + "grad_norm": 21.812307357788086, + "learning_rate": 1.0857784856122812e-07, + "loss": 0.8493, + "num_input_tokens_seen": 7275020288, + "step": 3469 + }, + { + "epoch": 0.01898562516951451, + "grad_norm": 14.81186580657959, + "learning_rate": 1.084995813068843e-07, + "loss": 0.3828, + "num_input_tokens_seen": 7277117440, + "step": 3470 + }, + { + "epoch": 0.01925684838622186, + "grad_norm": 16.475910186767578, + "learning_rate": 1.0842166935527716e-07, + "loss": 0.5323, + "num_input_tokens_seen": 7279214592, + "step": 3471 + }, + { + "epoch": 0.01952807160292921, + "grad_norm": 14.052254676818848, + "learning_rate": 1.0834411276910715e-07, + "loss": 0.3918, + "num_input_tokens_seen": 7281311744, + "step": 3472 + }, + { + "epoch": 0.019799294819636562, + "grad_norm": 17.08545684814453, + "learning_rate": 1.0826691161078895e-07, + "loss": 0.5653, + "num_input_tokens_seen": 7283408896, + "step": 3473 + }, + { + "epoch": 0.02007051803634391, + "grad_norm": 13.46341323852539, + "learning_rate": 1.0819006594245114e-07, + "loss": 0.3184, + "num_input_tokens_seen": 7285506048, + "step": 3474 + }, + { + "epoch": 0.02034174125305126, + "grad_norm": 18.920207977294922, + "learning_rate": 1.0811357582593613e-07, + "loss": 0.6962, + "num_input_tokens_seen": 7287603200, + "step": 3475 + }, + { + "epoch": 0.02061296446975861, + "grad_norm": 18.384702682495117, + "learning_rate": 1.0803744132280025e-07, + "loss": 0.8108, + "num_input_tokens_seen": 7289700352, + "step": 3476 + }, + { + "epoch": 0.020884187686465962, + "grad_norm": 19.510459899902344, + "learning_rate": 1.0796166249431371e-07, + "loss": 0.5891, + "num_input_tokens_seen": 7291797504, + "step": 3477 + }, + { + "epoch": 0.02115541090317331, + "grad_norm": 18.225069046020508, + "learning_rate": 1.0788623940146032e-07, + "loss": 0.5444, + "num_input_tokens_seen": 7293894656, + "step": 3478 + }, + { + "epoch": 0.02142663411988066, + "grad_norm": 13.333983421325684, + "learning_rate": 1.0781117210493781e-07, + "loss": 0.3763, + "num_input_tokens_seen": 7295991808, + "step": 3479 + }, + { + "epoch": 0.021697857336588012, + "grad_norm": 18.025068283081055, + "learning_rate": 1.0773646066515748e-07, + "loss": 0.6043, + "num_input_tokens_seen": 7298088960, + "step": 3480 + }, + { + "epoch": 0.021969080553295363, + "grad_norm": 11.711277961730957, + "learning_rate": 1.0766210514224419e-07, + "loss": 0.2852, + "num_input_tokens_seen": 7300186112, + "step": 3481 + }, + { + "epoch": 0.022240303770002714, + "grad_norm": 17.692590713500977, + "learning_rate": 1.0758810559603651e-07, + "loss": 0.4692, + "num_input_tokens_seen": 7302283264, + "step": 3482 + }, + { + "epoch": 0.02251152698671006, + "grad_norm": 13.046731948852539, + "learning_rate": 1.0751446208608642e-07, + "loss": 0.3152, + "num_input_tokens_seen": 7304380416, + "step": 3483 + }, + { + "epoch": 0.022782750203417412, + "grad_norm": 11.570886611938477, + "learning_rate": 1.0744117467165938e-07, + "loss": 0.4343, + "num_input_tokens_seen": 7306477568, + "step": 3484 + }, + { + "epoch": 0.023053973420124763, + "grad_norm": 13.106870651245117, + "learning_rate": 1.0736824341173442e-07, + "loss": 0.3972, + "num_input_tokens_seen": 7308574720, + "step": 3485 + }, + { + "epoch": 0.023325196636832114, + "grad_norm": 16.621774673461914, + "learning_rate": 1.0729566836500373e-07, + "loss": 0.4051, + "num_input_tokens_seen": 7310671872, + "step": 3486 + }, + { + "epoch": 0.023596419853539462, + "grad_norm": 18.08404541015625, + "learning_rate": 1.07223449589873e-07, + "loss": 0.4515, + "num_input_tokens_seen": 7312769024, + "step": 3487 + }, + { + "epoch": 0.023867643070246813, + "grad_norm": 10.743152618408203, + "learning_rate": 1.0715158714446109e-07, + "loss": 0.3258, + "num_input_tokens_seen": 7314866176, + "step": 3488 + }, + { + "epoch": 0.024138866286954164, + "grad_norm": 16.302248001098633, + "learning_rate": 1.0708008108660026e-07, + "loss": 0.4079, + "num_input_tokens_seen": 7316963328, + "step": 3489 + }, + { + "epoch": 0.024410089503661515, + "grad_norm": 10.805963516235352, + "learning_rate": 1.0700893147383582e-07, + "loss": 0.2726, + "num_input_tokens_seen": 7319060480, + "step": 3490 + }, + { + "epoch": 0.024681312720368862, + "grad_norm": 15.196581840515137, + "learning_rate": 1.069381383634263e-07, + "loss": 0.5156, + "num_input_tokens_seen": 7321157632, + "step": 3491 + }, + { + "epoch": 0.024952535937076213, + "grad_norm": 12.510798454284668, + "learning_rate": 1.0686770181234322e-07, + "loss": 0.4177, + "num_input_tokens_seen": 7323254784, + "step": 3492 + }, + { + "epoch": 0.025223759153783564, + "grad_norm": 14.379700660705566, + "learning_rate": 1.0679762187727129e-07, + "loss": 0.356, + "num_input_tokens_seen": 7325351936, + "step": 3493 + }, + { + "epoch": 0.025494982370490915, + "grad_norm": 10.159677505493164, + "learning_rate": 1.0672789861460818e-07, + "loss": 0.2734, + "num_input_tokens_seen": 7327449088, + "step": 3494 + }, + { + "epoch": 0.025766205587198263, + "grad_norm": 14.783156394958496, + "learning_rate": 1.0665853208046449e-07, + "loss": 0.5266, + "num_input_tokens_seen": 7329546240, + "step": 3495 + }, + { + "epoch": 0.026037428803905614, + "grad_norm": 16.891870498657227, + "learning_rate": 1.0658952233066381e-07, + "loss": 0.3847, + "num_input_tokens_seen": 7331643392, + "step": 3496 + }, + { + "epoch": 0.026308652020612965, + "grad_norm": 16.215871810913086, + "learning_rate": 1.0652086942074255e-07, + "loss": 0.5542, + "num_input_tokens_seen": 7333740544, + "step": 3497 + }, + { + "epoch": 0.026579875237320316, + "grad_norm": 11.77064323425293, + "learning_rate": 1.0645257340594988e-07, + "loss": 0.2651, + "num_input_tokens_seen": 7335837696, + "step": 3498 + }, + { + "epoch": 0.026851098454027666, + "grad_norm": 11.754627227783203, + "learning_rate": 1.06384634341248e-07, + "loss": 0.3247, + "num_input_tokens_seen": 7337934848, + "step": 3499 + }, + { + "epoch": 0.027122321670735014, + "grad_norm": 12.69980239868164, + "learning_rate": 1.0631705228131149e-07, + "loss": 0.3777, + "num_input_tokens_seen": 7340032000, + "step": 3500 + }, + { + "epoch": 0.027393544887442365, + "grad_norm": 13.112504959106445, + "learning_rate": 1.0624982728052795e-07, + "loss": 0.2684, + "num_input_tokens_seen": 7342129152, + "step": 3501 + }, + { + "epoch": 0.027664768104149716, + "grad_norm": 13.424850463867188, + "learning_rate": 1.0618295939299752e-07, + "loss": 0.3104, + "num_input_tokens_seen": 7344226304, + "step": 3502 + }, + { + "epoch": 0.027935991320857067, + "grad_norm": 17.421981811523438, + "learning_rate": 1.0611644867253284e-07, + "loss": 0.6812, + "num_input_tokens_seen": 7346323456, + "step": 3503 + }, + { + "epoch": 0.028207214537564414, + "grad_norm": 18.970033645629883, + "learning_rate": 1.0605029517265918e-07, + "loss": 0.6363, + "num_input_tokens_seen": 7348420608, + "step": 3504 + }, + { + "epoch": 0.028478437754271765, + "grad_norm": 35.321075439453125, + "learning_rate": 1.0598449894661445e-07, + "loss": 0.3609, + "num_input_tokens_seen": 7350517760, + "step": 3505 + }, + { + "epoch": 0.028749660970979116, + "grad_norm": 16.84616470336914, + "learning_rate": 1.0591906004734895e-07, + "loss": 0.6572, + "num_input_tokens_seen": 7352614912, + "step": 3506 + }, + { + "epoch": 0.029020884187686467, + "grad_norm": 18.14307403564453, + "learning_rate": 1.0585397852752544e-07, + "loss": 0.87, + "num_input_tokens_seen": 7354712064, + "step": 3507 + }, + { + "epoch": 0.029292107404393815, + "grad_norm": 16.01613426208496, + "learning_rate": 1.0578925443951895e-07, + "loss": 0.5537, + "num_input_tokens_seen": 7356809216, + "step": 3508 + }, + { + "epoch": 0.029563330621101166, + "grad_norm": 8.767593383789062, + "learning_rate": 1.0572488783541702e-07, + "loss": 0.2193, + "num_input_tokens_seen": 7358906368, + "step": 3509 + }, + { + "epoch": 0.029834553837808517, + "grad_norm": 11.045705795288086, + "learning_rate": 1.0566087876701941e-07, + "loss": 0.3346, + "num_input_tokens_seen": 7361003520, + "step": 3510 + }, + { + "epoch": 0.030105777054515868, + "grad_norm": 17.116188049316406, + "learning_rate": 1.0559722728583825e-07, + "loss": 0.5335, + "num_input_tokens_seen": 7363100672, + "step": 3511 + }, + { + "epoch": 0.030377000271223215, + "grad_norm": 10.3762845993042, + "learning_rate": 1.0553393344309775e-07, + "loss": 0.3105, + "num_input_tokens_seen": 7365197824, + "step": 3512 + }, + { + "epoch": 0.030648223487930566, + "grad_norm": 13.67977237701416, + "learning_rate": 1.054709972897344e-07, + "loss": 0.3106, + "num_input_tokens_seen": 7367294976, + "step": 3513 + }, + { + "epoch": 0.030919446704637917, + "grad_norm": 17.335601806640625, + "learning_rate": 1.0540841887639698e-07, + "loss": 0.6964, + "num_input_tokens_seen": 7369392128, + "step": 3514 + }, + { + "epoch": 0.031190669921345268, + "grad_norm": 15.248687744140625, + "learning_rate": 1.0534619825344596e-07, + "loss": 0.3862, + "num_input_tokens_seen": 7371489280, + "step": 3515 + }, + { + "epoch": 0.031461893138052616, + "grad_norm": 12.66186809539795, + "learning_rate": 1.052843354709543e-07, + "loss": 0.3237, + "num_input_tokens_seen": 7373586432, + "step": 3516 + }, + { + "epoch": 0.03173311635475997, + "grad_norm": 13.880027770996094, + "learning_rate": 1.0522283057870675e-07, + "loss": 0.3875, + "num_input_tokens_seen": 7375683584, + "step": 3517 + }, + { + "epoch": 0.03200433957146732, + "grad_norm": 18.82491683959961, + "learning_rate": 1.0516168362620013e-07, + "loss": 0.4796, + "num_input_tokens_seen": 7377780736, + "step": 3518 + }, + { + "epoch": 0.032275562788174665, + "grad_norm": 16.534194946289062, + "learning_rate": 1.0510089466264321e-07, + "loss": 0.4197, + "num_input_tokens_seen": 7379877888, + "step": 3519 + }, + { + "epoch": 0.03254678600488202, + "grad_norm": 10.575153350830078, + "learning_rate": 1.0504046373695648e-07, + "loss": 0.283, + "num_input_tokens_seen": 7381975040, + "step": 3520 + }, + { + "epoch": 0.03281800922158937, + "grad_norm": 16.936920166015625, + "learning_rate": 1.0498039089777265e-07, + "loss": 0.4884, + "num_input_tokens_seen": 7384072192, + "step": 3521 + }, + { + "epoch": 0.03308923243829672, + "grad_norm": 13.413094520568848, + "learning_rate": 1.0492067619343594e-07, + "loss": 0.4334, + "num_input_tokens_seen": 7386169344, + "step": 3522 + }, + { + "epoch": 0.03336045565500407, + "grad_norm": 9.609505653381348, + "learning_rate": 1.0486131967200254e-07, + "loss": 0.2049, + "num_input_tokens_seen": 7388266496, + "step": 3523 + }, + { + "epoch": 0.033631678871711417, + "grad_norm": 13.24195384979248, + "learning_rate": 1.048023213812403e-07, + "loss": 0.296, + "num_input_tokens_seen": 7390363648, + "step": 3524 + }, + { + "epoch": 0.03390290208841877, + "grad_norm": 15.955248832702637, + "learning_rate": 1.0474368136862876e-07, + "loss": 0.4047, + "num_input_tokens_seen": 7392460800, + "step": 3525 + }, + { + "epoch": 0.03417412530512612, + "grad_norm": 18.489961624145508, + "learning_rate": 1.0468539968135922e-07, + "loss": 0.6169, + "num_input_tokens_seen": 7394557952, + "step": 3526 + }, + { + "epoch": 0.034445348521833466, + "grad_norm": 15.667023658752441, + "learning_rate": 1.046274763663345e-07, + "loss": 0.4134, + "num_input_tokens_seen": 7396655104, + "step": 3527 + }, + { + "epoch": 0.03471657173854082, + "grad_norm": 20.19091796875, + "learning_rate": 1.045699114701691e-07, + "loss": 0.6246, + "num_input_tokens_seen": 7398752256, + "step": 3528 + }, + { + "epoch": 0.03498779495524817, + "grad_norm": 18.28851890563965, + "learning_rate": 1.0451270503918906e-07, + "loss": 0.5747, + "num_input_tokens_seen": 7400849408, + "step": 3529 + }, + { + "epoch": 0.03525901817195552, + "grad_norm": 20.15885353088379, + "learning_rate": 1.0445585711943205e-07, + "loss": 0.6841, + "num_input_tokens_seen": 7402946560, + "step": 3530 + }, + { + "epoch": 0.03553024138866287, + "grad_norm": 15.178775787353516, + "learning_rate": 1.0439936775664699e-07, + "loss": 0.5073, + "num_input_tokens_seen": 7405043712, + "step": 3531 + }, + { + "epoch": 0.03580146460537022, + "grad_norm": 13.089923858642578, + "learning_rate": 1.043432369962943e-07, + "loss": 0.3552, + "num_input_tokens_seen": 7407140864, + "step": 3532 + }, + { + "epoch": 0.03607268782207757, + "grad_norm": 12.704035758972168, + "learning_rate": 1.0428746488354606e-07, + "loss": 0.4504, + "num_input_tokens_seen": 7409238016, + "step": 3533 + }, + { + "epoch": 0.03634391103878492, + "grad_norm": 13.723061561584473, + "learning_rate": 1.0423205146328548e-07, + "loss": 0.3546, + "num_input_tokens_seen": 7411335168, + "step": 3534 + }, + { + "epoch": 0.03661513425549227, + "grad_norm": 15.719633102416992, + "learning_rate": 1.0417699678010708e-07, + "loss": 0.5327, + "num_input_tokens_seen": 7413432320, + "step": 3535 + }, + { + "epoch": 0.03688635747219962, + "grad_norm": 18.25145721435547, + "learning_rate": 1.0412230087831689e-07, + "loss": 0.607, + "num_input_tokens_seen": 7415529472, + "step": 3536 + }, + { + "epoch": 0.03715758068890697, + "grad_norm": 16.473398208618164, + "learning_rate": 1.0406796380193203e-07, + "loss": 0.4334, + "num_input_tokens_seen": 7417626624, + "step": 3537 + }, + { + "epoch": 0.03742880390561432, + "grad_norm": 14.576959609985352, + "learning_rate": 1.0401398559468098e-07, + "loss": 0.5355, + "num_input_tokens_seen": 7419723776, + "step": 3538 + }, + { + "epoch": 0.03770002712232167, + "grad_norm": 23.102252960205078, + "learning_rate": 1.0396036630000324e-07, + "loss": 0.4434, + "num_input_tokens_seen": 7421820928, + "step": 3539 + }, + { + "epoch": 0.03797125033902902, + "grad_norm": 14.231484413146973, + "learning_rate": 1.0390710596104965e-07, + "loss": 0.3967, + "num_input_tokens_seen": 7423918080, + "step": 3540 + }, + { + "epoch": 0.03824247355573637, + "grad_norm": 20.063039779663086, + "learning_rate": 1.0385420462068206e-07, + "loss": 0.556, + "num_input_tokens_seen": 7426015232, + "step": 3541 + }, + { + "epoch": 0.03851369677244372, + "grad_norm": 11.724411010742188, + "learning_rate": 1.0380166232147354e-07, + "loss": 0.308, + "num_input_tokens_seen": 7428112384, + "step": 3542 + }, + { + "epoch": 0.038784919989151075, + "grad_norm": 11.57480239868164, + "learning_rate": 1.0374947910570805e-07, + "loss": 0.3655, + "num_input_tokens_seen": 7430209536, + "step": 3543 + }, + { + "epoch": 0.03905614320585842, + "grad_norm": 13.075111389160156, + "learning_rate": 1.0369765501538067e-07, + "loss": 0.3675, + "num_input_tokens_seen": 7432306688, + "step": 3544 + }, + { + "epoch": 0.03932736642256577, + "grad_norm": 15.664528846740723, + "learning_rate": 1.0364619009219743e-07, + "loss": 0.3647, + "num_input_tokens_seen": 7434403840, + "step": 3545 + }, + { + "epoch": 0.039598589639273124, + "grad_norm": 16.314237594604492, + "learning_rate": 1.0359508437757544e-07, + "loss": 0.3922, + "num_input_tokens_seen": 7436500992, + "step": 3546 + }, + { + "epoch": 0.03986981285598047, + "grad_norm": 15.859907150268555, + "learning_rate": 1.0354433791264255e-07, + "loss": 0.4337, + "num_input_tokens_seen": 7438598144, + "step": 3547 + }, + { + "epoch": 0.04014103607268782, + "grad_norm": 21.789993286132812, + "learning_rate": 1.0349395073823768e-07, + "loss": 0.639, + "num_input_tokens_seen": 7440695296, + "step": 3548 + }, + { + "epoch": 0.040412259289395173, + "grad_norm": 13.827730178833008, + "learning_rate": 1.0344392289491038e-07, + "loss": 0.3828, + "num_input_tokens_seen": 7442792448, + "step": 3549 + }, + { + "epoch": 0.04068348250610252, + "grad_norm": 18.37379264831543, + "learning_rate": 1.0339425442292118e-07, + "loss": 0.7685, + "num_input_tokens_seen": 7444889600, + "step": 3550 + }, + { + "epoch": 0.040954705722809875, + "grad_norm": 14.437710762023926, + "learning_rate": 1.0334494536224146e-07, + "loss": 0.508, + "num_input_tokens_seen": 7446986752, + "step": 3551 + }, + { + "epoch": 0.04122592893951722, + "grad_norm": 15.604809761047363, + "learning_rate": 1.0329599575255321e-07, + "loss": 0.5133, + "num_input_tokens_seen": 7449083904, + "step": 3552 + }, + { + "epoch": 0.04149715215622457, + "grad_norm": 11.595867156982422, + "learning_rate": 1.0324740563324923e-07, + "loss": 0.311, + "num_input_tokens_seen": 7451181056, + "step": 3553 + }, + { + "epoch": 0.041768375372931925, + "grad_norm": 20.499563217163086, + "learning_rate": 1.0319917504343297e-07, + "loss": 0.8097, + "num_input_tokens_seen": 7453278208, + "step": 3554 + }, + { + "epoch": 0.04203959858963927, + "grad_norm": 15.316178321838379, + "learning_rate": 1.0315130402191866e-07, + "loss": 0.3584, + "num_input_tokens_seen": 7455375360, + "step": 3555 + }, + { + "epoch": 0.04231082180634662, + "grad_norm": 13.93494701385498, + "learning_rate": 1.0310379260723094e-07, + "loss": 0.4657, + "num_input_tokens_seen": 7457472512, + "step": 3556 + }, + { + "epoch": 0.042582045023053974, + "grad_norm": 18.21516227722168, + "learning_rate": 1.0305664083760532e-07, + "loss": 0.6825, + "num_input_tokens_seen": 7459569664, + "step": 3557 + }, + { + "epoch": 0.04285326823976132, + "grad_norm": 15.643766403198242, + "learning_rate": 1.0300984875098772e-07, + "loss": 0.5517, + "num_input_tokens_seen": 7461666816, + "step": 3558 + }, + { + "epoch": 0.043124491456468676, + "grad_norm": 19.415977478027344, + "learning_rate": 1.0296341638503458e-07, + "loss": 0.4861, + "num_input_tokens_seen": 7463763968, + "step": 3559 + }, + { + "epoch": 0.043395714673176024, + "grad_norm": 14.686422348022461, + "learning_rate": 1.029173437771129e-07, + "loss": 0.3778, + "num_input_tokens_seen": 7465861120, + "step": 3560 + }, + { + "epoch": 0.04366693788988337, + "grad_norm": 18.66650390625, + "learning_rate": 1.0287163096430024e-07, + "loss": 0.7157, + "num_input_tokens_seen": 7467958272, + "step": 3561 + }, + { + "epoch": 0.043938161106590726, + "grad_norm": 12.286962509155273, + "learning_rate": 1.0282627798338444e-07, + "loss": 0.331, + "num_input_tokens_seen": 7470055424, + "step": 3562 + }, + { + "epoch": 0.04420938432329807, + "grad_norm": 14.363128662109375, + "learning_rate": 1.0278128487086387e-07, + "loss": 0.5657, + "num_input_tokens_seen": 7472152576, + "step": 3563 + }, + { + "epoch": 0.04448060754000543, + "grad_norm": 15.838930130004883, + "learning_rate": 1.0273665166294735e-07, + "loss": 0.3759, + "num_input_tokens_seen": 7474249728, + "step": 3564 + }, + { + "epoch": 0.044751830756712775, + "grad_norm": 12.090499877929688, + "learning_rate": 1.0269237839555398e-07, + "loss": 0.3603, + "num_input_tokens_seen": 7476346880, + "step": 3565 + }, + { + "epoch": 0.04502305397342012, + "grad_norm": 11.031499862670898, + "learning_rate": 1.0264846510431307e-07, + "loss": 0.2938, + "num_input_tokens_seen": 7478444032, + "step": 3566 + }, + { + "epoch": 0.04529427719012748, + "grad_norm": 18.495397567749023, + "learning_rate": 1.0260491182456453e-07, + "loss": 0.8536, + "num_input_tokens_seen": 7480541184, + "step": 3567 + }, + { + "epoch": 0.045565500406834825, + "grad_norm": 12.062615394592285, + "learning_rate": 1.0256171859135826e-07, + "loss": 0.3868, + "num_input_tokens_seen": 7482638336, + "step": 3568 + }, + { + "epoch": 0.04583672362354217, + "grad_norm": 11.398265838623047, + "learning_rate": 1.0251888543945458e-07, + "loss": 0.3781, + "num_input_tokens_seen": 7484735488, + "step": 3569 + }, + { + "epoch": 0.04610794684024953, + "grad_norm": 11.258602142333984, + "learning_rate": 1.0247641240332397e-07, + "loss": 0.368, + "num_input_tokens_seen": 7486832640, + "step": 3570 + }, + { + "epoch": 0.046379170056956874, + "grad_norm": 12.560171127319336, + "learning_rate": 1.0243429951714714e-07, + "loss": 0.389, + "num_input_tokens_seen": 7488929792, + "step": 3571 + }, + { + "epoch": 0.04665039327366423, + "grad_norm": 13.790288925170898, + "learning_rate": 1.023925468148149e-07, + "loss": 0.4665, + "num_input_tokens_seen": 7491026944, + "step": 3572 + }, + { + "epoch": 0.046921616490371576, + "grad_norm": 8.38277530670166, + "learning_rate": 1.023511543299283e-07, + "loss": 0.1851, + "num_input_tokens_seen": 7493124096, + "step": 3573 + }, + { + "epoch": 0.047192839707078924, + "grad_norm": 13.927973747253418, + "learning_rate": 1.0231012209579831e-07, + "loss": 0.4549, + "num_input_tokens_seen": 7495221248, + "step": 3574 + }, + { + "epoch": 0.04746406292378628, + "grad_norm": 23.091039657592773, + "learning_rate": 1.0226945014544624e-07, + "loss": 0.6259, + "num_input_tokens_seen": 7497318400, + "step": 3575 + }, + { + "epoch": 0.047735286140493625, + "grad_norm": 17.690340042114258, + "learning_rate": 1.0222913851160335e-07, + "loss": 0.6754, + "num_input_tokens_seen": 7499415552, + "step": 3576 + }, + { + "epoch": 0.04800650935720097, + "grad_norm": 13.187427520751953, + "learning_rate": 1.0218918722671074e-07, + "loss": 0.4299, + "num_input_tokens_seen": 7501512704, + "step": 3577 + }, + { + "epoch": 0.04827773257390833, + "grad_norm": 15.680404663085938, + "learning_rate": 1.0214959632291984e-07, + "loss": 0.5009, + "num_input_tokens_seen": 7503609856, + "step": 3578 + }, + { + "epoch": 0.048548955790615675, + "grad_norm": 13.153005599975586, + "learning_rate": 1.0211036583209181e-07, + "loss": 0.3405, + "num_input_tokens_seen": 7505707008, + "step": 3579 + }, + { + "epoch": 0.04882017900732303, + "grad_norm": 11.65854549407959, + "learning_rate": 1.0207149578579789e-07, + "loss": 0.2888, + "num_input_tokens_seen": 7507804160, + "step": 3580 + }, + { + "epoch": 0.04909140222403038, + "grad_norm": 24.025188446044922, + "learning_rate": 1.0203298621531923e-07, + "loss": 0.4813, + "num_input_tokens_seen": 7509901312, + "step": 3581 + }, + { + "epoch": 0.049362625440737724, + "grad_norm": 22.057125091552734, + "learning_rate": 1.0199483715164687e-07, + "loss": 0.5313, + "num_input_tokens_seen": 7511998464, + "step": 3582 + }, + { + "epoch": 0.04963384865744508, + "grad_norm": 21.940088272094727, + "learning_rate": 1.0195704862548167e-07, + "loss": 0.8506, + "num_input_tokens_seen": 7514095616, + "step": 3583 + }, + { + "epoch": 0.049905071874152426, + "grad_norm": 14.703182220458984, + "learning_rate": 1.0191962066723448e-07, + "loss": 0.4072, + "num_input_tokens_seen": 7516192768, + "step": 3584 + }, + { + "epoch": 0.05017629509085978, + "grad_norm": 8.556319236755371, + "learning_rate": 1.0188255330702583e-07, + "loss": 0.2465, + "num_input_tokens_seen": 7518289920, + "step": 3585 + }, + { + "epoch": 0.05044751830756713, + "grad_norm": 18.820598602294922, + "learning_rate": 1.0184584657468615e-07, + "loss": 0.6037, + "num_input_tokens_seen": 7520387072, + "step": 3586 + }, + { + "epoch": 0.050718741524274476, + "grad_norm": 11.99779224395752, + "learning_rate": 1.018095004997556e-07, + "loss": 0.3087, + "num_input_tokens_seen": 7522484224, + "step": 3587 + }, + { + "epoch": 0.05098996474098183, + "grad_norm": 11.369059562683105, + "learning_rate": 1.0177351511148414e-07, + "loss": 0.2991, + "num_input_tokens_seen": 7524581376, + "step": 3588 + }, + { + "epoch": 0.05126118795768918, + "grad_norm": 16.702545166015625, + "learning_rate": 1.0173789043883147e-07, + "loss": 0.476, + "num_input_tokens_seen": 7526678528, + "step": 3589 + }, + { + "epoch": 0.051532411174396525, + "grad_norm": 13.388700485229492, + "learning_rate": 1.0170262651046687e-07, + "loss": 0.3989, + "num_input_tokens_seen": 7528775680, + "step": 3590 + }, + { + "epoch": 0.05180363439110388, + "grad_norm": 16.760042190551758, + "learning_rate": 1.0166772335476951e-07, + "loss": 0.476, + "num_input_tokens_seen": 7530872832, + "step": 3591 + }, + { + "epoch": 0.05207485760781123, + "grad_norm": 16.29047203063965, + "learning_rate": 1.0163318099982808e-07, + "loss": 0.4348, + "num_input_tokens_seen": 7532969984, + "step": 3592 + }, + { + "epoch": 0.05234608082451858, + "grad_norm": 14.072612762451172, + "learning_rate": 1.0159899947344094e-07, + "loss": 0.4399, + "num_input_tokens_seen": 7535067136, + "step": 3593 + }, + { + "epoch": 0.05261730404122593, + "grad_norm": 12.154025077819824, + "learning_rate": 1.0156517880311614e-07, + "loss": 0.3253, + "num_input_tokens_seen": 7537164288, + "step": 3594 + }, + { + "epoch": 0.05288852725793328, + "grad_norm": 14.722796440124512, + "learning_rate": 1.0153171901607118e-07, + "loss": 0.3324, + "num_input_tokens_seen": 7539261440, + "step": 3595 + }, + { + "epoch": 0.05315975047464063, + "grad_norm": 17.35401153564453, + "learning_rate": 1.0149862013923329e-07, + "loss": 0.6772, + "num_input_tokens_seen": 7541358592, + "step": 3596 + }, + { + "epoch": 0.05343097369134798, + "grad_norm": 13.542460441589355, + "learning_rate": 1.0146588219923917e-07, + "loss": 0.4358, + "num_input_tokens_seen": 7543455744, + "step": 3597 + }, + { + "epoch": 0.05370219690805533, + "grad_norm": 11.1475248336792, + "learning_rate": 1.0143350522243509e-07, + "loss": 0.3717, + "num_input_tokens_seen": 7545552896, + "step": 3598 + }, + { + "epoch": 0.05397342012476268, + "grad_norm": 14.331890106201172, + "learning_rate": 1.0140148923487675e-07, + "loss": 0.4832, + "num_input_tokens_seen": 7547650048, + "step": 3599 + }, + { + "epoch": 0.05424464334147003, + "grad_norm": 13.898726463317871, + "learning_rate": 1.0136983426232945e-07, + "loss": 0.4623, + "num_input_tokens_seen": 7549747200, + "step": 3600 + }, + { + "epoch": 0.05451586655817738, + "grad_norm": 9.847304344177246, + "learning_rate": 1.0133854033026789e-07, + "loss": 0.2438, + "num_input_tokens_seen": 7551844352, + "step": 3601 + }, + { + "epoch": 0.05478708977488473, + "grad_norm": 16.682790756225586, + "learning_rate": 1.0130760746387622e-07, + "loss": 0.2657, + "num_input_tokens_seen": 7553941504, + "step": 3602 + }, + { + "epoch": 0.05505831299159208, + "grad_norm": 12.895301818847656, + "learning_rate": 1.0127703568804805e-07, + "loss": 0.3056, + "num_input_tokens_seen": 7556038656, + "step": 3603 + }, + { + "epoch": 0.05532953620829943, + "grad_norm": 14.191023826599121, + "learning_rate": 1.0124682502738638e-07, + "loss": 0.368, + "num_input_tokens_seen": 7558135808, + "step": 3604 + }, + { + "epoch": 0.05560075942500678, + "grad_norm": 18.443326950073242, + "learning_rate": 1.0121697550620365e-07, + "loss": 0.5653, + "num_input_tokens_seen": 7560232960, + "step": 3605 + }, + { + "epoch": 0.055871982641714134, + "grad_norm": 15.984533309936523, + "learning_rate": 1.0118748714852156e-07, + "loss": 0.4285, + "num_input_tokens_seen": 7562330112, + "step": 3606 + }, + { + "epoch": 0.05614320585842148, + "grad_norm": 12.376087188720703, + "learning_rate": 1.011583599780712e-07, + "loss": 0.2091, + "num_input_tokens_seen": 7564427264, + "step": 3607 + }, + { + "epoch": 0.05641442907512883, + "grad_norm": 21.866743087768555, + "learning_rate": 1.01129594018293e-07, + "loss": 0.6654, + "num_input_tokens_seen": 7566524416, + "step": 3608 + }, + { + "epoch": 0.05668565229183618, + "grad_norm": 18.539138793945312, + "learning_rate": 1.0110118929233682e-07, + "loss": 0.4428, + "num_input_tokens_seen": 7568621568, + "step": 3609 + }, + { + "epoch": 0.05695687550854353, + "grad_norm": 19.51919174194336, + "learning_rate": 1.0107314582306156e-07, + "loss": 0.6288, + "num_input_tokens_seen": 7570718720, + "step": 3610 + }, + { + "epoch": 0.05722809872525088, + "grad_norm": 16.458036422729492, + "learning_rate": 1.0104546363303566e-07, + "loss": 0.5354, + "num_input_tokens_seen": 7572815872, + "step": 3611 + }, + { + "epoch": 0.05749932194195823, + "grad_norm": 18.08392906188965, + "learning_rate": 1.0101814274453661e-07, + "loss": 0.6547, + "num_input_tokens_seen": 7574913024, + "step": 3612 + }, + { + "epoch": 0.05777054515866558, + "grad_norm": 16.274539947509766, + "learning_rate": 1.0099118317955127e-07, + "loss": 0.5016, + "num_input_tokens_seen": 7577010176, + "step": 3613 + }, + { + "epoch": 0.058041768375372935, + "grad_norm": 16.817398071289062, + "learning_rate": 1.0096458495977564e-07, + "loss": 0.5801, + "num_input_tokens_seen": 7579107328, + "step": 3614 + }, + { + "epoch": 0.05831299159208028, + "grad_norm": 12.463071823120117, + "learning_rate": 1.0093834810661498e-07, + "loss": 0.3728, + "num_input_tokens_seen": 7581204480, + "step": 3615 + }, + { + "epoch": 0.05858421480878763, + "grad_norm": 11.567461967468262, + "learning_rate": 1.0091247264118372e-07, + "loss": 0.3667, + "num_input_tokens_seen": 7583301632, + "step": 3616 + }, + { + "epoch": 0.058855438025494984, + "grad_norm": 10.735450744628906, + "learning_rate": 1.0088695858430539e-07, + "loss": 0.2851, + "num_input_tokens_seen": 7585398784, + "step": 3617 + }, + { + "epoch": 0.05912666124220233, + "grad_norm": 13.669662475585938, + "learning_rate": 1.0086180595651278e-07, + "loss": 0.4644, + "num_input_tokens_seen": 7587495936, + "step": 3618 + }, + { + "epoch": 0.059397884458909686, + "grad_norm": 15.31719970703125, + "learning_rate": 1.0083701477804778e-07, + "loss": 0.5703, + "num_input_tokens_seen": 7589593088, + "step": 3619 + }, + { + "epoch": 0.059669107675617034, + "grad_norm": 14.730269432067871, + "learning_rate": 1.0081258506886134e-07, + "loss": 0.5415, + "num_input_tokens_seen": 7591690240, + "step": 3620 + }, + { + "epoch": 0.05994033089232438, + "grad_norm": 12.143908500671387, + "learning_rate": 1.0078851684861357e-07, + "loss": 0.3662, + "num_input_tokens_seen": 7593787392, + "step": 3621 + }, + { + "epoch": 0.060211554109031735, + "grad_norm": 12.585685729980469, + "learning_rate": 1.0076481013667376e-07, + "loss": 0.2725, + "num_input_tokens_seen": 7595884544, + "step": 3622 + }, + { + "epoch": 0.06048277732573908, + "grad_norm": 11.40467643737793, + "learning_rate": 1.0074146495212001e-07, + "loss": 0.307, + "num_input_tokens_seen": 7597981696, + "step": 3623 + }, + { + "epoch": 0.06075400054244643, + "grad_norm": 16.664461135864258, + "learning_rate": 1.0071848131373972e-07, + "loss": 0.7096, + "num_input_tokens_seen": 7600078848, + "step": 3624 + }, + { + "epoch": 0.061025223759153785, + "grad_norm": 11.929876327514648, + "learning_rate": 1.0069585924002924e-07, + "loss": 0.3869, + "num_input_tokens_seen": 7602176000, + "step": 3625 + }, + { + "epoch": 0.06129644697586113, + "grad_norm": 20.532058715820312, + "learning_rate": 1.0067359874919395e-07, + "loss": 0.7844, + "num_input_tokens_seen": 7604273152, + "step": 3626 + }, + { + "epoch": 0.06156767019256849, + "grad_norm": 10.287115097045898, + "learning_rate": 1.0065169985914826e-07, + "loss": 0.2573, + "num_input_tokens_seen": 7606370304, + "step": 3627 + }, + { + "epoch": 0.061838893409275834, + "grad_norm": 17.162843704223633, + "learning_rate": 1.0063016258751553e-07, + "loss": 0.6302, + "num_input_tokens_seen": 7608467456, + "step": 3628 + }, + { + "epoch": 0.06211011662598318, + "grad_norm": 14.848957061767578, + "learning_rate": 1.0060898695162816e-07, + "loss": 0.4309, + "num_input_tokens_seen": 7610564608, + "step": 3629 + }, + { + "epoch": 0.062381339842690536, + "grad_norm": 17.66904640197754, + "learning_rate": 1.005881729685275e-07, + "loss": 0.4711, + "num_input_tokens_seen": 7612661760, + "step": 3630 + }, + { + "epoch": 0.06265256305939788, + "grad_norm": 20.336261749267578, + "learning_rate": 1.0056772065496387e-07, + "loss": 0.8298, + "num_input_tokens_seen": 7614758912, + "step": 3631 + }, + { + "epoch": 0.06292378627610523, + "grad_norm": 22.26197052001953, + "learning_rate": 1.005476300273965e-07, + "loss": 0.6695, + "num_input_tokens_seen": 7616856064, + "step": 3632 + }, + { + "epoch": 0.06319500949281258, + "grad_norm": 14.296065330505371, + "learning_rate": 1.0052790110199348e-07, + "loss": 0.4288, + "num_input_tokens_seen": 7618953216, + "step": 3633 + }, + { + "epoch": 0.06346623270951994, + "grad_norm": 18.81780433654785, + "learning_rate": 1.0050853389463205e-07, + "loss": 0.6585, + "num_input_tokens_seen": 7621050368, + "step": 3634 + }, + { + "epoch": 0.06373745592622729, + "grad_norm": 11.14263916015625, + "learning_rate": 1.0048952842089805e-07, + "loss": 0.3255, + "num_input_tokens_seen": 7623147520, + "step": 3635 + }, + { + "epoch": 0.06400867914293464, + "grad_norm": 12.55125904083252, + "learning_rate": 1.0047088469608648e-07, + "loss": 0.315, + "num_input_tokens_seen": 7625244672, + "step": 3636 + }, + { + "epoch": 0.06427990235964198, + "grad_norm": 20.127553939819336, + "learning_rate": 1.00452602735201e-07, + "loss": 0.3163, + "num_input_tokens_seen": 7627341824, + "step": 3637 + }, + { + "epoch": 0.06455112557634933, + "grad_norm": 18.10858726501465, + "learning_rate": 1.0043468255295435e-07, + "loss": 0.6731, + "num_input_tokens_seen": 7629438976, + "step": 3638 + }, + { + "epoch": 0.06482234879305669, + "grad_norm": 15.758679389953613, + "learning_rate": 1.0041712416376795e-07, + "loss": 0.5683, + "num_input_tokens_seen": 7631536128, + "step": 3639 + }, + { + "epoch": 0.06509357200976404, + "grad_norm": 12.081664085388184, + "learning_rate": 1.0039992758177211e-07, + "loss": 0.2851, + "num_input_tokens_seen": 7633633280, + "step": 3640 + }, + { + "epoch": 0.06536479522647139, + "grad_norm": 12.943633079528809, + "learning_rate": 1.0038309282080596e-07, + "loss": 0.2524, + "num_input_tokens_seen": 7635730432, + "step": 3641 + }, + { + "epoch": 0.06563601844317873, + "grad_norm": 16.313081741333008, + "learning_rate": 1.0036661989441755e-07, + "loss": 0.5379, + "num_input_tokens_seen": 7637827584, + "step": 3642 + }, + { + "epoch": 0.06590724165988608, + "grad_norm": 14.559349060058594, + "learning_rate": 1.0035050881586364e-07, + "loss": 0.3989, + "num_input_tokens_seen": 7639924736, + "step": 3643 + }, + { + "epoch": 0.06617846487659344, + "grad_norm": 12.672063827514648, + "learning_rate": 1.0033475959810974e-07, + "loss": 0.3097, + "num_input_tokens_seen": 7642021888, + "step": 3644 + }, + { + "epoch": 0.06644968809330079, + "grad_norm": 12.448989868164062, + "learning_rate": 1.0031937225383036e-07, + "loss": 0.4511, + "num_input_tokens_seen": 7644119040, + "step": 3645 + }, + { + "epoch": 0.06672091131000814, + "grad_norm": 12.244915962219238, + "learning_rate": 1.0030434679540853e-07, + "loss": 0.2887, + "num_input_tokens_seen": 7646216192, + "step": 3646 + }, + { + "epoch": 0.06699213452671549, + "grad_norm": 16.625158309936523, + "learning_rate": 1.0028968323493623e-07, + "loss": 0.5713, + "num_input_tokens_seen": 7648313344, + "step": 3647 + }, + { + "epoch": 0.06726335774342283, + "grad_norm": 18.604345321655273, + "learning_rate": 1.0027538158421413e-07, + "loss": 0.6223, + "num_input_tokens_seen": 7650410496, + "step": 3648 + }, + { + "epoch": 0.06753458096013018, + "grad_norm": 17.25090217590332, + "learning_rate": 1.002614418547516e-07, + "loss": 0.6491, + "num_input_tokens_seen": 7652507648, + "step": 3649 + }, + { + "epoch": 0.06780580417683754, + "grad_norm": 17.5065975189209, + "learning_rate": 1.0024786405776686e-07, + "loss": 0.5556, + "num_input_tokens_seen": 7654604800, + "step": 3650 + }, + { + "epoch": 0.06807702739354489, + "grad_norm": 13.661022186279297, + "learning_rate": 1.0023464820418676e-07, + "loss": 0.451, + "num_input_tokens_seen": 7656701952, + "step": 3651 + }, + { + "epoch": 0.06834825061025224, + "grad_norm": 14.327960968017578, + "learning_rate": 1.00221794304647e-07, + "loss": 0.3287, + "num_input_tokens_seen": 7658799104, + "step": 3652 + }, + { + "epoch": 0.06861947382695958, + "grad_norm": 21.388771057128906, + "learning_rate": 1.0020930236949182e-07, + "loss": 0.363, + "num_input_tokens_seen": 7660896256, + "step": 3653 + }, + { + "epoch": 0.06889069704366693, + "grad_norm": 9.650550842285156, + "learning_rate": 1.0019717240877424e-07, + "loss": 0.253, + "num_input_tokens_seen": 7662993408, + "step": 3654 + }, + { + "epoch": 0.0691619202603743, + "grad_norm": 17.260183334350586, + "learning_rate": 1.001854044322561e-07, + "loss": 0.7714, + "num_input_tokens_seen": 7665090560, + "step": 3655 + }, + { + "epoch": 0.06943314347708164, + "grad_norm": 22.375675201416016, + "learning_rate": 1.0017399844940774e-07, + "loss": 0.4834, + "num_input_tokens_seen": 7667187712, + "step": 3656 + }, + { + "epoch": 0.06970436669378899, + "grad_norm": 13.567085266113281, + "learning_rate": 1.0016295446940827e-07, + "loss": 0.3653, + "num_input_tokens_seen": 7669284864, + "step": 3657 + }, + { + "epoch": 0.06997558991049634, + "grad_norm": 17.413497924804688, + "learning_rate": 1.001522725011455e-07, + "loss": 0.7221, + "num_input_tokens_seen": 7671382016, + "step": 3658 + }, + { + "epoch": 0.07024681312720368, + "grad_norm": 13.802844047546387, + "learning_rate": 1.0014195255321583e-07, + "loss": 0.4207, + "num_input_tokens_seen": 7673479168, + "step": 3659 + }, + { + "epoch": 0.07051803634391104, + "grad_norm": 13.901693344116211, + "learning_rate": 1.0013199463392433e-07, + "loss": 0.3311, + "num_input_tokens_seen": 7675576320, + "step": 3660 + }, + { + "epoch": 0.07078925956061839, + "grad_norm": 11.964288711547852, + "learning_rate": 1.0012239875128484e-07, + "loss": 0.2948, + "num_input_tokens_seen": 7677673472, + "step": 3661 + }, + { + "epoch": 0.07106048277732574, + "grad_norm": 16.391555786132812, + "learning_rate": 1.0011316491301973e-07, + "loss": 0.4533, + "num_input_tokens_seen": 7679770624, + "step": 3662 + }, + { + "epoch": 0.07133170599403309, + "grad_norm": 15.69762134552002, + "learning_rate": 1.0010429312656006e-07, + "loss": 0.5279, + "num_input_tokens_seen": 7681867776, + "step": 3663 + }, + { + "epoch": 0.07160292921074043, + "grad_norm": 18.312103271484375, + "learning_rate": 1.000957833990454e-07, + "loss": 0.6914, + "num_input_tokens_seen": 7683964928, + "step": 3664 + }, + { + "epoch": 0.0718741524274478, + "grad_norm": 11.982022285461426, + "learning_rate": 1.0008763573732421e-07, + "loss": 0.2825, + "num_input_tokens_seen": 7686062080, + "step": 3665 + }, + { + "epoch": 0.07214537564415514, + "grad_norm": 15.081099510192871, + "learning_rate": 1.0007985014795331e-07, + "loss": 0.4631, + "num_input_tokens_seen": 7688159232, + "step": 3666 + }, + { + "epoch": 0.07241659886086249, + "grad_norm": 14.211804389953613, + "learning_rate": 1.0007242663719824e-07, + "loss": 0.4301, + "num_input_tokens_seen": 7690256384, + "step": 3667 + }, + { + "epoch": 0.07268782207756984, + "grad_norm": 16.612119674682617, + "learning_rate": 1.0006536521103325e-07, + "loss": 0.5014, + "num_input_tokens_seen": 7692353536, + "step": 3668 + }, + { + "epoch": 0.07295904529427719, + "grad_norm": 16.976350784301758, + "learning_rate": 1.0005866587514106e-07, + "loss": 0.5847, + "num_input_tokens_seen": 7694450688, + "step": 3669 + }, + { + "epoch": 0.07323026851098453, + "grad_norm": 16.93351936340332, + "learning_rate": 1.0005232863491297e-07, + "loss": 0.5873, + "num_input_tokens_seen": 7696547840, + "step": 3670 + }, + { + "epoch": 0.0735014917276919, + "grad_norm": 19.781625747680664, + "learning_rate": 1.0004635349544907e-07, + "loss": 0.5883, + "num_input_tokens_seen": 7698644992, + "step": 3671 + }, + { + "epoch": 0.07377271494439924, + "grad_norm": 15.636222839355469, + "learning_rate": 1.0004074046155789e-07, + "loss": 0.3702, + "num_input_tokens_seen": 7700742144, + "step": 3672 + }, + { + "epoch": 0.07404393816110659, + "grad_norm": 20.640913009643555, + "learning_rate": 1.000354895377565e-07, + "loss": 0.7943, + "num_input_tokens_seen": 7702839296, + "step": 3673 + }, + { + "epoch": 0.07431516137781394, + "grad_norm": 16.68242073059082, + "learning_rate": 1.0003060072827073e-07, + "loss": 0.4785, + "num_input_tokens_seen": 7704936448, + "step": 3674 + }, + { + "epoch": 0.07458638459452128, + "grad_norm": 14.670442581176758, + "learning_rate": 1.0002607403703492e-07, + "loss": 0.3074, + "num_input_tokens_seen": 7707033600, + "step": 3675 + }, + { + "epoch": 0.07485760781122865, + "grad_norm": 19.64070701599121, + "learning_rate": 1.000219094676919e-07, + "loss": 0.7391, + "num_input_tokens_seen": 7709130752, + "step": 3676 + }, + { + "epoch": 0.075128831027936, + "grad_norm": 12.361337661743164, + "learning_rate": 1.0001810702359326e-07, + "loss": 0.3917, + "num_input_tokens_seen": 7711227904, + "step": 3677 + }, + { + "epoch": 0.07540005424464334, + "grad_norm": 11.465216636657715, + "learning_rate": 1.0001466670779896e-07, + "loss": 0.3472, + "num_input_tokens_seen": 7713325056, + "step": 3678 + }, + { + "epoch": 0.07567127746135069, + "grad_norm": 10.229789733886719, + "learning_rate": 1.000115885230777e-07, + "loss": 0.2105, + "num_input_tokens_seen": 7715422208, + "step": 3679 + }, + { + "epoch": 0.07594250067805804, + "grad_norm": 11.707015991210938, + "learning_rate": 1.0000887247190662e-07, + "loss": 0.3553, + "num_input_tokens_seen": 7717519360, + "step": 3680 + }, + { + "epoch": 0.0762137238947654, + "grad_norm": 17.815807342529297, + "learning_rate": 1.000065185564716e-07, + "loss": 0.7441, + "num_input_tokens_seen": 7719616512, + "step": 3681 + }, + { + "epoch": 0.07648494711147275, + "grad_norm": 16.822296142578125, + "learning_rate": 1.0000452677866691e-07, + "loss": 0.5363, + "num_input_tokens_seen": 7721713664, + "step": 3682 + }, + { + "epoch": 0.07675617032818009, + "grad_norm": 16.27877426147461, + "learning_rate": 1.0000289714009542e-07, + "loss": 0.5075, + "num_input_tokens_seen": 7723810816, + "step": 3683 + }, + { + "epoch": 0.07702739354488744, + "grad_norm": 14.539108276367188, + "learning_rate": 1.000016296420687e-07, + "loss": 0.4568, + "num_input_tokens_seen": 7725907968, + "step": 3684 + }, + { + "epoch": 0.07729861676159479, + "grad_norm": 16.22719383239746, + "learning_rate": 1.0000072428560674e-07, + "loss": 0.494, + "num_input_tokens_seen": 7728005120, + "step": 3685 + }, + { + "epoch": 0.07756983997830215, + "grad_norm": 15.531390190124512, + "learning_rate": 1.000001810714381e-07, + "loss": 0.4097, + "num_input_tokens_seen": 7730102272, + "step": 3686 + }, + { + "epoch": 0.0778410631950095, + "grad_norm": 10.840725898742676, + "learning_rate": 1e-07, + "loss": 0.3166, + "num_input_tokens_seen": 7732199424, + "step": 3687 + }, + { + "epoch": 0.0778410631950095, + "num_input_tokens_seen": 7732199424, + "step": 3687, + "total_flos": 4.352218066950134e+19, + "train_loss": 0.036126901625228955, + "train_runtime": 9429.4679, + "train_samples_per_second": 0.782, + "train_steps_per_second": 0.391 + } + ], + "logging_steps": 1.0, + "max_steps": 3687, + "num_input_tokens_seen": 7732199424, + "num_train_epochs": 1, + "save_steps": 200, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 4.352218066950134e+19, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +}