{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.18488560203374163, "eval_steps": 500, "global_step": 1200, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0015407133502811801, "grad_norm": 48.20575104273089, "learning_rate": 4.615384615384616e-06, "loss": 1.3971, "step": 10 }, { "epoch": 0.0030814267005623602, "grad_norm": 33.11975747134839, "learning_rate": 9.230769230769232e-06, "loss": 1.4053, "step": 20 }, { "epoch": 0.004622140050843541, "grad_norm": 39.826606299507084, "learning_rate": 1.3846153846153847e-05, "loss": 1.3472, "step": 30 }, { "epoch": 0.0061628534011247205, "grad_norm": 6.773742739301677, "learning_rate": 1.8461538461538465e-05, "loss": 1.1311, "step": 40 }, { "epoch": 0.007703566751405901, "grad_norm": 12.278712041952108, "learning_rate": 2.307692307692308e-05, "loss": 0.9715, "step": 50 }, { "epoch": 0.009244280101687081, "grad_norm": 3.776870082439811, "learning_rate": 2.7692307692307694e-05, "loss": 0.8937, "step": 60 }, { "epoch": 0.010784993451968261, "grad_norm": 6.812660145408396, "learning_rate": 2.9999955171465948e-05, "loss": 0.8472, "step": 70 }, { "epoch": 0.012325706802249441, "grad_norm": 14.1920325479403, "learning_rate": 2.9999596544801216e-05, "loss": 0.8418, "step": 80 }, { "epoch": 0.013866420152530621, "grad_norm": 4.760185564415112, "learning_rate": 2.999887930004599e-05, "loss": 0.8275, "step": 90 }, { "epoch": 0.015407133502811803, "grad_norm": 5.64069685533569, "learning_rate": 2.9997803454348518e-05, "loss": 0.8085, "step": 100 }, { "epoch": 0.01694784685309298, "grad_norm": 8.910755610595423, "learning_rate": 2.9996369033430674e-05, "loss": 0.8105, "step": 110 }, { "epoch": 0.018488560203374162, "grad_norm": 7.128375140746789, "learning_rate": 2.9994576071587345e-05, "loss": 0.7647, "step": 120 }, { "epoch": 0.020029273553655344, "grad_norm": 4.883749519467239, "learning_rate": 2.9992424611685575e-05, "loss": 0.7472, "step": 130 }, { "epoch": 0.021569986903936522, "grad_norm": 6.669698248447361, "learning_rate": 2.9989914705163582e-05, "loss": 0.7644, "step": 140 }, { "epoch": 0.023110700254217704, "grad_norm": 8.814324372744018, "learning_rate": 2.9987046412029506e-05, "loss": 0.7642, "step": 150 }, { "epoch": 0.024651413604498882, "grad_norm": 5.408352128962619, "learning_rate": 2.9983819800859976e-05, "loss": 0.7506, "step": 160 }, { "epoch": 0.026192126954780064, "grad_norm": 4.916006313310609, "learning_rate": 2.998023494879848e-05, "loss": 0.7086, "step": 170 }, { "epoch": 0.027732840305061242, "grad_norm": 6.4968532868316515, "learning_rate": 2.9976291941553508e-05, "loss": 0.7302, "step": 180 }, { "epoch": 0.029273553655342423, "grad_norm": 5.414282856209666, "learning_rate": 2.9971990873396512e-05, "loss": 0.7389, "step": 190 }, { "epoch": 0.030814267005623605, "grad_norm": 35.78019922037886, "learning_rate": 2.996733184715964e-05, "loss": 0.7247, "step": 200 }, { "epoch": 0.03235498035590478, "grad_norm": 4.344484335761467, "learning_rate": 2.9962314974233306e-05, "loss": 0.7239, "step": 210 }, { "epoch": 0.03389569370618596, "grad_norm": 4.761393885682967, "learning_rate": 2.995694037456349e-05, "loss": 0.7219, "step": 220 }, { "epoch": 0.03543640705646715, "grad_norm": 6.082748106236669, "learning_rate": 2.995120817664889e-05, "loss": 0.7036, "step": 230 }, { "epoch": 0.036977120406748325, "grad_norm": 7.925002223835823, "learning_rate": 2.9945118517537857e-05, "loss": 0.6795, "step": 240 }, { "epoch": 0.0385178337570295, "grad_norm": 4.33748938479485, "learning_rate": 2.9938671542825102e-05, "loss": 0.6894, "step": 250 }, { "epoch": 0.04005854710731069, "grad_norm": 3.419861222736367, "learning_rate": 2.993186740664821e-05, "loss": 0.674, "step": 260 }, { "epoch": 0.041599260457591866, "grad_norm": 4.418012569894865, "learning_rate": 2.9924706271683993e-05, "loss": 0.7091, "step": 270 }, { "epoch": 0.043139973807873044, "grad_norm": 5.5535932232942065, "learning_rate": 2.9917188309144548e-05, "loss": 0.7114, "step": 280 }, { "epoch": 0.04468068715815422, "grad_norm": 5.346051172390458, "learning_rate": 2.990931369877321e-05, "loss": 0.7092, "step": 290 }, { "epoch": 0.04622140050843541, "grad_norm": 5.04585392782648, "learning_rate": 2.9901082628840216e-05, "loss": 0.7079, "step": 300 }, { "epoch": 0.047762113858716586, "grad_norm": 5.3108067226217095, "learning_rate": 2.989249529613823e-05, "loss": 0.7044, "step": 310 }, { "epoch": 0.049302827208997764, "grad_norm": 12.936374781230974, "learning_rate": 2.9883551905977647e-05, "loss": 0.6795, "step": 320 }, { "epoch": 0.05084354055927895, "grad_norm": 4.969368758151501, "learning_rate": 2.987425267218164e-05, "loss": 0.7365, "step": 330 }, { "epoch": 0.05238425390956013, "grad_norm": 17.59963354249395, "learning_rate": 2.9864597817081083e-05, "loss": 0.6459, "step": 340 }, { "epoch": 0.053924967259841305, "grad_norm": 4.287709062556245, "learning_rate": 2.985458757150924e-05, "loss": 0.7151, "step": 350 }, { "epoch": 0.055465680610122484, "grad_norm": 2.353889201419343, "learning_rate": 2.9844222174796224e-05, "loss": 0.6982, "step": 360 }, { "epoch": 0.05700639396040367, "grad_norm": 6.487203178892158, "learning_rate": 2.983350187476328e-05, "loss": 0.6946, "step": 370 }, { "epoch": 0.05854710731068485, "grad_norm": 8.604226568596971, "learning_rate": 2.982242692771688e-05, "loss": 0.7024, "step": 380 }, { "epoch": 0.060087820660966025, "grad_norm": 3.7381224404208107, "learning_rate": 2.9810997598442558e-05, "loss": 0.6813, "step": 390 }, { "epoch": 0.06162853401124721, "grad_norm": 7.21302363512955, "learning_rate": 2.9799214160198622e-05, "loss": 0.6572, "step": 400 }, { "epoch": 0.06316924736152839, "grad_norm": 7.446657585655329, "learning_rate": 2.9787076894709592e-05, "loss": 0.6612, "step": 410 }, { "epoch": 0.06470996071180957, "grad_norm": 16.149479699980372, "learning_rate": 2.977458609215946e-05, "loss": 0.6823, "step": 420 }, { "epoch": 0.06625067406209074, "grad_norm": 3.654646693112352, "learning_rate": 2.9761742051184786e-05, "loss": 0.6941, "step": 430 }, { "epoch": 0.06779138741237192, "grad_norm": 7.468345726914931, "learning_rate": 2.9748545078867524e-05, "loss": 0.6254, "step": 440 }, { "epoch": 0.06933210076265311, "grad_norm": 6.559292545318311, "learning_rate": 2.9734995490727696e-05, "loss": 0.6977, "step": 450 }, { "epoch": 0.0708728141129343, "grad_norm": 8.749994768413615, "learning_rate": 2.9721093610715844e-05, "loss": 0.6742, "step": 460 }, { "epoch": 0.07241352746321547, "grad_norm": 14.746822726186753, "learning_rate": 2.9706839771205282e-05, "loss": 0.6986, "step": 470 }, { "epoch": 0.07395424081349665, "grad_norm": 2.1125271451210494, "learning_rate": 2.9692234312984156e-05, "loss": 0.6708, "step": 480 }, { "epoch": 0.07549495416377783, "grad_norm": 4.191787346884265, "learning_rate": 2.9677277585247296e-05, "loss": 0.6839, "step": 490 }, { "epoch": 0.077035667514059, "grad_norm": 6.099946769106883, "learning_rate": 2.9661969945587867e-05, "loss": 0.7253, "step": 500 }, { "epoch": 0.07857638086434018, "grad_norm": 8.272898482647978, "learning_rate": 2.9646311759988804e-05, "loss": 0.6972, "step": 510 }, { "epoch": 0.08011709421462138, "grad_norm": 12.38761120543235, "learning_rate": 2.9630303402814095e-05, "loss": 0.7174, "step": 520 }, { "epoch": 0.08165780756490255, "grad_norm": 7.826170900143438, "learning_rate": 2.961394525679979e-05, "loss": 0.7227, "step": 530 }, { "epoch": 0.08319852091518373, "grad_norm": 3.926770198625081, "learning_rate": 2.9597237713044888e-05, "loss": 0.6682, "step": 540 }, { "epoch": 0.08473923426546491, "grad_norm": 3.3408474311965692, "learning_rate": 2.9580181171001962e-05, "loss": 0.6634, "step": 550 }, { "epoch": 0.08627994761574609, "grad_norm": 15.954087186336258, "learning_rate": 2.956277603846761e-05, "loss": 0.7005, "step": 560 }, { "epoch": 0.08782066096602727, "grad_norm": 3.3366594091813475, "learning_rate": 2.9545022731572723e-05, "loss": 0.6752, "step": 570 }, { "epoch": 0.08936137431630845, "grad_norm": 4.359220549879328, "learning_rate": 2.9526921674772522e-05, "loss": 0.6985, "step": 580 }, { "epoch": 0.09090208766658964, "grad_norm": 5.585924353950877, "learning_rate": 2.95084733008364e-05, "loss": 0.6729, "step": 590 }, { "epoch": 0.09244280101687082, "grad_norm": 2.5002732127626075, "learning_rate": 2.94896780508376e-05, "loss": 0.6881, "step": 600 }, { "epoch": 0.093983514367152, "grad_norm": 8.22666868773466, "learning_rate": 2.9470536374142656e-05, "loss": 0.6918, "step": 610 }, { "epoch": 0.09552422771743317, "grad_norm": 14.050897991765508, "learning_rate": 2.9451048728400644e-05, "loss": 0.6584, "step": 620 }, { "epoch": 0.09706494106771435, "grad_norm": 3.117445397285045, "learning_rate": 2.9431215579532253e-05, "loss": 0.6495, "step": 630 }, { "epoch": 0.09860565441799553, "grad_norm": 7.000286096448409, "learning_rate": 2.9411037401718628e-05, "loss": 0.6568, "step": 640 }, { "epoch": 0.1001463677682767, "grad_norm": 3.09698455955492, "learning_rate": 2.939051467739006e-05, "loss": 0.7095, "step": 650 }, { "epoch": 0.1016870811185579, "grad_norm": 4.163429426840949, "learning_rate": 2.936964789721442e-05, "loss": 0.6573, "step": 660 }, { "epoch": 0.10322779446883908, "grad_norm": 3.282543980811516, "learning_rate": 2.934843756008546e-05, "loss": 0.6901, "step": 670 }, { "epoch": 0.10476850781912025, "grad_norm": 7.4918954929572115, "learning_rate": 2.932688417311085e-05, "loss": 0.6826, "step": 680 }, { "epoch": 0.10630922116940143, "grad_norm": 49.53066344289919, "learning_rate": 2.9304988251600084e-05, "loss": 0.6668, "step": 690 }, { "epoch": 0.10784993451968261, "grad_norm": 23.218366206451197, "learning_rate": 2.9282750319052154e-05, "loss": 0.6643, "step": 700 }, { "epoch": 0.10939064786996379, "grad_norm": 2.59944116748961, "learning_rate": 2.9260170907143012e-05, "loss": 0.6709, "step": 710 }, { "epoch": 0.11093136122024497, "grad_norm": 4.59006482180365, "learning_rate": 2.9237250555712887e-05, "loss": 0.6878, "step": 720 }, { "epoch": 0.11247207457052616, "grad_norm": 3.1508976338573507, "learning_rate": 2.9213989812753366e-05, "loss": 0.6512, "step": 730 }, { "epoch": 0.11401278792080734, "grad_norm": 4.5342197552006835, "learning_rate": 2.9190389234394285e-05, "loss": 0.6687, "step": 740 }, { "epoch": 0.11555350127108852, "grad_norm": 6.514898099534209, "learning_rate": 2.9166449384890446e-05, "loss": 0.6868, "step": 750 }, { "epoch": 0.1170942146213697, "grad_norm": 5.1251296980659005, "learning_rate": 2.9142170836608115e-05, "loss": 0.6869, "step": 760 }, { "epoch": 0.11863492797165087, "grad_norm": 5.896478552497953, "learning_rate": 2.9117554170011352e-05, "loss": 0.667, "step": 770 }, { "epoch": 0.12017564132193205, "grad_norm": 3.4672489526983483, "learning_rate": 2.909259997364811e-05, "loss": 0.6674, "step": 780 }, { "epoch": 0.12171635467221323, "grad_norm": 5.413662982418376, "learning_rate": 2.9067308844136193e-05, "loss": 0.6891, "step": 790 }, { "epoch": 0.12325706802249442, "grad_norm": 5.807164850831801, "learning_rate": 2.9041681386148966e-05, "loss": 0.6447, "step": 800 }, { "epoch": 0.1247977813727756, "grad_norm": 5.542855195057769, "learning_rate": 2.9015718212400918e-05, "loss": 0.6486, "step": 810 }, { "epoch": 0.12633849472305678, "grad_norm": 3.60243198052599, "learning_rate": 2.8989419943632992e-05, "loss": 0.6552, "step": 820 }, { "epoch": 0.12787920807333797, "grad_norm": 10.54124596601907, "learning_rate": 2.896278720859776e-05, "loss": 0.667, "step": 830 }, { "epoch": 0.12941992142361913, "grad_norm": 8.542824965925352, "learning_rate": 2.8935820644044398e-05, "loss": 0.697, "step": 840 }, { "epoch": 0.13096063477390033, "grad_norm": 6.327341591650607, "learning_rate": 2.890852089470343e-05, "loss": 0.65, "step": 850 }, { "epoch": 0.1325013481241815, "grad_norm": 7.687827002540841, "learning_rate": 2.888088861327135e-05, "loss": 0.6435, "step": 860 }, { "epoch": 0.13404206147446268, "grad_norm": 3.356453126127434, "learning_rate": 2.885292446039499e-05, "loss": 0.6721, "step": 870 }, { "epoch": 0.13558277482474385, "grad_norm": 8.406402473059597, "learning_rate": 2.8824629104655736e-05, "loss": 0.6694, "step": 880 }, { "epoch": 0.13712348817502504, "grad_norm": 11.653019434398818, "learning_rate": 2.8796003222553558e-05, "loss": 0.6531, "step": 890 }, { "epoch": 0.13866420152530623, "grad_norm": 6.371551478518258, "learning_rate": 2.8767047498490798e-05, "loss": 0.6568, "step": 900 }, { "epoch": 0.1402049148755874, "grad_norm": 26.71523757066426, "learning_rate": 2.8737762624755846e-05, "loss": 0.6857, "step": 910 }, { "epoch": 0.1417456282258686, "grad_norm": 4.417578021376778, "learning_rate": 2.8708149301506573e-05, "loss": 0.665, "step": 920 }, { "epoch": 0.14328634157614975, "grad_norm": 5.335327649767265, "learning_rate": 2.8678208236753577e-05, "loss": 0.7014, "step": 930 }, { "epoch": 0.14482705492643094, "grad_norm": 4.155520038033631, "learning_rate": 2.8647940146343278e-05, "loss": 0.6767, "step": 940 }, { "epoch": 0.1463677682767121, "grad_norm": 3.815046866792752, "learning_rate": 2.86173457539408e-05, "loss": 0.6557, "step": 950 }, { "epoch": 0.1479084816269933, "grad_norm": 3.7651811393538552, "learning_rate": 2.8586425791012648e-05, "loss": 0.675, "step": 960 }, { "epoch": 0.1494491949772745, "grad_norm": 4.382966553943605, "learning_rate": 2.8555180996809246e-05, "loss": 0.6313, "step": 970 }, { "epoch": 0.15098990832755566, "grad_norm": 10.103890347717131, "learning_rate": 2.8523612118347245e-05, "loss": 0.645, "step": 980 }, { "epoch": 0.15253062167783685, "grad_norm": 6.294074332584702, "learning_rate": 2.8491719910391685e-05, "loss": 0.659, "step": 990 }, { "epoch": 0.154071335028118, "grad_norm": 2.5808531227457565, "learning_rate": 2.845950513543791e-05, "loss": 0.6688, "step": 1000 }, { "epoch": 0.1556120483783992, "grad_norm": 2.927888770737132, "learning_rate": 2.842696856369338e-05, "loss": 0.6381, "step": 1010 }, { "epoch": 0.15715276172868037, "grad_norm": 14.062433268070832, "learning_rate": 2.8394110973059233e-05, "loss": 0.6946, "step": 1020 }, { "epoch": 0.15869347507896156, "grad_norm": 2.4470437840581054, "learning_rate": 2.8360933149111695e-05, "loss": 0.6844, "step": 1030 }, { "epoch": 0.16023418842924275, "grad_norm": 3.8078577514013343, "learning_rate": 2.8327435885083292e-05, "loss": 0.64, "step": 1040 }, { "epoch": 0.16177490177952392, "grad_norm": 25.680768915471432, "learning_rate": 2.8293619981843887e-05, "loss": 0.6329, "step": 1050 }, { "epoch": 0.1633156151298051, "grad_norm": 5.0369491995422715, "learning_rate": 2.8259486247881537e-05, "loss": 0.6604, "step": 1060 }, { "epoch": 0.16485632848008627, "grad_norm": 3.9026521516961608, "learning_rate": 2.8225035499283155e-05, "loss": 0.6564, "step": 1070 }, { "epoch": 0.16639704183036746, "grad_norm": 4.945417598915296, "learning_rate": 2.8190268559715017e-05, "loss": 0.6655, "step": 1080 }, { "epoch": 0.16793775518064863, "grad_norm": 3.222466850494984, "learning_rate": 2.815518626040304e-05, "loss": 0.6603, "step": 1090 }, { "epoch": 0.16947846853092982, "grad_norm": 6.539136587655263, "learning_rate": 2.811978944011293e-05, "loss": 0.7036, "step": 1100 }, { "epoch": 0.171019181881211, "grad_norm": 8.14211829139052, "learning_rate": 2.8084078945130117e-05, "loss": 0.6356, "step": 1110 }, { "epoch": 0.17255989523149218, "grad_norm": 4.1954200205175605, "learning_rate": 2.8048055629239543e-05, "loss": 0.6591, "step": 1120 }, { "epoch": 0.17410060858177337, "grad_norm": 4.333940585698679, "learning_rate": 2.8011720353705224e-05, "loss": 0.6575, "step": 1130 }, { "epoch": 0.17564132193205453, "grad_norm": 4.2805487065333, "learning_rate": 2.797507398724966e-05, "loss": 0.6624, "step": 1140 }, { "epoch": 0.17718203528233573, "grad_norm": 4.7173213185412, "learning_rate": 2.7938117406033086e-05, "loss": 0.623, "step": 1150 }, { "epoch": 0.1787227486326169, "grad_norm": 99.71383370833006, "learning_rate": 2.7900851493632508e-05, "loss": 0.6591, "step": 1160 }, { "epoch": 0.18026346198289808, "grad_norm": 3.6747989781213954, "learning_rate": 2.786327714102058e-05, "loss": 0.692, "step": 1170 }, { "epoch": 0.18180417533317927, "grad_norm": 2.5009166944220604, "learning_rate": 2.78253952465443e-05, "loss": 0.6614, "step": 1180 }, { "epoch": 0.18334488868346044, "grad_norm": 3.144011687958325, "learning_rate": 2.7787206715903543e-05, "loss": 0.6406, "step": 1190 }, { "epoch": 0.18488560203374163, "grad_norm": 4.063731315051197, "learning_rate": 2.7748712462129396e-05, "loss": 0.6444, "step": 1200 } ], "logging_steps": 10, "max_steps": 6490, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 400, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 1.0934282360979456e+19, "train_batch_size": 4, "trial_name": null, "trial_params": null }