{ "best_metric": null, "best_model_checkpoint": null, "epoch": 2.9985620955893846, "eval_steps": 500, "global_step": 2997, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0500140664561908, "grad_norm": 1.5154740810394287, "learning_rate": 1.6666666666666667e-06, "loss": 1.6338, "step": 50 }, { "epoch": 0.1000281329123816, "grad_norm": 0.7568167448043823, "learning_rate": 3.3333333333333333e-06, "loss": 1.1959, "step": 100 }, { "epoch": 0.1500421993685724, "grad_norm": 0.6318637728691101, "learning_rate": 5e-06, "loss": 0.8316, "step": 150 }, { "epoch": 0.2000562658247632, "grad_norm": 0.49124225974082947, "learning_rate": 6.666666666666667e-06, "loss": 0.7008, "step": 200 }, { "epoch": 0.25007033228095404, "grad_norm": 0.49193131923675537, "learning_rate": 8.333333333333334e-06, "loss": 0.6234, "step": 250 }, { "epoch": 0.3000843987371448, "grad_norm": 0.5137419700622559, "learning_rate": 1e-05, "loss": 0.5887, "step": 300 }, { "epoch": 0.35009846519333565, "grad_norm": 0.4855007231235504, "learning_rate": 1.1666666666666668e-05, "loss": 0.5523, "step": 350 }, { "epoch": 0.4001125316495264, "grad_norm": 0.4011429250240326, "learning_rate": 1.3333333333333333e-05, "loss": 0.5204, "step": 400 }, { "epoch": 0.45012659810571726, "grad_norm": 0.4894815683364868, "learning_rate": 1.5000000000000002e-05, "loss": 0.4922, "step": 450 }, { "epoch": 0.5001406645619081, "grad_norm": 0.4187089204788208, "learning_rate": 1.6666666666666667e-05, "loss": 0.4676, "step": 500 }, { "epoch": 0.5501547310180989, "grad_norm": 0.37147971987724304, "learning_rate": 1.8333333333333333e-05, "loss": 0.4564, "step": 550 }, { "epoch": 0.6001687974742896, "grad_norm": 0.36486196517944336, "learning_rate": 2e-05, "loss": 0.451, "step": 600 }, { "epoch": 0.6501828639304804, "grad_norm": 0.39784976840019226, "learning_rate": 1.999576008468646e-05, "loss": 0.4313, "step": 650 }, { "epoch": 0.7001969303866713, "grad_norm": 0.4364568293094635, "learning_rate": 1.9983043934122208e-05, "loss": 0.4132, "step": 700 }, { "epoch": 0.7502109968428621, "grad_norm": 0.7991167306900024, "learning_rate": 1.9961862331387545e-05, "loss": 0.4103, "step": 750 }, { "epoch": 0.8002250632990529, "grad_norm": 0.3835698068141937, "learning_rate": 1.9932233238122834e-05, "loss": 0.3919, "step": 800 }, { "epoch": 0.8502391297552436, "grad_norm": 0.40175750851631165, "learning_rate": 1.9894181779297323e-05, "loss": 0.3846, "step": 850 }, { "epoch": 0.9002531962114345, "grad_norm": 0.3462737202644348, "learning_rate": 1.984774022190361e-05, "loss": 0.3712, "step": 900 }, { "epoch": 0.9502672626676253, "grad_norm": 0.31011104583740234, "learning_rate": 1.9792947947595772e-05, "loss": 0.3672, "step": 950 }, { "epoch": 1.0002813291238162, "grad_norm": 0.34971198439598083, "learning_rate": 1.972985141929439e-05, "loss": 0.3635, "step": 1000 }, { "epoch": 1.050295395580007, "grad_norm": 0.3588427007198334, "learning_rate": 1.9658504141786775e-05, "loss": 0.3368, "step": 1050 }, { "epoch": 1.1003094620361977, "grad_norm": 0.32608523964881897, "learning_rate": 1.9578966616355823e-05, "loss": 0.3279, "step": 1100 }, { "epoch": 1.1503235284923885, "grad_norm": 0.3917510509490967, "learning_rate": 1.9491306289475957e-05, "loss": 0.3239, "step": 1150 }, { "epoch": 1.2003375949485793, "grad_norm": 0.3702155649662018, "learning_rate": 1.9395597495619634e-05, "loss": 0.3147, "step": 1200 }, { "epoch": 1.25035166140477, "grad_norm": 0.34320342540740967, "learning_rate": 1.9291921394223e-05, "loss": 0.3178, "step": 1250 }, { "epoch": 1.3003657278609608, "grad_norm": 0.38336554169654846, "learning_rate": 1.918036590086405e-05, "loss": 0.3195, "step": 1300 }, { "epoch": 1.3503797943171518, "grad_norm": 0.4073985517024994, "learning_rate": 1.90610256127117e-05, "loss": 0.308, "step": 1350 }, { "epoch": 1.4003938607733426, "grad_norm": 0.37673142552375793, "learning_rate": 1.8934001728309003e-05, "loss": 0.3096, "step": 1400 }, { "epoch": 1.4504079272295334, "grad_norm": 0.36145591735839844, "learning_rate": 1.8799401961758492e-05, "loss": 0.3121, "step": 1450 }, { "epoch": 1.5004219936857242, "grad_norm": 0.4171595275402069, "learning_rate": 1.865734045138245e-05, "loss": 0.3017, "step": 1500 }, { "epoch": 1.5511550123472226, "grad_norm": 0.32917237281799316, "learning_rate": 1.1144858589642251e-05, "loss": 0.2992, "step": 1550 }, { "epoch": 1.6011690788034134, "grad_norm": 0.3282073736190796, "learning_rate": 1.0564650370835772e-05, "loss": 0.293, "step": 1600 }, { "epoch": 1.6511831452596044, "grad_norm": 0.31431856751441956, "learning_rate": 9.982527302252135e-06, "loss": 0.2969, "step": 1650 }, { "epoch": 1.7011972117157952, "grad_norm": 0.3012774586677551, "learning_rate": 9.40046348731131e-06, "loss": 0.2954, "step": 1700 }, { "epoch": 1.751211278171986, "grad_norm": 0.2984926104545593, "learning_rate": 8.820432828491542e-06, "loss": 0.2885, "step": 1750 }, { "epoch": 1.8012253446281767, "grad_norm": 0.30490660667419434, "learning_rate": 8.244402333405252e-06, "loss": 0.2841, "step": 1800 }, { "epoch": 1.8512394110843675, "grad_norm": 0.3301903009414673, "learning_rate": 7.674325444256899e-06, "loss": 0.2935, "step": 1850 }, { "epoch": 1.9012534775405583, "grad_norm": 0.3045901954174042, "learning_rate": 7.112135413304042e-06, "loss": 0.2855, "step": 1900 }, { "epoch": 1.951267543996749, "grad_norm": 0.310871958732605, "learning_rate": 6.55973874678682e-06, "loss": 0.2831, "step": 1950 }, { "epoch": 2.00128161045294, "grad_norm": 0.3118513822555542, "learning_rate": 6.0190087395588596e-06, "loss": 0.2823, "step": 2000 }, { "epoch": 2.0512956769091306, "grad_norm": 0.2910955250263214, "learning_rate": 5.491779122345093e-06, "loss": 0.2576, "step": 2050 }, { "epoch": 2.1013097433653214, "grad_norm": 0.3105012774467468, "learning_rate": 4.979837843169959e-06, "loss": 0.2522, "step": 2100 }, { "epoch": 2.151323809821512, "grad_norm": 0.31710630655288696, "learning_rate": 4.484921004044509e-06, "loss": 0.2546, "step": 2150 }, { "epoch": 2.201337876277703, "grad_norm": 0.41157081723213196, "learning_rate": 4.008706973474391e-06, "loss": 0.2531, "step": 2200 }, { "epoch": 2.2513519427338937, "grad_norm": 0.3166573941707611, "learning_rate": 3.5528106947544626e-06, "loss": 0.2519, "step": 2250 }, { "epoch": 2.301366009190085, "grad_norm": 0.3242489695549011, "learning_rate": 3.118778209351808e-06, "loss": 0.254, "step": 2300 }, { "epoch": 2.3513800756462757, "grad_norm": 0.34881240129470825, "learning_rate": 2.7080814139495402e-06, "loss": 0.2526, "step": 2350 }, { "epoch": 2.4013941421024665, "grad_norm": 0.3267311751842499, "learning_rate": 2.322113068931391e-06, "loss": 0.251, "step": 2400 }, { "epoch": 2.4514082085586573, "grad_norm": 0.3213905394077301, "learning_rate": 1.9621820752343324e-06, "loss": 0.246, "step": 2450 }, { "epoch": 2.501422275014848, "grad_norm": 0.48464563488960266, "learning_rate": 1.629509035586484e-06, "loss": 0.2527, "step": 2500 }, { "epoch": 2.551436341471039, "grad_norm": 0.31706663966178894, "learning_rate": 1.3252221151830513e-06, "loss": 0.2465, "step": 2550 }, { "epoch": 2.6014504079272296, "grad_norm": 0.33972597122192383, "learning_rate": 1.0503532158376584e-06, "loss": 0.2483, "step": 2600 }, { "epoch": 2.6514644743834204, "grad_norm": 0.2904072105884552, "learning_rate": 8.058344765833171e-07, "loss": 0.2465, "step": 2650 }, { "epoch": 2.701478540839611, "grad_norm": 0.32852381467819214, "learning_rate": 5.924951125902545e-07, "loss": 0.2512, "step": 2700 }, { "epoch": 2.751492607295802, "grad_norm": 0.2948131263256073, "learning_rate": 4.11058603120511e-07, "loss": 0.2483, "step": 2750 }, { "epoch": 2.8015066737519927, "grad_norm": 0.33766064047813416, "learning_rate": 2.6214023805552826e-07, "loss": 0.2496, "step": 2800 }, { "epoch": 2.8515207402081835, "grad_norm": 0.29850533604621887, "learning_rate": 1.462450313169983e-07, "loss": 0.2444, "step": 2850 }, { "epoch": 2.9015348066643742, "grad_norm": 0.33009734749794006, "learning_rate": 6.376600825699463e-08, "loss": 0.2478, "step": 2900 }, { "epoch": 2.951548873120565, "grad_norm": 0.5311033725738525, "learning_rate": 1.49828728252277e-08, "loss": 0.2412, "step": 2950 }, { "epoch": 2.9985620955893846, "step": 2997, "total_flos": 2.5672199949139968e+17, "train_loss": 0.13155911801694273, "train_runtime": 22482.9111, "train_samples_per_second": 8.537, "train_steps_per_second": 0.133 } ], "logging_steps": 50, "max_steps": 2997, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 100, "total_flos": 2.5672199949139968e+17, "train_batch_size": 2, "trial_name": null, "trial_params": null }