{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 12, "global_step": 47, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.02127659574468085, "grad_norm": 0.42955278162899063, "learning_rate": 1.0000000000000002e-06, "loss": 1.1272, "step": 1 }, { "epoch": 0.02127659574468085, "eval_loss": 1.130263328552246, "eval_runtime": 46.1353, "eval_samples_per_second": 2.146, "eval_steps_per_second": 0.217, "step": 1 }, { "epoch": 0.0425531914893617, "grad_norm": 0.35905455037073136, "learning_rate": 2.0000000000000003e-06, "loss": 1.2237, "step": 2 }, { "epoch": 0.06382978723404255, "grad_norm": 0.3859662975436329, "learning_rate": 3e-06, "loss": 1.2109, "step": 3 }, { "epoch": 0.0851063829787234, "grad_norm": 0.3458897777003975, "learning_rate": 4.000000000000001e-06, "loss": 1.3168, "step": 4 }, { "epoch": 0.10638297872340426, "grad_norm": 0.297024954598104, "learning_rate": 5e-06, "loss": 1.1879, "step": 5 }, { "epoch": 0.1276595744680851, "grad_norm": 0.2974375637230888, "learning_rate": 6e-06, "loss": 1.1898, "step": 6 }, { "epoch": 0.14893617021276595, "grad_norm": 0.35483433691691424, "learning_rate": 7e-06, "loss": 1.2275, "step": 7 }, { "epoch": 0.1702127659574468, "grad_norm": 0.3817564026852017, "learning_rate": 8.000000000000001e-06, "loss": 1.2163, "step": 8 }, { "epoch": 0.19148936170212766, "grad_norm": 0.40183241982975176, "learning_rate": 9e-06, "loss": 1.0668, "step": 9 }, { "epoch": 0.2127659574468085, "grad_norm": 0.3748400230734179, "learning_rate": 1e-05, "loss": 1.2589, "step": 10 }, { "epoch": 0.23404255319148937, "grad_norm": 0.4732373123638865, "learning_rate": 9.981987442712634e-06, "loss": 1.209, "step": 11 }, { "epoch": 0.2553191489361702, "grad_norm": 0.34097014836642403, "learning_rate": 9.928079551738542e-06, "loss": 1.1222, "step": 12 }, { "epoch": 0.2553191489361702, "eval_loss": 1.1231276988983154, "eval_runtime": 45.5524, "eval_samples_per_second": 2.173, "eval_steps_per_second": 0.22, "step": 12 }, { "epoch": 0.2765957446808511, "grad_norm": 0.3865545405035091, "learning_rate": 9.838664734667496e-06, "loss": 1.2509, "step": 13 }, { "epoch": 0.2978723404255319, "grad_norm": 0.5747305954760134, "learning_rate": 9.714387227305422e-06, "loss": 1.1448, "step": 14 }, { "epoch": 0.3191489361702128, "grad_norm": 0.43342726982955576, "learning_rate": 9.55614245194068e-06, "loss": 1.1433, "step": 15 }, { "epoch": 0.3404255319148936, "grad_norm": 0.38621360188314374, "learning_rate": 9.365070565805941e-06, "loss": 1.2438, "step": 16 }, { "epoch": 0.3617021276595745, "grad_norm": 0.4586858888503175, "learning_rate": 9.142548246219212e-06, "loss": 1.217, "step": 17 }, { "epoch": 0.3829787234042553, "grad_norm": 0.4605021995799594, "learning_rate": 8.890178771592198e-06, "loss": 1.1023, "step": 18 }, { "epoch": 0.40425531914893614, "grad_norm": 0.4149444720869162, "learning_rate": 8.609780469772623e-06, "loss": 1.1057, "step": 19 }, { "epoch": 0.425531914893617, "grad_norm": 0.4776677990058847, "learning_rate": 8.303373616950408e-06, "loss": 1.1878, "step": 20 }, { "epoch": 0.44680851063829785, "grad_norm": 0.6083128734022943, "learning_rate": 7.973165881521435e-06, "loss": 1.0719, "step": 21 }, { "epoch": 0.46808510638297873, "grad_norm": 0.427407991852223, "learning_rate": 7.621536417786159e-06, "loss": 1.1414, "step": 22 }, { "epoch": 0.48936170212765956, "grad_norm": 0.41689291665199874, "learning_rate": 7.251018724088367e-06, "loss": 1.1651, "step": 23 }, { "epoch": 0.5106382978723404, "grad_norm": 0.5049216829996077, "learning_rate": 6.864282388901544e-06, "loss": 1.1099, "step": 24 }, { "epoch": 0.5106382978723404, "eval_loss": 1.0642483234405518, "eval_runtime": 45.5272, "eval_samples_per_second": 2.175, "eval_steps_per_second": 0.22, "step": 24 }, { "epoch": 0.5319148936170213, "grad_norm": 0.3940106585784572, "learning_rate": 6.464113856382752e-06, "loss": 0.9954, "step": 25 }, { "epoch": 0.5531914893617021, "grad_norm": 0.35275642793537804, "learning_rate": 6.053396349978632e-06, "loss": 1.0469, "step": 26 }, { "epoch": 0.574468085106383, "grad_norm": 0.5184071159973599, "learning_rate": 5.635089098734394e-06, "loss": 1.1481, "step": 27 }, { "epoch": 0.5957446808510638, "grad_norm": 0.4370219832674898, "learning_rate": 5.212206015980742e-06, "loss": 1.1364, "step": 28 }, { "epoch": 0.6170212765957447, "grad_norm": 0.36478900402279923, "learning_rate": 4.78779398401926e-06, "loss": 1.0762, "step": 29 }, { "epoch": 0.6382978723404256, "grad_norm": 0.40328712656577087, "learning_rate": 4.364910901265607e-06, "loss": 1.085, "step": 30 }, { "epoch": 0.6595744680851063, "grad_norm": 0.341826695457334, "learning_rate": 3.94660365002137e-06, "loss": 1.0013, "step": 31 }, { "epoch": 0.6808510638297872, "grad_norm": 0.4177223934587119, "learning_rate": 3.5358861436172487e-06, "loss": 1.0372, "step": 32 }, { "epoch": 0.7021276595744681, "grad_norm": 0.4218122745580511, "learning_rate": 3.1357176110984578e-06, "loss": 0.9904, "step": 33 }, { "epoch": 0.723404255319149, "grad_norm": 0.3778418743271345, "learning_rate": 2.748981275911633e-06, "loss": 1.0609, "step": 34 }, { "epoch": 0.7446808510638298, "grad_norm": 0.5024526255778048, "learning_rate": 2.3784635822138424e-06, "loss": 1.0193, "step": 35 }, { "epoch": 0.7659574468085106, "grad_norm": 0.29310915204897864, "learning_rate": 2.0268341184785674e-06, "loss": 1.105, "step": 36 }, { "epoch": 0.7659574468085106, "eval_loss": 1.0241957902908325, "eval_runtime": 45.4699, "eval_samples_per_second": 2.177, "eval_steps_per_second": 0.22, "step": 36 }, { "epoch": 0.7872340425531915, "grad_norm": 0.38019500098636755, "learning_rate": 1.6966263830495939e-06, "loss": 1.122, "step": 37 }, { "epoch": 0.8085106382978723, "grad_norm": 0.4515729624152646, "learning_rate": 1.390219530227378e-06, "loss": 1.1268, "step": 38 }, { "epoch": 0.8297872340425532, "grad_norm": 0.3174904808093991, "learning_rate": 1.1098212284078037e-06, "loss": 1.1366, "step": 39 }, { "epoch": 0.851063829787234, "grad_norm": 0.32824416887732244, "learning_rate": 8.574517537807897e-07, "loss": 1.085, "step": 40 }, { "epoch": 0.8723404255319149, "grad_norm": 0.3723789314131866, "learning_rate": 6.349294341940593e-07, "loss": 1.0074, "step": 41 }, { "epoch": 0.8936170212765957, "grad_norm": 0.3863665218570595, "learning_rate": 4.43857548059321e-07, "loss": 1.0786, "step": 42 }, { "epoch": 0.9148936170212766, "grad_norm": 0.28822889001191754, "learning_rate": 2.85612772694579e-07, "loss": 1.0286, "step": 43 }, { "epoch": 0.9361702127659575, "grad_norm": 0.4310630390723365, "learning_rate": 1.6133526533250566e-07, "loss": 0.8951, "step": 44 }, { "epoch": 0.9574468085106383, "grad_norm": 0.3151119202195982, "learning_rate": 7.192044826145772e-08, "loss": 0.9489, "step": 45 }, { "epoch": 0.9787234042553191, "grad_norm": 0.3577888537210098, "learning_rate": 1.8012557287367394e-08, "loss": 1.1146, "step": 46 }, { "epoch": 1.0, "grad_norm": 0.383122905902664, "learning_rate": 0.0, "loss": 1.031, "step": 47 } ], "logging_steps": 1, "max_steps": 47, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 500, "total_flos": 286172412641280.0, "train_batch_size": 2, "trial_name": null, "trial_params": null }