|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 6.0, |
|
"global_step": 35574, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 4.9297239556979816e-05, |
|
"loss": 3.4444, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 4.8594479113959636e-05, |
|
"loss": 3.339, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 4.789171867093945e-05, |
|
"loss": 3.3261, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 4.718895822791927e-05, |
|
"loss": 3.2812, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 4.648619778489909e-05, |
|
"loss": 3.25, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 4.5783437341878904e-05, |
|
"loss": 3.2331, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 4.508067689885872e-05, |
|
"loss": 3.2195, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 4.437791645583854e-05, |
|
"loss": 3.2063, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 4.367515601281835e-05, |
|
"loss": 3.2028, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 4.297239556979817e-05, |
|
"loss": 3.1897, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 4.2269635126777984e-05, |
|
"loss": 3.1792, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 4.15668746837578e-05, |
|
"loss": 3.1564, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 1.1, |
|
"learning_rate": 4.0864114240737625e-05, |
|
"loss": 3.0599, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"learning_rate": 4.016135379771744e-05, |
|
"loss": 3.0641, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"learning_rate": 3.945859335469725e-05, |
|
"loss": 3.0661, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 3.875583291167707e-05, |
|
"loss": 3.0731, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 1.43, |
|
"learning_rate": 3.8053072468656886e-05, |
|
"loss": 3.0555, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 1.52, |
|
"learning_rate": 3.73503120256367e-05, |
|
"loss": 3.0592, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"learning_rate": 3.664755158261652e-05, |
|
"loss": 3.0618, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 3.594479113959633e-05, |
|
"loss": 3.0592, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"learning_rate": 3.524203069657615e-05, |
|
"loss": 3.0496, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"learning_rate": 3.453927025355597e-05, |
|
"loss": 3.0396, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"learning_rate": 3.383650981053579e-05, |
|
"loss": 3.0479, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 3.313374936751561e-05, |
|
"loss": 3.0178, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 2.11, |
|
"learning_rate": 3.243098892449542e-05, |
|
"loss": 2.9502, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 2.19, |
|
"learning_rate": 3.1728228481475234e-05, |
|
"loss": 2.9585, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 2.28, |
|
"learning_rate": 3.1025468038455054e-05, |
|
"loss": 2.968, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 3.0322707595434867e-05, |
|
"loss": 2.9591, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 2.45, |
|
"learning_rate": 2.9619947152414684e-05, |
|
"loss": 2.9585, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"learning_rate": 2.89171867093945e-05, |
|
"loss": 2.9628, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 2.61, |
|
"learning_rate": 2.821442626637432e-05, |
|
"loss": 2.9712, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 2.7511665823354138e-05, |
|
"loss": 2.9583, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 2.78, |
|
"learning_rate": 2.6808905380333955e-05, |
|
"loss": 2.9491, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 2.87, |
|
"learning_rate": 2.610614493731377e-05, |
|
"loss": 2.9594, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"learning_rate": 2.5403384494293585e-05, |
|
"loss": 2.9673, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 3.04, |
|
"learning_rate": 2.4700624051273402e-05, |
|
"loss": 2.9344, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 3.12, |
|
"learning_rate": 2.3997863608253223e-05, |
|
"loss": 2.8771, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 3.2, |
|
"learning_rate": 2.3295103165233036e-05, |
|
"loss": 2.8863, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 3.29, |
|
"learning_rate": 2.2592342722212853e-05, |
|
"loss": 2.8845, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 3.37, |
|
"learning_rate": 2.188958227919267e-05, |
|
"loss": 2.9015, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 3.46, |
|
"learning_rate": 2.1186821836172487e-05, |
|
"loss": 2.8927, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 3.54, |
|
"learning_rate": 2.0484061393152303e-05, |
|
"loss": 2.889, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 3.63, |
|
"learning_rate": 1.978130095013212e-05, |
|
"loss": 2.9037, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 3.71, |
|
"learning_rate": 1.9078540507111937e-05, |
|
"loss": 2.8995, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 3.79, |
|
"learning_rate": 1.837578006409175e-05, |
|
"loss": 2.8922, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 3.88, |
|
"learning_rate": 1.767301962107157e-05, |
|
"loss": 2.8954, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 3.96, |
|
"learning_rate": 1.6970259178051388e-05, |
|
"loss": 2.9017, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 4.05, |
|
"learning_rate": 1.6267498735031205e-05, |
|
"loss": 2.8701, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 4.13, |
|
"learning_rate": 1.5564738292011018e-05, |
|
"loss": 2.8549, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 4.22, |
|
"learning_rate": 1.4861977848990838e-05, |
|
"loss": 2.8502, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 4.3, |
|
"learning_rate": 1.4159217405970653e-05, |
|
"loss": 2.846, |
|
"step": 25500 |
|
}, |
|
{ |
|
"epoch": 4.39, |
|
"learning_rate": 1.345645696295047e-05, |
|
"loss": 2.8497, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 4.47, |
|
"learning_rate": 1.2753696519930285e-05, |
|
"loss": 2.8284, |
|
"step": 26500 |
|
}, |
|
{ |
|
"epoch": 4.55, |
|
"learning_rate": 1.2050936076910104e-05, |
|
"loss": 2.8432, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 4.64, |
|
"learning_rate": 1.1348175633889919e-05, |
|
"loss": 2.8468, |
|
"step": 27500 |
|
}, |
|
{ |
|
"epoch": 4.72, |
|
"learning_rate": 1.0645415190869738e-05, |
|
"loss": 2.8425, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 4.81, |
|
"learning_rate": 9.942654747849553e-06, |
|
"loss": 2.8403, |
|
"step": 28500 |
|
}, |
|
{ |
|
"epoch": 4.89, |
|
"learning_rate": 9.23989430482937e-06, |
|
"loss": 2.8276, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 4.98, |
|
"learning_rate": 8.537133861809187e-06, |
|
"loss": 2.8549, |
|
"step": 29500 |
|
}, |
|
{ |
|
"epoch": 5.06, |
|
"learning_rate": 7.834373418789003e-06, |
|
"loss": 2.8061, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 5.14, |
|
"learning_rate": 7.13161297576882e-06, |
|
"loss": 2.8111, |
|
"step": 30500 |
|
}, |
|
{ |
|
"epoch": 5.23, |
|
"learning_rate": 6.428852532748637e-06, |
|
"loss": 2.8153, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 5.31, |
|
"learning_rate": 5.726092089728453e-06, |
|
"loss": 2.8129, |
|
"step": 31500 |
|
}, |
|
{ |
|
"epoch": 5.4, |
|
"learning_rate": 5.02333164670827e-06, |
|
"loss": 2.812, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 5.48, |
|
"learning_rate": 4.320571203688087e-06, |
|
"loss": 2.8134, |
|
"step": 32500 |
|
}, |
|
{ |
|
"epoch": 5.57, |
|
"learning_rate": 3.6178107606679037e-06, |
|
"loss": 2.8172, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 5.65, |
|
"learning_rate": 2.9150503176477205e-06, |
|
"loss": 2.809, |
|
"step": 33500 |
|
}, |
|
{ |
|
"epoch": 5.73, |
|
"learning_rate": 2.212289874627537e-06, |
|
"loss": 2.8115, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 5.82, |
|
"learning_rate": 1.5095294316073538e-06, |
|
"loss": 2.81, |
|
"step": 34500 |
|
}, |
|
{ |
|
"epoch": 5.9, |
|
"learning_rate": 8.067689885871704e-07, |
|
"loss": 2.8087, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 5.99, |
|
"learning_rate": 1.0400854556698713e-07, |
|
"loss": 2.799, |
|
"step": 35500 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"step": 35574, |
|
"total_flos": 1.8590405492736e+16, |
|
"train_loss": 2.969814732685026, |
|
"train_runtime": 9127.501, |
|
"train_samples_per_second": 3.897, |
|
"train_steps_per_second": 3.897 |
|
} |
|
], |
|
"max_steps": 35574, |
|
"num_train_epochs": 6, |
|
"total_flos": 1.8590405492736e+16, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|