{ "best_metric": 1.5939216613769531, "best_model_checkpoint": "opus-mt-zh-en-finetuned-zhen-checkpoints/checkpoint-25985", "epoch": 1.0, "eval_steps": 500, "global_step": 25985, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.02, "grad_norm": 2.777989149093628, "learning_rate": 1.9961593226861654e-05, "loss": 0.608, "step": 500 }, { "epoch": 0.04, "grad_norm": 2.571350336074829, "learning_rate": 1.9923109486242066e-05, "loss": 0.6031, "step": 1000 }, { "epoch": 0.06, "grad_norm": 6.669718265533447, "learning_rate": 1.9884625745622478e-05, "loss": 0.5823, "step": 1500 }, { "epoch": 0.08, "grad_norm": 3.766803503036499, "learning_rate": 1.984614200500289e-05, "loss": 0.625, "step": 2000 }, { "epoch": 0.1, "grad_norm": 2.617687225341797, "learning_rate": 1.9807735231864538e-05, "loss": 0.6324, "step": 2500 }, { "epoch": 0.12, "grad_norm": 4.26535177230835, "learning_rate": 1.976932845872619e-05, "loss": 0.6184, "step": 3000 }, { "epoch": 0.13, "grad_norm": 5.035365581512451, "learning_rate": 1.9730844718106602e-05, "loss": 0.6277, "step": 3500 }, { "epoch": 0.15, "grad_norm": 6.626140594482422, "learning_rate": 1.9692360977487014e-05, "loss": 0.6557, "step": 4000 }, { "epoch": 0.17, "grad_norm": 3.718008279800415, "learning_rate": 1.9653877236867426e-05, "loss": 0.6623, "step": 4500 }, { "epoch": 0.19, "grad_norm": 4.26861047744751, "learning_rate": 1.9615470463729075e-05, "loss": 0.6719, "step": 5000 }, { "epoch": 0.21, "grad_norm": 5.266055107116699, "learning_rate": 1.9576986723109487e-05, "loss": 0.636, "step": 5500 }, { "epoch": 0.23, "grad_norm": 6.086512565612793, "learning_rate": 1.95385029824899e-05, "loss": 0.6585, "step": 6000 }, { "epoch": 0.25, "grad_norm": 4.0940327644348145, "learning_rate": 1.950001924187031e-05, "loss": 0.6341, "step": 6500 }, { "epoch": 0.27, "grad_norm": 7.52503776550293, "learning_rate": 1.9461535501250722e-05, "loss": 0.6589, "step": 7000 }, { "epoch": 0.29, "grad_norm": 3.5400032997131348, "learning_rate": 1.9423051760631134e-05, "loss": 0.6627, "step": 7500 }, { "epoch": 0.31, "grad_norm": 3.525592565536499, "learning_rate": 1.9384568020011546e-05, "loss": 0.639, "step": 8000 }, { "epoch": 0.33, "grad_norm": 3.1940386295318604, "learning_rate": 1.9346084279391958e-05, "loss": 0.6433, "step": 8500 }, { "epoch": 0.35, "grad_norm": 5.362303733825684, "learning_rate": 1.930767750625361e-05, "loss": 0.6714, "step": 9000 }, { "epoch": 0.37, "grad_norm": 4.885318279266357, "learning_rate": 1.9269193765634022e-05, "loss": 0.6334, "step": 9500 }, { "epoch": 0.38, "grad_norm": 7.104074001312256, "learning_rate": 1.9230710025014434e-05, "loss": 0.6496, "step": 10000 }, { "epoch": 0.4, "grad_norm": 5.086849689483643, "learning_rate": 1.9192226284394843e-05, "loss": 0.6647, "step": 10500 }, { "epoch": 0.42, "grad_norm": 10.493572235107422, "learning_rate": 1.9153819511256495e-05, "loss": 0.6609, "step": 11000 }, { "epoch": 0.44, "grad_norm": 10.206048011779785, "learning_rate": 1.9115335770636907e-05, "loss": 0.6806, "step": 11500 }, { "epoch": 0.46, "grad_norm": 9.489506721496582, "learning_rate": 1.907685203001732e-05, "loss": 0.641, "step": 12000 }, { "epoch": 0.48, "grad_norm": 4.525002479553223, "learning_rate": 1.903844525687897e-05, "loss": 0.6863, "step": 12500 }, { "epoch": 0.5, "grad_norm": 7.507730007171631, "learning_rate": 1.9000038483740623e-05, "loss": 0.6829, "step": 13000 }, { "epoch": 0.52, "grad_norm": 4.827548980712891, "learning_rate": 1.8961554743121035e-05, "loss": 0.6597, "step": 13500 }, { "epoch": 0.54, "grad_norm": 6.409604072570801, "learning_rate": 1.8923071002501447e-05, "loss": 0.6445, "step": 14000 }, { "epoch": 0.56, "grad_norm": 9.57291316986084, "learning_rate": 1.8884664229363095e-05, "loss": 0.6882, "step": 14500 }, { "epoch": 0.58, "grad_norm": 9.129105567932129, "learning_rate": 1.8846180488743507e-05, "loss": 0.6719, "step": 15000 }, { "epoch": 0.6, "grad_norm": 8.071757316589355, "learning_rate": 1.880769674812392e-05, "loss": 0.6668, "step": 15500 }, { "epoch": 0.62, "grad_norm": 3.636784791946411, "learning_rate": 1.876921300750433e-05, "loss": 0.6773, "step": 16000 }, { "epoch": 0.63, "grad_norm": 6.6433515548706055, "learning_rate": 1.8730729266884743e-05, "loss": 0.6776, "step": 16500 }, { "epoch": 0.65, "grad_norm": 3.0176546573638916, "learning_rate": 1.8692245526265155e-05, "loss": 0.6587, "step": 17000 }, { "epoch": 0.67, "grad_norm": 7.913615703582764, "learning_rate": 1.8653761785645567e-05, "loss": 0.6665, "step": 17500 }, { "epoch": 0.69, "grad_norm": 5.335203647613525, "learning_rate": 1.861527804502598e-05, "loss": 0.6684, "step": 18000 }, { "epoch": 0.71, "grad_norm": 5.51749324798584, "learning_rate": 1.857679430440639e-05, "loss": 0.6577, "step": 18500 }, { "epoch": 0.73, "grad_norm": 4.254025936126709, "learning_rate": 1.8538310563786802e-05, "loss": 0.6776, "step": 19000 }, { "epoch": 0.75, "grad_norm": 3.3869874477386475, "learning_rate": 1.8499826823167214e-05, "loss": 0.6351, "step": 19500 }, { "epoch": 0.77, "grad_norm": 12.704727172851562, "learning_rate": 1.8461343082547626e-05, "loss": 0.6904, "step": 20000 }, { "epoch": 0.79, "grad_norm": 3.478996515274048, "learning_rate": 1.8422859341928038e-05, "loss": 0.6829, "step": 20500 }, { "epoch": 0.81, "grad_norm": 6.513260841369629, "learning_rate": 1.838437560130845e-05, "loss": 0.6615, "step": 21000 }, { "epoch": 0.83, "grad_norm": 5.242546081542969, "learning_rate": 1.8345891860688862e-05, "loss": 0.6859, "step": 21500 }, { "epoch": 0.85, "grad_norm": 6.638239860534668, "learning_rate": 1.8307408120069274e-05, "loss": 0.6485, "step": 22000 }, { "epoch": 0.87, "grad_norm": 8.32187557220459, "learning_rate": 1.8269001346930923e-05, "loss": 0.6567, "step": 22500 }, { "epoch": 0.89, "grad_norm": 3.9929823875427246, "learning_rate": 1.8230517606311334e-05, "loss": 0.6996, "step": 23000 }, { "epoch": 0.9, "grad_norm": 5.426717281341553, "learning_rate": 1.8192110833172987e-05, "loss": 0.6529, "step": 23500 }, { "epoch": 0.92, "grad_norm": 7.328232288360596, "learning_rate": 1.81536270925534e-05, "loss": 0.6711, "step": 24000 }, { "epoch": 0.94, "grad_norm": 6.648090362548828, "learning_rate": 1.811514335193381e-05, "loss": 0.7026, "step": 24500 }, { "epoch": 0.96, "grad_norm": 4.047011375427246, "learning_rate": 1.8076659611314222e-05, "loss": 0.6867, "step": 25000 }, { "epoch": 0.98, "grad_norm": 10.354930877685547, "learning_rate": 1.803825283817587e-05, "loss": 0.6718, "step": 25500 }, { "epoch": 1.0, "eval_loss": 1.5939216613769531, "eval_runtime": 11.3331, "eval_samples_per_second": 351.271, "eval_steps_per_second": 87.884, "step": 25985 } ], "logging_steps": 500, "max_steps": 259850, "num_input_tokens_seen": 0, "num_train_epochs": 10, "save_steps": 500, "total_flos": 1029890059665408.0, "train_batch_size": 4, "trial_name": null, "trial_params": null }